Downloading winning-args-corpus to /home/XXXXXXd/.convokit/downloads/winning-args-corpus
Downloading winning-args-corpus from http://zissou.infosci.cornell.edu/convokit/datasets/winning-args-corpus/winning-args-corpus.zip (73.7MB)... Done
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cnhre1n has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cnhs1jf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cn7mmnt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cn66mck has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmt6w97 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmsgxzm has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmsitjr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmqqd49 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmqsp80 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmr57l3 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cmihu76 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cm85qp1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cm6tktt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_clz6jxt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cm09icp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cly0oho has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cly8wzq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_clxpe30 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_clv6oas has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cls8zr2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_clj3tcp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_clhtmr0 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cler134 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_claiask has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl96qxj has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl92hgd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl92jdc has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl7lil7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_claaeq7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl6xkrl has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl6yu6h has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl6ywwj has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl6yxf2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl6u197 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl65sus has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl67wux has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl5n0e5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl55jtf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl53psw has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl54hde has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl5490t has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl38tmp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cl2rq86 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ckujlyq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ckujdib has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cklsm40 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ckmghf2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cknmezc has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ckj2g0i has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ck8j9e4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ck8cwb6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjxlz5l has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjntq7s has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjnp5hy has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjiqn1e has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjh7hr0 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjh6yab has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cjcd62j has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cj8l1kv has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cj8zjae has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cj7yiy4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cj7o50t has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cj50shk has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cizh6xb has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cj0ukb7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ciz9aka has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cize32v has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_civr4mn has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cip166t has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cil3zav has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cijy9lk has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cijcups has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cifu4zp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cid30or has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ci8wo4e has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ci9fbtw has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ci2pixf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ci2pcku has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ci2cc99 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chz0gqi has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chw609b has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chv6j3s has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chv26as has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_churmo0 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chswmlq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chsxa75 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chnlrqw has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chl8bsy has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chiw401 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chihwy1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chh0sby has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chbrqn7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ch7rh7p has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ch6watf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ch6ssui has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ch61v2p has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ch3ng8c has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgrm70r has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgrz99m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgsegx6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgnj6j9 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ch073vt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgkn5ae has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgiihu5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgieyuu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgig23g has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgie34t has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgh4ifl has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cggr63w has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cggd3v7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cggutk5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgeiql6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgej6dy has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cge3th4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cgedp3m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cg8kgbj has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfwc45l has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfv3enu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfpx424 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfl78p1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfilsz7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfj8wvg has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cffznn5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cffo8l7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfdq2z7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfc9c4b has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfcrvtf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfbxgor has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfb4q61 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cfb5l7x has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cf99z8w has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cf2h04o has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ceptv5v has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ceo2hjj has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ceorou4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cenyp1x has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cemye2y has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_celsimn has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ceehosv has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cebrznc has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ceconce has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cebnxzk has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cebd6e3 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cec3nwt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_chv7nyk has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cedktpa has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ce9dob3 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ce5q8tj has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ce4gt8u has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ce0p904 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cdosehh has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cdiarjx has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cda83lr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cd759lq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cd4pzb7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cd4p7s1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cd4zoy7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cdn4qk9 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccue3mg has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccphf6t has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccphsld has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccmc12f has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccjdtdf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cciu2c8 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccizwjs has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccivs3y has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccj0iqr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cciupg2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cch8ibq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccgulma has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccgulwv has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccgnj8i has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cch2tad has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cch9i4l has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cch9wwe has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccgvmn6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cce35j0 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ccdamor has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cce8w7z has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cce42qt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cc9x38t has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cc302nb has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbyfuze has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbvq8lh has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbrec2x has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbrmg6l has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbq8ij7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbp1a4b has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cboztg2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbou2b5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbpau3m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbokj1o has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cboz1er has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbmr775 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cblsags has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cbqpqwi has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb9lo66 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb9lrsq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb9zx65 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb8beut has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb7zk0m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb6c4sq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb33ath has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb1kfhi has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb0ahbi has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cb0bzr4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cawc0b9 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cawjc4y has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caxt2vf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caukmyp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caurnnc has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cavlgpq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caulblj has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cauridr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_casyucd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_catl342 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_camzwda has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_camssul has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_can6u3l has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_calik3u has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caljueg has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cajb0py has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cajbvqf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caajlk6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_caachmu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cabc5l7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9xkwd1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9za8qk has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9x7xos has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9zr42i has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9tlmhq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9r6a05 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ca0tpgp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9r7py2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9rnobm has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9rp0so has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9qubp4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9qwaar has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9pfs2o has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9lc68q has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9jlurv has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9hall9 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9gkb83 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c9bvjlq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c98s9ip has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c98o57y has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c99ahf3 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c990rx6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c97eoi9 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c97acob has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c97ah5c has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c95q7s3 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c95kdch has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_c95l9ml has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqzqi7q has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqzmedm has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cr13cnd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqzr1lp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqyom18 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqybmxe has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqy7ahu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqybfct has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqyuen2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqy83y1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqxoxi5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqn9nyq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqnhdk2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqn3mch has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqn7pac has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqmw3di has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqf8ryl has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqf8mpf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqfb5xa has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqfhjkq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqgarpy has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqdccrd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqctlqr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqcthds has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqkelp2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqd8j90 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqd8a8j has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cqcrdrw has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cq8xz69 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cq97x51 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cq8zb08 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cq959on has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cq9ci0g has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cq97u6v has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cptu8ww has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpstv77 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpsxhki has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpswzfn has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpuguvq has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpvfq3o has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cptriew has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpsxudr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpszfeh has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpqft77 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpqfs0u has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cprk62o has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpnulph has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpnn77o has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpnju02 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpn5p75 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpni19v has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpo4x6m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpbijg5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cpbfv95 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coz6meu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coz24er has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coz3omg has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_copv0tz has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cor1wk2 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coq0wpg has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coqe7nl has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cokywqo has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cokwcae has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cokx8ik has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cokyzdf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_col2sjk has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cokxscn has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cojw16m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coezsna has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cof1lb9 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cohpfyt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cof4o1h has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coeyzd4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coai23q has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coa8l6z has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coa80dz has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_coa8l84 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co68zqm has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co6a4yb has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co6ef8m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co6ba44 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co6rpqr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co69x71 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_co6a13x has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cnq3mwr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cnq4ycs has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cnqs4bp has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cunpy0j has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cuior6e has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cufvp2u has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cu7mdqo has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cu7xcz4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cu5cv9a has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cu3h1wa has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cu272pa has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cu0qopb has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctzro35 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctxzpb4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctwh70b has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctqswb6 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctqjqcd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctoncx7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctq5xed has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctjjz2v has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cti68mr has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctiqywc has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cthu9hx has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctdqpne has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctdkusf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctdhx1w has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctcfj2v has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct5ul8m has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct5m7v4 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct4klhi has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct3sgfd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct5bruu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct1rmby has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cstl2de has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cstboz7 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cstbdw5 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ct7dtjh has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_csk5kbs has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctam9yz has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_csh3dqw has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_csislrs has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_csed3sw has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctalrap has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_csernqu has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_csdn1ui has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cs8mmtf has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cs4re28 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_ctb2nf1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cs1tl9q has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_crul067 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_crusrpd has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_crjkijn has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cre72f1 has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_creewvt has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_creh7rb has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_crctrfy has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_crbkihy has been casted to a string.
[91mWARNING: [0mUtterance text must be a string: text of utterance with ID: t1_cra0ei4 has been casted to a string.
Number of threads in dataset: 293297
--------------EPOCH 1-------------
Test Accuracy: tensor(0.1622, device='cuda:0')
Loss: tensor(0.5829, device='cuda:0', grad_fn=<NllLossBackward>)
Saving model at iteration: 0
Loss: tensor(0.9501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.9170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.2785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3964, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5834, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] maximizing security &amp; happiness for yourself and others would be the obvious choice. [ENDQ] [NEWLINE] That is one big point. With being allowed to decided for yourself the means in which you can make yourself safe and happy are on you,<mask> you can make that happen<mask> you wish<mask><mask><mask> you don't interfere with others doing the same. This is<mask> the states choice of<mask> makes you safe, may not fit<mask> would be needed to make you safe,<mask> the state may choose your not worth making safe. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask>, natural evolution is slow and dumb. [ENDQ] [NEWLINE] Well lets look at states and<mask> fast they evolved under traditional monarchy's. States<mask> there was one major power that dictated everything efficiently... any have rapid advances? not<mask> much,<mask>?<mask> things are good for those in charge the way they are. No need to change, and change likely means more risks to power coming out to present threats.( max weber writes about this in detail<mask><mask>..) Under more free society's, people change and evolve more rapidly,<mask> nothing is necessarily sacred.<mask> massive booms to trade, to technology, and overall happiness increases<mask> a society is free to evolve naturally. [NEWLINE] [NEWLINE] <mask> Rome was a republic is never fell,<mask> you remember it fell<mask> an empire. That's<mask> republics are strong and able to adapt for more long term success, more<mask> than short term.<mask> Republics usually build wealth and incorporate more into their state and nation,<mask><mask> Empire is usually expending wealth to "keep the huns away from the gates"<mask> you will. Empires<mask> usually rely on creating an "Other" to target<mask> a threat to keep the populations mobilized with its rule. Republics usually are more welcoming of the other and incorporate it into the republic to make them part of the workings.<mask><mask> i have written that, you might see<mask> i place the current American system... and<mask> i worry. [NEWLINE] [NEWLINE] [STARTQ] the state guarantees peace of mind for minorities [ENDQ] [NEWLINE] I'm sorry i can't imagine you said that with a straight face. Have you seen that happens to minorities<mask> a strong state has power? Holocaust? Armenian genocide? Pogroms? The state only guarantee peace of mind for those who are part of the state,<mask> that make up of the state is against a race or minority, they get...... removed. In a free society those groups might be shit on,<mask> they can come together and make means to defend themselves<mask> much<mask> possible.<mask>
Label encoding: <s> [STARTQ] maximizing security &amp; happiness for yourself and others would be the obvious choice. [ENDQ] [NEWLINE] That is one big point. With being allowed to decided for yourself the means in which you can make yourself safe and happy are on you, so you can make that happen as you wish so long as you don't interfere with others doing the same. This is because the states choice of what makes you safe, may not fit what would be needed to make you safe, also the state may choose your not worth making safe. [NEWLINE] [NEWLINE] [STARTQ] In my opinion, natural evolution is slow and dumb. [ENDQ] [NEWLINE] Well lets look at states and how fast they evolved under traditional monarchy's. States where there was one major power that dictated everything efficiently... any have rapid advances? not as much, why? because things are good for those in charge the way they are. No need to change, and change likely means more risks to power coming out to present threats.( max weber writes about this in detail i think..) Under more free society's, people change and evolve more rapidly, as nothing is necessarily sacred. So massive booms to trade, to technology, and overall happiness increases when a society is free to evolve naturally. [NEWLINE] [NEWLINE] When Rome was a republic is never fell, if you remember it fell as an empire. That's because republics are strong and able to adapt for more long term success, more so than short term. Also Republics usually build wealth and incorporate more into their state and nation, where as Empire is usually expending wealth to "keep the huns away from the gates" if you will. Empires also usually rely on creating an "Other" to target as a threat to keep the populations mobilized with its rule. Republics usually are more welcoming of the other and incorporate it into the republic to make them part of the workings. Given that i have written that, you might see where i place the current American system... and why i worry. [NEWLINE] [NEWLINE] [STARTQ] the state guarantees peace of mind for minorities [ENDQ] [NEWLINE] I'm sorry i can't imagine you said that with a straight face. Have you seen that happens to minorities when a strong state has power? Holocaust? Armenian genocide? Pogroms? The state only guarantee peace of mind for those who are part of the state, if that make up of the state is against a race or minority, they get...... removed. In a free society those groups might be shit on, but they can come together and make means to defend themselves as much as possible. While
Loss: tensor(0.6126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4827, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Except that pushing the middle class further and further out substantially increases their commute time and fuel costs. We've seen a lot of sprawl without the accompanying infrastructure improvements we've had in the past (subways, light rail, etc), which results in longer commute time and congestion. [ENDQ] [NEWLINE] A lot of this can be attributed to poor urban planning and a lack of commitment to infrastructure by all levels of government,<mask> "we're going to repave the highway" really doesn't get votes. From<mask> I have seen in Canada and in other countries there is a lot of pressure from the opposition in government to invest in infrastructure<mask> it is necessary, and it<mask> means the government has less resources to put towards the flashy things that get them votes. And the general public is starting to see just<mask> bad our roads, rail, bridges, and even power grids, water and sewage are getting. [NEWLINE] [NEWLINE] Toronto has spent millions of dollars on cancelling transit expansions, new mayors coming up with new plans, people wanting subways instead of LRT's etc. Everyone wants increased services and service areas,<mask> no one wants to pay an increased rate to use the service, or new taxes or other costs to access it. [NEWLINE] [NEWLINE] It can<mask> be linked directly back to inflated lifestyle/standard of living. Everyone wants a suburban home on a quiet residential street, and then they complain that there are no amenities close to them and their commute is too long. You can't simultaneously move out to the suburbs to get away from the noise of downtown, and expect everything that used to be downtown right next door. [NEWLINE] [NEWLINE] [STARTQ] Leisure time is another good metric to measure quality. Add commute time + hours worked (which has increased) + vacation time (which has decreased), and it again doesn't look great. [ENDQ] [NEWLINE] We are starting to see push back to this in some European countries<mask> they are moving to 35 or 30 hour work weeks instead of 40 for government employees and the like. Whether or not that translates to parts of the economy<mask> a "full time 40 hour work week" actually means a 60+ hour week, I really don't know. Work/life balance is something a lot of employees and employers claim to be paying more attention to,<mask> most of that action is happening in higher paid non middle class professions. [NEWLINE] [NEWLINE] Productivity is up exponentially in almost every sector,<mask> that hasn't lead to increases in pay or decreases in work hours. A large part of that is cultural,<mask> we define ourselves by our work and
Label encoding: <s> [STARTQ] Except that pushing the middle class further and further out substantially increases their commute time and fuel costs. We've seen a lot of sprawl without the accompanying infrastructure improvements we've had in the past (subways, light rail, etc), which results in longer commute time and congestion. [ENDQ] [NEWLINE] A lot of this can be attributed to poor urban planning and a lack of commitment to infrastructure by all levels of government, as "we're going to repave the highway" really doesn't get votes. From what I have seen in Canada and in other countries there is a lot of pressure from the opposition in government to invest in infrastructure because it is necessary, and it also means the government has less resources to put towards the flashy things that get them votes. And the general public is starting to see just how bad our roads, rail, bridges, and even power grids, water and sewage are getting. [NEWLINE] [NEWLINE] Toronto has spent millions of dollars on cancelling transit expansions, new mayors coming up with new plans, people wanting subways instead of LRT's etc. Everyone wants increased services and service areas, but no one wants to pay an increased rate to use the service, or new taxes or other costs to access it. [NEWLINE] [NEWLINE] It can also be linked directly back to inflated lifestyle/standard of living. Everyone wants a suburban home on a quiet residential street, and then they complain that there are no amenities close to them and their commute is too long. You can't simultaneously move out to the suburbs to get away from the noise of downtown, and expect everything that used to be downtown right next door. [NEWLINE] [NEWLINE] [STARTQ] Leisure time is another good metric to measure quality. Add commute time + hours worked (which has increased) + vacation time (which has decreased), and it again doesn't look great. [ENDQ] [NEWLINE] We are starting to see push back to this in some European countries where they are moving to 35 or 30 hour work weeks instead of 40 for government employees and the like. Whether or not that translates to parts of the economy where a "full time 40 hour work week" actually means a 60+ hour week, I really don't know. Work/life balance is something a lot of employees and employers claim to be paying more attention to, but most of that action is happening in higher paid non middle class professions. [NEWLINE] [NEWLINE] Productivity is up exponentially in almost every sector, but that hasn't lead to increases in pay or decreases in work hours. A large part of that is cultural, where we define ourselves by our work and
Loss: tensor(0.7479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] edit: to clarify, I'm not talking about Java here - I've never used it, and<mask><mask> it is vilified<mask> much<mask> VS. [ENDQ] [NEWLINE] Quite true -- more<mask>, in some circles.<mask> the Java stack is now entirely open source, including the best IDEs,<mask> there's that. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> all these new languages are hype and most of you don't even know<mask> intelli-sense is. [ENDQ] [NEWLINE] First, define "new". Javascript was new in 1995. [NEWLINE] [NEWLINE] <mask><mask> you know those languages well, you know that many of their advantages are derived from exactly the sort of choices which make tooling difficult. You mention Rails -- well, Ruby has strong-<mask> -dynamic typing. Dynamic typing means an IDE can only do very limited analysis on your objects -- something like IntelliSense would be impossible. [NEWLINE] [NEWLINE] The tradeoff is, you don't have to define a huge hierarchy of classes, or even types. You get testability and mockability for free. [NEWLINE] [NEWLINE] And best of all, you get a REPL. The best of these gives you the equivalent of IntelliSense in tab-completion. [NEWLINE] [NEWLINE] [STARTQ] It's like going back 20 years...Well, at least 15 - at least we had debug, edit, and continue back then. [ENDQ] [NEWLINE] <mask>, in return, you get a commandline with which to explore the program. You can type "rails console" and inspect the model in real time, interactively, in a very free-form manner. Debuggers are great, and I don't think this can truly replace a debugger,<mask> I<mask> don't think debuggers replace this. (And I can *partially* replace a debugger -- I can invoke a REPL from inside some problematic code, instead of firing up a debugger, and type arbitrary Ruby commands to inspect the situation.) [NEWLINE] [NEWLINE] Look at the Chrome dev tools again -- a high-quality debugger, sure,<mask><mask> a console. Ever worked with XML in a language like C# or Java? Compare typing XQuery, running your main loop, edit, compile, run, edit, compile, run... Compare that with jQuery in the Chrome dev tools. Ctrl+Shift+J right here, Reddit uses jQuery, you could immediately try something like $('body').css('background-color', 'lime') -- or just $('body') to see<mask> that matches, hover over the result to see it highlighted on the page. And<mask> you want
Label encoding: <s> [STARTQ] edit: to clarify, I'm not talking about Java here - I've never used it, and I think it is vilified as much as VS. [ENDQ] [NEWLINE] Quite true -- more so, in some circles. But the Java stack is now entirely open source, including the best IDEs, so there's that. [NEWLINE] [NEWLINE] [STARTQ] I think all these new languages are hype and most of you don't even know what intelli-sense is. [ENDQ] [NEWLINE] First, define "new". Javascript was new in 1995. [NEWLINE] [NEWLINE] But if you know those languages well, you know that many of their advantages are derived from exactly the sort of choices which make tooling difficult. You mention Rails -- well, Ruby has strong- but -dynamic typing. Dynamic typing means an IDE can only do very limited analysis on your objects -- something like IntelliSense would be impossible. [NEWLINE] [NEWLINE] The tradeoff is, you don't have to define a huge hierarchy of classes, or even types. You get testability and mockability for free. [NEWLINE] [NEWLINE] And best of all, you get a REPL. The best of these gives you the equivalent of IntelliSense in tab-completion. [NEWLINE] [NEWLINE] [STARTQ] It's like going back 20 years...Well, at least 15 - at least we had debug, edit, and continue back then. [ENDQ] [NEWLINE] But, in return, you get a commandline with which to explore the program. You can type "rails console" and inspect the model in real time, interactively, in a very free-form manner. Debuggers are great, and I don't think this can truly replace a debugger, but I also don't think debuggers replace this. (And I can *partially* replace a debugger -- I can invoke a REPL from inside some problematic code, instead of firing up a debugger, and type arbitrary Ruby commands to inspect the situation.) [NEWLINE] [NEWLINE] Look at the Chrome dev tools again -- a high-quality debugger, sure, but also a console. Ever worked with XML in a language like C# or Java? Compare typing XQuery, running your main loop, edit, compile, run, edit, compile, run... Compare that with jQuery in the Chrome dev tools. Ctrl+Shift+J right here, Reddit uses jQuery, you could immediately try something like $('body').css('background-color', 'lime') -- or just $('body') to see what that matches, hover over the result to see it highlighted on the page. And if you want
Loss: tensor(0.5398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.9036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Psychology is the scientific study of the brain, mind, behavior, and nervous system.<mask> do you mean<mask> you say it's not a "real science"? The whole point of psychology is that only the scientific studies are considered psychology. [NEWLINE] [NEWLINE] [STARTQ] The scientific method has the foundation of fact, and it's constructed by logic. [ENDQ] [NEWLINE] This is a view within philosophy of science<mask> I don't think it's universally accepted. Ultimately, it comes down to whether you're a rationalist, an empiricist, or a little bit of both.<mask> science is ultimately a study of sensory experience and<mask> heavily leans towards empiricism.<mask> no, it's not "constructed by logic",<mask> rather logic is constructed by our sensory experience.<mask> is or is not logical ultimately depends on our sensory experience. One example is the idea that two different objects can exist in the exact same time and space. It used to be "illogical" for that to be true. And<mask>, quantum mechanics tells us that two different electrons can be in the exact same time and space.<mask> now, it's no longer considered "illogical". [NEWLINE] [NEWLINE] [STARTQ] Psychology attempts to construct with logic,<mask> lacks the foundation of fact. It's based on assumption, probability, and inconsistent statistics. [ENDQ] [NEWLINE] All science is based on assumptions, probabilities, and inconsistent statistics. The reason<mask> is<mask> (1) humans are fallible, and (2) our sensory experiences are not uniform,<mask> rather change over time. [NEWLINE] [NEWLINE] [STARTQ] The reason for this lack of fact is the random variable: choice, and conscious will. [ENDQ] [NEWLINE] The problem is that psychologists have studied these things and have concluded, based on good evidence, that<mask> people think of<mask> "choice" or "conscious will" are<mask><mask> determined by a person's internal structures, such<mask> genes, brain structures, and brain chemistry. [NEWLINE] [NEWLINE] [STARTQ] There are oxymoronic terms used, too, like: personality disorder. [ENDQ] [NEWLINE] A personality disorder<mask> your behavior, cognition or inner experience cause strain on either yourself or the society that you live in. There's nothing oxymoronic about it. [NEWLINE] [NEWLINE] [STARTQ] <mask> personality is unique to every individual, it's shaped by their choices and experiences. Some might argue, like me, that core personality never changes through out life. You can't wrap it up into a box and label it.<mask> to call something, that has no clear shape to begin with, disorderly is an oxymoron. [ENDQ] [NEWLINE] Except this is contradicted by
Label encoding: <s>Psychology is the scientific study of the brain, mind, behavior, and nervous system. What do you mean when you say it's not a "real science"? The whole point of psychology is that only the scientific studies are considered psychology. [NEWLINE] [NEWLINE] [STARTQ] The scientific method has the foundation of fact, and it's constructed by logic. [ENDQ] [NEWLINE] This is a view within philosophy of science but I don't think it's universally accepted. Ultimately, it comes down to whether you're a rationalist, an empiricist, or a little bit of both. But science is ultimately a study of sensory experience and thus heavily leans towards empiricism. So no, it's not "constructed by logic", but rather logic is constructed by our sensory experience. What is or is not logical ultimately depends on our sensory experience. One example is the idea that two different objects can exist in the exact same time and space. It used to be "illogical" for that to be true. And yet, quantum mechanics tells us that two different electrons can be in the exact same time and space. So now, it's no longer considered "illogical". [NEWLINE] [NEWLINE] [STARTQ] Psychology attempts to construct with logic, but lacks the foundation of fact. It's based on assumption, probability, and inconsistent statistics. [ENDQ] [NEWLINE] All science is based on assumptions, probabilities, and inconsistent statistics. The reason why is because (1) humans are fallible, and (2) our sensory experiences are not uniform, but rather change over time. [NEWLINE] [NEWLINE] [STARTQ] The reason for this lack of fact is the random variable: choice, and conscious will. [ENDQ] [NEWLINE] The problem is that psychologists have studied these things and have concluded, based on good evidence, that what people think of as "choice" or "conscious will" are in fact determined by a person's internal structures, such as genes, brain structures, and brain chemistry. [NEWLINE] [NEWLINE] [STARTQ] There are oxymoronic terms used, too, like: personality disorder. [ENDQ] [NEWLINE] A personality disorder when your behavior, cognition or inner experience cause strain on either yourself or the society that you live in. There's nothing oxymoronic about it. [NEWLINE] [NEWLINE] [STARTQ] So personality is unique to every individual, it's shaped by their choices and experiences. Some might argue, like me, that core personality never changes through out life. You can't wrap it up into a box and label it. So to call something, that has no clear shape to begin with, disorderly is an oxymoron. [ENDQ] [NEWLINE] Except this is contradicted by
Loss: tensor(0.4222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3958, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6596, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Serious attempt to understand the logic and come to a final conclusion here. Backstory: there's women that have been in my life over the past few years that have accused me of being a misogynist for a variety of reasons. I had a discussion today that resulted in me, again, being called a misogynist.<mask> of right now I tend to assume it's just an insult that's being cast at me<mask> I've "won" an argument,<mask> being used<mask> an easy-out.<mask>,<mask><mask> it's true?<mask><mask>, I should probably work on some self-improvement right? [NEWLINE] [NEWLINE] Here's<mask><mask><mask> I *might* be a misogynist: [NEWLINE] [NEWLINE] * Let's just start by saying I absolutely despise the mother of my child<mask> it seems to be a frequent reason I'm interpreted<mask> a "woman hater." It would be false to say that I haven't let my experiences with her have an affect on me,<mask> it would<mask> be false to say that<mask> I hate her I must hate all women. [NEWLINE] [NEWLINE] * Second: perhaps<mask> I would see<mask> my closest-to-misogynistic trait. It's more difficult for women to gain my trust. This generally only applies to women who I'm considering a relationship with. Probably leads into my next point.<mask>, I don't have a problem being friends with women or associating with them<mask><mask> much do you *really* need to be able to trust a friend or associate? On that level,<mask><mask> distrust everyone fairly equally. For that matter, I don't really try to get into relationships with men<mask>... maybe I would distrust men in the same way? Not sure. [NEWLINE] [NEWLINE] * In all likelihood<mask> of the above point I don't usually give two flying fucks about having sex or especially getting into a committed relationship that extends beyond dating. [NEWLINE] [NEWLINE] * On occasions outside of the above: I do from time to time partake in "no strings" or "friends with benefits" arrangements. I make my intentions clear from<mask> early<mask> possible,<mask><mask> asked about escalating to a legitimate relationship I tend to back out of the arrangement and put an end to it. I see this<mask> my best effort to not "lead someone on" or "tie them down"<mask> clearly we don't share the same end-game.<mask> I wanted a relationship that's<mask> I would have aimed for from the beginning. [NEWLINE] [NEWLINE] * I have standards. The few times I have been in relationships my
Label encoding: <s>Serious attempt to understand the logic and come to a final conclusion here. Backstory: there's women that have been in my life over the past few years that have accused me of being a misogynist for a variety of reasons. I had a discussion today that resulted in me, again, being called a misogynist. As of right now I tend to assume it's just an insult that's being cast at me because I've "won" an argument, thus being used as an easy-out. However, what if it's true? If so, I should probably work on some self-improvement right? [NEWLINE] [NEWLINE] Here's why I think I *might* be a misogynist: [NEWLINE] [NEWLINE] * Let's just start by saying I absolutely despise the mother of my child as it seems to be a frequent reason I'm interpreted as a "woman hater." It would be false to say that I haven't let my experiences with her have an affect on me, but it would also be false to say that because I hate her I must hate all women. [NEWLINE] [NEWLINE] * Second: perhaps what I would see as my closest-to-misogynistic trait. It's more difficult for women to gain my trust. This generally only applies to women who I'm considering a relationship with. Probably leads into my next point. However, I don't have a problem being friends with women or associating with them because how much do you *really* need to be able to trust a friend or associate? On that level, I think distrust everyone fairly equally. For that matter, I don't really try to get into relationships with men so... maybe I would distrust men in the same way? Not sure. [NEWLINE] [NEWLINE] * In all likelihood because of the above point I don't usually give two flying fucks about having sex or especially getting into a committed relationship that extends beyond dating. [NEWLINE] [NEWLINE] * On occasions outside of the above: I do from time to time partake in "no strings" or "friends with benefits" arrangements. I make my intentions clear from as early as possible, but when asked about escalating to a legitimate relationship I tend to back out of the arrangement and put an end to it. I see this as my best effort to not "lead someone on" or "tie them down" because clearly we don't share the same end-game. If I wanted a relationship that's what I would have aimed for from the beginning. [NEWLINE] [NEWLINE] * I have standards. The few times I have been in relationships my
Loss: tensor(0.4982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.4409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.3421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3649, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyone agrees you cannot guarantee that governments or corporations aren't evil, or that they are doing the best they know<mask> to,<mask> I stand ready for a coming communist utopia instead of hoping and wishing against the world ending<mask> some guy tried to make a cold-fusion yacht and it worked<mask> one of his underpaid overworked foreign employees blew it up and accidentally imploded time space<mask> we know it,,, [NEWLINE] [NEWLINE] (of course this ungrateful vampire of an employee could have been a communist saboteur who wanted the right to be a leech on society more than he cared about the current balance of existence/status quo of reality,<mask> we will never know<mask> the yacht being private property it would be wrong to ask about its' trade secret design and of course the man should be given a tax break even<mask> we discover its' primary use is<mask> a doomsday device which<mask> happened to power his hyper-space yacht- we would be forced to do right by his right to privacy and copyright<mask> would be required to help him hunt down any pirates who stole the technology<mask><mask><mask> he has the legal right to seize and destroy these illegal copies we wouldn't want people to say that we encouraged him to be a vigilante,<mask><mask> he currently uses his power to threaten the universe itself in capitalist terminology he has used his hard work and ingenuity to come up with a product which gives him nigh unlimited power and<mask> graciously offers us the'service' of both not destroying the universe<mask><mask> protecting us from those who would... [NEWLINE] [NEWLINE] [STARTQ] tl;dr: Governments will always wield more dangerous powers<mask> always provide the mechanisms for wider and wider distribution of those powers to the many. Corporations will always wield less dangerous power (<mask> sometimes only slightly less dangerous) and will always minimize distributed decision making<mask> it is antithetical to the goal of maximal efficiency. [ENDQ] [NEWLINE] + [URL] [NEWLINE] This'stand up' comedian pointing out<mask> cannot be commodified, like having a great job<mask> you just try to be funny,<mask> I respect that it is OK to get<mask> rich<mask> you want, I feel that the priority will always be distributing needs and that a rational society will enforce the governments supremacy over the individual for the individual, this meaning the comic might be forced to work in a mine, for low wages<mask> a guaranteed roof or at least bed, and food. Whereas in our current system we say that it is ridiculous that such absolute power would not be absolutely corrupted,<mask> think it is funny<mask> this guy
Label encoding: <s>Everyone agrees you cannot guarantee that governments or corporations aren't evil, or that they are doing the best they know how to, but I stand ready for a coming communist utopia instead of hoping and wishing against the world ending because some guy tried to make a cold-fusion yacht and it worked but one of his underpaid overworked foreign employees blew it up and accidentally imploded time space as we know it,,, [NEWLINE] [NEWLINE] (of course this ungrateful vampire of an employee could have been a communist saboteur who wanted the right to be a leech on society more than he cared about the current balance of existence/status quo of reality, but we will never know because the yacht being private property it would be wrong to ask about its' trade secret design and of course the man should be given a tax break even if we discover its' primary use is as a doomsday device which also happened to power his hyper-space yacht- we would be forced to do right by his right to privacy and copyright but would be required to help him hunt down any pirates who stole the technology since even though he has the legal right to seize and destroy these illegal copies we wouldn't want people to say that we encouraged him to be a vigilante, even though he currently uses his power to threaten the universe itself in capitalist terminology he has used his hard work and ingenuity to come up with a product which gives him nigh unlimited power and also graciously offers us the'service' of both not destroying the universe but also protecting us from those who would... [NEWLINE] [NEWLINE] [STARTQ] tl;dr: Governments will always wield more dangerous powers but always provide the mechanisms for wider and wider distribution of those powers to the many. Corporations will always wield less dangerous power ( though sometimes only slightly less dangerous) and will always minimize distributed decision making as it is antithetical to the goal of maximal efficiency. [ENDQ] [NEWLINE] + [URL] [NEWLINE] This'stand up' comedian pointing out what cannot be commodified, like having a great job where you just try to be funny, while I respect that it is OK to get as rich as you want, I feel that the priority will always be distributing needs and that a rational society will enforce the governments supremacy over the individual for the individual, this meaning the comic might be forced to work in a mine, for low wages but a guaranteed roof or at least bed, and food. Whereas in our current system we say that it is ridiculous that such absolute power would not be absolutely corrupted, but think it is funny how this guy
Loss: tensor(0.3674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2839, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This has been tackled beautifully by /u/dougie_g and I will reproduce xis post below. All credit to xim: [NEWLINE] [NEWLINE] [STARTQ] Caution, big law essay coming up. [ENDQ] In the UK<mask><mask> we have the balance right, and I'm not convinced that the standard is different in the States,<mask> please correct me<mask> I'm wrong. This question comes up a lot on reddit, and it seems<mask><mask> people believe it is automatically rape to have sex with a drunk person,<mask> this is demonstrably untrue. [NEWLINE] Crime has two elements - actus reus and mens rea. Actus reus is the unlawful act itself. In the case of rape, the actus reus is having sex with a person who has not consented, or is unable to consent (due to certain factors such<mask> age), to having sex with you. The mens rea is the mental element - the state of mind of the perpetrator. The mens rea of rape is that the rapist knew, or reasonably should have known, that the victim had not consented to sex. [NEWLINE] [NEWLINE] [STARTQ] Intoxication can affect both. First, did the victim<mask><mask> offer valid consent? It is possible they did not<mask> they were intoxicated. More on this in a moment. Second,<mask> the person was too intoxicated to consent, did the defendant know, or should the defendant have known that the victim was too intoxicated to offer valid consent? [ENDQ] [NEWLINE] [STARTQ] Basically, there are two kinds of intoxication: voluntary, and involuntary. An example of involuntary intoxication is<mask> you have your drink spiked. It isn't enough to say 'well I only had ten beers and normally I'm fine with that'. You still drank ten beers and should expect to be intoxicated.<mask><mask> you only have one small beer and you are completely incapable of functioning (and that is very unusual for you) then you may be involuntarily intoxicated. [ENDQ] [NEWLINE] [STARTQ] Voluntary intoxication is pretty self explanatory. [ENDQ] [NEWLINE] [STARTQ] Now,<mask> voluntarily intoxicated, you can still technically consent to anything you can consent to<mask> sober.<mask> a drunk person couldn't give legal consent then you could get drunk before signing a contract, then back out of that contract later on by proving you were drunk at the time. This does not,<mask>, extend to<mask> you are completely mentally and/or physically incapacitated.<mask> a girl passes out drunk at a male friend's place, it might be rape for him to have sex with her, unless they
Label encoding: <s>This has been tackled beautifully by /u/dougie_g and I will reproduce xis post below. All credit to xim: [NEWLINE] [NEWLINE] [STARTQ] Caution, big law essay coming up. [ENDQ] In the UK I think we have the balance right, and I'm not convinced that the standard is different in the States, but please correct me if I'm wrong. This question comes up a lot on reddit, and it seems as if people believe it is automatically rape to have sex with a drunk person, but this is demonstrably untrue. [NEWLINE] Crime has two elements - actus reus and mens rea. Actus reus is the unlawful act itself. In the case of rape, the actus reus is having sex with a person who has not consented, or is unable to consent (due to certain factors such as age), to having sex with you. The mens rea is the mental element - the state of mind of the perpetrator. The mens rea of rape is that the rapist knew, or reasonably should have known, that the victim had not consented to sex. [NEWLINE] [NEWLINE] [STARTQ] Intoxication can affect both. First, did the victim in fact offer valid consent? It is possible they did not if they were intoxicated. More on this in a moment. Second, if the person was too intoxicated to consent, did the defendant know, or should the defendant have known that the victim was too intoxicated to offer valid consent? [ENDQ] [NEWLINE] [STARTQ] Basically, there are two kinds of intoxication: voluntary, and involuntary. An example of involuntary intoxication is when you have your drink spiked. It isn't enough to say 'well I only had ten beers and normally I'm fine with that'. You still drank ten beers and should expect to be intoxicated. But if you only have one small beer and you are completely incapable of functioning (and that is very unusual for you) then you may be involuntarily intoxicated. [ENDQ] [NEWLINE] [STARTQ] Voluntary intoxication is pretty self explanatory. [ENDQ] [NEWLINE] [STARTQ] Now, while voluntarily intoxicated, you can still technically consent to anything you can consent to when sober. If a drunk person couldn't give legal consent then you could get drunk before signing a contract, then back out of that contract later on by proving you were drunk at the time. This does not, however, extend to when you are completely mentally and/or physically incapacitated. If a girl passes out drunk at a male friend's place, it might be rape for him to have sex with her, unless they
Loss: tensor(0.3766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2941, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.5572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4992, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is<mask> true for Wal-Mart,<mask> not all corporations are built like that[1], and at least some of those do manage to be profitable businesses. [ENDQ] [NEWLINE] Certainly, there are many examples of B's<mask> well<mask> co-op's and other organizational structures which act towards mutual benefit instead of pure profit. [NEWLINE] [NEWLINE] The reason these serve wealthier markets is precisely<mask> its more expensive to run a business this way, Wal-Mart would not be able to become a B corp<mask> still serving the extreme low-income market it does<mask> it would require them to raise prices. [NEWLINE] [NEWLINE] [STARTQ] You could<mask> see the problem<mask> an undersupply of well-paid jobs. [ENDQ] [NEWLINE] Which wouldn't fit the data, we have had near continuous labor shortage in many high-skilled industries for well over a decade now. We certainly have fewer semi-skilled positions then in the past<mask> this is true through all advanced economies, in those economies labor have changed to meet the new skills requirements<mask> we have not. [NEWLINE] [NEWLINE] <mask> is occurring in labor markets is the growth in high-skilled &amp; high-income roles, semi-skilled &amp; mid-income jobs are going away and all others are remaining constant. The entry requirements for the new high-skilled jobs are actually relatively low<mask> American's are simply not retraining and instead are holding out in the vain hope the old semi-skilled jobs return. [NEWLINE] [NEWLINE] [STARTQ] We<mask> have plenty of work which needs to be done [ENDQ] [NEWLINE] This is incorrect, the organization which publishes this data is the organization who represents civil engineers throughout the country and uses phrases like "red list" to make things sound scary<mask> they are not. [NEWLINE] [NEWLINE] <mask> a good example of<mask> the problem exists in this issue any bridge constructed prior to 1972 which has not effectively been structurally redesigned will appear on the DOT's "red list". These bridges are not unsafe and will continue to perform safely until the end of their operation lifetime<mask> simply no longer meet the federal standards for bridges due to the introduction of new standards. [NEWLINE] [NEWLINE] Replacing this infrastructure would not provide us with an economic benefit, replacement would have an extremely low multiplier effect. [NEWLINE] [NEWLINE] [STARTQ] and it's very, very cheap to borrow money [ENDQ] [NEWLINE] Its only cheap to borrow<mask> we repay it at maturity, which we don't and haven't for many decades, otherwise at maturity it will rollover in to new debt at a higher yield. [NEWLINE] [NEWLINE] The long term inability of the federal government to manage debt effectively is a fairly
Label encoding: <s> [STARTQ] This is indeed true for Wal-Mart, but not all corporations are built like that[1], and at least some of those do manage to be profitable businesses. [ENDQ] [NEWLINE] Certainly, there are many examples of B's as well as co-op's and other organizational structures which act towards mutual benefit instead of pure profit. [NEWLINE] [NEWLINE] The reason these serve wealthier markets is precisely because its more expensive to run a business this way, Wal-Mart would not be able to become a B corp while still serving the extreme low-income market it does as it would require them to raise prices. [NEWLINE] [NEWLINE] [STARTQ] You could also see the problem as an undersupply of well-paid jobs. [ENDQ] [NEWLINE] Which wouldn't fit the data, we have had near continuous labor shortage in many high-skilled industries for well over a decade now. We certainly have fewer semi-skilled positions then in the past but this is true through all advanced economies, in those economies labor have changed to meet the new skills requirements while we have not. [NEWLINE] [NEWLINE] What is occurring in labor markets is the growth in high-skilled &amp; high-income roles, semi-skilled &amp; mid-income jobs are going away and all others are remaining constant. The entry requirements for the new high-skilled jobs are actually relatively low but American's are simply not retraining and instead are holding out in the vain hope the old semi-skilled jobs return. [NEWLINE] [NEWLINE] [STARTQ] We also have plenty of work which needs to be done [ENDQ] [NEWLINE] This is incorrect, the organization which publishes this data is the organization who represents civil engineers throughout the country and uses phrases like "red list" to make things sound scary when they are not. [NEWLINE] [NEWLINE] As a good example of where the problem exists in this issue any bridge constructed prior to 1972 which has not effectively been structurally redesigned will appear on the DOT's "red list". These bridges are not unsafe and will continue to perform safely until the end of their operation lifetime but simply no longer meet the federal standards for bridges due to the introduction of new standards. [NEWLINE] [NEWLINE] Replacing this infrastructure would not provide us with an economic benefit, replacement would have an extremely low multiplier effect. [NEWLINE] [NEWLINE] [STARTQ] and it's very, very cheap to borrow money [ENDQ] [NEWLINE] Its only cheap to borrow if we repay it at maturity, which we don't and haven't for many decades, otherwise at maturity it will rollover in to new debt at a higher yield. [NEWLINE] [NEWLINE] The long term inability of the federal government to manage debt effectively is a fairly
Loss: tensor(0.4553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3905, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.3027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I run a company and we maintain a work dress code for professionals that don't interact with customers (We have a dress code for everyone,<mask> obviously it slightly differs to accommodate<mask> manual labor is involved, etc) for one very simple reason: I don't want to deal with bullshit. [NEWLINE] [NEWLINE] Consider the following: [NEWLINE] [NEWLINE] Bill: Earl, John is mocking my religion. I shouldn't have to work in an environment<mask> my religion is being mocked. [NEWLINE] [NEWLINE] John: I'm not mocking his religious beliefs, I'm expressing mine. [NEWLINE] [NEWLINE] Bill: He's wearing a flying spaghetti monster shirt, Earl. That's not a real belief. He's just making fun of me<mask> I'm a Christian. [NEWLINE] [NEWLINE] John: It's a valid belief, just<mask> valid<mask> his. He wears his Jesus shirt all the time, I have the same rights<mask> him to wear a shirt about my beliefs. [NEWLINE] [NEWLINE] Me: Okay, no religious wear is allowed anymore to avoid offending anymore. [NEWLINE] [NEWLINE] Bill: Well<mask> I can't wear my Jesus shirt, Gary shouldn't be allowed to wear his shirt with Satanist symbols on it. [NEWLINE] [NEWLINE] Gary: It's not Satanist. It's a shirt of my favorite band's album. [NEWLINE] [NEWLINE] Bill: It looks like a star of Satan to me, Earl. [NEWLINE] [NEWLINE] Gary: You don't know<mask> you're talking about. You're thinking of a pentagram, which isn't even historically Satanist, and this isn't even a pentagram. [NEWLINE] [NEWLINE] Bill: It looks like a star of Satan to me, Earl. [NEWLINE] [NEWLINE] Me: Okay, everyone just wear plain shirts with no logos or writing on them. [NEWLINE] [NEWLINE] Frank: Well don't do that,<mask> then Jenny is going to wear her tight little white t-shirts every day which are really distracting. [NEWLINE] [NEWLINE] Jenny:<mask>? <mask> is my wearing a white t-shirt have anything to do with your religion debate? [NEWLINE] [NEWLINE] Frank: you know<mask> you're doing. You wear those shirts that show off your chest, which is distracting. It's reverse sexual harassment, Earl. Equal rights go both ways. [NEWLINE] [NEWLINE] Jenny: Don't blame me just<mask> you're a pervert. You walk around staring at my legs all the time. I'm the one being harassed. [NEWLINE] [NEWLINE] Frank: See she knows<mask> she's doing. She dresses sexy on purpose. It makes me uncomfortable and isn't professional. [NEWLINE] [NEWLINE] Jenny: You wear shorts all the time
Label encoding: <s>I run a company and we maintain a work dress code for professionals that don't interact with customers (We have a dress code for everyone, though obviously it slightly differs to accommodate if manual labor is involved, etc) for one very simple reason: I don't want to deal with bullshit. [NEWLINE] [NEWLINE] Consider the following: [NEWLINE] [NEWLINE] Bill: Earl, John is mocking my religion. I shouldn't have to work in an environment where my religion is being mocked. [NEWLINE] [NEWLINE] John: I'm not mocking his religious beliefs, I'm expressing mine. [NEWLINE] [NEWLINE] Bill: He's wearing a flying spaghetti monster shirt, Earl. That's not a real belief. He's just making fun of me because I'm a Christian. [NEWLINE] [NEWLINE] John: It's a valid belief, just as valid as his. He wears his Jesus shirt all the time, I have the same rights as him to wear a shirt about my beliefs. [NEWLINE] [NEWLINE] Me: Okay, no religious wear is allowed anymore to avoid offending anymore. [NEWLINE] [NEWLINE] Bill: Well if I can't wear my Jesus shirt, Gary shouldn't be allowed to wear his shirt with Satanist symbols on it. [NEWLINE] [NEWLINE] Gary: It's not Satanist. It's a shirt of my favorite band's album. [NEWLINE] [NEWLINE] Bill: It looks like a star of Satan to me, Earl. [NEWLINE] [NEWLINE] Gary: You don't know what you're talking about. You're thinking of a pentagram, which isn't even historically Satanist, and this isn't even a pentagram. [NEWLINE] [NEWLINE] Bill: It looks like a star of Satan to me, Earl. [NEWLINE] [NEWLINE] Me: Okay, everyone just wear plain shirts with no logos or writing on them. [NEWLINE] [NEWLINE] Frank: Well don't do that, because then Jenny is going to wear her tight little white t-shirts every day which are really distracting. [NEWLINE] [NEWLINE] Jenny: What?  How is my wearing a white t-shirt have anything to do with your religion debate? [NEWLINE] [NEWLINE] Frank: you know what you're doing. You wear those shirts that show off your chest, which is distracting. It's reverse sexual harassment, Earl. Equal rights go both ways. [NEWLINE] [NEWLINE] Jenny: Don't blame me just because you're a pervert. You walk around staring at my legs all the time. I'm the one being harassed. [NEWLINE] [NEWLINE] Frank: See she knows what she's doing. She dresses sexy on purpose. It makes me uncomfortable and isn't professional. [NEWLINE] [NEWLINE] Jenny: You wear shorts all the time
Loss: tensor(0.6168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2911, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off: Thank you for the very reasonable and well-thought response.<mask><mask>, especially given sensitive subject matter like this, it is too easy to personally attack someone or become overly emotional about a subject<mask> honest questions are posed; you did neither and I really appreciate it. [NEWLINE] [NEWLINE] <mask> : The article I linked to was less an avocation for its overall point, and more just a justification for pointing out that animals do die due to a vegetarian diet—something that I have overhead some evangelical vegetarian/vegans argue isn’t possible or is simply made up. You not only provided a delightful bit of rebuttal reading for tonight,<mask><mask> acknowledge the intended fact without problems, and again I salute your reasonableness! [NEWLINE] [NEWLINE] Now, onto<mask> I hope is further amicable discussion. [NEWLINE] [NEWLINE] [STARTQ] A philosophical point would be that there is (from my point of view, at least) an ethical difference between animal deaths that occur<mask> a unintended secondary effect, and the ones that are deliberately and directly caused to happen. Essentially, it's a difference of intent;<mask> I eat vegetables, I don't intend to cause the deaths of any animals, and I have no control over them. These deaths are certainly not inherent to eating vegetables; [ENDQ] [NEWLINE] <mask> we assume KerSan’s point of “There is no such thing<mask> ethical murder” then we are left both of us being nothing<mask> simple killers for our food.  (He/she might very well be one<mask> well given our discussion point,<mask> that is a different discussion.) This gets both of us nowhere,<mask> let’s focus on your point. [NEWLINE] [NEWLINE] I don’t see a philosophical difference between the intentional killing of animals for my food, and killing that is unintended<mask> a known byproduct of my diet. In both cases “your” food supply is marked by the death of animals,<mask> in the omnivorous diet, at least those animals are useful.<mask> you know that animals die specifically to supply “you” with food, you are morally equivalent. These deaths ARE inherent to eating vegetables,<mask> much<mask> death is inherent to eating steak. “You” just don’t have to be faced with that moral quandary directly at your butcher counter. [NEWLINE] [NEWLINE] For meat-eaters, at least their animal’s deaths are for a purpose. Farmland vegetarian diet killing is, to me, a callous by-product that foregoes direct purpose
Label encoding: <s>First off: Thank you for the very reasonable and well-thought response. I think, especially given sensitive subject matter like this, it is too easy to personally attack someone or become overly emotional about a subject when honest questions are posed; you did neither and I really appreciate it. [NEWLINE] [NEWLINE] Secondly : The article I linked to was less an avocation for its overall point, and more just a justification for pointing out that animals do die due to a vegetarian diet—something that I have overhead some evangelical vegetarian/vegans argue isn’t possible or is simply made up. You not only provided a delightful bit of rebuttal reading for tonight, but also acknowledge the intended fact without problems, and again I salute your reasonableness! [NEWLINE] [NEWLINE] Now, onto what I hope is further amicable discussion. [NEWLINE] [NEWLINE] [STARTQ] A philosophical point would be that there is (from my point of view, at least) an ethical difference between animal deaths that occur as a unintended secondary effect, and the ones that are deliberately and directly caused to happen. Essentially, it's a difference of intent; if I eat vegetables, I don't intend to cause the deaths of any animals, and I have no control over them. These deaths are certainly not inherent to eating vegetables; [ENDQ] [NEWLINE] If we assume KerSan’s point of “There is no such thing as ethical murder” then we are left both of us being nothing but simple killers for our food.  (He/she might very well be one as well given our discussion point, but that is a different discussion.) This gets both of us nowhere, so let’s focus on your point. [NEWLINE] [NEWLINE] I don’t see a philosophical difference between the intentional killing of animals for my food, and killing that is unintended yet a known byproduct of my diet. In both cases “your” food supply is marked by the death of animals, but in the omnivorous diet, at least those animals are useful. If you know that animals die specifically to supply “you” with food, you are morally equivalent. These deaths ARE inherent to eating vegetables, as much as death is inherent to eating steak. “You” just don’t have to be faced with that moral quandary directly at your butcher counter. [NEWLINE] [NEWLINE] For meat-eaters, at least their animal’s deaths are for a purpose. Farmland vegetarian diet killing is, to me, a callous by-product that foregoes direct purpose
Loss: tensor(0.2797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1998, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2532, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that people may be poor judges of their own happiness, purpose, etc. I<mask> agree that<mask> we feel at one time doesn't necessarily mean we'll continue to feel that way. We're irrational and imperfect.<mask> by this argument,<mask> should we ever be allowed to make a decision?<mask> can a judge sentence a criminal to imprisonment? Couldn't he change his mind later? Couldn't he be wrong? Yes,<mask> nonetheless we can only act<mask><mask> our *best* interpretation of things. And<mask> your *best* interpretation of the quality of your life is that it is not worth living, that is<mask> valid a determination<mask> can be made and that is the feeling on which you should act. (Another example:<mask> I decide I want to have a child and procreate, couldn't I change my mind in a few days and not want a child anymore? Couldn't I merely be misestimating the challenges parenthood will entail? Of course,<mask> it doesn't seem like you'd be willing to<mask><mask> I'm wrong or unjustified in choosing to be a parent<mask> my determination is that I want to be and am suited to be a parent). [NEWLINE] [NEWLINE] [NEWLINE] Suffering is only "useless"<mask> a concept in the sense that it is an imperfect concept. We make imperfect decisions, and they're based on imperfect analyses. "Good" and "evil" or "right" and "wrong" are<mask> imperfect ideas, impossible to fully define, and impossible to have actions fully assigned to one or the other.<mask> we still need to choose our actions based on some metric, even<mask> the metric is a bit fuzzy. Pleasure vs. suffering such a metric, and I see no reason<mask> it's not valid to consider. I can't *measure*<mask> much suffering I will cause someone by kicking their teeth in,<mask> I still know I shouldn't do it. [NEWLINE] [NEWLINE] [NEWLINE] I don't want to get too far off track here,<mask> regarding the necessity of simplifying philosophy: every academic discipline simplifies its subject matter to make modeling, discussing, and thinking about it possible. Is it a "made up game" to discuss DNA replication without naming every single bonding interaction between every molecule in DNA and transcriptase? Is it wrong to discuss classical or operant conditioning<mask> we don't know every neurotransmitter released in every synapse of the brain being conditioned? With philosophy, we cannot discuss every possible thing that makes life seem good or bad to every possible
Label encoding: <s>I agree that people may be poor judges of their own happiness, purpose, etc. I also agree that how we feel at one time doesn't necessarily mean we'll continue to feel that way. We're irrational and imperfect. But by this argument, why should we ever be allowed to make a decision? Why can a judge sentence a criminal to imprisonment? Couldn't he change his mind later? Couldn't he be wrong? Yes, but nonetheless we can only act according to our *best* interpretation of things. And if your *best* interpretation of the quality of your life is that it is not worth living, that is as valid a determination as can be made and that is the feeling on which you should act. (Another example: if I decide I want to have a child and procreate, couldn't I change my mind in a few days and not want a child anymore? Couldn't I merely be misestimating the challenges parenthood will entail? Of course, but it doesn't seem like you'd be willing to argue that I'm wrong or unjustified in choosing to be a parent if my determination is that I want to be and am suited to be a parent). [NEWLINE] [NEWLINE] [NEWLINE] Suffering is only "useless" as a concept in the sense that it is an imperfect concept. We make imperfect decisions, and they're based on imperfect analyses. "Good" and "evil" or "right" and "wrong" are also imperfect ideas, impossible to fully define, and impossible to have actions fully assigned to one or the other. But we still need to choose our actions based on some metric, even if the metric is a bit fuzzy. Pleasure vs. suffering such a metric, and I see no reason why it's not valid to consider. I can't *measure* how much suffering I will cause someone by kicking their teeth in, but I still know I shouldn't do it. [NEWLINE] [NEWLINE] [NEWLINE] I don't want to get too far off track here, but regarding the necessity of simplifying philosophy: every academic discipline simplifies its subject matter to make modeling, discussing, and thinking about it possible. Is it a "made up game" to discuss DNA replication without naming every single bonding interaction between every molecule in DNA and transcriptase? Is it wrong to discuss classical or operant conditioning because we don't know every neurotransmitter released in every synapse of the brain being conditioned? With philosophy, we cannot discuss every possible thing that makes life seem good or bad to every possible
Loss: tensor(0.2374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4573, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] And even then... it's data.<mask> do you stop it from being duplicated? [ENDQ] [NEWLINE] I apologize<mask> my comment is too technical,<mask> I will try to keep it simple. I am aware the learning curve is quite high for bitcoin, especially<mask> you are not computer-savvy. [NEWLINE] [NEWLINE] Bitcoin relies heavily on cryptography. Essentially this means you can take a string of letters of numbers and you can manipulate them 'in one direction' to give another string of letter or numbers, reproducibly.<mask> you cannot do it easily in the other direction, to the point<mask> its computationally next to impossible and probably forever will be. [NEWLINE] [NEWLINE] <mask> you download a bitcoin client that handles your bitcoin 'wallet' or you sign up for a free 'wallet' service online (eg blockchain.info) then you are usually provided with just one address. You can think of an address like a compartment in your real physical wallet or purse. It holds value (money) in the form of bitcoins, which are created and destroyed instantaneously in accordance with the public ledger (the blockchain) which others have described in other comments. [NEWLINE] [NEWLINE] Now, the 'address' I mention is actually whats called a public key. It is a cryptographic component that you can freely share with anybody, it is *NOT* a secret. [NEWLINE] [NEWLINE] <mask> an example, a randomly generated address I just made is: [NEWLINE] [NEWLINE] 1Axvy2YXdfantGNiS6oDVTgrSwoCkuJRXU [NEWLINE] [NEWLINE] Anybody can have your address, the public key, and they can't spend your bitcoins. [NEWLINE] [NEWLINE] In order for you to spend your bitcoins, you need another cryptographic component. You need to know the private key associated with that address. *Private keys are not stored in the blockchain (public ledger).* The private key should be known to only - and I really mean only - by the person who the bitcoins belong to. Thanks to some clever maths, you can always derive a full address (<mask> shown above) from the private key. [NEWLINE] [NEWLINE] It is the cryptographic difficulty in going from a public key - [STARTQ] private key that makes the ownership of that address secure. <mask> secure that even<mask> all the computers on earth were trying to 'crack' it, it would take millions of years for them to do<mask>. Going in the other direction, private key -&gt; public key, takes mere milliseconds and uses nil computational power. [ENDQ] [NEWLINE] <mask> you were wondering, the corresponding private key
Label encoding: <s> [STARTQ] And even then... it's data. How do you stop it from being duplicated? [ENDQ] [NEWLINE] I apologize if my comment is too technical, but I will try to keep it simple. I am aware the learning curve is quite high for bitcoin, especially if you are not computer-savvy. [NEWLINE] [NEWLINE] Bitcoin relies heavily on cryptography. Essentially this means you can take a string of letters of numbers and you can manipulate them 'in one direction' to give another string of letter or numbers, reproducibly. But you cannot do it easily in the other direction, to the point where its computationally next to impossible and probably forever will be. [NEWLINE] [NEWLINE] When you download a bitcoin client that handles your bitcoin 'wallet' or you sign up for a free 'wallet' service online (eg blockchain.info) then you are usually provided with just one address. You can think of an address like a compartment in your real physical wallet or purse. It holds value (money) in the form of bitcoins, which are created and destroyed instantaneously in accordance with the public ledger (the blockchain) which others have described in other comments. [NEWLINE] [NEWLINE] Now, the 'address' I mention is actually whats called a public key. It is a cryptographic component that you can freely share with anybody, it is *NOT* a secret. [NEWLINE] [NEWLINE] As an example, a randomly generated address I just made is: [NEWLINE] [NEWLINE] 1Axvy2YXdfantGNiS6oDVTgrSwoCkuJRXU [NEWLINE] [NEWLINE] Anybody can have your address, the public key, and they can't spend your bitcoins. [NEWLINE] [NEWLINE] In order for you to spend your bitcoins, you need another cryptographic component. You need to know the private key associated with that address. *Private keys are not stored in the blockchain (public ledger).* The private key should be known to only - and I really mean only - by the person who the bitcoins belong to. Thanks to some clever maths, you can always derive a full address ( as shown above) from the private key. [NEWLINE] [NEWLINE] It is the cryptographic difficulty in going from a public key - [STARTQ] private key that makes the ownership of that address secure.  So secure that even if all the computers on earth were trying to 'crack' it, it would take millions of years for them to do so. Going in the other direction, private key -&gt; public key, takes mere milliseconds and uses nil computational power. [ENDQ] [NEWLINE] If you were wondering, the corresponding private key
Loss: tensor(0.2914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3501, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Taxation, for better or worse, is attempting to get the maximum amount of milk for the minimum amount of moo.  The simple fact is that the government has bills to pay, and those bills are determined by the elected representatives.  Congress.  Specific governmental expenditures are made by people who lobby very hard to GET those particular government expenditures, and will fight like hell to KEEP them.  Cutting spending will create a LARGE amount of moo. <mask> that, the growth of government over the past decade has been at historic lows.  On the state and local level, many governments have actually SHRUNK. <mask>, shrinking the government more,<mask> a goal that many people have, is not realistic,<mask> the elected representatives will not get RE-elected<mask> they keep cutting. [NEWLINE] <mask> remember that government taxation is very much a form of incentive.  Incentive to work, incentive to invest, incentive to buy a house, incentive to not smoke, incentive to delay your retirement, all sorts of different incentives to do things that the government thinks are "good for you." [NEWLINE] Except they're NOT taxing you based on<mask>'s good for YOU.  They're taxing you based on<mask>'s good for SOCIETY. <mask> should you pay taxes to pay for public schools<mask> you don't have kids? <mask> society benefits greatly from having a well-educated workforce. [NEWLINE] <mask> should you pay for military spending<mask> you think we're involved in too many wars? <mask> the government is working in the best interests of the society, of the companies that do business overseas, and<mask> it benefits world peace and stability<mask> there is a world policeman to deal with the rogue countries out there. [NEWLINE] I'm not saying<mask><mask> with all government expenditures.  Far from it. <mask> SOMEBODY made a big stick to get that expenditure made in the first place. [NEWLINE] The average cost of government spending per person in the US is something like $16,000 a year per person. [NEWLINE] <mask> do we divvy up the tax bill? [NEWLINE] Equally?  Okay, then a family of four owes $64,000 a year.  Period.  The average family of four owes ten grand more in taxes than the average household makes in a year.  And forget about having kids. [NEWLINE] Divvied up among those who have jobs?  It gets even worse.  It jumps up to around $35,000 a taxpayer.  Two income household?  Congrats
Label encoding: <s>Taxation, for better or worse, is attempting to get the maximum amount of milk for the minimum amount of moo.  The simple fact is that the government has bills to pay, and those bills are determined by the elected representatives.  Congress.  Specific governmental expenditures are made by people who lobby very hard to GET those particular government expenditures, and will fight like hell to KEEP them.  Cutting spending will create a LARGE amount of moo.  Despite that, the growth of government over the past decade has been at historic lows.  On the state and local level, many governments have actually SHRUNK.  So, shrinking the government more, while a goal that many people have, is not realistic, as the elected representatives will not get RE-elected if they keep cutting. [NEWLINE] Also remember that government taxation is very much a form of incentive.  Incentive to work, incentive to invest, incentive to buy a house, incentive to not smoke, incentive to delay your retirement, all sorts of different incentives to do things that the government thinks are "good for you." [NEWLINE] Except they're NOT taxing you based on what's good for YOU.  They're taxing you based on what's good for SOCIETY.  Why should you pay taxes to pay for public schools when you don't have kids?  Because society benefits greatly from having a well-educated workforce. [NEWLINE] Why should you pay for military spending when you think we're involved in too many wars?  Because the government is working in the best interests of the society, of the companies that do business overseas, and because it benefits world peace and stability when there is a world policeman to deal with the rogue countries out there. [NEWLINE] I'm not saying I agree with all government expenditures.  Far from it.  But SOMEBODY made a big stick to get that expenditure made in the first place. [NEWLINE] The average cost of government spending per person in the US is something like $16,000 a year per person. [NEWLINE] How do we divvy up the tax bill? [NEWLINE] Equally?  Okay, then a family of four owes $64,000 a year.  Period.  The average family of four owes ten grand more in taxes than the average household makes in a year.  And forget about having kids. [NEWLINE] Divvied up among those who have jobs?  It gets even worse.  It jumps up to around $35,000 a taxpayer.  Two income household?  Congrats
Loss: tensor(0.2437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3979, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2905, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1964, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I want to start by pointing out that his points are broad strokes,<mask> pointing out a counterexample isn't really sufficient to disprove it; you'd need to show that it's false in a large number of cases. Further,<mask> you're trying to show that the left does these things, you're probably going to make better cases by pointing to instances of the left doing it *<mask> they're currently in power*. The metaphor of the zombie apocalypse for the virtues of conservatism is an even better metaphor for the actions of desperate people,<mask> showing that leftists do these things<mask> their values are threatened doesn't really prove anything. You want to show that these are the virtues that the left *aspires to*. That's the whole point of the exercise; to show<mask> each ideology *aspires to*. [NEWLINE] [NEWLINE] [STARTQ] Gun control is not a decisive issue between the kind of right and left Scott means. The revolutionary left usually thinks guns in the hands of the people are a good idea. [ENDQ] [NEWLINE] Are you really trying to<mask><mask> (at least in America) gun rights are broadly supported by the right and not by the left? I mean, I know a lot of (non-revolutionary) liberals who don't support strict gun regulations,<mask> really?<mask>, revolutionary *anyone* thinks guns are a good thing<mask> they see them<mask> a means to their goals. I see this<mask> different from thinking guns are an end in and of themselves. See point 1. [NEWLINE] [NEWLINE] [STARTQ] That doesn't follow from the praying. [ENDQ] [NEWLINE] No,<mask> take someone who is already *very* religious. Place them in a disaster. Many of them will think<mask><mask><mask><mask><mask><mask> people doing things that offend god. You may think this is crazy (<mask> it is)<mask> *politicians* already do this every time there's a major *hurricane*! Increased religion will cause more of this kind of thing. [NEWLINE] [NEWLINE] [STARTQ] I The left is extremely suspicious of outsiders. That's<mask> Gulags are for. They just don't define outsider by skin colour or privation,<mask> ideology. [ENDQ] [NEWLINE] Distrusting people for having a different ideology is different for distrusting them for having a different skin color. One is making an assessment of a person based on qualities that will likely affect<mask> you interact with them.<mask><mask> the far left should be more tolerant of other viewpoints,<mask> that's true of any "far" ideological standpoint.<mask>,<mask><mask> it's hard to describe Stalinist Russia<mask> being
Label encoding: <s>I want to start by pointing out that his points are broad strokes, so pointing out a counterexample isn't really sufficient to disprove it; you'd need to show that it's false in a large number of cases. Further, if you're trying to show that the left does these things, you're probably going to make better cases by pointing to instances of the left doing it * when they're currently in power*. The metaphor of the zombie apocalypse for the virtues of conservatism is an even better metaphor for the actions of desperate people, so showing that leftists do these things when their values are threatened doesn't really prove anything. You want to show that these are the virtues that the left *aspires to*. That's the whole point of the exercise; to show what each ideology *aspires to*. [NEWLINE] [NEWLINE] [STARTQ] Gun control is not a decisive issue between the kind of right and left Scott means. The revolutionary left usually thinks guns in the hands of the people are a good idea. [ENDQ] [NEWLINE] Are you really trying to argue that (at least in America) gun rights are broadly supported by the right and not by the left? I mean, I know a lot of (non-revolutionary) liberals who don't support strict gun regulations, but really? Also, revolutionary *anyone* thinks guns are a good thing because they see them as a means to their goals. I see this as different from thinking guns are an end in and of themselves. See point 1. [NEWLINE] [NEWLINE] [STARTQ] That doesn't follow from the praying. [ENDQ] [NEWLINE] No, but take someone who is already *very* religious. Place them in a disaster. Many of them will think the reason is the reason is people doing things that offend god. You may think this is crazy ( because it is) but *politicians* already do this every time there's a major *hurricane*! Increased religion will cause more of this kind of thing. [NEWLINE] [NEWLINE] [STARTQ] I The left is extremely suspicious of outsiders. That's what Gulags are for. They just don't define outsider by skin colour or privation, but ideology. [ENDQ] [NEWLINE] Distrusting people for having a different ideology is different for distrusting them for having a different skin color. One is making an assessment of a person based on qualities that will likely affect how you interact with them. I think the far left should be more tolerant of other viewpoints, but that's true of any "far" ideological standpoint. Also, I think it's hard to describe Stalinist Russia as being
Loss: tensor(0.3397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2744, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is not correct. In 1880 the Zionist movement started, with the stated mission of creating a Jewish homeland in Israel. At the time, the Palestine mandate was a part of the Ottoman Empire and was basically a wasteland, except for some orange and olive farming.<mask> Jews started moving there, they bought the land from the Ottomans, and brought modern farming technniques with them. These techniques brought huge areas in the country to agriculture, leading to increased food supplies for everyone. [NEWLINE] [NEWLINE] Not only did the Jews bring in farming techniques,<mask> they<mask> brought modern medicine, schooling, and trade.  The native Arab population grew due to access to these facilities,<mask> they were far out-numbered by the Arab immmigrants coming from the surrounding areas looking for job opportunities and access to the modern world. [NEWLINE] [NEWLINE] Now, after WWI, the Ottoman Empire falls, and both Jews and Arabs are under British control.  The British issue the Balfour Declaration, saying they favor the establishing of a Jewish state in Palestine,<mask> Arab leaders were not happy with this, saying that their efforts should have earned them a nation of their own.  This leads to an increase in violence in the region, and, eventually, the British split the Palestine mandate into two parts, Palestine and Jordan. Jordan is meant to be the nation for the Arab population, and Palestine the Jews. [NEWLINE] [NEWLINE] This solution doesn't work,<mask> Palestinians were left out of the leadership in Jordan in favor of a pro-England leader, and for a few other reasons, namely Jordan was mostly untouched by (Jewish) efforts at modernization. Violence continues to escalate. [NEWLINE] [NEWLINE] Now we come to WWII, and the Holocaust.  By this point, in an attempt to assuage the Arab population, the Brits issue a series of White Papers which limit the immigration of Jews to Palestine. This couldn't have happened at a worse time,<mask> now millions of Jews are stuck in Europe and exposed to the horrors of<mask> followed.  This led to a huge amount of resentment against British rule, and massive, illegal immigration of Jews attempting to escape the violence. The Arabs in Palestine are now resentful,<mask> they believe that the Brits are helping the Jews to continue the immigration policies, depsite their promises. <mask> the Brits are stuck in between 2 increasingly hostile factions, both of whom feel wronged. [NEWLINE] [NEWLINE] Following WWII, you find a Britain losing its empire piece by piece. Fed up with trying to maintain peace among two groups using violence
Label encoding: <s>This is not correct. In 1880 the Zionist movement started, with the stated mission of creating a Jewish homeland in Israel. At the time, the Palestine mandate was a part of the Ottoman Empire and was basically a wasteland, except for some orange and olive farming. When Jews started moving there, they bought the land from the Ottomans, and brought modern farming technniques with them. These techniques brought huge areas in the country to agriculture, leading to increased food supplies for everyone. [NEWLINE] [NEWLINE] Not only did the Jews bring in farming techniques, but they also brought modern medicine, schooling, and trade.  The native Arab population grew due to access to these facilities, but they were far out-numbered by the Arab immmigrants coming from the surrounding areas looking for job opportunities and access to the modern world. [NEWLINE] [NEWLINE] Now, after WWI, the Ottoman Empire falls, and both Jews and Arabs are under British control.  The British issue the Balfour Declaration, saying they favor the establishing of a Jewish state in Palestine, but Arab leaders were not happy with this, saying that their efforts should have earned them a nation of their own.  This leads to an increase in violence in the region, and, eventually, the British split the Palestine mandate into two parts, Palestine and Jordan. Jordan is meant to be the nation for the Arab population, and Palestine the Jews. [NEWLINE] [NEWLINE] This solution doesn't work, because Palestinians were left out of the leadership in Jordan in favor of a pro-England leader, and for a few other reasons, namely Jordan was mostly untouched by (Jewish) efforts at modernization. Violence continues to escalate. [NEWLINE] [NEWLINE] Now we come to WWII, and the Holocaust.  By this point, in an attempt to assuage the Arab population, the Brits issue a series of White Papers which limit the immigration of Jews to Palestine. This couldn't have happened at a worse time, as now millions of Jews are stuck in Europe and exposed to the horrors of what followed.  This led to a huge amount of resentment against British rule, and massive, illegal immigration of Jews attempting to escape the violence. The Arabs in Palestine are now resentful, because they believe that the Brits are helping the Jews to continue the immigration policies, depsite their promises.  So the Brits are stuck in between 2 increasingly hostile factions, both of whom feel wronged. [NEWLINE] [NEWLINE] Following WWII, you find a Britain losing its empire piece by piece. Fed up with trying to maintain peace among two groups using violence
Loss: tensor(0.2092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2662, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 2) There are businesses everywhere, and all of them are at the cost of higher overhead and more operating expenses than 99% of North America. I don't see a wealth of opportunity, I see a shitload of expensive investments that have to be extremely revenue-rich to be profitable. [ENDQ] [NEWLINE] I couldn't disagree more.  [NYC has the highest GDP in the US]( [URL] ) at $1.35BN.  That's *almost twice<mask> high*<mask> the second largest city, Los Angeles at $765BN. <mask><mask> rent in Boston (for comparison) might be 30% cheaper it is<mask> an economy that has $336BN, or *1/4* of NYCs GDP.  With all due respect to anyone out there, *it's just not even close*. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 3)<mask><mask> you want to drive? Own a horse? Sure, there's lots of stuff you can do in NYC,<mask> it's not significantly better than any other large American city (&gt;5M in population) in that regard. [ENDQ] [NEWLINE] My argument was that NYC is a good city.  It *was not* that everything is better in NYC.  You *can* go horseback riding in several places in NYC. <mask><mask> you prefer a place in the country, then you should.  I'm only saying that these things are viable options. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 4) Restaurants may have been true 20 years ago; now it's a tired talking point that doesn't really hold true. Just about every major metropolitan area in the US has a wealth of cultural diversity and,<mask><mask><mask>, representative foods and cuisine that are considered world-class. [ENDQ] [NEWLINE] I don't even see<mask> that's remotely accurate.  Whenever I leave the city I am reminded that I can't just go out for a 2:00 AM burger<mask> most places are closed. <mask>, just<mask><mask><mask> of having fewer people it would be just about impossible to have this much variety. [NEWLINE] [NEWLINE] [STARTQ] 5) This is true,<mask> again, most large cities include some measure of diversity;<mask> is that a selling point, anyway? You mention that NYC is diverse,<mask> I don't see that<mask> a strict positive. NYC's extreme diversity often leads to crime and civil unrest; the diversity of the city is a double-edged sword. [ENDQ] [NEWLINE] [NEWLINE] NYC has record low crime rates.  They've been dropping for about 15 years. <mask> it *usually
Label encoding: <s> [STARTQ] 2) There are businesses everywhere, and all of them are at the cost of higher overhead and more operating expenses than 99% of North America. I don't see a wealth of opportunity, I see a shitload of expensive investments that have to be extremely revenue-rich to be profitable. [ENDQ] [NEWLINE] I couldn't disagree more.  [NYC has the highest GDP in the US]( [URL] ) at $1.35BN.  That's *almost twice as high* as the second largest city, Los Angeles at $765BN.  So while rent in Boston (for comparison) might be 30% cheaper it is also an economy that has $336BN, or *1/4* of NYCs GDP.  With all due respect to anyone out there, *it's just not even close*. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 3) What if you want to drive? Own a horse? Sure, there's lots of stuff you can do in NYC, but it's not significantly better than any other large American city (&gt;5M in population) in that regard. [ENDQ] [NEWLINE] My argument was that NYC is a good city.  It *was not* that everything is better in NYC.  You *can* go horseback riding in several places in NYC.  But if you prefer a place in the country, then you should.  I'm only saying that these things are viable options. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 4) Restaurants may have been true 20 years ago; now it's a tired talking point that doesn't really hold true. Just about every major metropolitan area in the US has a wealth of cultural diversity and, as a result, representative foods and cuisine that are considered world-class. [ENDQ] [NEWLINE] I don't even see how that's remotely accurate.  Whenever I leave the city I am reminded that I can't just go out for a 2:00 AM burger because most places are closed.  Also, just as a result of having fewer people it would be just about impossible to have this much variety. [NEWLINE] [NEWLINE] [STARTQ] 5) This is true, but again, most large cities include some measure of diversity; why is that a selling point, anyway? You mention that NYC is diverse, but I don't see that as a strict positive. NYC's extreme diversity often leads to crime and civil unrest; the diversity of the city is a double-edged sword. [ENDQ] [NEWLINE] [NEWLINE] NYC has record low crime rates.  They've been dropping for about 15 years.  What it *usually
Loss: tensor(0.2922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.9160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your excuse for misogynistic language is that it's especially effective at causing emotional harm to women? Am I reading that correctly? [ENDQ] [NEWLINE] I am asking that you be consistent and follow your logic to it's end. Self-censorship is still censorship. Do you think all insults should be censored? Do you think all classifications that could be seen<mask> derogatory should be censored? Do you think (self)censorship will ultimately prevent me and the rest of society from identifying who is different from my prefered group and pointing that difference out to someone who is more akin to me? The power does not lie in the words I use, nigger, faggot, idiot are just old classifications that was turned into insults. Black people, gays and less-abled people still exist and one day those classifications will be seen<mask> insulting too<mask> they differentiate someone from the group. [NEWLINE] [NEWLINE] More likely you want special treatment for "minority groups"<mask> reserving the right to deem which minority groups are minority groups<mask> giving the majority group a classification to use on the minority group and defeating your own purpose (<mask> you have not thought this through). [NEWLINE] [NEWLINE] [STARTQ] Whether you think<mask> or not, such behaviour<mask> you've described is bigoted, and is a major part of the reason<mask> videogames are taken seriously by<mask> few people, which upsets me. The culture is like an oroboros. It demands to be taken seriously<mask> then act in a way that precludes from being taken seriously. [ENDQ] [NEWLINE] No it is not, I just want to play games with people who want to play games. Turn another medium into a shitty sjw playground, I do not care about gender, sex or skincolor (get enough of that bullshit the rest of the day). I'm not saying you should leave,<mask> turn the ideology down, feminists/sjws are like the jehovas of the internet. [NEWLINE] [NEWLINE] [STARTQ] It's a shame for all of us who actually care about games<mask> a medium of human and artistic expression. All the social capital collected is regularly spent on useless diatribes about<mask> women are "ruining videogames" and, for lack of a better way to put it, an ongoing battle for games to remain a boy's club. [ENDQ] [NEWLINE] <mask><mask>, the people complaining should stfu and settle the dispute by playing videogames<mask> we have always done. Whining about<mask> girls aren't welcome,<mask> you identify with whatever arbitrary shit you identify with that much then
Label encoding: <s> [STARTQ] Your excuse for misogynistic language is that it's especially effective at causing emotional harm to women? Am I reading that correctly? [ENDQ] [NEWLINE] I am asking that you be consistent and follow your logic to it's end. Self-censorship is still censorship. Do you think all insults should be censored? Do you think all classifications that could be seen as derogatory should be censored? Do you think (self)censorship will ultimately prevent me and the rest of society from identifying who is different from my prefered group and pointing that difference out to someone who is more akin to me? The power does not lie in the words I use, nigger, faggot, idiot are just old classifications that was turned into insults. Black people, gays and less-abled people still exist and one day those classifications will be seen as insulting too because they differentiate someone from the group. [NEWLINE] [NEWLINE] More likely you want special treatment for "minority groups" while reserving the right to deem which minority groups are minority groups thus giving the majority group a classification to use on the minority group and defeating your own purpose ( because you have not thought this through). [NEWLINE] [NEWLINE] [STARTQ] Whether you think so or not, such behaviour as you've described is bigoted, and is a major part of the reason why videogames are taken seriously by so few people, which upsets me. The culture is like an oroboros. It demands to be taken seriously but then act in a way that precludes from being taken seriously. [ENDQ] [NEWLINE] No it is not, I just want to play games with people who want to play games. Turn another medium into a shitty sjw playground, I do not care about gender, sex or skincolor (get enough of that bullshit the rest of the day). I'm not saying you should leave, but turn the ideology down, feminists/sjws are like the jehovas of the internet. [NEWLINE] [NEWLINE] [STARTQ] It's a shame for all of us who actually care about games as a medium of human and artistic expression. All the social capital collected is regularly spent on useless diatribes about how women are "ruining videogames" and, for lack of a better way to put it, an ongoing battle for games to remain a boy's club. [ENDQ] [NEWLINE] I agree, the people complaining should stfu and settle the dispute by playing videogames as we have always done. Whining about how girls aren't welcome, if you identify with whatever arbitrary shit you identify with that much then
Loss: tensor(0.2395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1699, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**based on the response I got from Subtle150,<mask><mask> I should probably preface this opinion with the absolutely sincere declaration that I am 100% open to any argument against<mask> I say here, and I'm totally open to having my mind changed about this topic. I've put a lot of thought in to this,<mask> it means a lot to me, and laid it all out here in the most detailed way I could, not to be pretentious,<mask> to express everything<mask><mask> about the topic. Feel free to pick apart anything you think is wrong with it, I am open to it** [NEWLINE] [NEWLINE] Belief in God(s) is ~~fairly useless~~ incredibly illogical, and is not necessary to provide the benefits attributed to belief in it(them) (Thanks Failedhall for pointing out the original was kind of arrogant). It is absolutely impossible to prove or disprove the existence of God and none of the defenses people give for believing in God without evidence have any grounding. [NEWLINE] [NEWLINE] No one has been able to, can now, or ever will be able to prove or disprove the existence of God. (<mask>, neither can anyone prove or disprove the existence of Santa Clause, unicorns, leprechauns, the tooth fairy, Zeus, Thor, Huitzilopochtli, daemons, witches or anything else of the sort - common sense helps us to weigh in on the existence of those beings.) [NEWLINE] [NEWLINE] <mask> even<mask> you gift to theists the claim that God(s) exists, I would argue they gain nothing of value anyway. [NEWLINE] [NEWLINE] Let us say that it is absolutely, irrefutably, true that there is an omnipotent, omniscient, benevolent, creator of the universe (or multiple). You still can't move on from this claim to anything useful. [NEWLINE] [NEWLINE] Merely conceding that God(s) exists only gives a proprietor and instigator to the Universe and Natural Laws. To do anything further, you would have to<mask><mask> human beings can interpret the will and function of God(s). [NEWLINE] [NEWLINE] <mask> could (decidedly) finite beings such<mask> ourselves ever even hope to discern the will or means of an infinite being(s)? [NEWLINE] [NEWLINE] Let us even say that God(s) exists and that it(they) is(are) the embodiment and command of everything that is good, just, and right in this Universe. You are **still** left with nothing. We still could not post
Label encoding: <s>**based on the response I got from Subtle150, I think I should probably preface this opinion with the absolutely sincere declaration that I am 100% open to any argument against what I say here, and I'm totally open to having my mind changed about this topic. I've put a lot of thought in to this, because it means a lot to me, and laid it all out here in the most detailed way I could, not to be pretentious, but to express everything I think about the topic. Feel free to pick apart anything you think is wrong with it, I am open to it** [NEWLINE] [NEWLINE] Belief in God(s) is ~~fairly useless~~ incredibly illogical, and is not necessary to provide the benefits attributed to belief in it(them) (Thanks Failedhall for pointing out the original was kind of arrogant). It is absolutely impossible to prove or disprove the existence of God and none of the defenses people give for believing in God without evidence have any grounding. [NEWLINE] [NEWLINE] No one has been able to, can now, or ever will be able to prove or disprove the existence of God. ( However, neither can anyone prove or disprove the existence of Santa Clause, unicorns, leprechauns, the tooth fairy, Zeus, Thor, Huitzilopochtli, daemons, witches or anything else of the sort - common sense helps us to weigh in on the existence of those beings.) [NEWLINE] [NEWLINE] But even if you gift to theists the claim that God(s) exists, I would argue they gain nothing of value anyway. [NEWLINE] [NEWLINE] Let us say that it is absolutely, irrefutably, true that there is an omnipotent, omniscient, benevolent, creator of the universe (or multiple). You still can't move on from this claim to anything useful. [NEWLINE] [NEWLINE] Merely conceding that God(s) exists only gives a proprietor and instigator to the Universe and Natural Laws. To do anything further, you would have to argue that human beings can interpret the will and function of God(s). [NEWLINE] [NEWLINE] How could (decidedly) finite beings such as ourselves ever even hope to discern the will or means of an infinite being(s)? [NEWLINE] [NEWLINE] Let us even say that God(s) exists and that it(they) is(are) the embodiment and command of everything that is good, just, and right in this Universe. You are **still** left with nothing. We still could not post
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I'm gonna open this one with a ∆<mask> you've complexified my understanding of the Enlightenment. Thanks for taking the time to formulate this thoughtful response. You've given me a lot to think about. [NEWLINE] [NEWLINE] [STARTQ] Okay,<mask><mask> this is "the logic of modern capitalism",<mask> does "logic" mean? I know<mask> it means<mask> you say "unfettered exchange is the logic of capitalism"; you mean something like the reason capitalism functions the way it does, sustains itself, and has particular pros and cons that differentiate it from other economic systems are all connected through the way that free exchange functions, and its conceptual pros and cons. [ENDQ] [NEWLINE] I could have been more specific. One purpose of modern capitalism is the creation of surplus value and the concentration of that surplus into the hands of investors(<mask> opposed to, say, subsistence farming). Production is only viable<mask> it will make a profit. Free labour means an exponential increase in profits. The original iteration of capitalism relied primarily on free labour.<mask> did other modes of production which pre-date capitalism, like you state. My argument wasn't that modernity was singularly evil, just evil (or more accurately, **reliant on unfree labour regimes<mask> lauding itself<mask> freedom incarnate**). [NEWLINE] [NEWLINE] [STARTQ] <mask> Adorno's conclusion wasn't that Enlightenment was evil, it was that we need to push forward to a fairer world or the unhappiness our half-breed culture - rational enough to make us want justice,<mask> not rational enough to deliver it to us - will continue to produce horrors like the Second World War. [ENDQ] [NEWLINE] <mask><mask> with Adorno.<mask><mask> that pushing forward to a fairer world doesn't necessitate liberal norms and I'm not convinced they are particularly useful in producing equality. They are not rational enough to deliver justice. I didn't mean to attack enlightenment thinkers<mask> to make us question the dominant moral position that liberalism occupies<mask> the skeletons in its closet. I'm saying that, due to its violent birth and the continuing violence which it perpetuates, liberalism (the project of modernity, related to<mask> not encapsulated entirely by the term 'the Enlightenment') ought not be lauded<mask> the epitome of moral governance. [NEWLINE] [NEWLINE] [STARTQ] <mask> for the Enlightenment, it's a calumny against the Enlightenment to claim that scientific racism was a project of the Enlightenment (it came later) or that it was ever a central part of Western philosophy or scientific thought. [ENDQ] [NEWLINE] Keeping in mind that my target was modern
Label encoding: <s>Well I'm gonna open this one with a ∆ because you've complexified my understanding of the Enlightenment. Thanks for taking the time to formulate this thoughtful response. You've given me a lot to think about. [NEWLINE] [NEWLINE] [STARTQ] Okay, so if this is "the logic of modern capitalism", what does "logic" mean? I know what it means if you say "unfettered exchange is the logic of capitalism"; you mean something like the reason capitalism functions the way it does, sustains itself, and has particular pros and cons that differentiate it from other economic systems are all connected through the way that free exchange functions, and its conceptual pros and cons. [ENDQ] [NEWLINE] I could have been more specific. One purpose of modern capitalism is the creation of surplus value and the concentration of that surplus into the hands of investors( as opposed to, say, subsistence farming). Production is only viable if it will make a profit. Free labour means an exponential increase in profits. The original iteration of capitalism relied primarily on free labour. As did other modes of production which pre-date capitalism, like you state. My argument wasn't that modernity was singularly evil, just evil (or more accurately, **reliant on unfree labour regimes while lauding itself as freedom incarnate**). [NEWLINE] [NEWLINE] [STARTQ] But Adorno's conclusion wasn't that Enlightenment was evil, it was that we need to push forward to a fairer world or the unhappiness our half-breed culture - rational enough to make us want justice, but not rational enough to deliver it to us - will continue to produce horrors like the Second World War. [ENDQ] [NEWLINE] I agree with Adorno. I think that pushing forward to a fairer world doesn't necessitate liberal norms and I'm not convinced they are particularly useful in producing equality. They are not rational enough to deliver justice. I didn't mean to attack enlightenment thinkers but to make us question the dominant moral position that liberalism occupies despite the skeletons in its closet. I'm saying that, due to its violent birth and the continuing violence which it perpetuates, liberalism (the project of modernity, related to but not encapsulated entirely by the term 'the Enlightenment') ought not be lauded as the epitome of moral governance. [NEWLINE] [NEWLINE] [STARTQ] As for the Enlightenment, it's a calumny against the Enlightenment to claim that scientific racism was a project of the Enlightenment (it came later) or that it was ever a central part of Western philosophy or scientific thought. [ENDQ] [NEWLINE] Keeping in mind that my target was modern
Loss: tensor(0.1883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1921, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Nobody is afraid of women.<mask> I go out late at night, there has to be a trusted man with me. There are rapists at night, or<mask> I get into an accident, who's going to change my tire? [ENDQ] [NEWLINE] This is just bizarre.<mask> you don't learn to change a tire it's no ones fault<mask> yourself. Public approval is not necessary in order for you to learn this. Not only that,<mask> random night rapists are not the most common for of rape. Males who go out at night have to worry about being mugged. Sure, that's a lesser issue,<mask> it's not like they have no fears.. [NEWLINE] [NEWLINE] [STARTQ] <mask> I sleep around, I am a slut.<mask> a man sleeps around, he is a stud. [ENDQ] [NEWLINE] You are willfully choosing to ignore the fact that the reverse is<mask> true.<mask> a female doesn't have much sexual experience it's a choice.<mask> a male doesn't it means they can't, and attempts to assert otherwise make them even more pathetic. Not only that,<mask> many people<mask> hate males for sleeping around. The demonization of females who do does not even remotely approach the fact that males get judged either way. [NEWLINE] [NEWLINE] [STARTQ] No one gets very angry<mask> a man abandons the baby he fathered. [ENDQ] [NEWLINE] I love<mask> you literally stated something exactly opposite of the truth, and<mask> act like it's a point. Most males are extremely demonized at the idea of them being a "deadbeat dad."<mask><mask> females have the get out of jail free card that's arguably a lot worse,<mask> less people are against. [NEWLINE] [NEWLINE] [STARTQ] Men have freedom of opportunity. There are still many lucrative professions that are anti-women. Tech and Engineering are the most popular ones. [ENDQ] [NEWLINE] You do know that the people who complain about females not being in these are the ones who keep them out of them, right? The type of people who legitimately think being "female" is<mask> bad<mask> being black in this day and age are mostly postmodernists. They do not value objectivity very highly, and<mask> anyone associated with their ideas will be demonized in fields<mask> that is important. Sure, there's<mask> the lonely virgin engineer stereotype,<mask> this person being like this is hardly a benefit for them,<mask> it's an indication that they're bitter<mask> their life is shitty. [NEWLINE] [NEWLINE] [STARTQ] Men aren't objectified.<mask> I go anywhere looking fashionable and attractive, I will get harassed.<mask> a man goes out
Label encoding: <s> [STARTQ] Nobody is afraid of women. If I go out late at night, there has to be a trusted man with me. There are rapists at night, or if I get into an accident, who's going to change my tire? [ENDQ] [NEWLINE] This is just bizarre. If you don't learn to change a tire it's no ones fault but yourself. Public approval is not necessary in order for you to learn this. Not only that, but random night rapists are not the most common for of rape. Males who go out at night have to worry about being mugged. Sure, that's a lesser issue, but it's not like they have no fears.. [NEWLINE] [NEWLINE] [STARTQ] If I sleep around, I am a slut. If a man sleeps around, he is a stud. [ENDQ] [NEWLINE] You are willfully choosing to ignore the fact that the reverse is also true. If a female doesn't have much sexual experience it's a choice. If a male doesn't it means they can't, and attempts to assert otherwise make them even more pathetic. Not only that, but many people also hate males for sleeping around. The demonization of females who do does not even remotely approach the fact that males get judged either way. [NEWLINE] [NEWLINE] [STARTQ] No one gets very angry if a man abandons the baby he fathered. [ENDQ] [NEWLINE] I love how you literally stated something exactly opposite of the truth, and yet act like it's a point. Most males are extremely demonized at the idea of them being a "deadbeat dad." Where as females have the get out of jail free card that's arguably a lot worse, but less people are against. [NEWLINE] [NEWLINE] [STARTQ] Men have freedom of opportunity. There are still many lucrative professions that are anti-women. Tech and Engineering are the most popular ones. [ENDQ] [NEWLINE] You do know that the people who complain about females not being in these are the ones who keep them out of them, right? The type of people who legitimately think being "female" is as bad as being black in this day and age are mostly postmodernists. They do not value objectivity very highly, and so anyone associated with their ideas will be demonized in fields where that is important. Sure, there's also the lonely virgin engineer stereotype, but this person being like this is hardly a benefit for them, since it's an indication that they're bitter since their life is shitty. [NEWLINE] [NEWLINE] [STARTQ] Men aren't objectified. If I go anywhere looking fashionable and attractive, I will get harassed. If a man goes out
Loss: tensor(0.2639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.9551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.8721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2756, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Who pays for the schools and roads and police now? TAXPAYERS. The government doesn't actually pay for anything. It cannot create wealth, it can only print money, collect taxes, and borrow. [ENDQ] [NEWLINE] Tax payers pay the government, who then pays for schools and roads. [NEWLINE] [NEWLINE] [STARTQ] <mask> most people think public schools are a good idea? let them contribute to it voluntarily. Don't like public schools, well then you don't have to support them. [ENDQ] [NEWLINE] The problem comes<mask> plenty of people LIKE public schools,<mask> don't contribute<mask> they either *can't* or just *don't want to*. The amount of money needed to pay for public schools far exceeds<mask> people would willingly give.  The system would collapse. [NEWLINE] [NEWLINE] [STARTQ] Perhaps without the government, we wouldn't have<mask> many roads that aren't economically viable - IE they don't help enough business to make them viable. In a voluntary society, that road would not be built, rather than having taxpayers be forced to waste money on it. [ENDQ] [NEWLINE] Roads are extremely expensive, especially interstate highways.  Someone else has already made the counterargument for this in this thread. [NEWLINE] [NEWLINE] [STARTQ] The free market absolutely has the interests of the people in mind.<mask> you want to form a successful business, you MUST give customers<mask> they want.<mask> you sell a product to a consumer and they give you money for it voluntarily, it is win-win. [ENDQ] [NEWLINE] It has the greater interests of its *customers* in mind, sure.  Everyone else can go fuck themselves.  For example, without government regulations on the amount of pollution a factory can generate, we'll end up with the pollution problems that China is dealing with right now.  Shitty working conditions at a Foxconn factory were<mask> bad that people are committing suicide.  Foxconn makes parts for iPhones and other things, and<mask> people still flock to the Apple store to get their new iPhone.  People don't give two shits about the harm manufacturers do without government regulations.  A perfectly free market can benefit customers, sure,<mask> at<mask> cost? [NEWLINE] [NEWLINE] The government doesn't have to produce a profit, or be efficient,<mask> they don't have to make money. They just take it. The government has zero incentive to produce a cost-efficient product. [NEWLINE] [NEWLINE] [STARTQ] The government doesn't have to produce a profit, or be efficient,<mask> they don't have to make money. They just take it. The government has zero incentive to
Label encoding: <s> [STARTQ] Who pays for the schools and roads and police now? TAXPAYERS. The government doesn't actually pay for anything. It cannot create wealth, it can only print money, collect taxes, and borrow. [ENDQ] [NEWLINE] Tax payers pay the government, who then pays for schools and roads. [NEWLINE] [NEWLINE] [STARTQ] So most people think public schools are a good idea? let them contribute to it voluntarily. Don't like public schools, well then you don't have to support them. [ENDQ] [NEWLINE] The problem comes when plenty of people LIKE public schools, but don't contribute because they either *can't* or just *don't want to*. The amount of money needed to pay for public schools far exceeds what people would willingly give.  The system would collapse. [NEWLINE] [NEWLINE] [STARTQ] Perhaps without the government, we wouldn't have as many roads that aren't economically viable - IE they don't help enough business to make them viable. In a voluntary society, that road would not be built, rather than having taxpayers be forced to waste money on it. [ENDQ] [NEWLINE] Roads are extremely expensive, especially interstate highways.  Someone else has already made the counterargument for this in this thread. [NEWLINE] [NEWLINE] [STARTQ] The free market absolutely has the interests of the people in mind. If you want to form a successful business, you MUST give customers what they want. When you sell a product to a consumer and they give you money for it voluntarily, it is win-win. [ENDQ] [NEWLINE] It has the greater interests of its *customers* in mind, sure.  Everyone else can go fuck themselves.  For example, without government regulations on the amount of pollution a factory can generate, we'll end up with the pollution problems that China is dealing with right now.  Shitty working conditions at a Foxconn factory were so bad that people are committing suicide.  Foxconn makes parts for iPhones and other things, and yet people still flock to the Apple store to get their new iPhone.  People don't give two shits about the harm manufacturers do without government regulations.  A perfectly free market can benefit customers, sure, but at what cost? [NEWLINE] [NEWLINE] The government doesn't have to produce a profit, or be efficient, because they don't have to make money. They just take it. The government has zero incentive to produce a cost-efficient product. [NEWLINE] [NEWLINE] [STARTQ] The government doesn't have to produce a profit, or be efficient, because they don't have to make money. They just take it. The government has zero incentive to
Loss: tensor(0.2158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1972, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] To stretch your analogy,<mask>,<mask> you hire someone to clean toilets and then tell him to fight bears instead threatening to withhold his livelihood<mask> he does not, you absolutely are engaging in financial coercion. [ENDQ] [NEWLINE] Sure.<mask> an employer does not have the moral or legal right to make a janitor fight bears.<mask> they don't have that right, that makes it coercion. [NEWLINE] [NEWLINE] <mask><mask>, a man refusing to pay for a child he never agreed to is something that he has a moral right to. [NEWLINE] [NEWLINE] [STARTQ] The state mandates that men who give up parental resposibilities pay child support. [ENDQ] [NEWLINE] Yes, and the argument is that the state is wrong for doing<mask>. [NEWLINE] [NEWLINE] [STARTQ] You're essentially arguing that victims of theft, say, force people to go to jail for stealing from them, this is simply not true. [ENDQ] [NEWLINE] No, I am not. Your arguments are incredibly dishonest. [NEWLINE] [NEWLINE] Stealing is undeniably wrong, and it's obvious that people should not have the right to do it. [NEWLINE] [NEWLINE] <mask><mask>, refusing to raise a child you never wanted to have is not wrong. [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask><mask><mask><mask> I take out a million dollar line of credit and refuse to pay it back unless [ENDQ] [NEWLINE] You don't seem to understand.<mask> comes after the "unless" doesn't matter. It is quite irrelevant. [NEWLINE] [NEWLINE] The fact is, someone who takes out a loan and doesn't pay it back is in the wrong, both morally and legally. They are obligated to pay back their debts - whatever conditions they state is quite irrelevant to that. [NEWLINE] [NEWLINE] You keep equating a man not paying for a child he never wanted to have with obviously unethical acts - that is quite a dishonest argument. [NEWLINE] [NEWLINE] [STARTQ] <mask> until the father acts to withdraw support, support is assumed in the same way that until demands are leveled repayment is assumed. [ENDQ] [NEWLINE] <mask>??? You are saying that all men are **assumed** to support any resulting kids<mask> they have sex, **even<mask> they explicitly say they don't want kids, use birth control, and/or the woman sabotages the birth control or lies about it**? [NEWLINE] [NEWLINE] That is quite a ridiculous argument. [NEWLINE] [NEWLINE] [STARTQ] You're assuming offering a conditional favour is the same<mask> leveling a conditional penalty and it's just not. [ENDQ] [NEWLINE] A man not giving a woman his money is not a "penalty". That is his right,<mask><mask> the reason. Likewise, a woman aborting her child is not a "
Label encoding: <s> [STARTQ] To stretch your analogy, however, if you hire someone to clean toilets and then tell him to fight bears instead threatening to withhold his livelihood if he does not, you absolutely are engaging in financial coercion. [ENDQ] [NEWLINE] Sure. Because an employer does not have the moral or legal right to make a janitor fight bears. Since they don't have that right, that makes it coercion. [NEWLINE] [NEWLINE] In contrast, a man refusing to pay for a child he never agreed to is something that he has a moral right to. [NEWLINE] [NEWLINE] [STARTQ] The state mandates that men who give up parental resposibilities pay child support. [ENDQ] [NEWLINE] Yes, and the argument is that the state is wrong for doing so. [NEWLINE] [NEWLINE] [STARTQ] You're essentially arguing that victims of theft, say, force people to go to jail for stealing from them, this is simply not true. [ENDQ] [NEWLINE] No, I am not. Your arguments are incredibly dishonest. [NEWLINE] [NEWLINE] Stealing is undeniably wrong, and it's obvious that people should not have the right to do it. [NEWLINE] [NEWLINE] In contrast, refusing to raise a child you never wanted to have is not wrong. [NEWLINE] [NEWLINE] [STARTQ] If, on the other hand I take out a million dollar line of credit and refuse to pay it back unless [ENDQ] [NEWLINE] You don't seem to understand. What comes after the "unless" doesn't matter. It is quite irrelevant. [NEWLINE] [NEWLINE] The fact is, someone who takes out a loan and doesn't pay it back is in the wrong, both morally and legally. They are obligated to pay back their debts - whatever conditions they state is quite irrelevant to that. [NEWLINE] [NEWLINE] You keep equating a man not paying for a child he never wanted to have with obviously unethical acts - that is quite a dishonest argument. [NEWLINE] [NEWLINE] [STARTQ] Because until the father acts to withdraw support, support is assumed in the same way that until demands are leveled repayment is assumed. [ENDQ] [NEWLINE] What??? You are saying that all men are **assumed** to support any resulting kids if they have sex, **even if they explicitly say they don't want kids, use birth control, and/or the woman sabotages the birth control or lies about it**? [NEWLINE] [NEWLINE] That is quite a ridiculous argument. [NEWLINE] [NEWLINE] [STARTQ] You're assuming offering a conditional favour is the same as leveling a conditional penalty and it's just not. [ENDQ] [NEWLINE] A man not giving a woman his money is not a "penalty". That is his right, regardless of the reason. Likewise, a woman aborting her child is not a "
Loss: tensor(0.2936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd like to point out that<mask> you have identified is the major difference between nationalism and cosmopolitanism. This is a sharply divisive topic within the political science community, with both sides having excellent points. Nationalist thinkers believe that co-nationals have more of a moral claim on each other than co-humans do, and Cosmopolitan thinkers believe that the only justifiable relationship dynamic is one of co-humans. [NEWLINE] [NEWLINE] Cosmopolitans views nationalism<mask> morally untenable. Their main questions for nationalists are those of **justification and arbitrariness:**<mask> do you feel that your fellow citizens deserve more loyalty based on arbitrary luck--<mask> they were born,<mask> they gained citizenship, etc? [NEWLINE] [NEWLINE] Nationalism views Cosmopolitanism<mask> unrealistic.Their main questions for cosmopolitans are those of **motivation and attainability:**<mask> do you motivate citizens without an appeal to a common group identity?<mask> can nations pursue a cosmopolitan ideal of co-humanism<mask> there is<mask> much divisiveness/hate within a nation? [NEWLINE] [NEWLINE] <mask> you seem to be arguing from a cosmopolitan perspective, I'll answer from a nationalist one. First (you identified this<mask> a flaw), a national identity is beneficial to the average citizen precisely<mask> it motivates people to action. Yes it motivates people to fight in wars,<mask> wars have been incredibly beneficial to some average citizens. just look at America, they fought WW1, WW2, the Cold War, etc precisely<mask> of nationalism. That has greatly increased the wealth and livelihood of the American people. And<mask> about voting? People vote<mask> they feel proud to part of an country which has such free and liberal institutions. Nationalism motivates a populace, it's true that it can be used for ill,<mask> on a co-human scale,<mask> would we use to motivate people? **<mask> modern nationalists argue is that nationalism doesn't motivate through an us vs. them dichotomy,<mask> rather it motivates through a people's love of the particular institutions of freedom which exist in a country.** [NEWLINE] [NEWLINE] <mask>, in terms of morality, it completely makes sense to prefer co-nationals to foreigners. It all has to do with in-group/out-group relationships, which are not necessarily a bad thing. It's true that you had no choice in your nationality,<mask> you didn't have choice in a lot of things. [NEWLINE] [NEWLINE] Pretty much<mask> it boils down to is this: I feel closer with my brother than I do with
Label encoding: <s>I'd like to point out that what you have identified is the major difference between nationalism and cosmopolitanism. This is a sharply divisive topic within the political science community, with both sides having excellent points. Nationalist thinkers believe that co-nationals have more of a moral claim on each other than co-humans do, and Cosmopolitan thinkers believe that the only justifiable relationship dynamic is one of co-humans. [NEWLINE] [NEWLINE] Cosmopolitans views nationalism as morally untenable. Their main questions for nationalists are those of **justification and arbitrariness:** Why do you feel that your fellow citizens deserve more loyalty based on arbitrary luck-- where they were born, how they gained citizenship, etc? [NEWLINE] [NEWLINE] Nationalism views Cosmopolitanism as unrealistic.Their main questions for cosmopolitans are those of **motivation and attainability:** How do you motivate citizens without an appeal to a common group identity? How can nations pursue a cosmopolitan ideal of co-humanism if there is so much divisiveness/hate within a nation? [NEWLINE] [NEWLINE] Because you seem to be arguing from a cosmopolitan perspective, I'll answer from a nationalist one. First (you identified this as a flaw), a national identity is beneficial to the average citizen precisely because it motivates people to action. Yes it motivates people to fight in wars, but wars have been incredibly beneficial to some average citizens. just look at America, they fought WW1, WW2, the Cold War, etc precisely because of nationalism. That has greatly increased the wealth and livelihood of the American people. And what about voting? People vote because they feel proud to part of an country which has such free and liberal institutions. Nationalism motivates a populace, it's true that it can be used for ill, but on a co-human scale, what would we use to motivate people? ** What modern nationalists argue is that nationalism doesn't motivate through an us vs. them dichotomy, but rather it motivates through a people's love of the particular institutions of freedom which exist in a country.** [NEWLINE] [NEWLINE] Secondly, in terms of morality, it completely makes sense to prefer co-nationals to foreigners. It all has to do with in-group/out-group relationships, which are not necessarily a bad thing. It's true that you had no choice in your nationality, but you didn't have choice in a lot of things. [NEWLINE] [NEWLINE] Pretty much what it boils down to is this: I feel closer with my brother than I do with
Loss: tensor(0.1621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1825, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is an interesting comment<mask> most people don't acknowledge that transexuals have changed sex or gender. "Whether or not there is a modification means nothing." isn't a popular opinion outside of Tumblr. [NEWLINE] [NEWLINE] The uni bathroom is a great idea,<mask> then it doesn't matter<mask> you are, you just go in. These are limited,<mask>,<mask> they aren't universal bathrooms. And they only seat one person.<mask> they didn't, we could have women and men using the same bathroom and OP's problem doesn't exist. [NEWLINE] [NEWLINE] I get that the line of thinking "<mask> you have determines<mask> you are." makes the bathroom thing easier categorically,<mask> in practice there are weird situations. You might have seen the picture of the cowboy in the ladies room. This is a person who has accepted a male gender,<mask> has a vagina. They've taken therapy to help deal with clinical issues like body dysmorphia. They've practiced taking on traits of the gender they want to assume. The biggest part of everything involved with transitioning is the hormone supplements leading up to transitioning. [NEWLINE] [NEWLINE] <mask> you want to talk to scientists about the physical representations of sex and the mental recognition of gender, talk to Tumblr. They've researched and educated their community on three main components that affect<mask> sex you are.<mask> you said, the most obvious is genitalia,<mask> of course no one ever sees other people's genitalia,<mask> it's kind of hard to use that. The next thing we can use is body structure, which is heavily influenced by hormones. [NEWLINE] [NEWLINE] Transgender people taking hormone supplements (before transitioning, thats<mask> I'm calling them transgender) will look like the opposite gender. You will call them by the opposite gender, not<mask> you are a nice, polite, super political correct person,<mask><mask> they have attained the body structure of the other gender and you can't see their genitals. This is<mask> they are called transgender. They don't mentally identify themselves<mask><mask> their genitals claim they are (and neither will you,<mask> they can help it). That dysmorphia has physical and mental evidence that is constantly studied by people interested in the development and condition of human brains. [NEWLINE] [NEWLINE] Tumblr researches those studies, sends them around the community, debates on the usefulness and validity of those studies, and generally comes to a consensus on<mask> gender and sex is. This is the preferred way of approaching education, rather than reading the definitions written clearly on every dictionary and then claiming social context changes<mask> gender and sex
Label encoding: <s>This is an interesting comment because most people don't acknowledge that transexuals have changed sex or gender. "Whether or not there is a modification means nothing." isn't a popular opinion outside of Tumblr. [NEWLINE] [NEWLINE] The uni bathroom is a great idea, because then it doesn't matter what you are, you just go in. These are limited, though, because they aren't universal bathrooms. And they only seat one person. If they didn't, we could have women and men using the same bathroom and OP's problem doesn't exist. [NEWLINE] [NEWLINE] I get that the line of thinking " What you have determines what you are." makes the bathroom thing easier categorically, but in practice there are weird situations. You might have seen the picture of the cowboy in the ladies room. This is a person who has accepted a male gender, but has a vagina. They've taken therapy to help deal with clinical issues like body dysmorphia. They've practiced taking on traits of the gender they want to assume. The biggest part of everything involved with transitioning is the hormone supplements leading up to transitioning. [NEWLINE] [NEWLINE] If you want to talk to scientists about the physical representations of sex and the mental recognition of gender, talk to Tumblr. They've researched and educated their community on three main components that affect what sex you are. As you said, the most obvious is genitalia, but of course no one ever sees other people's genitalia, so it's kind of hard to use that. The next thing we can use is body structure, which is heavily influenced by hormones. [NEWLINE] [NEWLINE] Transgender people taking hormone supplements (before transitioning, thats why I'm calling them transgender) will look like the opposite gender. You will call them by the opposite gender, not because you are a nice, polite, super political correct person, but because they have attained the body structure of the other gender and you can't see their genitals. This is why they are called transgender. They don't mentally identify themselves as what their genitals claim they are (and neither will you, if they can help it). That dysmorphia has physical and mental evidence that is constantly studied by people interested in the development and condition of human brains. [NEWLINE] [NEWLINE] Tumblr researches those studies, sends them around the community, debates on the usefulness and validity of those studies, and generally comes to a consensus on what gender and sex is. This is the preferred way of approaching education, rather than reading the definitions written clearly on every dictionary and then claiming social context changes how gender and sex
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1397, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>EDIT: I have awarded /u/huadpe my delta. They have convinced me that I have misunderstood the argument<mask> well<mask> others misusing the argument.<mask> people talk about birth control being a right, they are not talking about (or shouldn't be at least) a basic right<mask> rather a contractual one for health care to cover meaningful expenses. Thanks everyone for their input and discussion! [NEWLINE] [NEWLINE] EDIT 2: I'm still continuing this discussion<mask><mask> my view has been changed on part of this issue. I might just create another CMV<mask> the issues I now raise are different<mask> I will do my best to continue to respond to comments<mask> I have time between studying and my final today. [NEWLINE] [NEWLINE] <mask> Lena Dunham has once again caused another stir in the social justice world by asking twitter<mask> they need birth control.<mask><mask><mask>, even<mask> a woman, I find that an extremely childish point of view. To "need" birth control. (I would like to preface this by saying that yes I am aware of the various conditions that birth control aides, and for the sake of this CMV, those are not<mask> I'm talking about here). For all non-medical reasons, no one needs birth control. [NEWLINE] [NEWLINE] I am a huge advocate for safe sex, and the prevention of pregnancy and the spread of STDs through birth control and condoms. That being said, I don't think sex is a basic human right. No one needs to have sex every three days or they'll face death, maiming, etc. Sex is pleasurable and sex is fun and sex can create life.<mask> by no means is it a necessity in the way that many claim it is. Furthermore, claiming that your healthcare should be required to cover birth control outside of health concerns<mask> it's a basic right is absolutely ridiculous. [NEWLINE] [NEWLINE] <mask> you are not prepared to make birth control a priority by budgeting for it, then you need to make the decision to either practice unprotected sex and face the consequences or abstain until you can afford to make birth control a priority. Here are some of the main reasons<mask><mask><mask><mask> : [NEWLINE] [NEWLINE] * Birth Control doesn't have to be super expensive,<mask> low<mask> $15 a month. [NEWLINE] [NEWLINE] *<mask> you can't afford $15 a month for birth control you definitely can't afford a kid and should be able to abstain until you are more financially stable. [NEWLINE] [NEWLINE] * There are other non-penetrative methods for sexual release [NEWLINE] [NEWLINE] *<mask> you can't
Label encoding: <s>EDIT: I have awarded /u/huadpe my delta. They have convinced me that I have misunderstood the argument as well as others misusing the argument. When people talk about birth control being a right, they are not talking about (or shouldn't be at least) a basic right but rather a contractual one for health care to cover meaningful expenses. Thanks everyone for their input and discussion! [NEWLINE] [NEWLINE] EDIT 2: I'm still continuing this discussion even though my view has been changed on part of this issue. I might just create another CMV since the issues I now raise are different but I will do my best to continue to respond to comments as I have time between studying and my final today. [NEWLINE] [NEWLINE] So Lena Dunham has once again caused another stir in the social justice world by asking twitter why they need birth control. First of all, even as a woman, I find that an extremely childish point of view. To "need" birth control. (I would like to preface this by saying that yes I am aware of the various conditions that birth control aides, and for the sake of this CMV, those are not what I'm talking about here). For all non-medical reasons, no one needs birth control. [NEWLINE] [NEWLINE] I am a huge advocate for safe sex, and the prevention of pregnancy and the spread of STDs through birth control and condoms. That being said, I don't think sex is a basic human right. No one needs to have sex every three days or they'll face death, maiming, etc. Sex is pleasurable and sex is fun and sex can create life. But by no means is it a necessity in the way that many claim it is. Furthermore, claiming that your healthcare should be required to cover birth control outside of health concerns because it's a basic right is absolutely ridiculous. [NEWLINE] [NEWLINE] If you are not prepared to make birth control a priority by budgeting for it, then you need to make the decision to either practice unprotected sex and face the consequences or abstain until you can afford to make birth control a priority. Here are some of the main reasons why I think so : [NEWLINE] [NEWLINE] * Birth Control doesn't have to be super expensive, as low as $15 a month. [NEWLINE] [NEWLINE] * If you can't afford $15 a month for birth control you definitely can't afford a kid and should be able to abstain until you are more financially stable. [NEWLINE] [NEWLINE] * There are other non-penetrative methods for sexual release [NEWLINE] [NEWLINE] * If you can't
Loss: tensor(0.1346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2984, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're saying want approval to judge and punish other people based only on the shape of their body?  <mask> is that kind of bigotry any different than racism or sexism?  <mask> bigot hasn't justified his attacks on his target group by looking down on them and saying they don't know<mask> inferior they really are? [NEWLINE] [NEWLINE] I could point out a million reasons<mask> it's wrong to be a hateful bigot towards people who don't look the way you like. <mask> I'm not sure I can can change your view.  You have worked hard building your own self esteem by distancing yourself from<mask> you used to be.  It's unimaginable to you that other people might have that quality you used to and be happy productive people. [NEWLINE] [NEWLINE] The real problem is that it's arrogant to walk around thinking you know<mask> know<mask> is best for other people.  You take a glance at a stranger and can out perform any doctor by diagnosing their health and happiness levels?  You have made yourself the ultimate authority on<mask> looks and feels good to humankind,  without considering not everyone has your beliefs and goals. [NEWLINE] [NEWLINE] Here's the deal: [NEWLINE] -  fat people are still people and deserve the same respect given to all people [NEWLINE] [NEWLINE] - you are not in charge of anyone else's health or body choices.  You are not the food police or the fashion police. [NEWLINE] [NEWLINE] - the only thing you can be sure of<mask> looking at a fat person is that they're fat. [NEWLINE] [NEWLINE] - there is absolutely no evidence that supports "fat shaming" or any verbal abuse changes a fat person's body shape or health.   Verbal abuse about appearance does cause stress and may trigger eating disordered behavior.  Did you know the #1 most lethal mental illness in the us is eating disorders? [NEWLINE] [NEWLINE] -  there is no evidence that long term any weight loss plan works for most people.  The evidence shows that long term (5 year mark) [95% of the time weight loss interventions fail]( [URL] /) and for some people they gain back additional weight. <mask> can you shame others into not doing weight loss<mask> you haven't invented a weight loss intervention that WORKS? [NEWLINE] [NEWLINE] - fat doesn't not equal unhealthy; skinny does not equal healthy.   Skinny people get diabetes, arthritis,  heart attacks, and  respiratory problems.   You're<mask> ignoring the unhealthy reasons<mask> a person might look skinny:  eating disorders, drug or alcohol addiction,  mental illness, cancer,
Label encoding: <s>You're saying want approval to judge and punish other people based only on the shape of their body?   How is that kind of bigotry any different than racism or sexism?   What bigot hasn't justified his attacks on his target group by looking down on them and saying they don't know how inferior they really are? [NEWLINE] [NEWLINE] I could point out a million reasons why it's wrong to be a hateful bigot towards people who don't look the way you like.  But I'm not sure I can can change your view.  You have worked hard building your own self esteem by distancing yourself from what you used to be.  It's unimaginable to you that other people might have that quality you used to and be happy productive people. [NEWLINE] [NEWLINE] The real problem is that it's arrogant to walk around thinking you know what know what is best for other people.  You take a glance at a stranger and can out perform any doctor by diagnosing their health and happiness levels?  You have made yourself the ultimate authority on what looks and feels good to humankind,  without considering not everyone has your beliefs and goals. [NEWLINE] [NEWLINE] Here's the deal: [NEWLINE] -  fat people are still people and deserve the same respect given to all people [NEWLINE] [NEWLINE] - you are not in charge of anyone else's health or body choices.  You are not the food police or the fashion police. [NEWLINE] [NEWLINE] - the only thing you can be sure of when looking at a fat person is that they're fat. [NEWLINE] [NEWLINE] - there is absolutely no evidence that supports "fat shaming" or any verbal abuse changes a fat person's body shape or health.   Verbal abuse about appearance does cause stress and may trigger eating disordered behavior.  Did you know the #1 most lethal mental illness in the us is eating disorders? [NEWLINE] [NEWLINE] -  there is no evidence that long term any weight loss plan works for most people.  The evidence shows that long term (5 year mark) [95% of the time weight loss interventions fail]( [URL] /) and for some people they gain back additional weight.  How can you shame others into not doing weight loss when you haven't invented a weight loss intervention that WORKS? [NEWLINE] [NEWLINE] - fat doesn't not equal unhealthy; skinny does not equal healthy.   Skinny people get diabetes, arthritis,  heart attacks, and  respiratory problems.   You're also ignoring the unhealthy reasons why a person might look skinny:  eating disorders, drug or alcohol addiction,  mental illness, cancer,
Loss: tensor(0.1092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2941, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[Cellulite source.]( [URL] ;rct=j&amp;q=&amp;esrc=s&amp;source=web&amp;cd=1&amp;ved=0CB0QFjAA&amp;url=http%3A%2F%2Fwww.springer.com%2Fcda%2Fcontent%2Fdocument%2Fcda_downloaddocument%2F9783642340284-c1.pdf%3FSGWID%3D0-0-45-1384009-p174676255&amp;ei=U7DGU4KXOcO3igL17oHoCQ&amp;usg=AFQjCNHeN1djm7TPJm6p4mPAOWx2DH829w&amp;sig2=O6U_vmtbPVxOkYXlPpwWKw&amp;bvm=bv.71126742,d.cGE) [NEWLINE] [NEWLINE] I could not find a source on tuberous breasts in the general population, which is<mask> I didn't say anything about it,<mask> there are sources on women who come in for [breast-related surgery]( [URL] ): "In total, 57.1 percent of all reduction mammaplasties (n = 92) and 83.2 percent of all augmentation mammaplasties (n = 178) had asymmetry with tuberous deformity." That's almost certainly higher than the general population,<mask> I would speculate that such high numbers would not occur<mask> the rates in the general population are negligible. [NEWLINE] [NEWLINE] <mask>, please keep in mind that tuberous breasts are only one example of a way that your proportions can not fit the beauty ideal. It came to mind<mask> I spend a lot of time on /r/abrathatfits and I can think of a lot of examples of people coming in and saying things like "I have the ugliest breasts" or whatever (and they are, admittedly, harder to fit). I personally have very, very short legs. (Short enough in proportion to my body that it is notable.) Women have different waist to hip ratios, some of which will look less feminine and less attractive. You have lovely hourglass shapes in your photo,<mask> the same body fat percentage might not be<mask> visually appealing on an [apple shaped
Label encoding: <s>[Cellulite source.]( [URL] ;rct=j&amp;q=&amp;esrc=s&amp;source=web&amp;cd=1&amp;ved=0CB0QFjAA&amp;url=http%3A%2F%2Fwww.springer.com%2Fcda%2Fcontent%2Fdocument%2Fcda_downloaddocument%2F9783642340284-c1.pdf%3FSGWID%3D0-0-45-1384009-p174676255&amp;ei=U7DGU4KXOcO3igL17oHoCQ&amp;usg=AFQjCNHeN1djm7TPJm6p4mPAOWx2DH829w&amp;sig2=O6U_vmtbPVxOkYXlPpwWKw&amp;bvm=bv.71126742,d.cGE) [NEWLINE] [NEWLINE] I could not find a source on tuberous breasts in the general population, which is why I didn't say anything about it, but there are sources on women who come in for [breast-related surgery]( [URL] ): "In total, 57.1 percent of all reduction mammaplasties (n = 92) and 83.2 percent of all augmentation mammaplasties (n = 178) had asymmetry with tuberous deformity." That's almost certainly higher than the general population, but I would speculate that such high numbers would not occur if the rates in the general population are negligible. [NEWLINE] [NEWLINE] Also, please keep in mind that tuberous breasts are only one example of a way that your proportions can not fit the beauty ideal. It came to mind because I spend a lot of time on /r/abrathatfits and I can think of a lot of examples of people coming in and saying things like "I have the ugliest breasts" or whatever (and they are, admittedly, harder to fit). I personally have very, very short legs. (Short enough in proportion to my body that it is notable.) Women have different waist to hip ratios, some of which will look less feminine and less attractive. You have lovely hourglass shapes in your photo, but the same body fat percentage might not be as visually appealing on an [apple shaped
Loss: tensor(0.1489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1990, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2905, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Over the past year, the idea of raising the minimum wage in Canada (<mask> I live) and in the USA has gained popularity. Many states and provinces are either mandating, or in the process of mandating a minimum wage of ~$15/hour.<mask> on the surface, this seems like a good policy for strengthening the poor/middle class I believe this will ultimately HURT, not help the public, and especially the poor. [NEWLINE] [NEWLINE] Here are several common arguments made in favour of the raising minimum wage and my response: [NEWLINE] [NEWLINE] * "Average wages haven't increased<mask> an increase in production and profit in the economy" [NEWLINE] [NEWLINE] It's a common misunderstanding that wages are/should be directly tied to production. In reality wages are determined just like prices of anything else, through supply and demand. Most minimum wage jobs are unskilled and easily replaceable (through either humans or automation) making the supply of labour far larger than the demand, which equals a lower wage. The rise in technology that has resulted in higher production from labour has little affect on wages. [NEWLINE] [NEWLINE] * "Current minimum wages are<mask> low that they do not pay a livable annual salary, keeping the working poor in a perpetual state of poverty" [NEWLINE] [NEWLINE] I would challenge the idea that most minimum wage jobs were ever meant to be sustainable for living. Most minimum wage jobs are unskilled labour (ex. factory work, assembly lines) or service industry jobs (cooks, janitorial staff, retail jobs etc..). At least to me, these seem like temporary work done by people in the process of finding higher paying employment (students, recent graduates, new immigrants etc...), and not intended<mask> a career at all,<mask> a livable wage is not required. [NEWLINE] [NEWLINE] * "A higher minimum wage would increase the income of many in the population, boosting spending, saving and the overall economy." [NEWLINE] [NEWLINE] <mask> I said previously, the vast majority of minimum wage jobs are easily replaceable, outsourced or automated.<mask> large companies are forced by government to pay high minimum wages, they will seek cheaper alternatives and replacements instead. For example, a McDonald's hamburger could very easily be made faster, cheaper, cleaner and better by a machine, [like this one, which makes a burger every 10 seconds ]( [URL] ).<mask> the minimum wage gets high enough that it costs McDonald's more to pay 10 people to flip burgers than one machine, they will simply replace the human labour altogether. This will result in an increase in unemployment.
Label encoding: <s>Over the past year, the idea of raising the minimum wage in Canada ( where I live) and in the USA has gained popularity. Many states and provinces are either mandating, or in the process of mandating a minimum wage of ~$15/hour. While on the surface, this seems like a good policy for strengthening the poor/middle class I believe this will ultimately HURT, not help the public, and especially the poor. [NEWLINE] [NEWLINE] Here are several common arguments made in favour of the raising minimum wage and my response: [NEWLINE] [NEWLINE] * "Average wages haven't increased despite an increase in production and profit in the economy" [NEWLINE] [NEWLINE] It's a common misunderstanding that wages are/should be directly tied to production. In reality wages are determined just like prices of anything else, through supply and demand. Most minimum wage jobs are unskilled and easily replaceable (through either humans or automation) making the supply of labour far larger than the demand, which equals a lower wage. The rise in technology that has resulted in higher production from labour has little affect on wages. [NEWLINE] [NEWLINE] * "Current minimum wages are so low that they do not pay a livable annual salary, keeping the working poor in a perpetual state of poverty" [NEWLINE] [NEWLINE] I would challenge the idea that most minimum wage jobs were ever meant to be sustainable for living. Most minimum wage jobs are unskilled labour (ex. factory work, assembly lines) or service industry jobs (cooks, janitorial staff, retail jobs etc..). At least to me, these seem like temporary work done by people in the process of finding higher paying employment (students, recent graduates, new immigrants etc...), and not intended as a career at all, so a livable wage is not required. [NEWLINE] [NEWLINE] * "A higher minimum wage would increase the income of many in the population, boosting spending, saving and the overall economy." [NEWLINE] [NEWLINE] As I said previously, the vast majority of minimum wage jobs are easily replaceable, outsourced or automated. If large companies are forced by government to pay high minimum wages, they will seek cheaper alternatives and replacements instead. For example, a McDonald's hamburger could very easily be made faster, cheaper, cleaner and better by a machine, [like this one, which makes a burger every 10 seconds ]( [URL] ). If the minimum wage gets high enough that it costs McDonald's more to pay 10 people to flip burgers than one machine, they will simply replace the human labour altogether. This will result in an increase in unemployment.
Loss: tensor(0.2445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Almost everyone I know carries around a tablet and a laptop<mask> they have both. One of my friends brings 3 devices regularly (not including phone in any situation) [ENDQ] [NEWLINE] This kind of seems like a personal problem and I'm not really convinced that a Surface would necessarily solve the problem.  The 11" Air is the same weight<mask> the Surface 3, and the iPad Air is only a pound...that's not an advantage to the Surface<mask> you only carry one or the other.  In my experience people usually carry one device around campus. [NEWLINE] [NEWLINE] [STARTQ] The pen comes in handy for those who need to take notes that involve things that are difficult to simply input by keyboard. Almost any upper-level math course, all organic chemistry courses, and some linguistics courses will benefit from this. Not to mention the art that a large amount of people produce for fun [ENDQ] [NEWLINE] Pen and paper is still king.  The number of people I saw using tablets rather than paper taking notes in my engineering courses was pitiful. <mask>, just<mask> something is handy doesn't mean that it'll be used often.  From an art perspective I'll throw that back under the "niche" category,<mask> you're argument is based on the general population I wouldn't count this<mask> an advantage there. [NEWLINE] [NEWLINE] [STARTQ] No, we need 1440p screens on 11" laptops :) [ENDQ] [NEWLINE] Simply put, no...we don't. [NEWLINE] [NEWLINE] [STARTQ] The better DPI on an HD laptop really reduces eyestrain.<mask>, the lack of HD (or even IPS technology) really makes MacBook Airs look overpriced [ENDQ] [NEWLINE] I'll admit that Apple hasn't really been nice to the Air's screen...<mask> at the ranges you're using a laptop over a phone coupled with an 11-13" screen the DPI isn't really an issue for the same reason that a 24" 1080p monitor isn't really a problem.  The battery life saved and the extra horsepower (read: heat, money) to power the display vs the extra DPI on a screen that you shouldn't be pressing against your nose is a wash<mask><mask>. [NEWLINE] [NEWLINE] [STARTQ] I meant by synchronization that having one device with everything on it is superior to any greater number of devices. [ENDQ] [NEWLINE] In that respect the MBA is just<mask> capable<mask> the Surface in general.  The only thing it lacks is touchscreen input, which isn't necessary. [NEWLINE] [NEWLINE] [STARTQ] I actually should have removed this. I haven't tried applications like Word on the iPad,<mask> I don't
Label encoding: <s> [STARTQ] Almost everyone I know carries around a tablet and a laptop if they have both. One of my friends brings 3 devices regularly (not including phone in any situation) [ENDQ] [NEWLINE] This kind of seems like a personal problem and I'm not really convinced that a Surface would necessarily solve the problem.  The 11" Air is the same weight as the Surface 3, and the iPad Air is only a pound...that's not an advantage to the Surface if you only carry one or the other.  In my experience people usually carry one device around campus. [NEWLINE] [NEWLINE] [STARTQ] The pen comes in handy for those who need to take notes that involve things that are difficult to simply input by keyboard. Almost any upper-level math course, all organic chemistry courses, and some linguistics courses will benefit from this. Not to mention the art that a large amount of people produce for fun [ENDQ] [NEWLINE] Pen and paper is still king.  The number of people I saw using tablets rather than paper taking notes in my engineering courses was pitiful.  Also, just because something is handy doesn't mean that it'll be used often.  From an art perspective I'll throw that back under the "niche" category, since you're argument is based on the general population I wouldn't count this as an advantage there. [NEWLINE] [NEWLINE] [STARTQ] No, we need 1440p screens on 11" laptops :) [ENDQ] [NEWLINE] Simply put, no...we don't. [NEWLINE] [NEWLINE] [STARTQ] The better DPI on an HD laptop really reduces eyestrain. Also, the lack of HD (or even IPS technology) really makes MacBook Airs look overpriced [ENDQ] [NEWLINE] I'll admit that Apple hasn't really been nice to the Air's screen... however at the ranges you're using a laptop over a phone coupled with an 11-13" screen the DPI isn't really an issue for the same reason that a 24" 1080p monitor isn't really a problem.  The battery life saved and the extra horsepower (read: heat, money) to power the display vs the extra DPI on a screen that you shouldn't be pressing against your nose is a wash IMO. [NEWLINE] [NEWLINE] [STARTQ] I meant by synchronization that having one device with everything on it is superior to any greater number of devices. [ENDQ] [NEWLINE] In that respect the MBA is just as capable as the Surface in general.  The only thing it lacks is touchscreen input, which isn't necessary. [NEWLINE] [NEWLINE] [STARTQ] I actually should have removed this. I haven't tried applications like Word on the iPad, so I don't
Loss: tensor(0.3170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2886, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem here is this isn't something that's happened lately, it's just something that's become more prevalent lately. This element of feminism isn't anything new, it's just become contemporarily popular. Sort of like swing dancing in the 90s. [NEWLINE] [NEWLINE] You're right that it's not representative of all of feminism,<mask><mask><mask> it's *inevitable*<mask> you're working with feminism<mask> it's been structured. We're talking about a model of sexism that,<mask> highlighting the disparity in perceived agency between men and women, completely disregards that disparity in their solutions to sexism.<mask> sexism manifests<mask> a lack of respect for women<mask> agents and a lack of respect for men<mask> objects,<mask> on Earth are you going to remedy this by denying the agency of women?<mask> feminism wants to empower women the first thing it needs to do is stop *coddling* women. It's great to focus on instances of legitimate inequality,<mask> it's *terrible* to make people more hung up on<mask> basically amounts to chivalry than they already are. Men not telling offensive jokes around women isn't a right that women won, it's the way it's been for ages. We can't offend the women folk, they're delicate. They might go into hysterics. [NEWLINE] [NEWLINE] <mask> feminism focuses on protecting the delicate sensibilities of women, it is a regressive force of sexism. It may pretend to be stamping that very thing out,<mask><mask> it's really doing is *adjusting* sexism<mask> that it's more in line with<mask>'s comfortable for women in a more opulent society. Even from a sexist perspective, in which women are helpless, we quite clearly don't need to keep a stranglehold on them to ensure the safety of our offspring and the continuation our civilization. They've got plenty to do to take up their time and a whole society focused on keeping them safe. That's sexism. [NEWLINE] [NEWLINE] <mask> anti-sexism ought to be is the reversal and negation of assumptions of imbalanced agency. People of both sexes need to realize that women can and do fend for themselves and men can and do need help sometimes. This is a bit difficult,<mask> women are capable of pregnancy and are themselves loaded with neotenous traits<mask> men are equipped with extra muscle and loaded with accelerated traits. It's quite natural for us to make the sexist assumptions we do. It *seems*<mask><mask> it was once a rather adaptive model. In the context of our current situation,
Label encoding: <s>The problem here is this isn't something that's happened lately, it's just something that's become more prevalent lately. This element of feminism isn't anything new, it's just become contemporarily popular. Sort of like swing dancing in the 90s. [NEWLINE] [NEWLINE] You're right that it's not representative of all of feminism, but I think it's *inevitable* when you're working with feminism as it's been structured. We're talking about a model of sexism that, while highlighting the disparity in perceived agency between men and women, completely disregards that disparity in their solutions to sexism. If sexism manifests as a lack of respect for women as agents and a lack of respect for men as objects, how on Earth are you going to remedy this by denying the agency of women? If feminism wants to empower women the first thing it needs to do is stop *coddling* women. It's great to focus on instances of legitimate inequality, but it's *terrible* to make people more hung up on what basically amounts to chivalry than they already are. Men not telling offensive jokes around women isn't a right that women won, it's the way it's been for ages. We can't offend the women folk, they're delicate. They might go into hysterics. [NEWLINE] [NEWLINE] When feminism focuses on protecting the delicate sensibilities of women, it is a regressive force of sexism. It may pretend to be stamping that very thing out, but what it's really doing is *adjusting* sexism so that it's more in line with what's comfortable for women in a more opulent society. Even from a sexist perspective, in which women are helpless, we quite clearly don't need to keep a stranglehold on them to ensure the safety of our offspring and the continuation our civilization. They've got plenty to do to take up their time and a whole society focused on keeping them safe. That's sexism. [NEWLINE] [NEWLINE] What anti-sexism ought to be is the reversal and negation of assumptions of imbalanced agency. People of both sexes need to realize that women can and do fend for themselves and men can and do need help sometimes. This is a bit difficult, as women are capable of pregnancy and are themselves loaded with neotenous traits while men are equipped with extra muscle and loaded with accelerated traits. It's quite natural for us to make the sexist assumptions we do. It *seems* as though it was once a rather adaptive model. In the context of our current situation,
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1905, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a rather unique perspective on this issue<mask> I am a former Coast Guardsman and I used to pick up the bodies of people who commit suicide off the Golden Gate Bridge on a weekly basis. One of them was even a 13 year old girl. I have dozens<mask> not hundreds of stories about this exact topic. [NEWLINE] [NEWLINE] I am not here to tell you that there is anything morally wrong with suicide. I am not judging the people who kill themselves or telling them<mask> they should do with their lives. It's none of my business, frankly. I'm a pro-freedom type,<mask> you want to kill yourself, kill yourself. [NEWLINE] [NEWLINE] <mask> I am saying, is that out of the hundred or<mask> people I've seen jump off that bridge and had to pick up, two of them survived. Both of those people are on record for saying that it was the worst mistake they've ever made in their entire lives and that the second they set foot off of that bridge, they realized that. [NEWLINE] [NEWLINE] It stands to reason that most<mask> not all of the people who commit suicide this way think the same in their last moments. Suicide,<mask> has been famously put, is a permanent solution to a temporary problem. It is, for the vast majority of people, a huge mistake. Jumping off of a bridge with a low rail is a really quick and easy thing to do, making it really easy to make this huge mistake<mask> you happen to be in a state of emotional distress.<mask> you remove suicide by jumping from a bridge<mask> an option by raising the rails on a bridge, the people who don't kill themselves by jumping off of it are prevented from making a horrible mistake that they don't get to learn from<mask> now they're dead and instead have a chance at actually fixing the problems in their life instead of running from them. Some of these people will find other ways of killing themselves,<mask> some of them won't. [Studies have shown that well-designed suicide barriers not only stop people from jumping at a particular site,<mask><mask> decrease the overall suicide rate in the surrounding area.]( [URL] ) [NEWLINE] [NEWLINE] Imagine<mask> that 13 year old girl didn't have the option of jumping off of the bridge, nor access to any other means to kill herself and actually had to deal with her life. I can't imagine that the problems in her life,<mask> serious they might have been, were something that couldn't have been resolved. It's much more probable that she did it out of being in a situation of temporary emotional
Label encoding: <s>I have a rather unique perspective on this issue because I am a former Coast Guardsman and I used to pick up the bodies of people who commit suicide off the Golden Gate Bridge on a weekly basis. One of them was even a 13 year old girl. I have dozens if not hundreds of stories about this exact topic. [NEWLINE] [NEWLINE] I am not here to tell you that there is anything morally wrong with suicide. I am not judging the people who kill themselves or telling them what they should do with their lives. It's none of my business, frankly. I'm a pro-freedom type, if you want to kill yourself, kill yourself. [NEWLINE] [NEWLINE] What I am saying, is that out of the hundred or so people I've seen jump off that bridge and had to pick up, two of them survived. Both of those people are on record for saying that it was the worst mistake they've ever made in their entire lives and that the second they set foot off of that bridge, they realized that. [NEWLINE] [NEWLINE] It stands to reason that most if not all of the people who commit suicide this way think the same in their last moments. Suicide, as has been famously put, is a permanent solution to a temporary problem. It is, for the vast majority of people, a huge mistake. Jumping off of a bridge with a low rail is a really quick and easy thing to do, making it really easy to make this huge mistake if you happen to be in a state of emotional distress. If you remove suicide by jumping from a bridge as an option by raising the rails on a bridge, the people who don't kill themselves by jumping off of it are prevented from making a horrible mistake that they don't get to learn from because now they're dead and instead have a chance at actually fixing the problems in their life instead of running from them. Some of these people will find other ways of killing themselves, but some of them won't. [Studies have shown that well-designed suicide barriers not only stop people from jumping at a particular site, but also decrease the overall suicide rate in the surrounding area.]( [URL] ) [NEWLINE] [NEWLINE] Imagine if that 13 year old girl didn't have the option of jumping off of the bridge, nor access to any other means to kill herself and actually had to deal with her life. I can't imagine that the problems in her life, however serious they might have been, were something that couldn't have been resolved. It's much more probable that she did it out of being in a situation of temporary emotional
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0961, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1905, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1858, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2512, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think<mask> people are trying to convey is the following: [NEWLINE] [NEWLINE] Traditionally chess was male dominated. Mainly<mask> of prejudices and societal norms about<mask> are the relative roles of men and women.<mask>, a very small number of women at that time engaged in chess or ever thought that chess was something they could be interested in. [NEWLINE] [NEWLINE] <mask> things started to change, chess was still male dominated. Women are still less exposed to chess (<mask>, have less interest), less welcomed in the chess playing community etc, etc.<mask>, much more males play chess than females, and the population from which male professional chess players are drawn is still much bigger than the population from which female professional chess players are drawn.  The fact that men are drawn from a bigger population, with more support and resources, means that **on average** a randomly selected male player will be playing at a greater level than a randomly selected female player. This is<mask> *I hope* /u/0xstev3 was trying to say with: [NEWLINE] [NEWLINE] [STARTQ] they'd be more likely to lose apparently? [ENDQ] [NEWLINE] Again: this has nothing to do with innate capacities of women for chess,<mask> only with statistics. There are overwhelmingly more male players (<mask> of societal norms, prejudice, etc, etc),<mask>, even<mask> innate ability and opportunity were gender neutral, there would be overwhelmingly more really good male players. [NEWLINE] [NEWLINE] To find a person that can play at the grand master level is a extremely rare event. That person went to lots of work, training and<mask> have some natural talents that are rare to find. And the number of females that end up reaching that level is even smaller for reasons already discussed. This implies that you have a low chance of ever seeing a women in a tournament at that level, simply<mask> women are numerically overwhelmed by men in the population that can qualify for those tournaments. [NEWLINE] [NEWLINE] For you to find a female in a tournament you'd have to observe a two rare events simultaneously: not only you'd have to find a player that will beat most of the contesters in qualification<mask> it<mask> must be a female professional chess player at a very high level. Such rarity creates a feedback loop: there are few women playing chess,<mask> there are few women grand masters,<mask> there are few women at tournaments,<mask> there are few women noticing the fact that chess is an interesting thing for women to engage with,<mask> there are few women willing to face societal barriers to become a chess player, <mask> there are few women playing
Label encoding: <s>I think what people are trying to convey is the following: [NEWLINE] [NEWLINE] Traditionally chess was male dominated. Mainly because of prejudices and societal norms about what are the relative roles of men and women. Thus, a very small number of women at that time engaged in chess or ever thought that chess was something they could be interested in. [NEWLINE] [NEWLINE] When things started to change, chess was still male dominated. Women are still less exposed to chess ( thus, have less interest), less welcomed in the chess playing community etc, etc. So, much more males play chess than females, and the population from which male professional chess players are drawn is still much bigger than the population from which female professional chess players are drawn.  The fact that men are drawn from a bigger population, with more support and resources, means that **on average** a randomly selected male player will be playing at a greater level than a randomly selected female player. This is what *I hope* /u/0xstev3 was trying to say with: [NEWLINE] [NEWLINE] [STARTQ] they'd be more likely to lose apparently? [ENDQ] [NEWLINE] Again: this has nothing to do with innate capacities of women for chess, but only with statistics. There are overwhelmingly more male players ( because of societal norms, prejudice, etc, etc), thus, even if innate ability and opportunity were gender neutral, there would be overwhelmingly more really good male players. [NEWLINE] [NEWLINE] To find a person that can play at the grand master level is a extremely rare event. That person went to lots of work, training and also have some natural talents that are rare to find. And the number of females that end up reaching that level is even smaller for reasons already discussed. This implies that you have a low chance of ever seeing a women in a tournament at that level, simply because women are numerically overwhelmed by men in the population that can qualify for those tournaments. [NEWLINE] [NEWLINE] For you to find a female in a tournament you'd have to observe a two rare events simultaneously: not only you'd have to find a player that will beat most of the contesters in qualification but it also must be a female professional chess player at a very high level. Such rarity creates a feedback loop: there are few women playing chess, thus there are few women grand masters, thus there are few women at tournaments, thus there are few women noticing the fact that chess is an interesting thing for women to engage with, thus there are few women willing to face societal barriers to become a chess player,  thus there are few women playing
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a good collection,<mask> of course you've cherry-picked only the opposing papers. [NEWLINE] [NEWLINE] I much prefer the links from this site: [Circumcision:<mask> Does Science Say]( [URL] /), in which the sum over *all* science is that there really isn't much case either way. It is purely a choice issue. Your anecdote aside, the sensation issue<mask> doesn't fit the science. Some studies suggest [there is]( [URL] ) some minor loss of sensation, and some studies suggest [there is not]( [URL] /). The aggregate appears to be that the evidence doesn't swing far one way or the other, that statistically<mask> there is a measurable phenomenon that it is a very weak one. [NEWLINE] [NEWLINE] Further, none of the scientific arguments on either side, which seem to amount to "meh", address the actual reason circumcision survives<mask> a cultural phenomenon. It isn't religion; it's sexual selection. That is, [many women]( [URL] ) find an uncircumcised penis to be unattractive, gross, smelly, and unsatisfactory. [NEWLINE] [NEWLINE] This may be a conditioning issue,<mask> that is irrelevant. That is, it may be that women brought up in cultures with mostly circumcised men will tend to dislike uncircumcised penises,<mask> that doesn't change the fact that they do. An uncircumcised male in such a culture will be at a disadvantage from being selected<mask> a mate or having a satisfied female lover later in life. [NEWLINE] [NEWLINE] I have gotten the opinion of several dozen women in my area and all of them dislike uncircumcised penises. For several of them, it's a deal breaker -- meaning they won't continue to date an uncircumcised male. None are religious. These include my wife and many of her friends,<mask> it is not a random sample at all, nor independent opinions. Rather,<mask> we're trading anecdotes I use it here only to illustrate the point that there is a cost to uncircumcised males. [NEWLINE] [NEWLINE] In that sense, it is a self-perpetuating phenomenon much like the way sexual selection works.<mask> there were no circumcisions then it wouldn't start on its own<mask> a preference,<mask><mask> it does exist it has become a point of selection. The standard example is the peacocks tail. Like circumcision, a peacock's tail makes no sense on paper.<mask> once a longer, more ornate tail started being attractive to peahens, it was the males with those features that
Label encoding: <s>This is a good collection, but of course you've cherry-picked only the opposing papers. [NEWLINE] [NEWLINE] I much prefer the links from this site: [Circumcision: What Does Science Say]( [URL] /), in which the sum over *all* science is that there really isn't much case either way. It is purely a choice issue. Your anecdote aside, the sensation issue also doesn't fit the science. Some studies suggest [there is]( [URL] ) some minor loss of sensation, and some studies suggest [there is not]( [URL] /). The aggregate appears to be that the evidence doesn't swing far one way or the other, that statistically if there is a measurable phenomenon that it is a very weak one. [NEWLINE] [NEWLINE] Further, none of the scientific arguments on either side, which seem to amount to "meh", address the actual reason circumcision survives as a cultural phenomenon. It isn't religion; it's sexual selection. That is, [many women]( [URL] ) find an uncircumcised penis to be unattractive, gross, smelly, and unsatisfactory. [NEWLINE] [NEWLINE] This may be a conditioning issue, but that is irrelevant. That is, it may be that women brought up in cultures with mostly circumcised men will tend to dislike uncircumcised penises, but that doesn't change the fact that they do. An uncircumcised male in such a culture will be at a disadvantage from being selected as a mate or having a satisfied female lover later in life. [NEWLINE] [NEWLINE] I have gotten the opinion of several dozen women in my area and all of them dislike uncircumcised penises. For several of them, it's a deal breaker -- meaning they won't continue to date an uncircumcised male. None are religious. These include my wife and many of her friends, so it is not a random sample at all, nor independent opinions. Rather, since we're trading anecdotes I use it here only to illustrate the point that there is a cost to uncircumcised males. [NEWLINE] [NEWLINE] In that sense, it is a self-perpetuating phenomenon much like the way sexual selection works. If there were no circumcisions then it wouldn't start on its own as a preference, but because it does exist it has become a point of selection. The standard example is the peacocks tail. Like circumcision, a peacock's tail makes no sense on paper. But once a longer, more ornate tail started being attractive to peahens, it was the males with those features that
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0979, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2941, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1743, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're making<mask>umptions, it's not bad on itself<mask> I'd like to express myself. [NEWLINE] [NEWLINE] I do believe american kids doing that is,<mask> not barbaric, at least very wrong. It gives blind faith and everything I hate about patriotism and proud on a country. Like yeah, fuck that and everything about indoctrinate children. Anyone doing it.<mask> I feel the need to, at a basic level, say: that is wrong. No matter your context, no matter<mask> you come from, that is wrong. There are bigger and more important stuff than that, for sure.<mask><mask> we discuss the topic, I'm against it. [NEWLINE] [NEWLINE] Second point, yeah in some cases it might not be different from women in america (or western world,<mask><mask><mask> I know is similar) forced to wear certain outfits or makeup. And that ONLY makes that thing in the western world barbaric and backward. It's judging someone on their looks, and many, many people is past that.<mask> you're judged<mask> you have tattoos, yeah shit happens,<mask> it's still SHIT. Shit is not going to taste better<mask> somewhere else shit is the only thing you can eat. [NEWLINE] [NEWLINE] About nuns and doble standard, yeah it is. It still doesn't make it right, and one does not justify the other, in any way or mean. And yeah, many people have doble standard, and a good amount<mask> feel both things are at the same level and can be equal or similar. [NEWLINE] [NEWLINE] And<mask><mask>, the biggest difference in this regard?<mask> I live and<mask> I've been<mask> far in live, people think about<mask> nuns and catholics and other groups of mostly, usually old people do, as backward. Not only old fashioned,<mask> something they can't wait till it dies. It's holding people back, it's an aspect of our culture many people isn't proud or rejects.<mask> they know<mask> does it come from. From<mask> women weren't able to go to school,<mask> they had to go to the church. From<mask> they weren't able to vote, to go take money from a bank without their husband present. That's the type of thing,<mask> I've been and lived and visited, comes from. It's a thing from the past, that is mostly dieing. [NEWLINE] [NEWLINE] Will people still do it? Sure. Feel free to cover.<mask><mask> of you. After that, you go fuck, vote, take the money, and do whatever
Label encoding: <s>You're making asumptions, it's not bad on itself but I'd like to express myself. [NEWLINE] [NEWLINE] I do believe american kids doing that is, if not barbaric, at least very wrong. It gives blind faith and everything I hate about patriotism and proud on a country. Like yeah, fuck that and everything about indoctrinate children. Anyone doing it. Because I feel the need to, at a basic level, say: that is wrong. No matter your context, no matter where you come from, that is wrong. There are bigger and more important stuff than that, for sure. But if we discuss the topic, I'm against it. [NEWLINE] [NEWLINE] Second point, yeah in some cases it might not be different from women in america (or western world, as far as I know is similar) forced to wear certain outfits or makeup. And that ONLY makes that thing in the western world barbaric and backward. It's judging someone on their looks, and many, many people is past that. If you're judged because you have tattoos, yeah shit happens, but it's still SHIT. Shit is not going to taste better because somewhere else shit is the only thing you can eat. [NEWLINE] [NEWLINE] About nuns and doble standard, yeah it is. It still doesn't make it right, and one does not justify the other, in any way or mean. And yeah, many people have doble standard, and a good amount also feel both things are at the same level and can be equal or similar. [NEWLINE] [NEWLINE] And imo, the biggest difference in this regard? Where I live and where I've been so far in live, people think about what nuns and catholics and other groups of mostly, usually old people do, as backward. Not only old fashioned, but something they can't wait till it dies. It's holding people back, it's an aspect of our culture many people isn't proud or rejects. Because they know where does it come from. From when women weren't able to go to school, so they had to go to the church. From when they weren't able to vote, to go take money from a bank without their husband present. That's the type of thing, where I've been and lived and visited, comes from. It's a thing from the past, that is mostly dieing. [NEWLINE] [NEWLINE] Will people still do it? Sure. Feel free to cover. But because of you. After that, you go fuck, vote, take the money, and do whatever
Loss: tensor(0.1669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1478, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> the part of your post I intend to challenge is specifically the "college was only created to make money" aspect. [NEWLINE] [NEWLINE] I know all you young whippersnappers on Reddit have all grown up in this crazy world<mask> glorified multinational industrial Capitalism has all<mask> swallowed the planet whole and is quickly bankrupting its resources and destroying the social contracts the middle class had worked for years to build,<mask> believe it not college wasn't invented to "make money." [NEWLINE] [NEWLINE] Initially, places of learning were more closely tied to the clergy,<mask> being a "scholar" used to be reserved specifically for those who could read and write, and they most generally read and wrote about... you guessed it, religion and religious texts. [NEWLINE] [NEWLINE] Eventually, public schooling became a liberal ideal of a world which benefited the "greater good" by educating<mask> much<mask> the populace<mask> possible.  Honestly, up until recently, a college degree _wasn't_ viewed<mask> a necessity.  In the 1940's you were set for a good life providing you at least graduated high school.  Beyond that, up until the 1970's, it was technically feasible to support yourself and go to school with nothing more than a part time job (for a single person supporting only themselves) without resorting to financial aid or student loans.  Back then, college wasn't about securing a "good" job, you already had that secured with a high school diploma.  College was about expanding your horizons<mask> a human being, which is<mask> college _used_ to have much more variety and focus on things like philosophy and the arts, and now it seems like nothing<mask> STEM majors and a few stragglers in the humanities that everyone rips on<mask> it doesn't "gaurantee them a job." [NEWLINE] [NEWLINE] So welcome to the "privatized" and "<mask> much better" future<mask> for-profit colleges reign, and public colleges slowly privatize<mask> they can to stay afloat. <mask> it seems like you are paying too much for too little gain in college, you are right, you are getting very little gain in terms of marketplace value,<mask> college wasn't ever originally aimed at producing people who could pay off insane student loans.  Originally student loans were nonexistent and public funds plus student tuition was enough to run a successful institution of learning.  Only recently have banks been allowed to turn them into nothing<mask> a rip-off, asking exorbitant amounts of money for an education you could likely get on your own by visiting the
Label encoding: <s>So the part of your post I intend to challenge is specifically the "college was only created to make money" aspect. [NEWLINE] [NEWLINE] I know all you young whippersnappers on Reddit have all grown up in this crazy world where glorified multinational industrial Capitalism has all but swallowed the planet whole and is quickly bankrupting its resources and destroying the social contracts the middle class had worked for years to build, but believe it not college wasn't invented to "make money." [NEWLINE] [NEWLINE] Initially, places of learning were more closely tied to the clergy, as being a "scholar" used to be reserved specifically for those who could read and write, and they most generally read and wrote about... you guessed it, religion and religious texts. [NEWLINE] [NEWLINE] Eventually, public schooling became a liberal ideal of a world which benefited the "greater good" by educating as much as the populace as possible.  Honestly, up until recently, a college degree _wasn't_ viewed as a necessity.  In the 1940's you were set for a good life providing you at least graduated high school.  Beyond that, up until the 1970's, it was technically feasible to support yourself and go to school with nothing more than a part time job (for a single person supporting only themselves) without resorting to financial aid or student loans.  Back then, college wasn't about securing a "good" job, you already had that secured with a high school diploma.  College was about expanding your horizons as a human being, which is why college _used_ to have much more variety and focus on things like philosophy and the arts, and now it seems like nothing but STEM majors and a few stragglers in the humanities that everyone rips on because it doesn't "gaurantee them a job." [NEWLINE] [NEWLINE] So welcome to the "privatized" and " so much better" future where for-profit colleges reign, and public colleges slowly privatize what they can to stay afloat.  If it seems like you are paying too much for too little gain in college, you are right, you are getting very little gain in terms of marketplace value, but college wasn't ever originally aimed at producing people who could pay off insane student loans.  Originally student loans were nonexistent and public funds plus student tuition was enough to run a successful institution of learning.  Only recently have banks been allowed to turn them into nothing but a rip-off, asking exorbitant amounts of money for an education you could likely get on your own by visiting the
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can't comment on the actual quality of his writing, all I know is that I and many others really enjoy it. Thats really all I need to consider a writer 'good'.<mask> you don't enjoy his writing style then I can't really change that.<mask> I have noticed about myself and is presumably true of others is that<mask> I don't like the themes of a book I am much more critical of its writing style (lookin' at you Ayn Rand). I<mask> know I see people criticising both style and content at the same time more often than I see people criticise one and praise the other. Maybe you do this too? [NEWLINE] [NEWLINE] On to the themes: [NEWLINE] [NEWLINE] [STARTQ] rape scenes, references to rape, gang rape, threats of rape, senseless nudity, high percentage of female characters who are prostitutes, sexual assault on any character<mask> especially those who are minors, [ENDQ] [NEWLINE] Martin doesn't want to hide parts of his characters lives from us. People have a lot of sex, and<mask> do Martins characters. And unfortunately rape happens. Especially during wartime, and especially in societies<mask> women are denied agency. I don't see<mask> Martin can express this tragedy better than by describing it. I mean, there are plenty of strongly feminist novels that deal with rape and describe rape and we hardly call them mysogynist. Prostitution was pretty common during the Middle Ages,<mask> its inclusion. Keep in mind that during the Middle Ages there was no conception of a minor. People certainly wern't having sex with children,<mask> teenagers were considered adults.<mask> Joffrey is<mask> we would consider a minor. [NEWLINE] [NEWLINE] Plenty of fantasy presents a whitewashed medieval setting full of rightful princes, love marriages and wars<mask> all the suffering is done on the battlefield. Martin offers something a little closer to reality even<mask> his setting is only pseudo-medieval. [NEWLINE] [NEWLINE] [STARTQ] constant references to a non-gender-conformist character<mask> ugly and homely [ENDQ] [NEWLINE] Brienne right? I actually think thats one of the strengths of Briennes character. Too often in fantasy I feel like female characters are allowed to be skilled warriors or talented mages or clever theives<mask><mask><mask> they are still hot. They arn't given the muscles you need to weild a heavy metal sword and wear heavy metal armour and they are expected to contort themselves in such a way<mask> to show off their arse and boobs at the same time<mask> fighting off skeletons or whatever. Briennes character defies
Label encoding: <s>I can't comment on the actual quality of his writing, all I know is that I and many others really enjoy it. Thats really all I need to consider a writer 'good'. If you don't enjoy his writing style then I can't really change that. What I have noticed about myself and is presumably true of others is that when I don't like the themes of a book I am much more critical of its writing style (lookin' at you Ayn Rand). I also know I see people criticising both style and content at the same time more often than I see people criticise one and praise the other. Maybe you do this too? [NEWLINE] [NEWLINE] On to the themes: [NEWLINE] [NEWLINE] [STARTQ] rape scenes, references to rape, gang rape, threats of rape, senseless nudity, high percentage of female characters who are prostitutes, sexual assault on any character but especially those who are minors, [ENDQ] [NEWLINE] Martin doesn't want to hide parts of his characters lives from us. People have a lot of sex, and so do Martins characters. And unfortunately rape happens. Especially during wartime, and especially in societies where women are denied agency. I don't see how Martin can express this tragedy better than by describing it. I mean, there are plenty of strongly feminist novels that deal with rape and describe rape and we hardly call them mysogynist. Prostitution was pretty common during the Middle Ages, hence its inclusion. Keep in mind that during the Middle Ages there was no conception of a minor. People certainly wern't having sex with children, but teenagers were considered adults. Also Joffrey is what we would consider a minor. [NEWLINE] [NEWLINE] Plenty of fantasy presents a whitewashed medieval setting full of rightful princes, love marriages and wars where all the suffering is done on the battlefield. Martin offers something a little closer to reality even if his setting is only pseudo-medieval. [NEWLINE] [NEWLINE] [STARTQ] constant references to a non-gender-conformist character as ugly and homely [ENDQ] [NEWLINE] Brienne right? I actually think thats one of the strengths of Briennes character. Too often in fantasy I feel like female characters are allowed to be skilled warriors or talented mages or clever theives as long as they are still hot. They arn't given the muscles you need to weild a heavy metal sword and wear heavy metal armour and they are expected to contort themselves in such a way as to show off their arse and boobs at the same time while fighting off skeletons or whatever. Briennes character defies
Loss: tensor(0.2197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1864, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would definitely agree that Hugo Chavez was a "Bolivarian revolutionary" and that his system of government had strong socialist leanings,<mask> keep in mind that his government's hostility toward the United States had less to do with ideology than the knowledge of U.S. interventionism in Latin America.<mask> of his resistance to U.S. interests in the region, the U.S. government (and its counterpart in the loyal beltway media) preferred to represent him<mask> a strongman/authoritarian/dictator<mask> in reality the elections in Venezuela were essentially free and fair, and Chavez maintained popular support throughout his reign, especially among the working classes in that nation. The limits of Chavez's electoral power can be seen,<mask> well,<mask> he lost the 2007 referendum to implement the more controversial aspects of his socialist agenda, and he accepted that defeat<mask> the will of the people (not quite the tactic taken by a dictator). [NEWLINE] [NEWLINE] The problem with Castro,<mask><mask><mask> I'm concerned, is that he failed to focus on domestic concerns<mask> of the (serious) security threat posed by the United States during the first years of his regime.<mask>, he was dealing with crippling economic sanctions and a deeply skeptical ally in the Soviet Union, which ultimately removed his access to defensive nuclear weapons capabilities to avoid escalation between the USSR and the US. With that kind of serious security threat (an aggressive superpower ninety miles from your coast), and the crippling economic sanctions imposed on an island nation just out of a destabilizing revolution, combined to make successful improvements in the lives of Cubans all<mask> impossible for decades following the revolution. Not that<mask><mask> Castro was a phenomenal leader,<mask> much of the failures we attribute to his communist regime were really caused by factors outside of his control.<mask>, had the US (right up to the present day) not taken such a harsh anti-Castro position on Cuba, to the point of being irrational (the Cold War ended twenty five years ago,<mask> we still maintain an embargo), Cuba would likely be a much more free and open society today. [NEWLINE] [NEWLINE] Even the USSR only had one strongman dictator, in Stalin. Following his death in 1953, the process of "destalinization" and serious efforts at reform under Khrushchev, and then Brezhnev, and ultimately Gorbachev, increasingly brought the USSR into a more free and open society than had existed in the decades after the Russian revolution. Of course, the USSR backed multiple dictators in eastern Europe,<mask> that crime was committed
Label encoding: <s>I would definitely agree that Hugo Chavez was a "Bolivarian revolutionary" and that his system of government had strong socialist leanings, but keep in mind that his government's hostility toward the United States had less to do with ideology than the knowledge of U.S. interventionism in Latin America. Because of his resistance to U.S. interests in the region, the U.S. government (and its counterpart in the loyal beltway media) preferred to represent him as a strongman/authoritarian/dictator when in reality the elections in Venezuela were essentially free and fair, and Chavez maintained popular support throughout his reign, especially among the working classes in that nation. The limits of Chavez's electoral power can be seen, as well, when he lost the 2007 referendum to implement the more controversial aspects of his socialist agenda, and he accepted that defeat as the will of the people (not quite the tactic taken by a dictator). [NEWLINE] [NEWLINE] The problem with Castro, as far as I'm concerned, is that he failed to focus on domestic concerns because of the (serious) security threat posed by the United States during the first years of his regime. Additionally, he was dealing with crippling economic sanctions and a deeply skeptical ally in the Soviet Union, which ultimately removed his access to defensive nuclear weapons capabilities to avoid escalation between the USSR and the US. With that kind of serious security threat (an aggressive superpower ninety miles from your coast), and the crippling economic sanctions imposed on an island nation just out of a destabilizing revolution, combined to make successful improvements in the lives of Cubans all but impossible for decades following the revolution. Not that I think Castro was a phenomenal leader, but much of the failures we attribute to his communist regime were really caused by factors outside of his control. Additionally, had the US (right up to the present day) not taken such a harsh anti-Castro position on Cuba, to the point of being irrational (the Cold War ended twenty five years ago, yet we still maintain an embargo), Cuba would likely be a much more free and open society today. [NEWLINE] [NEWLINE] Even the USSR only had one strongman dictator, in Stalin. Following his death in 1953, the process of "destalinization" and serious efforts at reform under Khrushchev, and then Brezhnev, and ultimately Gorbachev, increasingly brought the USSR into a more free and open society than had existed in the decades after the Russian revolution. Of course, the USSR backed multiple dictators in eastern Europe, but that crime was committed
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Hick responds to this by suggesting that evil is a necessity for the ascension of man. It is better to earn something through hard work and personal growth rather than simply being ascribed it, similar to<mask> we view people who "pulled themselves up by their bootstraps" in a more positive light than we view people born into money. By enduring and responding to evil, we grow and become virtuous in a way that is impossible<mask> we were simply granted these attributes. [ENDQ] [NEWLINE] Did god have to earn his/her virtue through hard work and personal growth?<mask> not, then<mask> does he/she demand it of humans? And hardship is not the same<mask> evil. You can have hardship and personal growth without people being murdered and raped. [NEWLINE] [NEWLINE] [STARTQ] A question that is always raised<mask> this species of discussion occurs is the problem of<mask> bad things happen to good people (<mask> much<mask> that nearly and entire book of the bible deals with solving this problem). Hick responds by asserting that,<mask> bad things only happened to bad people, people would be good only<mask> they did not want to risk divine punishment rather than doing<mask><mask> being good is good in itself, negating the lessons and growth prompted by evil in the first place. [ENDQ] [NEWLINE] Isn't this the entire premise of heaven and hell? Aren't they supposed to be a divine reward/punishment for being good/bad?<mask> god doesn't want people to be good out of fear of divine consequences, then<mask> create heaven and hell? [NEWLINE] [NEWLINE] And<mask> this were true, then shouldn't we<mask> apply this reasoning to our laws? Laws<mask> create rewards and punishments for good and bad behavior.<mask> god doesn't want people to behave morally out of fear of consequences, then we should get rid of all of our laws,<mask> trying to influence moral behavior with punishment goes against god's will. [NEWLINE] [NEWLINE] [STARTQ] A less common response to Hick is that,<mask> some evil may be necessary, far too terrible and far too much evil is present in the world for Hick's argument to justify it. Hick responds by suggesting that the human perception of evil is a matter of relativity.<mask> the most evil thing in the universe that humans know of, say EVILX, were removed, the second most evil thing, EVILX-1 would now be perceived<mask> just<mask> bad an evil<mask> EVILX, and<mask> on, until we arrive at EVIL0, that is, no evil at all, interfering with the creation of the goodly people. [ENDQ]
Label encoding: <s> [STARTQ] Hick responds to this by suggesting that evil is a necessity for the ascension of man. It is better to earn something through hard work and personal growth rather than simply being ascribed it, similar to how we view people who "pulled themselves up by their bootstraps" in a more positive light than we view people born into money. By enduring and responding to evil, we grow and become virtuous in a way that is impossible if we were simply granted these attributes. [ENDQ] [NEWLINE] Did god have to earn his/her virtue through hard work and personal growth? If not, then why does he/she demand it of humans? And hardship is not the same as evil. You can have hardship and personal growth without people being murdered and raped. [NEWLINE] [NEWLINE] [STARTQ] A question that is always raised when this species of discussion occurs is the problem of why bad things happen to good people ( so much so that nearly and entire book of the bible deals with solving this problem). Hick responds by asserting that, if bad things only happened to bad people, people would be good only because they did not want to risk divine punishment rather than doing so because being good is good in itself, negating the lessons and growth prompted by evil in the first place. [ENDQ] [NEWLINE] Isn't this the entire premise of heaven and hell? Aren't they supposed to be a divine reward/punishment for being good/bad? If god doesn't want people to be good out of fear of divine consequences, then why create heaven and hell? [NEWLINE] [NEWLINE] And if this were true, then shouldn't we also apply this reasoning to our laws? Laws also create rewards and punishments for good and bad behavior. If god doesn't want people to behave morally out of fear of consequences, then we should get rid of all of our laws, since trying to influence moral behavior with punishment goes against god's will. [NEWLINE] [NEWLINE] [STARTQ] A less common response to Hick is that, while some evil may be necessary, far too terrible and far too much evil is present in the world for Hick's argument to justify it. Hick responds by suggesting that the human perception of evil is a matter of relativity. If the most evil thing in the universe that humans know of, say EVILX, were removed, the second most evil thing, EVILX-1 would now be perceived as just as bad an evil as EVILX, and so on, until we arrive at EVIL0, that is, no evil at all, interfering with the creation of the goodly people. [ENDQ]
Loss: tensor(0.2767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0961, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, people with more money buy more stuff.<mask>'s your point? [NEWLINE] <mask> economic activity is based entirely on<mask> gets bought. <mask> people do not get more money to buy stuff, the economy stagnates. <mask> 'd the economy crash in 2008?  Credit dried up, and people had to make up the negative savings rate with wages and not buying stuff.  GDP went severely negative, basically until enough people had paid off enough debt that they could start buying stuff again.  The fact that you can't make the connection between people buying stuff and people having enough MONEY to buy stuff is astounding. [NEWLINE] Here's another fun fact.  Your pay has little to nothing to do with the value you add to a product.  Your pay is determined by your NEGOTIATING POSITION. <mask> you can add millions of dollars of value,<mask> have a crappy negotiating position, you're not going to get paid squat (Nicola Tesla). <mask><mask><mask><mask>,<mask> you are marginal in<mask> you do,<mask> you can convince a company to shell out big bucks for you, then you're going to be rolling in dough.  After all, CEO pay has almost ZERO correlation with company performance. [NEWLINE] Commodity prices. <mask> are they an indicator of utilization of the means of production?  Supply and demand. <mask> mines worldwide are able to produce X amount of iron ore a year,<mask> companies only need x-25%, then there will be mines that are closed, and the price of iron ore (and iron, and steel) will be correspondingly low,<mask> well. <mask>,<mask> demand for iron ore is x+ 25%, then the price of iron ore, steel, and iron will be sky high, ALL the mines that can produce iron ore will be running<mask> hard<mask> they can, and commodity prices will be high. [NEWLINE] <mask> they're not. [NEWLINE] The market is underutilized. [NEWLINE] Wages weren't stagnant for anyone. [NEWLINE] You see, you're going to have to show<mask> you're getting your numbers.  I'm getting mine from Pew. [NEWLINE] [URL] / [NEWLINE] Other measures of income inequality are nasty,<mask> well. [NEWLINE] [URL] [NEWLINE] Rich people spend just<mask> much of their income<mask> poor people. [NEWLINE] No.  Period.  They don't.  Period.  One of the fundamentals ways you get rich is by spending LESS than you make. [NEWLINE] The fundamental way you stay poor is you spend EVERYTHING you make.
Label encoding: <s>So, people with more money buy more stuff. What's your point? [NEWLINE] Because economic activity is based entirely on what gets bought.  If people do not get more money to buy stuff, the economy stagnates.  Why 'd the economy crash in 2008?  Credit dried up, and people had to make up the negative savings rate with wages and not buying stuff.  GDP went severely negative, basically until enough people had paid off enough debt that they could start buying stuff again.  The fact that you can't make the connection between people buying stuff and people having enough MONEY to buy stuff is astounding. [NEWLINE] Here's another fun fact.  Your pay has little to nothing to do with the value you add to a product.  Your pay is determined by your NEGOTIATING POSITION.  If you can add millions of dollars of value, but have a crappy negotiating position, you're not going to get paid squat (Nicola Tesla).  On the other hand, if you are marginal in what you do, but you can convince a company to shell out big bucks for you, then you're going to be rolling in dough.  After all, CEO pay has almost ZERO correlation with company performance. [NEWLINE] Commodity prices.  How are they an indicator of utilization of the means of production?  Supply and demand.  If mines worldwide are able to produce X amount of iron ore a year, but companies only need x-25%, then there will be mines that are closed, and the price of iron ore (and iron, and steel) will be correspondingly low, as well.  But, if demand for iron ore is x+ 25%, then the price of iron ore, steel, and iron will be sky high, ALL the mines that can produce iron ore will be running as hard as they can, and commodity prices will be high. [NEWLINE] But they're not. [NEWLINE] The market is underutilized. [NEWLINE] Wages weren't stagnant for anyone. [NEWLINE] You see, you're going to have to show where you're getting your numbers.  I'm getting mine from Pew. [NEWLINE] [URL] / [NEWLINE] Other measures of income inequality are nasty, as well. [NEWLINE] [URL] [NEWLINE] Rich people spend just as much of their income as poor people. [NEWLINE] No.  Period.  They don't.  Period.  One of the fundamentals ways you get rich is by spending LESS than you make. [NEWLINE] The fundamental way you stay poor is you spend EVERYTHING you make.
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1963, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The automation issue really depends on<mask> well government can actually operate in the best interests of the people, and/or the richest are sufficiently superrational to recognize their imediate self-interests work against their ultimate best interests, and bind themselves to a common economic good. [NEWLINE] [NEWLINE] <mask>, it is very likely that all jobs will eventually become obsolete. Once machines are capable of doing *everything* that a human being is capable of doing, and doing it better and more efficiently, then there isn't a single thing that a human can offer to do that can outdo the machine.<mask> you think there is such a task, remember that human beings are just machines<mask> well; we are simply machines built by a slow, complex process. And we're pretty inefficient for most things. We simply don't have the technology<mask> for the vast majority of human capabilities.<mask> it will happen. [NEWLINE] [NEWLINE] Given the range of capabilities of humans, and even the range of the maximum *potential* across the population,<mask> technology gets better and better it starts to squeeze us into fewer and fewer capabilities at the top of our potential skillset. Even before we get there a large portion of the population will be below that threshold and simply cannot do anything of value that a machine can't do more efficiently (at lower cost in terms of even calories/energy).<mask> it takes us years to get an education for decent adult knowledge, and a machine can copy that knowledge in a fraction of a second, there just isn't any hope for investing in human work. [NEWLINE] [NEWLINE] <mask> there are two sides to this coin. One side is that there will be no jobs. The other side will be that anything and everything we need can be provided ultra-cheap for almost no cost<mask> there is no human labour involved. The machines are perpetual slaves producing for us, including generating their own power. [NEWLINE] [NEWLINE] <mask> one version says we all live in a utopia<mask> everything is provided for us without lifting a finger. We see the movement toward this with things like shorter work weeks<mask> still having enough income for a decent lifestyle, or in movements toward universal/basic income to replace welfare. [NEWLINE] [NEWLINE] The other version says that the rich will own all of the automation for their own purposes and the bulk of the population will be clawing at the gates of their castles for a morsel of food. We see movement in this direction by libertarian economic movements and disdain for anybody who can't be productive. [NEWLINE] [NEWLINE] Which way it ends up will be interesting. The utopian
Label encoding: <s>The automation issue really depends on how well government can actually operate in the best interests of the people, and/or the richest are sufficiently superrational to recognize their imediate self-interests work against their ultimate best interests, and bind themselves to a common economic good. [NEWLINE] [NEWLINE] Indeed, it is very likely that all jobs will eventually become obsolete. Once machines are capable of doing *everything* that a human being is capable of doing, and doing it better and more efficiently, then there isn't a single thing that a human can offer to do that can outdo the machine. If you think there is such a task, remember that human beings are just machines as well; we are simply machines built by a slow, complex process. And we're pretty inefficient for most things. We simply don't have the technology yet for the vast majority of human capabilities. But it will happen. [NEWLINE] [NEWLINE] Given the range of capabilities of humans, and even the range of the maximum *potential* across the population, as technology gets better and better it starts to squeeze us into fewer and fewer capabilities at the top of our potential skillset. Even before we get there a large portion of the population will be below that threshold and simply cannot do anything of value that a machine can't do more efficiently (at lower cost in terms of even calories/energy). When it takes us years to get an education for decent adult knowledge, and a machine can copy that knowledge in a fraction of a second, there just isn't any hope for investing in human work. [NEWLINE] [NEWLINE] But there are two sides to this coin. One side is that there will be no jobs. The other side will be that anything and everything we need can be provided ultra-cheap for almost no cost because there is no human labour involved. The machines are perpetual slaves producing for us, including generating their own power. [NEWLINE] [NEWLINE] So one version says we all live in a utopia where everything is provided for us without lifting a finger. We see the movement toward this with things like shorter work weeks while still having enough income for a decent lifestyle, or in movements toward universal/basic income to replace welfare. [NEWLINE] [NEWLINE] The other version says that the rich will own all of the automation for their own purposes and the bulk of the population will be clawing at the gates of their castles for a morsel of food. We see movement in this direction by libertarian economic movements and disdain for anybody who can't be productive. [NEWLINE] [NEWLINE] Which way it ends up will be interesting. The utopian
Loss: tensor(0.2048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0979, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would actually make the case that bullying in general is given far too much attention. For context, I am a 16 year old male, who was bullied pretty seriously through all of middle school, and the first year and a half or<mask> of highschool. [NEWLINE] [NEWLINE] Here is<mask> I would say: yes, being bullied sucks. It's uncomfortable,  harsh, and generally bad. That doesn't mean it's a problem of epic proportions that deems recognition at a national level,<mask> the anti-bullying programs and related media would have us believe. [NEWLINE] [NEWLINE] Frankly, discrimination of one sort or another has taken place in our society pretty much<mask> its inception. It's a shame that people pick on each other to establish a social pecking order,<mask> it happens. It would be great<mask> we could wave a magic wand and make it go away,<mask> we can't.<mask><mask><mask> I can tell, the only strategy put forth by anti-bullying groups is to make verbal harassment punishable in schools. [NEWLINE] [NEWLINE] The whole idea of limiting acceptable speech like this is completely preposterous. I am part of school that instituted such a policy, with pernicious results. First and foremost, it didn't work; people (including me) were still picked on<mask> matter of habit. The idea of "bullying" was<mask> determined to be defined by the victim, which created an influx of students who made up accusations to get back at kids they hated, and allowed allowed obnoxious students who had been excluded from some party or get together to punish their peers. [NEWLINE] [NEWLINE] The rule<mask> pertained to "Cyber Bullying", which was decided to be anybody being mean on the internet. I, who was being bullied at the time, found this to be patently ridiculous.<mask>, the school had no way of knowing who was doing the alleged bullying,<mask> people have the luxury of anonymity on the internet. Attempts to track down perpetrators were completely hopeless.<mask>, I don't believe the school had any right to punish students for the things they did outside the classroom. They exist to educate children, and<mask> they may have to be disciplinarians, it is only for the purpose of keeping children focused in the classroom. They do not exist to nanny teenager's lives. [NEWLINE] [NEWLINE] I didn't believe, and still don't, that cyber bullying is anything more than a media scare. I have spent much of my teenage years going through various forums, and have been insulted a thousand and one ways by random
Label encoding: <s>I would actually make the case that bullying in general is given far too much attention. For context, I am a 16 year old male, who was bullied pretty seriously through all of middle school, and the first year and a half or so of highschool. [NEWLINE] [NEWLINE] Here is what I would say: yes, being bullied sucks. It's uncomfortable,  harsh, and generally bad. That doesn't mean it's a problem of epic proportions that deems recognition at a national level, as the anti-bullying programs and related media would have us believe. [NEWLINE] [NEWLINE] Frankly, discrimination of one sort or another has taken place in our society pretty much since its inception. It's a shame that people pick on each other to establish a social pecking order, but it happens. It would be great if we could wave a magic wand and make it go away, but we can't. Insofar as I can tell, the only strategy put forth by anti-bullying groups is to make verbal harassment punishable in schools. [NEWLINE] [NEWLINE] The whole idea of limiting acceptable speech like this is completely preposterous. I am part of school that instituted such a policy, with pernicious results. First and foremost, it didn't work; people (including me) were still picked on as matter of habit. The idea of "bullying" was also determined to be defined by the victim, which created an influx of students who made up accusations to get back at kids they hated, and allowed allowed obnoxious students who had been excluded from some party or get together to punish their peers. [NEWLINE] [NEWLINE] The rule also pertained to "Cyber Bullying", which was decided to be anybody being mean on the internet. I, who was being bullied at the time, found this to be patently ridiculous. Firstly, the school had no way of knowing who was doing the alleged bullying, as people have the luxury of anonymity on the internet. Attempts to track down perpetrators were completely hopeless. Secondly, I don't believe the school had any right to punish students for the things they did outside the classroom. They exist to educate children, and while they may have to be disciplinarians, it is only for the purpose of keeping children focused in the classroom. They do not exist to nanny teenager's lives. [NEWLINE] [NEWLINE] I didn't believe, and still don't, that cyber bullying is anything more than a media scare. I have spent much of my teenage years going through various forums, and have been insulted a thousand and one ways by random
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>On paper, it might sound good to have a free game that draws you in and gives you perks for monetary donations<mask> you like<mask> you see.<mask> in practice, it completely changes a video game into a psychologically aggressive environment which is always reminding you<mask> you're not getting<mask> you don't pay real money. These games are essentially built to bully you out of your money and give nothing back of any value. [NEWLINE] [NEWLINE] This goes double for "pay to win" games,<mask><mask> applies<mask> it's only ascetic. [NEWLINE] [NEWLINE] Someone once told me an excellent analogy. Pretend you're shopping at a department store and you find some awesome clothes that are ridiculously cheap,<mask> you decide to buy them.<mask><mask> you get to the checkout counter, you find out that you're not allowed to take the clothes out of the store.<mask> they're "yours", you can only own them within the confines of the store.<mask> you find yourself coming back to the store everyday<mask> you really want to wear those awesome clothes you purchased.<mask><mask> you're walking around the store wearing your awesome clothes, you have lots of time to see everything else that's on sale until you're tempted to buy more.<mask>, other shoppers will see you wearing your awesome clothes and be jealous of you, and want to buy those clothes for themselves.<mask> in the end, the act of purchasing your clothes gave you nothing of value (<mask> let's be honest, you never actually owned anything), and only enabled the store to further advertize to you and indirectly to your friends. It's like the store didn't even sell anything,<mask> still wanted your money and used empty promises and peer pressure to get it out of you. In short, a scam. [NEWLINE] [NEWLINE] Another good analogy is that paying for stuff in a F2P game is like creating a horcrux. It doesn't have any actual value on its own; you give it value by paying for something, which is like putting part of your soul into the game that you can't take back out. Once you do it, you find it hard to separate yourself from the game without feeling like you've lost something important. You get emotionally invested and you can't let it go,<mask><mask> the only real value in the game is the artificial value you gave it by depositing your money into it. From there, it creates a self-perpetuating "Gambler's Fallacy" which preys on your psychological weakness and makes you spend more money. [NEWLINE] [NEWLINE] I
Label encoding: <s>On paper, it might sound good to have a free game that draws you in and gives you perks for monetary donations if you like what you see. But in practice, it completely changes a video game into a psychologically aggressive environment which is always reminding you what you're not getting if you don't pay real money. These games are essentially built to bully you out of your money and give nothing back of any value. [NEWLINE] [NEWLINE] This goes double for "pay to win" games, but also applies when it's only ascetic. [NEWLINE] [NEWLINE] Someone once told me an excellent analogy. Pretend you're shopping at a department store and you find some awesome clothes that are ridiculously cheap, so you decide to buy them. But when you get to the checkout counter, you find out that you're not allowed to take the clothes out of the store. While they're "yours", you can only own them within the confines of the store. So you find yourself coming back to the store everyday because you really want to wear those awesome clothes you purchased. So while you're walking around the store wearing your awesome clothes, you have lots of time to see everything else that's on sale until you're tempted to buy more. Also, other shoppers will see you wearing your awesome clothes and be jealous of you, and want to buy those clothes for themselves. So in the end, the act of purchasing your clothes gave you nothing of value ( because let's be honest, you never actually owned anything), and only enabled the store to further advertize to you and indirectly to your friends. It's like the store didn't even sell anything, but still wanted your money and used empty promises and peer pressure to get it out of you. In short, a scam. [NEWLINE] [NEWLINE] Another good analogy is that paying for stuff in a F2P game is like creating a horcrux. It doesn't have any actual value on its own; you give it value by paying for something, which is like putting part of your soul into the game that you can't take back out. Once you do it, you find it hard to separate yourself from the game without feeling like you've lost something important. You get emotionally invested and you can't let it go, even though the only real value in the game is the artificial value you gave it by depositing your money into it. From there, it creates a self-perpetuating "Gambler's Fallacy" which preys on your psychological weakness and makes you spend more money. [NEWLINE] [NEWLINE] I
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1929, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm in the military.  I deployed to Iraq.  I've read your entire post and your edits<mask> not any comment.  I just want to drop my two cents about programs like this.  Especially in your fourth comment about people have mentioned these programs do currently exist and have mixed results of success. [NEWLINE] [NEWLINE] The problem with establishing or revising a program to help people re-integrate is that no program will ever work 100%.  Each person has to deal with the reintegration into society in their own way.  You may want to spend time with friends and family.  Me?  I want to spend time with a good book and computer games for a few days.  Others may want to go to a location.  Some may want to avoid family and friends and any stimulus and just be alone and have silence to deal with things on their own.  The ways that people deal are<mask> varied<mask> people are different.  No program can handle that.  The programs that do exist do their best. <mask> I came back from Iraq there was a 4 hour briefing and I got two weeks off.  That was it.  I believe its more involved that now.  It worked for me. <mask> they wanted me to sit in a group setting or with a counselor I would have been miserable and it would have delayed or maybe prohibited my ability to re-integrate into normal society. [NEWLINE] [NEWLINE] You mention the suicide rates are high.  Are the suicides directly linked to deployments?  I don't think<mask>.  I am certain there is some percentage of them, yes,<mask> not all of them.  This year the entire military is going through a draw-down<mask> our government does not value a large, strong military.  The Air Force is downsizing by 35K, the Army by 50-60K, the Navy has gone through it recently.  You want to talk about stress?  You've commited 13 years in service to your country and you're about to get involuntarily kicked out and the only thing you are going to get is a lump sum check that's laughable.  Other stressors include just the day-to-day work.  Some people work in rooms doing top secret work.  That work can be stressful day to day.  To go home and not be able to talk to your family or friends about<mask> you do.  Just a simple, I'm going to work and I do "stuff".  Stress.  This type of
Label encoding: <s>I'm in the military.  I deployed to Iraq.  I've read your entire post and your edits but not any comment.  I just want to drop my two cents about programs like this.  Especially in your fourth comment about people have mentioned these programs do currently exist and have mixed results of success. [NEWLINE] [NEWLINE] The problem with establishing or revising a program to help people re-integrate is that no program will ever work 100%.  Each person has to deal with the reintegration into society in their own way.  You may want to spend time with friends and family.  Me?  I want to spend time with a good book and computer games for a few days.  Others may want to go to a location.  Some may want to avoid family and friends and any stimulus and just be alone and have silence to deal with things on their own.  The ways that people deal are as varied as people are different.  No program can handle that.  The programs that do exist do their best.  When I came back from Iraq there was a 4 hour briefing and I got two weeks off.  That was it.  I believe its more involved that now.  It worked for me.  If they wanted me to sit in a group setting or with a counselor I would have been miserable and it would have delayed or maybe prohibited my ability to re-integrate into normal society. [NEWLINE] [NEWLINE] You mention the suicide rates are high.  Are the suicides directly linked to deployments?  I don't think so.  I am certain there is some percentage of them, yes, but not all of them.  This year the entire military is going through a draw-down because our government does not value a large, strong military.  The Air Force is downsizing by 35K, the Army by 50-60K, the Navy has gone through it recently.  You want to talk about stress?  You've commited 13 years in service to your country and you're about to get involuntarily kicked out and the only thing you are going to get is a lump sum check that's laughable.  Other stressors include just the day-to-day work.  Some people work in rooms doing top secret work.  That work can be stressful day to day.  To go home and not be able to talk to your family or friends about what you do.  Just a simple, I'm going to work and I do "stuff".  Stress.  This type of
Loss: tensor(0.0844, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The idea of using the more specific language is to point out<mask> makes it typical. It's both typical for Males to avoid floral perfume and<mask> to have a larger shoulder-width to height ratio. The reason for the former is<mask> it is normative, the latter biological. Of course, 'typical' is a correct word. [NEWLINE] [NEWLINE] My point (well, admittedly, it's not really *my* point,<mask> a piece of an overall feminist philosophy) is that it's important to account for these reasons (that is, say, gender norms) in communicating, especially<mask> ignoring them is widely seen<mask> part of the problem in the first place. [NEWLINE] [NEWLINE] And that problem arises in the one case you seem to miss: [NEWLINE] [NEWLINE] [STARTQ]...men shouldn't wear floral perfume.<mask> both parties understand this, then the words typical and heteronormative mean the same thing, agreed? The same is true<mask> neither party understands this. [ENDQ] [NEWLINE] Agreed,<mask> :<mask> about the case<mask> one party understands and the other does not? This tends to be way it works in the case of marginalized populations. The marginalized individual vividly understands that they are not 'typical' purely<mask> they do/can not conform to the unwritten<mask> strictly enforced norms of their society. The cis-gendered, heteronormative observer might see that person<mask> atypical in a sense that actually warrants a certain level of criticism, dismissal etc.; or, they may recognize that the *reasons* they see this person<mask> atypical are not substantive, and<mask> accept them<mask> a person, even<mask> they violate the norms they are used to. The aim is to perpetuate the latter way of thinking, by involving talk of norms wherever norms are in play in deciding/evaluating/acting. Eventually, maybe, the'made-up word' won't be necessary anymore,<mask> there won't be a problem.<mask> for now, there is. [NEWLINE] [NEWLINE] [STARTQ] It perpetuates this idea that some people have that atypical must be taboo. [ENDQ] [NEWLINE] <mask><mask>,<mask><mask> it is actually an attempt to *deflate* the idea that some people have that atypical must be taboo. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> it's bad practice to make up new words to skip over concepts that are themselves simple and would be valuable to learn [ENDQ] [NEWLINE] <mask> I was trying to say is that using that made-up word is necessary to bring focus to the concept of norms,<mask> it's applicable. That's all.
Label encoding: <s>The idea of using the more specific language is to point out what makes it typical. It's both typical for Males to avoid floral perfume and also to have a larger shoulder-width to height ratio. The reason for the former is because it is normative, the latter biological. Of course, 'typical' is a correct word. [NEWLINE] [NEWLINE] My point (well, admittedly, it's not really *my* point, but a piece of an overall feminist philosophy) is that it's important to account for these reasons (that is, say, gender norms) in communicating, especially when ignoring them is widely seen as part of the problem in the first place. [NEWLINE] [NEWLINE] And that problem arises in the one case you seem to miss: [NEWLINE] [NEWLINE] [STARTQ]...men shouldn't wear floral perfume. If both parties understand this, then the words typical and heteronormative mean the same thing, agreed? The same is true if neither party understands this. [ENDQ] [NEWLINE] Agreed, but : What about the case when one party understands and the other does not? This tends to be way it works in the case of marginalized populations. The marginalized individual vividly understands that they are not 'typical' purely because they do/can not conform to the unwritten but strictly enforced norms of their society. The cis-gendered, heteronormative observer might see that person as atypical in a sense that actually warrants a certain level of criticism, dismissal etc.; or, they may recognize that the *reasons* they see this person as atypical are not substantive, and therefore accept them as a person, even if they violate the norms they are used to. The aim is to perpetuate the latter way of thinking, by involving talk of norms wherever norms are in play in deciding/evaluating/acting. Eventually, maybe, the'made-up word' won't be necessary anymore, because there won't be a problem. But for now, there is. [NEWLINE] [NEWLINE] [STARTQ] It perpetuates this idea that some people have that atypical must be taboo. [ENDQ] [NEWLINE] In fact, I think it is actually an attempt to *deflate* the idea that some people have that atypical must be taboo. [NEWLINE] [NEWLINE] [STARTQ] I think it's bad practice to make up new words to skip over concepts that are themselves simple and would be valuable to learn [ENDQ] [NEWLINE] What I was trying to say is that using that made-up word is necessary to bring focus to the concept of norms, when it's applicable. That's all.
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> this supports the OP's point. The idea of Asian homogeneity is itself racist, and no one pays attention. In the United States, the category "Asian" is the most diverse group. Indians, Indonesians, and Koreans are definitely very different people,<mask> they're all lumped together<mask> Asian. Even people from Northern China and Southern China are very different from each other. Furthermore, the poverty rate of the Hmong Americans (an Asian group you've probably never even heard about) is the same<mask> the poverty rate for African Americans and Hispanics (about 30%). [ENDQ] [NEWLINE] This argument was incorrectly answered. First "homogeneity" refers to the fact that the African American culture in the United States and throughout the world is accentuated much by the diaspora and fragmentation that other cultures faced, like Native Americans or people of Jewish ancestry. It is not to say that all Asians can be grouped together and minimized. [NEWLINE] [NEWLINE] [STARTQ] And this is<mask> Asians need more attention. There are many untrue stereotypes of Asians, such<mask> the homogeneity that you believe.<mask>, there are many disadvantaged Asians (such<mask> the Hmong and many others) who suffer even more<mask> they don't have the "social justice" support network that blacks and Hispanics have. [ENDQ] [NEWLINE] [STARTQ] <mask>,<mask><mask> with the fact that xenophobia and racism are different. Even<mask> they were, xenophobia can be just<mask> damaging. [ENDQ] [NEWLINE] <mask><mask> that a distinction should be made. Xenophobia is a fear and/or mischaracterization of the other that is inherently different from a more outward expression of dehumanization. [NEWLINE] [NEWLINE] [STARTQ] Just consider this:<mask> China rises in world influence, will Americans of Chinese descent need to worry that they will be put into camps just like the Japanese Americans were in WWII? [ENDQ] [NEWLINE] This lacks context and is conjecture. [NEWLINE] [NEWLINE] [STARTQ] Different, yes.<mask> different is not necessarily better or worse. My parents were immigrants from China, and<mask> I was born, they were sharing a 1 bedroom apartment with ten other people.<mask><mask> this apartment was in San Francisco, is that somehow better than growing up in the projects? [ENDQ] [NEWLINE] *Yes.* [NEWLINE] [NEWLINE] My story is anecdotal,<mask> I don't think these conditions are typical for immigrants from Asia. [NEWLINE] [NEWLINE] Fair. [NEWLINE] [NEWLINE] [STARTQ] Finally, I want to mention that you can't point out successful Asians to discount the suffering of less fortunate Asians. This is tantamount to pointing at successful African Americans in the NBA and concluding that African Americans are not disadvantaged.
Label encoding: <s> [STARTQ] I think this supports the OP's point. The idea of Asian homogeneity is itself racist, and no one pays attention. In the United States, the category "Asian" is the most diverse group. Indians, Indonesians, and Koreans are definitely very different people, but they're all lumped together as Asian. Even people from Northern China and Southern China are very different from each other. Furthermore, the poverty rate of the Hmong Americans (an Asian group you've probably never even heard about) is the same as the poverty rate for African Americans and Hispanics (about 30%). [ENDQ] [NEWLINE] This argument was incorrectly answered. First "homogeneity" refers to the fact that the African American culture in the United States and throughout the world is accentuated much by the diaspora and fragmentation that other cultures faced, like Native Americans or people of Jewish ancestry. It is not to say that all Asians can be grouped together and minimized. [NEWLINE] [NEWLINE] [STARTQ] And this is why Asians need more attention. There are many untrue stereotypes of Asians, such as the homogeneity that you believe. Also, there are many disadvantaged Asians (such as the Hmong and many others) who suffer even more because they don't have the "social justice" support network that blacks and Hispanics have. [ENDQ] [NEWLINE] [STARTQ] Also, I disagree with the fact that xenophobia and racism are different. Even if they were, xenophobia can be just as damaging. [ENDQ] [NEWLINE] I think that a distinction should be made. Xenophobia is a fear and/or mischaracterization of the other that is inherently different from a more outward expression of dehumanization. [NEWLINE] [NEWLINE] [STARTQ] Just consider this: as China rises in world influence, will Americans of Chinese descent need to worry that they will be put into camps just like the Japanese Americans were in WWII? [ENDQ] [NEWLINE] This lacks context and is conjecture. [NEWLINE] [NEWLINE] [STARTQ] Different, yes. But different is not necessarily better or worse. My parents were immigrants from China, and when I was born, they were sharing a 1 bedroom apartment with ten other people. Even though this apartment was in San Francisco, is that somehow better than growing up in the projects? [ENDQ] [NEWLINE] *Yes.* [NEWLINE] [NEWLINE] My story is anecdotal, but I don't think these conditions are typical for immigrants from Asia. [NEWLINE] [NEWLINE] Fair. [NEWLINE] [NEWLINE] [STARTQ] Finally, I want to mention that you can't point out successful Asians to discount the suffering of less fortunate Asians. This is tantamount to pointing at successful African Americans in the NBA and concluding that African Americans are not disadvantaged.
Loss: tensor(0.1538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From one nonreligious person to another. [NEWLINE] [NEWLINE] I have no issue with words that are considered profane. I don't use them much,<mask> I have no reservations about using them and have no issues with people using them (except for<mask> it is excessive,<mask> that is rare). [NEWLINE] [NEWLINE] Now here is<mask> I see there being an issue with swear words:<mask><mask> it does not offend you, someone in your audience (and this just means someone to whom you are talking or in your direct vicinity) might take offense. That person might take offense. I know you place the [NEWLINE] [NEWLINE] "just<mask> you're offended doesn't make you right." [NEWLINE] [NEWLINE] Which is true.<mask> keep in mind that<mask><mask> your morality states that profanity is fine, that will not matter much,<mask> at all, to a religious person whose morality dictates that profanity is unacceptable. [NEWLINE] [NEWLINE] <mask> you said, you try to prevent your children from swearing. I do not see you<mask> a hypocrite<mask> you have a great reason for it: they might get into the habit. [NEWLINE] [NEWLINE] My moral code is simple: does this action harm another individual in some way?<mask><mask>, analyze the action and determine<mask> the fault lies in the action or the person being harmed by it. [NEWLINE] [NEWLINE] It is a pretty good moral code,<mask> it ignores the fact that some have less control over their morality. At the risk of generalizing, it is more common for a religious person's morality to be guided by more than just oneself and the social situation. They have a more religiously-oriented moral code.<mask> you interpret<mask> their fault ("they are too sensitive"), they could interpret<mask> your fault ("he is trying to offend me"). [NEWLINE] [NEWLINE] Ultimately, profanity is perfectly fine, I am not arguing that. I AM saying that<mask> children do not<mask> have the full cognitive understanding that there are words that are inappropriate to say to certain people and/or during certain situations. Once they learn<mask> it is unacceptable (<mask> that<mask> they go to their friend's cousin's Bar Mitzvah or the such, they should not use profanity<mask> at the Synagogue), then it no longer is an issue. [NEWLINE] [NEWLINE] Let's consider for a moment other words. Not words that are profane<mask> ones which are inappropriate during a given situation.<mask>, FSM forbid, one of your children is kidnapped and killed, it would be EXTREMELY inappropriate for someone to tell you that "God has a plan for everyone and your kid is
Label encoding: <s>From one nonreligious person to another. [NEWLINE] [NEWLINE] I have no issue with words that are considered profane. I don't use them much, but I have no reservations about using them and have no issues with people using them (except for when it is excessive, but that is rare). [NEWLINE] [NEWLINE] Now here is where I see there being an issue with swear words: even though it does not offend you, someone in your audience (and this just means someone to whom you are talking or in your direct vicinity) might take offense. That person might take offense. I know you place the [NEWLINE] [NEWLINE] "just because you're offended doesn't make you right." [NEWLINE] [NEWLINE] Which is true. But keep in mind that even though your morality states that profanity is fine, that will not matter much, if at all, to a religious person whose morality dictates that profanity is unacceptable. [NEWLINE] [NEWLINE] As you said, you try to prevent your children from swearing. I do not see you as a hypocrite because you have a great reason for it: they might get into the habit. [NEWLINE] [NEWLINE] My moral code is simple: does this action harm another individual in some way? If so, analyze the action and determine if the fault lies in the action or the person being harmed by it. [NEWLINE] [NEWLINE] It is a pretty good moral code, but it ignores the fact that some have less control over their morality. At the risk of generalizing, it is more common for a religious person's morality to be guided by more than just oneself and the social situation. They have a more religiously-oriented moral code. What you interpret as their fault ("they are too sensitive"), they could interpret as your fault ("he is trying to offend me"). [NEWLINE] [NEWLINE] Ultimately, profanity is perfectly fine, I am not arguing that. I AM saying that since children do not yet have the full cognitive understanding that there are words that are inappropriate to say to certain people and/or during certain situations. Once they learn when it is unacceptable ( so that if they go to their friend's cousin's Bar Mitzvah or the such, they should not use profanity while at the Synagogue), then it no longer is an issue. [NEWLINE] [NEWLINE] Let's consider for a moment other words. Not words that are profane but ones which are inappropriate during a given situation. If, FSM forbid, one of your children is kidnapped and killed, it would be EXTREMELY inappropriate for someone to tell you that "God has a plan for everyone and your kid is
Loss: tensor(0.3774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1821, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'l begin by saying it has never happened to me, and you make some good points.<mask> I'm addicted to CMV,<mask> here we go. [NEWLINE] [NEWLINE] [STARTQ] sex is a very important part of a relationship,<mask> not the most important part. [ENDQ] [NEWLINE] <mask> it is one of the most important parts. To break it would be to break one of the most important parts of a marriage. I'll continue. [NEWLINE] [NEWLINE] *edit: even<mask> you value it not at all, cheating breaks other important parts of the relationship. Like trust. Loyalty. Companionship. [NEWLINE] [NEWLINE] [STARTQ] Nobody makes mistakes ever,<mask><mask> they do, the relationship should end. [ENDQ] [NEWLINE] You are correct that this is false.<mask>, cheating on a spouse is  *pretty big mistake*. Cheating,<mask> I define it, is a choice. A bad choice which breaks one of the cornerstones of marriage. [NEWLINE] [NEWLINE] [STARTQ] Cheaters will always cheat again [ENDQ] [NEWLINE] You are correct that this is false.<mask>, I would say that *cheaters are more likely to cheat again*. The circumstances that resulted in cheating the first time may repeat. [NEWLINE] [NEWLINE] [STARTQ] Cheaters cheat<mask> of underlying issues with the relationship. My belief:<mask> that's true, there are a lot of possibilities - not just breaking up - that could fix the problem, such<mask>... talking about the problem. [ENDQ] [NEWLINE] Yes.<mask> "talking about the problem" is an ideal situation. It often doesn't work out<mask> well that way. People get stuck in a pattern... living in the same house, married to the same person you stopped really caring about. Cheating is evidence that *something* must be wrong. [NEWLINE] [NEWLINE] <mask>,<mask> far we can say cheating [NEWLINE] [NEWLINE] * is a bad choice which breaks one of the cornerstones of a relationship [NEWLINE] [NEWLINE] * is an indication that such a thing will happen again [NEWLINE] [NEWLINE] * is an indication that there is a problem with the relationship [NEWLINE] [NEWLINE] Are there no situations in which the cheaters should be forgiven? I'm trying to<mask><mask><mask> you find your spouse cheating, breaking up with them is the most logical thing to do, whatever the circumstances. [NEWLINE] [NEWLINE] I point out, cheating has a profound negative emotional effect of the person cheated on. You have every reason to break up. [NEWLINE] [NEWLINE] In my mind, marriage is a commitment.<mask> you break that commitment, the marriage is broken. One partner isn't fulfilling all of his/her obligations (i.e. not to cheat)<mask>
Label encoding: <s>I'l begin by saying it has never happened to me, and you make some good points. But I'm addicted to CMV, so here we go. [NEWLINE] [NEWLINE] [STARTQ] sex is a very important part of a relationship, but not the most important part. [ENDQ] [NEWLINE] So it is one of the most important parts. To break it would be to break one of the most important parts of a marriage. I'll continue. [NEWLINE] [NEWLINE] *edit: even if you value it not at all, cheating breaks other important parts of the relationship. Like trust. Loyalty. Companionship. [NEWLINE] [NEWLINE] [STARTQ] Nobody makes mistakes ever, so when they do, the relationship should end. [ENDQ] [NEWLINE] You are correct that this is false. However, cheating on a spouse is  *pretty big mistake*. Cheating, as I define it, is a choice. A bad choice which breaks one of the cornerstones of marriage. [NEWLINE] [NEWLINE] [STARTQ] Cheaters will always cheat again [ENDQ] [NEWLINE] You are correct that this is false. However, I would say that *cheaters are more likely to cheat again*. The circumstances that resulted in cheating the first time may repeat. [NEWLINE] [NEWLINE] [STARTQ] Cheaters cheat because of underlying issues with the relationship. My belief: if that's true, there are a lot of possibilities - not just breaking up - that could fix the problem, such as... talking about the problem. [ENDQ] [NEWLINE] Yes. But "talking about the problem" is an ideal situation. It often doesn't work out so well that way. People get stuck in a pattern... living in the same house, married to the same person you stopped really caring about. Cheating is evidence that *something* must be wrong. [NEWLINE] [NEWLINE] So, so far we can say cheating [NEWLINE] [NEWLINE] * is a bad choice which breaks one of the cornerstones of a relationship [NEWLINE] [NEWLINE] * is an indication that such a thing will happen again [NEWLINE] [NEWLINE] * is an indication that there is a problem with the relationship [NEWLINE] [NEWLINE] Are there no situations in which the cheaters should be forgiven? I'm trying to argue that if you find your spouse cheating, breaking up with them is the most logical thing to do, whatever the circumstances. [NEWLINE] [NEWLINE] I point out, cheating has a profound negative emotional effect of the person cheated on. You have every reason to break up. [NEWLINE] [NEWLINE] In my mind, marriage is a commitment. If you break that commitment, the marriage is broken. One partner isn't fulfilling all of his/her obligations (i.e. not to cheat) So
Loss: tensor(0.1951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0977, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0996, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0934, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0918, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0819, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Until someone posts something showing that atheism is more commonly used to describe "lack of,<mask> not disbelief in God" [ENDQ] [NEWLINE] There's the trouble.  At the heart of the issue is that the atheist thinks that lack of belief is, essentially, disbelief. [NEWLINE] [NEWLINE] <mask> I said elsewhere: Do you believe in Santa Claus?<mask> would you say you don't and<mask> isn't the burden of proof on you to disprove every potential explanation for the existence of Santa Claus. [NEWLINE] [NEWLINE] You might be tempted to back up and say that you're agnostic about Santa Claus,<mask> I don't think that would be completely honest.  At no time in your life<mask> childhood did you ever behave in a way that would allow for the existence of Santa.  You've never counted on him to provide a present for someone you'd have otherwise bought a present for.  You've never felt bad about not leaving out cookies, and<mask> a person matching his description was in your house on Christmas Eve, you'd immediately call the police. <mask> you can't disprove Santa, you don't behave in a manner that allows for his existence.  You're saying that you don't know, that it's inconclusive,<mask> you act<mask><mask> you are very certain of his non-existence. [NEWLINE] [NEWLINE] The reason for this instinctive reaction is that ruling a thing false or untrue is always a matter of likelihood given evidence and never a matter of all logically possible explanation being ruled out.  'Can't be disproven' is a trait every last coherent idea has and isn't really worth anything in the day to day of your life, and you behave<mask> such. [NEWLINE] [NEWLINE] The Agnostic might feel that he's in a more honest philosophical position,<mask> he isn't. He's allowing God far more epistemological leeway than he does with any other judgement of falsehood.<mask> the best he can do is fail to disprove it, that should be good enough for<mask> we'd agree is disbelief<mask> everything you and I would agree is false at all would<mask> fail to be disproved by that standard with which you ask the atheist to disprove God. [NEWLINE] [NEWLINE] <mask><mask> I say that lack of belief is equivalent to disbelief, I<mask> imply that<mask> you were being honest, you'd agree with me<mask> evidenced by your behavior with regard to a number of things you can't disprove.  Lack of evidence asserting the things existence is more than enough of a standard for every other thing you'd ever say you disbel
Label encoding: <s> [STARTQ] Until someone posts something showing that atheism is more commonly used to describe "lack of, but not disbelief in God" [ENDQ] [NEWLINE] There's the trouble.  At the heart of the issue is that the atheist thinks that lack of belief is, essentially, disbelief. [NEWLINE] [NEWLINE] As I said elsewhere: Do you believe in Santa Claus? Why would you say you don't and why isn't the burden of proof on you to disprove every potential explanation for the existence of Santa Claus. [NEWLINE] [NEWLINE] You might be tempted to back up and say that you're agnostic about Santa Claus, but I don't think that would be completely honest.  At no time in your life since childhood did you ever behave in a way that would allow for the existence of Santa.  You've never counted on him to provide a present for someone you'd have otherwise bought a present for.  You've never felt bad about not leaving out cookies, and if a person matching his description was in your house on Christmas Eve, you'd immediately call the police.  While you can't disprove Santa, you don't behave in a manner that allows for his existence.  You're saying that you don't know, that it's inconclusive, but you act as if you are very certain of his non-existence. [NEWLINE] [NEWLINE] The reason for this instinctive reaction is that ruling a thing false or untrue is always a matter of likelihood given evidence and never a matter of all logically possible explanation being ruled out.  'Can't be disproven' is a trait every last coherent idea has and isn't really worth anything in the day to day of your life, and you behave as such. [NEWLINE] [NEWLINE] The Agnostic might feel that he's in a more honest philosophical position, but he isn't. He's allowing God far more epistemological leeway than he does with any other judgement of falsehood. If the best he can do is fail to disprove it, that should be good enough for what we'd agree is disbelief because everything you and I would agree is false at all would also fail to be disproved by that standard with which you ask the atheist to disprove God. [NEWLINE] [NEWLINE] So when I say that lack of belief is equivalent to disbelief, I also imply that if you were being honest, you'd agree with me as evidenced by your behavior with regard to a number of things you can't disprove.  Lack of evidence asserting the things existence is more than enough of a standard for every other thing you'd ever say you disbel
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2996, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1819, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm posting this here<mask> it got removed<mask> a direct response for not specifically challenging the post,<mask><mask><mask> it's relevant to your stance: [NEWLINE] [NEWLINE] I'm not going to try to change your mind concerning legalisation of drugs,<mask><mask><mask> with you.<mask>, I will offer you something to think about<mask> it comes to debating this topic and<mask> arguments work<mask> talking to others about it. [NEWLINE] [NEWLINE] You say you get a lot of hostile reactions<mask> you advocate legalisation from the standpoint of personal liberty and this reminded me immediately of this [fantastic discussion between Sam Harris and Johann Hari]( [URL] ) I read recently. I suggest you read it too, it's a brilliant look into the drug war debate and Hari's book looks to be even more incisive. [NEWLINE] [NEWLINE] The relevant section reads<mask> : [NEWLINE] [NEWLINE] *JH: "<mask><mask> one of the really important things, particularly in winning the debate in America, is to look at<mask> arguments won in these places and<mask> arguments didn’t. We found that in the places that successfully decriminalized or legalized, liberty-based arguments for ending the drug war were very unpopular. I’m philosophically sympathetic to the argument that it’s your body and you’ve got a right to do<mask> you want with it.<mask> it turns out that’s a politically toxic argument—people really don’t like it, and it only works with people who already agree. [NEWLINE] The arguments that work well in persuading the people we still want to reach are order-based arguments.<mask><mask> the Swiss heroin referenda are good models for that. Basically,<mask> they said was drug war means chaos. It means unknown criminals selling unknown chemicals to unknown users, all in the dark, in our public places, filled with disease and chaos. Legalization is a way of imposing regulation and order on this anarchy. It’s about taking it away from criminal gangs and giving it to doctors and pharmacists, and making sure it happens in nice clean clinics, and we get our nice parks back, and we reduce crime. That’s the argument that will win. And it’s not like it’s a rhetorical trick—it’s true. That is<mask> happens."* [NEWLINE] [NEWLINE] … [NEWLINE] [NEWLINE] *SH: "<mask><mask> it’s a great insight to emphasize the pragmatic case for legalization,<mask> opposed to the ethical one. It is always tempting to try to lead people through
Label encoding: <s>I'm posting this here as it got removed as a direct response for not specifically challenging the post, but I think it's relevant to your stance: [NEWLINE] [NEWLINE] I'm not going to try to change your mind concerning legalisation of drugs, as I agree with you. However, I will offer you something to think about when it comes to debating this topic and what arguments work when talking to others about it. [NEWLINE] [NEWLINE] You say you get a lot of hostile reactions when you advocate legalisation from the standpoint of personal liberty and this reminded me immediately of this [fantastic discussion between Sam Harris and Johann Hari]( [URL] ) I read recently. I suggest you read it too, it's a brilliant look into the drug war debate and Hari's book looks to be even more incisive. [NEWLINE] [NEWLINE] The relevant section reads thus : [NEWLINE] [NEWLINE] *JH: " I think one of the really important things, particularly in winning the debate in America, is to look at what arguments won in these places and what arguments didn’t. We found that in the places that successfully decriminalized or legalized, liberty-based arguments for ending the drug war were very unpopular. I’m philosophically sympathetic to the argument that it’s your body and you’ve got a right to do what you want with it. But it turns out that’s a politically toxic argument—people really don’t like it, and it only works with people who already agree. [NEWLINE] The arguments that work well in persuading the people we still want to reach are order-based arguments. I think the Swiss heroin referenda are good models for that. Basically, what they said was drug war means chaos. It means unknown criminals selling unknown chemicals to unknown users, all in the dark, in our public places, filled with disease and chaos. Legalization is a way of imposing regulation and order on this anarchy. It’s about taking it away from criminal gangs and giving it to doctors and pharmacists, and making sure it happens in nice clean clinics, and we get our nice parks back, and we reduce crime. That’s the argument that will win. And it’s not like it’s a rhetorical trick—it’s true. That is what happens."* [NEWLINE] [NEWLINE] … [NEWLINE] [NEWLINE] *SH: " I think it’s a great insight to emphasize the pragmatic case for legalization, as opposed to the ethical one. It is always tempting to try to lead people through
Loss: tensor(0.1271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I looked through the bios of all the head people and it made it clear BuzzFeed is all about getting things viral. [ENDQ] [NEWLINE] Well, obviously they're not going to come out and say "We're making listicles to make a profit." They need to be able to spin it positively - making viral content is a lot more understandable than just saying "We need money to fund other stuff." [NEWLINE] [NEWLINE] [STARTQ] Now everyone sees all this stupid viral list/quiz stuff and it has the same weight<mask> real news. [ENDQ] [NEWLINE] <mask><mask> this is<mask> you're wrong to include Buzzfeed. There are some sites which do "stupid viral" stuff in order to garner clicks that's presented in the form of "real news." Off the top of my head, that's sites like Gawker, Thinkprogress, Elite Daily, and occasionally HuffPost and Upworthy. Their viral content is very politically charged, often designed to press the "anger" button of their users just enough<mask> that they'll hit the "Share" button almost immediately. The way they cover issues which are really deep-rooted and require carefully-constructed views in order to take down - sexism, racism, homophobia/transphobia, etc. - is, I would argue, exceptionally harmful for the climate we're in,<mask> their coverage<mask> lacks nuance or care that all it does is preach to the choir and turn away anyone who doesn't already agree, furthering rifts which don't really need to be there in the first place. [NEWLINE] [NEWLINE] Buzzfeed, by comparison, is harmless. I don't think anyone will view ["28 Struggles Only People With Big Butts Will Understand"]( [URL] #.ixB8XaDn6)<mask> serious journalism in the same way that someone might view a click-baity io9 article<mask> "serious journalism,"<mask> Buzzfeed does a very good job of separating that content. The great thing about having click-bait and more hard-hitting reporting being<mask> separate is that it doesn't harm journalism and journalistic integrity in the same way that merging the two (<mask> other sites have done) harm it. It's<mask> allows the author of the Big Butts piece (Rega Jha) to turn around and [write a 4,000-word piece on sexual assault in India]( [URL] #.yjY3eDnKv) -<mask> they were kept separate, there's little to no big-butt cross-pollination in the long-form article. The low-
Label encoding: <s> [STARTQ] I looked through the bios of all the head people and it made it clear BuzzFeed is all about getting things viral. [ENDQ] [NEWLINE] Well, obviously they're not going to come out and say "We're making listicles to make a profit." They need to be able to spin it positively - making viral content is a lot more understandable than just saying "We need money to fund other stuff." [NEWLINE] [NEWLINE] [STARTQ] Now everyone sees all this stupid viral list/quiz stuff and it has the same weight as real news. [ENDQ] [NEWLINE] I think this is where you're wrong to include Buzzfeed. There are some sites which do "stupid viral" stuff in order to garner clicks that's presented in the form of "real news." Off the top of my head, that's sites like Gawker, Thinkprogress, Elite Daily, and occasionally HuffPost and Upworthy. Their viral content is very politically charged, often designed to press the "anger" button of their users just enough so that they'll hit the "Share" button almost immediately. The way they cover issues which are really deep-rooted and require carefully-constructed views in order to take down - sexism, racism, homophobia/transphobia, etc. - is, I would argue, exceptionally harmful for the climate we're in, because their coverage so lacks nuance or care that all it does is preach to the choir and turn away anyone who doesn't already agree, furthering rifts which don't really need to be there in the first place. [NEWLINE] [NEWLINE] Buzzfeed, by comparison, is harmless. I don't think anyone will view ["28 Struggles Only People With Big Butts Will Understand"]( [URL] #.ixB8XaDn6) as serious journalism in the same way that someone might view a click-baity io9 article as "serious journalism," because Buzzfeed does a very good job of separating that content. The great thing about having click-bait and more hard-hitting reporting being so separate is that it doesn't harm journalism and journalistic integrity in the same way that merging the two ( as other sites have done) harm it. It's what allows the author of the Big Butts piece (Rega Jha) to turn around and [write a 4,000-word piece on sexual assault in India]( [URL] #.yjY3eDnKv) - since they were kept separate, there's little to no big-butt cross-pollination in the long-form article. The low-
Loss: tensor(0.1850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, paying workers in the Third World more would not result in much of an increase in prices.  In your scenario, you are assuming a 100% increase in the cost of a shirt. <mask>,<mask><mask> [this study]( [URL].cgi?article=1012&amp;context=peri_workingpapers) done by University of Massachusetts which looked at garment production in Mexico, a 100% increase in workers wages only resulted in a 2-6% increase in the retail price of clothing. <mask> your $2 shirt would cost around $2.12 instead.  Clearly, the negative effects of supporting better working conditions are negligible to you. [NEWLINE] [NEWLINE] Second,<mask> about positive effects?  There are some philosophical arguments to be made here.  From a moral standpoint, we tend to feel guilty and displeased<mask> we see or hear about poverty, suffering, and exploitation.  Oscar Wilde discusses this in his 1891 essay [*The Soul of Man Under Socialism*]( [URL] /).  These negative emotions that tend to afflict us<mask> of other people's poverty and suffering decrease our happiness;<mask>, we have an incentive to solve these issues and gain utility from 1) not feeling bad about the existence of poverty, and 2) feeling proud that we helped solve these issues. <mask>, by helping people in the Third World not have to work in sweatshops, we can decrease our guilt and increase our pride and self-worth. [NEWLINE] [NEWLINE] There are<mask> positive effects in direct material gains, in that improving conditions in the Third World make it less likely that our own economy in the West will suffer from unemployment.  The closer wages in the Third World are on par with wages in the First, the less likely it is that we see manufacturing and other solid middle-class jobs get offshored and outsourced.  These jobs are more likely to stay here, and<mask> improve our economy.  Even<mask> you don't want or need one of these jobs, an improved economy will still benefit you in the way of reduced crime, more and better quality goods and services, better schools, better public services, and<mask> forth. [NEWLINE] [NEWLINE] There are<mask> positive effects in the political scene.  Improved worker wages in the Third World reduces the profits of the large multinationals that exploit these workers.  These are often the same multinationals that control legislation and lobbyists here in the First World, and lobby to try to erode worker's rights here, jump around environmental regulations, and [siphon taxpayer money
Label encoding: <s>First off, paying workers in the Third World more would not result in much of an increase in prices.  In your scenario, you are assuming a 100% increase in the cost of a shirt.  However, according to [this study]( [URL].cgi?article=1012&amp;context=peri_workingpapers) done by University of Massachusetts which looked at garment production in Mexico, a 100% increase in workers wages only resulted in a 2-6% increase in the retail price of clothing.  So your $2 shirt would cost around $2.12 instead.  Clearly, the negative effects of supporting better working conditions are negligible to you. [NEWLINE] [NEWLINE] Second, what about positive effects?  There are some philosophical arguments to be made here.  From a moral standpoint, we tend to feel guilty and displeased when we see or hear about poverty, suffering, and exploitation.  Oscar Wilde discusses this in his 1891 essay [*The Soul of Man Under Socialism*]( [URL] /).  These negative emotions that tend to afflict us because of other people's poverty and suffering decrease our happiness; therefore, we have an incentive to solve these issues and gain utility from 1) not feeling bad about the existence of poverty, and 2) feeling proud that we helped solve these issues.  Thus, by helping people in the Third World not have to work in sweatshops, we can decrease our guilt and increase our pride and self-worth. [NEWLINE] [NEWLINE] There are also positive effects in direct material gains, in that improving conditions in the Third World make it less likely that our own economy in the West will suffer from unemployment.  The closer wages in the Third World are on par with wages in the First, the less likely it is that we see manufacturing and other solid middle-class jobs get offshored and outsourced.  These jobs are more likely to stay here, and thus improve our economy.  Even if you don't want or need one of these jobs, an improved economy will still benefit you in the way of reduced crime, more and better quality goods and services, better schools, better public services, and so forth. [NEWLINE] [NEWLINE] There are also positive effects in the political scene.  Improved worker wages in the Third World reduces the profits of the large multinationals that exploit these workers.  These are often the same multinationals that control legislation and lobbyists here in the First World, and lobby to try to erode worker's rights here, jump around environmental regulations, and [siphon taxpayer money
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The brain is a pattern maker and pattern matcher. And the patterns it makes, it finds in the outside world with ease. And the more deeply held and treasured a pattern (belief), the more it must find them out there in the world,<mask> that they can be matched are validated to be true. And the more invested it becomes, the more it sees<mask> it believes, even<mask> "out there" is the very poorest of matches. [NEWLINE] [NEWLINE] To challenge your own beliefs is to threaten your very identity! For it's our patterns/beliefs/values that make us who we are, and to change them for something else is an act of self-destruction without the guarantee of catharsis at the other end!<mask> kudos to you! [NEWLINE] [NEWLINE] My Dad was a mathematician by training and a hobby astrologer/astronomer, and he made thousands of charts. He sort a statistical proof of astrology for about a decade, collecting thousands of birth dates/times/locations for people in different professions. I remember there was a famous sportsmen set, and I remember he had a huge amount of data on volcanic eruptions and earthquakes. He eventually discounted the popular field<mask> completely rubbish, and was left with two statistical correlations (mars in some house for sportsmen and for earthquakes!) That was 20 years ago, and<mask> remains from my perspective, and conversations with him, is his spiritual sense that "existential causation" is "top down", not "bottom-up"... [NEWLINE] [NEWLINE] Ahh, Dad. [NEWLINE] [NEWLINE] Anyhow, you discount confirmation bias,<mask><mask> sure are you with the validity of your self-diagnosis? It's<mask> a small part of known [cognitive biases]( [URL] ) -<mask> many ways for the mind to trick itself to retain it's values/beliefs! [NEWLINE] [NEWLINE] You love science and thinking critically,<mask> no doubt you have a hunch that there is a contradiction you need to resolve. [NEWLINE] [NEWLINE] Science doesn't do<mask> well proving the non-existence of something,<mask> that something doesn't exist or is untestable  -<mask> there is no proof to find! There's nothing for science to point to and say "Oh, there it isn't!"<mask><mask> you make a very specific claim,<mask> your words have set meanings and definitions, then Science or just well applied logic can find proof of absence (or proof of impossibility) much easier. For example, for Science, you have to make a specific claim like
Label encoding: <s>The brain is a pattern maker and pattern matcher. And the patterns it makes, it finds in the outside world with ease. And the more deeply held and treasured a pattern (belief), the more it must find them out there in the world, so that they can be matched are validated to be true. And the more invested it becomes, the more it sees what it believes, even if "out there" is the very poorest of matches. [NEWLINE] [NEWLINE] To challenge your own beliefs is to threaten your very identity! For it's our patterns/beliefs/values that make us who we are, and to change them for something else is an act of self-destruction without the guarantee of catharsis at the other end! So kudos to you! [NEWLINE] [NEWLINE] My Dad was a mathematician by training and a hobby astrologer/astronomer, and he made thousands of charts. He sort a statistical proof of astrology for about a decade, collecting thousands of birth dates/times/locations for people in different professions. I remember there was a famous sportsmen set, and I remember he had a huge amount of data on volcanic eruptions and earthquakes. He eventually discounted the popular field as completely rubbish, and was left with two statistical correlations (mars in some house for sportsmen and for earthquakes!) That was 20 years ago, and what remains from my perspective, and conversations with him, is his spiritual sense that "existential causation" is "top down", not "bottom-up"... [NEWLINE] [NEWLINE] Ahh, Dad. [NEWLINE] [NEWLINE] Anyhow, you discount confirmation bias, but how sure are you with the validity of your self-diagnosis? It's but a small part of known [cognitive biases]( [URL] ) - so many ways for the mind to trick itself to retain it's values/beliefs! [NEWLINE] [NEWLINE] You love science and thinking critically, so no doubt you have a hunch that there is a contradiction you need to resolve. [NEWLINE] [NEWLINE] Science doesn't do so well proving the non-existence of something, if that something doesn't exist or is untestable  - because there is no proof to find! There's nothing for science to point to and say "Oh, there it isn't!" But if you make a very specific claim, where your words have set meanings and definitions, then Science or just well applied logic can find proof of absence (or proof of impossibility) much easier. For example, for Science, you have to make a specific claim like
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**I work for a Diamond wholesaler<mask> I'll bite this....I'm still very new BTW.** [NEWLINE] [NEWLINE] *Let's tear down this "ridiculously overpriced" or "debeers scam" argument *<mask> you want dirt cheap diamonds, go *buy them USED*<mask> there is one thing on planet earth that literally last forever, it's rocks and minerals. Gold can be easily easily be worked on and diamonds can be reset/cleaned<mask> you can modify the design. Used engagement rings go for VERY cheap online. BUY ONE<mask> you need one online then everyone is happy! [NEWLINE] [NEWLINE] Here's one! Just remember to haggle cuz they have no one else to sell too :) [NEWLINE] [NEWLINE] [URL] up.<mask> don't get apprehensive about used....many of your wives are wearing used gold whether you know it or not. [NEWLINE] [NEWLINE] **<mask> for high prices**....it's called comparison shopping which some people seem to magically forget<mask> it comes to jewels and luxury goods. Someplace will charge 100-300% premiums simply<mask> the consumer won't walk down the street to compare....or heaven forbid....buy a used diamond [roll eyes]. Restaurants would charge 100% premiums<mask> they could get away with it.<mask>, RESEARCH diamonds....the vendors have a science behind stone grading which few seem to research: Don't be that guy who comes in saying " I want a pretty one". Do you go into a car dealer saying  "Gimme a fast one?" [NEWLINE] [NEWLINE] [NEWLINE] **<mask> for the "useless" diamonds argument**. Have you ever visited a REAL jewelry vendor? Not the generic kay jewelers and mall shops....nice, GIA certified diamonds are mind blowingly beautiful in person and the right light. Those stores have regularly customers who love jewels the same way nerds love their video games. They're not useless<mask> it shows other people you have class and sophistication. There was a thread I started asking<mask> guys don't receive more complements. I always received tons<mask> a guy<mask> it's<mask> I tend to wear "nice, useless" things like expensive shoes, fountain pens, nice watches, hand made messenger bags, bespoke suits and coats, etc. Some balk at the money I spent<mask> enough people(both genders; strangers usually) appreciate it enough to tell me about it. [NEWLINE] [NEWLINE] For some reason other guys think nice jewelry are a waste of money<mask> blow $30-100k+ on their dream car which coincidentally many
Label encoding: <s>**I work for a Diamond wholesaler so I'll bite this....I'm still very new BTW.** [NEWLINE] [NEWLINE] *Let's tear down this "ridiculously overpriced" or "debeers scam" argument * If you want dirt cheap diamonds, go *buy them USED* If there is one thing on planet earth that literally last forever, it's rocks and minerals. Gold can be easily easily be worked on and diamonds can be reset/cleaned so you can modify the design. Used engagement rings go for VERY cheap online. BUY ONE if you need one online then everyone is happy! [NEWLINE] [NEWLINE] Here's one! Just remember to haggle cuz they have no one else to sell too :) [NEWLINE] [NEWLINE] [URL] up. So don't get apprehensive about used....many of your wives are wearing used gold whether you know it or not. [NEWLINE] [NEWLINE] ** As for high prices**....it's called comparison shopping which some people seem to magically forget when it comes to jewels and luxury goods. Someplace will charge 100-300% premiums simply because the consumer won't walk down the street to compare....or heaven forbid....buy a used diamond [roll eyes]. Restaurants would charge 100% premiums if they could get away with it. Lastly, RESEARCH diamonds....the vendors have a science behind stone grading which few seem to research: Don't be that guy who comes in saying " I want a pretty one". Do you go into a car dealer saying  "Gimme a fast one?" [NEWLINE] [NEWLINE] [NEWLINE] ** As for the "useless" diamonds argument**. Have you ever visited a REAL jewelry vendor? Not the generic kay jewelers and mall shops....nice, GIA certified diamonds are mind blowingly beautiful in person and the right light. Those stores have regularly customers who love jewels the same way nerds love their video games. They're not useless because it shows other people you have class and sophistication. There was a thread I started asking why guys don't receive more complements. I always received tons as a guy but it's because I tend to wear "nice, useless" things like expensive shoes, fountain pens, nice watches, hand made messenger bags, bespoke suits and coats, etc. Some balk at the money I spent but enough people(both genders; strangers usually) appreciate it enough to tell me about it. [NEWLINE] [NEWLINE] For some reason other guys think nice jewelry are a waste of money but blow $30-100k+ on their dream car which coincidentally many
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You never responded to the points I made above about<mask><mask><mask> philosophy is valuable to understanding science and the implications of the knowledge we gain from it. I guess you're saying that people will be able to figure that out on their own without having to dedicate resources to it? [NEWLINE] [NEWLINE] People can and will continue to think about science and engineering even<mask> they don't do it<mask> their day job. And that's probably one of the reasons we have a lot of armchair climate scientists who deny that global warming is real, along with amateur structural engineers who think the Twin Towers came down in a controlled demolition, and backseat biologists who deny evolution. Perhaps most frighteningly, there are people who refuse to vaccinate their children<mask> they believe it causes autism. In other words, you need training to understand things and to be able to make judgments properly. [NEWLINE] [NEWLINE] My point is that,<mask> those examples demonstrate, people who speculate on science in their free time, even<mask> it's not their day job (to use your words), are frequently tremendously wrong.<mask> would you think it's different for other fields? [NEWLINE] [NEWLINE] I trust you would be skeptical of crossing a bridge that was designed by an amateur, just<mask> you probably would not want to undergo surgery performed by an amateur.<mask>,<mask> it's your day in court, do you want to be punished based on an amateur understanding of law and criminology?<mask> your government makes laws, do you want the resulting policies to be informed by a reasoned, rational understanding of<mask> makes society tick and<mask> to fix social, economic and political problems? Or would you rather those laws be based on the gut feelings of some bureaucrats whose day job is not to think about sociology, political science, anthropology and psychology? And<mask> our leaders make decisions about whether we go to war or try to maintain peace, should their choices be informed by a real understanding of history, or should they just wing it? [NEWLINE] [NEWLINE] <mask> we insist,<mask> we should, that a bridge should be engineered correctly by trained engineers, shouldn't we agree that a law—might deprive someone of liberty and even life—should be engineered by trained people<mask> well?<mask> we insist that doctors should have a deep knowledge of the human body and medicine, shouldn't we<mask> insist that our school teachers have a deep knowledge of<mask> people learn and<mask> they behave? [NEWLINE] [NEWLINE] In short, reading a Wikipedia page on WWII makes you an historian just<mask> much<mask> reading the Wikipedia page on cells makes me a biologist. Watching FOX News or
Label encoding: <s>You never responded to the points I made above about why I think philosophy is valuable to understanding science and the implications of the knowledge we gain from it. I guess you're saying that people will be able to figure that out on their own without having to dedicate resources to it? [NEWLINE] [NEWLINE] People can and will continue to think about science and engineering even if they don't do it as their day job. And that's probably one of the reasons we have a lot of armchair climate scientists who deny that global warming is real, along with amateur structural engineers who think the Twin Towers came down in a controlled demolition, and backseat biologists who deny evolution. Perhaps most frighteningly, there are people who refuse to vaccinate their children because they believe it causes autism. In other words, you need training to understand things and to be able to make judgments properly. [NEWLINE] [NEWLINE] My point is that, as those examples demonstrate, people who speculate on science in their free time, even if it's not their day job (to use your words), are frequently tremendously wrong. Why would you think it's different for other fields? [NEWLINE] [NEWLINE] I trust you would be skeptical of crossing a bridge that was designed by an amateur, just as you probably would not want to undergo surgery performed by an amateur. So, when it's your day in court, do you want to be punished based on an amateur understanding of law and criminology? When your government makes laws, do you want the resulting policies to be informed by a reasoned, rational understanding of what makes society tick and how to fix social, economic and political problems? Or would you rather those laws be based on the gut feelings of some bureaucrats whose day job is not to think about sociology, political science, anthropology and psychology? And when our leaders make decisions about whether we go to war or try to maintain peace, should their choices be informed by a real understanding of history, or should they just wing it? [NEWLINE] [NEWLINE] If we insist, as we should, that a bridge should be engineered correctly by trained engineers, shouldn't we agree that a law—might deprive someone of liberty and even life—should be engineered by trained people as well? If we insist that doctors should have a deep knowledge of the human body and medicine, shouldn't we also insist that our school teachers have a deep knowledge of how people learn and how they behave? [NEWLINE] [NEWLINE] In short, reading a Wikipedia page on WWII makes you an historian just as much as reading the Wikipedia page on cells makes me a biologist. Watching FOX News or
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As I reflected over an earlier conversation I had with someone who battled with depression, I came to think that mental health facilities weren't serving her.<mask> I say they should have similar facilities, I specifically mean: [NEWLINE] [NEWLINE] * **"Emergency Room" with staff available 24/7** to treat people with immediate concerns, such<mask> wanting to harm themselves/others or extreme psychotic episodes. [NEWLINE] [NEWLINE] * **Mobile mental health "ambulances"** who can respond quickly to emergencies (like those previously mentioned) and bring patients to the Mental Health Emergency Room<mask> necessary. [NEWLINE] [NEWLINE] * **General practitioners** who *most people* should see once or twice a year to "check up" and some may need to see a little more often<mask> under particular distress [NEWLINE] [NEWLINE] * **Specialists** who see patients regularly who have diagnosed, chronic mental health issues [NEWLINE] [NEWLINE] It seems to me that the mental health world is predominantly "specialists" with "general practitioners" here and there. The only emergency staff tend to be at dedicated facilities<mask> patients are committed. My friend (with depression/self-harm issues) explained that she didn't think her visits with counselors/therapists were particularly helpful<mask><mask> **she saw them was<mask> she felt "normal" and not in the state of distress**. [NEWLINE] [NEWLINE] <mask> I pointed her toward hotlines, it is extremely limited to<mask> much you can help over the phone and often they only referred her to set up an appointment with a local professional in the next few weeks.<mask> that is good, **it doesn't prevent her from self-harming in those "emergency" situations**.<mask><mask> having both "ER" and "Ambulance" type services for mental health would be extremely beneficial in these scenarios<mask> they can provide care<mask> it is most needed. [NEWLINE] [NEWLINE] I believe that having these facilities available would give people suffering from mental health issues a resource<mask> they feel like they don't have anyone who can help them. Ultimately, making these services available<mask><mask> would make it less taboo to seek out mental health care and make it similar to "regular" health care. [NEWLINE] [NEWLINE] <mask>, I don't really have any first-hand experience with mental health care,<mask> I'm open to change my mind. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL]
Label encoding: <s>As I reflected over an earlier conversation I had with someone who battled with depression, I came to think that mental health facilities weren't serving her. When I say they should have similar facilities, I specifically mean: [NEWLINE] [NEWLINE] * **"Emergency Room" with staff available 24/7** to treat people with immediate concerns, such as wanting to harm themselves/others or extreme psychotic episodes. [NEWLINE] [NEWLINE] * **Mobile mental health "ambulances"** who can respond quickly to emergencies (like those previously mentioned) and bring patients to the Mental Health Emergency Room if necessary. [NEWLINE] [NEWLINE] * **General practitioners** who *most people* should see once or twice a year to "check up" and some may need to see a little more often when under particular distress [NEWLINE] [NEWLINE] * **Specialists** who see patients regularly who have diagnosed, chronic mental health issues [NEWLINE] [NEWLINE] It seems to me that the mental health world is predominantly "specialists" with "general practitioners" here and there. The only emergency staff tend to be at dedicated facilities where patients are committed. My friend (with depression/self-harm issues) explained that she didn't think her visits with counselors/therapists were particularly helpful because when **she saw them was when she felt "normal" and not in the state of distress**. [NEWLINE] [NEWLINE] While I pointed her toward hotlines, it is extremely limited to how much you can help over the phone and often they only referred her to set up an appointment with a local professional in the next few weeks. While that is good, **it doesn't prevent her from self-harming in those "emergency" situations**. I think having both "ER" and "Ambulance" type services for mental health would be extremely beneficial in these scenarios because they can provide care when it is most needed. [NEWLINE] [NEWLINE] I believe that having these facilities available would give people suffering from mental health issues a resource when they feel like they don't have anyone who can help them. Ultimately, making these services available I think would make it less taboo to seek out mental health care and make it similar to "regular" health care. [NEWLINE] [NEWLINE] However, I don't really have any first-hand experience with mental health care, so I'm open to change my mind. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL]
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0998, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0979, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0977, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People *do* make that kind of deal about a lot of other food. I mean, acidity isn't a characteristic of steak,<mask> there are tons of enthusiasts who take steak [very seriously](/r/steak) (don't even get them started on Kobe beef). I am sure there are a few who are enthusiasts of oranges (likely farmers),<mask> the difference between grapes and oranges is that grapes can transform into wine, and *every aspect* of wine production changes the overall flavor *dramatically*. Oranges... don't ferment well or really do anything interesting. There isn't a lot to be enthused about. [NEWLINE] [NEWLINE] Vinegar smell or flavour is the iconic mark of cheap wine. Wine quite literally [turns into vinegar]( [URL] #Wine)<mask> it spoils (grapes are seriously magic).<mask> wine tastes sour, vinegary, or has a strong alcohol taste, it's bad. The problem is that most wine brought to parties and the like, especially by younger/apathetic people, is "whatever cheap bottle looks nice enough". Cost doesn't [*always*]( [URL] ) correlate with quality,<mask> it [*usually*]( [URL] ) does. [NEWLINE] [NEWLINE] <mask> you're actually interested in<mask> wine can really offer, I recommend going either to a winery for a tasting (usually costs only a couple bucks, or is free<mask> you buy a bottle)(for the love of god avoid big wineries and opt for smaller, family-run ones), or to an actual "wine bar" which will typically have tasting options with [charcuterie](/r/charcuterie) pairings. Or, just make friends with a wine snob and get your early education in for free<mask> they [enthusiastically try to sway you to the dark side]( [URL] %27s_Big_Wine_Adventure). [NEWLINE] [NEWLINE] <mask><mask> another thing worth note is that alcohol in general is going to always be a little unpleasant tasting. North Americans have a diet culture which involves [a lot of sugar]( [URL] +consumption+in+the+us),<mask> whenever anything isn't extremely sweet, we tend to find it offputting. Getting past that and discovering that not all food is or needs to be sugar-laden to be delicious is a hurdle, and a lot of people don't bother (which is<mask> stuff like "light beers", "mocha-frappa-whatevers" and "mcburger now with
Label encoding: <s>People *do* make that kind of deal about a lot of other food. I mean, acidity isn't a characteristic of steak, but there are tons of enthusiasts who take steak [very seriously](/r/steak) (don't even get them started on Kobe beef). I am sure there are a few who are enthusiasts of oranges (likely farmers), but the difference between grapes and oranges is that grapes can transform into wine, and *every aspect* of wine production changes the overall flavor *dramatically*. Oranges... don't ferment well or really do anything interesting. There isn't a lot to be enthused about. [NEWLINE] [NEWLINE] Vinegar smell or flavour is the iconic mark of cheap wine. Wine quite literally [turns into vinegar]( [URL] #Wine) when it spoils (grapes are seriously magic). If wine tastes sour, vinegary, or has a strong alcohol taste, it's bad. The problem is that most wine brought to parties and the like, especially by younger/apathetic people, is "whatever cheap bottle looks nice enough". Cost doesn't [*always*]( [URL] ) correlate with quality, but it [*usually*]( [URL] ) does. [NEWLINE] [NEWLINE] If you're actually interested in what wine can really offer, I recommend going either to a winery for a tasting (usually costs only a couple bucks, or is free if you buy a bottle)(for the love of god avoid big wineries and opt for smaller, family-run ones), or to an actual "wine bar" which will typically have tasting options with [charcuterie](/r/charcuterie) pairings. Or, just make friends with a wine snob and get your early education in for free while they [enthusiastically try to sway you to the dark side]( [URL] %27s_Big_Wine_Adventure). [NEWLINE] [NEWLINE] I think another thing worth note is that alcohol in general is going to always be a little unpleasant tasting. North Americans have a diet culture which involves [a lot of sugar]( [URL] +consumption+in+the+us), so whenever anything isn't extremely sweet, we tend to find it offputting. Getting past that and discovering that not all food is or needs to be sugar-laden to be delicious is a hurdle, and a lot of people don't bother (which is why stuff like "light beers", "mocha-frappa-whatevers" and "mcburger now with
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I personally do not understand the value in spending in the tens to hundreds of thousands of dollars on IVF, surrogacy, etc. to facilitate having a child with your genes<mask> there are<mask> many children throughout the entire world deprived of a family and in need of adoption.<mask> is it about YOUR genes that is<mask> special? It's not just a couple of rounds of IVF in a 30-something patient who is having trouble conceiving that bothers me the most,<mask> I guess the more complicated/expensive it becomes the more frustrated I get with the concept. For example, having brothers or sisters donating gametes<mask> yours do not work due to age or medical issue/<mask> you're a homosexual couple in need of gametes from the opposite gender.<mask> your motivation stems purely from the sense of purpose that parenthood would bring, then<mask> does it matter<mask> the child's exact genome is? I understand that this might be touchy, especially in the homosexual community for which assisted reproductive technology is the only way of bearing children of your own genetic lineage,<mask> I guess I don't understand<mask> this should be a right, and something that needs to be facilitated<mask> it is not "naturally" possible. I realize my feelings about this are purely feelings, and I'd never want anything written into law banning these procedures - I realize this would ostracize entire communities.<mask> it really bothers me that people would turn to this instead of to adoption. Would you really love a child with your genes more than an adopted child? Privilege check: I'm posting this<mask> someone who would most likely be able to reproduce without such assistance,<mask><mask> it turned out that I couldn't and I was at a place in life<mask> I wanted to be a parent, I'm pretty sure I'd choose adoption over anything more "complicated." Please change my view! I'd particularly like to hear from parents who have chosen to conceive via IVF, surrogacy, etc. Please know I respect your decision and am happy for all children raised by loving parents, and<mask><mask> this should remain a legal possibility,<mask> to be honest I have a personal bias and would like that to be changed. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is
Label encoding: <s>I personally do not understand the value in spending in the tens to hundreds of thousands of dollars on IVF, surrogacy, etc. to facilitate having a child with your genes when there are so many children throughout the entire world deprived of a family and in need of adoption. What is it about YOUR genes that is so special? It's not just a couple of rounds of IVF in a 30-something patient who is having trouble conceiving that bothers me the most, but I guess the more complicated/expensive it becomes the more frustrated I get with the concept. For example, having brothers or sisters donating gametes if yours do not work due to age or medical issue/ if you're a homosexual couple in need of gametes from the opposite gender. If your motivation stems purely from the sense of purpose that parenthood would bring, then why does it matter what the child's exact genome is? I understand that this might be touchy, especially in the homosexual community for which assisted reproductive technology is the only way of bearing children of your own genetic lineage, but I guess I don't understand why this should be a right, and something that needs to be facilitated if it is not "naturally" possible. I realize my feelings about this are purely feelings, and I'd never want anything written into law banning these procedures - I realize this would ostracize entire communities. But it really bothers me that people would turn to this instead of to adoption. Would you really love a child with your genes more than an adopted child? Privilege check: I'm posting this as someone who would most likely be able to reproduce without such assistance, but if it turned out that I couldn't and I was at a place in life where I wanted to be a parent, I'm pretty sure I'd choose adoption over anything more "complicated." Please change my view! I'd particularly like to hear from parents who have chosen to conceive via IVF, surrogacy, etc. Please know I respect your decision and am happy for all children raised by loving parents, and I think this should remain a legal possibility, but to be honest I have a personal bias and would like that to be changed. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is
Loss: tensor(0.1584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0990, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok,<mask> you're off to fight your enemy. [NEWLINE] [NEWLINE] <mask><mask> *is* the person you're fighting?  They're a person, who has a set of beliefs or ideals that, for whatever reason, conflicts with yours.<mask> did you get your beliefs?  From the place you grew up and the people that you grew up around, and your education, tempered by the things that happen to occur in the course of life. <mask> 'd they get theirs?  Same thing.  Different place, different people, different events, different beliefs, same process. <mask> I might not agree with their views, I can try to understand them.  Their vantage point is different from mine, maybe they see something I can't. <mask> I can't empathize, I can never learn from others, and I'll just keep making the same errors over and over. [NEWLINE] [NEWLINE] They're the contingent result of their environment, a product of the same forces that made you. There are people at home that need them, and that they need.  They like some foods and don't like others.  There's "something that it's like to *be them*", they're a subjective viewpoint in the world, that's exclusive to them, that no one else can know, and can never be re-created ever, in the entire course of the universe.  This utter uniqueness seems to make everyone, even the people I find the most disagreeable, to be inherently valuable. [NEWLINE] [NEWLINE] It's easy to see the person you're in conflict with<mask> this "other" that appears fully formed to be beaten,<mask> that's not really the most accurate conception.  They're just someone.  You're just someone. <mask> the course of events in life had run differently, you both could be coworkers, or comrades, or friends, or partners.  They're just<mask> important<mask> you, just to different people. [NEWLINE] [NEWLINE] <mask> you see your enemy across the battlefield, they're seeing their enemy too.  Of course, both are *actually* not seeing<mask>'s really across from them, which is a person.  It's the fact that you're opposing them (and them you) that makes you enemies, and that's just perspectival. [NEWLINE] [NEWLINE] Looking at it this way makes it kinda hard *not* to empathize.  I'm not saying I'm any good at this stuff,<mask> I try to keep it in mind, and<mask> I do manage to act or think
Label encoding: <s>Ok, so you're off to fight your enemy. [NEWLINE] [NEWLINE] But what *is* the person you're fighting?  They're a person, who has a set of beliefs or ideals that, for whatever reason, conflicts with yours. How did you get your beliefs?  From the place you grew up and the people that you grew up around, and your education, tempered by the things that happen to occur in the course of life.  How 'd they get theirs?  Same thing.  Different place, different people, different events, different beliefs, same process.  While I might not agree with their views, I can try to understand them.  Their vantage point is different from mine, maybe they see something I can't.  If I can't empathize, I can never learn from others, and I'll just keep making the same errors over and over. [NEWLINE] [NEWLINE] They're the contingent result of their environment, a product of the same forces that made you. There are people at home that need them, and that they need.  They like some foods and don't like others.  There's "something that it's like to *be them*", they're a subjective viewpoint in the world, that's exclusive to them, that no one else can know, and can never be re-created ever, in the entire course of the universe.  This utter uniqueness seems to make everyone, even the people I find the most disagreeable, to be inherently valuable. [NEWLINE] [NEWLINE] It's easy to see the person you're in conflict with as this "other" that appears fully formed to be beaten, but that's not really the most accurate conception.  They're just someone.  You're just someone.  If the course of events in life had run differently, you both could be coworkers, or comrades, or friends, or partners.  They're just as important as you, just to different people. [NEWLINE] [NEWLINE] When you see your enemy across the battlefield, they're seeing their enemy too.  Of course, both are *actually* not seeing what's really across from them, which is a person.  It's the fact that you're opposing them (and them you) that makes you enemies, and that's just perspectival. [NEWLINE] [NEWLINE] Looking at it this way makes it kinda hard *not* to empathize.  I'm not saying I'm any good at this stuff, but I try to keep it in mind, and when I do manage to act or think
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't disagree with you about<mask> merit means for each type of community.<mask> resources are vital for any society. And it is a control upon them I'm proposing. We already have many controls, like taxes, I'm just saying that another one should be done to have a meritocratic society. And yes, the merit I mean is resources (money).<mask><mask> you pay attention I said several times in different comments that changing that heritage subject, people would start to prioritize culture and education on their offspring. Being the only way they could help them,<mask> a natural thing done by any mammal (protect their offspring). [NEWLINE] [NEWLINE] Many things can define happiness,<mask> material propriety is a tangible thing a community should control. And it is the thing that gives power and influence,<mask> that's the reason we might be careful with that.<mask> I'm worried to whom that power (resources) should be given. It should be for someone who deserves, not someone who inherits. [NEWLINE] [NEWLINE] I won't discuss about<mask> we could avoid someone sneaking money to somebody else, it would be an IRS problem. I'm not on the technicalities of that for now. [NEWLINE] [NEWLINE] [STARTQ] Even<mask> we could do it somehow, people who were successful will always be giving unfair advantages to their children. They will have grown up in an affluent home and have more access to the best nutrition and education. [ENDQ] [NEWLINE] No and yes. They will grown up in an affluent home, probably best educated parents that will teach them better on somethings.<mask> the main point is the SAME education. Cutting out heritage is to have ALL children starting with the same conditions. And a good education to all. (remember, government would be collecting all heritages to finance the future of all children). [NEWLINE] [NEWLINE] [STARTQ] More to the point, the harm of wealth disparity isn't the rich kid getting a Ferrarri, it's the poor kid not getting to go to school, or eat well, or have a place to live<mask> she's looking for a job. [ENDQ] [NEWLINE] That's a way to try to solve the problem in the way it is now.<mask> I'm proposing is something<mask> there isn't the rich kid.<mask> the rich dad! And both, the kid with poor dad and rich dad, will go to the same school, have the same medicare, etc.<mask> both want to be like the rich dad, well, the better will win.<mask> was the rich dad son, well, maybe it is the genes... or
Label encoding: <s>I don't disagree with you about what merit means for each type of community. But resources are vital for any society. And it is a control upon them I'm proposing. We already have many controls, like taxes, I'm just saying that another one should be done to have a meritocratic society. And yes, the merit I mean is resources (money). But if you pay attention I said several times in different comments that changing that heritage subject, people would start to prioritize culture and education on their offspring. Being the only way they could help them, as a natural thing done by any mammal (protect their offspring). [NEWLINE] [NEWLINE] Many things can define happiness, but material propriety is a tangible thing a community should control. And it is the thing that gives power and influence, so that's the reason we might be careful with that. So I'm worried to whom that power (resources) should be given. It should be for someone who deserves, not someone who inherits. [NEWLINE] [NEWLINE] I won't discuss about how we could avoid someone sneaking money to somebody else, it would be an IRS problem. I'm not on the technicalities of that for now. [NEWLINE] [NEWLINE] [STARTQ] Even if we could do it somehow, people who were successful will always be giving unfair advantages to their children. They will have grown up in an affluent home and have more access to the best nutrition and education. [ENDQ] [NEWLINE] No and yes. They will grown up in an affluent home, probably best educated parents that will teach them better on somethings. But the main point is the SAME education. Cutting out heritage is to have ALL children starting with the same conditions. And a good education to all. (remember, government would be collecting all heritages to finance the future of all children). [NEWLINE] [NEWLINE] [STARTQ] More to the point, the harm of wealth disparity isn't the rich kid getting a Ferrarri, it's the poor kid not getting to go to school, or eat well, or have a place to live while she's looking for a job. [ENDQ] [NEWLINE] That's a way to try to solve the problem in the way it is now. What I'm proposing is something where there isn't the rich kid. But the rich dad! And both, the kid with poor dad and rich dad, will go to the same school, have the same medicare, etc. If both want to be like the rich dad, well, the better will win. If was the rich dad son, well, maybe it is the genes... or
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>With the advent of self-driving vehicles, the unavoidable will become clear: people are terrible drivers, and operating your own car is unacceptably reckless<mask> a better alternative exists.  I see the coming timeline like this: (copied from a reply to another post) [NEWLINE] [NEWLINE] [STARTQ] 2-5 years: The last major technological hurdles (driving in rural/poorly documented areas, driving in adverse conditions, cost) are resolved. Cars are now demonstratively better drivers than humans in all situations. (note: may be a very liberal estimate.) [ENDQ] 4-6 years: The first round of legal cases involving driverless cars is settled, producing a precedent that makes driving your own car very risky. A collision between two vehicles, one self driving the other not, almost always results in fault to the driver. Causing an accident<mask> operating a car with unused self-driving capability makes drivers extremely vulnerable to being sued. [NEWLINE] 5-10 years: Safety studies, overwhelmingly favorable to self-driving cars, lead to the option becoming mandatory on all new vehicles. insurance companies, burned by litigation, offer premium rates to those who never switch off the driverless option,<mask> increasing rates on drivers who elect to operate their cars manually. Soon the difference between these rates becomes enormous. [NEWLINE] 10-15 years: Commercial driving is entirely automated. Cabs, buses, trucks, trains, "driver" becomes an obsolete profession. The savings in both wages and liability is simply too tremendous to allow any non-automated fleet to remain competitive. [NEWLINE] 15-20 years: Studies conclusively show that the only traffic casualties that still occur are exclusively due to human operator error. It becomes evident that driving your own car is unthinkably dangerous, like drunk driving at night with no headlights or seatbelts. Safety laws are passed that effectively outlaw operating your own vehicle. [NEWLINE] [NEWLINE] By the time my nephew is 15-16, controlling a car will be something that only hobbyists do, and never on public roads.  Very few cars will be privately owned, rather they will be operated by private or municipal transportation services. [NEWLINE] The age of the personal automobile is ending. CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than
Label encoding: <s>With the advent of self-driving vehicles, the unavoidable will become clear: people are terrible drivers, and operating your own car is unacceptably reckless if a better alternative exists.  I see the coming timeline like this: (copied from a reply to another post) [NEWLINE] [NEWLINE] [STARTQ] 2-5 years: The last major technological hurdles (driving in rural/poorly documented areas, driving in adverse conditions, cost) are resolved. Cars are now demonstratively better drivers than humans in all situations. (note: may be a very liberal estimate.) [ENDQ] 4-6 years: The first round of legal cases involving driverless cars is settled, producing a precedent that makes driving your own car very risky. A collision between two vehicles, one self driving the other not, almost always results in fault to the driver. Causing an accident while operating a car with unused self-driving capability makes drivers extremely vulnerable to being sued. [NEWLINE] 5-10 years: Safety studies, overwhelmingly favorable to self-driving cars, lead to the option becoming mandatory on all new vehicles. insurance companies, burned by litigation, offer premium rates to those who never switch off the driverless option, while increasing rates on drivers who elect to operate their cars manually. Soon the difference between these rates becomes enormous. [NEWLINE] 10-15 years: Commercial driving is entirely automated. Cabs, buses, trucks, trains, "driver" becomes an obsolete profession. The savings in both wages and liability is simply too tremendous to allow any non-automated fleet to remain competitive. [NEWLINE] 15-20 years: Studies conclusively show that the only traffic casualties that still occur are exclusively due to human operator error. It becomes evident that driving your own car is unthinkably dangerous, like drunk driving at night with no headlights or seatbelts. Safety laws are passed that effectively outlaw operating your own vehicle. [NEWLINE] [NEWLINE] By the time my nephew is 15-16, controlling a car will be something that only hobbyists do, and never on public roads.  Very few cars will be privately owned, rather they will be operated by private or municipal transportation services. [NEWLINE] The age of the personal automobile is ending. CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than
Loss: tensor(0.1146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've outright said "no" before to women I was interested in. Sometimes, it was in bed with them, completely naked. Somehow, I ended up having sex anyway. [NEWLINE] [NEWLINE] With a few girls I've seen over the course of my life, I've taken them back to either my or their apartment, got into bed, started feeling each other up, and then stopped them<mask> I didn't want to have sex. I've then gotten pressured by them to continue. [NEWLINE] [NEWLINE] I am clearly aroused--I just don't want to have sex for personal reasons. I like to get to know someone pretty well before I sleep with them, just to make sure there's not gonna be problems<mask> we don't work out at some point (that's happened in the past). [NEWLINE] [NEWLINE] <mask> some girls don't like hearing that. One girl started coercing me, saying, "Come on, I do yoga. Don't you want to see<mask> flexible I am?" and she started rubbing on me. I say I really shouldn't do this,<mask> I *am* getting turned on.<mask> eventually I just say fuck it (without explicit consent) and go for it. Another time, a girl just put the condom on me, and I was like, well, let's just get this over with. [NEWLINE] [NEWLINE] <mask>, here's<mask> I'm gonna get controversial with this. I know there is the "Don't blame the victim" mentality,<mask> few things are black and white to me. No undeniably means no,<mask> there are things I can do to not send mixed signals to a partner, which, objectively speaking I did. I've taken steps to stop sending those signals. [NEWLINE] [NEWLINE] <mask> I mean,<mask> I said no, that's by definition rape is it not? [NEWLINE] [NEWLINE] Here's another situation that's a bit sketchy:<mask> about<mask> I'm with my current girlfriend, whom I've had sex with many times. I've told her outright no before<mask> I'm not turned on,<mask> I care for her, and I care for her needs.<mask>, without explicit consent after saying no, I have sex with her just<mask> I care about satisfying her. Is that rape, too,<mask> I didn't want it? [NEWLINE] [NEWLINE] At the very least, I don't consider the situations I've described worth reporting to anyone. I mean, it's my body we're talking about here, aren't we--not the law's, who is sometimes less than trust worthy? I'm not
Label encoding: <s>I've outright said "no" before to women I was interested in. Sometimes, it was in bed with them, completely naked. Somehow, I ended up having sex anyway. [NEWLINE] [NEWLINE] With a few girls I've seen over the course of my life, I've taken them back to either my or their apartment, got into bed, started feeling each other up, and then stopped them because I didn't want to have sex. I've then gotten pressured by them to continue. [NEWLINE] [NEWLINE] I am clearly aroused--I just don't want to have sex for personal reasons. I like to get to know someone pretty well before I sleep with them, just to make sure there's not gonna be problems if we don't work out at some point (that's happened in the past). [NEWLINE] [NEWLINE] But some girls don't like hearing that. One girl started coercing me, saying, "Come on, I do yoga. Don't you want to see how flexible I am?" and she started rubbing on me. I say I really shouldn't do this, but I *am* getting turned on. So eventually I just say fuck it (without explicit consent) and go for it. Another time, a girl just put the condom on me, and I was like, well, let's just get this over with. [NEWLINE] [NEWLINE] So, here's where I'm gonna get controversial with this. I know there is the "Don't blame the victim" mentality, but few things are black and white to me. No undeniably means no, but there are things I can do to not send mixed signals to a partner, which, objectively speaking I did. I've taken steps to stop sending those signals. [NEWLINE] [NEWLINE] But I mean, if I said no, that's by definition rape is it not? [NEWLINE] [NEWLINE] Here's another situation that's a bit sketchy: how about when I'm with my current girlfriend, whom I've had sex with many times. I've told her outright no before when I'm not turned on, but I care for her, and I care for her needs. So, without explicit consent after saying no, I have sex with her just because I care about satisfying her. Is that rape, too, when I didn't want it? [NEWLINE] [NEWLINE] At the very least, I don't consider the situations I've described worth reporting to anyone. I mean, it's my body we're talking about here, aren't we--not the law's, who is sometimes less than trust worthy? I'm not
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0998, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0929, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0987, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am extremely angered<mask> I hear things like Michael Brown was executed. Cops are not secretly serial killers who are just waiting for a chance to get away with killing someone.<mask> that's<mask> these morons are portraying the police. I refuse to believe that a member of the police performed a public execution under scrutiny of other bystanders. The cop isn't mentally ill. His actions must done alongside of some kind of self interest. Shoot an innocent person who poses no danger in the head execution style in public? Really?<mask> possibly way of reasoning could lead to that conclusion? I was 100% confident<mask> the very break of the news that the cop was unlikely to have done anything wrong. These protesters automatically assumed the cop was in the wrong and refused to acknowledge new evidence. [NEWLINE] [NEWLINE] I get pissed off that people think Michael Brown is still a saint after the video of him robbing the store was released. Retards insist that it has nothing to do with the shooting. It has everything to do with it. The reality is that Michael probably thought he got caught, and<mask>, his interactions with the cop was likely to be extremely aggressive. I imagine it went something like- [NEWLINE] [NEWLINE] Officer - Hey you're blocking traffic [NEWLINE] [NEWLINE] Michael - I ain't rob no store you fucking pig. [NEWLINE] [NEWLINE] Officer - I didnt say... [NEWLINE] [NEWLINE] Michael - These cigars ain't from the store. [NEWLINE] [NEWLINE] Officer - I didn't ask about.... [NEWLINE] [NEWLINE] Michael - Fuck all ya. Always tryin a hold me down [NEWLINE] [NEWLINE] Officer - Ok put your hands in the air! [NEWLINE] [NEWLINE] Michael - Fly high or die tryin!! YOLO!! AHHH [charges head first at cop and gets shot like a retard] [NEWLINE] [NEWLINE] Seriously,<mask> the cop was a sick fuck and gets a boner shooting up black people, he could have done<mask> in a secluded area at night<mask> no one was around. It would have been his word against the world, no witnesses. Somebody please explain to me<mask> people protesting in favor of this dead thief isn't a moron. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views
Label encoding: <s>I am extremely angered when I hear things like Michael Brown was executed. Cops are not secretly serial killers who are just waiting for a chance to get away with killing someone. But that's how these morons are portraying the police. I refuse to believe that a member of the police performed a public execution under scrutiny of other bystanders. The cop isn't mentally ill. His actions must done alongside of some kind of self interest. Shoot an innocent person who poses no danger in the head execution style in public? Really? What possibly way of reasoning could lead to that conclusion? I was 100% confident since the very break of the news that the cop was unlikely to have done anything wrong. These protesters automatically assumed the cop was in the wrong and refused to acknowledge new evidence. [NEWLINE] [NEWLINE] I get pissed off that people think Michael Brown is still a saint after the video of him robbing the store was released. Retards insist that it has nothing to do with the shooting. It has everything to do with it. The reality is that Michael probably thought he got caught, and therefore, his interactions with the cop was likely to be extremely aggressive. I imagine it went something like- [NEWLINE] [NEWLINE] Officer - Hey you're blocking traffic [NEWLINE] [NEWLINE] Michael - I ain't rob no store you fucking pig. [NEWLINE] [NEWLINE] Officer - I didnt say... [NEWLINE] [NEWLINE] Michael - These cigars ain't from the store. [NEWLINE] [NEWLINE] Officer - I didn't ask about.... [NEWLINE] [NEWLINE] Michael - Fuck all ya. Always tryin a hold me down [NEWLINE] [NEWLINE] Officer - Ok put your hands in the air! [NEWLINE] [NEWLINE] Michael - Fly high or die tryin!! YOLO!! AHHH [charges head first at cop and gets shot like a retard] [NEWLINE] [NEWLINE] Seriously, if the cop was a sick fuck and gets a boner shooting up black people, he could have done so in a secluded area at night when no one was around. It would have been his word against the world, no witnesses. Somebody please explain to me why people protesting in favor of this dead thief isn't a moron. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi, thank you for commenting! [NEWLINE] [NEWLINE] [1]Specialist wikis were brought up before, and<mask><mask> that is a hole in my view,<mask> they aren't the focus of it<mask> adding more detail than a general encyclopedia would allow or require (ie about specific areas or fictional universes) does not create 'competition' with the main wiki. [NEWLINE] [NEWLINE] [2]<mask> for the definition of a wiki, I would<mask><mask> my definition is relatively accurate for *today*. I am relatively young and did not see the early days of the Internet,<mask> my perspective is mostly from the modern era of the total dominance and mostly-trustworthiness of WP. [NEWLINE] [NEWLINE] [3][4]<mask><mask> it would be better to shut down Usenet in favor of Reddit. :P [NEWLINE] [NEWLINE] <mask> on point, I do think that is a major problem and deserves addressing. My favorite SRs are DebateAnX subs<mask> well<mask> CMV itself,<mask><mask><mask> the *primary* advantage of the internet is putting ordinary people in contact with other ordinary people all across the world who have very different perspectives - giving everyone, in essence, a perspective well-rounded by worldwide contact. [NEWLINE] [NEWLINE] <mask>, I didn't mean to imply that I support the unilateral shutting down of sites like RW or CP - just that they should be ignored. [NEWLINE] [NEWLINE] <mask> for newspapers,<mask><mask> the Wiki is a different type of media<mask> it allows editing by everyone, whereas newspapers are, well, newspapers. That said,<mask><mask> it would be *great*<mask> we could have just one really good newspaper with all the radicalism of Fox News and Huffpost filtering each other out,<mask> we could all get a balanced view of world events for once. [NEWLINE] [NEWLINE] [5]Premise 5 is based on Premise 1. [NEWLINE] [NEWLINE] <mask> for RW<mask> a'shadow-wiki', I am well aware that they are different places; perhaps most obvious is that Wikipedia makes an attempt to be serious<mask> RW is a bit tongue-in cheek.<mask>, I believe that it still in a large way attempts to replace parts of WP it sees<mask> insufficiently correct. Let me repost something from below to clarify: [NEWLINE] [NEWLINE] [STARTQ] That may be<mask>,<mask> it still qualifies in my view<mask> a 'competitor' to WP and CP<mask> its content is not unique. I say this for two reasons: [ENDQ] In an extremely scientific study, I clicked the 'random page' button 20 times on RW and made a search for
Label encoding: <s>Hi, thank you for commenting! [NEWLINE] [NEWLINE] [1]Specialist wikis were brought up before, and I agree that is a hole in my view, but they aren't the focus of it because adding more detail than a general encyclopedia would allow or require (ie about specific areas or fictional universes) does not create 'competition' with the main wiki. [NEWLINE] [NEWLINE] [2] As for the definition of a wiki, I would argue that my definition is relatively accurate for *today*. I am relatively young and did not see the early days of the Internet, so my perspective is mostly from the modern era of the total dominance and mostly-trustworthiness of WP. [NEWLINE] [NEWLINE] [3][4] I think it would be better to shut down Usenet in favor of Reddit. :P [NEWLINE] [NEWLINE] Though on point, I do think that is a major problem and deserves addressing. My favorite SRs are DebateAnX subs as well as CMV itself, because I think the *primary* advantage of the internet is putting ordinary people in contact with other ordinary people all across the world who have very different perspectives - giving everyone, in essence, a perspective well-rounded by worldwide contact. [NEWLINE] [NEWLINE] Also, I didn't mean to imply that I support the unilateral shutting down of sites like RW or CP - just that they should be ignored. [NEWLINE] [NEWLINE] As for newspapers, I think the Wiki is a different type of media because it allows editing by everyone, whereas newspapers are, well, newspapers. That said, I think it would be *great* if we could have just one really good newspaper with all the radicalism of Fox News and Huffpost filtering each other out, so we could all get a balanced view of world events for once. [NEWLINE] [NEWLINE] [5]Premise 5 is based on Premise 1. [NEWLINE] [NEWLINE] As for RW as a'shadow-wiki', I am well aware that they are different places; perhaps most obvious is that Wikipedia makes an attempt to be serious while RW is a bit tongue-in cheek. However, I believe that it still in a large way attempts to replace parts of WP it sees as insufficiently correct. Let me repost something from below to clarify: [NEWLINE] [NEWLINE] [STARTQ] That may be so, but it still qualifies in my view as a 'competitor' to WP and CP because its content is not unique. I say this for two reasons: [ENDQ] In an extremely scientific study, I clicked the 'random page' button 20 times on RW and made a search for
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0964, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0858, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0999, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0977, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1474, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Politicians are elected not for a single viewpoint,<mask> a *composition* of many viewpoints. And there are correlations between viewpoints. This doesn't matter at the two-party stages of an election,<mask> in primaries and<mask> on, it does. [NEWLINE] [NEWLINE] Let's say there are 4 politicians competing for a democratic position. And there are only two issues everyone cares about: legalizing marijuana and same-sex marriage. [NEWLINE] [NEWLINE] Now, pretend of the democratic voters, 5% support neither marijuana legalization nor same-sex marriage. 30% support same-sex marriage and marijuana legalization. 40% support marijuana legalization<mask> not same-sex marriage. And 25% support same-sex<mask> not marijuana legalization. [NEWLINE] [NEWLINE] Each of the 4 politicians hold those 4 different view combinations.<mask> here's the important part: the politician that supports marijuana<mask> not same-sex marriage is the one who gets elected, with the highest percentage (40%).<mask> 55% of the voters support same-sex marriage! It just<mask> happens that a lot of those who support same-sex marriage do not<mask> support marijuana legalization. The politician who holds the two majority viewpoints, pro-same-sex marriage and marijuana legalization, wouldn't get elected. [NEWLINE] [NEWLINE] And then, after being elected, the politician who claimed to support marijuana legalization<mask> not same-sex marriage changes to support same-sex marriage, "based on public opinion" (the majority viewpoint). Is that okay? [NEWLINE] [NEWLINE] <mask> here, it depends<mask> you mean by "OK". On the one hand, ethically it might make sense for the politician to try to get elected,<mask> after election take an entirely different viewpoint<mask> the politician is now<mask> representing the people that didn't vote for her. [NEWLINE] [NEWLINE] <mask> it's<mask> "cheating" in a way - even<mask> the politician changed their views based on public opinion, they might not have won the election with the new public opinion and their new views. [NEWLINE] [NEWLINE] In my example,<mask> the politician who campaigned on legalizing marijuana<mask> not same-sex marriage gets elected, and then immediately changes to being pro-same-sex marriage "based on public opinion",<mask><mask> that would not be okay.<mask>, I don't think it's "cheating"<mask> the politician changes their view based on the changing public opinion of the people who voted for that politician. [NEWLINE] [NEWLINE] Another way to look at it would be to invert it and look at it from a personal perspective.<mask> would you feel about a politician you voted for with unpopular beliefs that
Label encoding: <s>Politicians are elected not for a single viewpoint, but a *composition* of many viewpoints. And there are correlations between viewpoints. This doesn't matter at the two-party stages of an election, but in primaries and so on, it does. [NEWLINE] [NEWLINE] Let's say there are 4 politicians competing for a democratic position. And there are only two issues everyone cares about: legalizing marijuana and same-sex marriage. [NEWLINE] [NEWLINE] Now, pretend of the democratic voters, 5% support neither marijuana legalization nor same-sex marriage. 30% support same-sex marriage and marijuana legalization. 40% support marijuana legalization but not same-sex marriage. And 25% support same-sex but not marijuana legalization. [NEWLINE] [NEWLINE] Each of the 4 politicians hold those 4 different view combinations. So here's the important part: the politician that supports marijuana but not same-sex marriage is the one who gets elected, with the highest percentage (40%). But 55% of the voters support same-sex marriage! It just so happens that a lot of those who support same-sex marriage do not also support marijuana legalization. The politician who holds the two majority viewpoints, pro-same-sex marriage and marijuana legalization, wouldn't get elected. [NEWLINE] [NEWLINE] And then, after being elected, the politician who claimed to support marijuana legalization but not same-sex marriage changes to support same-sex marriage, "based on public opinion" (the majority viewpoint). Is that okay? [NEWLINE] [NEWLINE] So here, it depends what you mean by "OK". On the one hand, ethically it might make sense for the politician to try to get elected, but after election take an entirely different viewpoint as the politician is now also representing the people that didn't vote for her. [NEWLINE] [NEWLINE] But it's also "cheating" in a way - even if the politician changed their views based on public opinion, they might not have won the election with the new public opinion and their new views. [NEWLINE] [NEWLINE] In my example, if the politician who campaigned on legalizing marijuana but not same-sex marriage gets elected, and then immediately changes to being pro-same-sex marriage "based on public opinion", I think that would not be okay. However, I don't think it's "cheating" if the politician changes their view based on the changing public opinion of the people who voted for that politician. [NEWLINE] [NEWLINE] Another way to look at it would be to invert it and look at it from a personal perspective. How would you feel about a politician you voted for with unpopular beliefs that
Loss: tensor(0.1215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6588, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, it troubles me that you're focusing on women in particular here. You seem to be implying that women behave in a way that is more shallow than men. You can't honestly think women are more interested in<mask> a man looks than men are the way a woman looks. [NEWLINE] [NEWLINE] Women are socialized to be attracted to a few different things; success chief among them. Successful men tend to be confident, and confident men tend to be funny (sense of humor and confidence are, of course, two things women find most attractive in a man). [NEWLINE] [NEWLINE] Men are socialized to appreciate beauty in a woman, and little else.<mask> we're talking about shallowness (I know you didn't mention that specifically) then which do you think is more shallow? Being attracted merely to beauty or merely to success? Success tells you a lot about a person. Beauty tells you nothing. [NEWLINE] [NEWLINE] Moving on. [NEWLINE] [NEWLINE] [STARTQ] [Jocks] get the girls (good looking ones or any girl they want) [ENDQ] [NEWLINE] This is clearly not true. I remember my time in high school and there was plenty of rejection to go around for everyone. [NEWLINE] [NEWLINE] [STARTQ] move on to college same thing (popular frat guys and college athletes) [ENDQ] [NEWLINE] Do you really believe that there aren't countless women of all shapes and sizes, many of whom meet and exceed our cultural standards of beauty, who enthusiastically reject these college athletes and frat boys? Do you realize that there are thousands of stunningly beautiful women who find frat boys unattractive merely by being frat boys? [NEWLINE] [NEWLINE] [STARTQ] and finally in the real world high end businessmen, athletes and anyone remotely successful. [ENDQ] [NEWLINE] <mask> you had a hypothetical choice between two girls with precisely equivalent personalities, one of whom was beautiful and the other was less<mask>, which would you pick? [NEWLINE] [NEWLINE] There are likely tens of thousands of women you could form a happy relationship with,<mask> you would pick "pretty" over "not<mask> pretty." [NEWLINE] [NEWLINE] [STARTQ] It would seem to me that most women are more interested in<mask> they will appear to be with someone rather than actually being in love. [ENDQ] [NEWLINE] I have absolutely no idea<mask> you reached this conclusion, or<mask> it has anything to do with<mask> you've said<mask> far. [NEWLINE] [NEWLINE] [STARTQ] Clarification: This is based on<mask> many women are attracted to these individuals and<mask> "good looking" they are. [ENDQ] [NEWLINE] Women are far less attracted to looks than men are, at least in western society. Regardless,<mask> you would choose a
Label encoding: <s>Firstly, it troubles me that you're focusing on women in particular here. You seem to be implying that women behave in a way that is more shallow than men. You can't honestly think women are more interested in how a man looks than men are the way a woman looks. [NEWLINE] [NEWLINE] Women are socialized to be attracted to a few different things; success chief among them. Successful men tend to be confident, and confident men tend to be funny (sense of humor and confidence are, of course, two things women find most attractive in a man). [NEWLINE] [NEWLINE] Men are socialized to appreciate beauty in a woman, and little else. If we're talking about shallowness (I know you didn't mention that specifically) then which do you think is more shallow? Being attracted merely to beauty or merely to success? Success tells you a lot about a person. Beauty tells you nothing. [NEWLINE] [NEWLINE] Moving on. [NEWLINE] [NEWLINE] [STARTQ] [Jocks] get the girls (good looking ones or any girl they want) [ENDQ] [NEWLINE] This is clearly not true. I remember my time in high school and there was plenty of rejection to go around for everyone. [NEWLINE] [NEWLINE] [STARTQ] move on to college same thing (popular frat guys and college athletes) [ENDQ] [NEWLINE] Do you really believe that there aren't countless women of all shapes and sizes, many of whom meet and exceed our cultural standards of beauty, who enthusiastically reject these college athletes and frat boys? Do you realize that there are thousands of stunningly beautiful women who find frat boys unattractive merely by being frat boys? [NEWLINE] [NEWLINE] [STARTQ] and finally in the real world high end businessmen, athletes and anyone remotely successful. [ENDQ] [NEWLINE] If you had a hypothetical choice between two girls with precisely equivalent personalities, one of whom was beautiful and the other was less so, which would you pick? [NEWLINE] [NEWLINE] There are likely tens of thousands of women you could form a happy relationship with, but you would pick "pretty" over "not as pretty." [NEWLINE] [NEWLINE] [STARTQ] It would seem to me that most women are more interested in how they will appear to be with someone rather than actually being in love. [ENDQ] [NEWLINE] I have absolutely no idea how you reached this conclusion, or how it has anything to do with what you've said so far. [NEWLINE] [NEWLINE] [STARTQ] Clarification: This is based on how many women are attracted to these individuals and how "good looking" they are. [ENDQ] [NEWLINE] Women are far less attracted to looks than men are, at least in western society. Regardless, if you would choose a
Loss: tensor(0.1086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0918, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll go from the back up for the post, hope you don't mind... [NEWLINE] [NEWLINE] [STARTQ] By saying that "it's insane to give the government more power," you imply that's a power we<mask> citizens have. That means the government can't be that powerful,<mask> it still depends on us for support.<mask> we can feel that empowered to discuss to<mask> degree we should choose to trust and empower our government,<mask> despotic can it really be? [ENDQ] [NEWLINE] I wouldn't say that the government is despotic, I do mean that we have the power to contract and expand governmental power, I do believe the democratic checks and principles are important. I am saying that lying to your citizens should not be a thing to in a democratic nation and trumps the fundamental principles on which democracy is based. [NEWLINE] [NEWLINE] [STARTQ] Not every case of a government misleading its citizens is a case<mask> "government elites" are profiting by betraying the interests of the governed. [ENDQ] [NEWLINE] Irrelevant, it's a matter of public trust.<mask> here comes the argument that the U.S. gov disclosed it at all, which is something almost no other government on Earth does. Which I guess does bring up the point that the American government is more trustworthy than others, which does influence my point to a degree,<mask> ∆<mask> I have to contest that transparency on these issue might be preferable. Democracies do not work on the principle of an "optimal" government,<mask> rather on government of the people,<mask> inefficiency is accepted<mask> long the people get to decide. [NEWLINE] [NEWLINE] [STARTQ] Not every government system is equal<mask> it comes to the degree with which it is guilty of dishonesty to its citizens, and some lies are worse than others. [ENDQ] [NEWLINE] Well yes,<mask> just<mask> someone somewhere is worse, doesn't mean that this is acceptable. For example, the murderer is not more acceptable, just<mask> they are serial killers out there. [NEWLINE] [NEWLINE] ^<mask><mask> this is<mask> adequate for the example of WW1. [NEWLINE] [NEWLINE] [STARTQ] The CIA can't let the media know every time it's started to hunt down a suspected espionage scheme, the FBI can't announce publicly that they're tracking the movements of a terrorist cell, the IRS can't broadcast that they suspect a large company has committed fraud before they have proof. Certain vital functions of the government would cease to work entirely<mask> forced to be transparent, and would<mask><mask> become counterproductive to the well-being of the nation. It's true that these more covert aspects of government function
Label encoding: <s>I'll go from the back up for the post, hope you don't mind... [NEWLINE] [NEWLINE] [STARTQ] By saying that "it's insane to give the government more power," you imply that's a power we as citizens have. That means the government can't be that powerful, as it still depends on us for support. If we can feel that empowered to discuss to what degree we should choose to trust and empower our government, how despotic can it really be? [ENDQ] [NEWLINE] I wouldn't say that the government is despotic, I do mean that we have the power to contract and expand governmental power, I do believe the democratic checks and principles are important. I am saying that lying to your citizens should not be a thing to in a democratic nation and trumps the fundamental principles on which democracy is based. [NEWLINE] [NEWLINE] [STARTQ] Not every case of a government misleading its citizens is a case where "government elites" are profiting by betraying the interests of the governed. [ENDQ] [NEWLINE] Irrelevant, it's a matter of public trust. Although here comes the argument that the U.S. gov disclosed it at all, which is something almost no other government on Earth does. Which I guess does bring up the point that the American government is more trustworthy than others, which does influence my point to a degree, so ∆ but I have to contest that transparency on these issue might be preferable. Democracies do not work on the principle of an "optimal" government, but rather on government of the people, so inefficiency is accepted as long the people get to decide. [NEWLINE] [NEWLINE] [STARTQ] Not every government system is equal when it comes to the degree with which it is guilty of dishonesty to its citizens, and some lies are worse than others. [ENDQ] [NEWLINE] Well yes, but just because someone somewhere is worse, doesn't mean that this is acceptable. For example, the murderer is not more acceptable, just because they are serial killers out there. [NEWLINE] [NEWLINE] ^ I think this is also adequate for the example of WW1. [NEWLINE] [NEWLINE] [STARTQ] The CIA can't let the media know every time it's started to hunt down a suspected espionage scheme, the FBI can't announce publicly that they're tracking the movements of a terrorist cell, the IRS can't broadcast that they suspect a large company has committed fraud before they have proof. Certain vital functions of the government would cease to work entirely if forced to be transparent, and would in fact become counterproductive to the well-being of the nation. It's true that these more covert aspects of government function
Loss: tensor(0.1367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1479, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, for instance, our day: [NEWLINE] [NEWLINE] 4:30 AM - Baby wakes up, hungry, angry, crying.  Feed baby. [NEWLINE] [NEWLINE] 4:45 AM - Baby goes back to sleep for half an hour.  You don't. [NEWLINE] [NEWLINE] 5:15 AM - Baby wakes up again.  Wants to get up.  Day has started.  Entertain baby.  (Note: "playing" with the baby is not super-fun for adults; it's mostly doing whatever you can to keep him from crying, which he'll sometimes do anyway,<mask> your best efforts,<mask> that's<mask> babies do.) [NEWLINE] [NEWLINE] 6:30 AM - Feed baby breakfast.  Takes about half an hour on a good day,<mask> he feels like eating<mask> you offer him.  More entertaining baby. [NEWLINE] [NEWLINE] 8:00 AM - Nanny arrives.  I have nine hours of not having to worry about the baby.  He is cared for, he is fed, he is happy, he is entertained. [NEWLINE] [NEWLINE] 5:00 PM - Nanny leaves.  Entertain baby. [NEWLINE] [NEWLINE] 5:45 PM - Feed baby dinner.  Like breakfast,<mask> with more prep time involved. [NEWLINE] [NEWLINE] 6:30 PM - Begin bedtime.  Read stories, bathtime.  Feed baby bottle to try to get him to sleep. [NEWLINE] [NEWLINE] 7:00 PM - Baby is asleep.  You've got about 2-3 hours to be an adult before you're too tired to stay awake.  Unless he wakes up at night, which is often the case.  Then all bets are off. [NEWLINE] [NEWLINE] Repeat. [NEWLINE] [NEWLINE] <mask>, even with paying somebody else to take care of the baby, you're basically always tired and have 2-3 hours a day to yourself.  Those are the good days.  On the weekends, the only thing that changes is the nine hours in the middle of the day are filled with more caring for the baby.  The rest is the same. [NEWLINE] [NEWLINE] Having experienced both, I can safely say that caring for the baby is harder, more tiring, more frustrating, and less interesting than doing my job. <mask> it's<mask> far more rewarding. <mask> there's a trade-off there. [NEWLINE] [NEWLINE] Every parent's situation is different, every baby is different, and every family is different. <mask> works for my neighbors might not work for me, and vice-versa.  That doesn't mean that one of us is
Label encoding: <s>Well, for instance, our day: [NEWLINE] [NEWLINE] 4:30 AM - Baby wakes up, hungry, angry, crying.  Feed baby. [NEWLINE] [NEWLINE] 4:45 AM - Baby goes back to sleep for half an hour.  You don't. [NEWLINE] [NEWLINE] 5:15 AM - Baby wakes up again.  Wants to get up.  Day has started.  Entertain baby.  (Note: "playing" with the baby is not super-fun for adults; it's mostly doing whatever you can to keep him from crying, which he'll sometimes do anyway, despite your best efforts, because that's what babies do.) [NEWLINE] [NEWLINE] 6:30 AM - Feed baby breakfast.  Takes about half an hour on a good day, when he feels like eating what you offer him.  More entertaining baby. [NEWLINE] [NEWLINE] 8:00 AM - Nanny arrives.  I have nine hours of not having to worry about the baby.  He is cared for, he is fed, he is happy, he is entertained. [NEWLINE] [NEWLINE] 5:00 PM - Nanny leaves.  Entertain baby. [NEWLINE] [NEWLINE] 5:45 PM - Feed baby dinner.  Like breakfast, but with more prep time involved. [NEWLINE] [NEWLINE] 6:30 PM - Begin bedtime.  Read stories, bathtime.  Feed baby bottle to try to get him to sleep. [NEWLINE] [NEWLINE] 7:00 PM - Baby is asleep.  You've got about 2-3 hours to be an adult before you're too tired to stay awake.  Unless he wakes up at night, which is often the case.  Then all bets are off. [NEWLINE] [NEWLINE] Repeat. [NEWLINE] [NEWLINE] So, even with paying somebody else to take care of the baby, you're basically always tired and have 2-3 hours a day to yourself.  Those are the good days.  On the weekends, the only thing that changes is the nine hours in the middle of the day are filled with more caring for the baby.  The rest is the same. [NEWLINE] [NEWLINE] Having experienced both, I can safely say that caring for the baby is harder, more tiring, more frustrating, and less interesting than doing my job.  But it's also far more rewarding.  So there's a trade-off there. [NEWLINE] [NEWLINE] Every parent's situation is different, every baby is different, and every family is different.  What works for my neighbors might not work for me, and vice-versa.  That doesn't mean that one of us is
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3968, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This mainly is in response to the fact that many people hold the view that the German people<mask> a whole deserved being raped, having their private property stolen, forced from their homes, firebombed, and being indiscriminately killed during the final years of the war. [NEWLINE] [NEWLINE] A majority of voting eligible Germans did not even vote for the NSDAP to come to power, and even those who did, hardly any voted with the intention of starting a war/committing atrocities. [NEWLINE] [NEWLINE] <mask><mask> this is best looked at through a modern lens. Terrorists commonly justify their acts against civilian Westerners by stating that<mask> the government is 'for the people, by the people' then all people have blood on their hands. [NEWLINE] [NEWLINE] <mask> stated in the title,<mask><mask><mask><mask> that any US-American can be justifiably killed just<mask> some 'idiots' decided to vote for Obama, and now he is dropping bombs on the Middle East. Even for those that did vote for Obama, I highly doubt that they did<mask> with the atrocities he committed in mind. Keep in mind this is a president that has assassinated more people than any president before. This was known already before his re-election. [NEWLINE] [NEWLINE] The common response is that we are all perpetrators<mask> we do not actively fight against the acts of our government.<mask>, change does not happen overnight. Millions of people protested the Vietnam War and only after 12 years (depending<mask> you look at the start date) did the US stop.  There are countless people that are against Obama's foreign policy that results in the killing of innocent Muslims.<mask> one must not only fight an uphill battle against the government to stop these actions,<mask><mask> try to convince millions that either see it<mask> a non-issue or even support it. [NEWLINE] [NEWLINE] In a totalitarian regime this becomes even more difficult, especially<mask> one can be killed for undertaking such actions. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[
Label encoding: <s>This mainly is in response to the fact that many people hold the view that the German people as a whole deserved being raped, having their private property stolen, forced from their homes, firebombed, and being indiscriminately killed during the final years of the war. [NEWLINE] [NEWLINE] A majority of voting eligible Germans did not even vote for the NSDAP to come to power, and even those who did, hardly any voted with the intention of starting a war/committing atrocities. [NEWLINE] [NEWLINE] I think this is best looked at through a modern lens. Terrorists commonly justify their acts against civilian Westerners by stating that if the government is 'for the people, by the people' then all people have blood on their hands. [NEWLINE] [NEWLINE] As stated in the title, I do not think that any US-American can be justifiably killed just because some 'idiots' decided to vote for Obama, and now he is dropping bombs on the Middle East. Even for those that did vote for Obama, I highly doubt that they did so with the atrocities he committed in mind. Keep in mind this is a president that has assassinated more people than any president before. This was known already before his re-election. [NEWLINE] [NEWLINE] The common response is that we are all perpetrators if we do not actively fight against the acts of our government. However, change does not happen overnight. Millions of people protested the Vietnam War and only after 12 years (depending when you look at the start date) did the US stop.  There are countless people that are against Obama's foreign policy that results in the killing of innocent Muslims. Yet one must not only fight an uphill battle against the government to stop these actions, but also try to convince millions that either see it as a non-issue or even support it. [NEWLINE] [NEWLINE] In a totalitarian regime this becomes even more difficult, especially when one can be killed for undertaking such actions. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1601, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is just the "equalizer" myth that's always hung around gun advocates. The problem is that guns don't equalize, they escalate, and<mask> put *everyone*, including the women with the guns, at greater risk than<mask> guns were hard to get. On a sense, this does equalize things,<mask> by making everybody worse off. [NEWLINE] [NEWLINE] The problem people promoting this concept keep running into is that the equalizer argument relies on the implicit assumption that *all other things remain constant*, which of course they don't. First is the Hobbesian escalation problem. That is, perhaps a thief just wants to rob you,<mask> either seeing you have a gun or knowing the risk you might have a gun (in a society that makes it easier to carry guns), now they have to worry you might<mask> shoot them, even<mask> you don't want to.<mask> they have to kill you first before you kill them. Likewise, even<mask> you don't want to kill them, and you may even think they really don't want to kill you, you do have think that they might think exactly this last statement above and try to kill you first,<mask> you have to kill them first. It's a Hobbesian trap that inceases the odds of everyone getting killed<mask> they otherwise would not. [NEWLINE] [NEWLINE] Second, the ubiquity of guns increases the need for theives and robbers (and other criminals) to bring guns, just in case. [NEWLINE] [NEWLINE] Third, the ubiquity of guns<mask> makes it easier for criminals to get guns,<mask> there are more of them available, including more of them to steal or buy on a black marjet. Plus, the ubiquity of guns increases supply which brings down the price for illegal guns (<mask> they are easier to get). Making it more attractive for them to have. [NEWLINE] [NEWLINE] Fourth, it increases the odds of crimes or risks of the moment that otherwise would not have happened, including accidental shootings, drunken/drugs reducing inhibitions of pulling out a gun and/or shooting, and irrational emtional responses like anger or fear that cause people to pull out guns and maybe shoot out of anger or, cause a Hobbesian escalation. [NEWLINE] [NEWLINE] Finally, this "equality" argument fails statistically<mask> well in that men are far more likely to be victims of violence than women, with or without guns,<mask> it doesn't actually equalize anything. [NEWLINE] [NEWLINE] In short, this line of reasoning cannot survive in the face of the reality. This "equality" argument,
Label encoding: <s>This is just the "equalizer" myth that's always hung around gun advocates. The problem is that guns don't equalize, they escalate, and hence put *everyone*, including the women with the guns, at greater risk than if guns were hard to get. On a sense, this does equalize things, but by making everybody worse off. [NEWLINE] [NEWLINE] The problem people promoting this concept keep running into is that the equalizer argument relies on the implicit assumption that *all other things remain constant*, which of course they don't. First is the Hobbesian escalation problem. That is, perhaps a thief just wants to rob you, but either seeing you have a gun or knowing the risk you might have a gun (in a society that makes it easier to carry guns), now they have to worry you might also shoot them, even if you don't want to. So they have to kill you first before you kill them. Likewise, even if you don't want to kill them, and you may even think they really don't want to kill you, you do have think that they might think exactly this last statement above and try to kill you first, therefore you have to kill them first. It's a Hobbesian trap that inceases the odds of everyone getting killed when they otherwise would not. [NEWLINE] [NEWLINE] Second, the ubiquity of guns increases the need for theives and robbers (and other criminals) to bring guns, just in case. [NEWLINE] [NEWLINE] Third, the ubiquity of guns also makes it easier for criminals to get guns, because there are more of them available, including more of them to steal or buy on a black marjet. Plus, the ubiquity of guns increases supply which brings down the price for illegal guns ( since they are easier to get). Making it more attractive for them to have. [NEWLINE] [NEWLINE] Fourth, it increases the odds of crimes or risks of the moment that otherwise would not have happened, including accidental shootings, drunken/drugs reducing inhibitions of pulling out a gun and/or shooting, and irrational emtional responses like anger or fear that cause people to pull out guns and maybe shoot out of anger or, cause a Hobbesian escalation. [NEWLINE] [NEWLINE] Finally, this "equality" argument fails statistically as well in that men are far more likely to be victims of violence than women, with or without guns, so it doesn't actually equalize anything. [NEWLINE] [NEWLINE] In short, this line of reasoning cannot survive in the face of the reality. This "equality" argument,
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a one time vegetarian (even a vegan for a few years<mask> I was younger), for me the morality or immorality of eating meat was conditional. My reasoning was basically that animals feel pain and seem to feel emotions like fear (different animals to different extents).<mask> there is no need for me to cause pain to an animal capable of experiencing it in order for me to meet my nutritional requirements then it is immoral for me to do<mask>. I can<mask><mask> get all the nutrition required for healthy living without consuming products that cause animals to suffer,<mask> I feel it is most moral for me to do<mask>. [NEWLINE] [NEWLINE] <mask><mask> I still feel that<mask> I wrote above holds<mask><mask> I eat meat now (we're all flawed I guess). My first girlfriend was horribly allergic to soy, and lots of other things including chocolate. She would be hard pressed to get enough protein in her diet without meat due to her dietary restrictions. I never considered it immoral for her to eat meat<mask><mask> I felt it was immoral for me to do<mask>.<mask> I were to try to enforce my beliefs about moral dieting on others (something I wouldn't do btw) it would have to be very conditional. I might say that<mask><mask> it would be best and most moral<mask> humans who live in areas<mask> plant based nutrients can efficiently provide for sufficient nutrition for people to be healthy to avoid consuming animal products and to stop maintaining a massive factory farming system for the production of animals for slaughter.<mask><mask> this would be good for the environment, good for human health and ethical with regards to<mask> humans treat animals. [NEWLINE] [NEWLINE] This wouldn't be practical for many regions and many cultures at the moment and<mask> it wouldn't make sense to view them<mask> immoral for their dietary choices and food producing practices. Even<mask> I held to an absolute moral conviction that humans eating meat is wrong, it still wouldn't make sense to apply that to animals. The predator prey relationships that make up life for wild animals is key to the balance of lots of different biological and natural systems. Forcibly removing all predation from the wild would cause massive harm to even cute cuddly herbivores everywhere within a few animal generations,<mask> it wouldn't necessarily be moral.<mask> killing all the predators would be its own moral conundrum. [NEWLINE] [NEWLINE] tl;dr summation would be that the morality of consuming meat needs to be conditional and qualified to really make any sense at all, and that being the case it just doesn't make sense to apply it to animals or
Label encoding: <s>As a one time vegetarian (even a vegan for a few years when I was younger), for me the morality or immorality of eating meat was conditional. My reasoning was basically that animals feel pain and seem to feel emotions like fear (different animals to different extents). If there is no need for me to cause pain to an animal capable of experiencing it in order for me to meet my nutritional requirements then it is immoral for me to do so. I can in fact get all the nutrition required for healthy living without consuming products that cause animals to suffer, so I feel it is most moral for me to do so. [NEWLINE] [NEWLINE] In fact I still feel that what I wrote above holds even though I eat meat now (we're all flawed I guess). My first girlfriend was horribly allergic to soy, and lots of other things including chocolate. She would be hard pressed to get enough protein in her diet without meat due to her dietary restrictions. I never considered it immoral for her to eat meat even though I felt it was immoral for me to do so. If I were to try to enforce my beliefs about moral dieting on others (something I wouldn't do btw) it would have to be very conditional. I might say that I think it would be best and most moral if humans who live in areas where plant based nutrients can efficiently provide for sufficient nutrition for people to be healthy to avoid consuming animal products and to stop maintaining a massive factory farming system for the production of animals for slaughter. I think this would be good for the environment, good for human health and ethical with regards to how humans treat animals. [NEWLINE] [NEWLINE] This wouldn't be practical for many regions and many cultures at the moment and so it wouldn't make sense to view them as immoral for their dietary choices and food producing practices. Even if I held to an absolute moral conviction that humans eating meat is wrong, it still wouldn't make sense to apply that to animals. The predator prey relationships that make up life for wild animals is key to the balance of lots of different biological and natural systems. Forcibly removing all predation from the wild would cause massive harm to even cute cuddly herbivores everywhere within a few animal generations, so it wouldn't necessarily be moral. Also killing all the predators would be its own moral conundrum. [NEWLINE] [NEWLINE] tl;dr summation would be that the morality of consuming meat needs to be conditional and qualified to really make any sense at all, and that being the case it just doesn't make sense to apply it to animals or
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh joy, another "let's legislate women's bodies" post.... [NEWLINE] [NEWLINE] Ok,<mask> look. Even<mask> breastfeeding is preferable (and there are [sibling studies]( [URL] ) which indicate that there is little to no measurable benefit), the benefits to any individual child are *minuscule*. 4 IQ points? That may be statistically significant across large populations,<mask> for an individual it's *nothing* - take the test two times in a row and you'll see those sorts of deviations. [NEWLINE] [NEWLINE] It's formula, FFS, not vodka laced with cyanide. It may not be totally optimal,<mask> it's a *perfectly viable alternative* which will not harm your child and will not make the difference between teaching Harvard and cleaning bathrooms at the local Stop-n-Shop. [NEWLINE] [NEWLINE] Look at any classroom - look any any classroom *of infants*, not even talking about middle- or high-schoolers. You will *never* be able to tell the formula-fed kids from the breast-fed. There are just no appreciable differences on an individual level. [NEWLINE] [NEWLINE] Parenthood is fucking *tough*. It's tough enough without these appalling parent-shaming-wars and without further curtailing the resources available to parents.<mask> a woman doesn't want to breastfeed *for any reason*, being forced to do<mask> is going to, at best, cause resentment. Is that *really*<mask> you want mothers to feel toward their kids? Is breastmilk really *that much more* beneficial than a loving and mentally stable parent? Yeah. No. [NEWLINE] [NEWLINE] Parenthood is a series of compromises. Parents make *all sorts of* decisions that are less-than-optimal. You think we should ban formula?<mask> about mac-n-cheese? Surely *that's* bad for their health too. Oh, and all sugar, of course - cupcakes are the absolute devil! And cars: cars are like *super*-dangerous, better not use them (and never mind the utility trade-offs).<mask> is formula for babies<mask> much worse than soda for teenagers, in your book? [NEWLINE] [NEWLINE] <mask> else do you want to be available by prescription only? [NEWLINE] [NEWLINE] (Just<mask> an addendum: My experience with lactation consultants has made me deeply antagonistic toward the whole movement. I acknowledge my biases and am striving to work past them,<mask> seriously - fuck anybody who makes a struggling new mother feel like shit about herself and her body and her choices
Label encoding: <s>Oh joy, another "let's legislate women's bodies" post.... [NEWLINE] [NEWLINE] Ok, so look. Even if breastfeeding is preferable (and there are [sibling studies]( [URL] ) which indicate that there is little to no measurable benefit), the benefits to any individual child are *minuscule*. 4 IQ points? That may be statistically significant across large populations, but for an individual it's *nothing* - take the test two times in a row and you'll see those sorts of deviations. [NEWLINE] [NEWLINE] It's formula, FFS, not vodka laced with cyanide. It may not be totally optimal, but it's a *perfectly viable alternative* which will not harm your child and will not make the difference between teaching Harvard and cleaning bathrooms at the local Stop-n-Shop. [NEWLINE] [NEWLINE] Look at any classroom - look any any classroom *of infants*, not even talking about middle- or high-schoolers. You will *never* be able to tell the formula-fed kids from the breast-fed. There are just no appreciable differences on an individual level. [NEWLINE] [NEWLINE] Parenthood is fucking *tough*. It's tough enough without these appalling parent-shaming-wars and without further curtailing the resources available to parents. If a woman doesn't want to breastfeed *for any reason*, being forced to do so is going to, at best, cause resentment. Is that *really* how you want mothers to feel toward their kids? Is breastmilk really *that much more* beneficial than a loving and mentally stable parent? Yeah. No. [NEWLINE] [NEWLINE] Parenthood is a series of compromises. Parents make *all sorts of* decisions that are less-than-optimal. You think we should ban formula? What about mac-n-cheese? Surely *that's* bad for their health too. Oh, and all sugar, of course - cupcakes are the absolute devil! And cars: cars are like *super*-dangerous, better not use them (and never mind the utility trade-offs). Why is formula for babies so much worse than soda for teenagers, in your book? [NEWLINE] [NEWLINE] What else do you want to be available by prescription only? [NEWLINE] [NEWLINE] (Just as an addendum: My experience with lactation consultants has made me deeply antagonistic toward the whole movement. I acknowledge my biases and am striving to work past them, but seriously - fuck anybody who makes a struggling new mother feel like shit about herself and her body and her choices
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly let me say that this is my first post in this sub, and I'm hoping that it's the correct place.  I would like to see the counter argument to my view(s)<mask> that I can objectively reanalyze my stance on them and decide<mask> I still feel this way.  I was referred here by another poster and I've read through the rules and it seems like the perfect sub for a good debate :).  That being said: [NEWLINE] [NEWLINE] I don't understand<mask> it's inappropriate to use profanity or even for children to be exposed to or even use profanity.  "Swear words" are literally just words, and most of them are considered slang or profanity<mask> they originated from the Anglo-Saxon words.  I do not allow my children to swear,<mask> every time that I tell them not to I feel like a hypocrite.  The only reasons I don't allow them to swear is<mask> I don't want them to get into the habit, then let a swear word slip at school and end up in trouble.  Society seems to really look down on children using profanity,<mask> I<mask> don't want an investigation from DCF or anything like that. [NEWLINE] [NEWLINE] I understand that some people find them offensive from a religious standpoint,<mask><mask> a non religious person myself, this doesn't apply to me and I don't feel like it's a good "anti-profanity argument"<mask> it doesn't coincide with my morality.  My moral code is simple: does this action harm another individual in some way? <mask><mask>, analyze the action and determine<mask> the fault lies in the action or the person being harmed by it.  Someone taking offense to profanity falls in the category of "just<mask> you're offended doesn't make you right."  Change my view? [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions
Label encoding: <s>Firstly let me say that this is my first post in this sub, and I'm hoping that it's the correct place.  I would like to see the counter argument to my view(s) so that I can objectively reanalyze my stance on them and decide if I still feel this way.  I was referred here by another poster and I've read through the rules and it seems like the perfect sub for a good debate :).  That being said: [NEWLINE] [NEWLINE] I don't understand why it's inappropriate to use profanity or even for children to be exposed to or even use profanity.  "Swear words" are literally just words, and most of them are considered slang or profanity since they originated from the Anglo-Saxon words.  I do not allow my children to swear, but every time that I tell them not to I feel like a hypocrite.  The only reasons I don't allow them to swear is because I don't want them to get into the habit, then let a swear word slip at school and end up in trouble.  Society seems to really look down on children using profanity, so I also don't want an investigation from DCF or anything like that. [NEWLINE] [NEWLINE] I understand that some people find them offensive from a religious standpoint, but as a non religious person myself, this doesn't apply to me and I don't feel like it's a good "anti-profanity argument" since it doesn't coincide with my morality.  My moral code is simple: does this action harm another individual in some way?  If so, analyze the action and determine if the fault lies in the action or the person being harmed by it.  Someone taking offense to profanity falls in the category of "just because you're offended doesn't make you right."  Change my view? [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] For several thousand years it has been the duty of clergy and religious institutions to formalize monogamous relationships into marriage, which is still the case in several countries. Only recently, with the advent of supra-religious states brought about by the French Revolution has it been made into a secular issue. [ENDQ] [NEWLINE] Let's get something straight here.<mask><mask><mask>, people have been doing monogamous relationships for far longer than any religion has existed. Yes, we have done polygamous relationships<mask> well,<mask> there is no reason to think that religion has a monopoly on unity between two people. Second, the "duty" you talk about was performed most commonly not for love,<mask> for duty, treaties, merging of properties, etc.<mask> you think that marriage has recently been "corrupted", then certainly that would apply to the idea that it is about *love* at all,<mask> this certainly wasn't the case in most early societies. [NEWLINE] [NEWLINE] Now to the most important point... [NEWLINE] [NEWLINE] Nobody is *forcing* you to get married, or even register your marriage with the state. People make the argument you are making, and I genuinely have to wonder<mask> they understand<mask> marriage works in our society. [NEWLINE] [NEWLINE] You are not prohibited from living and loving any person of the opposite gender. The government is not going to arrest Tony<mask> he loves Andrea, or<mask> they live together. [NEWLINE] [NEWLINE] <mask> you don't want to give the State "power", then you don't have to apply for Marriage status, and you can just tell everyone you are married without having it be recognized by the state. [NEWLINE] [NEWLINE] The State *does* have a vested interest in Marriage<mask>, and that is<mask> it provides a legal shortcut for couples who want to register. Again, couples **DO NOT HAVE TO REGISTER<mask> THEY<mask> CHOOSE.** [NEWLINE] [NEWLINE] There are tons of these "legal shortcuts" that Marriage provides, which is one of the reasons<mask> Homosexual Marriage is gaining more steam. There are *tangible* benefits that can be gained by registering your marriage with the State. [NEWLINE] [NEWLINE] Don't think of it<mask> the state saying "We are officially sanctioning your relationship." Instead, think of it<mask> the state saying the following... [NEWLINE] [NEWLINE] "We understand that you two people are in a relationship that combines many legal aspects of your persons together.<mask>,<mask> you wish to register with us, we can make it much easier for you to combine your persons legally, instead of both of you having to do each part individually." [NEWLINE] [NEWLINE]
Label encoding: <s> [STARTQ] For several thousand years it has been the duty of clergy and religious institutions to formalize monogamous relationships into marriage, which is still the case in several countries. Only recently, with the advent of supra-religious states brought about by the French Revolution has it been made into a secular issue. [ENDQ] [NEWLINE] Let's get something straight here. First of all, people have been doing monogamous relationships for far longer than any religion has existed. Yes, we have done polygamous relationships as well, but there is no reason to think that religion has a monopoly on unity between two people. Second, the "duty" you talk about was performed most commonly not for love, but for duty, treaties, merging of properties, etc. If you think that marriage has recently been "corrupted", then certainly that would apply to the idea that it is about *love* at all, when this certainly wasn't the case in most early societies. [NEWLINE] [NEWLINE] Now to the most important point... [NEWLINE] [NEWLINE] Nobody is *forcing* you to get married, or even register your marriage with the state. People make the argument you are making, and I genuinely have to wonder if they understand how marriage works in our society. [NEWLINE] [NEWLINE] You are not prohibited from living and loving any person of the opposite gender. The government is not going to arrest Tony because he loves Andrea, or because they live together. [NEWLINE] [NEWLINE] If you don't want to give the State "power", then you don't have to apply for Marriage status, and you can just tell everyone you are married without having it be recognized by the state. [NEWLINE] [NEWLINE] The State *does* have a vested interest in Marriage though, and that is why it provides a legal shortcut for couples who want to register. Again, couples **DO NOT HAVE TO REGISTER IF THEY SO CHOOSE.** [NEWLINE] [NEWLINE] There are tons of these "legal shortcuts" that Marriage provides, which is one of the reasons why Homosexual Marriage is gaining more steam. There are *tangible* benefits that can be gained by registering your marriage with the State. [NEWLINE] [NEWLINE] Don't think of it as the state saying "We are officially sanctioning your relationship." Instead, think of it as the state saying the following... [NEWLINE] [NEWLINE] "We understand that you two people are in a relationship that combines many legal aspects of your persons together. Therefore, if you wish to register with us, we can make it much easier for you to combine your persons legally, instead of both of you having to do each part individually." [NEWLINE] [NEWLINE]
Loss: tensor(0.1349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The natural objection to your view is that the court can only rule on cases that are brought to them<mask> it severely limits the absolute power they wield in terms of making final decisions on all manner of legal questions.  Points 1, 2, and 4 of yours<mask>, certainly show that the court has quite a lot of formidable power to wield,<mask> considering all of those powers can only be used in reaction to other entities exercizing their own powers first, then being sued for it, followed by multiple appeals, it is hard to view<mask> essentially is a veto power<mask> equal to those they are ruling over.   Add to that the fact that even that veto power is constrained by the majority of justices having to agree<mask> to rule and it's even harder to consider the power they wield to be unrivaled in the country. [NEWLINE] [NEWLINE] For number 5, you never compared the likelihood of removal from the Supreme Court to the other branches.   In neither of the other branches is it easy or common for an official to be forcibly removed from office<mask> a president has never been removed either and only 20 of over 12,000 congressmen have ever been removed.   Of course they still don't have to face public reelection<mask><mask> their job does not involve representing the people's will or interests<mask> they see fit anyway, it really makes no difference.  Their job is to apply the laws which have already been written (supposedly) to thay effect. [NEWLINE] [NEWLINE] The meat of your case seems to be 3<mask>, in that they can effectively can create laws.  Characterizing the case you sited in support of that notion<mask> the Court creating a law looks to me like an enormous stretch<mask>.  Based on<mask> I read, it seems perhaps more accurate to say that they established Congress' right to act with new powers that were not specifically granted them by the Constitution,<mask> were instead implied.  Taking your characterization [NEWLINE] for granted<mask>, they still can "create new laws" that Congress has sought to use and someone else has fought to prevent them from using. [NEWLINE] [NEWLINE] **TL;DR** [NEWLINE] All of the unique advantages the Supreme Court has that you list are fundamentally restricted in scope compared to the advantages the other branches have.  The most basic and important power common to both of the other branches is simply the ability to use their powers<mask> they choose to.  That basic ability affords them far more flexibility than the Supreme Court to effect political change.  <mask> Congress collectively wanted to make a law to
Label encoding: <s>The natural objection to your view is that the court can only rule on cases that are brought to them as it severely limits the absolute power they wield in terms of making final decisions on all manner of legal questions.  Points 1, 2, and 4 of yours however, certainly show that the court has quite a lot of formidable power to wield, but considering all of those powers can only be used in reaction to other entities exercizing their own powers first, then being sued for it, followed by multiple appeals, it is hard to view what essentially is a veto power as equal to those they are ruling over.   Add to that the fact that even that veto power is constrained by the majority of justices having to agree how to rule and it's even harder to consider the power they wield to be unrivaled in the country. [NEWLINE] [NEWLINE] For number 5, you never compared the likelihood of removal from the Supreme Court to the other branches.   In neither of the other branches is it easy or common for an official to be forcibly removed from office as a president has never been removed either and only 20 of over 12,000 congressmen have ever been removed.   Of course they still don't have to face public reelection but since their job does not involve representing the people's will or interests as they see fit anyway, it really makes no difference.  Their job is to apply the laws which have already been written (supposedly) to thay effect. [NEWLINE] [NEWLINE] The meat of your case seems to be 3 though, in that they can effectively can create laws.  Characterizing the case you sited in support of that notion as the Court creating a law looks to me like an enormous stretch though.  Based on what I read, it seems perhaps more accurate to say that they established Congress' right to act with new powers that were not specifically granted them by the Constitution, but were instead implied.  Taking your characterization [NEWLINE] for granted though, they still can "create new laws" that Congress has sought to use and someone else has fought to prevent them from using. [NEWLINE] [NEWLINE] **TL;DR** [NEWLINE] All of the unique advantages the Supreme Court has that you list are fundamentally restricted in scope compared to the advantages the other branches have.  The most basic and important power common to both of the other branches is simply the ability to use their powers when they choose to.  That basic ability affords them far more flexibility than the Supreme Court to effect political change.   If Congress collectively wanted to make a law to
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>,<mask> is fat shaming the right answer to that problem, or are there other solutions we could put in place that would reduce obesity without being judgmental against the individuals who don't maintain a healthy weight. [ENDQ] [NEWLINE] I never said it was. Personally, I'm against such abrasive interactions except in the most extreme cases,<mask> I would only be the one to intervene<mask> I knew the person well. I would be for taxes on junk food (<mask> only we could quantify<mask> junk food is) and other incentives. I am<mask> for gettimg to know people who are fat, and encouraging them to improve themselves in a positive mannar. My point was that it is still partly my business of someone is fat,<mask> they will be weighing down my pocketbook and quality of life down the road. [NEWLINE] [NEWLINE] <mask><mask><mask> the comparison to base jumping - sure, not everyone will enjoy it (<mask> I can't understand<mask> they wouldn't). And yes, junk food can be enjoyable, too (I have a particular fondness for Nutty Bars, myself).<mask>, people don't get fat from the occasional indulgence.<mask> junk food makes you fat, it is<mask> you are eating it regularly - at which point, it seems like less of a joy and more of a nasty habit, like smoking. And there are fat people who only buy organic, too. You can't<mask><mask> every fat person is fat<mask> they love eating any sort of food, either - some fat people couldn't care less about<mask> their food tastes, and would rather not be fat,<mask> well.<mask><mask><mask><mask>, *every* base jumper actively seeks out their thrill. Base jumping is a concious choice made for improving life happiness - having a poor diet, and the consequent fatness, is only half active. No one plans a trip to the grocery store a week in advance, prepares their grocery list only after having a competent shopper check it the previous 15 times, and lets out primal shouts of joy<mask> they exit the checkout line. Or,<mask> they do, I would bet that they are not eating junk. Having a poor diet is a passive decision. It is something that happens<mask> you din't know, you don't care, or you've given up.<mask>, I would<mask><mask> it would be silly to try to talk a base jumper into a less risky life. I doubt anyone will believe that saying "c'mon dude, let's go to the bar -<mask> we do it regularly,
Label encoding: <s> [STARTQ] Indeed, but is fat shaming the right answer to that problem, or are there other solutions we could put in place that would reduce obesity without being judgmental against the individuals who don't maintain a healthy weight. [ENDQ] [NEWLINE] I never said it was. Personally, I'm against such abrasive interactions except in the most extreme cases, where I would only be the one to intervene if I knew the person well. I would be for taxes on junk food ( if only we could quantify what junk food is) and other incentives. I am also for gettimg to know people who are fat, and encouraging them to improve themselves in a positive mannar. My point was that it is still partly my business of someone is fat, because they will be weighing down my pocketbook and quality of life down the road. [NEWLINE] [NEWLINE] As far as the comparison to base jumping - sure, not everyone will enjoy it ( although I can't understand why they wouldn't). And yes, junk food can be enjoyable, too (I have a particular fondness for Nutty Bars, myself). However, people don't get fat from the occasional indulgence. If junk food makes you fat, it is because you are eating it regularly - at which point, it seems like less of a joy and more of a nasty habit, like smoking. And there are fat people who only buy organic, too. You can't argue that every fat person is fat because they love eating any sort of food, either - some fat people couldn't care less about how their food tastes, and would rather not be fat, as well. On the other hand, *every* base jumper actively seeks out their thrill. Base jumping is a concious choice made for improving life happiness - having a poor diet, and the consequent fatness, is only half active. No one plans a trip to the grocery store a week in advance, prepares their grocery list only after having a competent shopper check it the previous 15 times, and lets out primal shouts of joy when they exit the checkout line. Or, if they do, I would bet that they are not eating junk. Having a poor diet is a passive decision. It is something that happens because you din't know, you don't care, or you've given up. Thus, I would argue that it would be silly to try to talk a base jumper into a less risky life. I doubt anyone will believe that saying "c'mon dude, let's go to the bar - if we do it regularly,
Loss: tensor(0.1106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0999, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0941, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People's talking points about their beliefs and their underlying reasoning don't always line up.<mask><mask>, people aren't always consciously aware of the reasons for their beliefs. [NEWLINE] [NEWLINE] I have a buddy who often gets into little arguments with his girlfriend. After a<mask> they realized that they weren't actually upset about the issues they were fighting about or each other. They were just forgetting to eat lunch and getting irritable from low blood sugar. [NEWLINE] [NEWLINE] We don't have a mind reading device,<mask> we can never truly know anyone's innermost thoughts,<mask> we can get some good clues from a number of sources. [NEWLINE] [NEWLINE] <mask> much<mask> the conservative narrative is that this is an issue about the fetus,<mask> not the woman's rights, a lot of the details of<mask> conservatives approach the dialogue makes me doubt that. [NEWLINE] [NEWLINE] [The large majority of conservatives support an exception for rape and incest]( [URL] ). Think about that. It means that the life of the fetus is no longer the first priority<mask> it was not the woman's choice to have sex. That's very much about the woman's choice, isn't it? [NEWLINE] [NEWLINE] Responses to moral thought experiments are<mask> telling. In western society, the right to life does not trump bodily autonomy. I can't be forced to donate my kidney<mask> you'll die without one. From the conservative perspective, people shouldn't even be forced to part with money to save others lives. [NEWLINE] [NEWLINE] [There's a thought experiment about a famous violinist]( [URL] /). Basically the idea is, you wake up to find that you have been drugged and<mask> you were under, your kidneys have been hooked up to an ailing famous violinist to keep him alive. He will die<mask> you disconnect. The question is whether it is ethical to do<mask>, or<mask> one must be forced to remain connected. [NEWLINE] [NEWLINE] Very very few people say that one has an ethical duty to remain connected. People stick by the notion that bodily autonomy is given precedence over the lives of others.<mask>, pro-life conservatives generally have a number of objections to the applicability of the thought experiment to abortion. These objections show that, to conservatives, this isn't ultimately only about the fetus's life. They object that a woman having sex "knew the risk" and someone hijacked for a violinist would not. They<mask><mask> a woman has a special duty to her child that does not exist between other humans. [NEWLINE] [NEWLINE] All of these objections show that it isn't the preservation of life that's paramount
Label encoding: <s>People's talking points about their beliefs and their underlying reasoning don't always line up. In fact, people aren't always consciously aware of the reasons for their beliefs. [NEWLINE] [NEWLINE] I have a buddy who often gets into little arguments with his girlfriend. After a while they realized that they weren't actually upset about the issues they were fighting about or each other. They were just forgetting to eat lunch and getting irritable from low blood sugar. [NEWLINE] [NEWLINE] We don't have a mind reading device, so we can never truly know anyone's innermost thoughts, but we can get some good clues from a number of sources. [NEWLINE] [NEWLINE] As much as the conservative narrative is that this is an issue about the fetus, but not the woman's rights, a lot of the details of how conservatives approach the dialogue makes me doubt that. [NEWLINE] [NEWLINE] [The large majority of conservatives support an exception for rape and incest]( [URL] ). Think about that. It means that the life of the fetus is no longer the first priority when it was not the woman's choice to have sex. That's very much about the woman's choice, isn't it? [NEWLINE] [NEWLINE] Responses to moral thought experiments are also telling. In western society, the right to life does not trump bodily autonomy. I can't be forced to donate my kidney if you'll die without one. From the conservative perspective, people shouldn't even be forced to part with money to save others lives. [NEWLINE] [NEWLINE] [There's a thought experiment about a famous violinist]( [URL] /). Basically the idea is, you wake up to find that you have been drugged and while you were under, your kidneys have been hooked up to an ailing famous violinist to keep him alive. He will die if you disconnect. The question is whether it is ethical to do so, or if one must be forced to remain connected. [NEWLINE] [NEWLINE] Very very few people say that one has an ethical duty to remain connected. People stick by the notion that bodily autonomy is given precedence over the lives of others. So, pro-life conservatives generally have a number of objections to the applicability of the thought experiment to abortion. These objections show that, to conservatives, this isn't ultimately only about the fetus's life. They object that a woman having sex "knew the risk" and someone hijacked for a violinist would not. They argue that a woman has a special duty to her child that does not exist between other humans. [NEWLINE] [NEWLINE] All of these objections show that it isn't the preservation of life that's paramount
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0996, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0949, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly I want to say that I share the same view<mask> you. I see no point in reading fiction anymore. I am getting much more enjoyment in reading scientific, social science and philosophical literature. [NEWLINE] [NEWLINE] My father<mask> says he doesn't understand<mask> would people read something you learn nothing from such<mask> 50 shades, Game of Thrones etc.<mask> he's pretty much anti-entertainment in all forms and shapes. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, there are fictions from which you can learn something such<mask> 1984 by George Orwell who<mask> masterfully explains<mask> would happen<mask> we would live in a police state. Brave New World by Aldous Huxley tackles many problems. Primarily,<mask> would happen<mask> people are given infinite amount of entertainment and<mask> drugs are free and legal.<mask>, it asks an ethical question whether is it ethical to socially engineer humans into happiness. In other words,<mask> you are conditioned all your life that you are a peasant, that you should do meaningless labor work, that you are stupid<mask> you actually are happy in the adulthood<mask> you don't remember the abuse in early childhood, is that ethical? [NEWLINE] [NEWLINE] Those are two of mine favorite fiction books<mask> I'm slightly biased cause I love the dystopian SF genre. Nevertheless, they are classics every man and woman should read. [NEWLINE] [NEWLINE] <mask><mask> that I would rather watch a TV show or a movie than read a book cause you get much more enjoyment in less time with it, and you can share it with someone. For example, I like to watch Game of Thrones with my girlfriend and see her react: "Oh my god, no! NO NO NO!"<mask> someone important is about to die. It's such a heartrending<mask><mask> cute moment with her :) [NEWLINE] [NEWLINE] <mask> I have to object here with saying:<mask> about awesome fiction which hasn't been televised<mask>? For instance, the [Miles Vorkosigan saga]( [URL] ). It's an SF far in the future which follows the life of a very intelligent young man who barely passes the military training to get into army,<mask> he has physical deformations. It has action scenes that will make you read 40 pages in 15 minutes without realizing it, plot twists which will make your head explode, love scenes which will make you go "Awww...", and satisfying endings. It's basically Game of Thrones in the future without gore and rape.<mask> you want to get started with it, I recommend doing it with the prequels: Barrayar and Shards of Honor (
Label encoding: <s>Firstly I want to say that I share the same view as you. I see no point in reading fiction anymore. I am getting much more enjoyment in reading scientific, social science and philosophical literature. [NEWLINE] [NEWLINE] My father also says he doesn't understand why would people read something you learn nothing from such as 50 shades, Game of Thrones etc. but he's pretty much anti-entertainment in all forms and shapes. [NEWLINE] [NEWLINE] On the other hand, there are fictions from which you can learn something such as 1984 by George Orwell who so masterfully explains what would happen if we would live in a police state. Brave New World by Aldous Huxley tackles many problems. Primarily, what would happen if people are given infinite amount of entertainment and if drugs are free and legal. Also, it asks an ethical question whether is it ethical to socially engineer humans into happiness. In other words, if you are conditioned all your life that you are a peasant, that you should do meaningless labor work, that you are stupid but you actually are happy in the adulthood because you don't remember the abuse in early childhood, is that ethical? [NEWLINE] [NEWLINE] Those are two of mine favorite fiction books but I'm slightly biased cause I love the dystopian SF genre. Nevertheless, they are classics every man and woman should read. [NEWLINE] [NEWLINE] I agree that I would rather watch a TV show or a movie than read a book cause you get much more enjoyment in less time with it, and you can share it with someone. For example, I like to watch Game of Thrones with my girlfriend and see her react: "Oh my god, no! NO NO NO!" when someone important is about to die. It's such a heartrending but also cute moment with her :) [NEWLINE] [NEWLINE] However I have to object here with saying: What about awesome fiction which hasn't been televised yet? For instance, the [Miles Vorkosigan saga]( [URL] ). It's an SF far in the future which follows the life of a very intelligent young man who barely passes the military training to get into army, because he has physical deformations. It has action scenes that will make you read 40 pages in 15 minutes without realizing it, plot twists which will make your head explode, love scenes which will make you go "Awww...", and satisfying endings. It's basically Game of Thrones in the future without gore and rape. If you want to get started with it, I recommend doing it with the prequels: Barrayar and Shards of Honor (
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's take a soccer player: Lionel Messi -20 Mil / year salary. [NEWLINE] Is he really worth that ammount of money? To you no. Then who is paying him that ammount and<mask> that much? [NEWLINE] [NEWLINE] [NEWLINE] The one paying him is his club-FC Barcelona.<mask>? [NEWLINE] Barcelona is making 80.000 fans show up at their stadium every week and sometimes 2 time a week pay tickets for seeng the team perform.<mask> does that have to do with Messi? Well he's a big part of the club's success. People come to see a show(which Messi is very good at delivering 80% of the cases) and to see their team win( Messi has broken 95% of records that a single soccer player can break) which again, Messi has a big addition to. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Let's say an avarage ticket for Barcelona matches is 30$. Multiply by 80.000 seats  =2.4 million$ per week. Multiply by (rough estimate) 50 matches per season = 120M/season. Just from tickets sales. [NEWLINE] [NEWLINE] <mask> he generates money. Not by himself, of course,<mask> his team mates have big salaries<mask> well. They portray an image that the club sells, and sells very well. In 2014 they had an income of 700M $. And<mask> you have watched soccer even the slightest, you know Messi is the brick on which the team is built,<mask> the brick that allows this high of an income. Selling, tickets, jerseys, promotional materials, ads, naming rights, winning a customer base is done by having players like these. Hell, even big corporations pay to have their name attributed to one sportsman or another. [NEWLINE] [NEWLINE] English Premier League TV rights were sold for 1 BILLION POUNDS. That is money that the teams get for allowing TV to broadcast their team's matches. No one would care about that matches<mask> the players were not giving a good show. Soccer and 90% of the other sports is a business. And<mask> you're in multi-billion business, you can ask to have a paycheck of several million. [NEWLINE] [NEWLINE] [NEWLINE] <mask><mask> : Money makes money, and a sportsman brings his club and its association and it's sport much more money than they are paid. The high salary is just the product of<mask> much money a business makes off that person.<mask> they decline is their ability, salary drops, focus gets on other players and the one that can attract more money in the long
Label encoding: <s>Let's take a soccer player: Lionel Messi -20 Mil / year salary. [NEWLINE] Is he really worth that ammount of money? To you no. Then who is paying him that ammount and why that much? [NEWLINE] [NEWLINE] [NEWLINE] The one paying him is his club-FC Barcelona. Why? [NEWLINE] Barcelona is making 80.000 fans show up at their stadium every week and sometimes 2 time a week pay tickets for seeng the team perform. What does that have to do with Messi? Well he's a big part of the club's success. People come to see a show(which Messi is very good at delivering 80% of the cases) and to see their team win( Messi has broken 95% of records that a single soccer player can break) which again, Messi has a big addition to. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Let's say an avarage ticket for Barcelona matches is 30$. Multiply by 80.000 seats  =2.4 million$ per week. Multiply by (rough estimate) 50 matches per season = 120M/season. Just from tickets sales. [NEWLINE] [NEWLINE] So he generates money. Not by himself, of course, but his team mates have big salaries as well. They portray an image that the club sells, and sells very well. In 2014 they had an income of 700M $. And if you have watched soccer even the slightest, you know Messi is the brick on which the team is built, therefore the brick that allows this high of an income. Selling, tickets, jerseys, promotional materials, ads, naming rights, winning a customer base is done by having players like these. Hell, even big corporations pay to have their name attributed to one sportsman or another. [NEWLINE] [NEWLINE] English Premier League TV rights were sold for 1 BILLION POUNDS. That is money that the teams get for allowing TV to broadcast their team's matches. No one would care about that matches if the players were not giving a good show. Soccer and 90% of the other sports is a business. And when you're in multi-billion business, you can ask to have a paycheck of several million. [NEWLINE] [NEWLINE] [NEWLINE] TLDR : Money makes money, and a sportsman brings his club and its association and it's sport much more money than they are paid. The high salary is just the product of how much money a business makes off that person. If they decline is their ability, salary drops, focus gets on other players and the one that can attract more money in the long
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The issue with taxes are the impact it has on the person you're taxing.<mask>, a 15% flat tax would be a significant decrease in tax revenues,<mask> that's unrealistically low. [NEWLINE] [NEWLINE] The issue with taxes is<mask> activity are the people you're taxing giving up to pay for the taxes. The more someone earns, the more likely that paying taxes is only cutting into luxury expenses. Currently, the bottom 20% of earners pay a total tax rate of about 8%,<mask> you're essentially doubling the taxes on poor people. This is taking money out of budgets for food, housing, and basic expenses.<mask> you're making $20k per year, losing another 8% of your income to taxes is devastating.<mask> you're making $500k per year, a tax cut means you maybe take extra nice vacations or buy the biggest Mercedes instead of the 2nd biggest.<mask>, spending on basic stuff is most beneficial to the economy. Tax cuts for lower income people generate more economic activity than tax cuts for wealthy people. Any time you're generating economic growth, you're<mask> growing future tax revenues,<mask> cutting taxes for poor people ultimately allows you to lower the overall tax burden. [NEWLINE] [NEWLINE] Taxing lower income people has a much larger effect on children. Lowering a families ability to invest in essential things for their kids leads to lower educational achievement, higher rates of behavioral problems and crime, etc. These all have a steep economic cost down the road. [NEWLINE] [NEWLINE] [STARTQ] It seems to me that it is punishing somebody for making more money, making it more desirable to stay in the middle class instead of encouraging people to work their way up into a higher class. [ENDQ] [NEWLINE] There is no point in our tax system<mask> making more money results in you taking less home. The top marginal rate is less than 40%. You get to keep a bit less of that extra dollar,<mask> you always have more to spend<mask> you make more money.<mask>, consider that after a certain income all your basic expenses are met, and any additional money you make is 100% disposable. [NEWLINE] [NEWLINE] [STARTQ] <mask> would still be encouraged to leave their capital in the US economy instead of finding ways around the tax code [ENDQ] [NEWLINE] Investments are taxed<mask> capital gains, which are flat<mask> you have described. A flat tax does nothing to discourage seeking ways around taxes.<mask> you have $100 k, and the tax is 15%,<mask><mask><mask> you can avoid the tax for less than $15k, you're going to do it<mask> at the end of the
Label encoding: <s>The issue with taxes are the impact it has on the person you're taxing. Firstly, a 15% flat tax would be a significant decrease in tax revenues, so that's unrealistically low. [NEWLINE] [NEWLINE] The issue with taxes is what activity are the people you're taxing giving up to pay for the taxes. The more someone earns, the more likely that paying taxes is only cutting into luxury expenses. Currently, the bottom 20% of earners pay a total tax rate of about 8%, so you're essentially doubling the taxes on poor people. This is taking money out of budgets for food, housing, and basic expenses. When you're making $20k per year, losing another 8% of your income to taxes is devastating. If you're making $500k per year, a tax cut means you maybe take extra nice vacations or buy the biggest Mercedes instead of the 2nd biggest. Additionally, spending on basic stuff is most beneficial to the economy. Tax cuts for lower income people generate more economic activity than tax cuts for wealthy people. Any time you're generating economic growth, you're also growing future tax revenues, so cutting taxes for poor people ultimately allows you to lower the overall tax burden. [NEWLINE] [NEWLINE] Taxing lower income people has a much larger effect on children. Lowering a families ability to invest in essential things for their kids leads to lower educational achievement, higher rates of behavioral problems and crime, etc. These all have a steep economic cost down the road. [NEWLINE] [NEWLINE] [STARTQ] It seems to me that it is punishing somebody for making more money, making it more desirable to stay in the middle class instead of encouraging people to work their way up into a higher class. [ENDQ] [NEWLINE] There is no point in our tax system where making more money results in you taking less home. The top marginal rate is less than 40%. You get to keep a bit less of that extra dollar, but you always have more to spend when you make more money. Additionally, consider that after a certain income all your basic expenses are met, and any additional money you make is 100% disposable. [NEWLINE] [NEWLINE] [STARTQ] but would still be encouraged to leave their capital in the US economy instead of finding ways around the tax code [ENDQ] [NEWLINE] Investments are taxed as capital gains, which are flat as you have described. A flat tax does nothing to discourage seeking ways around taxes. If you have $100 k, and the tax is 15%, as long as you can avoid the tax for less than $15k, you're going to do it because at the end of the
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1576, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1. It's completely arbitrary. I have never heard any reason<mask> to<mask> the "power plus prejudice" definition is preferable to the simple "prejudice" definition that was not circular reasoning. The only real reason I have heard is that one cannot be racist against white people or sexist against men<mask> those groups have power,<mask> "power plus prejudice" is a better definition;<mask>, this reason requires that "racism/sexism equals power plus prejudice" is true in the first place and is<mask> a fallacious argument. [ENDQ] [NEWLINE] You didnt make an argument just then you just stated your view. [NEWLINE] [NEWLINE] Its not arbitrary. Racism and racial discrimination are seperate concepts, racism is a structural entity,<mask> you talk about racism youre talking about power relations within society, youre talking about a dominant normative race and minority races which are oppressed. [NEWLINE] [NEWLINE] Racial discrimination<mask> is a single instance. You can discriminate against my race<mask><mask> im white I cant experience racism. [NEWLINE] [NEWLINE] This isnt about whats in the dictionary. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 2. It results in an inconsistent definition of<mask> is sexist/racist, even<mask> the same person holds the same views. The premise for this is that power must be taken in the context of immediate social surroundings,<mask> people do not meaningfully interact with every person in the country on a daily basis.<mask> a lot of people across a country share exactly the same views,<mask> they do not interact with each other in a meaningful way, the views do not have power<mask><mask><mask> many people share them, on the grounds that to have power, an idea must have structured support, and to have structured support, there must be interaction between believers of the idea. [ENDQ] [NEWLINE] Are you denying that minority races experience institutionalised racism? Just<mask> all the racists of america arent getting together for bi-weekly meetings doesnt mean their views dont have power. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Now, by moving people around, inconsistencies start to show up.<mask> we moved a man from group A (who is prejudiced) to group D, the same views that were once considered sexist now are not. This is<mask> he used to be in a group<mask> his prejudiced ideas had support,<mask> now that he is in group D,<mask> the idea that everyone is equal has majority support, the idea no longer has power and is<mask> not sexist. The same principle applies<mask> we moved a woman from group B to group C. [ENDQ] [NEWLINE] Whats your point?<mask> do you
Label encoding: <s> [STARTQ] 1. It's completely arbitrary. I have never heard any reason as to why the "power plus prejudice" definition is preferable to the simple "prejudice" definition that was not circular reasoning. The only real reason I have heard is that one cannot be racist against white people or sexist against men because those groups have power, therefore "power plus prejudice" is a better definition; however, this reason requires that "racism/sexism equals power plus prejudice" is true in the first place and is thus a fallacious argument. [ENDQ] [NEWLINE] You didnt make an argument just then you just stated your view. [NEWLINE] [NEWLINE] Its not arbitrary. Racism and racial discrimination are seperate concepts, racism is a structural entity, when you talk about racism youre talking about power relations within society, youre talking about a dominant normative race and minority races which are oppressed. [NEWLINE] [NEWLINE] Racial discrimination though is a single instance. You can discriminate against my race but since im white I cant experience racism. [NEWLINE] [NEWLINE] This isnt about whats in the dictionary. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 2. It results in an inconsistent definition of what is sexist/racist, even when the same person holds the same views. The premise for this is that power must be taken in the context of immediate social surroundings, because people do not meaningfully interact with every person in the country on a daily basis. If a lot of people across a country share exactly the same views, but they do not interact with each other in a meaningful way, the views do not have power regardless of how many people share them, on the grounds that to have power, an idea must have structured support, and to have structured support, there must be interaction between believers of the idea. [ENDQ] [NEWLINE] Are you denying that minority races experience institutionalised racism? Just because all the racists of america arent getting together for bi-weekly meetings doesnt mean their views dont have power. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Now, by moving people around, inconsistencies start to show up. If we moved a man from group A (who is prejudiced) to group D, the same views that were once considered sexist now are not. This is because he used to be in a group where his prejudiced ideas had support, but now that he is in group D, where the idea that everyone is equal has majority support, the idea no longer has power and is therefore not sexist. The same principle applies if we moved a woman from group B to group C. [ENDQ] [NEWLINE] Whats your point? What do you
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me start by saying that I am a feminist and I respect a parent's decision to work or stay at home. And I understand that most families need both parents working to support themselves. [NEWLINE] [NEWLINE] My mother was able to stay at home with me until I went to kindergarten. I feel I benefited from this experience immensely. My mother gave me lots of one-on-one attention and affection,<mask> well<mask> thoughtful educational experiences. I am able to vividly remember far back into my early childhood, which I believe is<mask> of the memorable experiences I had. My mother went back to work after I enrolled in school full-time, and I respect her for this. [NEWLINE] [NEWLINE] I don't have children<mask>,<mask> I am engaged and at the stage of my life<mask> I am planning my future family. I am almost done with school and about to enter into a career path<mask> staying at home long-term will not be an option. I would like to raise my children<mask> my mother did me,<mask> I cannot afford it,<mask> some form of daycare will be necessary. [NEWLINE] [NEWLINE] The money I make from my job will benefit my future children,<mask> I get upset thinking about<mask> much better off my children might be<mask> I could stay home and bond with them. I am<mask> bothered by my view<mask> I am opposed to rhetoric which pressures women into giving up their careers to be stay-at-home moms. Please change my view. [NEWLINE] [NEWLINE] Edit: A couple more points: My fiance cannot leave his career either, and we will need the dual income to afford having kids,<mask> him being a stay-at-home dad is not an option.<mask>, I see the value of socialization,<mask> this could come from a very part-time preschool (maybe 3-4 hours per day), or play-dates. After all, starting in kindergarten a child will be socialized full-time for 13 years straight, at least. [NEWLINE] [NEWLINE] I'm not sure there is anything I can do to change my situation. I just can't help<mask> feel like it is a sad situation, and my children would be better off spending all of that time with either me or my fiance, rather than with daycare providers. Many of my colleagues have said that they don't understand<mask> stay-at-home parents do all day, and I don't understand<mask> they don't understand<mask> kind of opportunities they are missing. It's not that<mask><mask> anyone needs to leave their job, it's just that I see it
Label encoding: <s>Let me start by saying that I am a feminist and I respect a parent's decision to work or stay at home. And I understand that most families need both parents working to support themselves. [NEWLINE] [NEWLINE] My mother was able to stay at home with me until I went to kindergarten. I feel I benefited from this experience immensely. My mother gave me lots of one-on-one attention and affection, as well as thoughtful educational experiences. I am able to vividly remember far back into my early childhood, which I believe is because of the memorable experiences I had. My mother went back to work after I enrolled in school full-time, and I respect her for this. [NEWLINE] [NEWLINE] I don't have children yet, but I am engaged and at the stage of my life where I am planning my future family. I am almost done with school and about to enter into a career path where staying at home long-term will not be an option. I would like to raise my children as my mother did me, but I cannot afford it, so some form of daycare will be necessary. [NEWLINE] [NEWLINE] The money I make from my job will benefit my future children, but I get upset thinking about how much better off my children might be if I could stay home and bond with them. I am also bothered by my view because I am opposed to rhetoric which pressures women into giving up their careers to be stay-at-home moms. Please change my view. [NEWLINE] [NEWLINE] Edit: A couple more points: My fiance cannot leave his career either, and we will need the dual income to afford having kids, so him being a stay-at-home dad is not an option. Also, I see the value of socialization, but this could come from a very part-time preschool (maybe 3-4 hours per day), or play-dates. After all, starting in kindergarten a child will be socialized full-time for 13 years straight, at least. [NEWLINE] [NEWLINE] I'm not sure there is anything I can do to change my situation. I just can't help but feel like it is a sad situation, and my children would be better off spending all of that time with either me or my fiance, rather than with daycare providers. Many of my colleagues have said that they don't understand what stay-at-home parents do all day, and I don't understand why they don't understand what kind of opportunities they are missing. It's not that I think anyone needs to leave their job, it's just that I see it
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0996, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Does Canada have an equivalent of the Army Corps of Engineers? [ENDQ] [NEWLINE] We have engineers,<mask> I can't comment on whether they're administrated the same way. [NEWLINE] [NEWLINE] [STARTQ] The US<mask> had a carrier battle group off of Indonesia after the Tsunami helping with airlifts, same with the Haiti earthquakes,a and the Japanese reactor/earthquakes. And domestically in every hurricane season. [ENDQ] [NEWLINE] We do the same thing; relief efforts both domestic and international. More than once the army has been put to shoveling snow. [NEWLINE] [NEWLINE] [STARTQ] The reason the US is able to do those things is<mask> they are already there, and have invested in the much larger footprint of forces then the rest of NATO. Canada doesnt have the ships to go out and help anywhere that needs it. [ENDQ] [NEWLINE] Obviously our navy or military generally doesn't hold a candle to the USA in terms of size,<mask> Canada is still capable and does show up internationally for various events including<mask> not limited to disaster relief and piracy patrols. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> is the situation like with Canada's reserve forces, is there an equivalent to the US National Guard<mask> each province/state has control of the forces which can then be nationalized? [ENDQ] [NEWLINE] The military is a national concern. Reserve forces exist and, IIRC, outnumber our full-time soldiers. Theoretically the province could order them about<mask> it become impossible to contact the DND, PMO or other relevant bodies,<mask> of course that's not an expected situation. [NEWLINE] [NEWLINE] [STARTQ] In the US those formations are made up in large part of engineering, disaster response, medical, and support formations.<mask> having them spread around the country, training a few weeks a year, and on call is useful<mask> a disaster strikes anywhere.<mask><mask> Canada doesnt have<mask> many then of course you are going to need to pull one of the few active duty units who do the same thing away from either training, or another mission to then move possible thousands of miles, instead of calling in local guys who are in the community to go respond. [ENDQ] [NEWLINE] Well Canada has more land and far fewer people<mask> naturally we have more gaps between our military facilities,<mask> for the most part every reasonably populated area has some military around somewhere, and the further out communities need to get things trucked or even flown in<mask> an emergency arises. [NEWLINE] [NEWLINE] Out in the far North the only real military presence on land would be the Canadian Rangers who aren't really useful for disaster relief considering they're generally operating in small
Label encoding: <s> [STARTQ] Does Canada have an equivalent of the Army Corps of Engineers? [ENDQ] [NEWLINE] We have engineers, but I can't comment on whether they're administrated the same way. [NEWLINE] [NEWLINE] [STARTQ] The US also had a carrier battle group off of Indonesia after the Tsunami helping with airlifts, same with the Haiti earthquakes,a and the Japanese reactor/earthquakes. And domestically in every hurricane season. [ENDQ] [NEWLINE] We do the same thing; relief efforts both domestic and international. More than once the army has been put to shoveling snow. [NEWLINE] [NEWLINE] [STARTQ] The reason the US is able to do those things is because they are already there, and have invested in the much larger footprint of forces then the rest of NATO. Canada doesnt have the ships to go out and help anywhere that needs it. [ENDQ] [NEWLINE] Obviously our navy or military generally doesn't hold a candle to the USA in terms of size, but Canada is still capable and does show up internationally for various events including but not limited to disaster relief and piracy patrols. [NEWLINE] [NEWLINE] [STARTQ] Also what is the situation like with Canada's reserve forces, is there an equivalent to the US National Guard where each province/state has control of the forces which can then be nationalized? [ENDQ] [NEWLINE] The military is a national concern. Reserve forces exist and, IIRC, outnumber our full-time soldiers. Theoretically the province could order them about if it become impossible to contact the DND, PMO or other relevant bodies, but of course that's not an expected situation. [NEWLINE] [NEWLINE] [STARTQ] In the US those formations are made up in large part of engineering, disaster response, medical, and support formations. Because having them spread around the country, training a few weeks a year, and on call is useful when a disaster strikes anywhere. But if Canada doesnt have as many then of course you are going to need to pull one of the few active duty units who do the same thing away from either training, or another mission to then move possible thousands of miles, instead of calling in local guys who are in the community to go respond. [ENDQ] [NEWLINE] Well Canada has more land and far fewer people so naturally we have more gaps between our military facilities, but for the most part every reasonably populated area has some military around somewhere, and the further out communities need to get things trucked or even flown in if an emergency arises. [NEWLINE] [NEWLINE] Out in the far North the only real military presence on land would be the Canadian Rangers who aren't really useful for disaster relief considering they're generally operating in small
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just for clarification, do you mean it's destined to fail forever? Or simply that it won't catch on in the near future? [NEWLINE] [NEWLINE] Most of your technical qualms are things that increased power, engineering and miniaturization will take on in the near future.<mask> hardware designers think it's a priority, then we aren't too far away from wireless VR goggles that aren't much bigger than regular sunglasses,<mask> bulky isn't an issue. [NEWLINE] [NEWLINE] <mask> for control, we're already going down the path with kinect that you can use your body<mask> a control. For FPS, hold the gun and pull the trigger.<mask>'s better than that?<mask> for expensive and confining. The cost will go down<mask> it becomes popular. The same thing that happens with every broad entertainment technology. And gaming is already a confined activity. People don't mind sitting on their couches or at desks to play games. It's wildly popular. [NEWLINE] [NEWLINE] That,<mask><mask><mask> I can see, takes care of #1 and #5. Maybe not tomorrow or next year,<mask> within a decade<mask> people put research into it, almost certainly. [NEWLINE] [NEWLINE] <mask> for #2, you could say the same thing about any number of wildly popular technologies. Some grandparents love their smartphones and facebook, some have no idea<mask> they do. I don't see VR<mask> meaningfully different from a lot of technologies that have thrived in that regard. [NEWLINE] [NEWLINE] <mask> for #3 and #4,<mask><mask>, concerts and chat won't be huge markets for VR. I don't think this kills the technology, it's focus is more likely to be games and other kinds of entertainment. [NEWLINE] [NEWLINE] For #6 I thought the same thing about 3d movies,<mask> they seem to be here to stay. The novelty may wear off,<mask> the pleasure of the experience isn't fully reliant on novelty. Adding quality 3d vision to games gives players a better appreciation for the physical space their avatar is occupying, alowing depth and distance to become more of a factor. It<mask> makes environments more immersive.  Snow swirls around you etc etc. The 3ds was a novelty<mask> it used 3d mostly<mask> a novelty. A more powerful system can use it to real effect. [NEWLINE] [NEWLINE] I may share your view that technology that's rolling out right now isn't going to make a huge market impact,<mask> I feel fairly certain that we'll see another push in the fairly near future<mask> technological advancement has ironed the kinks that make VR
Label encoding: <s>Just for clarification, do you mean it's destined to fail forever? Or simply that it won't catch on in the near future? [NEWLINE] [NEWLINE] Most of your technical qualms are things that increased power, engineering and miniaturization will take on in the near future. If hardware designers think it's a priority, then we aren't too far away from wireless VR goggles that aren't much bigger than regular sunglasses, so bulky isn't an issue. [NEWLINE] [NEWLINE] As for control, we're already going down the path with kinect that you can use your body as a control. For FPS, hold the gun and pull the trigger. What's better than that? As for expensive and confining. The cost will go down when it becomes popular. The same thing that happens with every broad entertainment technology. And gaming is already a confined activity. People don't mind sitting on their couches or at desks to play games. It's wildly popular. [NEWLINE] [NEWLINE] That, as far as I can see, takes care of #1 and #5. Maybe not tomorrow or next year, but within a decade if people put research into it, almost certainly. [NEWLINE] [NEWLINE] As for #2, you could say the same thing about any number of wildly popular technologies. Some grandparents love their smartphones and facebook, some have no idea what they do. I don't see VR as meaningfully different from a lot of technologies that have thrived in that regard. [NEWLINE] [NEWLINE] As for #3 and #4, I agree, concerts and chat won't be huge markets for VR. I don't think this kills the technology, it's focus is more likely to be games and other kinds of entertainment. [NEWLINE] [NEWLINE] For #6 I thought the same thing about 3d movies, but they seem to be here to stay. The novelty may wear off, but the pleasure of the experience isn't fully reliant on novelty. Adding quality 3d vision to games gives players a better appreciation for the physical space their avatar is occupying, alowing depth and distance to become more of a factor. It also makes environments more immersive.  Snow swirls around you etc etc. The 3ds was a novelty because it used 3d mostly as a novelty. A more powerful system can use it to real effect. [NEWLINE] [NEWLINE] I may share your view that technology that's rolling out right now isn't going to make a huge market impact, but I feel fairly certain that we'll see another push in the fairly near future when technological advancement has ironed the kinks that make VR
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0961, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0918, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can only speak for the us.   In my country we still struggle with a rape culture and male privilege.   Women are still too fearful to share a bathroom with men and trust they won't make threatening comments or actions. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 3. I need to work on my use of sex and gender. In my language they are both the same word, we don't have that difference,<mask> I mess them up a lot. I will work on that. [ENDQ] [NEWLINE] Sex and gender can mean the same thing, <mask> in male versus female.  In the right context you could use the words interchangeably. [NEWLINE] [NEWLINE] <mask> sex has many meanings. [NEWLINE] [NEWLINE] [STARTQ] 4. People are surprisingly passionate about who they share a bathroom with. [ENDQ] [NEWLINE] People are passionate about who they expose their most vulnerable activities with in general.   Using a bathroom can make someone feel very exposed. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask>, view changed? Kind of. I recognize that my experience is not universal and that it would be a much bigger problem in other places. I still think that unisex bathrooms would work fine around here and I heard many examples<mask> they do work fine.<mask> in other parts of the world,<mask> the real and/or the perceived danger of sexual harassment is higher it is different. [ENDQ] [NEWLINE] In my area of the us, we're starting to see a 3rd type of bathroom in busy locations.   Not only is there women's women's and men's<mask> there's a single stall private room.  Sometimes it's labeled "family" and sometimes it simply has symbols of man/woman/wheelchair.   It's used by families (mom can bring stroller in and change baby).  They're typically very large<mask> it may work for handicapped.   anyone who wants it can use, for example transgendered. [NEWLINE] [NEWLINE] People seek privacy<mask> they attend to bodily functions.   After all,<mask> they didn't, there wouldn't be doors or stalls in any bathroom.   Surely you can agree you'd prefer to have a door you can close<mask> you're pooping? [NEWLINE] [NEWLINE] <mask> much privacy is simply a function of the culture and traditions of an area.   You may not understand<mask> things are done a particular way,<mask> you do need to respect it's their way.  <mask> it's not causing you harm (which this is not), it shouldn't bother you<mask> bathrooms are segregated. </s>
Label encoding: <s>I can only speak for the us.   In my country we still struggle with a rape culture and male privilege.   Women are still too fearful to share a bathroom with men and trust they won't make threatening comments or actions. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] 3. I need to work on my use of sex and gender. In my language they are both the same word, we don't have that difference, so I mess them up a lot. I will work on that. [ENDQ] [NEWLINE] Sex and gender can mean the same thing,  as in male versus female.  In the right context you could use the words interchangeably. [NEWLINE] [NEWLINE] But sex has many meanings. [NEWLINE] [NEWLINE] [STARTQ] 4. People are surprisingly passionate about who they share a bathroom with. [ENDQ] [NEWLINE] People are passionate about who they expose their most vulnerable activities with in general.   Using a bathroom can make someone feel very exposed. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] So, view changed? Kind of. I recognize that my experience is not universal and that it would be a much bigger problem in other places. I still think that unisex bathrooms would work fine around here and I heard many examples where they do work fine. But in other parts of the world, where the real and/or the perceived danger of sexual harassment is higher it is different. [ENDQ] [NEWLINE] In my area of the us, we're starting to see a 3rd type of bathroom in busy locations.   Not only is there women's women's and men's but there's a single stall private room.  Sometimes it's labeled "family" and sometimes it simply has symbols of man/woman/wheelchair.   It's used by families (mom can bring stroller in and change baby).  They're typically very large so it may work for handicapped.   anyone who wants it can use, for example transgendered. [NEWLINE] [NEWLINE] People seek privacy when they attend to bodily functions.   After all, if they didn't, there wouldn't be doors or stalls in any bathroom.   Surely you can agree you'd prefer to have a door you can close when you're pooping? [NEWLINE] [NEWLINE] How much privacy is simply a function of the culture and traditions of an area.   You may not understand why things are done a particular way, but you do need to respect it's their way.   If it's not causing you harm (which this is not), it shouldn't bother you how bathrooms are segregated. </s>
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> these points have been brought up a lot - [NEWLINE] [NEWLINE] from my other comments: [NEWLINE] [NEWLINE] The question would read:<mask> is your party's opinion on x? [NEWLINE] A) X is good [NEWLINE] B) X is bad [NEWLINE] C) X is something we need to deal with<mask> not a priority [NEWLINE] D) we have no plans for X [NEWLINE] [NEWLINE] so each of those statments would be from pamphlets made by the parties taken word for word that you are given<mask> part of the ballot. You can look up the answers to the questions<mask> there's no reason for people to get them wrong anymore than accidentally spoiling a ballot now. You are just picking your party out of a line up and you have all the information in your hand. The pamphlet would read; [NEWLINE] we have no plans for X [NEWLINE] we think Y is bad [NEWLINE] we think Z is the biggest challenge we're facing currently. [NEWLINE] each party would have a statement for X, Y and Z and the quiz would ask about X, Y and Z; you just have to pick them out. it's just ensuring that people read the stuff they are given<mask> part of the ballot. [NEWLINE] [NEWLINE] i want to mention other things too; [NEWLINE] in the uk we already have a fair bit to read on ballots,<mask> this reading will not further disadvantage anyone. we already have things in place to help with comprehension [NEWLINE] [NEWLINE] the topics would be very vague, in the UK our parties release manifestos that contain the party's 'personality' and their general stances on things. think like news sites; economy, healthcare, education etc -<mask> the questions would be '<mask> is your chosen party's view on education' the pamphlets would have a phrase about their view on education, this phrase would be one of the options (among the phrases from the other parties) and you just have to match them up. it only serves to ensure that people are reading the general views of the party<mask> aren't taken by surprise<mask> the party does something in line with that view. [NEWLINE] [NEWLINE] u/caw81 came up with a simpler, better method for ensuring this<mask> which would require people signing that they have read a summary of all the parties they could vote for's manifestos (which would be attached to the signature line) probably a few weeks before the election<mask> part of registering to vote or something. it could<mask> be done at the voting poll,<mask> all the signed sheets would have to be included for the vote to count.</s>
Label encoding: <s>so these points have been brought up a lot - [NEWLINE] [NEWLINE] from my other comments: [NEWLINE] [NEWLINE] The question would read: what is your party's opinion on x? [NEWLINE] A) X is good [NEWLINE] B) X is bad [NEWLINE] C) X is something we need to deal with but not a priority [NEWLINE] D) we have no plans for X [NEWLINE] [NEWLINE] so each of those statments would be from pamphlets made by the parties taken word for word that you are given as part of the ballot. You can look up the answers to the questions so there's no reason for people to get them wrong anymore than accidentally spoiling a ballot now. You are just picking your party out of a line up and you have all the information in your hand. The pamphlet would read; [NEWLINE] we have no plans for X [NEWLINE] we think Y is bad [NEWLINE] we think Z is the biggest challenge we're facing currently. [NEWLINE] each party would have a statement for X, Y and Z and the quiz would ask about X, Y and Z; you just have to pick them out. it's just ensuring that people read the stuff they are given as part of the ballot. [NEWLINE] [NEWLINE] i want to mention other things too; [NEWLINE] in the uk we already have a fair bit to read on ballots, so this reading will not further disadvantage anyone. we already have things in place to help with comprehension [NEWLINE] [NEWLINE] the topics would be very vague, in the UK our parties release manifestos that contain the party's 'personality' and their general stances on things. think like news sites; economy, healthcare, education etc - so the questions would be'what is your chosen party's view on education' the pamphlets would have a phrase about their view on education, this phrase would be one of the options (among the phrases from the other parties) and you just have to match them up. it only serves to ensure that people are reading the general views of the party so aren't taken by surprise when the party does something in line with that view. [NEWLINE] [NEWLINE] u/caw81 came up with a simpler, better method for ensuring this though which would require people signing that they have read a summary of all the parties they could vote for's manifestos (which would be attached to the signature line) probably a few weeks before the election as part of registering to vote or something. it could also be done at the voting poll, but all the signed sheets would have to be included for the vote to count.</s>
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm not sure there is anything I can do to change my situation. [ENDQ] [NEWLINE] <mask><mask> you just wait a few years before you start having kids? Pay down some debt, downsize the house/cars, maybe move to a place with a lower cost of living. I believe that<mask> you really want to make it work, you can make it work. [NEWLINE] [NEWLINE] [STARTQ] I would like to raise my children<mask> my mother did me,<mask> I cannot afford it,<mask> some form of daycare will be necessary. [ENDQ] [NEWLINE] I don't know<mask> career you're going into,<mask> I was surprised to find out<mask> my first child was on the way, that it would cost almost<mask> my wife was making just to pay for day care... that is,<mask> she were to continue working we wouldn't be financially better-off than we would<mask> she quit to stay at home with the baby. (Which she did,<mask> we had already been planning to do<mask>.) Check out those numbers of actual costs; you may be surprised. (<mask> it's a matter of benefits rather than finances, consider Obamacare. With a kid and only one working parent I'm almost certain you'd be eligible for subsidies) [NEWLINE] [NEWLINE] You may<mask> consider working nights or weekends<mask> your career can accommodate it, or working part-time from home. I have a few friends who are stay-at-home parents and<mask> not working full-time in their careers, they are able to make money (and keep their skills up for future work-force re-entry) by working part-time or on weekends. There are a lot of options here for (I believe) many different careers -- one friend is a dental hygenist who works 1 shift a week and blogs about dental hygiene, another is a former schoolteacher who has some private tutoring jobs for extra income, and a third is a freaking eye doctor, who stays at home with her kids during the week and works on weekends<mask> her husband is home. I<mask> know of a single mom who worked nights and weekends<mask> a nurse<mask> she could spend most of the week with her kids and leave them with family friends on the weekends. [NEWLINE] [NEWLINE] Obviously in such a situation you'd make less than you would full-time,<mask> it could be that not having the expense of day-care or correlated expense like additional commuting, parking, car maintenance etc. could be made up for or better. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I'm not sure there is anything I can do to change my situation. [ENDQ] [NEWLINE] What if you just wait a few years before you start having kids? Pay down some debt, downsize the house/cars, maybe move to a place with a lower cost of living. I believe that if you really want to make it work, you can make it work. [NEWLINE] [NEWLINE] [STARTQ] I would like to raise my children as my mother did me, but I cannot afford it, so some form of daycare will be necessary. [ENDQ] [NEWLINE] I don't know what career you're going into, but I was surprised to find out when my first child was on the way, that it would cost almost what my wife was making just to pay for day care... that is, if she were to continue working we wouldn't be financially better-off than we would if she quit to stay at home with the baby. (Which she did, but we had already been planning to do so.) Check out those numbers of actual costs; you may be surprised. ( If it's a matter of benefits rather than finances, consider Obamacare. With a kid and only one working parent I'm almost certain you'd be eligible for subsidies) [NEWLINE] [NEWLINE] You may also consider working nights or weekends if your career can accommodate it, or working part-time from home. I have a few friends who are stay-at-home parents and while not working full-time in their careers, they are able to make money (and keep their skills up for future work-force re-entry) by working part-time or on weekends. There are a lot of options here for (I believe) many different careers -- one friend is a dental hygenist who works 1 shift a week and blogs about dental hygiene, another is a former schoolteacher who has some private tutoring jobs for extra income, and a third is a freaking eye doctor, who stays at home with her kids during the week and works on weekends while her husband is home. I also know of a single mom who worked nights and weekends as a nurse so she could spend most of the week with her kids and leave them with family friends on the weekends. [NEWLINE] [NEWLINE] Obviously in such a situation you'd make less than you would full-time, but it could be that not having the expense of day-care or correlated expense like additional commuting, parking, car maintenance etc. could be made up for or better. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I may get shit for this,<mask><mask><mask> it only exists<mask> that people can use it<mask> a scapegoat for not becoming successful (or even employed). Lots of people find it much easier to create a fictional reality in which something (i.e. privilege, The Illuminati, The Government, God, etc.) is keeping them "down."<mask> that's not<mask> life works. There are people that are able to succeed at a variety of things from an even wider variety of backgrounds. Someone's perceived level of 'privileged' shouldn't matter in the slightest. There are drug addicts who went to Harvard, and there are millionaires that didn't even go to college. [A large percentage of success]( [URL] ) is based on [luck anyways.]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] Specifically, I believe that you should treat everyone equally<mask><mask> sex, gender, race, or anything else, and that even<mask> privilege did not exist, you should still treat all people equally [ENDQ] [NEWLINE] Totally agree with that. [NEWLINE] [NEWLINE] [STARTQ] In short,<mask> it comes to my actions, I don't see any situation in which being aware of privilege affects my behavior in a manner different than that of the desire to treat everyone equally. [ENDQ] [NEWLINE] Don't worry, it doesn't. [NEWLINE] [NEWLINE] I tried to understand /r/anarchism and /r/SRS's stanch on this whole thing,<mask> I really can't understand it. I mean, I guess it just comes down to not pissing people off<mask> you're discussing things with people you've never met on reddit. I've never had someone say 'check your privilege' to me in real life,<mask> I guess I'm doing it right anyways (and yes, I hang out with anarchists, feminists, etc). [It sure does come up a lot over there<mask>.]( [URL] ;restrict_sr=on&amp;sort=hot&amp;t=all) Apparently, people are even [getting harassed about it!]( [URL] ) (<mask><mask> that poster didn't show any evidence of<mask> that harassment encompassed...) [NEWLINE] [NEWLINE] <mask>, [even most rich people have no clue<mask> to<mask> happiness works,]( [URL] )<mask> complaining about other people is likely just a detriment to your own happiness, *one* facet of your life that you do have control over.<mask> rather than make yourself unhappy by complaining about privilege, people can just stop complaining and instantly be happier.</s>
Label encoding: <s>I may get shit for this, but I think it only exists so that people can use it as a scapegoat for not becoming successful (or even employed). Lots of people find it much easier to create a fictional reality in which something (i.e. privilege, The Illuminati, The Government, God, etc.) is keeping them "down." But that's not how life works. There are people that are able to succeed at a variety of things from an even wider variety of backgrounds. Someone's perceived level of 'privileged' shouldn't matter in the slightest. There are drug addicts who went to Harvard, and there are millionaires that didn't even go to college. [A large percentage of success]( [URL] ) is based on [luck anyways.]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] Specifically, I believe that you should treat everyone equally regardless of sex, gender, race, or anything else, and that even if privilege did not exist, you should still treat all people equally [ENDQ] [NEWLINE] Totally agree with that. [NEWLINE] [NEWLINE] [STARTQ] In short, when it comes to my actions, I don't see any situation in which being aware of privilege affects my behavior in a manner different than that of the desire to treat everyone equally. [ENDQ] [NEWLINE] Don't worry, it doesn't. [NEWLINE] [NEWLINE] I tried to understand /r/anarchism and /r/SRS's stanch on this whole thing, but I really can't understand it. I mean, I guess it just comes down to not pissing people off when you're discussing things with people you've never met on reddit. I've never had someone say 'check your privilege' to me in real life, so I guess I'm doing it right anyways (and yes, I hang out with anarchists, feminists, etc). [It sure does come up a lot over there though.]( [URL] ;restrict_sr=on&amp;sort=hot&amp;t=all) Apparently, people are even [getting harassed about it!]( [URL] ) ( even though that poster didn't show any evidence of what that harassment encompassed...) [NEWLINE] [NEWLINE] Besides, [even most rich people have no clue as to how happiness works,]( [URL] ) so complaining about other people is likely just a detriment to your own happiness, *one* facet of your life that you do have control over. So rather than make yourself unhappy by complaining about privilege, people can just stop complaining and instantly be happier.</s>
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1405, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe you should win the game, by winning the actual game [ENDQ] [NEWLINE] There was once a time<mask> after normal match time AND overtime, the two teams would play "golden goal"<mask> the first team to score would win. Now I believe FIFA made away with this in the early 2000 for a few reasons [NEWLINE] [NEWLINE] 1. Teams would play defensively<mask> to stop their opponent from winning.<mask> you have two teams that are 95% defense oriented,<mask> a boring game that would be, eh? [NEWLINE] [NEWLINE] 2. Have you ever played 120 minutes of soccer? FIFA regulations only allow 3 substitutions. Not to mention 120 minutes of soccer in a hot climate? It's hard. Hell, it's super tough for the last 10 minutes in normal time. Prolonging a game unnecessesarily will not bring out the best in both teams. Personally, I don't think you can measure<mask> good a team is<mask> they are playing for 120 minutes in a 95% defense oriented positioning. [NEWLINE] [NEWLINE] [STARTQ] I don't understand<mask> there can't be an NHL playoffs style overtime [ENDQ] [NEWLINE] <mask> I am not mistaken, ice hockey has infinite substitutions,<mask> you can substitute fatigued players at will. Soccer, you're going to have at least 8 players (from each team) that have been on the pitch for those full 90 minutes. [NEWLINE] [NEWLINE] [STARTQ] <mask> player fatigue is an issue, create separate play periods with rest time inbetween [ENDQ] [NEWLINE] Scheduling, times, money, etc. It just doesn't work like that. [NEWLINE] [NEWLINE] [STARTQ] At worst, just schedule a rematch the next day [ENDQ] [NEWLINE] Fatigue doesn't go away after a night, plus scheduling. Especially in the world cup<mask> you're going to have games going on in the stadiums are certain times PLUS travel time. [NEWLINE] [NEWLINE] Soccer players are not invincible. Asking them to play 120 minutes and then the next day play another 90 minutes is a hell of a lot. Not to mention teams have pretty strict schedules, especially during championchips and league games. [NEWLINE] [NEWLINE] I know<mask> you're coming from becaues I, too, wanted the golden goal (or silver goal) to be the rule,<mask> player fatigue is just too much of a concern, and lets face it,<mask> players are fatigued, they are not going to play their best, and the team<mask> a whole is going to play defense, which makes from a pretty boring game.</s>
Label encoding: <s> [STARTQ] I believe you should win the game, by winning the actual game [ENDQ] [NEWLINE] There was once a time when after normal match time AND overtime, the two teams would play "golden goal" where the first team to score would win. Now I believe FIFA made away with this in the early 2000 for a few reasons [NEWLINE] [NEWLINE] 1. Teams would play defensively as to stop their opponent from winning. If you have two teams that are 95% defense oriented, what a boring game that would be, eh? [NEWLINE] [NEWLINE] 2. Have you ever played 120 minutes of soccer? FIFA regulations only allow 3 substitutions. Not to mention 120 minutes of soccer in a hot climate? It's hard. Hell, it's super tough for the last 10 minutes in normal time. Prolonging a game unnecessesarily will not bring out the best in both teams. Personally, I don't think you can measure how good a team is when they are playing for 120 minutes in a 95% defense oriented positioning. [NEWLINE] [NEWLINE] [STARTQ] I don't understand why there can't be an NHL playoffs style overtime [ENDQ] [NEWLINE] If I am not mistaken, ice hockey has infinite substitutions, so you can substitute fatigued players at will. Soccer, you're going to have at least 8 players (from each team) that have been on the pitch for those full 90 minutes. [NEWLINE] [NEWLINE] [STARTQ] If player fatigue is an issue, create separate play periods with rest time inbetween [ENDQ] [NEWLINE] Scheduling, times, money, etc. It just doesn't work like that. [NEWLINE] [NEWLINE] [STARTQ] At worst, just schedule a rematch the next day [ENDQ] [NEWLINE] Fatigue doesn't go away after a night, plus scheduling. Especially in the world cup where you're going to have games going on in the stadiums are certain times PLUS travel time. [NEWLINE] [NEWLINE] Soccer players are not invincible. Asking them to play 120 minutes and then the next day play another 90 minutes is a hell of a lot. Not to mention teams have pretty strict schedules, especially during championchips and league games. [NEWLINE] [NEWLINE] I know where you're coming from becaues I, too, wanted the golden goal (or silver goal) to be the rule, but player fatigue is just too much of a concern, and lets face it, when players are fatigued, they are not going to play their best, and the team as a whole is going to play defense, which makes from a pretty boring game.</s>
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0934, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me preface by saying that pornography is a 14 billion dollar industry and is responsible for such technological advances<mask> T1 internet, Video streaming, and torrenting amongst others. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The message that a man's pleasure during sex is the only thing that matters, and that female pleasure is irrelevant, which leads to women feeling like their only purpose during a sexual interaction is to bring the man to ejaculate [NEWLINE] [NEWLINE] I would<mask><mask> a majority of porn is about pleasing a woman. I understand<mask> you are coming from with this<mask> there is no way you can just generalize porn like this unless you are for example solely looking at blowjob videos. Most porns I have seen have a guy pounding away on a woman who is moaning for X amount of minutes<mask> they can have a sploosh and a grunt and be done with it after having put forth<mask> much effort. [NEWLINE] [NEWLINE] [NEWLINE] The degradation and objectification of women, again showing them that they are there to ensure the man experiences pleasure [NEWLINE] [NEWLINE] Refer to the first point. Objectification<mask><mask><mask><mask> - happens in all forms of media. I might even<mask><mask> porn is more objectifying of the men,<mask> they commonly are a disembodied penis. [NEWLINE] [NEWLINE] The abuse of individuals who either didn't consent to being in pornography in the first place, or who later regret their decision<mask> can't escape their past [NEWLINE] [NEWLINE] <mask> about the tons of people who have jobs<mask><mask><mask>? For every person who regrets being in porn, I am sure there are 3 people who enjoy making home videos for their own thrills and maybe every 2 people who regret being in porn there is a pornstar who enjoys it. [NEWLINE] [NEWLINE] Pornography addiction, which affects many men and can lead to erectile dysfunction, depression, and more [NEWLINE] [NEWLINE] Pornography addiction is a real thing, and like all things porn should be used in moderation. [NEWLINE] [NEWLINE] The harmful delusions held by men who believe that porn can act<mask> a sort of education process for picking up and sleeping with women, which leads to frustration, depression and possibly aggression/violence<mask> it turns out to be untrue in the real world [NEWLINE] [NEWLINE] No one thinks this. Porn is ridiculous. There may be some sort of sex tutorial porn site or something really specific<mask> anyone who is searching for that sort of thing is probably going to be off the mark no matter<mask> you spin it. </s>
Label encoding: <s>Let me preface by saying that pornography is a 14 billion dollar industry and is responsible for such technological advances as T1 internet, Video streaming, and torrenting amongst others. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The message that a man's pleasure during sex is the only thing that matters, and that female pleasure is irrelevant, which leads to women feeling like their only purpose during a sexual interaction is to bring the man to ejaculate [NEWLINE] [NEWLINE] I would argue that a majority of porn is about pleasing a woman. I understand where you are coming from with this but there is no way you can just generalize porn like this unless you are for example solely looking at blowjob videos. Most porns I have seen have a guy pounding away on a woman who is moaning for X amount of minutes so they can have a sploosh and a grunt and be done with it after having put forth so much effort. [NEWLINE] [NEWLINE] [NEWLINE] The degradation and objectification of women, again showing them that they are there to ensure the man experiences pleasure [NEWLINE] [NEWLINE] Refer to the first point. Objectification on the other hand - happens in all forms of media. I might even argue that porn is more objectifying of the men, where they commonly are a disembodied penis. [NEWLINE] [NEWLINE] The abuse of individuals who either didn't consent to being in pornography in the first place, or who later regret their decision but can't escape their past [NEWLINE] [NEWLINE] How about the tons of people who have jobs as a result? For every person who regrets being in porn, I am sure there are 3 people who enjoy making home videos for their own thrills and maybe every 2 people who regret being in porn there is a pornstar who enjoys it. [NEWLINE] [NEWLINE] Pornography addiction, which affects many men and can lead to erectile dysfunction, depression, and more [NEWLINE] [NEWLINE] Pornography addiction is a real thing, and like all things porn should be used in moderation. [NEWLINE] [NEWLINE] The harmful delusions held by men who believe that porn can act as a sort of education process for picking up and sleeping with women, which leads to frustration, depression and possibly aggression/violence when it turns out to be untrue in the real world [NEWLINE] [NEWLINE] No one thinks this. Porn is ridiculous. There may be some sort of sex tutorial porn site or something really specific but anyone who is searching for that sort of thing is probably going to be off the mark no matter how you spin it. </s>
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Edit: OK<mask> I give. My view is probably more a matter of preserving my romantic notion of the pursuit of truth rather than the reality of having to learn from existing natural experiments either to generate more meaningful hypotheses or to have the rigorously gathered data needed to perform retrospective studies. [NEWLINE] [NEWLINE] My friend and I had an argument about<mask> is science. My initial position was that until you begin to formulate and test hypotheses<mask> you are doing isn't science. For instance, coming up with ideas, being rigorous, and being an expert does not make you a scientist. [NEWLINE] [NEWLINE] This is not to say that ideation, rigor, and expertise are not neccessary, just that you do not have science with without something to test and the means to test it. [NEWLINE] [NEWLINE] <mask> an aside, another concern I had was that we give credence to the creation of ideas without the process then you are labeling a process "science," with its associated authority,<mask> it should not have that authority. Ultimately, the value of science is in<mask> it tests our understanding of the world<mask><mask> should we allow ourselves to call the entirety of our research efforts science? [NEWLINE] [NEWLINE] Apparently, I've offended people with this view in the past<mask> please don't take this positon<mask> an affront. I actually am leaning towards my view being incorrect,<mask> I was hoping reddit might be interested in discussing the topic and<mask> might remove this sense of ambivalence that I have about the matter. I have done the usual search for answers,<mask> I haven't found a nice conclusion. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Edit: OK so I give. My view is probably more a matter of preserving my romantic notion of the pursuit of truth rather than the reality of having to learn from existing natural experiments either to generate more meaningful hypotheses or to have the rigorously gathered data needed to perform retrospective studies. [NEWLINE] [NEWLINE] My friend and I had an argument about what is science. My initial position was that until you begin to formulate and test hypotheses what you are doing isn't science. For instance, coming up with ideas, being rigorous, and being an expert does not make you a scientist. [NEWLINE] [NEWLINE] This is not to say that ideation, rigor, and expertise are not neccessary, just that you do not have science with without something to test and the means to test it. [NEWLINE] [NEWLINE] As an aside, another concern I had was that we give credence to the creation of ideas without the process then you are labeling a process "science," with its associated authority, when it should not have that authority. Ultimately, the value of science is in how it tests our understanding of the world so why should we allow ourselves to call the entirety of our research efforts science? [NEWLINE] [NEWLINE] Apparently, I've offended people with this view in the past so please don't take this positon as an affront. I actually am leaning towards my view being incorrect, but I was hoping reddit might be interested in discussing the topic and also might remove this sense of ambivalence that I have about the matter. I have done the usual search for answers, but I haven't found a nice conclusion. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0905, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0929, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1562, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This seems to me to be a little like the Fox 'war on Christmas' thing. [NEWLINE] [NEWLINE] The purpose of affirmative action in universities,<mask> I understand it, is to try and address longstanding social inequality in American society by giving historically victimised groups preferential access to education. The theory being that their disadvantage may thereby be more quickly erased. [NEWLINE] [NEWLINE] Conservatives suffer no disadvantage whatsoever *in the real world*.<mask><mask><mask>,<mask> a general rule they tend to enjoy disproportionate wealth and power - pardon the generalisation<mask> they are for the most part either white, rich, or associated with multinational corporations.<mask> injustice, other than the injustice of not being able to extend conservative dominance to absolutely every aspect of a society, would this kind of affirmative action be designed to address? [NEWLINE] [NEWLINE] It may not seem fair or just to you that there are more people in universities openly disagreeing with real-world orthodoxies than in society at large.<mask> you should be aware that the reason heterodox thoughts are<mask> common in universities is<mask> campuses are literally the only safe place to express them. In damn-near every other context, you have to genuflect to the (largely conservative btw) conventional wisdom or suffer real consequences. In my field, for instance, that means you can study Marxist/post-structuralist and/or constructivist approaches to international relations at university,<mask> you know that the world is run by Kissengerian realists. Peddle that shit outside a uni and you're a crazy idealist, not worth listening to. Try and keep that up and see<mask> it does for your career! [NEWLINE] [NEWLINE] In short, the 'injustice' affirmative action on behalf of conservative academics would seek to address is a purely illusory one. Universities are one of the few places<mask> people can safely express non-standard ideas (which incidentally serves a pretty valuable function in terms of containing political dissent).<mask> I would try to take a broader view of the context in which universities operate before making a judgement on whether you want to try and force some sort of 'balance' on them. [NEWLINE] [NEWLINE] p.s. don't think that 'non-conservative' is the same thing<mask> 'homogeneous'. It aint! Intellectual discourse cannot be summed up in terms of a conservative-liberal dichotomy or simplistic left-right models. That's lazy.</s>
Label encoding: <s>This seems to me to be a little like the Fox 'war on Christmas' thing. [NEWLINE] [NEWLINE] The purpose of affirmative action in universities, as I understand it, is to try and address longstanding social inequality in American society by giving historically victimised groups preferential access to education. The theory being that their disadvantage may thereby be more quickly erased. [NEWLINE] [NEWLINE] Conservatives suffer no disadvantage whatsoever *in the real world*. On the contrary, as a general rule they tend to enjoy disproportionate wealth and power - pardon the generalisation but they are for the most part either white, rich, or associated with multinational corporations. What injustice, other than the injustice of not being able to extend conservative dominance to absolutely every aspect of a society, would this kind of affirmative action be designed to address? [NEWLINE] [NEWLINE] It may not seem fair or just to you that there are more people in universities openly disagreeing with real-world orthodoxies than in society at large. But you should be aware that the reason heterodox thoughts are so common in universities is because campuses are literally the only safe place to express them. In damn-near every other context, you have to genuflect to the (largely conservative btw) conventional wisdom or suffer real consequences. In my field, for instance, that means you can study Marxist/post-structuralist and/or constructivist approaches to international relations at university, but you know that the world is run by Kissengerian realists. Peddle that shit outside a uni and you're a crazy idealist, not worth listening to. Try and keep that up and see what it does for your career! [NEWLINE] [NEWLINE] In short, the 'injustice' affirmative action on behalf of conservative academics would seek to address is a purely illusory one. Universities are one of the few places where people can safely express non-standard ideas (which incidentally serves a pretty valuable function in terms of containing political dissent). So I would try to take a broader view of the context in which universities operate before making a judgement on whether you want to try and force some sort of 'balance' on them. [NEWLINE] [NEWLINE] p.s. don't think that 'non-conservative' is the same thing as 'homogeneous'. It aint! Intellectual discourse cannot be summed up in terms of a conservative-liberal dichotomy or simplistic left-right models. That's lazy.</s>
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] are you asking<mask> bigotry is unfair? [ENDQ] [NEWLINE] Im asking<mask> intolerance toward a group with an irreparably different world view is unfair. [NEWLINE] [NEWLINE] Bigotry in and of itself is not a bad thing. Look at its meaning. [NEWLINE] [NEWLINE] **big·ot·ry** [NEWLINE] [NEWLINE] **ˈbiɡətrē/** [NEWLINE] [NEWLINE] **noun** [NEWLINE] [NEWLINE] **intolerance toward those who hold different opinions from oneself.** [NEWLINE] [NEWLINE] Im very bigoted towards hurtfull and hate filled groups. I personally think everyone should be intolerant of harmful groups. Would you tell someone waxing on about<mask> evil the Nazi party was that they were being bigoted?<mask> is this different? [NEWLINE] [NEWLINE] [STARTQ] To say that it is not compatible in totality is incorrect<mask><mask><mask>. Further, the intolerance is thankfully not common and neither is it irreparable. [ENDQ] [NEWLINE] Islam demands the following from its observers: [NEWLINE] [NEWLINE] * Death sentences for Homosexuals [NEWLINE] * Subjugation of Women [NEWLINE] * Men and Women are separated and unequal [NEWLINE] * Adulterers should be stoned to death [NEWLINE] * Death to anyone who leaves Islam. [NEWLINE] [NEWLINE] <mask> is ANY of that compatible with a Western World view? [NEWLINE] [NEWLINE] [STARTQ] Isamification of the west is<mask> ridiculous<mask> it sounds and further ignores<mask> I'm saying in my post [ENDQ] [NEWLINE] <mask><mask>, Islam is ridiculous for the westernized world. It would go against everything the western world holds dear.<mask> we agree here I dont think this points to my side being unfair or bigoted. [NEWLINE] [NEWLINE] [STARTQ] Is being an echo chamber okay? [ENDQ] [NEWLINE] It totally can be. It really depends on the wants of the majority doesnt it? [NEWLINE] [NEWLINE] [STARTQ] that isn't<mask> the world works in general<mask> you can't drown out opinions and views of the minority<mask> there are laws against such a thing [ENDQ] [NEWLINE] I fully disagree. In a democracy the majority will have its side heard and its wishes followed. [NEWLINE] [NEWLINE] [STARTQ] On reddit, you are free to downvote whenever you feel like. You can choose to ignore the rules of the subreddit and not be punished for it<mask><mask><mask>. [ENDQ] [NEWLINE] Thats not true at all. Each sub has moderation rules and guidelines. Some are very lenient others are harsh and swift in their moderation. It totally depends on<mask> you spend your time. Much like the real world. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] are you asking why bigotry is unfair? [ENDQ] [NEWLINE] Im asking why intolerance toward a group with an irreparably different world view is unfair. [NEWLINE] [NEWLINE] Bigotry in and of itself is not a bad thing. Look at its meaning. [NEWLINE] [NEWLINE] **big·ot·ry** [NEWLINE] [NEWLINE] **ˈbiɡətrē/** [NEWLINE] [NEWLINE] **noun** [NEWLINE] [NEWLINE] **intolerance toward those who hold different opinions from oneself.** [NEWLINE] [NEWLINE] Im very bigoted towards hurtfull and hate filled groups. I personally think everyone should be intolerant of harmful groups. Would you tell someone waxing on about how evil the Nazi party was that they were being bigoted? Why is this different? [NEWLINE] [NEWLINE] [STARTQ] To say that it is not compatible in totality is incorrect in my opinion. Further, the intolerance is thankfully not common and neither is it irreparable. [ENDQ] [NEWLINE] Islam demands the following from its observers: [NEWLINE] [NEWLINE] * Death sentences for Homosexuals [NEWLINE] * Subjugation of Women [NEWLINE] * Men and Women are separated and unequal [NEWLINE] * Adulterers should be stoned to death [NEWLINE] * Death to anyone who leaves Islam. [NEWLINE] [NEWLINE] How is ANY of that compatible with a Western World view? [NEWLINE] [NEWLINE] [STARTQ] Isamification of the west is as ridiculous as it sounds and further ignores what I'm saying in my post [ENDQ] [NEWLINE] I agree, Islam is ridiculous for the westernized world. It would go against everything the western world holds dear. Since we agree here I dont think this points to my side being unfair or bigoted. [NEWLINE] [NEWLINE] [STARTQ] Is being an echo chamber okay? [ENDQ] [NEWLINE] It totally can be. It really depends on the wants of the majority doesnt it? [NEWLINE] [NEWLINE] [STARTQ] that isn't how the world works in general as you can't drown out opinions and views of the minority as there are laws against such a thing [ENDQ] [NEWLINE] I fully disagree. In a democracy the majority will have its side heard and its wishes followed. [NEWLINE] [NEWLINE] [STARTQ] On reddit, you are free to downvote whenever you feel like. You can choose to ignore the rules of the subreddit and not be punished for it as a result. [ENDQ] [NEWLINE] Thats not true at all. Each sub has moderation rules and guidelines. Some are very lenient others are harsh and swift in their moderation. It totally depends on where you spend your time. Much like the real world. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.1298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0941, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well you can't really characterize the cold war<mask> perpetual foreign wars for the US. It was two wars, and not against the USSR,<mask> again attempts to control the periphery of the sphere of influence. After the cold war, the wars have been fought using volunteer soldiers and the american casualties are practically nothing compared to WW2 and prior major wars. [NEWLINE] [NEWLINE] In another response I see you saying that the US has been 'dragged into' many military quagmires,<mask> this again doesn't characterize any situations post-WW2,<mask> the US was pro-actively meddling in every situation. I see now that maybe you're trying to talk about the citizens, rather than the geopolitical strategy of the actual decision-makers,<mask> maybe the US citizens get dragged into wars by our own leaders? [NEWLINE] [NEWLINE] <mask> for any domestic impoverishment, or lack of comparable social welfare programs, this is again purely political and the result of<mask> the policy-makers choose to govern. The US has been the richest country by far post-WW2, and could afford to run the best social systems even<mask> engaging in military adventurism. For a country like the US, 'costs' are in real-terms (people employed, raw resources used), not in financial terms (money is created<mask> the US deficit spends, having their own non-convertible currency and central bank, especially post gold standard). A 'cartoonishly large military budget' in money terms is just stimulative; the actual cost to society is the number of soldiers, the people/resources building aircraft carriers, bombers, etc. And we've had nothing close to a war-command-economy<mask> that actually eclipses private enterprise<mask> WW2. [NEWLINE] [NEWLINE] Now<mask> you're making a more nuanced argument that winning world hegemony after WW2 set up the political conditions that made these more elite-driven outcomes inevitable (or at least highly likely), I'm definitely intrigued by the idea.<mask> I don't really see that fleshed out in the OP,<mask> you're mostly looking at the results, rather than reasons, and taking some of the propaganda at face value ('gosh we'd like to be peaceful,<mask> we just have to be the world policeman' &amp; 'gee we'd sure like to treat our domestic economy better,<mask> we just don't have any money').</s>
Label encoding: <s>Well you can't really characterize the cold war as perpetual foreign wars for the US. It was two wars, and not against the USSR, but again attempts to control the periphery of the sphere of influence. After the cold war, the wars have been fought using volunteer soldiers and the american casualties are practically nothing compared to WW2 and prior major wars. [NEWLINE] [NEWLINE] In another response I see you saying that the US has been 'dragged into' many military quagmires, but this again doesn't characterize any situations post-WW2, where the US was pro-actively meddling in every situation. I see now that maybe you're trying to talk about the citizens, rather than the geopolitical strategy of the actual decision-makers, so maybe the US citizens get dragged into wars by our own leaders? [NEWLINE] [NEWLINE] As for any domestic impoverishment, or lack of comparable social welfare programs, this is again purely political and the result of how the policy-makers choose to govern. The US has been the richest country by far post-WW2, and could afford to run the best social systems even while engaging in military adventurism. For a country like the US, 'costs' are in real-terms (people employed, raw resources used), not in financial terms (money is created as the US deficit spends, having their own non-convertible currency and central bank, especially post gold standard). A 'cartoonishly large military budget' in money terms is just stimulative; the actual cost to society is the number of soldiers, the people/resources building aircraft carriers, bombers, etc. And we've had nothing close to a war-command-economy where that actually eclipses private enterprise since WW2. [NEWLINE] [NEWLINE] Now if you're making a more nuanced argument that winning world hegemony after WW2 set up the political conditions that made these more elite-driven outcomes inevitable (or at least highly likely), I'm definitely intrigued by the idea. But I don't really see that fleshed out in the OP, where you're mostly looking at the results, rather than reasons, and taking some of the propaganda at face value ('gosh we'd like to be peaceful, but we just have to be the world policeman' &amp; 'gee we'd sure like to treat our domestic economy better, but we just don't have any money').</s>
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0973, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1236, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In my experience, refrigerators and washing machines already work like this. I've never moved and brought either of those items, and they've always remained installed in my new house/apartment. I don't know<mask> it counts<mask> arguing against your stated view (per rule #1) for me to point out that it is already the case,<mask> it should. Otherwise you could say "the sky should be blue" and no one would be allowed to argue against you. [NEWLINE] [NEWLINE] On another subject,<mask> about furniture like TV cabinets? I personally don't own a TV,<mask> I would be pissed off<mask> I had to sacrifice valuable space for one of those just<mask> it came with the house. Even<mask> I owned one, my TV wouldn't necessarily fit on someone else's furniture. Or<mask><mask> the TV cabinet is too tall and I want something lower? Beds present a similar problem. I'm a tall guy and I need a long bed. Most people's beds are too short for me. [NEWLINE] [NEWLINE] I assume the reply to that line of reasoning is that I'm free to replace my furniture<mask> I don't like it.<mask> that's just wasteful; the world doesn't need more discarded couches. And with my too-tall-for-most-beds example, that means a new apartment would cost me much more than a short person has to spend. [NEWLINE] [NEWLINE] You may say that<mask> an apartment has shitty furniture I should factor that into my decision to take the apartment,<mask> in many places (like<mask> I live) getting an apartment is hard and you have to take<mask> you can get. [NEWLINE] [NEWLINE] <mask>, you keep mentioning hotels<mask> a counterexample. This ignores the difference between a hotel and a home. I do not feel nearly<mask> comfortable in a hotel<mask> I do at home. [NEWLINE] [NEWLINE] Your argument about items like bathroom fixtures and ceiling fans is<mask> flawed. I don't know about you,<mask> I have never looked twice at my faucet - it's just a tool. A couch is more than a tool, it's a place<mask> you spend time.<mask>, tying back to my first paragraph, the cultural attitude you describe is just not the case. I have personally moved with light fixtures that held special significance. For that matter, I have left couches behind that I didn't like.</s>
Label encoding: <s>In my experience, refrigerators and washing machines already work like this. I've never moved and brought either of those items, and they've always remained installed in my new house/apartment. I don't know if it counts as arguing against your stated view (per rule #1) for me to point out that it is already the case, but it should. Otherwise you could say "the sky should be blue" and no one would be allowed to argue against you. [NEWLINE] [NEWLINE] On another subject, what about furniture like TV cabinets? I personally don't own a TV, so I would be pissed off if I had to sacrifice valuable space for one of those just because it came with the house. Even if I owned one, my TV wouldn't necessarily fit on someone else's furniture. Or what if the TV cabinet is too tall and I want something lower? Beds present a similar problem. I'm a tall guy and I need a long bed. Most people's beds are too short for me. [NEWLINE] [NEWLINE] I assume the reply to that line of reasoning is that I'm free to replace my furniture if I don't like it. But that's just wasteful; the world doesn't need more discarded couches. And with my too-tall-for-most-beds example, that means a new apartment would cost me much more than a short person has to spend. [NEWLINE] [NEWLINE] You may say that if an apartment has shitty furniture I should factor that into my decision to take the apartment, but in many places (like where I live) getting an apartment is hard and you have to take what you can get. [NEWLINE] [NEWLINE] Also, you keep mentioning hotels as a counterexample. This ignores the difference between a hotel and a home. I do not feel nearly as comfortable in a hotel as I do at home. [NEWLINE] [NEWLINE] Your argument about items like bathroom fixtures and ceiling fans is also flawed. I don't know about you, but I have never looked twice at my faucet - it's just a tool. A couch is more than a tool, it's a place where you spend time. Besides, tying back to my first paragraph, the cultural attitude you describe is just not the case. I have personally moved with light fixtures that held special significance. For that matter, I have left couches behind that I didn't like.</s>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0885, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for the response. [NEWLINE] [NEWLINE] [STARTQ] Arguing<mask> individual rights "come from" is above my pay-grade,<mask> I'll give it a shot. For the record, I typically fall into the legal realist camp,<mask> rights exist<mask> society chooses to enforce them. [ENDQ] [NEWLINE] I don't think we have to discuss meta-ethical/meta-legal questions to address this. [NEWLINE] [NEWLINE] <mask> for your argument,<mask><mask> with it,<mask><mask><mask> it assumes that abortion is immoral/illegal.  (For the purposes of this discussion, we're arguing about<mask> should be law based on<mask> is moral,<mask> I'll use the word "moral" and related words from here on.)  The following is<mask> we must assume that abortion is immoral for your argument to work. [NEWLINE] [NEWLINE] An argument for the morality of abortion almost always, AFAIK, comes from one or both of two beliefs.  The first is that a fetus in the womb is not a person with rights.  The second belief is that a fetus in the womb is a person with rights<mask> those rights are trumped by the right of the mother to not be bothered with a 9-month pregnancy (this is often argued with a metaphor about a violinist.) [NEWLINE] [NEWLINE] <mask> you hold the first belief, the act that creates a person with human rights is not conception,<mask> giving birth (or, on some variations of the belief, carrying the baby until a certain period of time, a heartbeat exists, the baby has the ability to feel pain, etc.).  Whether to give birth is the decision of the mother and no one else. <mask> it is her decision alone, the father is not morally obligated. [NEWLINE] [NEWLINE] <mask> you hold the second belief, then the mother has no moral obligation to bear the child. <mask> the mother doesn't have this obligation, then the father clearly has no obligation to support it after it is born,<mask>, again, the decision to allow a person with rights that needs care to continue existing was hers alone. [NEWLINE] [NEWLINE] <mask> I briefly alluded to in my above comment, I actually do not hold abortion to be a viable moral alternative and<mask> agree with your argument that both parents are responsible. <mask>,<mask> abortion is seen<mask> a viable moral alternative, then it follows that only the mother is morally responsible for the care of the child.</s>
Label encoding: <s>Thanks for the response. [NEWLINE] [NEWLINE] [STARTQ] Arguing where individual rights "come from" is above my pay-grade, but I'll give it a shot. For the record, I typically fall into the legal realist camp, where rights exist if society chooses to enforce them. [ENDQ] [NEWLINE] I don't think we have to discuss meta-ethical/meta-legal questions to address this. [NEWLINE] [NEWLINE] As for your argument, I agree with it, so long as it assumes that abortion is immoral/illegal.  (For the purposes of this discussion, we're arguing about what should be law based on what is moral, so I'll use the word "moral" and related words from here on.)  The following is why we must assume that abortion is immoral for your argument to work. [NEWLINE] [NEWLINE] An argument for the morality of abortion almost always, AFAIK, comes from one or both of two beliefs.  The first is that a fetus in the womb is not a person with rights.  The second belief is that a fetus in the womb is a person with rights but those rights are trumped by the right of the mother to not be bothered with a 9-month pregnancy (this is often argued with a metaphor about a violinist.) [NEWLINE] [NEWLINE] If you hold the first belief, the act that creates a person with human rights is not conception, but giving birth (or, on some variations of the belief, carrying the baby until a certain period of time, a heartbeat exists, the baby has the ability to feel pain, etc.).  Whether to give birth is the decision of the mother and no one else.  Since it is her decision alone, the father is not morally obligated. [NEWLINE] [NEWLINE] If you hold the second belief, then the mother has no moral obligation to bear the child.  If the mother doesn't have this obligation, then the father clearly has no obligation to support it after it is born, since, again, the decision to allow a person with rights that needs care to continue existing was hers alone. [NEWLINE] [NEWLINE] As I briefly alluded to in my above comment, I actually do not hold abortion to be a viable moral alternative and therefore agree with your argument that both parents are responsible.  However, if abortion is seen as a viable moral alternative, then it follows that only the mother is morally responsible for the care of the child.</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right, first things first: [NEWLINE] [NEWLINE] We have no clue<mask> many potentially habitable planets there are, and people won't ever have a discussion of "<mask> are the chances of us being on an inhabited planet" on an uninhabited planet, on account of the fact that there's no one inhabiting it. That's called "the anthropic principle". [NEWLINE] [NEWLINE] <mask> in theory, there might be hundreds of billions of planets with the potential for life, and due to chance, only ONE had life, and naturally the one with life is the one<mask> said life says "<mask> are the chances?", possibly not even knowing of other said planets. [NEWLINE] [NEWLINE] That said, there's more than one die. There has to have been billions of puddles over the course of the earth's existence,<mask> obviously only the ones before life existed are relevant. [NEWLINE] [NEWLINE] Now, just<mask> the die is rolled, doesn't mean it won't be rolled again. Before there was life, there was weather and<mask>not (see mercury and venus and jupiter and saturn and whatever other planets, for examples of weather without life) to shake things up and roll the die again. [NEWLINE] [NEWLINE] [STARTQ] Let's say that the universe is<mask><mask> this trillion sided die.<mask>, let's say that dice has only one place it could land<mask> everything would fall into place. Those are some hefty odds. [ENDQ] [NEWLINE] Right,<mask> we don't know<mask> the odds are without actually analysing it thoroughly. Which I'm not going to do right now<mask> it's 1AM. [NEWLINE] [NEWLINE] <mask>, your edit of the original post (specifically the "ex nihilo" comment) doesn't make sense - surely, by that same logic, you could ask "<mask> did god come from?" Either way you have an uncaused cause,<mask> it kinda falls flat.<mask>, the "it hasn't been around forever" is somewhat flawed - what<mask> there was a previous universe with an ultimate "big crunch", and after that "big crunch", the result was another big bang? And that was our universe,<mask> there have been an infinite number of universes preceding us? We ultimately don't know, there's a big question mark and ultimately there's only<mask> far we can hypothesise before we just *don't have enough data*.</s>
Label encoding: <s>Right, first things first: [NEWLINE] [NEWLINE] We have no clue how many potentially habitable planets there are, and people won't ever have a discussion of " what are the chances of us being on an inhabited planet" on an uninhabited planet, on account of the fact that there's no one inhabiting it. That's called "the anthropic principle". [NEWLINE] [NEWLINE] So in theory, there might be hundreds of billions of planets with the potential for life, and due to chance, only ONE had life, and naturally the one with life is the one where said life says " what are the chances?", possibly not even knowing of other said planets. [NEWLINE] [NEWLINE] That said, there's more than one die. There has to have been billions of puddles over the course of the earth's existence, although obviously only the ones before life existed are relevant. [NEWLINE] [NEWLINE] Now, just because the die is rolled, doesn't mean it won't be rolled again. Before there was life, there was weather and whatnot (see mercury and venus and jupiter and saturn and whatever other planets, for examples of weather without life) to shake things up and roll the die again. [NEWLINE] [NEWLINE] [STARTQ] Let's say that the universe is in fact this trillion sided die. However, let's say that dice has only one place it could land where everything would fall into place. Those are some hefty odds. [ENDQ] [NEWLINE] Right, but we don't know what the odds are without actually analysing it thoroughly. Which I'm not going to do right now because it's 1AM. [NEWLINE] [NEWLINE] Also, your edit of the original post (specifically the "ex nihilo" comment) doesn't make sense - surely, by that same logic, you could ask " where did god come from?" Either way you have an uncaused cause, so it kinda falls flat. Also, the "it hasn't been around forever" is somewhat flawed - what if there was a previous universe with an ultimate "big crunch", and after that "big crunch", the result was another big bang? And that was our universe, but there have been an infinite number of universes preceding us? We ultimately don't know, there's a big question mark and ultimately there's only so far we can hypothesise before we just *don't have enough data*.</s>
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Idk.  I just feel like (in my town at least) I see all sorts of "homeless" people who stand on the street corner begging for money.  On the strip of road that they beg on there are at least 35 businesses that hire all forms of work.  Sure, they're minimum wage,<mask> with the twelve hours a day you spend standing outside begging for money<mask> don't you get a fucking job and find yourself some housing? <mask>, they all drink and smoke.  Not sure<mask> that's the case everywhere,<mask> they literally don't seem to give a fuck. [NEWLINE] [NEWLINE] My view:  Homeless people are homeless by choice - sure, they may have had an incident that forces them out of their home,<mask> then they keep themselves homeless by giving up and falling into despair. [NEWLINE] [NEWLINE] I do not give money to the homeless. I do not believe sheltering them is the right thing to do. <mask><mask> giving them proper psychological counselling and a job application is more important. [NEWLINE] [NEWLINE] Am I wrong in thinking this way?  Does anyone here personally know some homeless folks and can attest to their reasons for being homeless?  I don't want just any old schmuck trying to change my view here - you don't know their circumstances any better than I do. [NEWLINE] [NEWLINE] CMV.  Or don't CMV.  Maybe people agree with me?  I guess I'm not sure<mask> I should feel about this. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Idk.  I just feel like (in my town at least) I see all sorts of "homeless" people who stand on the street corner begging for money.  On the strip of road that they beg on there are at least 35 businesses that hire all forms of work.  Sure, they're minimum wage, but with the twelve hours a day you spend standing outside begging for money why don't you get a fucking job and find yourself some housing?  Also, they all drink and smoke.  Not sure if that's the case everywhere, but they literally don't seem to give a fuck. [NEWLINE] [NEWLINE] My view:  Homeless people are homeless by choice - sure, they may have had an incident that forces them out of their home, but then they keep themselves homeless by giving up and falling into despair. [NEWLINE] [NEWLINE] I do not give money to the homeless. I do not believe sheltering them is the right thing to do.  I think giving them proper psychological counselling and a job application is more important. [NEWLINE] [NEWLINE] Am I wrong in thinking this way?  Does anyone here personally know some homeless folks and can attest to their reasons for being homeless?  I don't want just any old schmuck trying to change my view here - you don't know their circumstances any better than I do. [NEWLINE] [NEWLINE] CMV.  Or don't CMV.  Maybe people agree with me?  I guess I'm not sure how I should feel about this. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your post is kind of all over the place in terms of<mask> view you are holding. You've cited several different views. In the interest of a shorter post I will simply focus on your title. [NEWLINE] [NEWLINE] [STARTQ] CMV:There is an ethical limit to the amount of money a person should receive<mask> direct compensation for their work and this number is significantly lower than<mask> executives and athletes are currently paid. [ENDQ] [NEWLINE] I'm going to try and reword your view just to simplify it and make it crystal clear.<mask> I do this incorrectly or oversimplify it to the point that I have warped your view, please let me know. [NEWLINE] [NEWLINE] [STARTQ] CMV: There exists an ethical limit to<mask> much anyone should make. [ENDQ] [NEWLINE] You mention a couple times that<mask> hard someone works is not proportional to<mask> the market value's that work. [NEWLINE] [NEWLINE] [STARTQ] I'm not convinced that the athletes who make 10s of millions work or train 10 times harder than a teammate who makes 250k/yr. Or that the artist who manages to sell his painting for 1M worked a lot harder than the artist who sells it for $1,000 [ENDQ] [NEWLINE] Lets assume you are right in this point.<mask> I take away from it is that certain people are more efficient with their work. An athlete is able to train himself to a higher level than another during the same time period or an artist is able to master his technique to a higher level than other. These people are being more efficient, and<mask> producing greater value for the same effort- and the market is rewarding them<mask> such. A similar can be made in the corporate world.  An executive has better learned<mask> the business and market operates to a higher level than another employee, and<mask> has more value to the market. [NEWLINE] [NEWLINE] This really brings us to<mask><mask><mask> your CMV really focuses on. That is this. [NEWLINE] [NEWLINE] [STARTQ] CMV: Salary should be proportional to effort invested and not value produced. [ENDQ] [NEWLINE] I end my point here. I end it with the hope that you will accept<mask> I have reworded your original post, and that my final restatement of your view is enough to make you realize that your original view itself can be seen<mask> unethical. Maybe that will be enough for you to change your view. [NEWLINE] [NEWLINE] Cheers.</s>
Label encoding: <s>Your post is kind of all over the place in terms of what view you are holding. You've cited several different views. In the interest of a shorter post I will simply focus on your title. [NEWLINE] [NEWLINE] [STARTQ] CMV:There is an ethical limit to the amount of money a person should receive as direct compensation for their work and this number is significantly lower than what executives and athletes are currently paid. [ENDQ] [NEWLINE] I'm going to try and reword your view just to simplify it and make it crystal clear. If I do this incorrectly or oversimplify it to the point that I have warped your view, please let me know. [NEWLINE] [NEWLINE] [STARTQ] CMV: There exists an ethical limit to how much anyone should make. [ENDQ] [NEWLINE] You mention a couple times that how hard someone works is not proportional to how the market value's that work. [NEWLINE] [NEWLINE] [STARTQ] I'm not convinced that the athletes who make 10s of millions work or train 10 times harder than a teammate who makes 250k/yr. Or that the artist who manages to sell his painting for 1M worked a lot harder than the artist who sells it for $1,000 [ENDQ] [NEWLINE] Lets assume you are right in this point. What I take away from it is that certain people are more efficient with their work. An athlete is able to train himself to a higher level than another during the same time period or an artist is able to master his technique to a higher level than other. These people are being more efficient, and thus producing greater value for the same effort- and the market is rewarding them as such. A similar can be made in the corporate world.  An executive has better learned how the business and market operates to a higher level than another employee, and thus has more value to the market. [NEWLINE] [NEWLINE] This really brings us to what I think your CMV really focuses on. That is this. [NEWLINE] [NEWLINE] [STARTQ] CMV: Salary should be proportional to effort invested and not value produced. [ENDQ] [NEWLINE] I end my point here. I end it with the hope that you will accept how I have reworded your original post, and that my final restatement of your view is enough to make you realize that your original view itself can be seen as unethical. Maybe that will be enough for you to change your view. [NEWLINE] [NEWLINE] Cheers.</s>
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you immediately refute your point of "most guns used in crime are bought legally" [ENDQ] [NEWLINE]............ No..? [NEWLINE] [NEWLINE] First off, that's not<mask> I said. I said "the vast majority of guns were legally purchased initially".<mask> you are going to use quotes, actually quote. Don't just make something up and attribute it to someone else. [NEWLINE] [NEWLINE] Second, ***initially*** is the key word here. Most guns come from a dealer legally. Very rarely are guns smuggled into the US. [NEWLINE] [NEWLINE] [STARTQ] even in the US,<mask> you have ANY criminal record, you will show up even in the broadest of background checks, in most states. [ENDQ] [NEWLINE] Ya,<mask> that doesn't stop people without a criminal record buying the gun for someone with one. [NEWLINE] [NEWLINE] [STARTQ] And even people who sell privately through gun shows or craigslist can often pick out shady individuals. [ENDQ] [NEWLINE] Sure. And many of them don't care. [NEWLINE] [NEWLINE] Here's a good account of<mask> guns get into the hands of criminals: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Notice, that eliminating legal gun sales would eliminate nearly all of these sources. [NEWLINE] [NEWLINE] <mask> fun fact from that source: [NEWLINE] [NEWLINE] [STARTQ] ATF officials say that only about 8% of the nation's 124,000 retail gun dealers sell the majority of handguns that are used in crimes. [ENDQ] [NEWLINE] # [NEWLINE] [NEWLINE] [STARTQ] "without any risk" Well,<mask> you consider the class D felony for possession<mask> you get caught [ENDQ] [NEWLINE] "<mask> you get caught" being the key here.<mask> was the last time you had a car searched<mask> crossing state lines? [NEWLINE] [NEWLINE] [STARTQ] Yeah, and<mask> I drive a car, I'm far more likely to be killed by a car accident than<mask> I don't.<mask> is that gonna make me give up my car? [ENDQ] [NEWLINE] This is a complete false equivalence. Owning a gun increases your chances of getting killed *by someone else.* [NEWLINE] [NEWLINE] [STARTQ] I'd like a source for that claim of "far more likely" to be killed by a gun<mask> you own one. [ENDQ] [NEWLINE]..<mask> I "own" one? Uh ok.. [NEWLINE] [NEWLINE] * [URL].aspx?articleid=1814426 [NEWLINE] [NEWLINE] * [URL].refs (paywall) [NEWLINE] [NEWLINE] * [URL] [NEWLINE] [NEWLINE] * [URL] / [NEWLINE] </s>
Label encoding: <s> [STARTQ] So you immediately refute your point of "most guns used in crime are bought legally" [ENDQ] [NEWLINE]............ No..? [NEWLINE] [NEWLINE] First off, that's not what I said. I said "the vast majority of guns were legally purchased initially". If you are going to use quotes, actually quote. Don't just make something up and attribute it to someone else. [NEWLINE] [NEWLINE] Second, ***initially*** is the key word here. Most guns come from a dealer legally. Very rarely are guns smuggled into the US. [NEWLINE] [NEWLINE] [STARTQ] even in the US, if you have ANY criminal record, you will show up even in the broadest of background checks, in most states. [ENDQ] [NEWLINE] Ya, but that doesn't stop people without a criminal record buying the gun for someone with one. [NEWLINE] [NEWLINE] [STARTQ] And even people who sell privately through gun shows or craigslist can often pick out shady individuals. [ENDQ] [NEWLINE] Sure. And many of them don't care. [NEWLINE] [NEWLINE] Here's a good account of how guns get into the hands of criminals: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Notice, that eliminating legal gun sales would eliminate nearly all of these sources. [NEWLINE] [NEWLINE] Also fun fact from that source: [NEWLINE] [NEWLINE] [STARTQ] ATF officials say that only about 8% of the nation's 124,000 retail gun dealers sell the majority of handguns that are used in crimes. [ENDQ] [NEWLINE] # [NEWLINE] [NEWLINE] [STARTQ] "without any risk" Well, if you consider the class D felony for possession if you get caught [ENDQ] [NEWLINE] " If you get caught" being the key here. When was the last time you had a car searched when crossing state lines? [NEWLINE] [NEWLINE] [STARTQ] Yeah, and if I drive a car, I'm far more likely to be killed by a car accident than if I don't. But is that gonna make me give up my car? [ENDQ] [NEWLINE] This is a complete false equivalence. Owning a gun increases your chances of getting killed *by someone else.* [NEWLINE] [NEWLINE] [STARTQ] I'd like a source for that claim of "far more likely" to be killed by a gun if you own one. [ENDQ] [NEWLINE].. if I "own" one? Uh ok.. [NEWLINE] [NEWLINE] * [URL].aspx?articleid=1814426 [NEWLINE] [NEWLINE] * [URL].refs (paywall) [NEWLINE] [NEWLINE] * [URL] [NEWLINE] [NEWLINE] * [URL] / [NEWLINE] </s>
Loss: tensor(0.1266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People shouldn't be applauded for not being douchebags. Of course I've had good interactions,<mask> that's<mask> should be expected of people; to be good. Or at least neutral. [NEWLINE] [NEWLINE] [STARTQ] ust<mask> i haven't experienced judgement on those specific things, doesn't mean i haven't taken the time to learn about<mask> that kind of harassment can feel, and can't apply it to scenarios i have experienced. [ENDQ] [NEWLINE] I totally agree with you here. The key here,<mask>, is that<mask><mask> you might have an objective understanding of<mask> it is like to be part of X group, you won't *really* be able to truly *empathize* (different than sympathize) with the struggles of X group. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask> empathy is a great place to start. the people doing the harassing very often have problems of their own and belittling others is their (naive and ignorant) way of coping with it. [ENDQ] [NEWLINE] Yeah, and<mask><mask> that to some degree having a conversation about<mask> one particular co-worker in general makes really terrible rape jokes might benefit everyone in getting to the root of the problem.<mask> I'm not going to take the time out of my day to sit down with every asshole who yells obscenities at people and ask them<mask> went wrong in their childhood. That is something that needs to be addressed on a larger scale.<mask> my boss is being an asshole, I give him the benefit of the doubt and assume that maybe he is just having a bad day. Strangers don't get that benefit. [NEWLINE] [NEWLINE] <mask><mask> that tip-toeing around certain words and over-labeling every tiny demographic might be more trouble than it is worth. This seems divisive to me.<mask>, I don't think the solution to racism/sexism/transphobia/homophobia/xenophobia/whatever-else-phobia should rest on the shoulders of the people being tormented, I don't think the victims in these situations should have to change *their* outlook on life,<mask> the people being victimized seem to rarely be harassers themselves. This almost feels like you're arguing that people who are bullies,<mask> you don't condone it, don't need to change<mask> much<mask> the people who are being bullied. </s>
Label encoding: <s>People shouldn't be applauded for not being douchebags. Of course I've had good interactions, but that's what should be expected of people; to be good. Or at least neutral. [NEWLINE] [NEWLINE] [STARTQ] ust because i haven't experienced judgement on those specific things, doesn't mean i haven't taken the time to learn about how that kind of harassment can feel, and can't apply it to scenarios i have experienced. [ENDQ] [NEWLINE] I totally agree with you here. The key here, though, is that even though you might have an objective understanding of what it is like to be part of X group, you won't *really* be able to truly *empathize* (different than sympathize) with the struggles of X group. [NEWLINE] [NEWLINE] [STARTQ] but i think empathy is a great place to start. the people doing the harassing very often have problems of their own and belittling others is their (naive and ignorant) way of coping with it. [ENDQ] [NEWLINE] Yeah, and I think that to some degree having a conversation about why one particular co-worker in general makes really terrible rape jokes might benefit everyone in getting to the root of the problem. But I'm not going to take the time out of my day to sit down with every asshole who yells obscenities at people and ask them what went wrong in their childhood. That is something that needs to be addressed on a larger scale. When my boss is being an asshole, I give him the benefit of the doubt and assume that maybe he is just having a bad day. Strangers don't get that benefit. [NEWLINE] [NEWLINE] I think that tip-toeing around certain words and over-labeling every tiny demographic might be more trouble than it is worth. This seems divisive to me. However, I don't think the solution to racism/sexism/transphobia/homophobia/xenophobia/whatever-else-phobia should rest on the shoulders of the people being tormented, I don't think the victims in these situations should have to change *their* outlook on life, since the people being victimized seem to rarely be harassers themselves. This almost feels like you're arguing that people who are bullies, while you don't condone it, don't need to change as much as the people who are being bullied. </s>
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's immoral to take a product from someone--that they put time and effort into--<mask> they want you to pay for it, just<mask> you don't agree to their terms. The options are not solely "pay for it or pirate it", they're actually "pay for it, pirate it, or choose not to consume it. Boycotting is more moral<mask> it implies that you disapprove of their actions<mask> strongly, you won't even consume the content even<mask> you might want to. [NEWLINE] [NEWLINE] <mask> you want to<mask><mask> you don't think it matters<mask> you can easily dissociate yourself from<mask> it means and represents to the content creator, that's fine, you gotta do<mask> you gotta do.<mask> that's not really any sort of reasoning to support it being moral. [NEWLINE] [NEWLINE] <mask> your argument is that media companies' businesses or their current model is immoral (or, more accurately, unreasonable or less than ideal) then that's all well and good and a very understandable position to take.<mask> you use that platform to<mask><mask> piracy isn't great,<mask> just an understandable response to an archaic business model, I can understand that reasoning. And I actually agree that media companies are behind the times and need to adapt to the way consumers want to access content. [NEWLINE] [NEWLINE] <mask> that doesn't have to be a cover for all instances of piracy.<mask> the old media model is eventually demolished some day, and it's replaced by an ideal system<mask> content is easily accessible and it's easy to fairly and reasonable compensate creators across the entire entertainment spectrum, I really hope you wouldn't still defend it<mask> moral for people to choose not to compensate them. [NEWLINE] [NEWLINE] These are sort of tangential,<mask><mask> Neil Gaiman is entitled to feel<mask> he chooses about his efforts, his feelings don't have to apply to or be held by other creators. HBO is not *in favor* of piracy. They're taking a pragmatic approach of not making an "angry old man" style rant about it, and instead making a calculated PR move to garner goodwill<mask> they know people will respond positively to it. Notice that nowhere in the article do you see a quote from HBO saying, "Piracy is<mask> awesome, we actually wish fewer people paid for our content and more people pirated it!"</s>
Label encoding: <s>It's immoral to take a product from someone--that they put time and effort into-- if they want you to pay for it, just because you don't agree to their terms. The options are not solely "pay for it or pirate it", they're actually "pay for it, pirate it, or choose not to consume it. Boycotting is more moral because it implies that you disapprove of their actions so strongly, you won't even consume the content even if you might want to. [NEWLINE] [NEWLINE] If you want to argue that you don't think it matters because you can easily dissociate yourself from what it means and represents to the content creator, that's fine, you gotta do what you gotta do. But that's not really any sort of reasoning to support it being moral. [NEWLINE] [NEWLINE] If your argument is that media companies' businesses or their current model is immoral (or, more accurately, unreasonable or less than ideal) then that's all well and good and a very understandable position to take. If you use that platform to argue that piracy isn't great, but just an understandable response to an archaic business model, I can understand that reasoning. And I actually agree that media companies are behind the times and need to adapt to the way consumers want to access content. [NEWLINE] [NEWLINE] But that doesn't have to be a cover for all instances of piracy. If the old media model is eventually demolished some day, and it's replaced by an ideal system where content is easily accessible and it's easy to fairly and reasonable compensate creators across the entire entertainment spectrum, I really hope you wouldn't still defend it as moral for people to choose not to compensate them. [NEWLINE] [NEWLINE] These are sort of tangential, but though Neil Gaiman is entitled to feel however he chooses about his efforts, his feelings don't have to apply to or be held by other creators. HBO is not *in favor* of piracy. They're taking a pragmatic approach of not making an "angry old man" style rant about it, and instead making a calculated PR move to garner goodwill because they know people will respond positively to it. Notice that nowhere in the article do you see a quote from HBO saying, "Piracy is so awesome, we actually wish fewer people paid for our content and more people pirated it!"</s>
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Europe &amp; Africa aren't countries they are continents. That's the point he/she is making.<mask> an American is of European heritage<mask> they don't know whereabouts in Europe,<mask> are they not referred to<mask> European-American? [NEWLINE] [NEWLINE] <mask> it's a huge generalisation. People from Ireland are not the same<mask> people from Italy. People from Morocco are not the same<mask> people from Uganda. [NEWLINE] [NEWLINE] It's kind of offensive to the people who are actually from these countries or continents to say that you're from there too<mask> likely neither you nor the last 10 generations of your family have ever stepped foot there. [NEWLINE] [NEWLINE] Finally, the whole Irish/Italian/European/Moroccan/Ugandan/African-American thing is silly<mask> very rarely are the people who identify in this way actually from these places. Most of the people who say "I'm Irish-American" (Usually leaving out the American part) were born and brought up in the US by people who were born and brought up in the US, you have to go back to the 19th century or later to find an ancestor of their's who was actually Irish. [NEWLINE] [NEWLINE] For 'African-Americans' it's even more<mask>,<mask> to find an ancestor who was actually from Africa you'd have to go back at least 200 years, most likely. [NEWLINE] [NEWLINE] <mask> someone were from Uganda, for example, and they moved to the US, lived there for the rest of their life &amp; became a naturalised citizen. They would be able to call themselves an African-American or better<mask> a Ugandan-American. It makes no sense for a person to identify with a country that they have never been too, simply<mask> they know that hundreds of years ago their ancestors lived there. [NEWLINE] [NEWLINE] <mask> everyone did that you'd have people saying "Hi, I'm Spanish-French", Hi "I'm Irish-Welsh", "Hi I'm Norwegian-Dutch". With people saying "<mask> makes you Spanish/Irish/Norwegian?" "Oh my Great-Great-Great-Great-Great-Great-Great-Great-Great-Great-Great-Grandfather was from there, no one else, just him &amp;<mask> that makes me part Spanish/Irish/Norwegian."</s>
Label encoding: <s>Europe &amp; Africa aren't countries they are continents. That's the point he/she is making. If an American is of European heritage but they don't know whereabouts in Europe, why are they not referred to as European-American? [NEWLINE] [NEWLINE] Because it's a huge generalisation. People from Ireland are not the same as people from Italy. People from Morocco are not the same as people from Uganda. [NEWLINE] [NEWLINE] It's kind of offensive to the people who are actually from these countries or continents to say that you're from there too when likely neither you nor the last 10 generations of your family have ever stepped foot there. [NEWLINE] [NEWLINE] Finally, the whole Irish/Italian/European/Moroccan/Ugandan/African-American thing is silly when very rarely are the people who identify in this way actually from these places. Most of the people who say "I'm Irish-American" (Usually leaving out the American part) were born and brought up in the US by people who were born and brought up in the US, you have to go back to the 19th century or later to find an ancestor of their's who was actually Irish. [NEWLINE] [NEWLINE] For 'African-Americans' it's even more so, as to find an ancestor who was actually from Africa you'd have to go back at least 200 years, most likely. [NEWLINE] [NEWLINE] If someone were from Uganda, for example, and they moved to the US, lived there for the rest of their life &amp; became a naturalised citizen. They would be able to call themselves an African-American or better yet a Ugandan-American. It makes no sense for a person to identify with a country that they have never been too, simply because they know that hundreds of years ago their ancestors lived there. [NEWLINE] [NEWLINE] If everyone did that you'd have people saying "Hi, I'm Spanish-French", Hi "I'm Irish-Welsh", "Hi I'm Norwegian-Dutch". With people saying " What makes you Spanish/Irish/Norwegian?" "Oh my Great-Great-Great-Great-Great-Great-Great-Great-Great-Great-Great-Grandfather was from there, no one else, just him &amp; so that makes me part Spanish/Irish/Norwegian."</s>
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Just the other week a polce officer, in his car, asked a man who was already outside of his car for his liscense. The dash cam shows him pat his pockets and then reach into his car for his wallet. The officer fired four shots at him with a gas station not several yards behind him. I understand<mask> you are saying here,<mask>'sudden movements' can be whatever the officer wants it to be. [ENDQ] [NEWLINE] That happened in my home town of Columbia, SC.  The officer in question was fired and is facing charges.  The first time I watched that video, I saw the guy make a sudden and jerky move towards his vehicle and thought, "Shit, I'm siding with the officer on this one." [NEWLINE] [NEWLINE] Then the asshole started shooting--I went<mask> far<mask> to download the video and mark out a timeline for<mask> the shots were fired.  Former SC Trooper Groubert fired his first shot *before* the victim had even completely turned around.  There was no way he could have positively identified a weapon.  He then proceeded to fire 3 more panic shots in a crowded place, long after it was apparent the victim did not have a weapon.  The officer was in the wrong--not just wrong,<mask> VERY fucking wrong. <mask>, the state of South Carolina fired him and is taking appropriate measures against him. [NEWLINE] [NEWLINE] With that said, it does not change the fact that the victim gave the officer reason to be alarmed.  Sudden movements<mask> you don't inform the officer can result in bad things.  Should the officer have shot?  Absolutely not. <mask> he absolutely did have cause to elevate his security posture. [NEWLINE] [NEWLINE] [STARTQ] Not to be rude,<mask> are you American? Police usually don't even ask to see registration unless they suspect the car of being stolen. They usually want to see proof of insurance. The 'liscense and registration' thing is only in movies for routine stops. [ENDQ] [NEWLINE] Yes.  I have been pulled over by police maybe a dozen times in<mask> many years, and *every single one of those times,* the officer asked for my license AND registration.  Maybe police do things differently<mask> you're from<mask><mask><mask> of a differing state law.</s>
Label encoding: <s> [STARTQ] Just the other week a polce officer, in his car, asked a man who was already outside of his car for his liscense. The dash cam shows him pat his pockets and then reach into his car for his wallet. The officer fired four shots at him with a gas station not several yards behind him. I understand what you are saying here, but'sudden movements' can be whatever the officer wants it to be. [ENDQ] [NEWLINE] That happened in my home town of Columbia, SC.  The officer in question was fired and is facing charges.  The first time I watched that video, I saw the guy make a sudden and jerky move towards his vehicle and thought, "Shit, I'm siding with the officer on this one." [NEWLINE] [NEWLINE] Then the asshole started shooting--I went so far as to download the video and mark out a timeline for when the shots were fired.  Former SC Trooper Groubert fired his first shot *before* the victim had even completely turned around.  There was no way he could have positively identified a weapon.  He then proceeded to fire 3 more panic shots in a crowded place, long after it was apparent the victim did not have a weapon.  The officer was in the wrong--not just wrong, but VERY fucking wrong.  However, the state of South Carolina fired him and is taking appropriate measures against him. [NEWLINE] [NEWLINE] With that said, it does not change the fact that the victim gave the officer reason to be alarmed.  Sudden movements where you don't inform the officer can result in bad things.  Should the officer have shot?  Absolutely not.  But he absolutely did have cause to elevate his security posture. [NEWLINE] [NEWLINE] [STARTQ] Not to be rude, but are you American? Police usually don't even ask to see registration unless they suspect the car of being stolen. They usually want to see proof of insurance. The 'liscense and registration' thing is only in movies for routine stops. [ENDQ] [NEWLINE] Yes.  I have been pulled over by police maybe a dozen times in as many years, and *every single one of those times,* the officer asked for my license AND registration.  Maybe police do things differently where you're from as a result of a differing state law.</s>
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well the thing is<mask> I was into it, it wasn't anti-women at all really at least thats not<mask> I took from it. The PUA community<mask> I was into it was really divided. You had scummy people and more natural people, and I mostly prescribed to "juggler"'s style of social interaction which was centered around just being able to talk to people and connect with them. A lot of people that go for it(maybe most?) are actually just looking for a steady girlfriend not a bunch of one night stands like people like to think of them<mask><mask> a prevailing belief is essentially<mask> you can get laid you can get a GF, which is kind of true (maintaining a good relationship is a whole nother beast<mask> ). [NEWLINE] [NEWLINE] <mask> there is just a huge difference between some general self help stuff like fitness, and something that helps DIRECTLY with social interaction with the end goal finding a romantic partner or partners.<mask> the men that go to these communities aren't some average guy looking for tips, they are usually people HUGELY behind the dating game with no idea<mask> to do, no confidence, and with a lot of fears about socializing. Having a super obsessed and aggressive community is the difference between a person actually finding the courage within themselves to talk to strangers, and them chickening out. Talking to strangers is extremely terrifying<mask> I'm sure you know and you kind of need the extremes to get yourself to do it regularly especially<mask> you are coming from a place of almost no socializing. [NEWLINE] [NEWLINE] <mask>, I don't want this to be offensive<mask>, it is much easier for girls in the dating world than guys usually. Like you said your bf helped you become more confident etc. For guys its the opposite, you can't get a gf unless you are ALREADY great, confident, or have something going for you. Not to mention at least at first guys have to put in a lot more effort at least typically, there are plenty of aggressive girls<mask> typically its the guys that are expected to brave the tundra of starting conversations and escalating. Its harder for us and<mask> it makes sense some guys would need extreme communities to go from social anxiety to constantly meeting strangers and being able to lead conversations. </s>
Label encoding: <s>Well the thing is when I was into it, it wasn't anti-women at all really at least thats not what I took from it. The PUA community when I was into it was really divided. You had scummy people and more natural people, and I mostly prescribed to "juggler"'s style of social interaction which was centered around just being able to talk to people and connect with them. A lot of people that go for it(maybe most?) are actually just looking for a steady girlfriend not a bunch of one night stands like people like to think of them as although a prevailing belief is essentially if you can get laid you can get a GF, which is kind of true (maintaining a good relationship is a whole nother beast though ). [NEWLINE] [NEWLINE] Also there is just a huge difference between some general self help stuff like fitness, and something that helps DIRECTLY with social interaction with the end goal finding a romantic partner or partners. Also the men that go to these communities aren't some average guy looking for tips, they are usually people HUGELY behind the dating game with no idea what to do, no confidence, and with a lot of fears about socializing. Having a super obsessed and aggressive community is the difference between a person actually finding the courage within themselves to talk to strangers, and them chickening out. Talking to strangers is extremely terrifying as I'm sure you know and you kind of need the extremes to get yourself to do it regularly especially when you are coming from a place of almost no socializing. [NEWLINE] [NEWLINE] Also, I don't want this to be offensive but, it is much easier for girls in the dating world than guys usually. Like you said your bf helped you become more confident etc. For guys its the opposite, you can't get a gf unless you are ALREADY great, confident, or have something going for you. Not to mention at least at first guys have to put in a lot more effort at least typically, there are plenty of aggressive girls but typically its the guys that are expected to brave the tundra of starting conversations and escalating. Its harder for us and so it makes sense some guys would need extreme communities to go from social anxiety to constantly meeting strangers and being able to lead conversations. </s>
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0941, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0988, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[Japanese vending machines are actually pretty simple.]( [URL].5.jpg) [NEWLINE] [NEWLINE] They all have an array of beverages with plastic molds with an imitation of the product itself. This is surprisingly universal. It's actually pretty cool,<mask> this shows you the size and shape of the product you could purchase,<mask> value propositions are easy. Under each imitation there is a price, and an indication of whether the beverage is served hot or cold, most are color coded, or cold<mask> no indication. (This is about the only thing that is not obvious to a westerner) [NEWLINE] [NEWLINE] You put in your money, most Japanese vending machines are cash only. This is a bit of a strange thing about Japan, most people assume that they're leaps and bounds ahead of the west in many ways,<mask> this is one thing I found that defied my assumptions. The cash culture runs deep, and this has been slow to change in Japan,<mask><mask> some have adopted the NFC payments like the one above, not all have. [NEWLINE] [NEWLINE] then<mask> you put in the money, the beverages you can afford with<mask> you have put in will have their lights light up on the buttons. press the button and boom! A satisfying mechanical click then a satisfying drink. [NEWLINE] [NEWLINE] <mask> for all the writing in the picture above, pretty much all of it save the temp indicators and the instructions above the NFC reader are ads. And the products themselves? Look and you'll see that most have english on them, even the Japanese brands. English has sex appeal there and advertisers know it,<mask> this is<mask> a lot of Japanese brands have names that don't sound too Japanesey. [NEWLINE] [NEWLINE] The machines<mask> tend to group up a lot which is something I'm not used to coming from america. Imagine<mask> whenever you saw a group of vending machines, you would usually find Coka-Cola *and* Pepsi. This was good for me<mask> once you find something you like, it usually means you'll find it in a lot of places. I grew pretty fond of [Boss Coffee]( [URL].jpg), mostly<mask> I found their celebrity endorsement very funny,<mask> it did taste excellent. [NEWLINE] [NEWLINE] TL;DR More than you ever wanted to know about the simplicity of Japanese vending machines.</s>
Label encoding: <s>[Japanese vending machines are actually pretty simple.]( [URL].5.jpg) [NEWLINE] [NEWLINE] They all have an array of beverages with plastic molds with an imitation of the product itself. This is surprisingly universal. It's actually pretty cool, because this shows you the size and shape of the product you could purchase, so value propositions are easy. Under each imitation there is a price, and an indication of whether the beverage is served hot or cold, most are color coded, or cold if no indication. (This is about the only thing that is not obvious to a westerner) [NEWLINE] [NEWLINE] You put in your money, most Japanese vending machines are cash only. This is a bit of a strange thing about Japan, most people assume that they're leaps and bounds ahead of the west in many ways, but this is one thing I found that defied my assumptions. The cash culture runs deep, and this has been slow to change in Japan, so while some have adopted the NFC payments like the one above, not all have. [NEWLINE] [NEWLINE] then as you put in the money, the beverages you can afford with what you have put in will have their lights light up on the buttons. press the button and boom! A satisfying mechanical click then a satisfying drink. [NEWLINE] [NEWLINE] As for all the writing in the picture above, pretty much all of it save the temp indicators and the instructions above the NFC reader are ads. And the products themselves? Look and you'll see that most have english on them, even the Japanese brands. English has sex appeal there and advertisers know it, so this is why a lot of Japanese brands have names that don't sound too Japanesey. [NEWLINE] [NEWLINE] The machines also tend to group up a lot which is something I'm not used to coming from america. Imagine if whenever you saw a group of vending machines, you would usually find Coka-Cola *and* Pepsi. This was good for me since once you find something you like, it usually means you'll find it in a lot of places. I grew pretty fond of [Boss Coffee]( [URL].jpg), mostly because I found their celebrity endorsement very funny, but it did taste excellent. [NEWLINE] [NEWLINE] TL;DR More than you ever wanted to know about the simplicity of Japanese vending machines.</s>
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0949, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] And<mask> the end results of hyperrealistic art offers nothing beyond<mask> a photograph could offer - other than inflating the ego of the artist - it is truly pointless. [ENDQ] [NEWLINE] The best hyperrealist artists aim to offer something that is better and more perfect than<mask> a straight photograph can offer. Many of them even mix references to create scenes that literally do not exist and cannot<mask> be photographed.<mask> they use photographic references, they rely on their own experience in painting or drawing from life in order to enhance the photorealistic experience of a subject. [NEWLINE] [NEWLINE] For instance, there is a pretty big difference between [this drawing]( [URL].jpg) made with six different colors of Bic ballpoint pens and the [original reference image]( [URL].jpg). Overlaid in Photoshop, you can see that the artist made some changes to the facial structure of the model, took great artistic license with the colors in the original photograph, and used various techniques to simplify unimportant details. This particular piece isn't truly hyperrealistic due to the limitations inherent to using only six colors of Bic ballpoint pen,<mask> it is extremely realistic.<mask> you couldn't take *that* photograph<mask> it is depicted in this piece,<mask> that girl doesn't exist. She has been altered to suit the artist's vision. [NEWLINE] [NEWLINE] Unfortunately, not too many truly good hypperealists will show references on their artist pages,<mask> it's not easy for people to see<mask> the artist transforms the source material. [NEWLINE] [NEWLINE] In any event, for artists much of the work is about the process. You can take a decent portrait in about an hour or two, including lighting, makeup, and other setup details. A person who spends 30-50 hours painting or drawing pores, freckles, the galaxy of colors and shapes in a person's iris, etc., has many more opportunities to make the piece more personal, emotional, and become more intimate with their subject matter. Not on a personal level, necessarily,<mask> from the standpoint of *really seeing* something. The experience of such intense study can enhance non-photorealistic/hyperrealistic endeavors, too.<mask> it's not necessarily pointless or stupid or an exercise in ego-stroking.</s>
Label encoding: <s> [STARTQ] And since the end results of hyperrealistic art offers nothing beyond what a photograph could offer - other than inflating the ego of the artist - it is truly pointless. [ENDQ] [NEWLINE] The best hyperrealist artists aim to offer something that is better and more perfect than what a straight photograph can offer. Many of them even mix references to create scenes that literally do not exist and cannot therefore be photographed. Although they use photographic references, they rely on their own experience in painting or drawing from life in order to enhance the photorealistic experience of a subject. [NEWLINE] [NEWLINE] For instance, there is a pretty big difference between [this drawing]( [URL].jpg) made with six different colors of Bic ballpoint pens and the [original reference image]( [URL].jpg). Overlaid in Photoshop, you can see that the artist made some changes to the facial structure of the model, took great artistic license with the colors in the original photograph, and used various techniques to simplify unimportant details. This particular piece isn't truly hyperrealistic due to the limitations inherent to using only six colors of Bic ballpoint pen, but it is extremely realistic. But you couldn't take *that* photograph as it is depicted in this piece, because that girl doesn't exist. She has been altered to suit the artist's vision. [NEWLINE] [NEWLINE] Unfortunately, not too many truly good hypperealists will show references on their artist pages, so it's not easy for people to see how the artist transforms the source material. [NEWLINE] [NEWLINE] In any event, for artists much of the work is about the process. You can take a decent portrait in about an hour or two, including lighting, makeup, and other setup details. A person who spends 30-50 hours painting or drawing pores, freckles, the galaxy of colors and shapes in a person's iris, etc., has many more opportunities to make the piece more personal, emotional, and become more intimate with their subject matter. Not on a personal level, necessarily, but from the standpoint of *really seeing* something. The experience of such intense study can enhance non-photorealistic/hyperrealistic endeavors, too. So it's not necessarily pointless or stupid or an exercise in ego-stroking.</s>
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1700, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Batman is an odd example,<mask>. He walks a very fine line. Warning: Spoilers for Arkham City at the end of this... [NEWLINE] [NEWLINE] He *does* hurt people. He hurts people all the time. He leaves them beaten, bloody, and broken, tied up, outside, in the snow, slowly freezing to death,<mask> he can live with himself<mask> he didn't *kill* them, and<mask> they die of hypothermia before the police find them, that's somehow not his fault. [NEWLINE] [NEWLINE] And he has an image to uphold<mask> a hero,<mask> just<mask> important is his image of fear. "I am vengeance. I am the night." He strikes fear into the hearts of criminals, and not<mask> he's a hero. Superman could squish you like a bug,<mask> he mostly stands for hope, for "Truth, Justice, and the American Way," and no one really seems to be *afraid* of him. They take him seriously,<mask> they don't *cower* at his approach. Whereas in Gotham, your average street thug is jumping at shadows -- even<mask> you really know The Bat won't kill you, he could still have you eating through a straw for the rest of your life. [NEWLINE] [NEWLINE] And, especially in Batman's case, it's not at all a PR decision. Again, spoilers...<mask> Batman walks out carrying The Joker's body, he doesn't want to talk about it. He's perfectly happy letting the city believe that he finally did it. And, really, Harley Quinn is the only one sad to see him go.<mask>, Batman doesn't believe it's a PR issue, and no one in Gotham (except Harley) sees it<mask> a PR problem. [NEWLINE] [NEWLINE] With some heroes, like Superman, maybe it makes sense not to use lethal force, and maybe it makes sense that it's a PR campaign.<mask> I don't think that's true in Batman's case. The reason Batman doesn't kill The Joker, especially knowing he'll just get out again and kill more people, is entirely to do with his own issues, his near-psychotic refusal to ever take a human life, or use a gun even non-lethally.</s>
Label encoding: <s>Batman is an odd example, though. He walks a very fine line. Warning: Spoilers for Arkham City at the end of this... [NEWLINE] [NEWLINE] He *does* hurt people. He hurts people all the time. He leaves them beaten, bloody, and broken, tied up, outside, in the snow, slowly freezing to death, but he can live with himself because he didn't *kill* them, and if they die of hypothermia before the police find them, that's somehow not his fault. [NEWLINE] [NEWLINE] And he has an image to uphold as a hero, but just as important is his image of fear. "I am vengeance. I am the night." He strikes fear into the hearts of criminals, and not because he's a hero. Superman could squish you like a bug, but he mostly stands for hope, for "Truth, Justice, and the American Way," and no one really seems to be *afraid* of him. They take him seriously, but they don't *cower* at his approach. Whereas in Gotham, your average street thug is jumping at shadows -- even if you really know The Bat won't kill you, he could still have you eating through a straw for the rest of your life. [NEWLINE] [NEWLINE] And, especially in Batman's case, it's not at all a PR decision. Again, spoilers... When Batman walks out carrying The Joker's body, he doesn't want to talk about it. He's perfectly happy letting the city believe that he finally did it. And, really, Harley Quinn is the only one sad to see him go. So, Batman doesn't believe it's a PR issue, and no one in Gotham (except Harley) sees it as a PR problem. [NEWLINE] [NEWLINE] With some heroes, like Superman, maybe it makes sense not to use lethal force, and maybe it makes sense that it's a PR campaign. But I don't think that's true in Batman's case. The reason Batman doesn't kill The Joker, especially knowing he'll just get out again and kill more people, is entirely to do with his own issues, his near-psychotic refusal to ever take a human life, or use a gun even non-lethally.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think there are several layers of misconception going on here (and one of the layers is that thought on these matters is shifting, and I'm not certain that anyone has the "right" idea about mental illness at the moment). [NEWLINE] [NEWLINE] Layer 1: a depressed person is still responsible for their actions.<mask> they do something to hurt someone else, it is still their fault.<mask>'s important to recognize is that the depressed person is not depressed by choice: they cannot flip a switch and stop being depressed. They need help digging themselves out, and that help should usually take a professional, medical form. [NEWLINE] [NEWLINE] Perhaps analogy might help:<mask> you're driving your car around with a fuel leak, and do nothing to get it fixed, you're going to be responsible for any accidents you might cause. The principle isn't too different<mask> you're piloting a brain with some medical issues. It's your responsibility to fix the issue,<mask> you probably have to go to a professional to get it fixed. [NEWLINE] [NEWLINE] Layer 2: "lazy" isn't a medical diagnosis,<mask> depression and anxiety are, and<mask><mask> that most people who you might call lazy are suffering from some level of depression or anxiety. Lazy is a symptom; anxiety might be the cause. [NEWLINE] [NEWLINE] Layer 3: The boundary between "clinically diagnosable", and within the bounds of "normal", is a little bit hard to pin down. Everybody exhibits depressive or anxious behaviors sometimes. Diagnosis usually happens<mask> those behaviors begin to seriously interfere in that person's life.<mask><mask><mask> that there are some profound implications of the disease model of mental disorders that we haven't fully worked through<mask> a society --<mask><mask> that we assume that people can just power through some things that they really can't (this relates to obesity, school schedules and teen sleep deprivation, the influence of advertising and shady sales tactics, etc.) [NEWLINE] [NEWLINE] Layer 4: none of this absolves people of personal responsibility for their actions.<mask> it can inform the ways in which we seek to better ourselves, and the ways in which we see each others' "character", and "character flaws".<mask><mask> that we should do less judging of each other, and more offering help and support.</s>
Label encoding: <s>I think there are several layers of misconception going on here (and one of the layers is that thought on these matters is shifting, and I'm not certain that anyone has the "right" idea about mental illness at the moment). [NEWLINE] [NEWLINE] Layer 1: a depressed person is still responsible for their actions. If they do something to hurt someone else, it is still their fault. What's important to recognize is that the depressed person is not depressed by choice: they cannot flip a switch and stop being depressed. They need help digging themselves out, and that help should usually take a professional, medical form. [NEWLINE] [NEWLINE] Perhaps analogy might help: if you're driving your car around with a fuel leak, and do nothing to get it fixed, you're going to be responsible for any accidents you might cause. The principle isn't too different when you're piloting a brain with some medical issues. It's your responsibility to fix the issue, though you probably have to go to a professional to get it fixed. [NEWLINE] [NEWLINE] Layer 2: "lazy" isn't a medical diagnosis, but depression and anxiety are, and I think that most people who you might call lazy are suffering from some level of depression or anxiety. Lazy is a symptom; anxiety might be the cause. [NEWLINE] [NEWLINE] Layer 3: The boundary between "clinically diagnosable", and within the bounds of "normal", is a little bit hard to pin down. Everybody exhibits depressive or anxious behaviors sometimes. Diagnosis usually happens when those behaviors begin to seriously interfere in that person's life. But I think that there are some profound implications of the disease model of mental disorders that we haven't fully worked through as a society -- I think that we assume that people can just power through some things that they really can't (this relates to obesity, school schedules and teen sleep deprivation, the influence of advertising and shady sales tactics, etc.) [NEWLINE] [NEWLINE] Layer 4: none of this absolves people of personal responsibility for their actions. But it can inform the ways in which we seek to better ourselves, and the ways in which we see each others' "character", and "character flaws". I think that we should do less judging of each other, and more offering help and support.</s>
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi everyone!! [NEWLINE] [NEWLINE] Ive had many hours of debates with my fellow colleagues about this matter. I believe that hydro power is a lot better than wind power, for the following reasons: [NEWLINE] [NEWLINE] 1. **Cheaper**.<mask> it may cost a lot to build a dam, its a lot lot more cheaper in the long run with low service costs. [NEWLINE] [NEWLINE] 2. **More efficient**. Hydropower will produce a lot more electricity, for a lot lower price. It can produce electricity all year round, and it produces a lot more electricity too. [NEWLINE] [NEWLINE] 3. **Doesnt have to be shut down**. Wind power must be shut down<mask> there is too much wind. [NEWLINE] [NEWLINE] 4. **Not effected by extreme weather**. Storms can damage wind powers. [NEWLINE] [NEWLINE] 5. **Noisy and threat to wild life**. They are super loud, and can kill birds. Not to mention they are ugly to look at. [NEWLINE] [NEWLINE] 6. **Locational**. Wind power cant be places everywhere, they have to be put in hilly areas or coastline. [NEWLINE] [NEWLINE] 7. **Consistency** Wind doesnt always blow consistently,<mask> water always flows<mask> it should. [NEWLINE] [NEWLINE] <mask> dou you think, /r/changemyview? Am I wrong? Please change my view, i am really looking forward to hear<mask> you say. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Hi everyone!! [NEWLINE] [NEWLINE] Ive had many hours of debates with my fellow colleagues about this matter. I believe that hydro power is a lot better than wind power, for the following reasons: [NEWLINE] [NEWLINE] 1. **Cheaper**. While it may cost a lot to build a dam, its a lot lot more cheaper in the long run with low service costs. [NEWLINE] [NEWLINE] 2. **More efficient**. Hydropower will produce a lot more electricity, for a lot lower price. It can produce electricity all year round, and it produces a lot more electricity too. [NEWLINE] [NEWLINE] 3. **Doesnt have to be shut down**. Wind power must be shut down if there is too much wind. [NEWLINE] [NEWLINE] 4. **Not effected by extreme weather**. Storms can damage wind powers. [NEWLINE] [NEWLINE] 5. **Noisy and threat to wild life**. They are super loud, and can kill birds. Not to mention they are ugly to look at. [NEWLINE] [NEWLINE] 6. **Locational**. Wind power cant be places everywhere, they have to be put in hilly areas or coastline. [NEWLINE] [NEWLINE] 7. **Consistency** Wind doesnt always blow consistently, while water always flows as it should. [NEWLINE] [NEWLINE] What dou you think, /r/changemyview? Am I wrong? Please change my view, i am really looking forward to hear what you say. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I must not be explaining<mask> I'm saying very well. You seem to be ignoring the most important part of my argument: the labor. [NEWLINE] [NEWLINE] The physical matter that composes X is one part of two that composes X. There is matter, M, and labor, L, that went into the creation, or crafting, of X. We could have different variables, that you mention, logistics, travel, et cetera,<mask> we'll put those under M<mask> they're in the physical realm. L is, say, the writing (and the editing by the editor, et cetera) of the text X, or the programming or the designing of the software X, or the formulation for candy bar X. X is composed of M and L, together. M exists by itself without L,<mask> it takes the addition of L to create X. For some really informal logic: [NEWLINE] [NEWLINE] X = M + L. [NEWLINE] [NEWLINE] With your magic copying device, you copy X, and get its exact clone, which we'll call X₁. X₁ is the exact same object<mask> X. X₁ = X.<mask>, this one took zero material, M, to make.<mask> we take away the value of the material. [NEWLINE] [NEWLINE] X₁ = X - M [NEWLINE] [NEWLINE] X₁ = (M + L) - M [NEWLINE] [NEWLINE] X₁ = L. [NEWLINE] [NEWLINE] Even removing the physical matter from the formula, we're still left with a value that has already been set by the original price.<mask> labor carries value, X₁ still has a cost: whatever L is.<mask> you're copying X, you still have L left.<mask>, you still owe some amount of payment for the value of L.<mask> you don't pay for something that requires a price, you're stealing it. The owner of X loses money<mask> they didn't pay for the labor. [NEWLINE] [NEWLINE] Copying and stealing (a physical product), in this context, equal, eventually, the same thing: a cost,<mask><mask> the costs are disparate. That's<mask> I keep calling it stealing:<mask>, ultimately, pirating is stealing.</s>
Label encoding: <s>I must not be explaining what I'm saying very well. You seem to be ignoring the most important part of my argument: the labor. [NEWLINE] [NEWLINE] The physical matter that composes X is one part of two that composes X. There is matter, M, and labor, L, that went into the creation, or crafting, of X. We could have different variables, that you mention, logistics, travel, et cetera, but we'll put those under M because they're in the physical realm. L is, say, the writing (and the editing by the editor, et cetera) of the text X, or the programming or the designing of the software X, or the formulation for candy bar X. X is composed of M and L, together. M exists by itself without L, but it takes the addition of L to create X. For some really informal logic: [NEWLINE] [NEWLINE] X = M + L. [NEWLINE] [NEWLINE] With your magic copying device, you copy X, and get its exact clone, which we'll call X₁. X₁ is the exact same object as X. X₁ = X. However, this one took zero material, M, to make. So we take away the value of the material. [NEWLINE] [NEWLINE] X₁ = X - M [NEWLINE] [NEWLINE] X₁ = (M + L) - M [NEWLINE] [NEWLINE] X₁ = L. [NEWLINE] [NEWLINE] Even removing the physical matter from the formula, we're still left with a value that has already been set by the original price. Since labor carries value, X₁ still has a cost: whatever L is. When you're copying X, you still have L left. Therefore, you still owe some amount of payment for the value of L. When you don't pay for something that requires a price, you're stealing it. The owner of X loses money because they didn't pay for the labor. [NEWLINE] [NEWLINE] Copying and stealing (a physical product), in this context, equal, eventually, the same thing: a cost, even though the costs are disparate. That's why I keep calling it stealing: because, ultimately, pirating is stealing.</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0918, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0900, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1) People think we caused (or impacted at all) the changes [ENDQ] [NEWLINE] The reason that people think we are causing the climate to change is<mask> of simple physics. The problem is ultimately about conservation of energy, that is, the total energy incident on the Earth from the sun has to be accounted for completely. This is called "radiative balancing". [Here]( [URL].png) is a nice cartoon representation. [NEWLINE] [NEWLINE] Physics tells us that different components of the atmosphere interact with radiation, that is, heat and light, differently. "Heat" is mostly associated with infrared radiation, which is strongly absorbed by gases in the atmosphere. [Here]( [URL].png) is a plot which shows<mask> much radiation is transmitted and absorbed over the relevant wavelength range. The upshot is this - CO2 "traps" heat.<mask> you increase the amount of CO2 in the atmosphere, you will warm the surface. Humans have emitted a very large amount of CO2. [NEWLINE] [NEWLINE] We can look at this on a global scale. [Here]( [URL].gif) is a map of the measured longwave radiation ("heat") and [here]( [URL].jpg) is the global CO2 concentration<mask> measured by an orbiting NASA observatory. Note that<mask> CO2 concentration is high is<mask> the measured heat flux is low. This is one of many things you expect<mask> climate change is driven by CO2 emissions. [NEWLINE] [NEWLINE] [STARTQ] 2)<mask> people think these changes are catastrophic to our way of life [ENDQ] [NEWLINE] Most people don't think that the average redditor will be catastrophically impacted.<mask> most people, including economists and policy makers have recognized is that climate change will impact our ability to grow and distribute food, it will lead to migration away from coastlines, and associated infrastructure costs. For the average redditor in the US or Europe, your lunch might be a bit more expensive than it otherwise would be. It hurts the economy on a macro-scale,<mask> it won't be catastrophic.<mask> it will really hurt is in places like Asia and Africa,<mask> economies are less developed and people have less capital to be able to afford more expensive food and energy. </s>
Label encoding: <s> [STARTQ] 1) People think we caused (or impacted at all) the changes [ENDQ] [NEWLINE] The reason that people think we are causing the climate to change is because of simple physics. The problem is ultimately about conservation of energy, that is, the total energy incident on the Earth from the sun has to be accounted for completely. This is called "radiative balancing". [Here]( [URL].png) is a nice cartoon representation. [NEWLINE] [NEWLINE] Physics tells us that different components of the atmosphere interact with radiation, that is, heat and light, differently. "Heat" is mostly associated with infrared radiation, which is strongly absorbed by gases in the atmosphere. [Here]( [URL].png) is a plot which shows how much radiation is transmitted and absorbed over the relevant wavelength range. The upshot is this - CO2 "traps" heat. If you increase the amount of CO2 in the atmosphere, you will warm the surface. Humans have emitted a very large amount of CO2. [NEWLINE] [NEWLINE] We can look at this on a global scale. [Here]( [URL].gif) is a map of the measured longwave radiation ("heat") and [here]( [URL].jpg) is the global CO2 concentration as measured by an orbiting NASA observatory. Note that where CO2 concentration is high is where the measured heat flux is low. This is one of many things you expect if climate change is driven by CO2 emissions. [NEWLINE] [NEWLINE] [STARTQ] 2) Why people think these changes are catastrophic to our way of life [ENDQ] [NEWLINE] Most people don't think that the average redditor will be catastrophically impacted. What most people, including economists and policy makers have recognized is that climate change will impact our ability to grow and distribute food, it will lead to migration away from coastlines, and associated infrastructure costs. For the average redditor in the US or Europe, your lunch might be a bit more expensive than it otherwise would be. It hurts the economy on a macro-scale, but it won't be catastrophic. Where it will really hurt is in places like Asia and Africa, where economies are less developed and people have less capital to be able to afford more expensive food and energy. </s>
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Therefore it follows that<mask> one creates something, everything that that creation does it the fault of the creator? [NEWLINE] [NEWLINE] Disease (I'm grouping sickness and parasites into one group) functions to keep itself alive, in some cases by destroying the cells of the host body. It natural selection at work. God created microbes and cells and all matter, for that matter (see<mask> I did there?),<mask> blaming him for H1N1 is like blaming a potato farmer<mask> a spud gets shot through your window by a spud gun. Sure, the farmer planted and grew the potato,<mask> it took a 90 degree turn from its intended purpose. [NEWLINE] [NEWLINE] A disease is not inherently evil, nor is a natural predator. They are fulfilling their base function, which is keeping itself alive. Human suffering is not inherently evil either. Suffering is a part of life, being protected from all forms of suffering ever would leave us empty. Through suffering, people can discover other emotions and feelings, such<mask> love, charity or hope. Take away suffering, and there is no reason to hope or to be charitable. Love is cheapened,<mask> it no longer requires giving a part of one's self to another, it takes away the risk. In Christian theology, a lot of thought is placed on suffering, and a lot of very, very smart people have discussed it. Suffering, in and of itself, is not evil. Evil can lead to suffering for others (and the one committing the evil),<mask> it doesn't always. And suffering doesn't always have to follow from evil, it can follow from morally neutral events.<mask> you are of the opinion that suffering and evil are directly related, and that all suffering is evil, then we need to have an entirely different debate. [NEWLINE] [NEWLINE] The gift of free will is the ultimate expression of love. Without it, God would merely be a benevolent dictator.<mask>'s the point of going through life<mask> you can never make a decision for yourself? Sure, people are going to do bad things, and people are going to do good things. Blaming God<mask> people do bad things is like blaming your parents every time you screw up.</s>
Label encoding: <s>Therefore it follows that when one creates something, everything that that creation does it the fault of the creator? [NEWLINE] [NEWLINE] Disease (I'm grouping sickness and parasites into one group) functions to keep itself alive, in some cases by destroying the cells of the host body. It natural selection at work. God created microbes and cells and all matter, for that matter (see what I did there?), but blaming him for H1N1 is like blaming a potato farmer when a spud gets shot through your window by a spud gun. Sure, the farmer planted and grew the potato, but it took a 90 degree turn from its intended purpose. [NEWLINE] [NEWLINE] A disease is not inherently evil, nor is a natural predator. They are fulfilling their base function, which is keeping itself alive. Human suffering is not inherently evil either. Suffering is a part of life, being protected from all forms of suffering ever would leave us empty. Through suffering, people can discover other emotions and feelings, such as love, charity or hope. Take away suffering, and there is no reason to hope or to be charitable. Love is cheapened, because it no longer requires giving a part of one's self to another, it takes away the risk. In Christian theology, a lot of thought is placed on suffering, and a lot of very, very smart people have discussed it. Suffering, in and of itself, is not evil. Evil can lead to suffering for others (and the one committing the evil), but it doesn't always. And suffering doesn't always have to follow from evil, it can follow from morally neutral events. If you are of the opinion that suffering and evil are directly related, and that all suffering is evil, then we need to have an entirely different debate. [NEWLINE] [NEWLINE] The gift of free will is the ultimate expression of love. Without it, God would merely be a benevolent dictator. What's the point of going through life if you can never make a decision for yourself? Sure, people are going to do bad things, and people are going to do good things. Blaming God when people do bad things is like blaming your parents every time you screw up.</s>
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] One of the reasons I like it,<mask>, is<mask> of<mask> serious it does NOT take itself.<mask> people talk about rappers being deep, I laugh most of the time. [ENDQ] [NEWLINE] This is going to sound snobbish. There are quite a few very deep rappers and hip hop artists out there.  Largely outside of the mainstream. [NEWLINE] [NEWLINE] I would say you're entire argument is spot on for the pop side of things.  Most rap that's played on the radio, and most rap artists making big money, are not very serious.  Or deep.  That's just kind of the nature of pop music<mask>. [NEWLINE] [NEWLINE] [STARTQ] Comparing it to actual poetry I feel is a very immature argument, and usually made by people who never read actual poetry. [ENDQ] [NEWLINE] <mask>? Atmosphere and Aesop Rock come to mind<mask> pretty poetic rap.  That's just off the cuff. <mask> I dug around, I know I could find plenty more examples. [NEWLINE] [NEWLINE] A lot of rap is like poetry.  I don't see<mask> that's immature.  To me, it's just stating the obvious. [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask> someone who has played different musical instruments and rapped, rapping at an elementary level was always much easier than playing an instrument. [ENDQ] [NEWLINE] Wow, speaking words is much easier than playing an instrument? You don't say? [NEWLINE] [NEWLINE] Sure, it's easy to string together a some words, throw in a few rhymes.  Especially in comparison to<mask> it takes to play an instrument. <mask> just<mask> you can karaoke Purple Pills doesn't mean you're a good rapper. [NEWLINE] [NEWLINE] [STARTQ] Case in point, Eazy E became a famous rapper (even touring the white house) doing something he was terrible at. And rightfully<mask>,<mask> you can be a terrible rapper and still sell a ton of records and actually be considered great at<mask> you do, and that does not apply to other forms of music. [ENDQ] [NEWLINE] You can be a great pop singer without actually being able to sing these days too.  Plenty of studio only singers out there churning out chart topping singles.</s>
Label encoding: <s> [STARTQ] One of the reasons I like it, however, is because of how serious it does NOT take itself. When people talk about rappers being deep, I laugh most of the time. [ENDQ] [NEWLINE] This is going to sound snobbish. There are quite a few very deep rappers and hip hop artists out there.  Largely outside of the mainstream. [NEWLINE] [NEWLINE] I would say you're entire argument is spot on for the pop side of things.  Most rap that's played on the radio, and most rap artists making big money, are not very serious.  Or deep.  That's just kind of the nature of pop music though. [NEWLINE] [NEWLINE] [STARTQ] Comparing it to actual poetry I feel is a very immature argument, and usually made by people who never read actual poetry. [ENDQ] [NEWLINE] Why? Atmosphere and Aesop Rock come to mind as pretty poetic rap.  That's just off the cuff.  If I dug around, I know I could find plenty more examples. [NEWLINE] [NEWLINE] A lot of rap is like poetry.  I don't see how that's immature.  To me, it's just stating the obvious. [NEWLINE] [NEWLINE] [STARTQ] Also, as someone who has played different musical instruments and rapped, rapping at an elementary level was always much easier than playing an instrument. [ENDQ] [NEWLINE] Wow, speaking words is much easier than playing an instrument? You don't say? [NEWLINE] [NEWLINE] Sure, it's easy to string together a some words, throw in a few rhymes.  Especially in comparison to what it takes to play an instrument.  But just because you can karaoke Purple Pills doesn't mean you're a good rapper. [NEWLINE] [NEWLINE] [STARTQ] Case in point, Eazy E became a famous rapper (even touring the white house) doing something he was terrible at. And rightfully so, because you can be a terrible rapper and still sell a ton of records and actually be considered great at what you do, and that does not apply to other forms of music. [ENDQ] [NEWLINE] You can be a great pop singer without actually being able to sing these days too.  Plenty of studio only singers out there churning out chart topping singles.</s>
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Except our natural resources here on earth are already being stretched thin [ENDQ] [NEWLINE] They are at our current technology level, and have been for<mask> at our historical technology levels. We've been living on the edge<mask> new technology enables more efficient and new uses of resources<mask> at the same time we increase in resource demands. [NEWLINE] [NEWLINE] [STARTQ] all the nearby moons/planets we know of have little to no resources capable of sustaining life. [ENDQ] [NEWLINE] Absolutely not true. Or at least not the whole story. It just requires a ton of technology to transform the local resources into life sustaining forms. [NEWLINE] [NEWLINE] [STARTQ] consume our non-renewable resources even faster than we are now. [ENDQ] [NEWLINE] The definition of 'non-renewable' is a technological context specific. Fossil fuels are nonrenewable using natural methods (geologic scale decomposition of matter under pressure) under reasonable timeframes.<mask> is may be possible to use other methods such<mask> bioengineered algae to produce all the specific hydrocarbons we need. Fresh-water is reusable,<mask> even contaminated water can be purified and salt-water can be desalinated. The problem isn't the lack of resources. The problem is the lack of resources cheaply available with simple or no technology.<mask> the limit here really is energy.<mask> we had cheap, efficient, solar powered desalinators producing freshwater for coastal cities (which is already starting to happen) we are most the way there to removing water<mask> a resource obstacle. You can easily imagine similar things for space environments (such<mask> asteroid mining, or atmospheric condenser systems)<mask> we just aren't at a technology level were we can do these things cheaply -<mask>.<mask> we aren't that far away from robots that can make significant cost reduction. Imagine sending a robot mining/factory to the astroid belt or mars. It has two functions, mine resources need for robots and two build them. We probably are 10-20 years away from being able to do that - even<mask> we need a high degree of human interaction - the automation can still make it scalable.<mask> you bootstap a process that can fabricate human habitats.</s>
Label encoding: <s> [STARTQ] Except our natural resources here on earth are already being stretched thin [ENDQ] [NEWLINE] They are at our current technology level, and have been for while at our historical technology levels. We've been living on the edge where new technology enables more efficient and new uses of resources while at the same time we increase in resource demands. [NEWLINE] [NEWLINE] [STARTQ] all the nearby moons/planets we know of have little to no resources capable of sustaining life. [ENDQ] [NEWLINE] Absolutely not true. Or at least not the whole story. It just requires a ton of technology to transform the local resources into life sustaining forms. [NEWLINE] [NEWLINE] [STARTQ] consume our non-renewable resources even faster than we are now. [ENDQ] [NEWLINE] The definition of 'non-renewable' is a technological context specific. Fossil fuels are nonrenewable using natural methods (geologic scale decomposition of matter under pressure) under reasonable timeframes. But is may be possible to use other methods such as bioengineered algae to produce all the specific hydrocarbons we need. Fresh-water is reusable, but even contaminated water can be purified and salt-water can be desalinated. The problem isn't the lack of resources. The problem is the lack of resources cheaply available with simple or no technology. So the limit here really is energy. If we had cheap, efficient, solar powered desalinators producing freshwater for coastal cities (which is already starting to happen) we are most the way there to removing water as a resource obstacle. You can easily imagine similar things for space environments (such as asteroid mining, or atmospheric condenser systems) but we just aren't at a technology level were we can do these things cheaply - yet. But we aren't that far away from robots that can make significant cost reduction. Imagine sending a robot mining/factory to the astroid belt or mars. It has two functions, mine resources need for robots and two build them. We probably are 10-20 years away from being able to do that - even if we need a high degree of human interaction - the automation can still make it scalable. So you bootstap a process that can fabricate human habitats.</s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It goes both ways. Do we need consent from a baby to adopt him or her? Do we need consent to circumcise baby boys? Do we need consent<mask> we want to get our daughter's ears pierced at a very young age? Do we need to get a criminal's consent before executing him/her<mask> a punishment? No, no, no, and no. I personally think this entire argument is a little bit inconsistent (including,<mask> you pointed out, that the buck stops at bestiality - or child abuse, or, for some, abortion),<mask> that doesn't mean that something like sexual consent isn't more weighty than, say, pet adoption.<mask>, against your first point, I'd<mask><mask> people view sex<mask> more weighty than, say, adoption or neutering. [NEWLINE] [NEWLINE] Personally, I don't think they really should.<mask>,<mask><mask> the arguments against bestiality I've heard don't revolve<mask> much around the sex itself. Rather, most of them are in response to people arguing that legalizing gay marriage/polyamory/etc. will lead down a slippery slope to bestiality, necrophilia, or pedophilia. In that case,<mask><mask> bestiality should be illegal - marriage is defined<mask> between consenting adults, which is<mask> marrying an animal, corpse, or child would be illegal. [NEWLINE] [NEWLINE] I do think it's hypocritical to frown upon bestiality<mask> condoning the slaughter of<mask> many animals for food, say, or an abusive pet owner whose dog/cat/bird/etc. is clearly terrified and unhappy at any moment.<mask><mask> it partially stems from our puritanical notions of sex<mask> taboo, and<mask><mask> it's partially from the fact that we view sex<mask><mask> weighty (and rightly<mask>,<mask> it's tough to<mask><mask> it's more weighty than slaughter). That being said,<mask> I'm arguing against bestiality, I'm arguing against it<mask> a legal concept similar to marriage - an adult shouldn't be allowed to marry his cow for the same reason he shouldn't be allowed to marry his 10-year-old daughter.</s>
Label encoding: <s>It goes both ways. Do we need consent from a baby to adopt him or her? Do we need consent to circumcise baby boys? Do we need consent if we want to get our daughter's ears pierced at a very young age? Do we need to get a criminal's consent before executing him/her as a punishment? No, no, no, and no. I personally think this entire argument is a little bit inconsistent (including, as you pointed out, that the buck stops at bestiality - or child abuse, or, for some, abortion), but that doesn't mean that something like sexual consent isn't more weighty than, say, pet adoption. So, against your first point, I'd argue that people view sex as more weighty than, say, adoption or neutering. [NEWLINE] [NEWLINE] Personally, I don't think they really should. However, I think the arguments against bestiality I've heard don't revolve so much around the sex itself. Rather, most of them are in response to people arguing that legalizing gay marriage/polyamory/etc. will lead down a slippery slope to bestiality, necrophilia, or pedophilia. In that case, I think bestiality should be illegal - marriage is defined as between consenting adults, which is why marrying an animal, corpse, or child would be illegal. [NEWLINE] [NEWLINE] I do think it's hypocritical to frown upon bestiality while condoning the slaughter of so many animals for food, say, or an abusive pet owner whose dog/cat/bird/etc. is clearly terrified and unhappy at any moment. I think it partially stems from our puritanical notions of sex as taboo, and I think it's partially from the fact that we view sex as so weighty (and rightly so, though it's tough to argue that it's more weighty than slaughter). That being said, when I'm arguing against bestiality, I'm arguing against it as a legal concept similar to marriage - an adult shouldn't be allowed to marry his cow for the same reason he shouldn't be allowed to marry his 10-year-old daughter.</s>
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right off the bat, let me say two things. I'm not racist (consciously; maybe I'll find out I am racist).<mask> I say I don't have an "unfair* advantage, I do not mean I deserve an advantage, I mean something else which I'll elaborate on. [NEWLINE] [NEWLINE] <mask> basically I've been told that I'm benefiting from racism (and that I should feel guilty<mask> of it), and<mask><mask> that's bogus. More<mask> the part about me feeling guilty. [NEWLINE] [NEWLINE] I recognize that being white puts me ahead of minorities in certain situations. Don't get me wrong, I definitely recognize that minorities deal with mistreatment or judgment that I just don't have to deal with.<mask> I don't think I should feel guilty for being white, nor do<mask><mask> I'm actually benefiting from being white in an objective way (I know this sounds confusing, bare with me). [NEWLINE] [NEWLINE] Others are perhaps detriment-ed by not being white,<mask> that's not actually an advantage for me. To elaborate further, racism causes minorities to be treated poorly,<mask> in comparison to me, it appears that I have an advantage.<mask> I'm not actually being treated better than I ought to be treated, the problem is others are being treated worse. Basically we should treat minorities the same<mask> whites, not whites the same<mask> minorities. [NEWLINE] [NEWLINE] <mask><mask> I say I don't have an objective advantage, I'm saying my advantage only exists<mask> I'm competing against a minority (who is being treated unfairly). It's like saying runner #1 has an advantage<mask> runner #2 hurt his knee. [NEWLINE] [NEWLINE] <mask> this boils down to is I don't think I should feel bad, and I don't think it's fair to say I'm actually benefiting from racism. Others are hurt by racism,<mask> that's not automatically good for me. [NEWLINE] [NEWLINE] I had nothing to do with america's racism.<mask><mask> everyone who actually participated in slavery is dead. I really think we should be done shaming people and telling people they should feel bad for being white. We should focus on eliminating racism, not redirecting it.</s>
Label encoding: <s>Right off the bat, let me say two things. I'm not racist (consciously; maybe I'll find out I am racist). When I say I don't have an "unfair* advantage, I do not mean I deserve an advantage, I mean something else which I'll elaborate on. [NEWLINE] [NEWLINE] So basically I've been told that I'm benefiting from racism (and that I should feel guilty because of it), and I think that's bogus. More so the part about me feeling guilty. [NEWLINE] [NEWLINE] I recognize that being white puts me ahead of minorities in certain situations. Don't get me wrong, I definitely recognize that minorities deal with mistreatment or judgment that I just don't have to deal with. But I don't think I should feel guilty for being white, nor do I think I'm actually benefiting from being white in an objective way (I know this sounds confusing, bare with me). [NEWLINE] [NEWLINE] Others are perhaps detriment-ed by not being white, but that's not actually an advantage for me. To elaborate further, racism causes minorities to be treated poorly, so in comparison to me, it appears that I have an advantage. But I'm not actually being treated better than I ought to be treated, the problem is others are being treated worse. Basically we should treat minorities the same as whites, not whites the same as minorities. [NEWLINE] [NEWLINE] So when I say I don't have an objective advantage, I'm saying my advantage only exists when I'm competing against a minority (who is being treated unfairly). It's like saying runner #1 has an advantage because runner #2 hurt his knee. [NEWLINE] [NEWLINE] What this boils down to is I don't think I should feel bad, and I don't think it's fair to say I'm actually benefiting from racism. Others are hurt by racism, but that's not automatically good for me. [NEWLINE] [NEWLINE] I had nothing to do with america's racism. In fact everyone who actually participated in slavery is dead. I really think we should be done shaming people and telling people they should feel bad for being white. We should focus on eliminating racism, not redirecting it.</s>
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "Would using an ipod with songs that my friend had purchased be stealing? I guess it would be right?" [ENDQ] [NEWLINE] <mask> you don't give the iPod back to your friend, then yeah, you stole his iPod. Seriously<mask>, in this scenario, someone has paid for one copy of the music,<mask> it's fair enough<mask> that one copy changes hands. It's considerably less fair<mask> they pay for one copy and then give it to 10 or 100 or 1000 people, which is the usual definition of "piracy." [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> I want to go for a run and listen to music<mask> I can't stream some service<mask> I have an ipod and not a smart phone. [ENDQ] [NEWLINE] <mask> you like a song, you can buy it legally for less than $1 (it's most likely $0.99 on iTunes or Amazon or Google). In other words, for the price of a cup of coffee you can buy the four best songs on an album. For the price of a movie ticket you can buy the album. Then put your legally downloaded track on your iPod and listen to it<mask> much<mask> you want. Is that really<mask> inconvenient? [NEWLINE] [NEWLINE] [STARTQ] "I don't think downloading them does them harm<mask> I wouldn't have paid for them." [ENDQ] [NEWLINE] The harm is that you're devaluing music. You're saying you like a song enough to listen to it<mask> you run,<mask> not enough to spend a measly $0.99 on it.<mask>'s revealing to me is that you say not "I *couldn't* have paid for them"<mask> "I *wouldn't* have paid for them." [NEWLINE] [NEWLINE] In other words, you expect to benefit from the results of musicians' time and effort<mask> not to pay them for their work<mask> you use it. Would you like it<mask> your boss expected you to show up and do your job for free on the grounds that he wouldn't have paid you anyway? You'd find that pretty insulting and demoralizing, yes? Then<mask> do you expect the same of a musician's job? [NEWLINE] [NEWLINE] (Edited, clarity.)</s>
Label encoding: <s> [STARTQ] "Would using an ipod with songs that my friend had purchased be stealing? I guess it would be right?" [ENDQ] [NEWLINE] If you don't give the iPod back to your friend, then yeah, you stole his iPod. Seriously though, in this scenario, someone has paid for one copy of the music, so it's fair enough if that one copy changes hands. It's considerably less fair if they pay for one copy and then give it to 10 or 100 or 1000 people, which is the usual definition of "piracy." [NEWLINE] [NEWLINE] [STARTQ] What if I want to go for a run and listen to music but I can't stream some service because I have an ipod and not a smart phone. [ENDQ] [NEWLINE] If you like a song, you can buy it legally for less than $1 (it's most likely $0.99 on iTunes or Amazon or Google). In other words, for the price of a cup of coffee you can buy the four best songs on an album. For the price of a movie ticket you can buy the album. Then put your legally downloaded track on your iPod and listen to it as much as you want. Is that really so inconvenient? [NEWLINE] [NEWLINE] [STARTQ] "I don't think downloading them does them harm If I wouldn't have paid for them." [ENDQ] [NEWLINE] The harm is that you're devaluing music. You're saying you like a song enough to listen to it while you run, but not enough to spend a measly $0.99 on it. What's revealing to me is that you say not "I *couldn't* have paid for them" but "I *wouldn't* have paid for them." [NEWLINE] [NEWLINE] In other words, you expect to benefit from the results of musicians' time and effort but not to pay them for their work when you use it. Would you like it if your boss expected you to show up and do your job for free on the grounds that he wouldn't have paid you anyway? You'd find that pretty insulting and demoralizing, yes? Then why do you expect the same of a musician's job? [NEWLINE] [NEWLINE] (Edited, clarity.)</s>
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think there is something that make Jews particularly vulnerable,<mask>, in comparison to other commonly prosecuted groups *in North America*. [NEWLINE] [NEWLINE] Judaism is unique in that it is an Ethnoreligious group; it is both a racial minority and a religious minority simultaneously. This is a dangerous combination<mask> it comes to prejudice,<mask> bigots can use it<mask> "proof" that the Jewish race is even more of a threat than their other 'go-to's. [NEWLINE] [NEWLINE] Look at the reasoning a racist has for their hatred of black peoples, for example; it's not very complex, is it? It's a want for superiority more than anything else. The fear a racist can have is less about black people being intrinsically threatening than it is about a group they deem subhuman coming into the same social ranking<mask> them, and<mask> infringing on their ability to feel superior. The idea of a group being inferior works very much against the possibility of them being an organized, sophisticated threat. [NEWLINE] [NEWLINE] Prejudice against religious groups is, by contrast, often motivated by a self-centered fear. It's much more *tangible.* Catholics, for example, have been hated in the past<mask> in more nationalistic times, their allegiance to the Pope has been taken by others<mask> reason to believe that they would not be trustworthy. It's more misinterpretation than it is fabrication, which makes it more solid, more justified in the mind of the bigot and more resistant to modern ethics. At the same time, a religion is a changeable an acquired set of beliefs rather than an inborn trait,<mask> antireligious sentiment tends to be less vicious. [NEWLINE] [NEWLINE] Now *combine* underlying rationale for each of those things, and it's pretty much the bigot's worst nightmare. They're not just seen<mask> bad *or* oppositional, they're seen<mask> MILITANTLY bad. Have you noticed<mask> Antisemites fabricate hysterical conspiracy theories more than any other bigoted group, at least<mask><mask><mask> North America goes? Have you ever seen theories of *entire shadow governments* for any other group?</s>
Label encoding: <s>I think there is something that make Jews particularly vulnerable, though, in comparison to other commonly prosecuted groups *in North America*. [NEWLINE] [NEWLINE] Judaism is unique in that it is an Ethnoreligious group; it is both a racial minority and a religious minority simultaneously. This is a dangerous combination when it comes to prejudice, because bigots can use it as "proof" that the Jewish race is even more of a threat than their other 'go-to's. [NEWLINE] [NEWLINE] Look at the reasoning a racist has for their hatred of black peoples, for example; it's not very complex, is it? It's a want for superiority more than anything else. The fear a racist can have is less about black people being intrinsically threatening than it is about a group they deem subhuman coming into the same social ranking as them, and thus infringing on their ability to feel superior. The idea of a group being inferior works very much against the possibility of them being an organized, sophisticated threat. [NEWLINE] [NEWLINE] Prejudice against religious groups is, by contrast, often motivated by a self-centered fear. It's much more *tangible.* Catholics, for example, have been hated in the past because in more nationalistic times, their allegiance to the Pope has been taken by others as reason to believe that they would not be trustworthy. It's more misinterpretation than it is fabrication, which makes it more solid, more justified in the mind of the bigot and more resistant to modern ethics. At the same time, a religion is a changeable an acquired set of beliefs rather than an inborn trait, so antireligious sentiment tends to be less vicious. [NEWLINE] [NEWLINE] Now *combine* underlying rationale for each of those things, and it's pretty much the bigot's worst nightmare. They're not just seen as bad *or* oppositional, they're seen as MILITANTLY bad. Have you noticed how Antisemites fabricate hysterical conspiracy theories more than any other bigoted group, at least as far as North America goes? Have you ever seen theories of *entire shadow governments* for any other group?</s>
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Atheism is a broad topic with many included aspects, much like "politics". [NEWLINE] [NEWLINE] Antitheism is a specific element that is included in the broad topic of atheism, much like the second amendment is a popular topic in /r/politics. [NEWLINE] [NEWLINE] Atheistic humor is another such broad topic. In much the same way that we can find jokes about minorities (including silly ones like blondes) funny without being against minorities, atheists can find memes about<mask> silly religion is funny without necessarily being against religion. [NEWLINE] [NEWLINE] There are numerous posts in /r/atheism about other topics. One of the most common to come up is "I'm coming out<mask> atheist to my parents,<mask> should I do?". And the usual answer is "don't do that until you can safely live on your own"<mask>, sadly, that's really extremely good advice. There will be comments in such a discussion that are anti-theist, of course,<mask> there will be many others that are just sympathetic and still others that will be practical. [NEWLINE] [NEWLINE] Another extremely common topic in /r/atheism is "<mask> did you become atheist?". Another common topic is "I'm having doubts about my religion,<mask> books should I read". [NEWLINE] [NEWLINE] You're suffering from confirmation bias.<mask> you think that all atheist is antitheism, you're going to only see the anti-theist posts in /r/atheism. [NEWLINE] [NEWLINE] And,<mask>, they do exist, and are a large fraction of the postings,<mask> it does fit under the umbrella, and sadly most atheists on reddit live in countries<mask> people do a lot of harmful things in the name of religion (whether or not one can attribute them to the religion... a point that universally *always* comes up in the discussions about these threads... which is another reason it's not "just antitheism"). [NEWLINE] [NEWLINE] The topics that rise to the top in a subreddit with more than a million subscribers will always be the easily digestible stuff that people find funny or superficially poignant. That's by no means unique to /r/atheism. </s>
Label encoding: <s>Atheism is a broad topic with many included aspects, much like "politics". [NEWLINE] [NEWLINE] Antitheism is a specific element that is included in the broad topic of atheism, much like the second amendment is a popular topic in /r/politics. [NEWLINE] [NEWLINE] Atheistic humor is another such broad topic. In much the same way that we can find jokes about minorities (including silly ones like blondes) funny without being against minorities, atheists can find memes about how silly religion is funny without necessarily being against religion. [NEWLINE] [NEWLINE] There are numerous posts in /r/atheism about other topics. One of the most common to come up is "I'm coming out as atheist to my parents, what should I do?". And the usual answer is "don't do that until you can safely live on your own" because, sadly, that's really extremely good advice. There will be comments in such a discussion that are anti-theist, of course, but there will be many others that are just sympathetic and still others that will be practical. [NEWLINE] [NEWLINE] Another extremely common topic in /r/atheism is " why did you become atheist?". Another common topic is "I'm having doubts about my religion, what books should I read". [NEWLINE] [NEWLINE] You're suffering from confirmation bias. If you think that all atheist is antitheism, you're going to only see the anti-theist posts in /r/atheism. [NEWLINE] [NEWLINE] And, indeed, they do exist, and are a large fraction of the postings, because it does fit under the umbrella, and sadly most atheists on reddit live in countries where people do a lot of harmful things in the name of religion (whether or not one can attribute them to the religion... a point that universally *always* comes up in the discussions about these threads... which is another reason it's not "just antitheism"). [NEWLINE] [NEWLINE] The topics that rise to the top in a subreddit with more than a million subscribers will always be the easily digestible stuff that people find funny or superficially poignant. That's by no means unique to /r/atheism. </s>
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is not an argument from authority. This is exactly the part that's causing all the confusion. It is totally logically valid to say that someone with expertise on a given subject is more likely to be correct than a layperson.<mask> is not logically valid is to say that the expert is correct for that reason. It's a subtle<mask> very meaningful distinction. [ENDQ] [NEWLINE] Actually it still is.<mask> you haven't limited the scope enough to make the assertion not fallacious. It is<mask> not fallacious to say that someone with expertise on a given subject is more likely to be correct than a layperson OVER AN ARBITRARILY LARGE number of "tests" (ie. questions). It is NOT correct to do the same for a SINGLE test (ie. question)<mask> there it assumes COMPLETELY that the likelihood that an expert is correct is uniform across an ENTIRE subject. This assumption is almost NEVER satisfied and<mask> such you cannot come to such a conclusion. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> that<mask><mask> you're saying is<mask><mask> OP meant it, it is 100% fallacious,<mask> I don't think that's<mask> s/he meant at all. I interpreted that<mask> much more of a complaint about<mask> everybody with a black belt in google-fu thinks they're an expert on a given subject,<mask> in reality they generally simply don't have the depth and breadth of experience in that subject to offer a comparably well-evidenced and informed opinion to one offered by a true expert in the field. Like all the anti-vax cranks who claim to have "done their homework"<mask> in reality don't have the faintest idea<mask> the data actually says. [ENDQ] [NEWLINE] <mask> it were intended the way you describe, that would be a different topic.<mask> it is relatively evident from the way OP words the statements especially with the inclusion of the phrase "will allow themselves to form a strong opinion" that he is specifically committing to an argument from authority or ad hominem in one way or another (either that expert opinions are necessarily stronger or that lay opinions are necessarily weaker). </s>
Label encoding: <s> [STARTQ] This is not an argument from authority. This is exactly the part that's causing all the confusion. It is totally logically valid to say that someone with expertise on a given subject is more likely to be correct than a layperson. What is not logically valid is to say that the expert is correct for that reason. It's a subtle but very meaningful distinction. [ENDQ] [NEWLINE] Actually it still is. Because you haven't limited the scope enough to make the assertion not fallacious. It is indeed not fallacious to say that someone with expertise on a given subject is more likely to be correct than a layperson OVER AN ARBITRARILY LARGE number of "tests" (ie. questions). It is NOT correct to do the same for a SINGLE test (ie. question) because there it assumes COMPLETELY that the likelihood that an expert is correct is uniform across an ENTIRE subject. This assumption is almost NEVER satisfied and as such you cannot come to such a conclusion. [NEWLINE] [NEWLINE] [STARTQ] I agree that if what you're saying is indeed how OP meant it, it is 100% fallacious, but I don't think that's what s/he meant at all. I interpreted that as much more of a complaint about how everybody with a black belt in google-fu thinks they're an expert on a given subject, when in reality they generally simply don't have the depth and breadth of experience in that subject to offer a comparably well-evidenced and informed opinion to one offered by a true expert in the field. Like all the anti-vax cranks who claim to have "done their homework" but in reality don't have the faintest idea what the data actually says. [ENDQ] [NEWLINE] If it were intended the way you describe, that would be a different topic. But it is relatively evident from the way OP words the statements especially with the inclusion of the phrase "will allow themselves to form a strong opinion" that he is specifically committing to an argument from authority or ad hominem in one way or another (either that expert opinions are necessarily stronger or that lay opinions are necessarily weaker). </s>
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The thing is, most clients *don't* know<mask> they want, or<mask> they do, it would be extremely inadvisable to execute it the way they have in mind. Funnily enough (I can laugh at it now), I've actually had a client sit next to me and dictate, step by step,<mask> they wanted me to do. He was pretty proud of himself,<mask> let's just say it won't be making an appearance in my portfolio. [NEWLINE] [NEWLINE] Most professionals use the services of a graphic designer<mask> they don't know<mask> to communicate visually. They don't know<mask> makes an effective design. They don't know<mask> to use space, color, emphasis, etc. to convey information. They want<mask> they are familiar with, which usually involves Comic Sans and clip art. They may have an *idea* of<mask> they want,<mask> it is your job<mask> a professional to get to the bottom of<mask> they *actually* want, interpret that notion into something that achieves their purpose, and make it look good<mask> you're at it. [NEWLINE] [NEWLINE] True, you are not an artist in a purely self-expressive sense.<mask> there's definitely more to it than being a Photoshop jockey. The sheer volume of garbage produced by someone's nephew who pirated CS and thinks they're a designer should attest to that (see /r/CrappyDesign). And, true enough, at the beginning of your career there will be a lot of people giving you very specific direction.<mask><mask> you get further into your career, you will be in a position to have greater and greater creative control. [NEWLINE] [NEWLINE] In the mean time, even<mask> your day job doesn't satisfy your creativity, seek out other opportunities. Freelance whenever possible, and don't be afraid to professionally and firmly steer a client away from questionable design choices. There are<mask> many worthy non-profits that need people to donate their design services (the one exception to "never do free work"). In my experience, these organizations are often willing to give you great creative freedom and are grateful and gracious to boot.</s>
Label encoding: <s>The thing is, most clients *don't* know what they want, or if they do, it would be extremely inadvisable to execute it the way they have in mind. Funnily enough (I can laugh at it now), I've actually had a client sit next to me and dictate, step by step, what they wanted me to do. He was pretty proud of himself, but let's just say it won't be making an appearance in my portfolio. [NEWLINE] [NEWLINE] Most professionals use the services of a graphic designer because they don't know how to communicate visually. They don't know what makes an effective design. They don't know how to use space, color, emphasis, etc. to convey information. They want what they are familiar with, which usually involves Comic Sans and clip art. They may have an *idea* of what they want, but it is your job as a professional to get to the bottom of what they *actually* want, interpret that notion into something that achieves their purpose, and make it look good while you're at it. [NEWLINE] [NEWLINE] True, you are not an artist in a purely self-expressive sense. But there's definitely more to it than being a Photoshop jockey. The sheer volume of garbage produced by someone's nephew who pirated CS and thinks they're a designer should attest to that (see /r/CrappyDesign). And, true enough, at the beginning of your career there will be a lot of people giving you very specific direction. But as you get further into your career, you will be in a position to have greater and greater creative control. [NEWLINE] [NEWLINE] In the mean time, even if your day job doesn't satisfy your creativity, seek out other opportunities. Freelance whenever possible, and don't be afraid to professionally and firmly steer a client away from questionable design choices. There are also many worthy non-profits that need people to donate their design services (the one exception to "never do free work"). In my experience, these organizations are often willing to give you great creative freedom and are grateful and gracious to boot.</s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> I was 17, I got drunk at a party and was dancing with this guy. We kissed a few times, and he repeatedly asked me<mask> I wanted to have sex. Every time he asked, I said no. [NEWLINE] [NEWLINE] After he handed me a few more drinks, and I was<mask> drunk that I couldn't walk, he took me to a bedroom and told me to get naked. I was<mask> drunk that I had no idea<mask> was going on,<mask> I did<mask> he told me to. Then he laid down on the bed and told me to climb on top. Again, I simply did<mask> I was told to do<mask> I didn't know<mask> was going on. He then proceeded to tell me<mask> to do, and<mask> to move. Being a drunk, virginal 17 year old, I simply followed his directions. [NEWLINE] [NEWLINE] When a friend burst into the room, screaming at me, and asked me<mask> I was doing, THAT was the moment I realized I was having sex. I had no idea<mask> was even happening before she came in the room. [NEWLINE] [NEWLINE] I struggled internally dealing with<mask> happened to me that night for a year before my boyfriend told me that I had been raped after telling him that story. I had no idea that that is<mask> had happened to me<mask> I wasn't familiar with the term "date rape" at the time. I had struggled for a year, thinking that<mask> I had followed this "friend's" instructions, and done everything that he told me to do, that it somehow nullified my refusals to his sexual advances from earlier in the night. [NEWLINE] [NEWLINE] I understand that in my story, I was over intoxicated,<mask> by your logic,<mask> I didn't say no or refuse to follow his directions, one could say that I had "eventually given consent" which I absolutely did not. Being coerced into having sex is coercion no matter<mask> YOU feel about it.<mask> OP feels that he was violated, he's got every right to those feelings. He shouldn't have to deal with people shaming him or trying to invalidate his claims.</s>
Label encoding: <s>When I was 17, I got drunk at a party and was dancing with this guy. We kissed a few times, and he repeatedly asked me if I wanted to have sex. Every time he asked, I said no. [NEWLINE] [NEWLINE] After he handed me a few more drinks, and I was so drunk that I couldn't walk, he took me to a bedroom and told me to get naked. I was so drunk that I had no idea what was going on, so I did what he told me to. Then he laid down on the bed and told me to climb on top. Again, I simply did what I was told to do because I didn't know what was going on. He then proceeded to tell me what to do, and how to move. Being a drunk, virginal 17 year old, I simply followed his directions. [NEWLINE] [NEWLINE] When a friend burst into the room, screaming at me, and asked me what I was doing, THAT was the moment I realized I was having sex. I had no idea what was even happening before she came in the room. [NEWLINE] [NEWLINE] I struggled internally dealing with what happened to me that night for a year before my boyfriend told me that I had been raped after telling him that story. I had no idea that that is what had happened to me because I wasn't familiar with the term "date rape" at the time. I had struggled for a year, thinking that because I had followed this "friend's" instructions, and done everything that he told me to do, that it somehow nullified my refusals to his sexual advances from earlier in the night. [NEWLINE] [NEWLINE] I understand that in my story, I was over intoxicated, but by your logic, because I didn't say no or refuse to follow his directions, one could say that I had "eventually given consent" which I absolutely did not. Being coerced into having sex is coercion no matter how YOU feel about it. If OP feels that he was violated, he's got every right to those feelings. He shouldn't have to deal with people shaming him or trying to invalidate his claims.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0858, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>the terms'reader' and 'gamer' have two different meanings. One is the obvious meaning that they occasionally play games or occasionally read.<mask> the second is the label meaning those things are their hobby. [NEWLINE] [NEWLINE] My brother is the first kind of reader in that he occasionally reads books assigned at school. I am the 2nd kind of reader in that I would check out 20 books at a time from the library and need to go back a week or two later<mask> I had devoured them all<mask> I was a kid.<mask> I got insanely excited for those scholastic order forms in school (almost better than christmas.)<mask> my favorite place in the world is a book store. [NEWLINE] [NEWLINE] My sister is the first kind of gamer<mask> she occasionally plays candy crush on her phone. I am the second tyoe of gamer<mask> I spent hours a week playing them<mask> a kid.<mask> my first memories are sneaking out of my room at night to watch my dad beat the next temple in OOT or get the next star in super mario bros 64.<mask> I have 5 different systems and a shelf full of games. [NEWLINE] [NEWLINE] My brother doesn't consider reading a hobby, I've been able to find one series he actually likes and he still will only read it<mask> he has nothing else to do. My sister does not consider gaming a hobby<mask> she only does it<mask> bored.<mask> yes, by the general definitions of the terms 'gamer' and'reader' they meet the criteria.<mask> words can have multiple meaning. And neither meet the criteria for ether's ulterior definition. [NEWLINE] [NEWLINE] Those things aren't hobbies for them. They aren't things they do<mask> an end<mask> rather<mask> a means to an end. I set time out of my day to do both of those things, they only do them<mask> they have nothing else to do. They want to cure their temporary boredom or finish an assignment. I want know<mask> Alex Rider's mission will go or<mask> choice shepherd will have to make next. And That is the difference. [NEWLINE] [NEWLINE] Do you not believe that words can have multiple meanings? [NEWLINE] </s>
Label encoding: <s>the terms'reader' and 'gamer' have two different meanings. One is the obvious meaning that they occasionally play games or occasionally read. But the second is the label meaning those things are their hobby. [NEWLINE] [NEWLINE] My brother is the first kind of reader in that he occasionally reads books assigned at school. I am the 2nd kind of reader in that I would check out 20 books at a time from the library and need to go back a week or two later because I had devoured them all when I was a kid. Because I got insanely excited for those scholastic order forms in school (almost better than christmas.) Because my favorite place in the world is a book store. [NEWLINE] [NEWLINE] My sister is the first kind of gamer because she occasionally plays candy crush on her phone. I am the second tyoe of gamer because I spent hours a week playing them as a kid. Because my first memories are sneaking out of my room at night to watch my dad beat the next temple in OOT or get the next star in super mario bros 64. Because I have 5 different systems and a shelf full of games. [NEWLINE] [NEWLINE] My brother doesn't consider reading a hobby, I've been able to find one series he actually likes and he still will only read it if he has nothing else to do. My sister does not consider gaming a hobby because she only does it when bored. So yes, by the general definitions of the terms 'gamer' and'reader' they meet the criteria. But words can have multiple meaning. And neither meet the criteria for ether's ulterior definition. [NEWLINE] [NEWLINE] Those things aren't hobbies for them. They aren't things they do as an end but rather as a means to an end. I set time out of my day to do both of those things, they only do them if they have nothing else to do. They want to cure their temporary boredom or finish an assignment. I want know how Alex Rider's mission will go or what choice shepherd will have to make next. And That is the difference. [NEWLINE] [NEWLINE] Do you not believe that words can have multiple meanings? [NEWLINE] </s>
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Part of the problem is that we have a scenario like this: [NEWLINE] [NEWLINE] Kid one grabs all the toys and won't share.  The other kids are angry<mask> they don't have toys, and there's a lot of arguing.  Kid one says, "We're going to keep fighting<mask><mask><mask> we talk about who has more toys. <mask>, lets not mention it again". [NEWLINE] [NEWLINE] <mask> effective is that likely to be? [NEWLINE] [NEWLINE] The problem is that *even<mask> we all suddenly lost the perception to distinguish between races* there would be huge inequality. [NEWLINE] [NEWLINE] In America, money=opportunity.  You are exposed to better educational materials from birth on.  Your schools are better.  Your parents set better expectations and can help more.  You are provided better lessons and enrichment.  You get to experience more of the world first hand, through travel, cultural events, even better cable stations.  You have a safe place to do your homework, and you do it on a full stomach.  You don't need to get a job to put food on the table.  You can go to a better college.  You have connections from your college, from your parent's college friends, from their co-workers and neighbors who can all help you get a job. [NEWLINE] [NEWLINE] Yeah, institutional racism makes things worse,<mask> even without it, economic disparities would still make things uneven. [NEWLINE] [NEWLINE] <mask> for your point about being a minority in your town, no, it doesn't give you insight, sorry.  You don't get pulled over for driving<mask> black.  You don't worry that the cops are going to shoot you for no good reason.  You don't have to contend with getting your resume thrown away<mask> your name sounds too black, or your essay being graded lower<mask> of your name.  You don't live with the assumption that you're a lazy, welfare mooching thug simply<mask> the color of your skin.  Yes, there are some things that you experience<mask> a minority,<mask> not the soul-crushing realities of being black in America.</s>
Label encoding: <s>Part of the problem is that we have a scenario like this: [NEWLINE] [NEWLINE] Kid one grabs all the toys and won't share.  The other kids are angry because they don't have toys, and there's a lot of arguing.  Kid one says, "We're going to keep fighting as long as we talk about who has more toys.  So, lets not mention it again". [NEWLINE] [NEWLINE] How effective is that likely to be? [NEWLINE] [NEWLINE] The problem is that *even if we all suddenly lost the perception to distinguish between races* there would be huge inequality. [NEWLINE] [NEWLINE] In America, money=opportunity.  You are exposed to better educational materials from birth on.  Your schools are better.  Your parents set better expectations and can help more.  You are provided better lessons and enrichment.  You get to experience more of the world first hand, through travel, cultural events, even better cable stations.  You have a safe place to do your homework, and you do it on a full stomach.  You don't need to get a job to put food on the table.  You can go to a better college.  You have connections from your college, from your parent's college friends, from their co-workers and neighbors who can all help you get a job. [NEWLINE] [NEWLINE] Yeah, institutional racism makes things worse, but even without it, economic disparities would still make things uneven. [NEWLINE] [NEWLINE] As for your point about being a minority in your town, no, it doesn't give you insight, sorry.  You don't get pulled over for driving while black.  You don't worry that the cops are going to shoot you for no good reason.  You don't have to contend with getting your resume thrown away because your name sounds too black, or your essay being graded lower because of your name.  You don't live with the assumption that you're a lazy, welfare mooching thug simply because the color of your skin.  Yes, there are some things that you experience as a minority, but not the soul-crushing realities of being black in America.</s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your comment about a lineman [ENDQ] [NEWLINE] **<mask> in my post did I mention anything about linemen?** [NEWLINE] [NEWLINE] Seems to me you can recognise the lack of technique in your sport,<mask><mask> I never explicitly named the position(s) I was talking about. [NEWLINE] [NEWLINE] Some guy called "Richard Seymour" was the highest played player in the NFL? Well bless his heart. TIL. [NEWLINE] [NEWLINE] Get back to me<mask> he can play non-stop for 90 minutes, whilst<mask> having the tactical nous to play both offense and defense. [NEWLINE] [NEWLINE] I must admit that I do like that bit in American football<mask> the fat sumo-wrestler-type blokes push each other to the ground, and then one guy runs really fast. That bit<mask> a guy runs really fast and then catches a ball is<mask> kinda cool. [NEWLINE] [NEWLINE] <mask> there's very little "skill" involved in it. It's just a big muscly guy with a probable steroid addiction running really fast and possibly using his hands to catch a ball. [NEWLINE] [NEWLINE] I could catch a ball<mask> I was five years old.<mask> I lived for five hundred years I couldn't caress a ball like Messi. [NEWLINE] [NEWLINE] Again, the only player on an NFL field who is required to have any kind of passing ability is the QB. [NEWLINE] [NEWLINE] Lionel Messi wouldn't succeed at any American sport,<mask> American "sports fans" have a bizarre obsession with 6'8" 300-pound "athletes", whereas the rest of the world admires skilled "players". [NEWLINE] [NEWLINE] This is<mask> you fail. [NEWLINE] [NEWLINE] The way that Jozy Altidore's absence was seen<mask> a reason that the US lost to Belgium today by many American fans is a prime example of this. He's a big powerful unit,<mask> he has no technical ability or finesse. He's a joke. [NEWLINE] [NEWLINE] To put it another way, give me the talent of a Sachin Tendulkar over the brawn of a Mark McGwire or Sammy Sosa any day of the week. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Your comment about a lineman [ENDQ] [NEWLINE] ** Where in my post did I mention anything about linemen?** [NEWLINE] [NEWLINE] Seems to me you can recognise the lack of technique in your sport, even though I never explicitly named the position(s) I was talking about. [NEWLINE] [NEWLINE] Some guy called "Richard Seymour" was the highest played player in the NFL? Well bless his heart. TIL. [NEWLINE] [NEWLINE] Get back to me when he can play non-stop for 90 minutes, whilst also having the tactical nous to play both offense and defense. [NEWLINE] [NEWLINE] I must admit that I do like that bit in American football where the fat sumo-wrestler-type blokes push each other to the ground, and then one guy runs really fast. That bit where a guy runs really fast and then catches a ball is also kinda cool. [NEWLINE] [NEWLINE] But there's very little "skill" involved in it. It's just a big muscly guy with a probable steroid addiction running really fast and possibly using his hands to catch a ball. [NEWLINE] [NEWLINE] I could catch a ball when I was five years old. If I lived for five hundred years I couldn't caress a ball like Messi. [NEWLINE] [NEWLINE] Again, the only player on an NFL field who is required to have any kind of passing ability is the QB. [NEWLINE] [NEWLINE] Lionel Messi wouldn't succeed at any American sport, because American "sports fans" have a bizarre obsession with 6'8" 300-pound "athletes", whereas the rest of the world admires skilled "players". [NEWLINE] [NEWLINE] This is why you fail. [NEWLINE] [NEWLINE] The way that Jozy Altidore's absence was seen as a reason that the US lost to Belgium today by many American fans is a prime example of this. He's a big powerful unit, but he has no technical ability or finesse. He's a joke. [NEWLINE] [NEWLINE] To put it another way, give me the talent of a Sachin Tendulkar over the brawn of a Mark McGwire or Sammy Sosa any day of the week. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Isn't really true. Just<mask> other people have said here, there are a lot of CEO's who make more money than the President of the US does and<mask> there's anything I know about power it's that money is directly related to the amount of power you can have. [ENDQ] [NEWLINE] <mask><mask> then, you don't know much about power. [NEWLINE] [NEWLINE] Here's all the things that the president can do legally to American citizens, that Bill Gates or Warren Buffett can't: [NEWLINE] [NEWLINE] * He can have them investigated by the IRS, the SEC, the FDA, the ATF, the Treasury, the FBI, the CIA, the NSA, the DEA, etc.  He can have their phones tapped, their records audited, their emails copied (and leaked), their equipment seized, etc, etc, etc. [NEWLINE] [NEWLINE] * He can have their families slandered wholesale, their reputations besmirched.  Even the hint that an investigation is forthcoming can completely ruin the CEO's livelihood. [NEWLINE] [NEWLINE] * He can have their family and friends indefinitely detained by declaring them terrorists and have them conveniently die in captivity after being extraordinarily renditioned. [NEWLINE] [NEWLINE] * He can direct government contracts towards their rivals [NEWLINE] [NEWLINE] * He can have his minions drop hints to the media, and based on his party affiliation, certain parts of the media will pick up the hints and run with them [NEWLINE] [NEWLINE] * Ultimately, he could seize their assets by claiming some sort of emergency. [NEWLINE] [NEWLINE] No CEO can do any of those things legally, and anyone who tried illegally would be put down like a rabid dog. [NEWLINE] [NEWLINE] Now, the Pope can't do those things either -<mask> he could,<mask> he wanted to, spark massive riots and protests and shut down a significant chunk of the global economy. [NEWLINE] [NEWLINE] [NEWLINE] Money is absolutely a form of power.  <mask><mask> is large scale reverence and obedience, and the Pope and the President have the latter two in spades. [NEWLINE] [NEWLINE] [NEWLINE] P.S. _"scandals with the NSA... Obama didn't know anything about."_  You're adorable. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Isn't really true. Just as other people have said here, there are a lot of CEO's who make more money than the President of the US does and if there's anything I know about power it's that money is directly related to the amount of power you can have. [ENDQ] [NEWLINE] IMO then, you don't know much about power. [NEWLINE] [NEWLINE] Here's all the things that the president can do legally to American citizens, that Bill Gates or Warren Buffett can't: [NEWLINE] [NEWLINE] * He can have them investigated by the IRS, the SEC, the FDA, the ATF, the Treasury, the FBI, the CIA, the NSA, the DEA, etc.  He can have their phones tapped, their records audited, their emails copied (and leaked), their equipment seized, etc, etc, etc. [NEWLINE] [NEWLINE] * He can have their families slandered wholesale, their reputations besmirched.  Even the hint that an investigation is forthcoming can completely ruin the CEO's livelihood. [NEWLINE] [NEWLINE] * He can have their family and friends indefinitely detained by declaring them terrorists and have them conveniently die in captivity after being extraordinarily renditioned. [NEWLINE] [NEWLINE] * He can direct government contracts towards their rivals [NEWLINE] [NEWLINE] * He can have his minions drop hints to the media, and based on his party affiliation, certain parts of the media will pick up the hints and run with them [NEWLINE] [NEWLINE] * Ultimately, he could seize their assets by claiming some sort of emergency. [NEWLINE] [NEWLINE] No CEO can do any of those things legally, and anyone who tried illegally would be put down like a rabid dog. [NEWLINE] [NEWLINE] Now, the Pope can't do those things either - but he could, if he wanted to, spark massive riots and protests and shut down a significant chunk of the global economy. [NEWLINE] [NEWLINE] [NEWLINE] Money is absolutely a form of power.   But so is large scale reverence and obedience, and the Pope and the President have the latter two in spades. [NEWLINE] [NEWLINE] [NEWLINE] P.S. _"scandals with the NSA... Obama didn't know anything about."_  You're adorable. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ultimately it's about more things having to go right/be compensated for than anything else.  It is just a fact that prejudice still exists. <mask> that prejudice at different levels and places just require more work or luck for some groups to go from a to b to c etc than others. [NEWLINE] [NEWLINE] Obviously choices play a huge part and economic level/class arguably plays a more important role than simply race<mask> it's not unreasonable to<mask><mask> the cards can be stacked against some people due to their race or ethnicity. [NEWLINE] [NEWLINE] Examples of extra hindrances: [NEWLINE] [NEWLINE] Drug policies typically target poor minority communities. [NEWLINE] [NEWLINE] A white man with a criminal record is more likely to get the job than a.black man with no record<mask> the same qualifications. [NEWLINE] [NEWLINE] A person with an<mask>ian sounding name is about half<mask> likely to get an interview<mask> a white sounding name with the same qualifications. [NEWLINE] [NEWLINE] Large institutions such<mask> bank of america have specifically targeted poor minority communities with shifty practices. [NEWLINE] [NEWLINE] People with more unique names (more common in the black community) are at a disadvantage. [NEWLINE] [NEWLINE] Major universities have de facto quotas for<mask>ian students to "preserve the brand." [NEWLINE] [NEWLINE] People from underprivileged asian groups (such as cambodians and hmong) receive little attention and support<mask> they are lumped into the successful asian paradigm<mask> having among the highest levels of poverty. [NEWLINE] [NEWLINE] The traits that make a white man seem dominant in a business environment are disliked<mask> displayed by asian men<mask> it doesn't fit the expectation of<mask> they are supposed to act (eg quiet and submissive). [NEWLINE] [NEWLINE] Aspersions are cast on minorities as "token" or "affirmative action" benefactors without proof.  These assumptions undermine the work they put in and skills<mask> devaluing their ability.  A white man picks himself up by his bootstraps; a black man only got<mask> he is cause of white guilt and political correctness. [NEWLINE] [NEWLINE] Stereotype threat causes some to underperform. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] EDIT: a word</s>
Label encoding: <s>Ultimately it's about more things having to go right/be compensated for than anything else.  It is just a fact that prejudice still exists.  So that prejudice at different levels and places just require more work or luck for some groups to go from a to b to c etc than others. [NEWLINE] [NEWLINE] Obviously choices play a huge part and economic level/class arguably plays a more important role than simply race but it's not unreasonable to argue that the cards can be stacked against some people due to their race or ethnicity. [NEWLINE] [NEWLINE] Examples of extra hindrances: [NEWLINE] [NEWLINE] Drug policies typically target poor minority communities. [NEWLINE] [NEWLINE] A white man with a criminal record is more likely to get the job than a.black man with no record but the same qualifications. [NEWLINE] [NEWLINE] A person with an asian sounding name is about half as likely to get an interview as a white sounding name with the same qualifications. [NEWLINE] [NEWLINE] Large institutions such as bank of america have specifically targeted poor minority communities with shifty practices. [NEWLINE] [NEWLINE] People with more unique names (more common in the black community) are at a disadvantage. [NEWLINE] [NEWLINE] Major universities have de facto quotas for asian students to "preserve the brand." [NEWLINE] [NEWLINE] People from underprivileged asian groups (such as cambodians and hmong) receive little attention and support because they are lumped into the successful asian paradigm despite having among the highest levels of poverty. [NEWLINE] [NEWLINE] The traits that make a white man seem dominant in a business environment are disliked when displayed by asian men because it doesn't fit the expectation of how they are supposed to act (eg quiet and submissive). [NEWLINE] [NEWLINE] Aspersions are cast on minorities as "token" or "affirmative action" benefactors without proof.  These assumptions undermine the work they put in and skills while devaluing their ability.  A white man picks himself up by his bootstraps; a black man only got where he is cause of white guilt and political correctness. [NEWLINE] [NEWLINE] Stereotype threat causes some to underperform. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] EDIT: a word</s>
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I feel pretty bad for posting this,<mask> I know I'll be called out<mask> a bigot. That's<mask> I'm hoping some of you can CMV. [ENDQ] [NEWLINE] I feel bad for you too...<mask> you believe isn't uncommon, unfortunately. [NEWLINE] [NEWLINE] [STARTQ] It's practically a stereotype that Asian and Indian children are going to grow up too be doctors or engineers! [ENDQ] [NEWLINE] Whew, that was quick, you changed you view for me :-) [NEWLINE] [NEWLINE] <mask> that's the answer for<mask> "blacks just need to work harder". Stereotypes. Asians and (some) Indians are stereotyped<mask> super smart, and more capable in technical fields. Blacks are often stereotyped in the opposite direction: lazy, unmotivated, comfortable in their lower social status, etc. [NEWLINE] [NEWLINE] The reality is that sometimes racism comes down to social/cultural inertia: people think and do things a certain way<mask> That's<mask> their friends/family do. Naming their kids in a fashion that easily identifies them<mask> "black" (Cheniqua [<mask> ], Jamal, Fantasia, LaShondra, etc). Speaking in "Ebonics", which sounds uneducated to myself and others. [NEWLINE] [NEWLINE] Every person is an individual,<mask>, with different thoughts, dreams, and opinions too. Stereotyping is wrong<mask> it takes a perceived "average trait" about a group and makes it the norm for everyone. [NEWLINE] [NEWLINE] I won't deny racism in more pure form still exists: blacks and whites (and browns, and every other color) still struggle with it. Whites tip toe around it mostly,<mask> many blacks are pretty open about it...<mask> you mentioned about "not being the right minority" can ring true.<mask> that ignorance and foolishness, it can't be extrapolated out to an entire race of people. That single woman was a fool and should be called out for it. Thats<mask> we do this: attack it on a case by case basis.<mask> you reinforce your on preconceptions, you're part of the problem, not the solution.</s>
Label encoding: <s> [STARTQ] So I feel pretty bad for posting this, because I know I'll be called out as a bigot. That's why I'm hoping some of you can CMV. [ENDQ] [NEWLINE] I feel bad for you too... what you believe isn't uncommon, unfortunately. [NEWLINE] [NEWLINE] [STARTQ] It's practically a stereotype that Asian and Indian children are going to grow up too be doctors or engineers! [ENDQ] [NEWLINE] Whew, that was quick, you changed you view for me :-) [NEWLINE] [NEWLINE] Because that's the answer for why "blacks just need to work harder". Stereotypes. Asians and (some) Indians are stereotyped as super smart, and more capable in technical fields. Blacks are often stereotyped in the opposite direction: lazy, unmotivated, comfortable in their lower social status, etc. [NEWLINE] [NEWLINE] The reality is that sometimes racism comes down to social/cultural inertia: people think and do things a certain way because That's how their friends/family do. Naming their kids in a fashion that easily identifies them as "black" (Cheniqua [ so ], Jamal, Fantasia, LaShondra, etc). Speaking in "Ebonics", which sounds uneducated to myself and others. [NEWLINE] [NEWLINE] Every person is an individual, though, with different thoughts, dreams, and opinions too. Stereotyping is wrong because it takes a perceived "average trait" about a group and makes it the norm for everyone. [NEWLINE] [NEWLINE] I won't deny racism in more pure form still exists: blacks and whites (and browns, and every other color) still struggle with it. Whites tip toe around it mostly, while many blacks are pretty open about it... what you mentioned about "not being the right minority" can ring true. But that ignorance and foolishness, it can't be extrapolated out to an entire race of people. That single woman was a fool and should be called out for it. Thats how we do this: attack it on a case by case basis. If you reinforce your on preconceptions, you're part of the problem, not the solution.</s>
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] got you laid [ENDQ] [NEWLINE] I'm looking to marry and settle down; sex is the least of my concerns. The problem is that I'm a young person in a first-world country and the global bankers think that I'm disposable. Plus, I live in a society that demands hard work and I'll be demonized<mask> I decide that greed isn't in my best interests. I'd gladly remain a virgin<mask> the love of my life was asexual. [NEWLINE] [NEWLINE] [STARTQ] just like we normally treat both the symptoms of a disease and the disease itself [ENDQ] [NEWLINE] The problem is that with this disease of disadvantaging the vast majority of people must be cured or else one group of extremists will replace another. [NEWLINE] [NEWLINE] [STARTQ] the issues you lump together seem very heterogeneous [ENDQ] [NEWLINE] At the heart of them, the issues all boil down to the fact that the global system is not only unfair<mask> perverse; countries that invest in their citizens [like Canada or much of Europe]( [URL] /) not only are unable to help the US see the light thanks to the brainwashed media and pseudo-political system,<mask> are encouraged to fix<mask> ain't broken thanks to the great hoax of the "European" (actually has its origin in the 2000s housing bust in Florida and elsewhere) "debt" (explain<mask> countries like Canada, Australia, Sweden and Denmark not only aren't utopias<mask> destroying<mask> makes them<mask> much better off than the US<mask><mask><mask> smaller, more homogeneous populations, AAA credit ratings, and excellent economic fundamentals. It's about lowering the standards of the average worker<mask> that people feel pressure to indenture themselves) "crisis" (Spain had a debt-to-GDP ratio that was lower than the US before the downturn and only suffered<mask>, like all civilized countries, it extended aid to its unemployed). Western civilization has largely been a history of increased material comfort, and<mask> things have gone backwards it has been due to war or plague, not due to politicians (openly in the Nordic countries, for instance) deciding that their countries are too developed and too educated.</s>
Label encoding: <s> [STARTQ] got you laid [ENDQ] [NEWLINE] I'm looking to marry and settle down; sex is the least of my concerns. The problem is that I'm a young person in a first-world country and the global bankers think that I'm disposable. Plus, I live in a society that demands hard work and I'll be demonized if I decide that greed isn't in my best interests. I'd gladly remain a virgin if the love of my life was asexual. [NEWLINE] [NEWLINE] [STARTQ] just like we normally treat both the symptoms of a disease and the disease itself [ENDQ] [NEWLINE] The problem is that with this disease of disadvantaging the vast majority of people must be cured or else one group of extremists will replace another. [NEWLINE] [NEWLINE] [STARTQ] the issues you lump together seem very heterogeneous [ENDQ] [NEWLINE] At the heart of them, the issues all boil down to the fact that the global system is not only unfair but perverse; countries that invest in their citizens [like Canada or much of Europe]( [URL] /) not only are unable to help the US see the light thanks to the brainwashed media and pseudo-political system, but are encouraged to fix what ain't broken thanks to the great hoax of the "European" (actually has its origin in the 2000s housing bust in Florida and elsewhere) "debt" (explain why countries like Canada, Australia, Sweden and Denmark not only aren't utopias but destroying what makes them so much better off than the US in spite of smaller, more homogeneous populations, AAA credit ratings, and excellent economic fundamentals. It's about lowering the standards of the average worker so that people feel pressure to indenture themselves) "crisis" (Spain had a debt-to-GDP ratio that was lower than the US before the downturn and only suffered because, like all civilized countries, it extended aid to its unemployed). Western civilization has largely been a history of increased material comfort, and when things have gone backwards it has been due to war or plague, not due to politicians (openly in the Nordic countries, for instance) deciding that their countries are too developed and too educated.</s>
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've already explained my reasoning throughly. Multiple times. It's your burden to explain<mask> it is not racist, not mine to defend my view in every one of your comments.<mask> you can't reasonably argue your point without asking 20 clarifying question and changing the topic, than you should go to another subreddit. [NEWLINE] [NEWLINE] [STARTQ] <mask> I mean is,<mask> you don't know<mask> is behind someone's attraction, their lack of attraction for you should not be taken in any negative way,<mask><mask><mask> it's about, unless they don't actually go on and insult you. "I am not attracted to you" does not equal "you are ugly"<mask> you may find people I am attracted to ugly and be relieved you are not in that group. [ENDQ] [NEWLINE] The next time you respond. Just delete any sentence that has the word "attraction" in it<mask> it's irrelevant. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> there isn't any insult there,<mask> would it be racist? Simply for mentioning race? [ENDQ] [NEWLINE] I'm going to answer this<mask> you aren't reading any of my other comments. [NEWLINE] [NEWLINE] [STARTQ] hy do<mask><mask> this is racist? An individual in this scenario is outright say he/she does not wish to speak to someone on the basis of race and we're to assume he/she will ignore comments from individuals belonging to whatever group he/she is excluding. I find this exclusion to be racist and the public shamelessness involved in saying that you don't want to talk to or be contacted by [insert [racial/ethnic group here] is racist. I'd be appalled<mask> I heard someone say this to me in any other circumstance.<mask> do we make an exception for dating? [ENDQ] [NEWLINE] [STARTQ] In the slavery example, it is a good economic decision only<mask> you exclude the needs of slaves.<mask> part of the society suffers for benefit of others, it is clearly NOT a good decision. By me saying I am not attracted to someone, no one suffers. [ENDQ] [NEWLINE] You shame someone which makes them feel bad.<mask>?<mask> they were born a certain skin color.</s>
Label encoding: <s>I've already explained my reasoning throughly. Multiple times. It's your burden to explain why it is not racist, not mine to defend my view in every one of your comments. If you can't reasonably argue your point without asking 20 clarifying question and changing the topic, than you should go to another subreddit. [NEWLINE] [NEWLINE] [STARTQ] What I mean is, since you don't know what is behind someone's attraction, their lack of attraction for you should not be taken in any negative way, regardless of what it's about, unless they don't actually go on and insult you. "I am not attracted to you" does not equal "you are ugly" since you may find people I am attracted to ugly and be relieved you are not in that group. [ENDQ] [NEWLINE] The next time you respond. Just delete any sentence that has the word "attraction" in it because it's irrelevant. [NEWLINE] [NEWLINE] [STARTQ] So since there isn't any insult there, why would it be racist? Simply for mentioning race? [ENDQ] [NEWLINE] I'm going to answer this since you aren't reading any of my other comments. [NEWLINE] [NEWLINE] [STARTQ] hy do I think this is racist? An individual in this scenario is outright say he/she does not wish to speak to someone on the basis of race and we're to assume he/she will ignore comments from individuals belonging to whatever group he/she is excluding. I find this exclusion to be racist and the public shamelessness involved in saying that you don't want to talk to or be contacted by [insert [racial/ethnic group here] is racist. I'd be appalled if I heard someone say this to me in any other circumstance. Why do we make an exception for dating? [ENDQ] [NEWLINE] [STARTQ] In the slavery example, it is a good economic decision only if you exclude the needs of slaves. If part of the society suffers for benefit of others, it is clearly NOT a good decision. By me saying I am not attracted to someone, no one suffers. [ENDQ] [NEWLINE] You shame someone which makes them feel bad. Why? Because they were born a certain skin color.</s>
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure this is actually changing your view directly,<mask><mask><mask> it's important to eliminate the bi-modal separation of hetero/homo sexuality. It's not really an either-or thing,<mask> rather a spectrum that extends to pure homo/heterosexuality. [NEWLINE] [NEWLINE] **For the sake of clarity/simplification imagine a scale ranging from minus -50 to +50.** [NEWLINE] [NEWLINE] **<mask> you have a negative score that means you are more heterosexual than homosexual.  The more negative score you have the less likely you are to be attracted to and/or engage in any homosexual activity.** [NEWLINE] [NEWLINE] **Similarly,<mask> you have a positive score you are less likely to be attracted to the opposite sex.  The higher the score the stronger you sexual preference becomes.** [NEWLINE] [NEWLINE] Now - this part is important - **imagine instead that your'score' was<mask> a distribution like a Gaussian curve (a bell curve)** that has two IMPORTANT values that describe<mask> you own score looks like: the AVERAGE (<mask> the peak is on the scale of -50, to +50) and the WIDTH (<mask> much spread/wiggle room there is in your preference). [NEWLINE] [NEWLINE] To me this is a much better way of looking a sexuality. [NEWLINE] [NEWLINE] This means that you can have ALL types of people. [NEWLINE] [NEWLINE] <mask> some one has a average of 0,<mask> a width of 75, that means that on average they are attracted to both sexes,<mask> at any given time they could be more or less attracted to the opposite/same sex. [NEWLINE] [NEWLINE] <mask> someone has an average of (-)45,<mask> with a width of 5, that means that they strongly performance partners of the (opposite) same sex. [NEWLINE] [NEWLINE] Now using this model to describe homo sexuality.  My interpretation goes like this: [NEWLINE] [NEWLINE] It's likely that your average is fixed genetically and/or during the ultra early stages of life<mask> your width can change throughout your life (<mask> it doesn't change that much). [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I'm not sure this is actually changing your view directly, but I think it's important to eliminate the bi-modal separation of hetero/homo sexuality. It's not really an either-or thing, but rather a spectrum that extends to pure homo/heterosexuality. [NEWLINE] [NEWLINE] **For the sake of clarity/simplification imagine a scale ranging from minus -50 to +50.** [NEWLINE] [NEWLINE] ** If you have a negative score that means you are more heterosexual than homosexual.  The more negative score you have the less likely you are to be attracted to and/or engage in any homosexual activity.** [NEWLINE] [NEWLINE] **Similarly, if you have a positive score you are less likely to be attracted to the opposite sex.  The higher the score the stronger you sexual preference becomes.** [NEWLINE] [NEWLINE] Now - this part is important - **imagine instead that your'score' was ALSO a distribution like a Gaussian curve (a bell curve)** that has two IMPORTANT values that describe what you own score looks like: the AVERAGE ( where the peak is on the scale of -50, to +50) and the WIDTH ( how much spread/wiggle room there is in your preference). [NEWLINE] [NEWLINE] To me this is a much better way of looking a sexuality. [NEWLINE] [NEWLINE] This means that you can have ALL types of people. [NEWLINE] [NEWLINE] If some one has a average of 0, but a width of 75, that means that on average they are attracted to both sexes, but at any given time they could be more or less attracted to the opposite/same sex. [NEWLINE] [NEWLINE] If someone has an average of (-)45, but with a width of 5, that means that they strongly performance partners of the (opposite) same sex. [NEWLINE] [NEWLINE] Now using this model to describe homo sexuality.  My interpretation goes like this: [NEWLINE] [NEWLINE] It's likely that your average is fixed genetically and/or during the ultra early stages of life but your width can change throughout your life ( although it doesn't change that much). [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1006, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for responding! I do have some objections to this comment,<mask>. [NEWLINE] [NEWLINE] "<mask><mask> I called you a bitch right now, is it okay<mask> I didn't call you a female dog?<mask> I called you a dick, would it be okay? I'm not really calling you a penis, after all." [NEWLINE] [NEWLINE] Well,<mask> you called me a bitch, I would interpret that<mask> a criticism of me in particular, and not a criticism of female dogs,<mask> you are using that word in a certain context.<mask> you were to use that word in the context of a discussion about gender in dogs, I would interpret it differently. [NEWLINE] Similarly,<mask> you called me a dick,  I would interpret that<mask> a criticism of me in particular and not a criticism of all males.<mask> you were to use the word dick<mask> discussing urology, I would interpret that word to mean an actual penis. [NEWLINE] [NEWLINE] "Just<mask> a word is "divorced from its original meaning" doesn't mean that the use of the word is okay." [NEWLINE] [NEWLINE] A major criticism of using the word Gay<mask> an insult is that it demeans all Gay people. Recognizing that the semantic shell can carry different meanings refutes this particular objection,<mask> calling one meaning bad does not necessarily tarnish every other meaning in the same semantic shell. [NEWLINE] [NEWLINE] "Imagine<mask> the word hetero were to become synonymous with stupid or terrible tomorrow.<mask> would you feel?" [NEWLINE] I would not be offended.<mask> in this hypothetical, using the word hetero<mask> an insult would not necessarily be an attack on straight people, it would just add another meaning to a semantic shell. [NEWLINE] [NEWLINE] "Would you feel like people are calling you terrible simply<mask> of your sexuality?" [NEWLINE] No. And even<mask> I did feel that way, that wouldn't make me correct. [NEWLINE] [NEWLINE] "<mask> that's<mask> gay people take it<mask> people use "gay"<mask> an insult." [NEWLINE] And I can fully sympathize.<mask>, I don't think that is a logical way to take it. [NEWLINE] </s>
Label encoding: <s>Thank you for responding! I do have some objections to this comment, however. [NEWLINE] [NEWLINE] " So if I called you a bitch right now, is it okay because I didn't call you a female dog? If I called you a dick, would it be okay? I'm not really calling you a penis, after all." [NEWLINE] [NEWLINE] Well, if you called me a bitch, I would interpret that as a criticism of me in particular, and not a criticism of female dogs, because you are using that word in a certain context. If you were to use that word in the context of a discussion about gender in dogs, I would interpret it differently. [NEWLINE] Similarly, if you called me a dick,  I would interpret that as a criticism of me in particular and not a criticism of all males. If you were to use the word dick when discussing urology, I would interpret that word to mean an actual penis. [NEWLINE] [NEWLINE] "Just because a word is "divorced from its original meaning" doesn't mean that the use of the word is okay." [NEWLINE] [NEWLINE] A major criticism of using the word Gay as an insult is that it demeans all Gay people. Recognizing that the semantic shell can carry different meanings refutes this particular objection, because calling one meaning bad does not necessarily tarnish every other meaning in the same semantic shell. [NEWLINE] [NEWLINE] "Imagine if the word hetero were to become synonymous with stupid or terrible tomorrow. How would you feel?" [NEWLINE] I would not be offended. Because in this hypothetical, using the word hetero as an insult would not necessarily be an attack on straight people, it would just add another meaning to a semantic shell. [NEWLINE] [NEWLINE] "Would you feel like people are calling you terrible simply because of your sexuality?" [NEWLINE] No. And even if I did feel that way, that wouldn't make me correct. [NEWLINE] [NEWLINE] " Because that's how gay people take it when people use "gay" as an insult." [NEWLINE] And I can fully sympathize. However, I don't think that is a logical way to take it. [NEWLINE] </s>
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That sometimes a small act of violence can prevent more violence and<mask> decrease total violence.<mask> a small act of violence can be more moral than a non-violent act. [ENDQ] [NEWLINE] This claim only makes sense from the perspective of utilitarianism, which you haven't justified.<mask>, even<mask> we grant utilitarianism, we would not say that pacifism is immoral,<mask> rather that in certain situations it results in a net loss of utility. In other situations, from the perspective of utility pacifism might actually be morally obligatory. [NEWLINE] [NEWLINE] Further, one of the greatest criticisms of utilitarianism is that it's impossible to calculate expected utility in the moment of action, and that unintended consequences often outweigh expected utility. Pacifism seems to vibe with this pretty well--it's entirely possible to imagine a version of rule utilitarianism that incorporates this criticism to say that it's better to hedge and avoid violence rather than risk far more devastating unintended consequences. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] That an absolute pacifist relies on others to commit violence for their own continued safety, for example the subduing of a dangerous criminal by police forces may be violent, and that any personal code which requires other persons to break said code in order to be followed cannot be moral. [ENDQ] [NEWLINE] This argument doesn't make any sense. The pacifists doesn't require anyone else to do anything. It's entirely possible to imagine a pacifist who doesn't call the police, has no expectation of a national security force to protect them, and would be willing to martyr themselves in the face of violence.<mask> anything, the fact that people take these actions without interrogating the pacifist's beliefs is just demonstration that people are disrespecting the pacifist, not that the pacifist is relying on violence. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Furthermore that there can be a just war. [ENDQ] [NEWLINE] This is highly contentious and simply begs the question of pacifism in the first place. There's<mask> a significant distinction between war<mask> a necessary evil and a just war. You should probably edit to defend a comprehensive just war theory, rather than asserting this is true.</s>
Label encoding: <s> [STARTQ] That sometimes a small act of violence can prevent more violence and thus decrease total violence. Thus a small act of violence can be more moral than a non-violent act. [ENDQ] [NEWLINE] This claim only makes sense from the perspective of utilitarianism, which you haven't justified. However, even if we grant utilitarianism, we would not say that pacifism is immoral, but rather that in certain situations it results in a net loss of utility. In other situations, from the perspective of utility pacifism might actually be morally obligatory. [NEWLINE] [NEWLINE] Further, one of the greatest criticisms of utilitarianism is that it's impossible to calculate expected utility in the moment of action, and that unintended consequences often outweigh expected utility. Pacifism seems to vibe with this pretty well--it's entirely possible to imagine a version of rule utilitarianism that incorporates this criticism to say that it's better to hedge and avoid violence rather than risk far more devastating unintended consequences. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] That an absolute pacifist relies on others to commit violence for their own continued safety, for example the subduing of a dangerous criminal by police forces may be violent, and that any personal code which requires other persons to break said code in order to be followed cannot be moral. [ENDQ] [NEWLINE] This argument doesn't make any sense. The pacifists doesn't require anyone else to do anything. It's entirely possible to imagine a pacifist who doesn't call the police, has no expectation of a national security force to protect them, and would be willing to martyr themselves in the face of violence. If anything, the fact that people take these actions without interrogating the pacifist's beliefs is just demonstration that people are disrespecting the pacifist, not that the pacifist is relying on violence. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Furthermore that there can be a just war. [ENDQ] [NEWLINE] This is highly contentious and simply begs the question of pacifism in the first place. There's also a significant distinction between war as a necessary evil and a just war. You should probably edit to defend a comprehensive just war theory, rather than asserting this is true.</s>
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My BA is in economics, and I've had many people try to challenge me on very philosophical issues about<mask> economics works (or doesn't work). [NEWLINE] [NEWLINE] <mask> I engage them, I always first point them in the direction of Khan Academy's Economics videos<mask> they are amazing. [NEWLINE] [NEWLINE] <mask> they persist, I might try to answer their query. After that it always breaks down one of three ways: [NEWLINE] [NEWLINE] 1) They have such a low point-of-view of the study of economics that they won't let me get a word in edgewise and I wish I had never given the person the time of day. [NEWLINE] [NEWLINE] 2) I do get to all the problems that they have,<mask> it is often a difficult, combative conversation that I didn't want, seek, or ask for. Often, I am able to convince them that whatever problem they had was an improper judgement/generalization of an entire field of academia.<mask> it doesn't feel satisfying; it doesn't feel like they are more informed; it does not stimulate their curiosity in my field; it often doesn't change<mask> the other person talks about (disparages) my field in the future; and, it wipes me out. And I wish I had never given the person the time of day. [NEWLINE] [NEWLINE] 3) I'm honestly too tired. It's hard work to convince someone of something on short notice. It even harder<mask> you have dedicated large portions of your life to something (or, worse, are something like Jewish or American) and then<mask> you are half drunk be challenged to defend half-formed arguments based on ignorance and hate. And I wish I had never given the person the time of day. [NEWLINE] [NEWLINE] <mask> someone does do the research, put a fraction of the time in that I did to understand something, that means that we are having a debate/argument/dialogue/dialectic in good faith. It means that these situations mentioned above probably will not happen, and we can work together to find "truth" together. And that is truly beautiful.</s>
Label encoding: <s>My BA is in economics, and I've had many people try to challenge me on very philosophical issues about how economics works (or doesn't work). [NEWLINE] [NEWLINE] When I engage them, I always first point them in the direction of Khan Academy's Economics videos because they are amazing. [NEWLINE] [NEWLINE] If they persist, I might try to answer their query. After that it always breaks down one of three ways: [NEWLINE] [NEWLINE] 1) They have such a low point-of-view of the study of economics that they won't let me get a word in edgewise and I wish I had never given the person the time of day. [NEWLINE] [NEWLINE] 2) I do get to all the problems that they have, but it is often a difficult, combative conversation that I didn't want, seek, or ask for. Often, I am able to convince them that whatever problem they had was an improper judgement/generalization of an entire field of academia. But it doesn't feel satisfying; it doesn't feel like they are more informed; it does not stimulate their curiosity in my field; it often doesn't change how the other person talks about (disparages) my field in the future; and, it wipes me out. And I wish I had never given the person the time of day. [NEWLINE] [NEWLINE] 3) I'm honestly too tired. It's hard work to convince someone of something on short notice. It even harder when you have dedicated large portions of your life to something (or, worse, are something like Jewish or American) and then when you are half drunk be challenged to defend half-formed arguments based on ignorance and hate. And I wish I had never given the person the time of day. [NEWLINE] [NEWLINE] When someone does do the research, put a fraction of the time in that I did to understand something, that means that we are having a debate/argument/dialogue/dialectic in good faith. It means that these situations mentioned above probably will not happen, and we can work together to find "truth" together. And that is truly beautiful.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [STARTQ] <mask> then subjects like archeology, macro-evolution and astronomy are not hard-sciences. [ENDQ] [NEWLINE] Ok<mask><mask> I understand<mask> the confusion is. The "<mask> then" part is to indicate that<mask> we follow this reasoning (the testable prediction) then we come to this hard-to-defend conclusion (i.e. archeology etc are not hard-sciences).  It is not my stance that they are soft/hard sciences, my stance is that that the criteria given leads to hard-to-defend conclusions and I am just giving examples. [NEWLINE] [NEWLINE] [STARTQ] No, they are not tangible. They are ideas that are created<mask><mask><mask> of civilization and society. [ENDQ] [NEWLINE] That's an interesting new criteria.  (Note the following is not<mask> I believe<mask> just to show hard-to-defend conclusions based on the criteria "not tangible, tangible being the results of civilization and society.) [NEWLINE] [NEWLINE] * Things happen on a quantum level are not determined until an observer sees it. <mask> far we have only indicated that its a human observer. <mask> things don't happen on a quantum level<mask> its a result of people? [NEWLINE] [NEWLINE] * Industrial Engineering involves human factors and conditions. <mask> Industrial Engineering is a soft-science? [NEWLINE] [NEWLINE] * Civil Engineering involves urban design and transportation which are defined by civilization. <mask> Civil Engineering is a soft-science? [NEWLINE] [NEWLINE] [STARTQ] Evidence in hard-sciences are physical things that are grounded by natural laws. [ENDQ] [NEWLINE] <mask> mathematics are not hard-science?  Advanced bleeding-edge theoretical subjects are not hard-sciences?  (e.g magnetic monopoles have not been physically found) [NEWLINE] [NEWLINE] [STARTQ] GDP is based on the concept of wealth and income which are subjective<mask> attempts to make it objective it by using monetary values. [ENDQ] [NEWLINE] Wealth is relative "I am richer than you"<mask> not subjective "I can prove I am richer than you".  Income is not subjective, you either make $X/yr or you don't. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [NEWLINE] [STARTQ] But then subjects like archeology, macro-evolution and astronomy are not hard-sciences. [ENDQ] [NEWLINE] Ok I think I understand where the confusion is. The " But then" part is to indicate that if we follow this reasoning (the testable prediction) then we come to this hard-to-defend conclusion (i.e. archeology etc are not hard-sciences).  It is not my stance that they are soft/hard sciences, my stance is that that the criteria given leads to hard-to-defend conclusions and I am just giving examples. [NEWLINE] [NEWLINE] [STARTQ] No, they are not tangible. They are ideas that are created as a result of civilization and society. [ENDQ] [NEWLINE] That's an interesting new criteria.  (Note the following is not what I believe but just to show hard-to-defend conclusions based on the criteria "not tangible, tangible being the results of civilization and society.) [NEWLINE] [NEWLINE] * Things happen on a quantum level are not determined until an observer sees it.  So far we have only indicated that its a human observer.  So things don't happen on a quantum level because its a result of people? [NEWLINE] [NEWLINE] * Industrial Engineering involves human factors and conditions.  So Industrial Engineering is a soft-science? [NEWLINE] [NEWLINE] * Civil Engineering involves urban design and transportation which are defined by civilization.  So Civil Engineering is a soft-science? [NEWLINE] [NEWLINE] [STARTQ] Evidence in hard-sciences are physical things that are grounded by natural laws. [ENDQ] [NEWLINE] So mathematics are not hard-science?  Advanced bleeding-edge theoretical subjects are not hard-sciences?  (e.g magnetic monopoles have not been physically found) [NEWLINE] [NEWLINE] [STARTQ] GDP is based on the concept of wealth and income which are subjective but attempts to make it objective it by using monetary values. [ENDQ] [NEWLINE] Wealth is relative "I am richer than you" but not subjective "I can prove I am richer than you".  Income is not subjective, you either make $X/yr or you don't. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If the chance of life arising elsewhere in the universe is greater than zero, then the odds that it does exist out there climbs with the increasing number of places<mask> all the ingredients are present for some form of life to begin. [NEWLINE] [NEWLINE] 1.  There's nothing that says the conditions that are required for life on Earth are **the** requirements for life everywhere else - for all we know it is entirely possible for life to begin someplace based on some chemical process centered around methane or ammonia instead of oxygen/nitrogen/carbon.  Just like water isn't the only liquid that can be used for getting stains out of clothes (dry cleaning doesn't mean no liquids - it just means no water), there could be other chemical processes that support something that is a parallel to<mask> we recognize on Earth<mask> life. [NEWLINE] [NEWLINE] 2. Clumps of molecules arranging themselves into structures [is now understood to be part of thermodynamics]( [URL] /) -<mask> any environment with an abundant source of energy is a good candidate. [NEWLINE] [NEWLINE] 3.  ET life doesn't necessarily need to be anything like life on Earth - our imaginations tend to be limited in that respect to variations on<mask> we see around us. <mask> nature has produced a pretty wide variation of<mask> is possible given the conditions on Earth, we only have a sample of<mask> can work under those conditions and<mask> competing against the other forms that developed under those conditions. [NEWLINE] [NEWLINE] 4.  Nothing is guaranteed,<mask> the human mind isn't well-adapted to comprehending the kinds of numbers involved in a this question<mask><mask><mask> just our galaxy is concerned -<mask> we're still fairly confident that there are at least hundreds of billions of bodies orbiting stars that are in a position to be receiving enough energy to start structuring themselves to dissipate that energy. [NEWLINE] [NEWLINE] Life forming is probably a lot less like a key that can open a specific lock and more like a lottery<mask> millions of chemical arrangements are playing to be in the right place at the right time - and there can be hundreds of winners<mask> the right numbers come up.</s>
Label encoding: <s>If the chance of life arising elsewhere in the universe is greater than zero, then the odds that it does exist out there climbs with the increasing number of places where all the ingredients are present for some form of life to begin. [NEWLINE] [NEWLINE] 1.  There's nothing that says the conditions that are required for life on Earth are **the** requirements for life everywhere else - for all we know it is entirely possible for life to begin someplace based on some chemical process centered around methane or ammonia instead of oxygen/nitrogen/carbon.  Just like water isn't the only liquid that can be used for getting stains out of clothes (dry cleaning doesn't mean no liquids - it just means no water), there could be other chemical processes that support something that is a parallel to what we recognize on Earth as life. [NEWLINE] [NEWLINE] 2. Clumps of molecules arranging themselves into structures [is now understood to be part of thermodynamics]( [URL] /) - so any environment with an abundant source of energy is a good candidate. [NEWLINE] [NEWLINE] 3.  ET life doesn't necessarily need to be anything like life on Earth - our imaginations tend to be limited in that respect to variations on what we see around us.  Since nature has produced a pretty wide variation of what is possible given the conditions on Earth, we only have a sample of what can work under those conditions and when competing against the other forms that developed under those conditions. [NEWLINE] [NEWLINE] 4.  Nothing is guaranteed, but the human mind isn't well-adapted to comprehending the kinds of numbers involved in a this question as far as just our galaxy is concerned - but we're still fairly confident that there are at least hundreds of billions of bodies orbiting stars that are in a position to be receiving enough energy to start structuring themselves to dissipate that energy. [NEWLINE] [NEWLINE] Life forming is probably a lot less like a key that can open a specific lock and more like a lottery where millions of chemical arrangements are playing to be in the right place at the right time - and there can be hundreds of winners when the right numbers come up.</s>
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This gets into the "slippery slope" problem *really* quickly. There are multiple genocides occurring in the world  **right now**. We know about them, you are capable of doing something. Are you complicit for not doing something about each and every one? For standing by and letting it happen?<mask> about ongoing starvation?<mask> about ongoing slavery? I am not saying that it isn't good to do something about it, or that I don't want to try to help,<mask> it isn't my moral imperative to do<mask>. The same thing applies to any number of issues. It simply isn't possible to contribute to rectifying all the issues of which we are aware; there isn't enough time or resources to do<mask>. We are directly responsible for *our own actions*, and bonus life points for above and beyond.<mask> you have an issue that you personally relate to and want to help fix, that is great. I have those pet issues<mask> well and try to help<mask> I can,<mask> I don't believe that someone else's pet issues have to be mine simply<mask> they present the binary "with us or against us" argument. [NEWLINE] [NEWLINE] Your example isn't non-participation, it perpetuates.<mask> I keep a racist store in business by going there, I am doing something wrong. At a certain point it becomes beyond your control. I choose extreme examples<mask> it illustrates the point. At some microscopic level, you *are contributing* to the perpetuation of these issue<mask> you live in (presumably) a 1st world country that in some way, shape or form influences international events, etc.<mask> again, you have to draw the line somewhere. [NEWLINE] [NEWLINE] I fully support the idea of being aware of<mask> your actions contribute to at a macroscopic and microscopic level and choosing to live in such a way that minimizes your impact on negative events,<mask><mask><mask> with the assertion that it is my responsibility to do something about every single issue, particularly those that I do not directly affect,<mask> it simply isn't possible.</s>
Label encoding: <s>This gets into the "slippery slope" problem *really* quickly. There are multiple genocides occurring in the world  **right now**. We know about them, you are capable of doing something. Are you complicit for not doing something about each and every one? For standing by and letting it happen? What about ongoing starvation? What about ongoing slavery? I am not saying that it isn't good to do something about it, or that I don't want to try to help, but it isn't my moral imperative to do so. The same thing applies to any number of issues. It simply isn't possible to contribute to rectifying all the issues of which we are aware; there isn't enough time or resources to do so. We are directly responsible for *our own actions*, and bonus life points for above and beyond. If you have an issue that you personally relate to and want to help fix, that is great. I have those pet issues as well and try to help where I can, but I don't believe that someone else's pet issues have to be mine simply because they present the binary "with us or against us" argument. [NEWLINE] [NEWLINE] Your example isn't non-participation, it perpetuates. If I keep a racist store in business by going there, I am doing something wrong. At a certain point it becomes beyond your control. I choose extreme examples because it illustrates the point. At some microscopic level, you *are contributing* to the perpetuation of these issue because you live in (presumably) a 1st world country that in some way, shape or form influences international events, etc. But again, you have to draw the line somewhere. [NEWLINE] [NEWLINE] I fully support the idea of being aware of what your actions contribute to at a macroscopic and microscopic level and choosing to live in such a way that minimizes your impact on negative events, but I disagree with the assertion that it is my responsibility to do something about every single issue, particularly those that I do not directly affect, because it simply isn't possible.</s>
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your argument for minimum wage changed my view to a degree and<mask><mask> that we do not have even close to a free and competitive market and I'll probably check out the Card and Kreuger book for further information.<mask> for that,  ∆. [NEWLINE] [NEWLINE] About the flat tax<mask>,<mask><mask> that money has decreasing marginal utility<mask> it increases which would make sense for a progressive tax<mask> just<mask> it places the same burden on people of higher incomes doesn't mean it is necessarily the best for society. Revenue for government is another good point<mask> increased taxation is not the answer to financial burdens.<mask><mask>, the countries with the steepest tax rates such<mask> the U.S. and Japan have stagnating economies (not stagnant)<mask> those with less burdensome taxes have burgeoning economies. I'm not going to point at GDP<mask><mask><mask> it's a stupid and inaccurate measurement of growth and there are currently no completely accurate alternatives<mask> GNP confirms that less steep tax rates increase growth. (<mask> about your financial situation, there is no true flat tax<mask> it is accepted that people below a certain threshold cannot bear the same burden,<mask> I am arguing for is a marginal flat tax) [NEWLINE] [NEWLINE] A lot of people in this thread seem to be caught up on the fact that people aren't going to want to earn less money<mask> of a progressive tax. That is not my point which I guess I worded completely wrong (do I still award a delta<mask> you made me realize that my question is wrong?) I guess you (ppl in this thread) helped me realize a progressive income tax has value<mask><mask> I meant is a progressive tax on capital gains rather than personal income. Individuals will typically strive to earn the maximum profit<mask> corporations will seek to maximize their profits with the least allocation of resources which means progressive taxation on capital gains inhibits growth at that level. [NEWLINE] In regards to your final quote, the U.S. has one of the highest progressive tax rates in the industrialized world. Even Russia (once thought socialist, definitely not anymore) has a flat tax.</s>
Label encoding: <s>Your argument for minimum wage changed my view to a degree and I agree that we do not have even close to a free and competitive market and I'll probably check out the Card and Kreuger book for further information. So for that,  ∆. [NEWLINE] [NEWLINE] About the flat tax though, I agree that money has decreasing marginal utility as it increases which would make sense for a progressive tax but just because it places the same burden on people of higher incomes doesn't mean it is necessarily the best for society. Revenue for government is another good point but increased taxation is not the answer to financial burdens. In fact, the countries with the steepest tax rates such as the U.S. and Japan have stagnating economies (not stagnant) while those with less burdensome taxes have burgeoning economies. I'm not going to point at GDP because I think it's a stupid and inaccurate measurement of growth and there are currently no completely accurate alternatives but GNP confirms that less steep tax rates increase growth. ( Also about your financial situation, there is no true flat tax because it is accepted that people below a certain threshold cannot bear the same burden, what I am arguing for is a marginal flat tax) [NEWLINE] [NEWLINE] A lot of people in this thread seem to be caught up on the fact that people aren't going to want to earn less money because of a progressive tax. That is not my point which I guess I worded completely wrong (do I still award a delta if you made me realize that my question is wrong?) I guess you (ppl in this thread) helped me realize a progressive income tax has value but what I meant is a progressive tax on capital gains rather than personal income. Individuals will typically strive to earn the maximum profit but corporations will seek to maximize their profits with the least allocation of resources which means progressive taxation on capital gains inhibits growth at that level. [NEWLINE] In regards to your final quote, the U.S. has one of the highest progressive tax rates in the industrialized world. Even Russia (once thought socialist, definitely not anymore) has a flat tax.</s>
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>* They are [aftermarket]( [URL].aspx). It is<mask> the motorcycle owner's choice to [install]( [URL] ) it. All of the [2014]( [URL].com/en_CA/Motorcycles/compare-bikes.html) model Harleys come with mufflers installed, sometimes even dual mufflers (like the SuperLow). [NEWLINE] [NEWLINE] * In my view they are just a way to say "look at me". They serve no other purpose. It is selfish to expose other people to excessive noise just<mask> you think it is cool. [NEWLINE] [NEWLINE] * It will damage hearing.<mask> I am a pedestrian and I hear one go by - one turned right behind me after I crossed a light - it hurts my ears. I'm sure<mask> I had to hear that every day it would do some damage. (sarcasm) Maybe the owners have damaged their hearing such that their cycle sounds normal to them. (/sarcasm) Check out [slide five here, under non-occupational.]( [URL] &amp;HINFO/NoiseToolBoxes/Wear%20Your%20Hearing%20Protection.pdf) [NEWLINE] [NEWLINE] CMV. [NEWLINE] [NEWLINE] Edit: My view isn't changed completely. I understand that loud pipes are one *extra* thing to get motorcyclists noticed by,<mask><mask><mask> they still can use their horn<mask> needed, and just try to be aware and avoid bad situations, especially<mask> they are always loud (i.e. you don't need them on residential streets). Basically<mask><mask> their obnoxious-ness is still too much. People make bad judgement calls all the time, even<mask> they hear a motorcyclist around. They could still swing out of their lane without checking their blind spot. [NEWLINE] [NEWLINE] <mask> obligatory, wow front page of CMV! [NEWLINE] [NEWLINE] Edit 2: /u/Smiley_Black_Sheep changed my mind by giving a different take on the issue, and I gave a delta. Sure they're annoying,<mask> banning is a bit extreme. </s><pad>
Label encoding: <s>* They are [aftermarket]( [URL].aspx). It is therefore the motorcycle owner's choice to [install]( [URL] ) it. All of the [2014]( [URL].com/en_CA/Motorcycles/compare-bikes.html) model Harleys come with mufflers installed, sometimes even dual mufflers (like the SuperLow). [NEWLINE] [NEWLINE] * In my view they are just a way to say "look at me". They serve no other purpose. It is selfish to expose other people to excessive noise just because you think it is cool. [NEWLINE] [NEWLINE] * It will damage hearing. When I am a pedestrian and I hear one go by - one turned right behind me after I crossed a light - it hurts my ears. I'm sure if I had to hear that every day it would do some damage. (sarcasm) Maybe the owners have damaged their hearing such that their cycle sounds normal to them. (/sarcasm) Check out [slide five here, under non-occupational.]( [URL] &amp;HINFO/NoiseToolBoxes/Wear%20Your%20Hearing%20Protection.pdf) [NEWLINE] [NEWLINE] CMV. [NEWLINE] [NEWLINE] Edit: My view isn't changed completely. I understand that loud pipes are one *extra* thing to get motorcyclists noticed by, but I think they still can use their horn if needed, and just try to be aware and avoid bad situations, especially since they are always loud (i.e. you don't need them on residential streets). Basically I think their obnoxious-ness is still too much. People make bad judgement calls all the time, even if they hear a motorcyclist around. They could still swing out of their lane without checking their blind spot. [NEWLINE] [NEWLINE] Also obligatory, wow front page of CMV! [NEWLINE] [NEWLINE] Edit 2: /u/Smiley_Black_Sheep changed my mind by giving a different take on the issue, and I gave a delta. Sure they're annoying, but banning is a bit extreme. </s><pad>
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>She's become a lightning rod for the drama concerning a much larger issue which is Gaming Journalists not disclosing personal and financial relationships with people and businesses they are writing about or reviewing. [NEWLINE] [NEWLINE] - Nobody really cares that people are getting jobs<mask> they know people. [NEWLINE] [NEWLINE] - Nobody really hates women aside from the exceptionally rare hate breather or toxic hate spreading children who think its funny. [NEWLINE] [NEWLINE] - Most people don't care or are indifferent about objectifying women<mask><mask><mask> they are being respectful to them. <mask> an example most normal people find it pretty okay to think or say women are hot,<mask> they don't go around slapping girls on the butt or calling them 'hoes'. [NEWLINE] [NEWLINE] [STARTQ] Now,<mask> I was to tail the cars of child sex offenders and throw rocks at them and threaten to kill them, I'm sure some people would agree,<mask> is this really reasonable? [ENDQ] [NEWLINE] It is or can be reasonable<mask>'sex offenders' get away with it.  In our society<mask> that issue is handled in the courts. <mask> it's not appropriate for someone to take it upon themselves to enact justice a second time. <mask> I lived in a country<mask> raping kids was legal, then huffing a rock at a'sex offender' would seem a lot more reasonable. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>,<mask> all this stuff about Zoe is to be believed then it's perfectly reasonable for the community to bring it to light, chew her out, and otherwise shun her activities.  Especially due to the fact that among those accusations is she sabotaged a $400,000 game jam, abused the DMCA to remove youtube criticism of herself, and is herself taking peoples money in activities related to the issue. [NEWLINE] [NEWLINE] - It is not<mask> appropriate for people to hack / dox / visit her house / harass her. [NEWLINE] [NEWLINE] And I have to point out that harassment isn't simply saying unflattering truths/untruths about someone.  Harassment is sending and spamming her with hate speak.  Criticism and harassment are not equivalent.</s>
Label encoding: <s>She's become a lightning rod for the drama concerning a much larger issue which is Gaming Journalists not disclosing personal and financial relationships with people and businesses they are writing about or reviewing. [NEWLINE] [NEWLINE] - Nobody really cares that people are getting jobs because they know people. [NEWLINE] [NEWLINE] - Nobody really hates women aside from the exceptionally rare hate breather or toxic hate spreading children who think its funny. [NEWLINE] [NEWLINE] - Most people don't care or are indifferent about objectifying women as long as they are being respectful to them.  As an example most normal people find it pretty okay to think or say women are hot, but they don't go around slapping girls on the butt or calling them 'hoes'. [NEWLINE] [NEWLINE] [STARTQ] Now, if I was to tail the cars of child sex offenders and throw rocks at them and threaten to kill them, I'm sure some people would agree, but is this really reasonable? [ENDQ] [NEWLINE] It is or can be reasonable if'sex offenders' get away with it.  In our society however that issue is handled in the courts.  So it's not appropriate for someone to take it upon themselves to enact justice a second time.  If I lived in a country where raping kids was legal, then huffing a rock at a'sex offender' would seem a lot more reasonable. [NEWLINE] [NEWLINE] On the other hand, if all this stuff about Zoe is to be believed then it's perfectly reasonable for the community to bring it to light, chew her out, and otherwise shun her activities.  Especially due to the fact that among those accusations is she sabotaged a $400,000 game jam, abused the DMCA to remove youtube criticism of herself, and is herself taking peoples money in activities related to the issue. [NEWLINE] [NEWLINE] - It is not however appropriate for people to hack / dox / visit her house / harass her. [NEWLINE] [NEWLINE] And I have to point out that harassment isn't simply saying unflattering truths/untruths about someone.  Harassment is sending and spamming her with hate speak.  Criticism and harassment are not equivalent.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've listed all these different things that have different capabilities of the Xbox One,<mask> it's almost like you're failing to see that there's value in bundling them together at a fairly reasonable price. [NEWLINE] [NEWLINE] I've seen you mention the Raspberry Pi, Rokus, last gen consoles, PCs, etc. <mask><mask> I don't have/want those things? [NEWLINE] [NEWLINE] Truth be told, last gen consoles have almost every media capability I'd want out of a console.  Netflix, youtube, streaming torrented files from my computer, etc.  About the only thing they don't have is Spotify. <mask>, in 2 years, they won't have the newest games. [NEWLINE] [NEWLINE] PCs have all those things.  Newest games, all the media apps and websites I could want, and it can run productivity software. <mask> gaming PCs aren't<mask> accessible to the general populace<mask> a console.  They're<mask> a never-ending chase of new parts.  I paid $400 for my Xbox 4 years ago.  I've spent over $2000 on non-I/O parts (monitors, keyboards, mouses, etc.) on gaming rigs in the same time period.  My console is very much WYSIWYG,<mask> I'll never be content with running medium graphics settings on my gaming rig. [NEWLINE] [NEWLINE] <mask> instead of thinking about<mask> you could do with this device or that device and all, imagine starting with absolutely no media consuming devices. [NEWLINE] [NEWLINE] You have nothing on which you can play video games, watch Netflix, play blu-rays, etc. <mask> do you get to best maximize your cost to feature ratio? [NEWLINE] [NEWLINE] Personally, I'd get a console.  The decision between Xbone and PS4 is arbitrary, really, and comes down to small features and exclusives. [NEWLINE] [NEWLINE] <mask> at that point, the fact that Project Spark or Titanfall are on PC doesn't matter. <mask> matters is that they're on the Xbone and not the PS4.</s>
Label encoding: <s>You've listed all these different things that have different capabilities of the Xbox One, but it's almost like you're failing to see that there's value in bundling them together at a fairly reasonable price. [NEWLINE] [NEWLINE] I've seen you mention the Raspberry Pi, Rokus, last gen consoles, PCs, etc.  What if I don't have/want those things? [NEWLINE] [NEWLINE] Truth be told, last gen consoles have almost every media capability I'd want out of a console.  Netflix, youtube, streaming torrented files from my computer, etc.  About the only thing they don't have is Spotify.  But, in 2 years, they won't have the newest games. [NEWLINE] [NEWLINE] PCs have all those things.  Newest games, all the media apps and websites I could want, and it can run productivity software.  But gaming PCs aren't as accessible to the general populace as a console.  They're also a never-ending chase of new parts.  I paid $400 for my Xbox 4 years ago.  I've spent over $2000 on non-I/O parts (monitors, keyboards, mouses, etc.) on gaming rigs in the same time period.  My console is very much WYSIWYG, but I'll never be content with running medium graphics settings on my gaming rig. [NEWLINE] [NEWLINE] So instead of thinking about what you could do with this device or that device and all, imagine starting with absolutely no media consuming devices. [NEWLINE] [NEWLINE] You have nothing on which you can play video games, watch Netflix, play blu-rays, etc.  What do you get to best maximize your cost to feature ratio? [NEWLINE] [NEWLINE] Personally, I'd get a console.  The decision between Xbone and PS4 is arbitrary, really, and comes down to small features and exclusives. [NEWLINE] [NEWLINE] But at that point, the fact that Project Spark or Titanfall are on PC doesn't matter.  What matters is that they're on the Xbone and not the PS4.</s>
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The issue is probably more one of a criminal record than the race of the occupants.  Numerous studies have been done about the racial aspect.  A.R. Ward, an activist and blogger, [looked at the name study a<mask> back] ( [URL] /): [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the only reason for the disparity was one or two randomly high or randomly low call back percentages for each group. Aisha, which is apparently a black sounding name, got 2.2% call back which killed the overall call back percentage for black women. I can’t imagine employers racially discriminating against someone named Aisha (2.2%)<mask> not someone named Ebony (9.6%, the name literally means black); it’s more likely that Aisha just wasn’t very lucky,<mask> the authors don’t address this at all. [ENDQ] [NEWLINE]. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the title simply used different names within the same study then the title would convey the exact opposite, for example “Are Emily (7.9%) and Todd (5.9%) More Employable than Ebony (9.6%) and Jermaine (9.6%)?” [ENDQ] [NEWLINE]. [NEWLINE] [STARTQ] <mask><mask> these resumes were evenly split between each name, this means 63 white resumes were sent for each white name and 61 for each black name.<mask> taking the names used in the title of the study, Greg (7.8) and Jamal (6.6), we see a grand total difference in callbacks equalling (brace yourself)… 1. White name Greg received 5 callbacks<mask> black name Jamal got 4. This is representative of all the male names, the difference is<mask> small that making them into percentages and comparing them is extremely deceptive. [ENDQ] [NEWLINE] Ward<mask> namechecks [this study from California] ( [URL] ), which found "no negative causal impact of having a distinctively Black name on life outcomes." [NEWLINE] [NEWLINE] <mask> anything, the study you're looking at appears to be an outlier at best.</s>
Label encoding: <s>The issue is probably more one of a criminal record than the race of the occupants.  Numerous studies have been done about the racial aspect.  A.R. Ward, an activist and blogger, [looked at the name study a while back] ( [URL] /): [NEWLINE] [NEWLINE] [STARTQ] In fact the only reason for the disparity was one or two randomly high or randomly low call back percentages for each group. Aisha, which is apparently a black sounding name, got 2.2% call back which killed the overall call back percentage for black women. I can’t imagine employers racially discriminating against someone named Aisha (2.2%) but not someone named Ebony (9.6%, the name literally means black); it’s more likely that Aisha just wasn’t very lucky, but the authors don’t address this at all. [ENDQ] [NEWLINE]. [NEWLINE] [NEWLINE] [STARTQ] But if the title simply used different names within the same study then the title would convey the exact opposite, for example “Are Emily (7.9%) and Todd (5.9%) More Employable than Ebony (9.6%) and Jermaine (9.6%)?” [ENDQ] [NEWLINE]. [NEWLINE] [STARTQ] Assuming that these resumes were evenly split between each name, this means 63 white resumes were sent for each white name and 61 for each black name. So taking the names used in the title of the study, Greg (7.8) and Jamal (6.6), we see a grand total difference in callbacks equalling (brace yourself)… 1. White name Greg received 5 callbacks while black name Jamal got 4. This is representative of all the male names, the difference is so small that making them into percentages and comparing them is extremely deceptive. [ENDQ] [NEWLINE] Ward also namechecks [this study from California] ( [URL] ), which found "no negative causal impact of having a distinctively Black name on life outcomes." [NEWLINE] [NEWLINE] If anything, the study you're looking at appears to be an outlier at best.</s>
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Think of the sports teams. [NEWLINE] [NEWLINE] My team<mask> to state or got close nearly every year. [NEWLINE] [NEWLINE] My football coach would get really stressed<mask> he couldn't fit in enough practice time in<mask> it was dark.  That was with morning and weekend practices. [NEWLINE] [NEWLINE] I wasn't on the Basketball team<mask> I do know that there wasn't enough time that the gym was open.  The freshman boys practiced 7:30-10:00pm<mask> the freshman girls practiced from 5:30-7:30am. [NEWLINE] [NEWLINE] 9-5 would wipe out the desirable after school time.  And create problems<mask> the teams aren't getting enough sleep<mask> all the practices are before school. [NEWLINE] [NEWLINE] After football ended I was on the Swim team.  My school did not have its own pool,<mask> we had to rent out lanes at the public schools' Natatorium.  Scheduling was a huge pain, trying to fit 4, sometimes 5 schools, and a club team into this pool. [NEWLINE] [NEWLINE] Now morning swim practice was worse than morning football.   Coaches hated it, janitors hated it, lifeguards hated it, swimmers hated it. [NEWLINE] [NEWLINE] I was<mask> on the track team. [NEWLINE] [NEWLINE] I ran long-distance and threw. [NEWLINE] [NEWLINE] Before 9 o'clock it's often dark.  After 5 it's often dark.  Is it really such a good idea to send High-Schoolers out to run sidewalks in the dark‽  Have you ever thrown a javelin or tried to recover a discus in the dark‽ [NEWLINE] [NEWLINE] I was in Band, it's hard enough to work around sports.  I feel sorry for anyone who has to do it with wacky scheduling. [NEWLINE] [NEWLINE] In short it's just not practical to have school set with the work day.  At least<mask> extra activities are concerned.   School isn't a daycare.  It has been set at the time it is not arbitrarily,<mask> selected<mask> a balance or compromise that works the best consistently.</s>
Label encoding: <s>Think of the sports teams. [NEWLINE] [NEWLINE] My team when to state or got close nearly every year. [NEWLINE] [NEWLINE] My football coach would get really stressed when he couldn't fit in enough practice time in because it was dark.  That was with morning and weekend practices. [NEWLINE] [NEWLINE] I wasn't on the Basketball team but I do know that there wasn't enough time that the gym was open.  The freshman boys practiced 7:30-10:00pm while the freshman girls practiced from 5:30-7:30am. [NEWLINE] [NEWLINE] 9-5 would wipe out the desirable after school time.  And create problems when the teams aren't getting enough sleep because all the practices are before school. [NEWLINE] [NEWLINE] After football ended I was on the Swim team.  My school did not have its own pool, so we had to rent out lanes at the public schools' Natatorium.  Scheduling was a huge pain, trying to fit 4, sometimes 5 schools, and a club team into this pool. [NEWLINE] [NEWLINE] Now morning swim practice was worse than morning football.   Coaches hated it, janitors hated it, lifeguards hated it, swimmers hated it. [NEWLINE] [NEWLINE] I was also on the track team. [NEWLINE] [NEWLINE] I ran long-distance and threw. [NEWLINE] [NEWLINE] Before 9 o'clock it's often dark.  After 5 it's often dark.  Is it really such a good idea to send High-Schoolers out to run sidewalks in the dark‽  Have you ever thrown a javelin or tried to recover a discus in the dark‽ [NEWLINE] [NEWLINE] I was in Band, it's hard enough to work around sports.  I feel sorry for anyone who has to do it with wacky scheduling. [NEWLINE] [NEWLINE] In short it's just not practical to have school set with the work day.  At least when extra activities are concerned.   School isn't a daycare.  It has been set at the time it is not arbitrarily, but selected as a balance or compromise that works the best consistently.</s>
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because allowing, and encouraging, teachers to carry guns in schools will cause more deaths than the policy will ever, ever save. <mask> teachers are allowed to carry guns in schools, then, obviously, there will be more guns in schools.  I firmly believe that a school with more guns in it is much more dangerous than the same school with less guns in it. [NEWLINE] [NEWLINE] Yes, there will come a day<mask> there is a dangerous situation at school<mask> a teacher carrying a gun will save someone's life.  This benefit should not be over looked or downplayed, and it is difficult to<mask><mask> in the right situation, a good guy with a gun is exactly<mask> you need.  Let's say most schools start letting teachers carry guns - in the course of a year,<mask> many people would be protected<mask><mask><mask>?  I'm guessing around 10. [NEWLINE] [NEWLINE] <mask> the mere presence of guns makes situations more dangerous.  Owners of guns are more likely to commit suicide than people who do not have guns.  Having a gun present in any confrontation increases the chance that the confrontation will end violently.  Simply having more (or for that matter, ANY) guns circulating around a school VASTLY increases the chance that someone will be hurt from either the use or misuse of that weapon. [NEWLINE] [NEWLINE] <mask> you start letting teachers carry guns in schools, there will be a lot of guns in schools  Most of those guns will never, ever be used for good or for bad.  A very very small portion of them will be used properly used to save a life that would otherwise be lost.  A very very small portion of them will be stolen and used<mask> a weapon by someone else, misused and cause fatal accidents, used in a moment of despair to commit suicide, and be deliberately used by their disgruntled owners.  I have a very, very hard time seeing<mask> the number of lives saved by guns in the former category can come anywhere close to matching the number of lives lost in the latter category.  </s>
Label encoding: <s>Because allowing, and encouraging, teachers to carry guns in schools will cause more deaths than the policy will ever, ever save.  If teachers are allowed to carry guns in schools, then, obviously, there will be more guns in schools.  I firmly believe that a school with more guns in it is much more dangerous than the same school with less guns in it. [NEWLINE] [NEWLINE] Yes, there will come a day when there is a dangerous situation at school where a teacher carrying a gun will save someone's life.  This benefit should not be over looked or downplayed, and it is difficult to argue that in the right situation, a good guy with a gun is exactly what you need.  Let's say most schools start letting teachers carry guns - in the course of a year, how many people would be protected because of this?  I'm guessing around 10. [NEWLINE] [NEWLINE] But the mere presence of guns makes situations more dangerous.  Owners of guns are more likely to commit suicide than people who do not have guns.  Having a gun present in any confrontation increases the chance that the confrontation will end violently.  Simply having more (or for that matter, ANY) guns circulating around a school VASTLY increases the chance that someone will be hurt from either the use or misuse of that weapon. [NEWLINE] [NEWLINE] If you start letting teachers carry guns in schools, there will be a lot of guns in schools  Most of those guns will never, ever be used for good or for bad.  A very very small portion of them will be used properly used to save a life that would otherwise be lost.  A very very small portion of them will be stolen and used as a weapon by someone else, misused and cause fatal accidents, used in a moment of despair to commit suicide, and be deliberately used by their disgruntled owners.  I have a very, very hard time seeing how the number of lives saved by guns in the former category can come anywhere close to matching the number of lives lost in the latter category.  </s>
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've said that taking lives for **whatever reason** is wrong. This seems a little ridiculous. Let's say you've got a gun and you're in a room with a man with a knife who's in the midst of a killing spree. And let's say for arguments sake you could easily shoot this guy and kill him,<mask> saving the numerous innocents, some of which are children by the way. You have two choices: [NEWLINE] [NEWLINE] 1. Let this man kill everyone including you [NEWLINE] 2. Shoot the man and save innocent people's lives [NEWLINE] [NEWLINE] I believe that in this situation it would be morally correct to kill this man.<mask> taking lives is **sometimes and only sometimes** justifiable and morally correct. [NEWLINE] [NEWLINE] <mask> this is the case, then extending this justification to soldiers is relatively straightforward.<mask> a nation is under attack, soldiers are simply responding to threats and acts of violence against themselves and their fellow humans.<mask>, it's morally right for them to defend themselves and others against acts of violence even<mask> it involves taking lives,<mask> that's often<mask>'s necessary to deal with these threats. [NEWLINE] [NEWLINE] This is<mask> happened in WWII. Japan (allied with Nazi Germany at the time) attacked America. America fought back, until Japan and the Nazis surrendered and was no longer a threat. Yeah sure, the Japanese soldiers were evil,<mask><mask><mask> it's a bit of a stretch to say that the US military and its Allies were in the wrong for not letting Japan and the Nazis win the war and oppress many millions of people and murder lots of people. [NEWLINE] [NEWLINE] <mask> the reason that soldiers are not evil is<mask> they're employed to protect **you** from other countries that might want to hurt you. Now, whether a particular conflict or not is a moral one or not is a different matter,<mask> it's certainly the case that they would be the ones who would be the first to lay down their lives<mask> you can keep yours<mask> your country was attacked. They deserve some respect for that.</s>
Label encoding: <s>You've said that taking lives for **whatever reason** is wrong. This seems a little ridiculous. Let's say you've got a gun and you're in a room with a man with a knife who's in the midst of a killing spree. And let's say for arguments sake you could easily shoot this guy and kill him, thus saving the numerous innocents, some of which are children by the way. You have two choices: [NEWLINE] [NEWLINE] 1. Let this man kill everyone including you [NEWLINE] 2. Shoot the man and save innocent people's lives [NEWLINE] [NEWLINE] I believe that in this situation it would be morally correct to kill this man. Therefore taking lives is **sometimes and only sometimes** justifiable and morally correct. [NEWLINE] [NEWLINE] If this is the case, then extending this justification to soldiers is relatively straightforward. If a nation is under attack, soldiers are simply responding to threats and acts of violence against themselves and their fellow humans. Therefore, it's morally right for them to defend themselves and others against acts of violence even if it involves taking lives, because that's often what's necessary to deal with these threats. [NEWLINE] [NEWLINE] This is what happened in WWII. Japan (allied with Nazi Germany at the time) attacked America. America fought back, until Japan and the Nazis surrendered and was no longer a threat. Yeah sure, the Japanese soldiers were evil, but I think it's a bit of a stretch to say that the US military and its Allies were in the wrong for not letting Japan and the Nazis win the war and oppress many millions of people and murder lots of people. [NEWLINE] [NEWLINE] So the reason that soldiers are not evil is because they're employed to protect **you** from other countries that might want to hurt you. Now, whether a particular conflict or not is a moral one or not is a different matter, but it's certainly the case that they would be the ones who would be the first to lay down their lives so you can keep yours if your country was attacked. They deserve some respect for that.</s>
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm probably too late to the party,<mask> oh well. I'm dating myself a bit here,<mask> this argument/concern is old and tired.<mask> my mother was a child, they used to sing: [NEWLINE] [NEWLINE] *Eeny, meeny, miny, moe,* [NEWLINE] [NEWLINE] *Catch a nigger by the toe.* [NEWLINE] [NEWLINE] *<mask> he hollers, let him go,* [NEWLINE] [NEWLINE] *Eeny, meeny, miny, moe.* [NEWLINE] [NEWLINE] In the 80s there was a push to eliminate racism,<mask><mask> I learned that rhyme, it was "tigger"... And some people called it "political correctness". "It's just a child's nursery rhyme... it doesn't mean anything." [NEWLINE] [NEWLINE] <mask> I was a kid (I'm in my 30s), we still used: [NEWLINE] [NEWLINE] * Fireman (now firefighter) [NEWLINE] * Policeman (now police officer) [NEWLINE] * Stewardess (now flight attendant) [NEWLINE] * Waitress (now server) [NEWLINE] [NEWLINE] And<mask> this push to move towards the gender neutral job titles happened, it was considered a joke. People laughed and said that "political correctness had gone too far".<mask> the fact is that words matter. And the first step towards changing people's mindsets depends on the words they use. And now job titles aren't constantly reinforcing gender roles. [NEWLINE] [NEWLINE] [STARTQ] wether we should then be differentiating the difference between diabetics and nondiabetics and<mask> forth. [ENDQ] [NEWLINE] Yes.<mask><mask> we should move away from "normal"<mask> a term, "I am average" is likely much more accurate in most situations. "I'm normal" inherently implies that the other person is "not normal" or "weird". I'm not going to spend my time describing my person by all the things I'm not,<mask> in specific context, yes, I will say "No, I'm not diabetic", not "No, I'm normal". </s>
Label encoding: <s>I'm probably too late to the party, but oh well. I'm dating myself a bit here, but this argument/concern is old and tired. When my mother was a child, they used to sing: [NEWLINE] [NEWLINE] *Eeny, meeny, miny, moe,* [NEWLINE] [NEWLINE] *Catch a nigger by the toe.* [NEWLINE] [NEWLINE] * If he hollers, let him go,* [NEWLINE] [NEWLINE] *Eeny, meeny, miny, moe.* [NEWLINE] [NEWLINE] In the 80s there was a push to eliminate racism, so when I learned that rhyme, it was "tigger"... And some people called it "political correctness". "It's just a child's nursery rhyme... it doesn't mean anything." [NEWLINE] [NEWLINE] When I was a kid (I'm in my 30s), we still used: [NEWLINE] [NEWLINE] * Fireman (now firefighter) [NEWLINE] * Policeman (now police officer) [NEWLINE] * Stewardess (now flight attendant) [NEWLINE] * Waitress (now server) [NEWLINE] [NEWLINE] And when this push to move towards the gender neutral job titles happened, it was considered a joke. People laughed and said that "political correctness had gone too far". But the fact is that words matter. And the first step towards changing people's mindsets depends on the words they use. And now job titles aren't constantly reinforcing gender roles. [NEWLINE] [NEWLINE] [STARTQ] wether we should then be differentiating the difference between diabetics and nondiabetics and so forth. [ENDQ] [NEWLINE] Yes. I think we should move away from "normal" as a term, "I am average" is likely much more accurate in most situations. "I'm normal" inherently implies that the other person is "not normal" or "weird". I'm not going to spend my time describing my person by all the things I'm not, but in specific context, yes, I will say "No, I'm not diabetic", not "No, I'm normal". </s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Correct me<mask> I'm wrong,<mask> you seem to take<mask> a<mask><mask> these countries will develop just<mask> easily and quickly with more employee protection laws in place.  That is not<mask> happens in practice.  These laws would create more costs for the businesses operating in these countries, and these additional costs will take away the cost advantage of operating in these countries.  The companies will either move on to a new country with a better cost advantage, or bring production back home.  This would leave the country right back<mask> it started with no development opportunities. [NEWLINE] [NEWLINE] These working conditions are a necessary evil in facilitating the country's development.  Every country must go through this phase of development in order to one day be considered a "developed" rather than a perpetual "developing" country.  It is the only way for the country to attract outside investment, which is necessary due to the dearth of domestic investment. [NEWLINE] [NEWLINE] This goes back to a very basic truth in human interactions. <mask> you want someone else to do something that benefits you, the best way to accomplish this is by showing<mask> it would benefit them, or showing another way you can benefit them in return.  People rarely inconvenience themselves<mask> doing<mask> will not somehow improve their condition. <mask> these countries want money coming into their country, the best way to do<mask> is by showing people with money<mask> the country can help. <mask> the country has nothing to offer people with money, then the only money the country will receive will be through random acts of charity. <mask> the country wants more than random acts of charity, then the country will need to find a way to make themselves useful to people with money.  It does this by showing people with money that the country will reduce their costs and allow the people with money to make more money than<mask> the goods were produced in another country.  Doing this makes people with money spend money on production in the country, which gives the country the money they need for development.</s>
Label encoding: <s>Correct me if I'm wrong, but you seem to take as a given that these countries will develop just as easily and quickly with more employee protection laws in place.  That is not what happens in practice.  These laws would create more costs for the businesses operating in these countries, and these additional costs will take away the cost advantage of operating in these countries.  The companies will either move on to a new country with a better cost advantage, or bring production back home.  This would leave the country right back where it started with no development opportunities. [NEWLINE] [NEWLINE] These working conditions are a necessary evil in facilitating the country's development.  Every country must go through this phase of development in order to one day be considered a "developed" rather than a perpetual "developing" country.  It is the only way for the country to attract outside investment, which is necessary due to the dearth of domestic investment. [NEWLINE] [NEWLINE] This goes back to a very basic truth in human interactions.  If you want someone else to do something that benefits you, the best way to accomplish this is by showing how it would benefit them, or showing another way you can benefit them in return.  People rarely inconvenience themselves if doing so will not somehow improve their condition.  If these countries want money coming into their country, the best way to do so is by showing people with money how the country can help.  If the country has nothing to offer people with money, then the only money the country will receive will be through random acts of charity.  If the country wants more than random acts of charity, then the country will need to find a way to make themselves useful to people with money.  It does this by showing people with money that the country will reduce their costs and allow the people with money to make more money than if the goods were produced in another country.  Doing this makes people with money spend money on production in the country, which gives the country the money they need for development.</s>
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are two main issues. First, these 'tips' are almost always given to people who have already been victimized on<mask> they could have prevented a crime that was already committed against them. That's incredibly dismissive and condescending. [NEWLINE] [NEWLINE] <mask>, the type of advice you are talking about basically boils down to "There are things a woman can do which will provoke a man to rape her". It's important to realize that there is no reliable way for a woman to prevent being raped. And I mean it<mask> I say *woman*. We only ever give this... *advice*... to women. It isn't meant to apply to men. Just about everything I've heard in this vein are things i've done. I've gotten drunk and passed out at relative strangers houses. I've gotten drunk around people I don't know. I've accepted drinks from strangers. I've flirted with people I don't know well. None of those things means it's okay to rape me. [NEWLINE] [NEWLINE] There is a time and place to warn young women about the realities of this world. I absolutely plan to have a conversation with my daughter about rape. And in that conversation, I will warn her about situations<mask> women are typically raped, and advise her to avoid those situations.<mask> there's a **massive** chasm of difference between that conversation and saying "Well you shouldn't have been drinking around people you don't know!" after someone has been raped. At that point, you're basically doing nothing except shifting the blame from someone who committed a crime to someone who was a victim of a crime. [NEWLINE] [NEWLINE] Edit: I thought of more. It's<mask> important to consider the message we are sending.<mask> a young woman is raped and all anyone talks about is<mask> *she did wrong*, aren't we sending the clear message that it's her fault? Rather than sending the clear message that this man is potentially guilty of a terrible crime. We're punishing the victims rather than perpetrators. </s>
Label encoding: <s>There are two main issues. First, these 'tips' are almost always given to people who have already been victimized on how they could have prevented a crime that was already committed against them. That's incredibly dismissive and condescending. [NEWLINE] [NEWLINE] Secondly, the type of advice you are talking about basically boils down to "There are things a woman can do which will provoke a man to rape her". It's important to realize that there is no reliable way for a woman to prevent being raped. And I mean it when I say *woman*. We only ever give this... *advice*... to women. It isn't meant to apply to men. Just about everything I've heard in this vein are things i've done. I've gotten drunk and passed out at relative strangers houses. I've gotten drunk around people I don't know. I've accepted drinks from strangers. I've flirted with people I don't know well. None of those things means it's okay to rape me. [NEWLINE] [NEWLINE] There is a time and place to warn young women about the realities of this world. I absolutely plan to have a conversation with my daughter about rape. And in that conversation, I will warn her about situations where women are typically raped, and advise her to avoid those situations. But there's a **massive** chasm of difference between that conversation and saying "Well you shouldn't have been drinking around people you don't know!" after someone has been raped. At that point, you're basically doing nothing except shifting the blame from someone who committed a crime to someone who was a victim of a crime. [NEWLINE] [NEWLINE] Edit: I thought of more. It's also important to consider the message we are sending. If a young woman is raped and all anyone talks about is what *she did wrong*, aren't we sending the clear message that it's her fault? Rather than sending the clear message that this man is potentially guilty of a terrible crime. We're punishing the victims rather than perpetrators. </s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I said that, you're the one arguing doctors should be able to be discriminatory based on personal beliefs. [ENDQ] [NEWLINE] Do you mean discretionary? This does not qualify<mask> discrimination. It's not a matter of whether they should have discretion<mask> it comes to elective surgery -<mask> they simply do have the power to say no (and,<mask><mask>, they should). [NEWLINE] [NEWLINE] [STARTQ] My analogy works on a personal autonomy level. We can trash our bodies without the doctor stepping in,<mask> I feel that doctors should provide the services they were trained to do no matter<mask>. [ENDQ] [NEWLINE] <mask> someone comes in and asks the doctor to remove his leg for no reason, he should be obligated to do<mask>?<mask><mask> they ask to be killed? To have a unneccesary procedure that will only cause harm to their bodies? You have a right to do whatever stupid shit you wanna do to yourself - you do not have the right to have a doctor do it for you. [NEWLINE] [NEWLINE] [STARTQ] A gay couple doesn't absolutely need a wedding cake, and no one wants to sell them one<mask> they think the couple is making a huge mistake. By your own argument, the gay couple should just keep expanding their radius of search until they find one that will - rather than labeling a discriminatory practice<mask> such and dealing with it, you are saying the couple should just deal with it. That analogy is a direct mirror for that situation. [ENDQ] [NEWLINE] You are confusing 'discretion' with 'discrimination'.<mask> a doctor will provide a procedure to only white, or straight patients, that is illegal, and no one is arguing for that. They do have a right to consider a patients request and say yes or no based on the information before them. You need to realize the difference here<mask> this conversation will bear any fruit - discrimination based on sex, religion, and other characteristics is illegal - making a decision about your patients request for an optional surgery based on the information at hand is simply doing their job. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I said that, you're the one arguing doctors should be able to be discriminatory based on personal beliefs. [ENDQ] [NEWLINE] Do you mean discretionary? This does not qualify as discrimination. It's not a matter of whether they should have discretion when it comes to elective surgery - because they simply do have the power to say no (and, imo, they should). [NEWLINE] [NEWLINE] [STARTQ] My analogy works on a personal autonomy level. We can trash our bodies without the doctor stepping in, but I feel that doctors should provide the services they were trained to do no matter what. [ENDQ] [NEWLINE] If someone comes in and asks the doctor to remove his leg for no reason, he should be obligated to do so? What if they ask to be killed? To have a unneccesary procedure that will only cause harm to their bodies? You have a right to do whatever stupid shit you wanna do to yourself - you do not have the right to have a doctor do it for you. [NEWLINE] [NEWLINE] [STARTQ] A gay couple doesn't absolutely need a wedding cake, and no one wants to sell them one because they think the couple is making a huge mistake. By your own argument, the gay couple should just keep expanding their radius of search until they find one that will - rather than labeling a discriminatory practice as such and dealing with it, you are saying the couple should just deal with it. That analogy is a direct mirror for that situation. [ENDQ] [NEWLINE] You are confusing 'discretion' with 'discrimination'. If a doctor will provide a procedure to only white, or straight patients, that is illegal, and no one is arguing for that. They do have a right to consider a patients request and say yes or no based on the information before them. You need to realize the difference here if this conversation will bear any fruit - discrimination based on sex, religion, and other characteristics is illegal - making a decision about your patients request for an optional surgery based on the information at hand is simply doing their job. [NEWLINE] </s>
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, you should read my argument carefuly. I am not claiming anything about consent here. I understand, consent is merely agreeing to something. I am not going to over-glorify it, by assigning some other semantics to it. All, I am saying is that, a system based on consent is not just,<mask> of information asymmetry: [URL] [NEWLINE] In ancient days, colonists used to recruit people from Africa and Asia, with the promise a wonderful life,<mask> actually enslave them. Is it just to do that, just<mask> they consented? The entire evolution of legal principles on deception ( [URL] (criminal_law)), culls out the injustice that can be inherent in consent. Hell, you know about con artists!<mask> do you think about their victims? :) [NEWLINE] [NEWLINE] OP, unlike others i am not arguing that the current system is great and that you can move away<mask> you want (I already recognized that it is illegal). I am saying, even<mask> you form a different society,<mask> base its justice system on consent, i am not joining you! :) First,<mask> it is not just at all. Second,<mask> it doesn't scale and is extraordinarily unclear to base any justice principle on.<mask> I bring up my kids<mask> atheists and<mask> they grow up, they sue me: "I didn't consent to being an atheist. You screwed my life." Will I be tried in your court and be sentenced to jail for tort? Note, that unlike previous examples<mask> intent of malice could be proved, in this case there is only best interest in mind,<mask> i may have unwittingly screwed my kid's life,<mask> perceived by them. [NEWLINE] [NEWLINE] [STARTQ] Right now we say "well you can just leave."<mask> you can't. [ENDQ] [NEWLINE] <mask><mask> with you on this. The current social structure makes it incredibly hard to try out alternative ideas. State claims total hegemony over every interesting place, on which some interesting social structures can be tried out!</s>
Label encoding: <s>Well, you should read my argument carefuly. I am not claiming anything about consent here. I understand, consent is merely agreeing to something. I am not going to over-glorify it, by assigning some other semantics to it. All, I am saying is that, a system based on consent is not just, because of information asymmetry: [URL] [NEWLINE] In ancient days, colonists used to recruit people from Africa and Asia, with the promise a wonderful life, but actually enslave them. Is it just to do that, just because they consented? The entire evolution of legal principles on deception ( [URL] (criminal_law)), culls out the injustice that can be inherent in consent. Hell, you know about con artists! What do you think about their victims? :) [NEWLINE] [NEWLINE] OP, unlike others i am not arguing that the current system is great and that you can move away if you want (I already recognized that it is illegal). I am saying, even if you form a different society, but base its justice system on consent, i am not joining you! :) First, because it is not just at all. Second, because it doesn't scale and is extraordinarily unclear to base any justice principle on. If I bring up my kids as atheists and when they grow up, they sue me: "I didn't consent to being an atheist. You screwed my life." Will I be tried in your court and be sentenced to jail for tort? Note, that unlike previous examples where intent of malice could be proved, in this case there is only best interest in mind, although i may have unwittingly screwed my kid's life, as perceived by them. [NEWLINE] [NEWLINE] [STARTQ] Right now we say "well you can just leave." But you can't. [ENDQ] [NEWLINE] I agree with you on this. The current social structure makes it incredibly hard to try out alternative ideas. State claims total hegemony over every interesting place, on which some interesting social structures can be tried out!</s>
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0934, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That it's more valuable to the region that it's in than it would be to sell it to somewhere else. [NEWLINE] [NEWLINE] Oil can turn a profit. There's no amount of profit that would make one region or city willingly give up (or share) its water. [NEWLINE] [NEWLINE] In Texas right now, there's a huge argument over just adding in another neighborhood just outside of San Antonio<mask> locals aren't okay with the additional stress that *housing in the area (you know, grow the city, attract jobs and people, improve the economy)* would cause to the Edwards Aquifer. [NEWLINE] [NEWLINE] Plus, given the amount of water that would need to be imported, and the amount of water that people use (even by conservative estimates), many people wouldn't be able to afford the imported water, unless it's imported at a loss. [NEWLINE] [NEWLINE] Then you get into<mask> can rich people have water and not poor people, yada yada. [NEWLINE] [NEWLINE] Water will never be shared. It's too important. Yeah, modern society doesn't run without oil,<mask> literally nothing runs without water, including the oil industry. [NEWLINE] [NEWLINE] The way we use water in this country needs a complete overhaul, and importing water won't fix it. At best, importing would put a bandaid on it, and would ultimately cause further issues down the road, not to mention it not being profitable, and really no regions having water to spare. [NEWLINE] [NEWLINE] I mean,<mask> place has an overabundance of water that's logistically a reasonable place from which to import water to California? And that can be accomplished in a short period of time? [NEWLINE] [NEWLINE] Water is more valuable than oil. It's more valuable than gold, or diamonds, or Apple stocks, or whatever. The fact that you assume water is a commodity like any other shows<mask> little you appreciate it, and is a perfect example of the kind of attitude that leads to cavalier water consumption in the first place.</s>
Label encoding: <s>That it's more valuable to the region that it's in than it would be to sell it to somewhere else. [NEWLINE] [NEWLINE] Oil can turn a profit. There's no amount of profit that would make one region or city willingly give up (or share) its water. [NEWLINE] [NEWLINE] In Texas right now, there's a huge argument over just adding in another neighborhood just outside of San Antonio because locals aren't okay with the additional stress that *housing in the area (you know, grow the city, attract jobs and people, improve the economy)* would cause to the Edwards Aquifer. [NEWLINE] [NEWLINE] Plus, given the amount of water that would need to be imported, and the amount of water that people use (even by conservative estimates), many people wouldn't be able to afford the imported water, unless it's imported at a loss. [NEWLINE] [NEWLINE] Then you get into why can rich people have water and not poor people, yada yada. [NEWLINE] [NEWLINE] Water will never be shared. It's too important. Yeah, modern society doesn't run without oil, but literally nothing runs without water, including the oil industry. [NEWLINE] [NEWLINE] The way we use water in this country needs a complete overhaul, and importing water won't fix it. At best, importing would put a bandaid on it, and would ultimately cause further issues down the road, not to mention it not being profitable, and really no regions having water to spare. [NEWLINE] [NEWLINE] I mean, what place has an overabundance of water that's logistically a reasonable place from which to import water to California? And that can be accomplished in a short period of time? [NEWLINE] [NEWLINE] Water is more valuable than oil. It's more valuable than gold, or diamonds, or Apple stocks, or whatever. The fact that you assume water is a commodity like any other shows how little you appreciate it, and is a perfect example of the kind of attitude that leads to cavalier water consumption in the first place.</s>
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A reactionary movement to feminism would seek to undo all of feminism's effects - perhaps removing women's right to vote, eliminating them from the workforce, and<mask> on - in order to return society to a previous status quo. [NEWLINE] [NEWLINE] The MRM is not a reactionary movement, it is a _response_ to feminism - particularly feminism's blatant sexism, its inaccurate models for society, and its inability to effectively address men's issues or female privilege.<mask> we seek is not a way back to the 50s, or to the 19th century,<mask> a different way forward. [NEWLINE] [NEWLINE] [STARTQ] A lot of the rhetoric that I hear seems to be anti-feminism [ENDQ] [NEWLINE] Yes, the MRM is anti-feminist. This is<mask> feminism is wrong and it is hurting both men and women. I can hardly think of better reasons to oppose anything. [NEWLINE] [NEWLINE] [STARTQ] I don't think most of the issues that the Men's Rights Movement put<mask> its ideals are necessarily wrong or out place.<mask> to my mind it should be seen<mask> a branch of the feminist movement. [ENDQ] [NEWLINE] There might have been a time<mask> that could be the case. Certainly, the MRM has been influenced by feminism.<mask> feminism has been telling men for decades now that men have no issues, that men's issues don't matter, that men's role in feminism is to be silent, that for men to speak about their issues in feminist spaces is derailing, and -<mask> it's not remotely true - that feminism is handling men's issues. [NEWLINE] [NEWLINE] It is clear that the MRM does not have any place in feminism, and going forward,<mask><mask> we would prefer not to be lumped in with it<mask> some sort of accessory or crazy uncle. [NEWLINE] [NEWLINE] [STARTQ] issues of equality and even men's rights have been better handled by the feminsit movement. [ENDQ] [NEWLINE] No, they really have not, and that's the biggest reason<mask> the MRM exists.</s>
Label encoding: <s>A reactionary movement to feminism would seek to undo all of feminism's effects - perhaps removing women's right to vote, eliminating them from the workforce, and so on - in order to return society to a previous status quo. [NEWLINE] [NEWLINE] The MRM is not a reactionary movement, it is a _response_ to feminism - particularly feminism's blatant sexism, its inaccurate models for society, and its inability to effectively address men's issues or female privilege. What we seek is not a way back to the 50s, or to the 19th century, but a different way forward. [NEWLINE] [NEWLINE] [STARTQ] A lot of the rhetoric that I hear seems to be anti-feminism [ENDQ] [NEWLINE] Yes, the MRM is anti-feminist. This is because feminism is wrong and it is hurting both men and women. I can hardly think of better reasons to oppose anything. [NEWLINE] [NEWLINE] [STARTQ] I don't think most of the issues that the Men's Rights Movement put as its ideals are necessarily wrong or out place. But to my mind it should be seen as a branch of the feminist movement. [ENDQ] [NEWLINE] There might have been a time when that could be the case. Certainly, the MRM has been influenced by feminism. But feminism has been telling men for decades now that men have no issues, that men's issues don't matter, that men's role in feminism is to be silent, that for men to speak about their issues in feminist spaces is derailing, and - although it's not remotely true - that feminism is handling men's issues. [NEWLINE] [NEWLINE] It is clear that the MRM does not have any place in feminism, and going forward, I think we would prefer not to be lumped in with it as some sort of accessory or crazy uncle. [NEWLINE] [NEWLINE] [STARTQ] issues of equality and even men's rights have been better handled by the feminsit movement. [ENDQ] [NEWLINE] No, they really have not, and that's the biggest reason why the MRM exists.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm a woman who is a big fan of both the books and the TV series (currently at the end of A Storm of Swords).<mask><mask> the TV show can veer towards sexism in the ratio of female nudity to male nudity,<mask> gratuitous nudity isn't really a problem in the books themselves.<mask> the majority of main characters are male, GRRM does write many female characters, many of whom prove to be just<mask> strong, courageous, and intelligent<mask> their male counterparts. Arya, Brienne, Asha, and Daenerys are good examples, and only Brienne is described<mask> ugly. All four subvert traditional gender roles, and a good deal of their narratives explores their struggles in proving themselves<mask> equals in a patriarchal society.<mask> would GRRM devote<mask> much time to this aspect of their stories<mask> he was a misogynist? Sansa is<mask> another interesting female character, one who is taught to behave<mask> a proper lady of her time "should", and is then severely punished for it.<mask> the text was misogynistic, wouldn't Sansa instead get rewarded for ascribing to the inferior role in society she accepted? [NEWLINE] [NEWLINE] You<mask> mention the large amounts of rapes that take place in the books. To be honest, they make me cringe to read too,<mask> keep in mind, the use of systematic rape<mask> a war tactic has been around forever and continues to happen in some areas of the world today. I don't think GRRM intends for the reader to get any pleasure out of reading about the rapes. He is being brutally candid about the atrocities of war. Women are raped,<mask> men are<mask> tortured in terrible ways (see Theon). A big theme in the books is the motives and consequences of war. The game of thrones has very terrifying consequences for the people of Westeros. The description of the rapes is one way in which GRRM imagines war would destroy a society.</s>
Label encoding: <s>I'm a woman who is a big fan of both the books and the TV series (currently at the end of A Storm of Swords). I think the TV show can veer towards sexism in the ratio of female nudity to male nudity, but gratuitous nudity isn't really a problem in the books themselves. While the majority of main characters are male, GRRM does write many female characters, many of whom prove to be just as strong, courageous, and intelligent as their male counterparts. Arya, Brienne, Asha, and Daenerys are good examples, and only Brienne is described as ugly. All four subvert traditional gender roles, and a good deal of their narratives explores their struggles in proving themselves as equals in a patriarchal society. Why would GRRM devote so much time to this aspect of their stories if he was a misogynist? Sansa is also another interesting female character, one who is taught to behave as a proper lady of her time "should", and is then severely punished for it. If the text was misogynistic, wouldn't Sansa instead get rewarded for ascribing to the inferior role in society she accepted? [NEWLINE] [NEWLINE] You also mention the large amounts of rapes that take place in the books. To be honest, they make me cringe to read too, but keep in mind, the use of systematic rape as a war tactic has been around forever and continues to happen in some areas of the world today. I don't think GRRM intends for the reader to get any pleasure out of reading about the rapes. He is being brutally candid about the atrocities of war. Women are raped, but men are also tortured in terrible ways (see Theon). A big theme in the books is the motives and consequences of war. The game of thrones has very terrifying consequences for the people of Westeros. The description of the rapes is one way in which GRRM imagines war would destroy a society.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I totally disagree.  People saying hosting your own birthday party is crass and egotistical are completely out of touch with reality.  Birthday parties are a pain in the ass to host, and it's absolutely standard for the birthday person to take charge of organizing and hosting the event,<mask> their the only ones that know who they want to invite and<mask> kind of party they want,<mask> well<mask><mask> day works for them.  Everyone in circles I'm familiar with ASSUME that the birthday person will do<mask> they want, and that they will take charge and host something. <mask> no party happens, it's<mask> the birthday person doesn't want one.  This etiquettehell website seems like a pretentious, completely out-of-touch circle jerk. [NEWLINE] [NEWLINE] <mask>, From your parent comment [NEWLINE] [NEWLINE] [STARTQ] After a certain point couples are considered a social unit, in is inappropriate in most cases to invite one and not the other.<mask> at<mask> point are couples considered a social unit? A good rule of thumb was historically engaged or married, and a more modern view is engaged, married, living together for a significant amount of time. The point is they should be in a long-term committed relationship and have something showing that level of commitment. [ENDQ] [NEWLINE] <mask><mask> this definition has to be a lot more pragmatic for different social situations.  It's understandable to not invite a friend's girlfriend/boyfriend of 3 months to a wedding or other formal event<mask> plates are expensive and pictures will be taken and immortalize. <mask>,<mask> you're having a backyard BBQ and limit it to "serious couples only,"<mask> one of your friends can't bring his GF<mask> everyone else can makes you look like a total douche. [NEWLINE] [NEWLINE] The problem with the OPs situation is that they're making an informal event (a 21st birthday) much more formal than it needs to be,<mask> excluding people looks crass and tacky.</s>
Label encoding: <s>I totally disagree.  People saying hosting your own birthday party is crass and egotistical are completely out of touch with reality.  Birthday parties are a pain in the ass to host, and it's absolutely standard for the birthday person to take charge of organizing and hosting the event, since their the only ones that know who they want to invite and what kind of party they want, as well as what day works for them.  Everyone in circles I'm familiar with ASSUME that the birthday person will do what they want, and that they will take charge and host something.  If no party happens, it's because the birthday person doesn't want one.  This etiquettehell website seems like a pretentious, completely out-of-touch circle jerk. [NEWLINE] [NEWLINE] Also, From your parent comment [NEWLINE] [NEWLINE] [STARTQ] After a certain point couples are considered a social unit, in is inappropriate in most cases to invite one and not the other. So at what point are couples considered a social unit? A good rule of thumb was historically engaged or married, and a more modern view is engaged, married, living together for a significant amount of time. The point is they should be in a long-term committed relationship and have something showing that level of commitment. [ENDQ] [NEWLINE] I think this definition has to be a lot more pragmatic for different social situations.  It's understandable to not invite a friend's girlfriend/boyfriend of 3 months to a wedding or other formal event where plates are expensive and pictures will be taken and immortalize.  However, if you're having a backyard BBQ and limit it to "serious couples only," so one of your friends can't bring his GF when everyone else can makes you look like a total douche. [NEWLINE] [NEWLINE] The problem with the OPs situation is that they're making an informal event (a 21st birthday) much more formal than it needs to be, so excluding people looks crass and tacky.</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think people keep bringing up the same kind of points, and I am having a really hard time seeing<mask> it is you want out of people<mask> this kind of answer doesn't suffice, especially in regards to the weed-smoking thing. [NEWLINE] [NEWLINE] I have grown up in a 100% pro-choice community,<mask><mask><mask> I know, and a lot of people look at it very similar to the weed-smoking thing "I don't like it for myself,<mask> it is not my place." Not wanting something for yourself does not equal thinking something is immoral.<mask> you don't want an abortion for yourself, it might not be the "moral reasoning of supporting life"<mask> much<mask> a strong drive to pass on your genes, a strong desire to be a parent, etc. [NEWLINE] [NEWLINE] I would hope people wouldn't have abortions just<mask> it has been framed largely<mask> a moral debate,<mask> people end up feeling badly about it and I don't like people feeling bad. It is<mask> a costly procedure, much more<mask> than just being on birth control. Lots of people who get abortions are [below the poverty line]( [URL] /),<mask> it is usually a giant burden on them and many times the taxpayer. I<mask> think for all of the pro-lifers out there, it kind of sucks that [Medicaid will help fund the abortions in 15 states]( [URL] ),<mask> anyone in those states is subsidizing something that they whole-heartedly and deeply disagree with. [NEWLINE] [NEWLINE] <mask> I get pregnant, I do not want to get an abortion (I want to have kids).<mask><mask> people should be able to have abortions<mask> it is not my place (and definitely not the governments place) and people who get them will have less happy lives (otherwise, they would just have the kid). Plus unwanted children often get treated like shit and that sucks for the child, the parent, and society at large. [NEWLINE] [NEWLINE] edited for links</s>
Label encoding: <s>I think people keep bringing up the same kind of points, and I am having a really hard time seeing what it is you want out of people if this kind of answer doesn't suffice, especially in regards to the weed-smoking thing. [NEWLINE] [NEWLINE] I have grown up in a 100% pro-choice community, as far as I know, and a lot of people look at it very similar to the weed-smoking thing "I don't like it for myself, but it is not my place." Not wanting something for yourself does not equal thinking something is immoral. If you don't want an abortion for yourself, it might not be the "moral reasoning of supporting life" so much as a strong drive to pass on your genes, a strong desire to be a parent, etc. [NEWLINE] [NEWLINE] I would hope people wouldn't have abortions just because it has been framed largely as a moral debate, so people end up feeling badly about it and I don't like people feeling bad. It is also a costly procedure, much more so than just being on birth control. Lots of people who get abortions are [below the poverty line]( [URL] /), so it is usually a giant burden on them and many times the taxpayer. I also think for all of the pro-lifers out there, it kind of sucks that [Medicaid will help fund the abortions in 15 states]( [URL] ), so anyone in those states is subsidizing something that they whole-heartedly and deeply disagree with. [NEWLINE] [NEWLINE] If I get pregnant, I do not want to get an abortion (I want to have kids). I think people should be able to have abortions because it is not my place (and definitely not the governments place) and people who get them will have less happy lives (otherwise, they would just have the kid). Plus unwanted children often get treated like shit and that sucks for the child, the parent, and society at large. [NEWLINE] [NEWLINE] edited for links</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The difference being in your example was that it was his own son.   In OP's, it is merely friends of theirs.   I doubt any of OP's family would disown them for being friends with a gay person.   They might likely judge them for their choice of friends, perhaps even harshly,<mask> I seriously doubt OP would be in danger of being disowned. [NEWLINE] [NEWLINE] I have a good gay friend that was at my wedding.  Plenty of family members who could be known to judge harshly.  Lucky for me my buddy likes to dance with the ladies, not other guys.  At any social even you'll find him dancing and flirting with girls, not guys.   He just likes to have a good time and is very feminine.   My close family members already knew him and mostly just don't say anything<mask> most of the people at the wedding had no idea the guy was gay.   It<mask> really helps in the case of him and my family that he's a _really good_ guy.   Meaning, really stand up, shakes your hand, looks you in the guy, very personable, professional and kind person.  <mask> anyone that meets him generally has a very  hard time not liking the guy.   I've actually used him on several occasion to convince people that frown upon gay people to help change their minds.  He is aware I do this and has no problem playing the part.   He is not the kind of guy to be easily offended either<mask>, he easily shakes off judgement and goes on with his life and you would _never_ catch him calling anyone a bigot.   He would believe that would be unfair judgement of other people for their own opinions.   The world needs more people like him. [NEWLINE] [NEWLINE] Edit:  Called my friend an "it" instead of "him".  Typing is hard. </s>
Label encoding: <s>The difference being in your example was that it was his own son.   In OP's, it is merely friends of theirs.   I doubt any of OP's family would disown them for being friends with a gay person.   They might likely judge them for their choice of friends, perhaps even harshly, but I seriously doubt OP would be in danger of being disowned. [NEWLINE] [NEWLINE] I have a good gay friend that was at my wedding.  Plenty of family members who could be known to judge harshly.  Lucky for me my buddy likes to dance with the ladies, not other guys.  At any social even you'll find him dancing and flirting with girls, not guys.   He just likes to have a good time and is very feminine.   My close family members already knew him and mostly just don't say anything but most of the people at the wedding had no idea the guy was gay.   It also really helps in the case of him and my family that he's a _really good_ guy.   Meaning, really stand up, shakes your hand, looks you in the guy, very personable, professional and kind person.   So anyone that meets him generally has a very  hard time not liking the guy.   I've actually used him on several occasion to convince people that frown upon gay people to help change their minds.  He is aware I do this and has no problem playing the part.   He is not the kind of guy to be easily offended either though, he easily shakes off judgement and goes on with his life and you would _never_ catch him calling anyone a bigot.   He would believe that would be unfair judgement of other people for their own opinions.   The world needs more people like him. [NEWLINE] [NEWLINE] Edit:  Called my friend an "it" instead of "him".  Typing is hard. </s>
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you more or less were spot-on with some of your examples of "black culture",<mask> to be fair, it's hard to tell<mask> is real and<mask> is just coming from rappers in the media sometimes.<mask> try to think of your underlying assumptions<mask> you list the degenerate aspects. [NEWLINE] [NEWLINE] For instance, #5... the assumption here is "being polite/considerate is never a weakness".<mask><mask><mask> it is?<mask> you voluntarily tip a person, *you* lose money that you weren't actually obligated to spend.<mask> you go home, *you* go home slightly poorer than you could've been. I have never tipped in restaurants<mask><mask><mask><mask> they really wanted a "service fee" they'd just charge you for it explicitly. Call me an asshole, call me selfish,<mask><mask> it's all said and done I came away with more money than I would have<mask> I tipped. I've got to be looking out for myself, not some other guy/girl bringing us plates. [NEWLINE] [NEWLINE] Now, you listen some other situations with people who aren't really gaining anything physical with their rudeness.<mask>, even in these cases I don't think the person is doing it just to be a dick - again, you could call this selfishness at best. The guys in the theater their communication more important than the experience of the other patrons, the guys in the alley think they're more important than the alley traffic, etc... the payoffs have changed here,<mask> the underlying ideology is the same: it's all about me, not other people.<mask> you could call this behavior morally inferior, from a strong/weak perspective it benefits the individual,<mask> I would say #5 is<mask><mask> NOT necessarily a flaw. [NEWLINE] [NEWLINE] <mask> for the rest, I pretty much agree. Considering I'm black myself I guess that makes me an Uncle Tom/house nigga/Oreo/etc...</s>
Label encoding: <s>I think you more or less were spot-on with some of your examples of "black culture", though to be fair, it's hard to tell what is real and what is just coming from rappers in the media sometimes. But try to think of your underlying assumptions when you list the degenerate aspects. [NEWLINE] [NEWLINE] For instance, #5... the assumption here is "being polite/considerate is never a weakness". But what if it is? When you voluntarily tip a person, *you* lose money that you weren't actually obligated to spend. When you go home, *you* go home slightly poorer than you could've been. I have never tipped in restaurants because I think if they really wanted a "service fee" they'd just charge you for it explicitly. Call me an asshole, call me selfish, but when it's all said and done I came away with more money than I would have if I tipped. I've got to be looking out for myself, not some other guy/girl bringing us plates. [NEWLINE] [NEWLINE] Now, you listen some other situations with people who aren't really gaining anything physical with their rudeness. However, even in these cases I don't think the person is doing it just to be a dick - again, you could call this selfishness at best. The guys in the theater their communication more important than the experience of the other patrons, the guys in the alley think they're more important than the alley traffic, etc... the payoffs have changed here, but the underlying ideology is the same: it's all about me, not other people. While you could call this behavior morally inferior, from a strong/weak perspective it benefits the individual, so I would say #5 is in fact NOT necessarily a flaw. [NEWLINE] [NEWLINE] As for the rest, I pretty much agree. Considering I'm black myself I guess that makes me an Uncle Tom/house nigga/Oreo/etc...</s>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're advocating "war" on personal choices. It's not up to you or anybody to decide<mask>'s right for someone. Many "healthy" looking people live unhealthy lifestyles in ways you can't imagine. You<mask> seem to believe that this pressure isn't already there? A much heavier focus on trying to be healthy?<mask> the hell does that even mean? <mask> do you care? It's<mask> funny to me that people pretend to have other's best interests at heart<mask> all they really want to to stop looking at things they deem unsightly. I'm basically positing that your motives are selfish, not magnanimous. Trying to control the way people live their lives is a losing battle.<mask> not become a doctor<mask> you're<mask> passionate about educating people about their health?<mask> about we make sure everyone has a roof over their head before we start to focus on those who choose to let their bodies degrade? Ever hear of mental health problems? Those are much more prevalent and [NEWLINE] important to fix. Hell, most binge eating is 90 percent mental anyway. [NEWLINE] [NEWLINE] <mask><mask> you're<mask> totally ignoring the socio-economic factors at play in Mexico and America that have lead to a rise in obese people. It's easy to say we should "pressure" the family of five that can only afford the cheapest, least nutritional, foodstuffs.<mask> it starts to fall apart<mask> you look at the realities... Do you fine people for being overweight? Pay more for healthcare? Send them to jail? Slippery Slope much? Shouldn't we be taxing or regulating the sugar industry instead of decrying its' victims? [NEWLINE] [NEWLINE] Honestly, there's<mask> many issues I have with this viewpoint that I can't even keep my counter points straight. You're trying to make this a binary issue of fat=unhealthy=bad vs. skinny=healthy=good. And it's just<mask> much more complex then that. </s>
Label encoding: <s>You're advocating "war" on personal choices. It's not up to you or anybody to decide what's right for someone. Many "healthy" looking people live unhealthy lifestyles in ways you can't imagine. You also seem to believe that this pressure isn't already there? A much heavier focus on trying to be healthy? What the hell does that even mean?  Why do you care? It's so funny to me that people pretend to have other's best interests at heart when all they really want to to stop looking at things they deem unsightly. I'm basically positing that your motives are selfish, not magnanimous. Trying to control the way people live their lives is a losing battle. Why not become a doctor if you're so passionate about educating people about their health? How about we make sure everyone has a roof over their head before we start to focus on those who choose to let their bodies degrade? Ever hear of mental health problems? Those are much more prevalent and [NEWLINE] important to fix. Hell, most binge eating is 90 percent mental anyway. [NEWLINE] [NEWLINE] I think you're also totally ignoring the socio-economic factors at play in Mexico and America that have lead to a rise in obese people. It's easy to say we should "pressure" the family of five that can only afford the cheapest, least nutritional, foodstuffs. But it starts to fall apart when you look at the realities... Do you fine people for being overweight? Pay more for healthcare? Send them to jail? Slippery Slope much? Shouldn't we be taxing or regulating the sugar industry instead of decrying its' victims? [NEWLINE] [NEWLINE] Honestly, there's so many issues I have with this viewpoint that I can't even keep my counter points straight. You're trying to make this a binary issue of fat=unhealthy=bad vs. skinny=healthy=good. And it's just so much more complex then that. </s>
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] And<mask> the individual wanting to commit is suicidal than the armed guards would directly solve the problem and stop the would-be gunmen before the tragedy even starts. [ENDQ] [NEWLINE] Well, these guards wouldn't be psychic. [NEWLINE] [NEWLINE] Truth is, in a hypothetical situation<mask> the shooter is suicidal (<mask> seems to frequently be the case), he's going to kill a few people before the guards are able to take him down. [NEWLINE] [NEWLINE] Do the guards respond quickly enough that he kills less people than he would have otherwise? Maybe. [NEWLINE] [NEWLINE] <mask> you've now introduced a group of armed guards firing on the suspect, which means more bullets flying through the air to potentially hit innocent bystanders, and this is likely happening *before* a proper evacuation can take place. [NEWLINE] [NEWLINE] <mask><mask> kind of armed guards are we talking about? Can't be cops,<mask> cops already have jobs.<mask>... private security contractors?<mask> kind of training do they have? [NEWLINE] [NEWLINE] Let's<mask> think about<mask> it may take away from the primary goal of the school: education. Imagine<mask> much learning time gets interrupted and lost due to conversations about the men with guns outside. [NEWLINE] [NEWLINE] Bottom line is, shootings have happened in plenty of places<mask> armed guards *were* posted, and the guards were only able to respond once the subject had begun their assault. You can argue for the deterrent effect, that maybe some shootings are stopped in advance simply *<mask> * armed guards are present,<mask> it seems that more often than not, the subject is suicidal or prepared to go down in a hail of gunfire. [NEWLINE] [NEWLINE] Frankly, I don't think there's much of a solution for school shootings. The goose is cooked,<mask> to speak. We've allowed everyone and their grandmother to have a gun in this country, to the point<mask> there are more guns than there are people. Shootings are going to happen. Armed guards just mean more bullets in the air.</s>
Label encoding: <s> [STARTQ] And if the individual wanting to commit is suicidal than the armed guards would directly solve the problem and stop the would-be gunmen before the tragedy even starts. [ENDQ] [NEWLINE] Well, these guards wouldn't be psychic. [NEWLINE] [NEWLINE] Truth is, in a hypothetical situation where the shooter is suicidal ( as seems to frequently be the case), he's going to kill a few people before the guards are able to take him down. [NEWLINE] [NEWLINE] Do the guards respond quickly enough that he kills less people than he would have otherwise? Maybe. [NEWLINE] [NEWLINE] But you've now introduced a group of armed guards firing on the suspect, which means more bullets flying through the air to potentially hit innocent bystanders, and this is likely happening *before* a proper evacuation can take place. [NEWLINE] [NEWLINE] So what kind of armed guards are we talking about? Can't be cops, because cops already have jobs. So... private security contractors? What kind of training do they have? [NEWLINE] [NEWLINE] Let's also think about how it may take away from the primary goal of the school: education. Imagine how much learning time gets interrupted and lost due to conversations about the men with guns outside. [NEWLINE] [NEWLINE] Bottom line is, shootings have happened in plenty of places where armed guards *were* posted, and the guards were only able to respond once the subject had begun their assault. You can argue for the deterrent effect, that maybe some shootings are stopped in advance simply * because * armed guards are present, but it seems that more often than not, the subject is suicidal or prepared to go down in a hail of gunfire. [NEWLINE] [NEWLINE] Frankly, I don't think there's much of a solution for school shootings. The goose is cooked, so to speak. We've allowed everyone and their grandmother to have a gun in this country, to the point where there are more guns than there are people. Shootings are going to happen. Armed guards just mean more bullets in the air.</s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Art" porn is an interesting test case,<mask> in a more serious vein... [NEWLINE] [NEWLINE] Your statement is probably way over-generalized. [NEWLINE] [NEWLINE] <mask><mask>, for example, that there is art whose *purpose* is to shed light on the gender roles that cause one or the other gender to be attracted to it, and that it would<mask> be *failing* in its primary purpose in its time and place<mask> it did<mask> you are saying it "should" do. [NEWLINE] [NEWLINE] Of course, in the larger sense, it would be wildly successful art<mask> some day gender roles changed such that this were no longer the case...<mask> it would no longer be serving its purpose. [NEWLINE] [NEWLINE] <mask><mask> to your main point...<mask> a topic area is, for social reasons unrelated to art, attractive to one gender, you seem to be saying that it's impossible to have "good" art about that topic. For example, that it's impossible to have a stunning picture of a football play, or a fantastically good floral arrangement (<mask> only outside of Japan, oddly). [NEWLINE] [NEWLINE] In a sense, you're saying that art about a gender biased topic can't be good unless it's really, fundamentally, not about that gender biased topic. I would argue, that<mask> it "transcends" its category, it's failed in its mission to be about that category. It might be interesting art in its own way,<mask> it's no longer speaking about<mask> the artist was trying to speak about. [NEWLINE] [NEWLINE] And<mask>'s<mask> special about gender bias? Is it impossible for there to be "good rap music"<mask> (hypothetically) that appeals, statistically, more to black people? Or any good classical music<mask>, hypothetically, it appealed less to them? [NEWLINE] [NEWLINE] Can there be no good "genre" art? Does it have to stop being its genre to do<mask>?</s>
Label encoding: <s>"Art" porn is an interesting test case, but in a more serious vein... [NEWLINE] [NEWLINE] Your statement is probably way over-generalized. [NEWLINE] [NEWLINE] I think, for example, that there is art whose *purpose* is to shed light on the gender roles that cause one or the other gender to be attracted to it, and that it would indeed be *failing* in its primary purpose in its time and place if it did what you are saying it "should" do. [NEWLINE] [NEWLINE] Of course, in the larger sense, it would be wildly successful art if some day gender roles changed such that this were no longer the case... But it would no longer be serving its purpose. [NEWLINE] [NEWLINE] But as to your main point... if a topic area is, for social reasons unrelated to art, attractive to one gender, you seem to be saying that it's impossible to have "good" art about that topic. For example, that it's impossible to have a stunning picture of a football play, or a fantastically good floral arrangement ( but only outside of Japan, oddly). [NEWLINE] [NEWLINE] In a sense, you're saying that art about a gender biased topic can't be good unless it's really, fundamentally, not about that gender biased topic. I would argue, that if it "transcends" its category, it's failed in its mission to be about that category. It might be interesting art in its own way, but it's no longer speaking about what the artist was trying to speak about. [NEWLINE] [NEWLINE] And what's so special about gender bias? Is it impossible for there to be "good rap music" because (hypothetically) that appeals, statistically, more to black people? Or any good classical music if, hypothetically, it appealed less to them? [NEWLINE] [NEWLINE] Can there be no good "genre" art? Does it have to stop being its genre to do so?</s>
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Vulcans are individually strong and smart... and after 200 years they couldn't even comprehensively end a conflict with the Andorians. They had massively better technology than others in the area and the Tellarites were natural allies<mask> they<mask> hated the Andorians<mask> the Vulcans still couldn't get their act together and calm things down. [NEWLINE] [NEWLINE] The problem is,<mask> they are<mask> individually superior and logical, they have trouble interacting with other races and different stories go back and forth on whether they are arrogant or just seem that way to everyone else,<mask> neither is a good trait for working with other (<mask> you see in the above example of conflict). This is remarked upon<mask> one of humanities greatest traits (Babel One, These are the Voyages...) [NEWLINE] [NEWLINE] Further, they are<mask> inherently emotional they have to completely suppress it to not murder each other. This means they can't trust their intuition or the like, and<mask> cannot take advantage of the brains inherent ability to figure things out on a subconcious level. [NEWLINE] [NEWLINE] Finally, you can say humans are<mask> successful from greed.<mask>, after the Third World War, the [New World Economy]( [URL].wikia.com/wiki/New_World_Economy) replace possessions and wealth<mask> primary human pursuits for self-enrichment and the betterment of humanity. One can argue there is still "greed," greed for social acclaim by accomplishment,<mask> I would say that is good! They have turned a negative fact into a positive, and are now on the forefront of exploration of the galaxy, diplomacy, and defense of the Federation (one of the few "not jerk" governments encountered). [NEWLINE] [NEWLINE] Even<mask> Vulcans are better at baseball. Which, whatever. Nobody plays that game anymore anyway, except some weirdos on a farm planet and a run down Cardassian station in the middle of nowhere.</s>
Label encoding: <s>Vulcans are individually strong and smart... and after 200 years they couldn't even comprehensively end a conflict with the Andorians. They had massively better technology than others in the area and the Tellarites were natural allies since they also hated the Andorians but the Vulcans still couldn't get their act together and calm things down. [NEWLINE] [NEWLINE] The problem is, because they are so individually superior and logical, they have trouble interacting with other races and different stories go back and forth on whether they are arrogant or just seem that way to everyone else, though neither is a good trait for working with other ( as you see in the above example of conflict). This is remarked upon as one of humanities greatest traits (Babel One, These are the Voyages...) [NEWLINE] [NEWLINE] Further, they are so inherently emotional they have to completely suppress it to not murder each other. This means they can't trust their intuition or the like, and so cannot take advantage of the brains inherent ability to figure things out on a subconcious level. [NEWLINE] [NEWLINE] Finally, you can say humans are so successful from greed. However, after the Third World War, the [New World Economy]( [URL].wikia.com/wiki/New_World_Economy) replace possessions and wealth as primary human pursuits for self-enrichment and the betterment of humanity. One can argue there is still "greed," greed for social acclaim by accomplishment, but I would say that is good! They have turned a negative fact into a positive, and are now on the forefront of exploration of the galaxy, diplomacy, and defense of the Federation (one of the few "not jerk" governments encountered). [NEWLINE] [NEWLINE] Even if Vulcans are better at baseball. Which, whatever. Nobody plays that game anymore anyway, except some weirdos on a farm planet and a run down Cardassian station in the middle of nowhere.</s>
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your definitions are too narrow to relate to the way you are using the word 'consent' and 'tacit'.<mask>, let's first look at the complete definition of consent and tacit. [NEWLINE] [NEWLINE] [STARTQ] [Consent:]( [URL] ) [ENDQ] intr v. [NEWLINE] 1. To give assent,<mask> to the proposal of another; agree. [NEWLINE] 2. *Archaic* To be of the same mind or opinion. [NEWLINE] n. [NEWLINE] 1. Acceptance or approval of<mask> is planned or done by another; acquiescence. [NEWLINE] 2. Agreement<mask> to opinion or a course of action. [NEWLINE] [NEWLINE] [STARTQ] Assent: To agree,<mask> to a proposal; concur; consent [ENDQ] [NEWLINE] [STARTQ] [Tacit:]( [URL] ) [ENDQ] 1. Not spoken. [NEWLINE] 2. Implied by or inferred from actions or statements; Law Arising by operation of the law rather than through direct expression. [NEWLINE] 3. *Archaic* Not speaking; silent. [NEWLINE] [NEWLINE] <mask> we incorporate all of the definitions, 'tacit consent' becomes: [NEWLINE] * Implied acceptance of<mask> is planned by another. [NEWLINE] [NEWLINE] In the context of nations/governments: [NEWLINE] * Implied acceptance of laws enforced by the governing body by virtue of actions taken conforming to those laws. [NEWLINE] [NEWLINE] In order to be part of and function in a community, one must either explicitly or implicitly accept and follow the laws of that community.<mask> one tries to exist in a community without adhering to those laws, the community has two options: remove or punish the offender. Anyone taking part in the community must assume the other members are following (tacitly consenting) to the laws unless demonstrated otherwise. [NEWLINE] [NEWLINE] We can think of this more concretely too: All tourists tacitly consent to the host country's laws. Otherwise they face punishment or deportation. [NEWLINE] </s>
Label encoding: <s>I think your definitions are too narrow to relate to the way you are using the word 'consent' and 'tacit'. Therefore, let's first look at the complete definition of consent and tacit. [NEWLINE] [NEWLINE] [STARTQ] [Consent:]( [URL] ) [ENDQ] intr v. [NEWLINE] 1. To give assent, as to the proposal of another; agree. [NEWLINE] 2. *Archaic* To be of the same mind or opinion. [NEWLINE] n. [NEWLINE] 1. Acceptance or approval of what is planned or done by another; acquiescence. [NEWLINE] 2. Agreement as to opinion or a course of action. [NEWLINE] [NEWLINE] [STARTQ] Assent: To agree, as to a proposal; concur; consent [ENDQ] [NEWLINE] [STARTQ] [Tacit:]( [URL] ) [ENDQ] 1. Not spoken. [NEWLINE] 2. Implied by or inferred from actions or statements; Law Arising by operation of the law rather than through direct expression. [NEWLINE] 3. *Archaic* Not speaking; silent. [NEWLINE] [NEWLINE] If we incorporate all of the definitions, 'tacit consent' becomes: [NEWLINE] * Implied acceptance of what is planned by another. [NEWLINE] [NEWLINE] In the context of nations/governments: [NEWLINE] * Implied acceptance of laws enforced by the governing body by virtue of actions taken conforming to those laws. [NEWLINE] [NEWLINE] In order to be part of and function in a community, one must either explicitly or implicitly accept and follow the laws of that community. If one tries to exist in a community without adhering to those laws, the community has two options: remove or punish the offender. Anyone taking part in the community must assume the other members are following (tacitly consenting) to the laws unless demonstrated otherwise. [NEWLINE] [NEWLINE] We can think of this more concretely too: All tourists tacitly consent to the host country's laws. Otherwise they face punishment or deportation. [NEWLINE] </s>
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One of the best pieces of satire I've seen in recent memory was Black Mirror. The reason I am such a fan of Black Mirror is that instead of taking an approach that simply "generalizes and mocks" it managed to be extremely cutting without being one-sided. [NEWLINE] [NEWLINE] For instance, one episode (actually my least-favourite) centres on a comedian who voices an extremely unfunny talking bear whose job involves mocking politicians in an extremely crude and apathetic way. He is a professional troll.<mask> a joke, the producers decide to put his name forward<mask> a political candidate for a local election.<mask>, mid-way through one of his terrible "comedy" routines he finally snaps and goes on a massive off-script rant about the state of British politics, and actually manages to galvanise the public into voting for him<mask> a protest candidate. The problem is of course that he stands for nothing<mask> lazy, aggressive, nihilistic apathy, and<mask> he tries to back down his greedy producer takes control of the project and uses it to throw the British political system into chaos. Ultimately the elitist conservative politician he was trolling earlier is the one who ends up being shown in the better light: **"The system may not be perfect,<mask> it built these roads."** [NEWLINE] [NEWLINE] Black Mirror is incredible satire<mask> rather than simply pointing to politicians and calling them dimwits and scumbags, **it skewers the satirists and their crude, apathetic name-calling at the same time.** I've never seen a satire satirise other satire,<mask> Black Mirror did that. Other episodes are similarly ruthless,<mask> not necessarily about politics and apathy<mask> society in general. Often crude, and over the top,<mask> definitely some of the best satire I've seen at least<mask><mask><mask> its message goes. [NEWLINE] (Excuse my British spellings.)</s>
Label encoding: <s>One of the best pieces of satire I've seen in recent memory was Black Mirror. The reason I am such a fan of Black Mirror is that instead of taking an approach that simply "generalizes and mocks" it managed to be extremely cutting without being one-sided. [NEWLINE] [NEWLINE] For instance, one episode (actually my least-favourite) centres on a comedian who voices an extremely unfunny talking bear whose job involves mocking politicians in an extremely crude and apathetic way. He is a professional troll. As a joke, the producers decide to put his name forward as a political candidate for a local election. However, mid-way through one of his terrible "comedy" routines he finally snaps and goes on a massive off-script rant about the state of British politics, and actually manages to galvanise the public into voting for him as a protest candidate. The problem is of course that he stands for nothing but lazy, aggressive, nihilistic apathy, and when he tries to back down his greedy producer takes control of the project and uses it to throw the British political system into chaos. Ultimately the elitist conservative politician he was trolling earlier is the one who ends up being shown in the better light: **"The system may not be perfect, but it built these roads."** [NEWLINE] [NEWLINE] Black Mirror is incredible satire because rather than simply pointing to politicians and calling them dimwits and scumbags, **it skewers the satirists and their crude, apathetic name-calling at the same time.** I've never seen a satire satirise other satire, but Black Mirror did that. Other episodes are similarly ruthless, although not necessarily about politics and apathy but society in general. Often crude, and over the top, but definitely some of the best satire I've seen at least as far as its message goes. [NEWLINE] (Excuse my British spellings.)</s>
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[spring break]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;oq=spring+break&amp;aqs=chrome.0.69i59j0l5.1366j0j9&amp;sourceid=chrome&amp;es_sm=122&amp;ie=UTF-8) [NEWLINE] [NEWLINE] [carnival]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=carnival+rio&amp;revid=255166102) [NEWLINE] [NEWLINE] [mardi gras]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=mardi+gras) [NEWLINE] [NEWLINE] It has nothing to do with 'thinking differently'. It has to do with these 'holidays' frequently involve men and women wearing scantily clad outfits and costumes for each other.<mask> you have actually been around any of these parties, you'd know.</s><pad>
Label encoding: <s>[spring break]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;oq=spring+break&amp;aqs=chrome.0.69i59j0l5.1366j0j9&amp;sourceid=chrome&amp;es_sm=122&amp;ie=UTF-8) [NEWLINE] [NEWLINE] [carnival]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=carnival+rio&amp;revid=255166102) [NEWLINE] [NEWLINE] [mardi gras]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=mardi+gras) [NEWLINE] [NEWLINE] It has nothing to do with 'thinking differently'. It has to do with these 'holidays' frequently involve men and women wearing scantily clad outfits and costumes for each other. If you have actually been around any of these parties, you'd know.</s><pad>
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most of your reply focused on equality of opportunity (EoC), and I want to address that a little bit. I am<mask> concerned about EoC<mask><mask><mask> it's legitimately unfair to score people<mask> they started at different places. [NEWLINE] [NEWLINE] <mask> let's say that you've played the game<mask><mask> Capitalism's rulebook. You can imagine a society in which all people start with the same opportunities<mask> you like<mask> it isn't necessary for<mask> concerns me. Let's say that you've done alright for yourself. You have a $120,000 income, married with 2.2 kids and a mortgage.<mask> reason do you have to care about<mask> the CEO is making<mask> you have played the game by the rules and done alright by yourself. Sure you didn't get lucky and land the CEO job,<mask> your lot in life isn't that bad.<mask> should you be concerned about "rising levels of inequality in our society"? [NEWLINE] [NEWLINE] Yours is actually the most promising answer I've responded to<mask>,<mask> only<mask> of a single sentence in your post: [NEWLINE] [STARTQ] Those with power and wealth do things to increase their power and wealth, that kind of hegemony is arguably inevitable, and at the very least it's a persistent feature of societies with widespread inequality. [ENDQ] [NEWLINE] I'd like to explore this further<mask> it signals a reason to care about inequality beyond envy or a feeling that life in unfair. You're suggesting that people with a lot of money are going to use *the power that comes with their money* to coerce others to do<mask> benefits them. You used the word "hegemony" -- a term straight out of Political Realism that is bursting with connotations of an imbalance of *power*, not wealth. This interests me, for I don't care about differences in wealth,<mask> I do care about an even distribution of power (<mask> Democracy). </s>
Label encoding: <s>Most of your reply focused on equality of opportunity (EoC), and I want to address that a little bit. I am also concerned about EoC because I think it's legitimately unfair to score people if they started at different places. [NEWLINE] [NEWLINE] But let's say that you've played the game according to Capitalism's rulebook. You can imagine a society in which all people start with the same opportunities if you like but it isn't necessary for what concerns me. Let's say that you've done alright for yourself. You have a $120,000 income, married with 2.2 kids and a mortgage. What reason do you have to care about what the CEO is making if you have played the game by the rules and done alright by yourself. Sure you didn't get lucky and land the CEO job, but your lot in life isn't that bad. Why should you be concerned about "rising levels of inequality in our society"? [NEWLINE] [NEWLINE] Yours is actually the most promising answer I've responded to yet, but only because of a single sentence in your post: [NEWLINE] [STARTQ] Those with power and wealth do things to increase their power and wealth, that kind of hegemony is arguably inevitable, and at the very least it's a persistent feature of societies with widespread inequality. [ENDQ] [NEWLINE] I'd like to explore this further because it signals a reason to care about inequality beyond envy or a feeling that life in unfair. You're suggesting that people with a lot of money are going to use *the power that comes with their money* to coerce others to do what benefits them. You used the word "hegemony" -- a term straight out of Political Realism that is bursting with connotations of an imbalance of *power*, not wealth. This interests me, for I don't care about differences in wealth, but I do care about an even distribution of power ( because Democracy). </s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're making a mistake by positing that it has to be either (a) relevant or (b) character assassination. [NEWLINE] [NEWLINE] <mask><mask> that it is relevant<mask> it goes towards establishing<mask> Mr. Brown's state of mind might have been at the time of the altercation. To the extent that figuring out<mask> happened there depends on figuring out which elements of competing accounts are plausible, knowing that Mr. Brown might have committed a crime earlier seems relevant, even<mask> the police officer was not aware of this. [NEWLINE] [NEWLINE] <mask>, the way in which the authorities released the information seems to indicate that they might<mask> have had an interest in painting Mr. Brown in the worst possible light,<mask> making him seem like less of a victim. I would call that 'character assassination'. [NEWLINE] [NEWLINE] The local police have mishandled numerous elements of the investigation and the crowd-control challenges they have faced, they seem dug-in and defensive. That context is at least in relevant to understanding their state of mind<mask> releasing the video<mask> Mr. Brown's alleged crime is to understanding his state of mind during the incident that led to his death. [NEWLINE] [NEWLINE] <mask>, let me add that 'character assassination' wouldn't be my prefered description of<mask> the police are doing; rather,<mask><mask> they're trying to muddy the waters. Releasing the video has helped them shift the conversation away from the central question that should be asked about<mask> happened: did Mr. Brown pose an imminent threat to the police officer or others? Instead, they want us to ask: could Mr. Brown have been combative rather than,<mask> we were encouraged to believe early on, cooperative? [NEWLINE] [NEWLINE] <mask> a sworn law enforcement officer shoots an unarmed civilian<mask> the result of confrontation stemming from a civil infraction, that burden of proof should rest on that police officer. This video seems like a distraction from that.</s>
Label encoding: <s>I think you're making a mistake by positing that it has to be either (a) relevant or (b) character assassination. [NEWLINE] [NEWLINE] I agree that it is relevant because it goes towards establishing what Mr. Brown's state of mind might have been at the time of the altercation. To the extent that figuring out what happened there depends on figuring out which elements of competing accounts are plausible, knowing that Mr. Brown might have committed a crime earlier seems relevant, even if the police officer was not aware of this. [NEWLINE] [NEWLINE] However, the way in which the authorities released the information seems to indicate that they might also have had an interest in painting Mr. Brown in the worst possible light, therefore making him seem like less of a victim. I would call that 'character assassination'. [NEWLINE] [NEWLINE] The local police have mishandled numerous elements of the investigation and the crowd-control challenges they have faced, they seem dug-in and defensive. That context is at least in relevant to understanding their state of mind when releasing the video as Mr. Brown's alleged crime is to understanding his state of mind during the incident that led to his death. [NEWLINE] [NEWLINE] Lastly, let me add that 'character assassination' wouldn't be my prefered description of what the police are doing; rather, I think they're trying to muddy the waters. Releasing the video has helped them shift the conversation away from the central question that should be asked about what happened: did Mr. Brown pose an imminent threat to the police officer or others? Instead, they want us to ask: could Mr. Brown have been combative rather than, as we were encouraged to believe early on, cooperative? [NEWLINE] [NEWLINE] When a sworn law enforcement officer shoots an unarmed civilian as the result of confrontation stemming from a civil infraction, that burden of proof should rest on that police officer. This video seems like a distraction from that.</s>
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Typical<mask>ane lighter fluid would not store in a super<mask>aker<mask> it would leak too fast. It only stores<mask> a liquid at very low temperatures and/or very high pressure. Super soakers are notoriously porous.<mask>, ronsonol might store for long periods of time. [NEWLINE] [NEWLINE] <mask> here's my problem with your proposition. [NEWLINE] [NEWLINE] 1. Things can go wrong<mask> you rely on both your lighter and your super soaker. You might have a faulty lighter. Or you might have to click your lighter multiple times to ignite it. You have to prime your super soaker. It doesn't stay pressurized forever, so who knows<mask> often or<mask> you'll have to re-prime it. Either way, the extra step is an inconvenience. [NEWLINE] [NEWLINE] 2. It require 2 hands. It requires the concentration of lining up and maintaining the position of the 2 lines up.<mask>, by occupying both hands you prevent yourself from multitasking or using a free arm to defend/attack/grab. [NEWLINE] [NEWLINE] 3. The lighter gets hot.<mask> long can you hold a lit lighter? I realize<mask> you hold it right and everything goes right this shouldn't be a problem, but I've found that I typically have to release the lighter after a period of time. [NEWLINE] [NEWLINE] 4. Someone could blow the lighter out. In the time it takes to relight and align with the gun, you'll probably be disarmed. [NEWLINE] [NEWLINE] 5. It's possible you'll end up with fluid on yourself and setting yourself aflame. [NEWLINE] [NEWLINE] 6. Range. You might have a 15 foot range.<mask> another guy with a gun has much more. [NEWLINE] [NEWLINE] Cons: unwieldy, requires prep, faulty, vulnerable to lighter extinguishment, occupies both arms leaving no free arms causing more vulnerability, danger to self, range [NEWLINE] [NEWLINE] Pros: creativity</s>
Label encoding: <s>Typical butane lighter fluid would not store in a super soaker because it would leak too fast. It only stores as a liquid at very low temperatures and/or very high pressure. Super soakers are notoriously porous. However, ronsonol might store for long periods of time. [NEWLINE] [NEWLINE] But here's my problem with your proposition. [NEWLINE] [NEWLINE] 1. Things can go wrong when you rely on both your lighter and your super soaker. You might have a faulty lighter. Or you might have to click your lighter multiple times to ignite it. You have to prime your super soaker. It doesn't stay pressurized forever, so who knows how often or when you'll have to re-prime it. Either way, the extra step is an inconvenience. [NEWLINE] [NEWLINE] 2. It require 2 hands. It requires the concentration of lining up and maintaining the position of the 2 lines up. Also, by occupying both hands you prevent yourself from multitasking or using a free arm to defend/attack/grab. [NEWLINE] [NEWLINE] 3. The lighter gets hot. How long can you hold a lit lighter? I realize if you hold it right and everything goes right this shouldn't be a problem, but I've found that I typically have to release the lighter after a period of time. [NEWLINE] [NEWLINE] 4. Someone could blow the lighter out. In the time it takes to relight and align with the gun, you'll probably be disarmed. [NEWLINE] [NEWLINE] 5. It's possible you'll end up with fluid on yourself and setting yourself aflame. [NEWLINE] [NEWLINE] 6. Range. You might have a 15 foot range. But another guy with a gun has much more. [NEWLINE] [NEWLINE] Cons: unwieldy, requires prep, faulty, vulnerable to lighter extinguishment, occupies both arms leaving no free arms causing more vulnerability, danger to self, range [NEWLINE] [NEWLINE] Pros: creativity</s>
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Our ethics class was mainly focused classical points (<mask> the professor specialized in Aristotle)<mask> considering the course<mask> had to satisfy a college general requirement there were limits.<mask> the name was Adam Smith, and it was invisible hand, apologies. We did talk about it a bit jokingly<mask> well,<mask> it did enough good to help start some discussion with non-philosophy majors. (<mask> once again, most of our philosophy courses had to<mask> satisfy other requirements. It wasn't necessarily something the professors wanted to do.) [NEWLINE] [NEWLINE] I did<mask> in my own time discuss certain topics/people who did not come up in classes with my professors. They were not ignorant of these individuals- just not simply<mask> we focused on. Frankly, I don't specialize in those areas,<mask> they are out of my reach. I hope that I can expand my knowledge<mask> I continue forward in my studies<mask>. Things are<mask> taught with a different focus in USA departments it appears, and I'm not going to necessarily condemn my department for focusing on<mask> they chose to focus on. [NEWLINE] [NEWLINE] We did<mask> choose to study Rawls, Mill, Marx and others in said ethics course- Popper came up in a philosophy of science course. [NEWLINE] [NEWLINE] You<mask> should be aware that here, things like this aren't briefed in high school for the most part. The general "idea" on philosophy is rather skewed in the public's eye. It contributes to some of the problems in<mask> the subject may be understood and taught. Trust me, I wish we had more flexibility<mask> I studied,<mask> the departments are getting smaller and budgets are thinning. Philosophy is seen<mask> rather useless by many here. Which is unfortunate. Not trying to excuse my ignorance, trust me,<mask> your non-USA experience is much different than<mask> things are taught here. </s>
Label encoding: <s>Our ethics class was mainly focused classical points ( as the professor specialized in Aristotle) but considering the course also had to satisfy a college general requirement there were limits. Also the name was Adam Smith, and it was invisible hand, apologies. We did talk about it a bit jokingly as well, but it did enough good to help start some discussion with non-philosophy majors. ( As once again, most of our philosophy courses had to also satisfy other requirements. It wasn't necessarily something the professors wanted to do.) [NEWLINE] [NEWLINE] I did however in my own time discuss certain topics/people who did not come up in classes with my professors. They were not ignorant of these individuals- just not simply what we focused on. Frankly, I don't specialize in those areas, so they are out of my reach. I hope that I can expand my knowledge as I continue forward in my studies however. Things are indeed taught with a different focus in USA departments it appears, and I'm not going to necessarily condemn my department for focusing on what they chose to focus on. [NEWLINE] [NEWLINE] We did also choose to study Rawls, Mill, Marx and others in said ethics course- Popper came up in a philosophy of science course. [NEWLINE] [NEWLINE] You also should be aware that here, things like this aren't briefed in high school for the most part. The general "idea" on philosophy is rather skewed in the public's eye. It contributes to some of the problems in how the subject may be understood and taught. Trust me, I wish we had more flexibility where I studied, but the departments are getting smaller and budgets are thinning. Philosophy is seen as rather useless by many here. Which is unfortunate. Not trying to excuse my ignorance, trust me, but your non-USA experience is much different than how things are taught here. </s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There is nothing inherently cruel about raising animals for food [ENDQ] [NEWLINE] Well, is there anything "inherently" cruel about anything? This<mask> ignores the real life conditions of some farming situations, which<mask><mask><mask> could be considered quite cruel. [NEWLINE] [NEWLINE] [STARTQ] Vegetarians do not live longer. [ENDQ] [NEWLINE] And? Doesn't this just mean that it is, at least, **equal**? [NEWLINE] [NEWLINE] [STARTQ] Eating meat has been an essential part of human evolution for 2.3 million years. [ENDQ] [NEWLINE] I dislike doing the whole fallacy thing,<mask> you can't assume that just<mask> something has been an "essential" part of human evolution for a long time that it is necessarily good. [NEWLINE] [NEWLINE] [STARTQ] Eating meat is not cruel or unethical; it is a natural part of the cycle of life. [ENDQ] [NEWLINE] To quote The Dude, "that's just, like, your opinion man."<mask> is or isn't cruel or unethical is subjective and honestly in this situation just about up to the individual to decide for themselves. [NEWLINE] [NEWLINE] [STARTQ] <mask> omnivores it is natural we eat both meat and plants for food [ENDQ] [NEWLINE] This is similar to your last two points.<mask> omnivores we *can* eat both meat and plants,<mask><mask><mask> omnivores we can exclusively survive on just plants. We are cognizant humans, we don't have to do something just<mask> we can. [NEWLINE] [NEWLINE] [STARTQ] Meat farming is no more destructive then traditional agriculture [ENDQ] [NEWLINE] Depends on the type of farming, IIRC. [NEWLINE] [NEWLINE] [STARTQ] Everything must be taken in moderation [ENDQ] [NEWLINE] This point stuck out to me, I'm not sure<mask> it has to do with vegetarianism/veganism. Are you positing that "everything" in this statement is literal? And we must,<mask>, take *everything* in moderation?</s>
Label encoding: <s> [STARTQ] There is nothing inherently cruel about raising animals for food [ENDQ] [NEWLINE] Well, is there anything "inherently" cruel about anything? This also ignores the real life conditions of some farming situations, which in my opinion could be considered quite cruel. [NEWLINE] [NEWLINE] [STARTQ] Vegetarians do not live longer. [ENDQ] [NEWLINE] And? Doesn't this just mean that it is, at least, **equal**? [NEWLINE] [NEWLINE] [STARTQ] Eating meat has been an essential part of human evolution for 2.3 million years. [ENDQ] [NEWLINE] I dislike doing the whole fallacy thing, but you can't assume that just because something has been an "essential" part of human evolution for a long time that it is necessarily good. [NEWLINE] [NEWLINE] [STARTQ] Eating meat is not cruel or unethical; it is a natural part of the cycle of life. [ENDQ] [NEWLINE] To quote The Dude, "that's just, like, your opinion man." What is or isn't cruel or unethical is subjective and honestly in this situation just about up to the individual to decide for themselves. [NEWLINE] [NEWLINE] [STARTQ] As omnivores it is natural we eat both meat and plants for food [ENDQ] [NEWLINE] This is similar to your last two points. As omnivores we *can* eat both meat and plants, but also as omnivores we can exclusively survive on just plants. We are cognizant humans, we don't have to do something just because we can. [NEWLINE] [NEWLINE] [STARTQ] Meat farming is no more destructive then traditional agriculture [ENDQ] [NEWLINE] Depends on the type of farming, IIRC. [NEWLINE] [NEWLINE] [STARTQ] Everything must be taken in moderation [ENDQ] [NEWLINE] This point stuck out to me, I'm not sure what it has to do with vegetarianism/veganism. Are you positing that "everything" in this statement is literal? And we must, indeed, take *everything* in moderation?</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you've done the bulk of the view changing,<mask> I just want to respond to some of the OPs points more directly. [NEWLINE] [NEWLINE] [STARTQ] Part of this drop in the quality of music comes from this. In today's music, instead of having a writer who understands music theory, one just needs a computer, a program or two, and a singer who appeals to the broadest demographic. [ENDQ] [NEWLINE] That's just not true at all. First off, music theory doesn't even need to be understood in order to make music. There are many fantastic musicians who just go by ear and make great music. Music theory is important to learn,<mask> it's essentially a tool, and everyone uses it differently. Not following music theory does not mean a musician or a piece of music is bad. Similarly, following music theory does not make a masterpiece. [NEWLINE] [NEWLINE] [STARTQ] <mask> happened to writing a song about something that has happened to you, or made an impact on you or someone important to you? One of my favorite songs, "Hey Jude" by The Beatles, was written by Paul McCartney [1] to comfort John Lennon's five year old son, Julian. [ENDQ] [NEWLINE] Again, not true. You don't even need to leave pop music to see that.<mask> do you think half of Taylor Swift's songs are about? And<mask> you mention yourself, singing about sex and drugs was practically invented in the 70s. [NEWLINE] [NEWLINE] [STARTQ] Not to add that she can play an instrument (guitar). I'm looking at you, Katy Perry. [ENDQ] [NEWLINE] First off, Katy Perry actually can play guitar. She does<mask> live and<mask> composing songs. [NEWLINE] Second, you've completely discredited the voice<mask> an instrument. Just<mask> it's not a physical thing, that does not mean it is not instrument. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I think you've done the bulk of the view changing, so I just want to respond to some of the OPs points more directly. [NEWLINE] [NEWLINE] [STARTQ] Part of this drop in the quality of music comes from this. In today's music, instead of having a writer who understands music theory, one just needs a computer, a program or two, and a singer who appeals to the broadest demographic. [ENDQ] [NEWLINE] That's just not true at all. First off, music theory doesn't even need to be understood in order to make music. There are many fantastic musicians who just go by ear and make great music. Music theory is important to learn, but it's essentially a tool, and everyone uses it differently. Not following music theory does not mean a musician or a piece of music is bad. Similarly, following music theory does not make a masterpiece. [NEWLINE] [NEWLINE] [STARTQ] What happened to writing a song about something that has happened to you, or made an impact on you or someone important to you? One of my favorite songs, "Hey Jude" by The Beatles, was written by Paul McCartney [1] to comfort John Lennon's five year old son, Julian. [ENDQ] [NEWLINE] Again, not true. You don't even need to leave pop music to see that. What do you think half of Taylor Swift's songs are about? And as you mention yourself, singing about sex and drugs was practically invented in the 70s. [NEWLINE] [NEWLINE] [STARTQ] Not to add that she can play an instrument (guitar). I'm looking at you, Katy Perry. [ENDQ] [NEWLINE] First off, Katy Perry actually can play guitar. She does so live and when composing songs. [NEWLINE] Second, you've completely discredited the voice as an instrument. Just because it's not a physical thing, that does not mean it is not instrument. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> you actually know less than you seem to,<mask> your input was extremely well communicated, and actually does help me see the difference between<mask> I have a problem with (behavioral analytics) and the proper science aspect of Psychology, which I suppose would be called Heuristic Psychology. For that, have a delta. ∆ [NEWLINE] [NEWLINE] I gather now that heuristics would be precisely<mask> would make Psychology something that could be empirically verifiable,<mask> it's<mask> menacingly complex that it wouldn't be easy to do<mask>. With my limited research of dyslexia, I can completely understand<mask> that is a cognitive association process, and<mask> a person learns to read either by generalizing the shape of a word (fast, prone to mistakes) or individually processing letters first (precise<mask> likely slower) could impact their reading ability later in life. I do understand the merits of study in that field,<mask> it's<mask> vastly different from Behavioral Psychology that it seems unfair to lump them together under one banner. [NEWLINE] [NEWLINE] The largest problem I have with Psychiatry is the completely lax standards to which practitioners are held compared to other fields,<mask><mask> much direct influence those practices have on the lives of others. This, to me, is due to gross over-diagnosis of problems that literally anyone could have<mask><mask> the standards of the DSM; literally any child will have ADD/ADHD<mask> practitioners are allowed to cherry-pick and ballpark symptoms, any teen is manic depressive or bi-polar, all adults have developmental irregularities based on events in their formative years -<mask> an overwhelming majority would be<mask> diagnosed *only* due to unchecked confirmation bias. I do believe that such maladies are very real things, and that some can be treated successfully, I simply do not believe that it is the norm.</s>
Label encoding: <s>I don't know if you actually know less than you seem to, because your input was extremely well communicated, and actually does help me see the difference between what I have a problem with (behavioral analytics) and the proper science aspect of Psychology, which I suppose would be called Heuristic Psychology. For that, have a delta. ∆ [NEWLINE] [NEWLINE] I gather now that heuristics would be precisely what would make Psychology something that could be empirically verifiable, though it's so menacingly complex that it wouldn't be easy to do so. With my limited research of dyslexia, I can completely understand how that is a cognitive association process, and how a person learns to read either by generalizing the shape of a word (fast, prone to mistakes) or individually processing letters first (precise but likely slower) could impact their reading ability later in life. I do understand the merits of study in that field, but it's so vastly different from Behavioral Psychology that it seems unfair to lump them together under one banner. [NEWLINE] [NEWLINE] The largest problem I have with Psychiatry is the completely lax standards to which practitioners are held compared to other fields, despite how much direct influence those practices have on the lives of others. This, to me, is due to gross over-diagnosis of problems that literally anyone could have according to the standards of the DSM; literally any child will have ADD/ADHD when practitioners are allowed to cherry-pick and ballpark symptoms, any teen is manic depressive or bi-polar, all adults have developmental irregularities based on events in their formative years - but an overwhelming majority would be so diagnosed *only* due to unchecked confirmation bias. I do believe that such maladies are very real things, and that some can be treated successfully, I simply do not believe that it is the norm.</s>
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0965, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Reading through those... [NEWLINE] [NEWLINE] [STARTQ] “The problem with drinking alcohol during your pregnancy is that there is no amount that has been proven to be safe,” [ENDQ] [NEWLINE] Basically asking people to prove a negative (ie: "It cannot harm")<mask> that's useless to know. Another comment is along those lines<mask> well. It can't be proven that it can't harm,<mask> it might harm. Be careful. On the next page... [NEWLINE] [NEWLINE] [STARTQ] In that study, researchers in the U.K. reported that the 5-year-old children of women who drank up to one to two alcoholic drinks per week or per occasion<mask> pregnant were not at an increased risk of behavioral or cognitive problems [ENDQ] [NEWLINE] This seems pretty positive, and it is looking at the actual end results which is far better than the theorizing being done by the others. [NEWLINE] [NEWLINE] [STARTQ] “The way I see it is:<mask> you wouldn’t give a 2-month-old a glass of wine, then<mask> would you drink a glass of wine<mask> you’re pregnant?” Garry says. [ENDQ] [NEWLINE] (This guy is a fucking doctor?) [NEWLINE] [NEWLINE] Ok<mask> after the first link I'm still quite convinced that few drinks cause no harm. [NEWLINE] [NEWLINE] The second link is alarmism at best. It doesn't give ANY odds percentages or anything else along those lines, it merely lists everything that might have ever happened related to it, which is a huge number of things of course given the incredible number of people who've given birth. [NEWLINE] [NEWLINE] <mask> yea, not very compelling links at all. Certainly the only actual study referred (the one in the UK) backed up my point, not yours.<mask> you can't find a link that doesn't actually factually back my stance, you have a problem :P</s>
Label encoding: <s>Reading through those... [NEWLINE] [NEWLINE] [STARTQ] “The problem with drinking alcohol during your pregnancy is that there is no amount that has been proven to be safe,” [ENDQ] [NEWLINE] Basically asking people to prove a negative (ie: "It cannot harm") so that's useless to know. Another comment is along those lines as well. It can't be proven that it can't harm, hence it might harm. Be careful. On the next page... [NEWLINE] [NEWLINE] [STARTQ] In that study, researchers in the U.K. reported that the 5-year-old children of women who drank up to one to two alcoholic drinks per week or per occasion while pregnant were not at an increased risk of behavioral or cognitive problems [ENDQ] [NEWLINE] This seems pretty positive, and it is looking at the actual end results which is far better than the theorizing being done by the others. [NEWLINE] [NEWLINE] [STARTQ] “The way I see it is: If you wouldn’t give a 2-month-old a glass of wine, then why would you drink a glass of wine when you’re pregnant?” Garry says. [ENDQ] [NEWLINE] (This guy is a fucking doctor?) [NEWLINE] [NEWLINE] Ok so after the first link I'm still quite convinced that few drinks cause no harm. [NEWLINE] [NEWLINE] The second link is alarmism at best. It doesn't give ANY odds percentages or anything else along those lines, it merely lists everything that might have ever happened related to it, which is a huge number of things of course given the incredible number of people who've given birth. [NEWLINE] [NEWLINE] So yea, not very compelling links at all. Certainly the only actual study referred (the one in the UK) backed up my point, not yours. If you can't find a link that doesn't actually factually back my stance, you have a problem :P</s>
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] For example, the ACA had part of it ruled unconstitutional, and it was signed by a man who [taught constitutional law]( [URL] #University_of_Chicago_Law_School_and_civil_rights_attorney) at the University of Chicago for 12 years.  Are you really accusing Barack Obama of being unfamiliar with the Constitution?  Or did he just have a different legal opinion than the Supreme Court? [ENDQ] [NEWLINE] Sometimes presidents even sign laws they believe are unconstitutional. At the very least, both Bush and Obama have done<mask>. [NEWLINE] [NEWLINE] Just recently there was a post on the Volokh Conspiracy about [whether the president should sign a law he believes to be unconstitutional]( [URL] /). [NEWLINE] [NEWLINE] I found Prof. Baude's position interesting: [NEWLINE] [NEWLINE] [STARTQ] The Constitution does not contain a categorical duty to veto unconstitutional laws; and<mask> unconstitutional provisions are legally void, no constitutional violation happens<mask> they are enacted. That means that<mask> Obama and Bush did is not inherently unconstitutional. [ENDQ] [NEWLINE] [STARTQ] <mask>, I<mask> argued, signing an unconstitutional bill creates a risk – a risk that the unconstitutional provision will mistakenly be enforced in the future, whether by a court, a new president with different constitutional views, or some other way. The president’s oath to the Constitution does not allow him to take those constitutional risks lightly.<mask> a matter of structure and tradition, I argued, there must be some countervailing constitutional duty that justifies signing the bill — generally<mask> the bill contains other provisions that help to enforce constitutional rights or protect the country against invasion... [ENDQ] [NEWLINE] &gt;Without any constitutional benefits from the bill, I am not convinced there was any justification for signing it.<mask> the president<mask> believes that enforcing the bill could infringe his constitutional authority to receive ambassadors, he should have vetoed it</s>
Label encoding: <s> [STARTQ] For example, the ACA had part of it ruled unconstitutional, and it was signed by a man who [taught constitutional law]( [URL] #University_of_Chicago_Law_School_and_civil_rights_attorney) at the University of Chicago for 12 years.  Are you really accusing Barack Obama of being unfamiliar with the Constitution?  Or did he just have a different legal opinion than the Supreme Court? [ENDQ] [NEWLINE] Sometimes presidents even sign laws they believe are unconstitutional. At the very least, both Bush and Obama have done so. [NEWLINE] [NEWLINE] Just recently there was a post on the Volokh Conspiracy about [whether the president should sign a law he believes to be unconstitutional]( [URL] /). [NEWLINE] [NEWLINE] I found Prof. Baude's position interesting: [NEWLINE] [NEWLINE] [STARTQ] The Constitution does not contain a categorical duty to veto unconstitutional laws; and because unconstitutional provisions are legally void, no constitutional violation happens when they are enacted. That means that what Obama and Bush did is not inherently unconstitutional. [ENDQ] [NEWLINE] [STARTQ] But, I also argued, signing an unconstitutional bill creates a risk – a risk that the unconstitutional provision will mistakenly be enforced in the future, whether by a court, a new president with different constitutional views, or some other way. The president’s oath to the Constitution does not allow him to take those constitutional risks lightly. As a matter of structure and tradition, I argued, there must be some countervailing constitutional duty that justifies signing the bill — generally because the bill contains other provisions that help to enforce constitutional rights or protect the country against invasion... [ENDQ] [NEWLINE] &gt;Without any constitutional benefits from the bill, I am not convinced there was any justification for signing it. If the president indeed believes that enforcing the bill could infringe his constitutional authority to receive ambassadors, he should have vetoed it</s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe you misunderstand the purpose of college in the first place.  Many people are arguing against your conception of dance<mask> a non-academic subject.  I don't know enough about it to know one way or the other there,<mask> I'll take a different approach, which is that it doesn't matter whether it's academic or not. [NEWLINE] [NEWLINE] Let's take a comparison here:  Economics vs. Business.  All majors available at a great many universities.  A degree in Economics prepares you for a bunch of stuff, academic or applied.  You could teach, research, write, or otherwise contribute to the academic discipline, or you could become an economic adviser, or work in a firm or in business, doing applied work.  Business majors study many of the same concepts,<mask> in the context of preparing them to work in the business world.  Sure, there might be some academic options open to a business major,<mask> for the most part, it's basically like a trade school, preparing them for a specific career.  Many other degree paths are akin to trade school:  My undergrad school had a sports major degree, trained social workers, sculptors, musicians, and many others who went to college to prepare themselves for a specific career. [NEWLINE] [NEWLINE] <mask> should dance be different?  They're studying technique under the competent tutelage of experienced and knowledgeable dancers.  You could get that outside the academy, perhaps,<mask> you could say the same for the business major or artist. [NEWLINE] [NEWLINE] <mask>, I'm not trying to say dance is an academic discipline, I'm saying it doesn't matter whether it is or not.  College offers plenty of applied education, with many degree paths that amount to very expensive trade school. <mask>,<mask> should dance be any different from these?</s>
Label encoding: <s>I believe you misunderstand the purpose of college in the first place.  Many people are arguing against your conception of dance as a non-academic subject.  I don't know enough about it to know one way or the other there, so I'll take a different approach, which is that it doesn't matter whether it's academic or not. [NEWLINE] [NEWLINE] Let's take a comparison here:  Economics vs. Business.  All majors available at a great many universities.  A degree in Economics prepares you for a bunch of stuff, academic or applied.  You could teach, research, write, or otherwise contribute to the academic discipline, or you could become an economic adviser, or work in a firm or in business, doing applied work.  Business majors study many of the same concepts, but in the context of preparing them to work in the business world.  Sure, there might be some academic options open to a business major, but for the most part, it's basically like a trade school, preparing them for a specific career.  Many other degree paths are akin to trade school:  My undergrad school had a sports major degree, trained social workers, sculptors, musicians, and many others who went to college to prepare themselves for a specific career. [NEWLINE] [NEWLINE] Why should dance be different?  They're studying technique under the competent tutelage of experienced and knowledgeable dancers.  You could get that outside the academy, perhaps, but you could say the same for the business major or artist. [NEWLINE] [NEWLINE] So, I'm not trying to say dance is an academic discipline, I'm saying it doesn't matter whether it is or not.  College offers plenty of applied education, with many degree paths that amount to very expensive trade school.  So, why should dance be any different from these?</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You could say it's a good thing<mask> you never know who will win in baseball,<mask><mask><mask> it really just comes down more to luck and a few flakey games, whereas in basketball you really have to earn championships. [ENDQ] [NEWLINE] That's not a valid argument. There are no 'better' teams - just teams on the field/ice/court, and the teams on paper. [NEWLINE] [NEWLINE] You're saying that in Basketball, I could open my chequebook, and throw money at the best free agents, and expect to win a championship. [NEWLINE] [NEWLINE] That's not fair, nor is that right. [NEWLINE] [NEWLINE] There's something wrong with your sport,<mask> all it takes to win, and be competitive is having the biggest pocket book.<mask>'s the parity throughout the league?<mask> are the underdogs? [NEWLINE] [NEWLINE] Lets use hockey for an example; [NEWLINE] [NEWLINE] Last year the Pittsburgh Penguins opened their chequebook, and threw money at every single available good player that they could get their hands on. [NEWLINE] [NEWLINE] They almost lost in the first round of the playoffs at the hands of the Islanders - they were then completely embarassed again in the conference finals<mask> the Bruins systematically demolished their entire team, with nowhere near the same amount of'star' players on their roster. [NEWLINE] [NEWLINE] The Penguins tried to buy a championship,<mask> got their asses handed to them by a team of hardworking lesser-known players who were hand-built from the ground up. [NEWLINE] [NEWLINE] The Bruins then got beaten by a team that was even more carefully built. [NEWLINE] [NEWLINE] That's the true realities of sport - the team that wins, should be the team that works hardest, and plays their asses off night in and night out. Not the team with the biggest names on the scoresheet. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] You could say it's a good thing because you never know who will win in baseball, but I think it really just comes down more to luck and a few flakey games, whereas in basketball you really have to earn championships. [ENDQ] [NEWLINE] That's not a valid argument. There are no 'better' teams - just teams on the field/ice/court, and the teams on paper. [NEWLINE] [NEWLINE] You're saying that in Basketball, I could open my chequebook, and throw money at the best free agents, and expect to win a championship. [NEWLINE] [NEWLINE] That's not fair, nor is that right. [NEWLINE] [NEWLINE] There's something wrong with your sport, if all it takes to win, and be competitive is having the biggest pocket book. Where's the parity throughout the league? Where are the underdogs? [NEWLINE] [NEWLINE] Lets use hockey for an example; [NEWLINE] [NEWLINE] Last year the Pittsburgh Penguins opened their chequebook, and threw money at every single available good player that they could get their hands on. [NEWLINE] [NEWLINE] They almost lost in the first round of the playoffs at the hands of the Islanders - they were then completely embarassed again in the conference finals when the Bruins systematically demolished their entire team, with nowhere near the same amount of'star' players on their roster. [NEWLINE] [NEWLINE] The Penguins tried to buy a championship, but got their asses handed to them by a team of hardworking lesser-known players who were hand-built from the ground up. [NEWLINE] [NEWLINE] The Bruins then got beaten by a team that was even more carefully built. [NEWLINE] [NEWLINE] That's the true realities of sport - the team that wins, should be the team that works hardest, and plays their asses off night in and night out. Not the team with the biggest names on the scoresheet. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why are you trying to pretend that the Palestinian Arabs' opinions didn't matter?  Do you think their opinions didn't matter? [NEWLINE] [NEWLINE] The UN treated it like settling some argument between two kids arguing over the space in the back seat of the car, and that's not<mask> it was.  The UN didn't have the power to enforce their plan. [NEWLINE] [NEWLINE] The Palestinian Arabs didn't start a war with the nascent Israel.  The Palestinian Arabs didn't just decide to leave and then demand their property back.  They left<mask> there was going to be a war.  A war they didn't have the wherewithal to instigate OR stop. [NEWLINE] [NEWLINE] <mask> during that war Israel laid claim to the homes &amp; property those people had run away from (stupidly wanting to avoid getting killed) and gave that property to Jews. [NEWLINE] [NEWLINE] <mask> the UN was actually interested in settling the matter<mask> you are portraying it here, they would have forced Israel to give that property back to the people who had run away *to save their lives*. [NEWLINE] [NEWLINE] The manner in which all the property was "acquired" during the Six Day War is the very definition of squatting.  "Oh these properties were abandoned.  No one was here<mask> I got here." [NEWLINE] [NEWLINE] It is exactly like<mask> you went to work one day and some people just moved into your house and the local government said, "well, hey, it's their house now.  Tough shit." [NEWLINE] [NEWLINE] Well, no.  It's worse.  It's like someone threatened to firebomb randomly in your neighborhood, you took your family to safety at a shelter, and<mask> you came back, the city had set up some city employees in your house and refused to acknowledge your claim on that property.</s>
Label encoding: <s>Why are you trying to pretend that the Palestinian Arabs' opinions didn't matter?  Do you think their opinions didn't matter? [NEWLINE] [NEWLINE] The UN treated it like settling some argument between two kids arguing over the space in the back seat of the car, and that's not what it was.  The UN didn't have the power to enforce their plan. [NEWLINE] [NEWLINE] The Palestinian Arabs didn't start a war with the nascent Israel.  The Palestinian Arabs didn't just decide to leave and then demand their property back.  They left because there was going to be a war.  A war they didn't have the wherewithal to instigate OR stop. [NEWLINE] [NEWLINE] So during that war Israel laid claim to the homes &amp; property those people had run away from (stupidly wanting to avoid getting killed) and gave that property to Jews. [NEWLINE] [NEWLINE] If the UN was actually interested in settling the matter as you are portraying it here, they would have forced Israel to give that property back to the people who had run away *to save their lives*. [NEWLINE] [NEWLINE] The manner in which all the property was "acquired" during the Six Day War is the very definition of squatting.  "Oh these properties were abandoned.  No one was here when I got here." [NEWLINE] [NEWLINE] It is exactly like if you went to work one day and some people just moved into your house and the local government said, "well, hey, it's their house now.  Tough shit." [NEWLINE] [NEWLINE] Well, no.  It's worse.  It's like someone threatened to firebomb randomly in your neighborhood, you took your family to safety at a shelter, and when you came back, the city had set up some city employees in your house and refused to acknowledge your claim on that property.</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The dollar is used<mask> a global reserve currency<mask> it's stable and reliable.  That includes the assumption that the government isn't just going to take it; that<mask> you hold a dollar you hold legitimate value and purchasing power that isn't going to be arbitrarily taken away.  The taxes we have are predictable, low and a matter of public record. [NEWLINE] [NEWLINE] <mask> you take billions (possibly trillions) in cash and liquidated assets (which would be difficult to do given the currency fluctuation) from those who lawfully hold them, you've undermined the idea that the dollar is stable and reliable. <mask> it can be taken at whim "for the good of the people", it isn't either of those things. [NEWLINE] [NEWLINE] Combine that with the likelihood that a great deal of foreign investment would either be confiscated, massively devalued or would just flat disappear in the melee, and you've got a serious reason to lose confidence in the dollar. [NEWLINE] [NEWLINE] Right now most internationally purchased oil is purchased with dollars; most countries hold dollars assuming they will have consistent value.  That significantly increases the value of all dollars by creating demand for them. <mask> would anyone place that kind of trust in something that can just be confiscated at will?  And<mask> they can't trust the value of the dollar,<mask> would they keep it? <mask> wouldn't they divest themselves<mask> quickly<mask> possible to avoid significant losses? [NEWLINE] [NEWLINE] <mask> they did that, the market would be flooded with trillions of dollars that nobody<mask> Americans would want; which would result in a loss of value in the currency, which would cause price inflation at the same time that the economy is crumbling due to a complete divestment and inefficient decentralization of capital. [NEWLINE] [NEWLINE] <mask> everybody would have plenty of dollars, they just wouldn't be worth much.</s>
Label encoding: <s>The dollar is used as a global reserve currency because it's stable and reliable.  That includes the assumption that the government isn't just going to take it; that when you hold a dollar you hold legitimate value and purchasing power that isn't going to be arbitrarily taken away.  The taxes we have are predictable, low and a matter of public record. [NEWLINE] [NEWLINE] If you take billions (possibly trillions) in cash and liquidated assets (which would be difficult to do given the currency fluctuation) from those who lawfully hold them, you've undermined the idea that the dollar is stable and reliable.  If it can be taken at whim "for the good of the people", it isn't either of those things. [NEWLINE] [NEWLINE] Combine that with the likelihood that a great deal of foreign investment would either be confiscated, massively devalued or would just flat disappear in the melee, and you've got a serious reason to lose confidence in the dollar. [NEWLINE] [NEWLINE] Right now most internationally purchased oil is purchased with dollars; most countries hold dollars assuming they will have consistent value.  That significantly increases the value of all dollars by creating demand for them.  Why would anyone place that kind of trust in something that can just be confiscated at will?  And if they can't trust the value of the dollar, why would they keep it?  Why wouldn't they divest themselves as quickly as possible to avoid significant losses? [NEWLINE] [NEWLINE] When they did that, the market would be flooded with trillions of dollars that nobody but Americans would want; which would result in a loss of value in the currency, which would cause price inflation at the same time that the economy is crumbling due to a complete divestment and inefficient decentralization of capital. [NEWLINE] [NEWLINE] So everybody would have plenty of dollars, they just wouldn't be worth much.</s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're looking at the complaints (<mask> not all of them) from the wrong approach. [NEWLINE] [NEWLINE] The fact that people feel backed into a one or the other, lesser of two evils choice every four years is enough reason for disdain<mask> is,<mask> it is only compounded by a trend that seems to grow every election; executive or legislative. [NEWLINE] [NEWLINE] People are voting for a President or representative and expect them to be their voice in steering the country in the direction the people want.  These people get elected,<mask>, and do<mask> **they** see<mask> right.  People are complaining about that, not<mask> much the man himself. [NEWLINE] [NEWLINE] I never understood<mask> politicians told me<mask> they felt about certain issues.  I don't give a shit<mask> they feel, I want someone who will vote, or in the President's case veto/pass,<mask> the majority wants. [NEWLINE] [NEWLINE] <mask> they rarely do and people complain.  Is the complaining pointless? No. It gets attention. With enough attention change may come about, whether it leads to an internal reform of<mask> things are done, people voting a certain/different way or an all out revolution. [NEWLINE] [NEWLINE] Now to address specific complaints toward Obama and the desire to remove him, this goes back to the idea that<mask><mask> he was elected, he's not living up to the expectations of those who elected him. <mask> I had to buy a car and chose the 'best of<mask> was offered' and ended up with a car that didn't run at all after a week, I'd want a new car.  Just<mask> the salesman said it was great doesn't mean it was or that I should be obligated to keep it. [NEWLINE] [NEWLINE] tl;dr Complaining isn't pointless, it raises awareness and may lead to change.</s>
Label encoding: <s>I think you're looking at the complaints ( though not all of them) from the wrong approach. [NEWLINE] [NEWLINE] The fact that people feel backed into a one or the other, lesser of two evils choice every four years is enough reason for disdain as is, but it is only compounded by a trend that seems to grow every election; executive or legislative. [NEWLINE] [NEWLINE] People are voting for a President or representative and expect them to be their voice in steering the country in the direction the people want.  These people get elected, however, and do what **they** see as right.  People are complaining about that, not so much the man himself. [NEWLINE] [NEWLINE] I never understood why politicians told me how they felt about certain issues.  I don't give a shit how they feel, I want someone who will vote, or in the President's case veto/pass, what the majority wants. [NEWLINE] [NEWLINE] But they rarely do and people complain.  Is the complaining pointless? No. It gets attention. With enough attention change may come about, whether it leads to an internal reform of how things are done, people voting a certain/different way or an all out revolution. [NEWLINE] [NEWLINE] Now to address specific complaints toward Obama and the desire to remove him, this goes back to the idea that even though he was elected, he's not living up to the expectations of those who elected him.  If I had to buy a car and chose the 'best of what was offered' and ended up with a car that didn't run at all after a week, I'd want a new car.  Just because the salesman said it was great doesn't mean it was or that I should be obligated to keep it. [NEWLINE] [NEWLINE] tl;dr Complaining isn't pointless, it raises awareness and may lead to change.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There has been almost no better time and place to raise children in the *history* of humanity than today. There are enormous institutional problems today, of course,<mask> think about the following. [NEWLINE] [NEWLINE] Prejudice based on race, gender, and sexual orientation is the lowest it has been in centuries.<mask> institutional problems leading to income inequality we have a (somewhat effective) social safety net in place and regulations that prevent the horrible abuses towards workers that existed in the in the 19th century. Due to the development of nuclear weapons, total war is no longer a threat to the world (for the foreseeable future). Education is cheaper and more accessible than ever before thanks to the internet and I would expect that this would extend into colleges by the time your children are 18. Georgia Tech is currently trying a pilot program to offer an online Masters in Computer Science for just a few thousand dollars. We have better access to live saving medicine than ever before in history. We have a greater understanding of mental illness to the likelihood that your children will be saddled with crippling depression or anxiety is limited. [NEWLINE] [NEWLINE] I could go on and on. [NEWLINE] [NEWLINE] My real point is that<mask> you truly believe that it is irresponsible to have a child today then (barring two arguments I'll get to in a moment) it must have been irresponsible to have a child for *most of human history*. You might agree that that is the case,<mask> at least you must acknowledge that it is a pretty extreme view. [NEWLINE] [NEWLINE] The only arguments that work for modern times<mask> not past decades or centuries are overpopulation and climate change.<mask><mask> you can make a valid argument that we should be having fewer children<mask> of these problems (<mask> certainly not zero children)<mask> that isn't the argument you seem to be making. </s>
Label encoding: <s>There has been almost no better time and place to raise children in the *history* of humanity than today. There are enormous institutional problems today, of course, but think about the following. [NEWLINE] [NEWLINE] Prejudice based on race, gender, and sexual orientation is the lowest it has been in centuries. Despite institutional problems leading to income inequality we have a (somewhat effective) social safety net in place and regulations that prevent the horrible abuses towards workers that existed in the in the 19th century. Due to the development of nuclear weapons, total war is no longer a threat to the world (for the foreseeable future). Education is cheaper and more accessible than ever before thanks to the internet and I would expect that this would extend into colleges by the time your children are 18. Georgia Tech is currently trying a pilot program to offer an online Masters in Computer Science for just a few thousand dollars. We have better access to live saving medicine than ever before in history. We have a greater understanding of mental illness to the likelihood that your children will be saddled with crippling depression or anxiety is limited. [NEWLINE] [NEWLINE] I could go on and on. [NEWLINE] [NEWLINE] My real point is that if you truly believe that it is irresponsible to have a child today then (barring two arguments I'll get to in a moment) it must have been irresponsible to have a child for *most of human history*. You might agree that that is the case, but at least you must acknowledge that it is a pretty extreme view. [NEWLINE] [NEWLINE] The only arguments that work for modern times but not past decades or centuries are overpopulation and climate change. I think you can make a valid argument that we should be having fewer children because of these problems ( but certainly not zero children) but that isn't the argument you seem to be making. </s>
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Medical procedures, especially ones that are not necessarily for survival, should require a parent's permission.  After all, we have to ask for an adult's consent to give it to them<mask> well. [NEWLINE] [NEWLINE] <mask><mask>, a parent can even choose for their child not to have a life-improving or even life-saving medical procedure. <mask> you take away the ability of parents to be their child's health care proxy for something small like vaccines, the government will certainly move to make other decisions for parents<mask> well. [NEWLINE] [NEWLINE] Maybe a parent doesn't want to put their child through any more chemotherapy?  Or maybe the government decides to mandate that all boys be circumcised to reduce the spread of HIV?  After all, the risk of problems due to circumcision is minimal and it provides a public benefit by reducing HIV transmission,<mask><mask> not override the parents wishes?  You see<mask> I mean? [NEWLINE] [NEWLINE] Not to mention, there are many vaccines, many have different choices in terms of vaccine type, delivery mechanisms, and timing, which often has to be weighed against different factors of the child's health. <mask> there are vaccines that are not overly effective (flu) or are for obscure diseases (rabies). [NEWLINE] [NEWLINE] Even doctors eventually decide to stop vaccinating for some things (e.g., Smallpox)<mask> they believe the risk of the vaccine is greater than the risk of catching the disease. [NEWLINE] [NEWLINE] All this is just to say that vaccination is a complex set of decisions which parents should discuss with their doctors and then make the right call for themselves.  The government can step in to prevent exposure to others by requiring vaccinations for schools and public parks,<mask> it shouldn't be knocking down doors and taking children away to perform medical procedures on them that their parents think are wrong.</s>
Label encoding: <s>Medical procedures, especially ones that are not necessarily for survival, should require a parent's permission.  After all, we have to ask for an adult's consent to give it to them as well. [NEWLINE] [NEWLINE] In fact, a parent can even choose for their child not to have a life-improving or even life-saving medical procedure.  If you take away the ability of parents to be their child's health care proxy for something small like vaccines, the government will certainly move to make other decisions for parents as well. [NEWLINE] [NEWLINE] Maybe a parent doesn't want to put their child through any more chemotherapy?  Or maybe the government decides to mandate that all boys be circumcised to reduce the spread of HIV?  After all, the risk of problems due to circumcision is minimal and it provides a public benefit by reducing HIV transmission, so why not override the parents wishes?  You see what I mean? [NEWLINE] [NEWLINE] Not to mention, there are many vaccines, many have different choices in terms of vaccine type, delivery mechanisms, and timing, which often has to be weighed against different factors of the child's health.  Additionally there are vaccines that are not overly effective (flu) or are for obscure diseases (rabies). [NEWLINE] [NEWLINE] Even doctors eventually decide to stop vaccinating for some things (e.g., Smallpox) when they believe the risk of the vaccine is greater than the risk of catching the disease. [NEWLINE] [NEWLINE] All this is just to say that vaccination is a complex set of decisions which parents should discuss with their doctors and then make the right call for themselves.  The government can step in to prevent exposure to others by requiring vaccinations for schools and public parks, but it shouldn't be knocking down doors and taking children away to perform medical procedures on them that their parents think are wrong.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "<mask> the public is not aware of the details surrounding his plea" [ENDQ] [NEWLINE] <mask> you want to be you can get a copy of his plea agreement<mask> all agreements for felonies must be submitted in writing to the Circuit Court.  You can contact the Charlottesville Circuit Court Clerk [here]( [URL] ). Now he didn't get away with serving 3 months of an 8 year sentence he was literally sentenced to 8 years with all<mask> 3 months suspended.  In other words the judge sentenced him to a three month active sentence and can revoke the suspended period should he commit another violation during his period of good behavior. [NEWLINE] [NEWLINE] [STARTQ] he is not considered a threat to the public [ENDQ] [NEWLINE] He is a violent sex offender.  He has to let the police know<mask> he lives and works at all times<mask> they can keep tabs on him.  He is visited at a minimum of once a year by a State Police officer who verifies this information through a surprise visit. <mask><mask> he must provide the police any and all usernames on social media, email addresses, vehicles he regularly uses, and his dna.  [See Va Code 9.1-903]( [URL].1/chapter9/section9.1-903/) [NEWLINE] [NEWLINE] He wasn't convicted of rape he had the charge reduced to aggravated sexual battery under Va Code 18.2-67.3 from rape.  You can find it by looking him up [here in the Charlottesville Circuit Court]( [URL].do). <mask> the charge was reduced from rape the statement of facts which he admitted was accurate<mask> part of his plea agreement states that he anally and vaginally raped a girl.  You can get that from the clerk's office<mask> it was made part of the record. [NEWLINE] [NEWLINE] Edit: spelling</s>
Label encoding: <s> [STARTQ] " as the public is not aware of the details surrounding his plea" [ENDQ] [NEWLINE] If you want to be you can get a copy of his plea agreement as all agreements for felonies must be submitted in writing to the Circuit Court.  You can contact the Charlottesville Circuit Court Clerk [here]( [URL] ). Now he didn't get away with serving 3 months of an 8 year sentence he was literally sentenced to 8 years with all but 3 months suspended.  In other words the judge sentenced him to a three month active sentence and can revoke the suspended period should he commit another violation during his period of good behavior. [NEWLINE] [NEWLINE] [STARTQ] he is not considered a threat to the public [ENDQ] [NEWLINE] He is a violent sex offender.  He has to let the police know where he lives and works at all times so they can keep tabs on him.  He is visited at a minimum of once a year by a State Police officer who verifies this information through a surprise visit.  In addition he must provide the police any and all usernames on social media, email addresses, vehicles he regularly uses, and his dna.  [See Va Code 9.1-903]( [URL].1/chapter9/section9.1-903/) [NEWLINE] [NEWLINE] He wasn't convicted of rape he had the charge reduced to aggravated sexual battery under Va Code 18.2-67.3 from rape.  You can find it by looking him up [here in the Charlottesville Circuit Court]( [URL].do).  While the charge was reduced from rape the statement of facts which he admitted was accurate as part of his plea agreement states that he anally and vaginally raped a girl.  You can get that from the clerk's office as it was made part of the record. [NEWLINE] [NEWLINE] Edit: spelling</s>
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0900, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>History does not repeat itself,<mask> it does have a rhyme to it. Look back in history on the technology that has affected workers: [NEWLINE] [NEWLINE] All societies start out with most of their population farming. Eventually machines are developed to make farming more efficient and require less labor. This has 2 effects: [NEWLINE] [NEWLINE] 1. Many people become unemployed. This is actually a good thing over time. Those people are then able to better themselves in order to produce newer, better things.<mask> your time is not consumed farming you have time to develop new ideas, skills, etc. [NEWLINE] [NEWLINE] 2. The cost of goods goes down (both to the producer and the consumer). Someone suggested that business owners would simply pocket the cash,<mask> that is ludicrous. One business would cut prices to attract customers<mask> everyone would have to follow suit to remain competitive. [NEWLINE] [NEWLINE] Textiles are usually the next process to become industrialized. This again left many people unemployed and eventually they were absorbed back into the workforce. [NEWLINE] [NEWLINE] The cycle of industrialisation is ongoing to this day. [NEWLINE] [NEWLINE] <mask> industrialisation occurs it *grows* the pie from which we all take our varying slices. The surplus of workers left in its wake are given employment in the new industries that see able to be birthed<mask><mask><mask> of the previous round of industrialisation. [NEWLINE] [NEWLINE] Automatons don't consume goods<mask>. Consumers are a necessary part of an economy. Producers know it better than anyone. They cannot exist in absence of each other.<mask> until the question is "are automatons sentiment life?" we will not be economically phased out. [NEWLINE] [NEWLINE] **TL;DR:** it will not cause problems<mask> our economy will grow and absorb those unemployed in new ways just like it always has. Simple economics.</s>
Label encoding: <s>History does not repeat itself, but it does have a rhyme to it. Look back in history on the technology that has affected workers: [NEWLINE] [NEWLINE] All societies start out with most of their population farming. Eventually machines are developed to make farming more efficient and require less labor. This has 2 effects: [NEWLINE] [NEWLINE] 1. Many people become unemployed. This is actually a good thing over time. Those people are then able to better themselves in order to produce newer, better things. If your time is not consumed farming you have time to develop new ideas, skills, etc. [NEWLINE] [NEWLINE] 2. The cost of goods goes down (both to the producer and the consumer). Someone suggested that business owners would simply pocket the cash, but that is ludicrous. One business would cut prices to attract customers so everyone would have to follow suit to remain competitive. [NEWLINE] [NEWLINE] Textiles are usually the next process to become industrialized. This again left many people unemployed and eventually they were absorbed back into the workforce. [NEWLINE] [NEWLINE] The cycle of industrialisation is ongoing to this day. [NEWLINE] [NEWLINE] When industrialisation occurs it *grows* the pie from which we all take our varying slices. The surplus of workers left in its wake are given employment in the new industries that see able to be birthed as a result of the previous round of industrialisation. [NEWLINE] [NEWLINE] Automatons don't consume goods yet. Consumers are a necessary part of an economy. Producers know it better than anyone. They cannot exist in absence of each other. So until the question is "are automatons sentiment life?" we will not be economically phased out. [NEWLINE] [NEWLINE] **TL;DR:** it will not cause problems because our economy will grow and absorb those unemployed in new ways just like it always has. Simple economics.</s>
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] One person saying, "no, I won't buy that" isn't enough to affect the demand.<mask><mask><mask>, the same number of animals will be killed and that person's vegetarianism or veganism will make no difference. [ENDQ] [NEWLINE] <mask> a person decides to refrain from a particular activity<mask> a matter of personal ethics, it's about not putting something in your body that you feel is wrong.  Affecting demand is tangential. [NEWLINE] [NEWLINE] Further,<mask> 1-3% of the population is vegetarian/vegan, it does have an effect on demand, even<mask> that effect is negligible. [NEWLINE] [NEWLINE] [STARTQ] humans are essentially nothing more than the highest cognitively functioning animals.  We're not going to feed a lion "meat" made of soy protein and nobody has a problem with them mauling an antelope. [ENDQ] [NEWLINE] Lions are carnivores, humans are omnivores that can survive and thrive on a variety of diverse diets, including meat heavy or meat free.  Vegetarians do OK and vegans are fine with a little supplementation. [NEWLINE] [NEWLINE] Non-human primates were at one point the highest functioning cognitive animals, and many of them subsisted on mostly plant matter.  Do your teeth look like[ this]( [URL].jpg) or [this]( [URL]?image=9)?  The latter is a bonobo, who are largely fruit eaters who eat small amounts of animal matter. [NEWLINE] [NEWLINE] [STARTQ] I don't buy that dying equates to suffering.<mask> someone walks up to me and puts a bullet in my brainstem, I haven't felt pain nor have I suffered. I'm just dead. [ENDQ] [NEWLINE] Are you saying dying &gt; suffering?  You'd rather lose your life than spend some time being uncomfortable?</s><pad>
Label encoding: <s> [STARTQ] One person saying, "no, I won't buy that" isn't enough to affect the demand. As a result, the same number of animals will be killed and that person's vegetarianism or veganism will make no difference. [ENDQ] [NEWLINE] If a person decides to refrain from a particular activity as a matter of personal ethics, it's about not putting something in your body that you feel is wrong.  Affecting demand is tangential. [NEWLINE] [NEWLINE] Further, if 1-3% of the population is vegetarian/vegan, it does have an effect on demand, even if that effect is negligible. [NEWLINE] [NEWLINE] [STARTQ] humans are essentially nothing more than the highest cognitively functioning animals.  We're not going to feed a lion "meat" made of soy protein and nobody has a problem with them mauling an antelope. [ENDQ] [NEWLINE] Lions are carnivores, humans are omnivores that can survive and thrive on a variety of diverse diets, including meat heavy or meat free.  Vegetarians do OK and vegans are fine with a little supplementation. [NEWLINE] [NEWLINE] Non-human primates were at one point the highest functioning cognitive animals, and many of them subsisted on mostly plant matter.  Do your teeth look like[ this]( [URL].jpg) or [this]( [URL]?image=9)?  The latter is a bonobo, who are largely fruit eaters who eat small amounts of animal matter. [NEWLINE] [NEWLINE] [STARTQ] I don't buy that dying equates to suffering. If someone walks up to me and puts a bullet in my brainstem, I haven't felt pain nor have I suffered. I'm just dead. [ENDQ] [NEWLINE] Are you saying dying &gt; suffering?  You'd rather lose your life than spend some time being uncomfortable?</s><pad>
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, here, change my view: [NEWLINE] [NEWLINE] Climate change is happening. [NEWLINE] [NEWLINE] Humans are most likely the most significant contributing factor. [NEWLINE] [NEWLINE] CMV: We can't stop it from getting worse. [NEWLINE] [NEWLINE] The projections say that at a modest increase we will have severe repercussions. The measured CO2 levels are far in excess of modest. It takes time for the environment to react,<mask> we are already on the "slope" to a radically different environmental situation even<mask> we somehow stopped CO2 production above natural levels tomorrow. [NEWLINE] [NEWLINE] It's too difficult to maintain our way of life AND have reduced CO2 levels in the near term (especially with abundant fossil fuel and animal farming)<mask><mask> not just adapt? [NEWLINE] [NEWLINE] Sure, there will be more hurricanes, flooding, desertification...<mask> these are problems we can tackle. [NEWLINE] [NEWLINE] The way to make a difference is not to say "stop using energy" -- that's backwards and you can't convince everyone of it.<mask> you can do is make the preparations for<mask> is to come. Better water, better energy sources (including more nuclear research that lessens risk and waste) and most of all: Assistance to those in the third world who,<mask> they start to develop, require cheap sources of plentiful energy<mask> that they can provide for their own ways of dealing with the changing environment. [NEWLINE] [NEWLINE] China and the USA will never shrink their emissions to agreeable levels without viable alternatives. At present, no alternative is either viable, cost effective, or politically welcome. That's unlikely to change<mask> the face of green is someone wearing hemp (<mask> that is who is protesting the most visibly) versus the businessman who sees a profit opportunity in the latest coal seam discovery. [NEWLINE] [NEWLINE] <mask> yeah. Change MY view.</s><pad><pad>
Label encoding: <s>Ok, here, change my view: [NEWLINE] [NEWLINE] Climate change is happening. [NEWLINE] [NEWLINE] Humans are most likely the most significant contributing factor. [NEWLINE] [NEWLINE] CMV: We can't stop it from getting worse. [NEWLINE] [NEWLINE] The projections say that at a modest increase we will have severe repercussions. The measured CO2 levels are far in excess of modest. It takes time for the environment to react, so we are already on the "slope" to a radically different environmental situation even if we somehow stopped CO2 production above natural levels tomorrow. [NEWLINE] [NEWLINE] It's too difficult to maintain our way of life AND have reduced CO2 levels in the near term (especially with abundant fossil fuel and animal farming) so why not just adapt? [NEWLINE] [NEWLINE] Sure, there will be more hurricanes, flooding, desertification... But these are problems we can tackle. [NEWLINE] [NEWLINE] The way to make a difference is not to say "stop using energy" -- that's backwards and you can't convince everyone of it. What you can do is make the preparations for what is to come. Better water, better energy sources (including more nuclear research that lessens risk and waste) and most of all: Assistance to those in the third world who, when they start to develop, require cheap sources of plentiful energy so that they can provide for their own ways of dealing with the changing environment. [NEWLINE] [NEWLINE] China and the USA will never shrink their emissions to agreeable levels without viable alternatives. At present, no alternative is either viable, cost effective, or politically welcome. That's unlikely to change while the face of green is someone wearing hemp ( because that is who is protesting the most visibly) versus the businessman who sees a profit opportunity in the latest coal seam discovery. [NEWLINE] [NEWLINE] So yeah. Change MY view.</s><pad><pad>
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I definitely see your point. I was definitely guilty of a bunch of "fatlogic" type of thoughts<mask> I was really fat.<mask><mask><mask> for the vast majority of people these thoughts are just a small form of comfort for them in their current state, not actual ideals they live by. You said you believed a bunch of "fatlogic" type thoughts,<mask> you still ended up researching weight loss right? You didn't embrace these ideas wholeheartedly and decide you were completely happy the way you were.<mask><mask> part of you believed that stuff an even bigger part of you knew that you wanted to do something about your weight, which resulted in you looking into weight loss forums.<mask><mask> you took comfort in those thoughts you still wanted to lose weight at the end of the day. My whole point is just that I've never met anybody who was really overweight who was 100% satisfied with every lb they had and didn't want to lose weight at all,<mask><mask> people read those fatlogicky kind of things they take it at face value and really believe those people are happy being obese. [NEWLINE] [NEWLINE] And yea<mask><mask> the "equating attractiveness with worth" is a big part of it,<mask> I don't think that's all it boils down to. At the same time being overweight is tied to character flaws.<mask> you're fat you're mad at yourself that you can't control the food that goes into your body. You're insecure that your lazy and don't exercise enough. Being fat in itself makes you insecure about certain things and<mask> somebody points out that you're fat it feels like they're really saying "you're a piece of shit lazy glutton" even<mask> they really mean "dude you know you gained a couple pounds right"?</s>
Label encoding: <s>I definitely see your point. I was definitely guilty of a bunch of "fatlogic" type of thoughts when I was really fat. But I think for the vast majority of people these thoughts are just a small form of comfort for them in their current state, not actual ideals they live by. You said you believed a bunch of "fatlogic" type thoughts, but you still ended up researching weight loss right? You didn't embrace these ideas wholeheartedly and decide you were completely happy the way you were. Even though part of you believed that stuff an even bigger part of you knew that you wanted to do something about your weight, which resulted in you looking into weight loss forums. Even though you took comfort in those thoughts you still wanted to lose weight at the end of the day. My whole point is just that I've never met anybody who was really overweight who was 100% satisfied with every lb they had and didn't want to lose weight at all, but when people read those fatlogicky kind of things they take it at face value and really believe those people are happy being obese. [NEWLINE] [NEWLINE] And yea I think the "equating attractiveness with worth" is a big part of it, but I don't think that's all it boils down to. At the same time being overweight is tied to character flaws. When you're fat you're mad at yourself that you can't control the food that goes into your body. You're insecure that your lazy and don't exercise enough. Being fat in itself makes you insecure about certain things and when somebody points out that you're fat it feels like they're really saying "you're a piece of shit lazy glutton" even if they really mean "dude you know you gained a couple pounds right"?</s>
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm an 18 years old girl,<mask> I hope i still fit in the young girls category that should be<mask> sensitive about looks :P [NEWLINE] I completely agree with OP and I<mask> use the analogy with other skills like painting, sports or anything else. You should be more worried<mask> you don't have general knowledge than<mask> you don't look like a model,<mask> somehow society imposes those wrong values to the majority of people. [NEWLINE] I guess sensitivity about looks just depends on your intelligence, maturity and rationality, not really age. The annoying fact is that society prioritizes looks, status and other fake human values and try to protect younger people from everything, instead of teaching them to think in a more rational way. The biggest mistake in society is banning everything, instead of teaching you to realize its wrong. For example they make XXL barbies instead of teaching girls that it's a toy and not a model of<mask> you look, and that you should be more worried about your education than looking like Barbie. Do boys want to look like Batman? Do they have to make Batman look more like a regular man or is he just a superhero that doesn't exist in real life and people just have to deal with it. [NEWLINE] <mask> you say younger people don't think rationally like adults, the problem is<mask> don't you educate them to think in a more mature way, instead of just looking at them like they're stupid.<mask><mask> the more you know, the more you want to know and in order to have any opinions about anything around you, you should have some knowledge about it first. [NEWLINE] [NEWLINE] EDIT:<mask><mask> don't people complain about Einstein for making people look stupid like they complain about Adriana Lima making women look ugly. Double standards. [NEWLINE] </s>
Label encoding: <s>I'm an 18 years old girl, so I hope i still fit in the young girls category that should be so sensitive about looks :P [NEWLINE] I completely agree with OP and I also use the analogy with other skills like painting, sports or anything else. You should be more worried if you don't have general knowledge than if you don't look like a model, but somehow society imposes those wrong values to the majority of people. [NEWLINE] I guess sensitivity about looks just depends on your intelligence, maturity and rationality, not really age. The annoying fact is that society prioritizes looks, status and other fake human values and try to protect younger people from everything, instead of teaching them to think in a more rational way. The biggest mistake in society is banning everything, instead of teaching you to realize its wrong. For example they make XXL barbies instead of teaching girls that it's a toy and not a model of how you look, and that you should be more worried about your education than looking like Barbie. Do boys want to look like Batman? Do they have to make Batman look more like a regular man or is he just a superhero that doesn't exist in real life and people just have to deal with it. [NEWLINE] If you say younger people don't think rationally like adults, the problem is why don't you educate them to think in a more mature way, instead of just looking at them like they're stupid. I think the more you know, the more you want to know and in order to have any opinions about anything around you, you should have some knowledge about it first. [NEWLINE] [NEWLINE] EDIT: Also why don't people complain about Einstein for making people look stupid like they complain about Adriana Lima making women look ugly. Double standards. [NEWLINE] </s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0987, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi, [NEWLINE] [NEWLINE] Of course, undeniably,<mask> Cuba could export goods to the US, it would change things a lot.<mask> I don't think the root cause of poverty in Cuba is the embargo. The root cause is the system<mask> every decision is made by the government with an "every one must be equal" approach (and that is assuming, for the sake of the argument, that humans in Cuban government are all 100% selflessly devoted to the community, contrary to the rest of the world). [NEWLINE] [NEWLINE] In such a system, having a "mattress with memory foam" simply<mask> you worked more,<mask> more lucky in your choices of life, had healthier parents, lived in a safer area and were born more intelligent than others IS NOT ACCEPTABLE.<mask> nobody gets the mattress until everybody gets one... and HAS to get one, in the name of equity. [NEWLINE] [NEWLINE] Even with the US government budget, giving out memory foam mattresses to the whole population would be out of the question. Millions of other priorities (like wars, army, weapons and everything else that goes under the god-principle of national security) come before. <mask><mask> there's no foam mattresses for anyone, there's nobody creating a business to make them. That's the way totalitarian/egalitarian systems work,<mask> every decision in your life is made by a higher entity. [NEWLINE] [NEWLINE] Hear me well, by being against that, I am<mask> against a system like<mask> the US are becoming,<mask> literally 50 people in a room can determine who the next president will be, and<mask> laws will or won't pass.<mask> this is not about absence of control vs. communism, it is much more about democracy vs. autocracy. </s>
Label encoding: <s>Hi, [NEWLINE] [NEWLINE] Of course, undeniably, if Cuba could export goods to the US, it would change things a lot. But I don't think the root cause of poverty in Cuba is the embargo. The root cause is the system where every decision is made by the government with an "every one must be equal" approach (and that is assuming, for the sake of the argument, that humans in Cuban government are all 100% selflessly devoted to the community, contrary to the rest of the world). [NEWLINE] [NEWLINE] In such a system, having a "mattress with memory foam" simply because you worked more, where more lucky in your choices of life, had healthier parents, lived in a safer area and were born more intelligent than others IS NOT ACCEPTABLE. So nobody gets the mattress until everybody gets one... and HAS to get one, in the name of equity. [NEWLINE] [NEWLINE] Even with the US government budget, giving out memory foam mattresses to the whole population would be out of the question. Millions of other priorities (like wars, army, weapons and everything else that goes under the god-principle of national security) come before.  So if there's no foam mattresses for anyone, there's nobody creating a business to make them. That's the way totalitarian/egalitarian systems work, where every decision in your life is made by a higher entity. [NEWLINE] [NEWLINE] Hear me well, by being against that, I am also against a system like what the US are becoming, where literally 50 people in a room can determine who the next president will be, and what laws will or won't pass. So this is not about absence of control vs. communism, it is much more about democracy vs. autocracy. </s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But, and this is just a thought. Can you even "know" the Truth of<mask> God wants? I mean apart from even this topic... is that even possible? Yes, we have the Bible<mask> not only is the New Testament itself a selection from the Old,<mask> it is preached<mask> the preachers interpret it -- different sects of each religion believe God said different things. [NEWLINE] [NEWLINE] In English class<mask> we are presented with a book we are asked to interpret.<mask> the author is dead, there is no point in wondering "<mask> did the author mean by ___"<mask> there is no way of truly asking them in person, at least<mask> living. All we have is<mask> we view the text and<mask> it means to us<mask> a person. [NEWLINE] [NEWLINE] These lines you wish to draw I don't believe are God's lines... even by coming and creating this thread you are asking for our own interpretations. God can never give you a true answer in this life and I believe<mask> a Christian all you can do is form your own opinions.<mask> you personally believe that homosexuality is a sin in scripture, then to you and your relationship with God it is...<mask> you believe it is not a sin and that scripture is perhaps taken out of context in our modern society... then that is the "right" line. Every person's relationship with God is different, and like you said,<mask> you mean well and treat your fellow human beings well<mask><mask><mask> you do meet God he will teach you his real lines,<mask> a teacher he will educate you perhaps<mask> you fumbled, and congratulate you<mask> you succeeded. [NEWLINE] [NEWLINE] <mask> in conclusion, draw your own lines<mask> in this world<mask> you will never truly know<mask> God's "lines" are.</s>
Label encoding: <s>But, and this is just a thought. Can you even "know" the Truth of what God wants? I mean apart from even this topic... is that even possible? Yes, we have the Bible but not only is the New Testament itself a selection from the Old, but it is preached as the preachers interpret it -- different sects of each religion believe God said different things. [NEWLINE] [NEWLINE] In English class when we are presented with a book we are asked to interpret. If the author is dead, there is no point in wondering " what did the author mean by ___" because there is no way of truly asking them in person, at least while living. All we have is how we view the text and what it means to us as a person. [NEWLINE] [NEWLINE] These lines you wish to draw I don't believe are God's lines... even by coming and creating this thread you are asking for our own interpretations. God can never give you a true answer in this life and I believe as a Christian all you can do is form your own opinions. If you personally believe that homosexuality is a sin in scripture, then to you and your relationship with God it is... If you believe it is not a sin and that scripture is perhaps taken out of context in our modern society... then that is the "right" line. Every person's relationship with God is different, and like you said, if you mean well and treat your fellow human beings well I think when you do meet God he will teach you his real lines, as a teacher he will educate you perhaps where you fumbled, and congratulate you where you succeeded. [NEWLINE] [NEWLINE] So in conclusion, draw your own lines as in this world as you will never truly know what God's "lines" are.</s>
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**Prenatal hormones and sexual orientation**]( [URL] %20hormones%20and%20sexual%20orientation): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] [STARTQ] The hormonal theory of sexuality holds that, just<mask> exposure to certain hormones plays a role in fetal [sex differentiation]( [URL] ), such exposure<mask> influences the [sexual orientation]( [URL] ) that emerges later in the adult. __Prenatal hormones__ may be seen<mask> the primary determinant of adult sexual orientation, or a co-factor with genes, [biological factors]( [URL] ) and/or environmental and social conditions. [ENDQ] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Bisexuality]( [URL] ) ^| [^Sexual ^orientation]( [URL] ) ^| [^Biology ^and ^sexual ^orientation]( [URL] ) ^| [^Homosexuality]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cj91c3m) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cj91c3m)^. ^Will ^<mask> ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Label encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**Prenatal hormones and sexual orientation**]( [URL] %20hormones%20and%20sexual%20orientation): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] [STARTQ] The hormonal theory of sexuality holds that, just as exposure to certain hormones plays a role in fetal [sex differentiation]( [URL] ), such exposure also influences the [sexual orientation]( [URL] ) that emerges later in the adult. __Prenatal hormones__ may be seen as the primary determinant of adult sexual orientation, or a co-factor with genes, [biological factors]( [URL] ) and/or environmental and social conditions. [ENDQ] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Bisexuality]( [URL] ) ^| [^Sexual ^orientation]( [URL] ) ^| [^Biology ^and ^sexual ^orientation]( [URL] ) ^| [^Homosexuality]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cj91c3m) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cj91c3m)^. ^Will ^ also ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I certainly don't completely agree with the settlements. [NEWLINE] [NEWLINE] <mask> original settlements were built, the West Bank was enemy territory conquered in a war started by enemies, and Israel wanted to secure itself by keeping the land<mask> a buffer from it's enemies (Jordan, and the Palestinians). [NEWLINE] [NEWLINE] Eventually, there were peace accords, and the premise was land for peace.<mask> the commitment is to give the Palestinians back their land. Until there were peace agreements, Israel felt under no obligation to give back land,<mask> they were in a state of war. [NEWLINE] [NEWLINE] After the peace agreements, more and more of the control of the territory has been seceded to the Palestinian Authority, which controls most affairs of the West Bank.<mask>, the settlements remain, and are likely to remain<mask><mask><mask> a final peace deal is not hammered out.<mask> should Israel bear all the burden of peace? In Gaza, they withdrew all Jewish settlements and handed over the territory to the Palestinians, and<mask> they got was rockets being shot into the heart of their cities. Maybe they learned their lesson about unilaterally withdrawing without a sound security arrangement or comprehensive peace deal. [NEWLINE] [NEWLINE] I<mask> don't believe Israel is better than any country for religious reasons. I just think a country's first priority has to be to protect it's own citizens. Sometimes people use the proportion of casualties on the side of the Palestinians to make it seem<mask><mask> Israel doesn't have any reason to be concerned for it's security (during Cast Lead over 1000 Palestinians died vs. 13 Israelis,<mask> people think that<mask> the Palestinians are right and Israel is wrong). Israel has learned from the past, and will try to build peace with the Palestinians<mask> never compromising itself to the point<mask> it could be destroyed.</s>
Label encoding: <s>I certainly don't completely agree with the settlements. [NEWLINE] [NEWLINE] When original settlements were built, the West Bank was enemy territory conquered in a war started by enemies, and Israel wanted to secure itself by keeping the land as a buffer from it's enemies (Jordan, and the Palestinians). [NEWLINE] [NEWLINE] Eventually, there were peace accords, and the premise was land for peace. So the commitment is to give the Palestinians back their land. Until there were peace agreements, Israel felt under no obligation to give back land, as they were in a state of war. [NEWLINE] [NEWLINE] After the peace agreements, more and more of the control of the territory has been seceded to the Palestinian Authority, which controls most affairs of the West Bank. However, the settlements remain, and are likely to remain as long as a final peace deal is not hammered out. Why should Israel bear all the burden of peace? In Gaza, they withdrew all Jewish settlements and handed over the territory to the Palestinians, and what they got was rockets being shot into the heart of their cities. Maybe they learned their lesson about unilaterally withdrawing without a sound security arrangement or comprehensive peace deal. [NEWLINE] [NEWLINE] I also don't believe Israel is better than any country for religious reasons. I just think a country's first priority has to be to protect it's own citizens. Sometimes people use the proportion of casualties on the side of the Palestinians to make it seem as if Israel doesn't have any reason to be concerned for it's security (during Cast Lead over 1000 Palestinians died vs. 13 Israelis, so people think that therefore the Palestinians are right and Israel is wrong). Israel has learned from the past, and will try to build peace with the Palestinians while never compromising itself to the point where it could be destroyed.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Alimony has been gender neutral in the United States for almost 45 years now. [ENDQ] [NEWLINE] And<mask> does that change? See the massachusetts alimony reform. It's like saying a law on inveterate shoppers is gender-neutral. [NEWLINE] [NEWLINE] [STARTQ] <mask> its publication in 1985, sociologist Lenore Weitzman's The Divorce Revolution has had a critical role in shaping the national debate on divorce and its economic effects. In particular, the book's claim that in the year after divorce women's standard of living decreased by a whopping 73 percent<mask> men enjoyed an increase of 43 percent caught the attention of pundits, legislators, and judges. This statistic has become one of the philosophical bases for deciding child custody and property division in divorce cases. It has<mask> altered public perceptions of men, women, and divorce. It was cited hundreds of times in news stories, scholarly studies, and law review articles last year, and was regarded<mask> clearly<mask> holy writ that President Clinton cited it too in his budget proposal this year<mask> part of his attack on deadbeat dads. [ENDQ] [NEWLINE] [STARTQ] The only problem with this statistic,<mask><mask>, is that it turns out to be wrong. [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> people who don't conform to gender roles are deviant? [ENDQ] [NEWLINE] People who don't conform to the male and female paradigm are deviant. The<mask> -called rising from male and female constructs(and not merely masculine/feminine distinction) is not a 'rising' for those who are already that way. The lesbian excerpt was an example. [NEWLINE] [NEWLINE] [STARTQ] You believe that feminism is attempted to convert girls everywhere into lesbians? [ENDQ] [NEWLINE] <mask> feminism is about seeing women<mask> people and women are not people, then who are? </s>
Label encoding: <s> [STARTQ] Alimony has been gender neutral in the United States for almost 45 years now. [ENDQ] [NEWLINE] And what does that change? See the massachusetts alimony reform. It's like saying a law on inveterate shoppers is gender-neutral. [NEWLINE] [NEWLINE] [STARTQ] Since its publication in 1985, sociologist Lenore Weitzman's The Divorce Revolution has had a critical role in shaping the national debate on divorce and its economic effects. In particular, the book's claim that in the year after divorce women's standard of living decreased by a whopping 73 percent while men enjoyed an increase of 43 percent caught the attention of pundits, legislators, and judges. This statistic has become one of the philosophical bases for deciding child custody and property division in divorce cases. It has also altered public perceptions of men, women, and divorce. It was cited hundreds of times in news stories, scholarly studies, and law review articles last year, and was regarded so clearly as holy writ that President Clinton cited it too in his budget proposal this year as part of his attack on deadbeat dads. [ENDQ] [NEWLINE] [STARTQ] The only problem with this statistic, in fact, is that it turns out to be wrong. [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] So people who don't conform to gender roles are deviant? [ENDQ] [NEWLINE] People who don't conform to the male and female paradigm are deviant. The so -called rising from male and female constructs(and not merely masculine/feminine distinction) is not a 'rising' for those who are already that way. The lesbian excerpt was an example. [NEWLINE] [NEWLINE] [STARTQ] You believe that feminism is attempted to convert girls everywhere into lesbians? [ENDQ] [NEWLINE] If feminism is about seeing women as people and women are not people, then who are? </s>
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] He literally said "One major reason we believe we need firearms is the 2nd amendment." That's the statement I was disagreeing with. [ENDQ] [NEWLINE] Keyword in that statement is "believe." [NEWLINE] [NEWLINE] [STARTQ] We're talking about gun bans, [ENDQ] [NEWLINE] I don't see<mask> that makes my point wrong. [NEWLINE] [NEWLINE] [STARTQ] the second amendment is completely<mask> gun bans are unconstitutional. [ENDQ] [NEWLINE] That's<mask> we change the second amendment. [NEWLINE] [NEWLINE] [STARTQ] The right to free speech and the right to bear arms are both guaranteed in the constitution. [ENDQ] [NEWLINE] That does not mean that the 1st and 2nd amendment are a legitimate analogy. OP isn't saying we should get rid of free speech, or fair trial, he is saying we should get rid of guns. [NEWLINE] [NEWLINE] [STARTQ] <mask> we are talking about an issue like space exploration that isn't remotely covered by the Constitution, I'll agree with you. [ENDQ] [NEWLINE] [STARTQ] In terms of words they actually wrote on paper? I'll disagree heartily. [ENDQ] [NEWLINE] I don't think you know<mask> an appeal to antiquity is.<mask> I am saying is that<mask> happened in the 1700s is not relevant to today [NEWLINE] [NEWLINE] [STARTQ] <mask> the founder's thought about the issue is extremely significant,<mask> they wrote it, and they sure knew<mask> they meant. [ENDQ] [NEWLINE] You know<mask> they didn't know anything about? The 21st century. [NEWLINE] [NEWLINE] [STARTQ] All of that is irrelevant, which I'll cover in the next block. [ENDQ] [NEWLINE] <mask> it is relevant<mask> you bring it up,<mask> not<mask> I do? Okay then. [NEWLINE] [NEWLINE] [STARTQ] The debate is all about<mask> the Second Amendment means, [ENDQ] [NEWLINE] It is<mask> a debate about whether it is still relevant in this day and age.</s>
Label encoding: <s> [STARTQ] He literally said "One major reason we believe we need firearms is the 2nd amendment." That's the statement I was disagreeing with. [ENDQ] [NEWLINE] Keyword in that statement is "believe." [NEWLINE] [NEWLINE] [STARTQ] We're talking about gun bans, [ENDQ] [NEWLINE] I don't see how that makes my point wrong. [NEWLINE] [NEWLINE] [STARTQ] the second amendment is completely why gun bans are unconstitutional. [ENDQ] [NEWLINE] That's why we change the second amendment. [NEWLINE] [NEWLINE] [STARTQ] The right to free speech and the right to bear arms are both guaranteed in the constitution. [ENDQ] [NEWLINE] That does not mean that the 1st and 2nd amendment are a legitimate analogy. OP isn't saying we should get rid of free speech, or fair trial, he is saying we should get rid of guns. [NEWLINE] [NEWLINE] [STARTQ] If we are talking about an issue like space exploration that isn't remotely covered by the Constitution, I'll agree with you. [ENDQ] [NEWLINE] [STARTQ] In terms of words they actually wrote on paper? I'll disagree heartily. [ENDQ] [NEWLINE] I don't think you know what an appeal to antiquity is. What I am saying is that what happened in the 1700s is not relevant to today [NEWLINE] [NEWLINE] [STARTQ] What the founder's thought about the issue is extremely significant, because they wrote it, and they sure knew what they meant. [ENDQ] [NEWLINE] You know what they didn't know anything about? The 21st century. [NEWLINE] [NEWLINE] [STARTQ] All of that is irrelevant, which I'll cover in the next block. [ENDQ] [NEWLINE] So it is relevant when you bring it up, but not when I do? Okay then. [NEWLINE] [NEWLINE] [STARTQ] The debate is all about what the Second Amendment means, [ENDQ] [NEWLINE] It is also a debate about whether it is still relevant in this day and age.</s>
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They are always portrayed<mask> being right along side people that have actually done relevant things to contribute to society today. <mask><mask> that both culturally and musically they were not that big of a deal.  There are loads of better musicians and many more interesting people.  After all, they did start out<mask> an equivalent to today's boy bands; they were a figure head, something for girls to oogle at.  There are<mask> many conspiracies about The Beatles and I don't think there have been the same amount of cover bands for any other band in history (<mask> maybe The Dead,<mask> they were great musicians). [NEWLINE] [NEWLINE] I know this explanation is a bit scattered,<mask> it's late.  I mainly wanted to make this<mask> I wanted to get my point across.  I will further explain, in more lengthy detail, in comments. [NEWLINE] [NEWLINE] Change my view. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Label encoding: <s>They are always portrayed as being right along side people that have actually done relevant things to contribute to society today.  I think that both culturally and musically they were not that big of a deal.  There are loads of better musicians and many more interesting people.  After all, they did start out as an equivalent to today's boy bands; they were a figure head, something for girls to oogle at.  There are so many conspiracies about The Beatles and I don't think there have been the same amount of cover bands for any other band in history ( besides maybe The Dead, but they were great musicians). [NEWLINE] [NEWLINE] I know this explanation is a bit scattered, but it's late.  I mainly wanted to make this because I wanted to get my point across.  I will further explain, in more lengthy detail, in comments. [NEWLINE] [NEWLINE] Change my view. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>your reading comprehension is lacking very badly, my friend. that makes it very difficult to have any sort of serious conversation with you.<mask> let me give this one try. [NEWLINE] [NEWLINE] [STARTQ] <mask> the other guy is typically offering her is that he treats her like shit (in one form or another). [ENDQ] [NEWLINE] nope. some guys may<mask> do this.<mask><mask> someone who has actually dated someone really, really shitty who I definitely shouldn't have dated, I have to say that I remember the qualities that made me attracted to him. they didn't make up for his flaws,<mask><mask> him being such a complete shithead not worth dating, some "nice guy" could certainly have observed to an extent<mask> was going on and could have learned from the few things he was doing right and that made me attracted to him. [NEWLINE] [NEWLINE] [STARTQ] People may notice that other people are being nice,<mask> the assumption is that there is an ulterior motive. [ENDQ] [NEWLINE] nope.<mask> a guy is nice to me, I generally assume he's, you know, nice. And I notice.<mask>,<mask>, he asks me out, I reject him, and he insults me, then that's the moment<mask> it becomes clear that he wasn't really a nice guy, and is only superficially nice to people he wants to fuck.<mask> he were really nice, he'd continue to be nice (or at least neutral) after being rejected (and yes, this is possible and people do this. I've been rejected and continued to be civil, and I've rejected several guys who continued to be civil. this isn't much to ask, and<mask> someone were genuinely nice, it would come very easily to them). [NEWLINE] [NEWLINE] </s>
Label encoding: <s>your reading comprehension is lacking very badly, my friend. that makes it very difficult to have any sort of serious conversation with you. but let me give this one try. [NEWLINE] [NEWLINE] [STARTQ] What the other guy is typically offering her is that he treats her like shit (in one form or another). [ENDQ] [NEWLINE] nope. some guys may also do this. but as someone who has actually dated someone really, really shitty who I definitely shouldn't have dated, I have to say that I remember the qualities that made me attracted to him. they didn't make up for his flaws, but despite him being such a complete shithead not worth dating, some "nice guy" could certainly have observed to an extent what was going on and could have learned from the few things he was doing right and that made me attracted to him. [NEWLINE] [NEWLINE] [STARTQ] People may notice that other people are being nice, but the assumption is that there is an ulterior motive. [ENDQ] [NEWLINE] nope. If a guy is nice to me, I generally assume he's, you know, nice. And I notice. IF, however, he asks me out, I reject him, and he insults me, then that's the moment where it becomes clear that he wasn't really a nice guy, and is only superficially nice to people he wants to fuck. if he were really nice, he'd continue to be nice (or at least neutral) after being rejected (and yes, this is possible and people do this. I've been rejected and continued to be civil, and I've rejected several guys who continued to be civil. this isn't much to ask, and if someone were genuinely nice, it would come very easily to them). [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your employer isn't paying for your health insurance.  They are paying for your employment.  You, and your fellow employees, are paying for your health insurance through a group insurance policy.  Some of that expensive you pay directly in premiums deducted from your paychecks.  The portion that your employer "covers" was a agreed upon part of your compensation for working there.  Just like vacation or holiday pay.  Instead of paying you more dollars per hour directly, they instead offered other financial compensation in the form of a 401k, life insurance, paid leave, and a health insurance plan.  "Their" expense is coming out of your agreed upon financial compensation. [NEWLINE] [NEWLINE] In the United States, any employer of more than 50 employees must provide full-time employees the option of enrolling in health insurance.  It is factored in to your pay, just like 401k matching or retirement contributions are, and<mask> you aren't using it, it is a portion of your compensation that you voluntarily not receiving.  We require that they provide holiday pay.  We do not allow employers to not pay certain holidays<mask> they don't celebrate them.  We require that military member's be allowed leave for their duty.  We do not allow employers to discriminate against servicemen<mask> they are opposed to the war.  We require overtime pay.  We don't allow an employer to not pay overtime,<mask> "<mask> they grew up, 60 hours was expected of everyone."  We require that they provide a health insurance option<mask> part of my compensation,<mask> should they get to pick<mask> it covers? [NEWLINE] [NEWLINE] Personally, I don't think we should have our health insurance mixed with our employers.  </s>
Label encoding: <s>Your employer isn't paying for your health insurance.  They are paying for your employment.  You, and your fellow employees, are paying for your health insurance through a group insurance policy.  Some of that expensive you pay directly in premiums deducted from your paychecks.  The portion that your employer "covers" was a agreed upon part of your compensation for working there.  Just like vacation or holiday pay.  Instead of paying you more dollars per hour directly, they instead offered other financial compensation in the form of a 401k, life insurance, paid leave, and a health insurance plan.  "Their" expense is coming out of your agreed upon financial compensation. [NEWLINE] [NEWLINE] In the United States, any employer of more than 50 employees must provide full-time employees the option of enrolling in health insurance.  It is factored in to your pay, just like 401k matching or retirement contributions are, and if you aren't using it, it is a portion of your compensation that you voluntarily not receiving.  We require that they provide holiday pay.  We do not allow employers to not pay certain holidays because they don't celebrate them.  We require that military member's be allowed leave for their duty.  We do not allow employers to discriminate against servicemen because they are opposed to the war.  We require overtime pay.  We don't allow an employer to not pay overtime, because " where they grew up, 60 hours was expected of everyone."  We require that they provide a health insurance option as part of my compensation, why should they get to pick what it covers? [NEWLINE] [NEWLINE] Personally, I don't think we should have our health insurance mixed with our employers.  </s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask><mask> highschoolers benefit from school zone speed restrictions around their highschools. [NEWLINE] [NEWLINE] <mask> it's true that they're (hopefully) more cognitively aware of their surroundings than say elementary age kids, it is completely incorrect to assume that teenagers are actively paying attention to their surroundings at all times. Many times you'll see them milling about in small clusters, talking, yelling, laughing around the school. Chances are they're paying more attention to making their group of friends laugh, doing something stupid for attention or to show off than they are to the flow of traffic on the nearby street. [NEWLINE] [NEWLINE] Combine this lack of awareness with the density of foot traffic around a school and you have a perfect situation for someone to get hurt. [NEWLINE] [NEWLINE] Should drivers pay attention and not hit peds (even<mask> the peds are not paying attention themselves)? Of course. Should the peds pay attention<mask> they're in close proximity to a busy street? Again, of course they should. Does either group actually do this 100% of the time? No, they do not. [NEWLINE] [NEWLINE] For that reason, the school zone speed limits are in place<mask> a "hard mitigation" to the vehicle side of the equation. [NEWLINE] [NEWLINE] There's<mask> the fact that all these HS kiddos *are actually someone's offspring* (your "think of the kids!" argument). I'm sure<mask> you had a kid that you loved (even<mask> they were in their shithead teenager phase), you'd be devastated<mask> they were struck by a speeding car just outside their highschool. [NEWLINE] [NEWLINE] The laws are in place to protect something that's very valuable to us<mask> humans, namely, our offspring.</s>
Label encoding: <s>I would argue that highschoolers benefit from school zone speed restrictions around their highschools. [NEWLINE] [NEWLINE] While it's true that they're (hopefully) more cognitively aware of their surroundings than say elementary age kids, it is completely incorrect to assume that teenagers are actively paying attention to their surroundings at all times. Many times you'll see them milling about in small clusters, talking, yelling, laughing around the school. Chances are they're paying more attention to making their group of friends laugh, doing something stupid for attention or to show off than they are to the flow of traffic on the nearby street. [NEWLINE] [NEWLINE] Combine this lack of awareness with the density of foot traffic around a school and you have a perfect situation for someone to get hurt. [NEWLINE] [NEWLINE] Should drivers pay attention and not hit peds (even when the peds are not paying attention themselves)? Of course. Should the peds pay attention when they're in close proximity to a busy street? Again, of course they should. Does either group actually do this 100% of the time? No, they do not. [NEWLINE] [NEWLINE] For that reason, the school zone speed limits are in place as a "hard mitigation" to the vehicle side of the equation. [NEWLINE] [NEWLINE] There's also the fact that all these HS kiddos *are actually someone's offspring* (your "think of the kids!" argument). I'm sure if you had a kid that you loved (even if they were in their shithead teenager phase), you'd be devastated if they were struck by a speeding car just outside their highschool. [NEWLINE] [NEWLINE] The laws are in place to protect something that's very valuable to us as humans, namely, our offspring.</s>
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sure,<mask> gold is a pretty terrible investment. It's a great store of wealth,<mask> it's not an investment. On average, it keeps up with inflation,<mask> it swings quite heavily from one side to another<mask> there is uncertainty in the market. It can go double its "intrinsic value" during a market crash and can be had very cheaply<mask> the market is doing really well. [NEWLINE] [NEWLINE] This makes it an excellent hedge and an excellent diversifying tool, provided it is 10% or less of your entire portfolio.<mask><mask> in the world would you just take that tax hit and buy it all today? Instead, you should be investing 10% of your income every single paycheque into your retirement before you ever spend it on anything else. For the next little<mask>, invest in gold until you reach your 5% or 10% mark. It would be insane to divest of a healthy stock portfolio and take a tax penalty in order to buy gold today. [NEWLINE] [NEWLINE] Remember, the kinds of risks you are talking about aren't likely to happen and basically can't happen in the next five years.<mask>, it's worth noting that<mask> those things do happen, the people with gold will not control anything. The people with steel conveniently shaped into guns and bullets,<mask><mask><mask><mask>, will. [NEWLINE] [NEWLINE] <mask> gold is a fantastic hedge against swings in the market,<mask> is pretty much useless in the kinds of events you are talking about. [NEWLINE] [NEWLINE] <mask> you distrust stocks and intangible assets, I recommend getting into real estate. It won't help during an apocalypse,<mask> can give you real wealth increase and a real asset,<mask> it's a bigger pain in the ass than stocks. </s>
Label encoding: <s>Sure, but gold is a pretty terrible investment. It's a great store of wealth, but it's not an investment. On average, it keeps up with inflation, although it swings quite heavily from one side to another when there is uncertainty in the market. It can go double its "intrinsic value" during a market crash and can be had very cheaply when the market is doing really well. [NEWLINE] [NEWLINE] This makes it an excellent hedge and an excellent diversifying tool, provided it is 10% or less of your entire portfolio. But why in the world would you just take that tax hit and buy it all today? Instead, you should be investing 10% of your income every single paycheque into your retirement before you ever spend it on anything else. For the next little while, invest in gold until you reach your 5% or 10% mark. It would be insane to divest of a healthy stock portfolio and take a tax penalty in order to buy gold today. [NEWLINE] [NEWLINE] Remember, the kinds of risks you are talking about aren't likely to happen and basically can't happen in the next five years. Also, it's worth noting that if those things do happen, the people with gold will not control anything. The people with steel conveniently shaped into guns and bullets, on the other hand, will. [NEWLINE] [NEWLINE] So gold is a fantastic hedge against swings in the market, but is pretty much useless in the kinds of events you are talking about. [NEWLINE] [NEWLINE] If you distrust stocks and intangible assets, I recommend getting into real estate. It won't help during an apocalypse, but can give you real wealth increase and a real asset, although it's a bigger pain in the ass than stocks. </s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me put it this way: The other 51 weeks of the year are *all* "Straight Ski Week." (Well, not 51 in the case of skiing,<mask> you get the point.)<mask> a straight couple decides to go skiing at any other time of year, they'll be surrounded by [STARTQ] 90% straight people. There's nothing *wrong* with being around a bunch of people who share your sexuality. It certainly doesn't imply that you're going to have sex with all of them. [ENDQ] [NEWLINE] Unless you live in a gay mecca and make a serious and concerted effort to only frequent gay establishments, the majority of your life outside your home is going to consist of "Straight" activities. You can hang out with straight people any time you want. [NEWLINE] [NEWLINE] <mask> you go to Gay Ski Week, you'll be surrounded by...probably 30-40% gay people,<mask> this is like most other gay events. 50% would be exceptional. Gay events at otherwise-straight venues rarely produce a gay majority, and they certainly never manage the sort of overwhelming majority that straight people are accustomed to (even most gay bars rarely see [STARTQ] 90%). [ENDQ] [NEWLINE] Going to Gay Ski Week isn't about segregating yourselves. It's about having the chance to be around people you normally don't get to spend much time with, people who have something important in common with you and who<mask> probably have a range of diverse life experiences that you aren't often exposed to. And it's about creating a large gay presence<mask> that the straight people who are there - who will still be in the majority - have the rare opportunity to interact with large numbers of gay people in one place that isn't a gay bar.</s>
Label encoding: <s>Let me put it this way: The other 51 weeks of the year are *all* "Straight Ski Week." (Well, not 51 in the case of skiing, but you get the point.) If a straight couple decides to go skiing at any other time of year, they'll be surrounded by [STARTQ] 90% straight people. There's nothing *wrong* with being around a bunch of people who share your sexuality. It certainly doesn't imply that you're going to have sex with all of them. [ENDQ] [NEWLINE] Unless you live in a gay mecca and make a serious and concerted effort to only frequent gay establishments, the majority of your life outside your home is going to consist of "Straight" activities. You can hang out with straight people any time you want. [NEWLINE] [NEWLINE] If you go to Gay Ski Week, you'll be surrounded by...probably 30-40% gay people, if this is like most other gay events. 50% would be exceptional. Gay events at otherwise-straight venues rarely produce a gay majority, and they certainly never manage the sort of overwhelming majority that straight people are accustomed to (even most gay bars rarely see [STARTQ] 90%). [ENDQ] [NEWLINE] Going to Gay Ski Week isn't about segregating yourselves. It's about having the chance to be around people you normally don't get to spend much time with, people who have something important in common with you and who also probably have a range of diverse life experiences that you aren't often exposed to. And it's about creating a large gay presence so that the straight people who are there - who will still be in the majority - have the rare opportunity to interact with large numbers of gay people in one place that isn't a gay bar.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As I was writing the title of this post, I realized that the other circumstance under which it might be morally permissible is in a procedure referred to<mask> "reduction,"<mask>, in a case of multiple pregnancy (think octomom), one or more of the fetuses are killed in order to prevent the deaths of all the others through miscarriage. In any case, the principle remains that the only permissible reason to take a life is to save one or more lives. I do not support exceptions in cases of rape or incest; I simply don't see<mask> killing the offspring makes the situation more palatable. [NEWLINE] [NEWLINE] I welcome any comments regarding the point at which life begins,<mask> I warn you, I know biology. [NEWLINE] [NEWLINE] EDIT: I would like to point out that,<mask> the votes are not<mask> visible, I can tell from my profile page that I have received a huge number of downvotes from r/changemyview in the last few hours. Downvoting<mask> you disagree with an argument is really not in keeping with the spirit of this subreddit.<mask> you disagree, tell me<mask> my argument is flawed. I have not been insulting, demeaning, silly, illogical, low-effort, or anything truly deserving of a downvote. I have given<mask> cogent a defense of my position<mask> I am capable, and<mask> people have pointed out potential logical flaws, I have been happy to engage constructively. I have only downvoted posts that were truly inane, which I am glad to see were relatively few. Again, please don't downvote me just<mask> you disagree with my possibly incorrect--<mask> sure<mask> hell not poorly considered--position. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>As I was writing the title of this post, I realized that the other circumstance under which it might be morally permissible is in a procedure referred to as "reduction," where, in a case of multiple pregnancy (think octomom), one or more of the fetuses are killed in order to prevent the deaths of all the others through miscarriage. In any case, the principle remains that the only permissible reason to take a life is to save one or more lives. I do not support exceptions in cases of rape or incest; I simply don't see how killing the offspring makes the situation more palatable. [NEWLINE] [NEWLINE] I welcome any comments regarding the point at which life begins, although I warn you, I know biology. [NEWLINE] [NEWLINE] EDIT: I would like to point out that, although the votes are not yet visible, I can tell from my profile page that I have received a huge number of downvotes from r/changemyview in the last few hours. Downvoting because you disagree with an argument is really not in keeping with the spirit of this subreddit. If you disagree, tell me why my argument is flawed. I have not been insulting, demeaning, silly, illogical, low-effort, or anything truly deserving of a downvote. I have given as cogent a defense of my position as I am capable, and where people have pointed out potential logical flaws, I have been happy to engage constructively. I have only downvoted posts that were truly inane, which I am glad to see were relatively few. Again, please don't downvote me just because you disagree with my possibly incorrect-- but sure as hell not poorly considered--position. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your appear to believe consumers of porn view it<mask> some sort of educational video,<mask><mask> it is viewed<mask> a step by step guide on<mask> to behave and act. It isn't, and I sincerely don't believe anyone truelly thinks it is. [NEWLINE] [NEWLINE] Porn is fiction, just like regular movies. I'm sure you don't believe that we regard the behavior in general movies to be an accurate model of<mask> to behave in every day life,<mask><mask> is porn?<mask> makes you think that people don't go out shooting, robbing banks, kidnapping etc simply<mask> they saw it in a film,<mask> then go out and rape someone<mask> they saw a snuff film? [NEWLINE] [NEWLINE] <mask> for your points: [NEWLINE] [NEWLINE] a)<mask><mask> you're over analysing porn. There is no hidden meaning, subtle inferences or undertones. The Pizza guy isn't being paid more than his female coworker, the lonely housewife isn't forced to stay at home<mask> of the patriarchy, it's 2+ people fucking, nothing more. [NEWLINE] [NEWLINE] b) That argument is just<mask> valid in reverse. Men are nothing more than a penis with legs, with no value other than<mask> long they can fuck a woman for, or<mask> big their cock is. [NEWLINE] [NEWLINE] c)<mask> someone doesn't consent to being in a film, that isn't pornography, that is illegal.<mask> for later regretting it, we regret actions in probably every sphere of life, that's called life, not pornography. [NEWLINE] [NEWLINE] d) Porn has positive health benefits<mask>. [NEWLINE] [NEWLINE] e) Again, over analysing. People don't consider porn to be an accurate representation of society and more than other films.</s>
Label encoding: <s>Your appear to believe consumers of porn view it as some sort of educational video, as if it is viewed as a step by step guide on how to behave and act. It isn't, and I sincerely don't believe anyone truelly thinks it is. [NEWLINE] [NEWLINE] Porn is fiction, just like regular movies. I'm sure you don't believe that we regard the behavior in general movies to be an accurate model of how to behave in every day life, so why is porn? What makes you think that people don't go out shooting, robbing banks, kidnapping etc simply because they saw it in a film, but then go out and rape someone because they saw a snuff film? [NEWLINE] [NEWLINE] As for your points: [NEWLINE] [NEWLINE] a) I think you're over analysing porn. There is no hidden meaning, subtle inferences or undertones. The Pizza guy isn't being paid more than his female coworker, the lonely housewife isn't forced to stay at home because of the patriarchy, it's 2+ people fucking, nothing more. [NEWLINE] [NEWLINE] b) That argument is just as valid in reverse. Men are nothing more than a penis with legs, with no value other than how long they can fuck a woman for, or how big their cock is. [NEWLINE] [NEWLINE] c) If someone doesn't consent to being in a film, that isn't pornography, that is illegal. As for later regretting it, we regret actions in probably every sphere of life, that's called life, not pornography. [NEWLINE] [NEWLINE] d) Porn has positive health benefits also. [NEWLINE] [NEWLINE] e) Again, over analysing. People don't consider porn to be an accurate representation of society and more than other films.</s>
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The AR15 actually does not overpenetrate<mask> you would expect. It tends to tumble after a few walls of overpenetration.<mask> you should always be aware<mask> direction you are shooting, the tumbling of a round will cause it to lose power quicker than a 12 gauge **slug or 00 Buck.** [ENDQ] [NEWLINE] Sure,<mask> compared to the most over penetrating rounds the 12 gauge has to offer, then, yes, the AR isn't<mask> bad. [NEWLINE] [NEWLINE] [STARTQ] The AR15 is much softer recoiling than a shotgun and easier to control than a handgun. You can create an AR15 that's more compact and easier to use in a house buy buying a Short Barreled Rifle tax stamp from the ATF or buying an AR15 Pistol. [ENDQ] [NEWLINE] You can build a magazine fed shotgun with a pistol grip before even having to deal with the ATF and NFA. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> to accessories that may or may not make the firearm "better" the AR15 platform is extremely modular. The ability to change from a.223/5.56 to.300 blackout to.50 Beowulf to cailbers I've never heard of to meet the exact needs of anyone who needs a rifle makes this one platform good for anyone from farmers to competitive shooters to hardline survivalists. [ENDQ] [NEWLINE] This is a great argument for the AR15. [NEWLINE] [NEWLINE] [STARTQ] I really think that the AR15 is a good rifle that meets the needs and wants of a large number of people. I believe that it is the everyman's firearm and helps lessen the number of physical firearms that someone would need just by being modular. [ENDQ] [NEWLINE] <mask> a good argument. </s>
Label encoding: <s> [STARTQ] The AR15 actually does not overpenetrate as you would expect. It tends to tumble after a few walls of overpenetration. While you should always be aware what direction you are shooting, the tumbling of a round will cause it to lose power quicker than a 12 gauge **slug or 00 Buck.** [ENDQ] [NEWLINE] Sure, when compared to the most over penetrating rounds the 12 gauge has to offer, then, yes, the AR isn't as bad. [NEWLINE] [NEWLINE] [STARTQ] The AR15 is much softer recoiling than a shotgun and easier to control than a handgun. You can create an AR15 that's more compact and easier to use in a house buy buying a Short Barreled Rifle tax stamp from the ATF or buying an AR15 Pistol. [ENDQ] [NEWLINE] You can build a magazine fed shotgun with a pistol grip before even having to deal with the ATF and NFA. [NEWLINE] [NEWLINE] [STARTQ] In addition to accessories that may or may not make the firearm "better" the AR15 platform is extremely modular. The ability to change from a.223/5.56 to.300 blackout to.50 Beowulf to cailbers I've never heard of to meet the exact needs of anyone who needs a rifle makes this one platform good for anyone from farmers to competitive shooters to hardline survivalists. [ENDQ] [NEWLINE] This is a great argument for the AR15. [NEWLINE] [NEWLINE] [STARTQ] I really think that the AR15 is a good rifle that meets the needs and wants of a large number of people. I believe that it is the everyman's firearm and helps lessen the number of physical firearms that someone would need just by being modular. [ENDQ] [NEWLINE] Also a good argument. </s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm assuming this is in response to the shootings in IV. First, Go Gauchos. [NEWLINE] [NEWLINE] Second... now you care about the legality? No, it is not illegal to ( in the extremely gun-phobic state of CA) to smith your own gun,<mask><mask><mask> it can be picked up by a metal detector. There's a lot of debate about "Ghost guns" happening right now<mask> of the 80% lower receivers which do not require any type of documentation to acquire, ( [URL] ) [NEWLINE] [NEWLINE] Here's the real problem, by attempting to enforce gun control you are making me a criminal instead of fighting actual crime. I am not going to surrender my firearms, under any circumstances,<mask> instead of being a small business owner who owns guns, I will be a closet felon. Who does this benefit? No one. [NEWLINE] [NEWLINE] Look at<mask> this situation played out for the war on drugs, we sacrificed countless lives and liberties and accomplished<mask>? A campaign of misinformation and crony capitalism full of no-bid contracts for private prisons. [NEWLINE] [NEWLINE] This really is the classic "Better mouse trap" scenario; The majority of gun owners aren't inbred idiots or psychopaths. They are your coworkers, superiors, lawyers, legislators and doctors.<mask><mask><mask> any gun control measure you put into place will have loopholes that will be found and exploited, look at CA bullet button - Do you really think hardend criminals were the first ones to buy the CA legal AR-15s?<mask>, this being the case,<mask> measures are you willing to take in order to prevent the occasional mass shooting? Which, depending on<mask> you want to quantify, have gone down<mask> 1980.</s>
Label encoding: <s>I'm assuming this is in response to the shootings in IV. First, Go Gauchos. [NEWLINE] [NEWLINE] Second... now you care about the legality? No, it is not illegal to ( in the extremely gun-phobic state of CA) to smith your own gun, so long as it can be picked up by a metal detector. There's a lot of debate about "Ghost guns" happening right now because of the 80% lower receivers which do not require any type of documentation to acquire, ( [URL] ) [NEWLINE] [NEWLINE] Here's the real problem, by attempting to enforce gun control you are making me a criminal instead of fighting actual crime. I am not going to surrender my firearms, under any circumstances, so instead of being a small business owner who owns guns, I will be a closet felon. Who does this benefit? No one. [NEWLINE] [NEWLINE] Look at how this situation played out for the war on drugs, we sacrificed countless lives and liberties and accomplished what? A campaign of misinformation and crony capitalism full of no-bid contracts for private prisons. [NEWLINE] [NEWLINE] This really is the classic "Better mouse trap" scenario; The majority of gun owners aren't inbred idiots or psychopaths. They are your coworkers, superiors, lawyers, legislators and doctors. Because of this any gun control measure you put into place will have loopholes that will be found and exploited, look at CA bullet button - Do you really think hardend criminals were the first ones to buy the CA legal AR-15s? So, this being the case, what measures are you willing to take in order to prevent the occasional mass shooting? Which, depending on how you want to quantify, have gone down since 1980.</s>
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Has anyone pointed out that, for many people, the idea of sleeping with a close blood relative feels kind of repulsive? My sister is a gorgeous woman, and I love her deeply,<mask> the thought of anything sexual with her makes me feel very sick. I don't think this is conditioning. I grew up with her, I saw her in nappies, I remember her<mask> a child, I saw her tantrums,<mask><mask><mask> we both recognise that we now have a bond way stronger than anything sexual. I'm using my sister<mask> an example,<mask> she is the closest to me in age,<mask> I<mask> feel the same about other close blood relatives - love them deeply,<mask> sex? Ew no. I'm trying to<mask><mask><mask><mask> that this is a natural biological response, to encourage you to spread your genes and diversify your offsping, and<mask>, that the 'Ew' response to other people's incest is a kind of 'WTF?'<mask> they don't understand<mask> they can get past that natural 'ew' barrier that most of us seem to have. [NEWLINE] [NEWLINE] <mask>,<mask><mask> incest has been typically riddled with stuff that isn't really about sex. Particularly in parent/child incest, there seems to be a lot of power games and abusive behaviour going on. I have read a fair few first hand accounts on reddit of consensual sibling sexual relationships that have never hurt anyone, and<mask> that's the case, meh, I'm certainly not here to judge anyone esle. [NEWLINE] [NEWLINE] Having said all this, I'm from England,<mask> legally, first cousins can marry and have children,<mask><mask> do I know. :) </s><pad>
Label encoding: <s>Has anyone pointed out that, for many people, the idea of sleeping with a close blood relative feels kind of repulsive? My sister is a gorgeous woman, and I love her deeply, but the thought of anything sexual with her makes me feel very sick. I don't think this is conditioning. I grew up with her, I saw her in nappies, I remember her as a child, I saw her tantrums, but I think we both recognise that we now have a bond way stronger than anything sexual. I'm using my sister as an example, because she is the closest to me in age, but I also feel the same about other close blood relatives - love them deeply, but sex? Ew no. I'm trying to argue that I think that this is a natural biological response, to encourage you to spread your genes and diversify your offsping, and therefore, that the 'Ew' response to other people's incest is a kind of 'WTF?' because they don't understand how they can get past that natural 'ew' barrier that most of us seem to have. [NEWLINE] [NEWLINE] Additionally, I think incest has been typically riddled with stuff that isn't really about sex. Particularly in parent/child incest, there seems to be a lot of power games and abusive behaviour going on. I have read a fair few first hand accounts on reddit of consensual sibling sexual relationships that have never hurt anyone, and if that's the case, meh, I'm certainly not here to judge anyone esle. [NEWLINE] [NEWLINE] Having said all this, I'm from England, where legally, first cousins can marry and have children, so what do I know. :) </s><pad>
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Faith is a journey, and is different than religion.  Faith is something you have to figure out for yourself and can always be changing,<mask> can your own personal beliefs.  My beliefs have certainly changed in ways I never expected in the past 31 years, and my own faith is something I struggle with all the time. <mask><mask> I was raised to believe in God, I'm not always sure just<mask> I feel at this point in my life. <mask> I was a child I went to church with my parents and they taught me<mask> they believed and<mask> I believed the same thing.  And at that time in my life the church was the perfect place for me.  I was surrounded by people who had the same beliefs<mask> I did, and in that way the church was like a comfort blanket.  Religion is like a blanket of people who believe and think the same way that you do and who can help you<mask> you journey<mask> life. <mask> my beliefs changed I tended to shy away from the church<mask> it no longer fit with my life. <mask> my beliefs continue to change perhaps I'll find a religion and a church that fit me.  Religion is just a set of beliefs, that's all.  It's important to look past the term religion and the names of said religions.  Those Baptists who picket funerals are not the same Baptists that I've met.  Those Muslims who terrorize others are not the same Muslims who practice peace.  They really are their own separate religions. [NEWLINE] [NEWLINE] My point is, that it's not religion that drives people to be evil. <mask> its evil people who drive religion to perpetuate their evil.</s>
Label encoding: <s>Faith is a journey, and is different than religion.  Faith is something you have to figure out for yourself and can always be changing, as can your own personal beliefs.  My beliefs have certainly changed in ways I never expected in the past 31 years, and my own faith is something I struggle with all the time.  Even though I was raised to believe in God, I'm not always sure just how I feel at this point in my life.  When I was a child I went to church with my parents and they taught me what they believed and so I believed the same thing.  And at that time in my life the church was the perfect place for me.  I was surrounded by people who had the same beliefs as I did, and in that way the church was like a comfort blanket.  Religion is like a blanket of people who believe and think the same way that you do and who can help you as you journey though life.  As my beliefs changed I tended to shy away from the church because it no longer fit with my life.  As my beliefs continue to change perhaps I'll find a religion and a church that fit me.  Religion is just a set of beliefs, that's all.  It's important to look past the term religion and the names of said religions.  Those Baptists who picket funerals are not the same Baptists that I've met.  Those Muslims who terrorize others are not the same Muslims who practice peace.  They really are their own separate religions. [NEWLINE] [NEWLINE] My point is, that it's not religion that drives people to be evil.  But its evil people who drive religion to perpetuate their evil.</s>
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1998: Karsten Braasch vs. the Williams sisters[edit] [ENDQ] A fourth event dubbed a "Battle of the Sexes" took place during the 1998 Australian Open[29] between Karsten Braasch and the Williams sisters. Venus and Serena Williams, aged 17 and 16 respectively, had claimed that they could beat any male player ranked below 200,<mask> Braasch, then ranked 203rd, challenged them both. The matches took place on court number 12 in Melbourne Park.[30] Braasch first took on Serena and beat her 6–1. He then played Venus and won 6–2.[31] Braasch said afterwards, "500 and above, no chance." He added that he had played like someone ranked 600th in order to keep the game "fun."[32] [NEWLINE] [NEWLINE] [-Wikipedia]( [URL] \(tennis\)#1998:_Karsten_Braasch_vs._the_Williams_sisters) [NEWLINE] [NEWLINE] Men and women are equal,<mask> they are different. Men are better suited to physical athleticism than women. [NEWLINE] [NEWLINE] <mask> people want full equality, pay equality and everything like that across genders, then it is simple, eliminate segregation. Remove men and women's olympics or sports, and just have one single group for both sexes. The result will be that all the women will be paid a lot less,<mask> they won't even make the farm teams. [NEWLINE] [NEWLINE] I mean, merit is merit. I usually don't have much interest in women's sports,<mask> a lot of the time they are just not<mask> entertaining to watch. It's not sexism, it's just a fact of reality.</s><pad>
Label encoding: <s> [STARTQ] 1998: Karsten Braasch vs. the Williams sisters[edit] [ENDQ] A fourth event dubbed a "Battle of the Sexes" took place during the 1998 Australian Open[29] between Karsten Braasch and the Williams sisters. Venus and Serena Williams, aged 17 and 16 respectively, had claimed that they could beat any male player ranked below 200, so Braasch, then ranked 203rd, challenged them both. The matches took place on court number 12 in Melbourne Park.[30] Braasch first took on Serena and beat her 6–1. He then played Venus and won 6–2.[31] Braasch said afterwards, "500 and above, no chance." He added that he had played like someone ranked 600th in order to keep the game "fun."[32] [NEWLINE] [NEWLINE] [-Wikipedia]( [URL] \(tennis\)#1998:_Karsten_Braasch_vs._the_Williams_sisters) [NEWLINE] [NEWLINE] Men and women are equal, but they are different. Men are better suited to physical athleticism than women. [NEWLINE] [NEWLINE] If people want full equality, pay equality and everything like that across genders, then it is simple, eliminate segregation. Remove men and women's olympics or sports, and just have one single group for both sexes. The result will be that all the women will be paid a lot less, because they won't even make the farm teams. [NEWLINE] [NEWLINE] I mean, merit is merit. I usually don't have much interest in women's sports, because a lot of the time they are just not as entertaining to watch. It's not sexism, it's just a fact of reality.</s><pad>
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, I'm not debating wether PCs or Macs are cheaper. I'm not saying Macs are unconditionally better computers. My argument is that I'm getting<mask> I pay for with a mac (i.e. not paying for the apple logo on the front). I'm paying (about) $800 for an $800 screen. Nothing wrong with that. [NEWLINE] [NEWLINE] [STARTQ] will make a lot of icons and menus by default very small [ENDQ] [NEWLINE] Not an issue on a mac. [NEWLINE] [NEWLINE] [STARTQ] videos do not run higher than 1920 × 1080 [ENDQ] [NEWLINE] Are you saying I won't be able to find high resolution videos? Ever heard of 4K? Higher resolution videos are surprisingly wildly available now, and will only become more<mask> in the future. [NEWLINE] [NEWLINE] [STARTQ] A MUCH cheaper option is to buy a laptop with a lower resolution screen and purchase one or two 24" 1920x1080p screens (~$400) giving you a TON more real estate to use from a productivity standpoint and more importantly keeping the screen size large enough<mask> that you can keep using it without a ton of eye strain. [ENDQ] [NEWLINE] A couple major selling points of the display are that A)It is above 300ppi meaning that you cannot see individual pixels. B)THE DISPLAY IS BUILT INTO THE COMPUTER. I don't have a shitty display<mask> I go anywhere<mask> my home/workstation. [NEWLINE] [NEWLINE] Eyestrain is not an issue. [NEWLINE] [NEWLINE] <mask>,<mask> i'm paying $400 on top of maybe $200-300 for the built-in display, i'm spending nearly<mask> much<mask> I would be anyway, without either of the benefits mentioned above. </s>
Label encoding: <s>First off, I'm not debating wether PCs or Macs are cheaper. I'm not saying Macs are unconditionally better computers. My argument is that I'm getting what I pay for with a mac (i.e. not paying for the apple logo on the front). I'm paying (about) $800 for an $800 screen. Nothing wrong with that. [NEWLINE] [NEWLINE] [STARTQ] will make a lot of icons and menus by default very small [ENDQ] [NEWLINE] Not an issue on a mac. [NEWLINE] [NEWLINE] [STARTQ] videos do not run higher than 1920 × 1080 [ENDQ] [NEWLINE] Are you saying I won't be able to find high resolution videos? Ever heard of 4K? Higher resolution videos are surprisingly wildly available now, and will only become more so in the future. [NEWLINE] [NEWLINE] [STARTQ] A MUCH cheaper option is to buy a laptop with a lower resolution screen and purchase one or two 24" 1920x1080p screens (~$400) giving you a TON more real estate to use from a productivity standpoint and more importantly keeping the screen size large enough so that you can keep using it without a ton of eye strain. [ENDQ] [NEWLINE] A couple major selling points of the display are that A)It is above 300ppi meaning that you cannot see individual pixels. B)THE DISPLAY IS BUILT INTO THE COMPUTER. I don't have a shitty display if I go anywhere but my home/workstation. [NEWLINE] [NEWLINE] Eyestrain is not an issue. [NEWLINE] [NEWLINE] Also, if i'm paying $400 on top of maybe $200-300 for the built-in display, i'm spending nearly as much as I would be anyway, without either of the benefits mentioned above. </s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I remember Bush getting blasted over and over on Reddit for his "Now watch this drive" comment. It seems like a similar situation. It was front page for a long<mask> and was mocked for years. [NEWLINE] [NEWLINE] [Now watch this drive.]( [URL] ) [NEWLINE] [NEWLINE] <mask> it's true that Obama tends to get a pass for this sort of thing from Reddit and the media, it's not true that he's vacationing any more than another president. In a tally of declared vacation days, Bush is far ahead of Obama.<mask> that's misleading. Presidents are never truly off duty. They are still briefed daily on a variety of subjects and are still expected to issue decisions on<mask> to deal with different situations. [NEWLINE] [NEWLINE] <mask><mask> you say that it would have been on the front page of "reddit", are you saying that it would have been highly voted up in a default subreddit? It seems likely<mask> the majority of redditors are pretty young and young people are commonly progressive/liberal/Democrat. It's human nature to point out the mistakes of people you dislike and/or disagree with in an attempt to discredit them. I just don't see<mask> that's relevant. [NEWLINE] [NEWLINE] I don't subscribe to any subreddits that would vote such a thing to the front page,<mask><mask> political bent of the president. Maybe you should consider unsubscribing from subreddits that are politically slanted in a way that you dislike. I did years ago and I enjoy Reddit much more. I still poke my head into some of those subreddits from time to time to see<mask> they're talking about and engage in friendly discourse,<mask> it's much less aggravating to ignore them most of the time. </s>
Label encoding: <s>I remember Bush getting blasted over and over on Reddit for his "Now watch this drive" comment. It seems like a similar situation. It was front page for a long while and was mocked for years. [NEWLINE] [NEWLINE] [Now watch this drive.]( [URL] ) [NEWLINE] [NEWLINE] While it's true that Obama tends to get a pass for this sort of thing from Reddit and the media, it's not true that he's vacationing any more than another president. In a tally of declared vacation days, Bush is far ahead of Obama. But that's misleading. Presidents are never truly off duty. They are still briefed daily on a variety of subjects and are still expected to issue decisions on how to deal with different situations. [NEWLINE] [NEWLINE] But when you say that it would have been on the front page of "reddit", are you saying that it would have been highly voted up in a default subreddit? It seems likely since the majority of redditors are pretty young and young people are commonly progressive/liberal/Democrat. It's human nature to point out the mistakes of people you dislike and/or disagree with in an attempt to discredit them. I just don't see why that's relevant. [NEWLINE] [NEWLINE] I don't subscribe to any subreddits that would vote such a thing to the front page, regardless of political bent of the president. Maybe you should consider unsubscribing from subreddits that are politically slanted in a way that you dislike. I did years ago and I enjoy Reddit much more. I still poke my head into some of those subreddits from time to time to see what they're talking about and engage in friendly discourse, but it's much less aggravating to ignore them most of the time. </s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ISIS would not exist<mask> two things had not happened: [NEWLINE] [NEWLINE] 1 The Iraq war [NEWLINE] 2 The Syrian civil war. [NEWLINE] [NEWLINE] <mask> do these two wars have in common? No it's not that they both involved guys who like kebabs, it's the fact the western countries played major roles in both. [NEWLINE] [NEWLINE] Let's start with Iraq. Saddam Hussein' SECULAR regime was defeated by America. This caused the country to descend into chaos with militants fighting everyone who opposed them.<mask> Iraq were under Saddam Hussein's rule the terrorists would never be able to attain the power that they have now. [NEWLINE] [NEWLINE] Now on to the Syrian civil war. The western countries are backing rebel groups who are fighting Bashar Al-Assad's SECULAR regime. This means weapons poured in from the west and parts of Syria became lawless lands with no government control. [NEWLINE] [NEWLINE] <mask> you see that terrorist groups thrive in places with little to no government control. I'm not saying that Asad and Hussein were great rulers<mask> their repressive regimes kept extremists at bay. [NEWLINE] [NEWLINE] Same happened in Muammar Gaddafi's Libya. Now after his fall (caused by the western countries) the country has turned into a place filled with terrorists due to weak government control [NEWLINE] [NEWLINE] <mask> groups like ISIS survive and thrive is places with weak government control. The Taliban is an example<mask> they capture Afghanistan after it was left in shambles by USSR. [NEWLINE] [NEWLINE] <mask><mask> caused the governments of these Iraq and Syria to lose control over their lands? That's right, it's your friendly western countries, the houses of democracy, freedom and justice.<mask> this doesn't mean western countries are evil. Everyone makes mistakes.</s>
Label encoding: <s>ISIS would not exist if two things had not happened: [NEWLINE] [NEWLINE] 1 The Iraq war [NEWLINE] 2 The Syrian civil war. [NEWLINE] [NEWLINE] What do these two wars have in common? No it's not that they both involved guys who like kebabs, it's the fact the western countries played major roles in both. [NEWLINE] [NEWLINE] Let's start with Iraq. Saddam Hussein' SECULAR regime was defeated by America. This caused the country to descend into chaos with militants fighting everyone who opposed them. If Iraq were under Saddam Hussein's rule the terrorists would never be able to attain the power that they have now. [NEWLINE] [NEWLINE] Now on to the Syrian civil war. The western countries are backing rebel groups who are fighting Bashar Al-Assad's SECULAR regime. This means weapons poured in from the west and parts of Syria became lawless lands with no government control. [NEWLINE] [NEWLINE] So you see that terrorist groups thrive in places with little to no government control. I'm not saying that Asad and Hussein were great rulers but their repressive regimes kept extremists at bay. [NEWLINE] [NEWLINE] Same happened in Muammar Gaddafi's Libya. Now after his fall (caused by the western countries) the country has turned into a place filled with terrorists due to weak government control [NEWLINE] [NEWLINE] Therefore groups like ISIS survive and thrive is places with weak government control. The Taliban is an example as they capture Afghanistan after it was left in shambles by USSR. [NEWLINE] [NEWLINE] So what caused the governments of these Iraq and Syria to lose control over their lands? That's right, it's your friendly western countries, the houses of democracy, freedom and justice. But this doesn't mean western countries are evil. Everyone makes mistakes.</s>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is such a thing<mask> a natural monopoly.<mask> it is true that for most products and services it requires a lot in the way intervention, politics, and social engineering to force a monopoly into existence, that isn't true for all things. [NEWLINE] [NEWLINE] There is such a thing<mask> a natural monopoly, which is defined<mask> a monopoly that exists without a government's intervention. JamesdfStudent did a good job of explaining the basic types. Some things are used<mask> other people use it. There can't be competition<mask> competition defeats the point. Other things, require large stat up costs, and<mask> the number of chances we get to start up are small.<mask> losing a couple to anticompetitive business can put people off the scent for quite some time. [NEWLINE] [NEWLINE] <mask>,<mask> the difference between two burger places and two cable companies? Well, the wire that goes to your house. Who owns that? The owner isn't *just* the one who gets paid,<mask><mask> the guy who paid for it to be put in and is responsible for upkeep and matainence.<mask> there were two competitors they would either have to run multiple wires (which would cost billions and would only allow one additional competitor in the market) or share one set of wires (which lends itself to all kind of backstabbing and "preferential treatment"). The free market failed to resolve an identical issue<mask> it came to water pipes and electrical generation and distribution,<mask> the barriers to entry are simply absurd. [NEWLINE] [NEWLINE] The Government is bungling its handling of that market,<mask> it comes from not defining internet service in the same terms<mask> electrical utilities, not from artificially creating monopolistic situations.</s>
Label encoding: <s>There is such a thing as a natural monopoly. While it is true that for most products and services it requires a lot in the way intervention, politics, and social engineering to force a monopoly into existence, that isn't true for all things. [NEWLINE] [NEWLINE] There is such a thing as a natural monopoly, which is defined as a monopoly that exists without a government's intervention. JamesdfStudent did a good job of explaining the basic types. Some things are used because other people use it. There can't be competition because competition defeats the point. Other things, require large stat up costs, and so the number of chances we get to start up are small. So losing a couple to anticompetitive business can put people off the scent for quite some time. [NEWLINE] [NEWLINE] So, why the difference between two burger places and two cable companies? Well, the wire that goes to your house. Who owns that? The owner isn't *just* the one who gets paid, but also the guy who paid for it to be put in and is responsible for upkeep and matainence. If there were two competitors they would either have to run multiple wires (which would cost billions and would only allow one additional competitor in the market) or share one set of wires (which lends itself to all kind of backstabbing and "preferential treatment"). The free market failed to resolve an identical issue when it came to water pipes and electrical generation and distribution, because the barriers to entry are simply absurd. [NEWLINE] [NEWLINE] The Government is bungling its handling of that market, but it comes from not defining internet service in the same terms as electrical utilities, not from artificially creating monopolistic situations.</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's interesting. Your "have the kid and put it up for adoption" and our discussion about having higher IQ people breed more is getting deliciously eugenicist.<mask> perhaps not the most moral idea considering the thousands of kids waiting to be adopted already, it's certainly food for thought. [NEWLINE] [NEWLINE] About the "not knowing<mask> much you can handle" thing - surely it's better to think that you can't handle a child and decide not to have one, then to have a child and not have the appropriate means to provide for them? Many would<mask><mask> it's selfish to bring up a child in an impoverished situation,<mask> it seems that this "selfish" word pretty much means anything that isn't driven towards the betterment of humanity. Which, naturally, is very hard to gauge on a micro-scale. [NEWLINE] [NEWLINE] ∆ [NEWLINE] [NEWLINE] I award you a delta<mask> you definitely made me think about my position more, and that<mask> on an individual case-by-case basis it's not necessarily bad, the fact that it's becoming a massive trend among those with more ability and means is a very troubling thing<mask>. It's not even just among our society, all over the world it's shown that the more wealthy and educated people get, the less children they have. I guess all I can really hope is that the "best" of those among us suddenly get real maternal/paternal, them suddenly wanting children is the only really moral/ethical way around this situation. [NEWLINE] [NEWLINE] You haven't convinced me to suddenly get pregnant,<mask> I'll definitely be more supportive towards those around me who decide to go down the family route. :)</s>
Label encoding: <s>That's interesting. Your "have the kid and put it up for adoption" and our discussion about having higher IQ people breed more is getting deliciously eugenicist. While perhaps not the most moral idea considering the thousands of kids waiting to be adopted already, it's certainly food for thought. [NEWLINE] [NEWLINE] About the "not knowing how much you can handle" thing - surely it's better to think that you can't handle a child and decide not to have one, then to have a child and not have the appropriate means to provide for them? Many would argue that it's selfish to bring up a child in an impoverished situation, but it seems that this "selfish" word pretty much means anything that isn't driven towards the betterment of humanity. Which, naturally, is very hard to gauge on a micro-scale. [NEWLINE] [NEWLINE] ∆ [NEWLINE] [NEWLINE] I award you a delta because you definitely made me think about my position more, and that while on an individual case-by-case basis it's not necessarily bad, the fact that it's becoming a massive trend among those with more ability and means is a very troubling thing indeed. It's not even just among our society, all over the world it's shown that the more wealthy and educated people get, the less children they have. I guess all I can really hope is that the "best" of those among us suddenly get real maternal/paternal, them suddenly wanting children is the only really moral/ethical way around this situation. [NEWLINE] [NEWLINE] You haven't convinced me to suddenly get pregnant, but I'll definitely be more supportive towards those around me who decide to go down the family route. :)</s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL] /[1] [NEWLINE] [NEWLINE] The only reason this video is on the front page is<mask> men feel truly oppressed by<mask> they perceive to be a double standard that says they can't hit women even<mask> they're being attacked.<mask> you doubt that this video is purely about gender roles, ask yourself<mask> it would be on the front page<mask> all the parties involved were male. [NEWLINE] [NEWLINE] By feeling oppressed by this dynamic,<mask><mask> most men are fundamentally failing to understand~~ a)<mask> much more common it is for men to be violent towards women,~~ b) the culture of fear and intimidation that such violence breeds, and c)<mask> ridiculous it makes you look to cheer and go "Right on!" in the relatively rare case<mask> a man is justified in hitting a woman. [NEWLINE] [NEWLINE] Anyone remember the last time a story about a woman defending herself against a violent man was on the front page? Me neither. [NEWLINE] [NEWLINE] EDIT: A few people have accurately pointed out that the frequency of cases of women perpetrating violence against men is roughly the same<mask> men perpetrating violence against women. Others have<mask> pointed out that<mask> makes violence against women more pernicious is the greater damage that men are able to inflict. In making the initial point, I was attempting to demonstrate that violence is a way of asserting power or control, and in the case of men committing violence against women, that power dynamic is shifted heavily in their favor. The overall point about men equating violence against women with violence against men stands -- it is not a valid comparison, and most of the men making it seem to be fundamentally failing to understand<mask> those power dynamics make the two situations quite different.</s>
Label encoding: <s> [URL] /[1] [NEWLINE] [NEWLINE] The only reason this video is on the front page is because men feel truly oppressed by what they perceive to be a double standard that says they can't hit women even if they're being attacked. If you doubt that this video is purely about gender roles, ask yourself if it would be on the front page if all the parties involved were male. [NEWLINE] [NEWLINE] By feeling oppressed by this dynamic, I think most men are fundamentally failing to understand~~ a) how much more common it is for men to be violent towards women,~~ b) the culture of fear and intimidation that such violence breeds, and c) how ridiculous it makes you look to cheer and go "Right on!" in the relatively rare case where a man is justified in hitting a woman. [NEWLINE] [NEWLINE] Anyone remember the last time a story about a woman defending herself against a violent man was on the front page? Me neither. [NEWLINE] [NEWLINE] EDIT: A few people have accurately pointed out that the frequency of cases of women perpetrating violence against men is roughly the same as men perpetrating violence against women. Others have also pointed out that what makes violence against women more pernicious is the greater damage that men are able to inflict. In making the initial point, I was attempting to demonstrate that violence is a way of asserting power or control, and in the case of men committing violence against women, that power dynamic is shifted heavily in their favor. The overall point about men equating violence against women with violence against men stands -- it is not a valid comparison, and most of the men making it seem to be fundamentally failing to understand how those power dynamics make the two situations quite different.</s>
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sure. I mean,<mask> you take a reductionist view far enough, this line of thinking becomes pretty uncontroversial,<mask><mask> not interesting in the way you seem to think. At some point, its all just biology, chemistry and physics. Forget "self-interest".<mask> does that even mean? The "real" reason I do<mask> I do is<mask> of electric potentials in my brain causing neurons to fire!<mask> it would be bizarre to call a neuron "selfish" or "self-interested". You've called the realization that you summarized above about self interest<mask> "depressing" in other posts here.<mask><mask>? You've literally defined your terms such that "non-self interest" no longer exists *logically* outside of accidents or randomization. There's no reason for your view to be an unsettling revelation about the world. Its not even a revelation about the world at all. Its just a change in<mask> you're using language.<mask><mask> doing this, you've smuggled negative connotations about "acting in self interest" into a level of abstraction<mask> it no longer makes any sense (again, does it make any sense to call neurons self-interested?). [NEWLINE] [NEWLINE] These are words that literally only make sense<mask> you're dealing with a sufficiently high level of abstraction. And at that level (the level of talking about humans having "motivations"), I don't think your view really makes sense. It makes more sense to think of virtuous / selfless people<mask> those whose motivations and pleasure centers align with the well being of others. These are people that exist, and we should continue to encourage such behavior. Nothing depressing about that!</s>
Label encoding: <s>Sure. I mean, if you take a reductionist view far enough, this line of thinking becomes pretty uncontroversial, but also not interesting in the way you seem to think. At some point, its all just biology, chemistry and physics. Forget "self-interest". What does that even mean? The "real" reason I do what I do is because of electric potentials in my brain causing neurons to fire! But it would be bizarre to call a neuron "selfish" or "self-interested". You've called the realization that you summarized above about self interest as "depressing" in other posts here. But why? You've literally defined your terms such that "non-self interest" no longer exists *logically* outside of accidents or randomization. There's no reason for your view to be an unsettling revelation about the world. Its not even a revelation about the world at all. Its just a change in how you're using language. But while doing this, you've smuggled negative connotations about "acting in self interest" into a level of abstraction where it no longer makes any sense (again, does it make any sense to call neurons self-interested?). [NEWLINE] [NEWLINE] These are words that literally only make sense if you're dealing with a sufficiently high level of abstraction. And at that level (the level of talking about humans having "motivations"), I don't think your view really makes sense. It makes more sense to think of virtuous / selfless people as those whose motivations and pleasure centers align with the well being of others. These are people that exist, and we should continue to encourage such behavior. Nothing depressing about that!</s>
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a difference between *xie*, which has been in usage<mask> the 1990s and was coined specifically to sound like *he* and *she* (sure, it's got an X in it,<mask> it's short, rhymes with the existing gendered pronouns, and isn't that hard to say; it's pronounced *zee*) and *bun*, which was coined two months ago by a special-snowflake Tumblr fourteen-year-old who's probably going to feel ashamed of themself in a year's time and which is not a pronoun,<mask> a repurposed noun. [NEWLINE] [NEWLINE] It's<mask> pretty telling that you put *nonbinary* in quotation marks. Gender dysphoria is a documented scientific thing, and recent studies have shown that dysphoria can have a target body outside of the standard male-female dichotomy, and that it can even fluctuate. Nonbinary genders have existed for millennia and are valid, even<mask> some people use them<mask> an excuse to be idiots on the internet. People will use *anything*<mask> an excuse to be an idiot on the internet. Deal with it. [NEWLINE] [NEWLINE] You say that *they* exists, which is all well and good<mask> there are plenty of people out there who bitch and moan about using *they*<mask> a singular. It's valid<mask> a singular, of course,<mask> that doesn't stop the grammar purists from whining. Which puts nonbinary people in an awkward situation: Either they request *they* and get bitched at by grammar nazis, or they request something like *xie* and get bitched at by people like you.</s>
Label encoding: <s>There's a difference between *xie*, which has been in usage since the 1990s and was coined specifically to sound like *he* and *she* (sure, it's got an X in it, but it's short, rhymes with the existing gendered pronouns, and isn't that hard to say; it's pronounced *zee*) and *bun*, which was coined two months ago by a special-snowflake Tumblr fourteen-year-old who's probably going to feel ashamed of themself in a year's time and which is not a pronoun, but a repurposed noun. [NEWLINE] [NEWLINE] It's also pretty telling that you put *nonbinary* in quotation marks. Gender dysphoria is a documented scientific thing, and recent studies have shown that dysphoria can have a target body outside of the standard male-female dichotomy, and that it can even fluctuate. Nonbinary genders have existed for millennia and are valid, even if some people use them as an excuse to be idiots on the internet. People will use *anything* as an excuse to be an idiot on the internet. Deal with it. [NEWLINE] [NEWLINE] You say that *they* exists, which is all well and good but there are plenty of people out there who bitch and moan about using *they* as a singular. It's valid as a singular, of course, but that doesn't stop the grammar purists from whining. Which puts nonbinary people in an awkward situation: Either they request *they* and get bitched at by grammar nazis, or they request something like *xie* and get bitched at by people like you.</s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We are often told today that Islam is incomparable with democracy; this is the other side of the early 20th century coin which reads "Jews are evil". That bigotry led to one of the greatest crimes in history. We should be very sure that our new bigotry on anti-semitism's reverse face does not lead to the same thing. It has that potential. This is<mask> I am very, very scared of Islamophobia. [NEWLINE] [NEWLINE] In my view, lying is the root of all evil. Perhaps you are not lying,<mask> you have been lied to. Islam has virtually never been violent,<mask> it has almost never even been political. Look at the 17th century; in Europe, Christians are tearing each other apart in the greater religious war in history.<mask> do the persecuted Jews flee for safety? The tolerant, peaceful Middle East. There was not one single internal conflict in the Islamic world until the First World War (aside from the borders; the Ottoman Empire was still an Empire after all). [NEWLINE] [NEWLINE] <mask> it's silly to say that Islam is a violent religion. It's not remotely backed up by history. You'd have a hell of a lot more grounds for saying Christianity is a violent religion;<mask>, Christianity has been an extremely violent, political belief everywhere it exists. Political Islam, by contrast, is about 30 years old. In Afghanistan, for instance, the public turned to communism<mask> their natural politics, not Islam. Popular communism continued until it was overthrown by the then-US-backed Taliban in about 1995. [NEWLINE] [NEWLINE] Something has CHANGED<mask> the 17th century, around about 1980. Now<mask> could that be?</s><pad>
Label encoding: <s>We are often told today that Islam is incomparable with democracy; this is the other side of the early 20th century coin which reads "Jews are evil". That bigotry led to one of the greatest crimes in history. We should be very sure that our new bigotry on anti-semitism's reverse face does not lead to the same thing. It has that potential. This is why I am very, very scared of Islamophobia. [NEWLINE] [NEWLINE] In my view, lying is the root of all evil. Perhaps you are not lying, but you have been lied to. Islam has virtually never been violent, indeed it has almost never even been political. Look at the 17th century; in Europe, Christians are tearing each other apart in the greater religious war in history. Where do the persecuted Jews flee for safety? The tolerant, peaceful Middle East. There was not one single internal conflict in the Islamic world until the First World War (aside from the borders; the Ottoman Empire was still an Empire after all). [NEWLINE] [NEWLINE] So it's silly to say that Islam is a violent religion. It's not remotely backed up by history. You'd have a hell of a lot more grounds for saying Christianity is a violent religion; indeed, Christianity has been an extremely violent, political belief everywhere it exists. Political Islam, by contrast, is about 30 years old. In Afghanistan, for instance, the public turned to communism as their natural politics, not Islam. Popular communism continued until it was overthrown by the then-US-backed Taliban in about 1995. [NEWLINE] [NEWLINE] Something has CHANGED since the 17th century, around about 1980. Now what could that be?</s><pad>
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most contemporary psychological and medical professionals would not disagree with the definition of mental illness. <mask> they would disagree with is the idea that the current means of addressing the disorder (hormone therapies) is "enabling" the disorder instead of treating it.  The problem to be addressed is the fact that the person is appalled by their own physiology, not the fact that they would rather have been born the other sex.  Hormone therapy addresses that problem and is<mask> far the only reliable treatment. [NEWLINE] [NEWLINE] [STARTQ] And<mask> society starts treating transgender people<mask> having no mental issue, and accepting invasive surgery<mask> the standard treatment then that will slow research towards less drastic treatments. [ENDQ] [NEWLINE] I would argue the opposite.  Gender identity disorder, gender dysphoria, the transgender experience and the effects of medical transition are all still subjects of active research.  The difference now is that visibility of transgender issues allow these researchers to get more funding and visibility than they could before,<mask> next to nobody knew<mask> a "tranny" actually was. [NEWLINE] [NEWLINE] [STARTQ] Ideally, in the future,<mask> someone were to come into a doctor's office and say "I feel<mask> bad in my current body that I want hormonal treatment and invasive surgery" the doctor would be able to prescribe something that would just make the transgender person no longer feel terrible in their current body. [ENDQ] [NEWLINE] Again, that's an ideal future, and both transgender people and research regarding them exists in present reality.  Even<mask>, I don't see<mask> societal acceptance of transgender people would at all obstruct research into some treatment that somehow absolved a person of any and all gender dysphoria. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Most contemporary psychological and medical professionals would not disagree with the definition of mental illness.  What they would disagree with is the idea that the current means of addressing the disorder (hormone therapies) is "enabling" the disorder instead of treating it.  The problem to be addressed is the fact that the person is appalled by their own physiology, not the fact that they would rather have been born the other sex.  Hormone therapy addresses that problem and is so far the only reliable treatment. [NEWLINE] [NEWLINE] [STARTQ] And if society starts treating transgender people as having no mental issue, and accepting invasive surgery as the standard treatment then that will slow research towards less drastic treatments. [ENDQ] [NEWLINE] I would argue the opposite.  Gender identity disorder, gender dysphoria, the transgender experience and the effects of medical transition are all still subjects of active research.  The difference now is that visibility of transgender issues allow these researchers to get more funding and visibility than they could before, when next to nobody knew what a "tranny" actually was. [NEWLINE] [NEWLINE] [STARTQ] Ideally, in the future, if someone were to come into a doctor's office and say "I feel so bad in my current body that I want hormonal treatment and invasive surgery" the doctor would be able to prescribe something that would just make the transgender person no longer feel terrible in their current body. [ENDQ] [NEWLINE] Again, that's an ideal future, and both transgender people and research regarding them exists in present reality.  Even so, I don't see how societal acceptance of transgender people would at all obstruct research into some treatment that somehow absolved a person of any and all gender dysphoria. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a few problems with this. [NEWLINE] [NEWLINE] <mask><mask><mask>, to get out of a claim, you are postulating one which requires just<mask> much evidence.<mask> a skeptic supposedly,<mask> are you<mask> willing to accept the notion of infinite universes with<mask> little evidence,<mask><mask> it comes to God, you want solid evidence? [NEWLINE] [NEWLINE] Second of all, you're violating Occam's Razor, which says "don't multiply entities beyond necessity." You have invoked an infinite amount of entities to account for a phenomenon, whereas only one is needed<mask> an explanation. [NEWLINE] [NEWLINE] Third of all, an *infinite* amount of universes is extremely problematic meta-logically.<mask> you literally have an endless amount of matter, anything can happen. This means any arrangement of matter in any shape at any time, all the time, will arise constantly. This includes spontaneously generated brains that are pre-programmed with thoughts<mask> don't actually exist in relation to "reality." The probability of your brain arising<mask> it is with all your thoughts and memories intact, with a chemical sequence that will play-out for you to experience reality in the causal order that you do is infinitesimally small.<mask> in an infinite amount of universes, it will happen an infinite amount of times, all the time.<mask> you have thrown reason under the rug, and infinite universes are no different than solipsism,<mask> all of reality could easily be a big illusion. That is<mask> an *infinite* amount of universes is problematic in a logical argument. [NEWLINE] [NEWLINE] <mask>, the number infinity is problematic in reality and is much more philosophical than doable.</s>
Label encoding: <s>I have a few problems with this. [NEWLINE] [NEWLINE] First of all, to get out of a claim, you are postulating one which requires just as much evidence. As a skeptic supposedly, why are you so willing to accept the notion of infinite universes with so little evidence, but when it comes to God, you want solid evidence? [NEWLINE] [NEWLINE] Second of all, you're violating Occam's Razor, which says "don't multiply entities beyond necessity." You have invoked an infinite amount of entities to account for a phenomenon, whereas only one is needed as an explanation. [NEWLINE] [NEWLINE] Third of all, an *infinite* amount of universes is extremely problematic meta-logically. When you literally have an endless amount of matter, anything can happen. This means any arrangement of matter in any shape at any time, all the time, will arise constantly. This includes spontaneously generated brains that are pre-programmed with thoughts but don't actually exist in relation to "reality." The probability of your brain arising as it is with all your thoughts and memories intact, with a chemical sequence that will play-out for you to experience reality in the causal order that you do is infinitesimally small. But in an infinite amount of universes, it will happen an infinite amount of times, all the time. So you have thrown reason under the rug, and infinite universes are no different than solipsism, because all of reality could easily be a big illusion. That is why an *infinite* amount of universes is problematic in a logical argument. [NEWLINE] [NEWLINE] Lastly, the number infinity is problematic in reality and is much more philosophical than doable.</s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not going to challenge your view that networking is unfair,<mask> it totally is.<mask> I am going to challenge your view that it should be discouraged or eliminated. [NEWLINE] [NEWLINE] Businesses are and should be designed to be effective. Being fair is, at best, a secondary goal. Any company/society which encourages business practices that are not effective will quickly be out-competed by one that is effective, and having effective businesses in a society in general improves the quality of life of the people living there. There are some exceptions, that justify regulation, such<mask> environmental or fraud protection,<mask> I don't think networking is one of them, the reason being just<mask> expensive it is to find employees without using networking. Simply getting an employee in the door often costs about 1 years worth of their salary, with that money going to the recruiters that found the person, paying people to review their resume, interview them, do background checks, etc. Then, after they start working, they have to be trained, and get up to speed on their new job, which (depending on the job) can often take months.<mask> all of this time and money goes towards a bad hire, that is an extremely bad thing for the business, which means that it's extremely important to minimize the number of bad hires. The best way to do that is to hire people you've worked with before, and you know are good, or people your colleagues and friends have worked with before, who you trust<mask> they say that someone is good. Interviews are an extremely poor replacement actually working with someone, or for the opinion of a trusted colleague.</s>
Label encoding: <s>I'm not going to challenge your view that networking is unfair, because it totally is. But I am going to challenge your view that it should be discouraged or eliminated. [NEWLINE] [NEWLINE] Businesses are and should be designed to be effective. Being fair is, at best, a secondary goal. Any company/society which encourages business practices that are not effective will quickly be out-competed by one that is effective, and having effective businesses in a society in general improves the quality of life of the people living there. There are some exceptions, that justify regulation, such as environmental or fraud protection, but I don't think networking is one of them, the reason being just how expensive it is to find employees without using networking. Simply getting an employee in the door often costs about 1 years worth of their salary, with that money going to the recruiters that found the person, paying people to review their resume, interview them, do background checks, etc. Then, after they start working, they have to be trained, and get up to speed on their new job, which (depending on the job) can often take months. If all of this time and money goes towards a bad hire, that is an extremely bad thing for the business, which means that it's extremely important to minimize the number of bad hires. The best way to do that is to hire people you've worked with before, and you know are good, or people your colleagues and friends have worked with before, who you trust when they say that someone is good. Interviews are an extremely poor replacement actually working with someone, or for the opinion of a trusted colleague.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Being someone that prefers tall women (being short myself), having dated fat, medium and thin, promiscuous, virginal and in between I have a rather strong opinion in the matter. [NEWLINE] [NEWLINE] Preferences like that are rather shallow.  Extremes can be understandable<mask> in most topics (like having a hard time being attracted to a dwarf or an obese person),<mask><mask> you decide beforehand that anyone with more than X sexual partners is undesirable is shallow, more shallow than preferring someone with a certain physical attribute (or without another one). [NEWLINE] Sexual experience is something that does not impact your perception of the person, like skin colour, size, fatness, eye colour, etc.  Even personality, style, education and some values can be picked up rather quick and have an effect on someone's attractiveness. <mask> promiscuity is affecting something visible then your problem is not promiscuity<mask> the effects of it.  Maybe he/she flirts too much, or is too sexually demanding for your taste, or has some psychological damage, or maybe they talk too much about previous partners. <mask> having any preference to the *number* of partners (unless it's an extreme like 3 gang bangs per week for the past 3 years) is definitely a poor judgment parameter compared to others that affect your senses directly and<mask> your experience with the person.  Making a big deal out of this talks more about your insecurities, desire of ownership, mental torment and probably other issues. [NEWLINE] [NEWLINE] <mask>, public shaming of sexual experience only encourages sexual experience to be hidden and increases the taboo, leading to less healthy sex lives.  </s><pad>
Label encoding: <s>Being someone that prefers tall women (being short myself), having dated fat, medium and thin, promiscuous, virginal and in between I have a rather strong opinion in the matter. [NEWLINE] [NEWLINE] Preferences like that are rather shallow.  Extremes can be understandable as in most topics (like having a hard time being attracted to a dwarf or an obese person), but if you decide beforehand that anyone with more than X sexual partners is undesirable is shallow, more shallow than preferring someone with a certain physical attribute (or without another one). [NEWLINE] Sexual experience is something that does not impact your perception of the person, like skin colour, size, fatness, eye colour, etc.  Even personality, style, education and some values can be picked up rather quick and have an effect on someone's attractiveness.  If promiscuity is affecting something visible then your problem is not promiscuity but the effects of it.  Maybe he/she flirts too much, or is too sexually demanding for your taste, or has some psychological damage, or maybe they talk too much about previous partners.  However having any preference to the *number* of partners (unless it's an extreme like 3 gang bangs per week for the past 3 years) is definitely a poor judgment parameter compared to others that affect your senses directly and therefore your experience with the person.  Making a big deal out of this talks more about your insecurities, desire of ownership, mental torment and probably other issues. [NEWLINE] [NEWLINE] Also, public shaming of sexual experience only encourages sexual experience to be hidden and increases the taboo, leading to less healthy sex lives.  </s><pad>
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The Bible is a complex canon of letters written to the early churches. We can't just take one verse from here and one verse from there and form out beliefs on it. This leads to some pretty bad stuff. [NEWLINE] [NEWLINE] <mask><mask><mask> homosexuality goes, I'll take a more philisophical approach rather than specific verses,<mask> you could do a Google search on all of the various interpretations of the verses you know about. [NEWLINE] [NEWLINE] [STARTQ] My real question, then, is is homosexuality sin at all? [ENDQ] [NEWLINE] I will<mask><mask> it isn't. We see examples of sin throughout the Bible. It is always destructive to someone. Bearing false witness ruins someone's reputation, cheating on your spouse ruins your relationship, murdering someone doesn't need explanation. All sin follows some sort of a pattern. Homosexuality doesn't follow this pattern. It wouldn't make sense for God to have this be a sin<mask> it didn't follow in lines with the others. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Does it displease the Creator? [ENDQ] [NEWLINE] I don't think<mask>. We are here to love others and to bring Glory to God. I don't see<mask> someone who is in a homosexual relationship is incapable of doing those things. [NEWLINE] [NEWLINE] [STARTQ] Bonus points: Is there any defense for homosexuality being okay with God that doesn't abolish the institution of monogamy, or cause promiscuity to be okay? [ENDQ] [NEWLINE] Homosexuals and heterosexuals are held to the same relationship standards. Keep one spouse, serve each other in humility and serve and glorify God with your lives and marriage. Promiscuity is bad<mask> it damages your relationship with your future spouse.</s>
Label encoding: <s>The Bible is a complex canon of letters written to the early churches. We can't just take one verse from here and one verse from there and form out beliefs on it. This leads to some pretty bad stuff. [NEWLINE] [NEWLINE] As far as homosexuality goes, I'll take a more philisophical approach rather than specific verses, since you could do a Google search on all of the various interpretations of the verses you know about. [NEWLINE] [NEWLINE] [STARTQ] My real question, then, is is homosexuality sin at all? [ENDQ] [NEWLINE] I will argue that it isn't. We see examples of sin throughout the Bible. It is always destructive to someone. Bearing false witness ruins someone's reputation, cheating on your spouse ruins your relationship, murdering someone doesn't need explanation. All sin follows some sort of a pattern. Homosexuality doesn't follow this pattern. It wouldn't make sense for God to have this be a sin if it didn't follow in lines with the others. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Does it displease the Creator? [ENDQ] [NEWLINE] I don't think so. We are here to love others and to bring Glory to God. I don't see how someone who is in a homosexual relationship is incapable of doing those things. [NEWLINE] [NEWLINE] [STARTQ] Bonus points: Is there any defense for homosexuality being okay with God that doesn't abolish the institution of monogamy, or cause promiscuity to be okay? [ENDQ] [NEWLINE] Homosexuals and heterosexuals are held to the same relationship standards. Keep one spouse, serve each other in humility and serve and glorify God with your lives and marriage. Promiscuity is bad as it damages your relationship with your future spouse.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Honestly,<mask> he took isn't the important part. It wasn't shoplifting, it was robbery.<mask><mask> wikipedia, "Robbery is the crime of taking or attempting to take something of value by force or threat of force or by putting the victim in fear."* [ENDQ] [NEWLINE] Taking 10 dollars worth of cigars is vastly different than taking 1000 dollars worth of DVDs. It does matter<mask> he stole. Taking 10 dollars worth of something is petty, it's essentially meaningless. [NEWLINE] [NEWLINE] [STARTQ] I believe this is highly relevant<mask> he got into a physical confrontation moments later with a law enforcement officer. [ENDQ] [NEWLINE] <mask><mask>, it is a completely different situation. He knows he can't intimidate a police officer with his size. The little 5'5" convience store owner? Yeah. A cop with a taser and a gun? Not a chance. [NEWLINE] [NEWLINE] [STARTQ] <mask> you could show references to those statistics, I would like to see them. (Not accusing you of making it up,<mask> I really am interested in seeing) [ENDQ] [NEWLINE] It was on the [second page]( [URL].2.html) of the link I gave you. I misquoted and said 90%<mask> it was only 86%. It's<mask> 92% of all searches. [NEWLINE] [NEWLINE] [STARTQ] That being said, here are some interesting snippets from the article: [ENDQ] [NEWLINE] <mask>,<mask> that article was saying was: Well, Ferguson isn't any worse than the rest of St. Louis;<mask> there's nothing to see here. Move along. [NEWLINE] [NEWLINE] <mask><mask> that just proves<mask> fucked up STL is in regards to race and the police...</s>
Label encoding: <s> [STARTQ] Honestly, what he took isn't the important part. It wasn't shoplifting, it was robbery. According to wikipedia, "Robbery is the crime of taking or attempting to take something of value by force or threat of force or by putting the victim in fear."* [ENDQ] [NEWLINE] Taking 10 dollars worth of cigars is vastly different than taking 1000 dollars worth of DVDs. It does matter what he stole. Taking 10 dollars worth of something is petty, it's essentially meaningless. [NEWLINE] [NEWLINE] [STARTQ] I believe this is highly relevant since he got into a physical confrontation moments later with a law enforcement officer. [ENDQ] [NEWLINE] I disagree, it is a completely different situation. He knows he can't intimidate a police officer with his size. The little 5'5" convience store owner? Yeah. A cop with a taser and a gun? Not a chance. [NEWLINE] [NEWLINE] [STARTQ] If you could show references to those statistics, I would like to see them. (Not accusing you of making it up, but I really am interested in seeing) [ENDQ] [NEWLINE] It was on the [second page]( [URL].2.html) of the link I gave you. I misquoted and said 90% when it was only 86%. It's also 92% of all searches. [NEWLINE] [NEWLINE] [STARTQ] That being said, here are some interesting snippets from the article: [ENDQ] [NEWLINE] So, what that article was saying was: Well, Ferguson isn't any worse than the rest of St. Louis; so there's nothing to see here. Move along. [NEWLINE] [NEWLINE] I think that just proves how fucked up STL is in regards to race and the police...</s>
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are many kinds of "strong atheist"--meaning someone who states that there is no god, more than merely having no belief in god--<mask> they have good reason. [NEWLINE] [NEWLINE] One of the first is the abundant evidence that humans make up gods and mythologies all the time. There are thousands of them, and tens of thousands of variants of them. A Christian organization once conducted a census that identified over 40,000 denominations of Christianity alone, each with its own interpretation of scripture. Many religions are deeply incompatible with the others.<mask> <mask> there really was a god sending prophets to Earth, we are forced into two conclusions: [NEWLINE] [NEWLINE] 1. God is incompetent, in which case we have to question His ability to create the universe in the first place. [NEWLINE] [NEWLINE] 2. God is intent on sewing confusion and strife that leads to conflict, death and misery. [NEWLINE] [NEWLINE] We don't consider it unreasonable to say that there is no tooth fairy,<mask><mask> you can't prove it. It is much more harmful to the healthy mental development of children and adults to insist that the tooth fairy's existence ought to be taken seriously just<mask> there is no proof of the negative. [NEWLINE] [NEWLINE] At some point you reach a tipping stage<mask> it becomes *more* arrogant to insist on the existence of something that can't be proven. At some point, you need to fire the airplane mechanic who insists the damage was caused by gremlins,<mask> he's endangering lives. Religion has made a mess of an awful lot of lives, and many atheists see it<mask> a serious problem that people cling to their beliefs *and evangelize them*.</s>
Label encoding: <s>There are many kinds of "strong atheist"--meaning someone who states that there is no god, more than merely having no belief in god-- but they have good reason. [NEWLINE] [NEWLINE] One of the first is the abundant evidence that humans make up gods and mythologies all the time. There are thousands of them, and tens of thousands of variants of them. A Christian organization once conducted a census that identified over 40,000 denominations of Christianity alone, each with its own interpretation of scripture. Many religions are deeply incompatible with the others. So  if there really was a god sending prophets to Earth, we are forced into two conclusions: [NEWLINE] [NEWLINE] 1. God is incompetent, in which case we have to question His ability to create the universe in the first place. [NEWLINE] [NEWLINE] 2. God is intent on sewing confusion and strife that leads to conflict, death and misery. [NEWLINE] [NEWLINE] We don't consider it unreasonable to say that there is no tooth fairy, even though you can't prove it. It is much more harmful to the healthy mental development of children and adults to insist that the tooth fairy's existence ought to be taken seriously just because there is no proof of the negative. [NEWLINE] [NEWLINE] At some point you reach a tipping stage where it becomes *more* arrogant to insist on the existence of something that can't be proven. At some point, you need to fire the airplane mechanic who insists the damage was caused by gremlins, because he's endangering lives. Religion has made a mess of an awful lot of lives, and many atheists see it as a serious problem that people cling to their beliefs *and evangelize them*.</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not particularly well versed in beers per se,<mask> my statement isn't restricted necessarily to craft beers. [NEWLINE] [NEWLINE] You find it upsetting that the general public is stereotyping all craft beers<mask> having to be bitter. <mask><mask>,<mask>, that this could just be a natural progression in the taste of beer.  I'll relate it to something I know better, music. <mask> jazz became popular, classical music buffs were extremely offended that this new music started to disenfranchise<mask> hundreds of years of music theory said is 'correct'.  Then Elvis came along and everyone hated<mask> sexual his music/dancing was.  The same is happening now with rap, people hate it for being too aggressive, not lyrical, etc. [NEWLINE] [NEWLINE] <mask>, I guess the point of all this is that yes, perhaps the general public is starting to view craft beers<mask> more and more bitter.  And yes, maybe those who know a thing or two about beer find this frustrating (like yourself). <mask>,<mask> it's a popular trend, then it won't slow down for any reason other than its own insustainability, and it may well become the norm soon, just<mask> the norm now is for not-<mask> -bitter beers. [NEWLINE] [NEWLINE] I'm not arguing that this shift isn't occurring, God knows I don't know enough to make that claim. Just that there's a different perspective<mask> well. <mask> anything, this is exciting.  The microbrewery movement in the past thirty years is bound to change the scene, and<mask> this is just one of the ways that's happening.</s>
Label encoding: <s>I'm not particularly well versed in beers per se, but my statement isn't restricted necessarily to craft beers. [NEWLINE] [NEWLINE] You find it upsetting that the general public is stereotyping all craft beers as having to be bitter.  I think, though, that this could just be a natural progression in the taste of beer.  I'll relate it to something I know better, music.  When jazz became popular, classical music buffs were extremely offended that this new music started to disenfranchise what hundreds of years of music theory said is 'correct'.  Then Elvis came along and everyone hated how sexual his music/dancing was.  The same is happening now with rap, people hate it for being too aggressive, not lyrical, etc. [NEWLINE] [NEWLINE] So, I guess the point of all this is that yes, perhaps the general public is starting to view craft beers as more and more bitter.  And yes, maybe those who know a thing or two about beer find this frustrating (like yourself).  But, if it's a popular trend, then it won't slow down for any reason other than its own insustainability, and it may well become the norm soon, just as the norm now is for not- as -bitter beers. [NEWLINE] [NEWLINE] I'm not arguing that this shift isn't occurring, God knows I don't know enough to make that claim. Just that there's a different perspective as well.  If anything, this is exciting.  The microbrewery movement in the past thirty years is bound to change the scene, and so this is just one of the ways that's happening.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While it is neutral from the point of view of the child to be aborted, it is better for society that the children of the unmarried and poor are aborted. [NEWLINE] [NEWLINE] From [URL] [NEWLINE] [NEWLINE] [STARTQ] <mask> liberals are completely wrong. Poverty is about low IQ, low future-time orientation, and bad values. The first item on the list is all about genes: people are born with low IQ and there’s not much you can do about it. I suspect that low future-time orientation<mask> has a large genetic component,<mask><mask><mask> it’s possible that it can be raised with proper training. Values are, by definition, learned attitudes and behaviors. [ENDQ] [NEWLINE] &gt; The Half Sigma approach to eliminate poverty is a two-pronged attack which I guarantee would achieve good results<mask> it were ever implemented. (1) we need to reduce the number of people with poor-people genes from being born with eugenic policies; and (2) we should focus on teaching middle-class values to poor children. The first half of the Half Sigma approach is not likely to be implemented in my lifetime<mask> eugenics is considered to be the most evil thing in the world. There’s a possibility that we might come around to acknowledging the importance of values,<mask> even that’s not very likely<mask> it touches too close to the truth of HBD. Suggesting that poor people need better values tends to outrage liberals, who say that you are blaming poor people for their poverty<mask> liberals know that poverty is caused by Republicans who are too cheap to spend the money needed to end it.</s>
Label encoding: <s>While it is neutral from the point of view of the child to be aborted, it is better for society that the children of the unmarried and poor are aborted. [NEWLINE] [NEWLINE] From [URL] [NEWLINE] [NEWLINE] [STARTQ] But liberals are completely wrong. Poverty is about low IQ, low future-time orientation, and bad values. The first item on the list is all about genes: people are born with low IQ and there’s not much you can do about it. I suspect that low future-time orientation also has a large genetic component, but I think it’s possible that it can be raised with proper training. Values are, by definition, learned attitudes and behaviors. [ENDQ] [NEWLINE] &gt; The Half Sigma approach to eliminate poverty is a two-pronged attack which I guarantee would achieve good results if it were ever implemented. (1) we need to reduce the number of people with poor-people genes from being born with eugenic policies; and (2) we should focus on teaching middle-class values to poor children. The first half of the Half Sigma approach is not likely to be implemented in my lifetime because eugenics is considered to be the most evil thing in the world. There’s a possibility that we might come around to acknowledging the importance of values, but even that’s not very likely because it touches too close to the truth of HBD. Suggesting that poor people need better values tends to outrage liberals, who say that you are blaming poor people for their poverty when liberals know that poverty is caused by Republicans who are too cheap to spend the money needed to end it.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] [Life begins] at cell division [ENDQ] [NEWLINE] <mask> you would agree with medication that could stop this cell division in some way, functioning<mask> a form of morning after pill? [NEWLINE] [NEWLINE] [STARTQ] This argument is similar to saying "<mask> you shoot an arrow at a target, and each time it moves, it only moves half the distance to the target, then it never reaches the target" isn't it? [ENDQ] [NEWLINE] I don't see<mask> that relates to this situation,<mask> then I've never studied logic,<mask> I'm not an expert,<mask> it sounds rather like [Zeno's paradoxes]( [URL] %27s_paradoxes).<mask><mask> it's more like a gray-scale fade from one end to the other: we're trying to draw a line between white and black, viable human life and non-viable human life,<mask> actually none exists. [NEWLINE] [NEWLINE] <mask> do you say cell division and not the pairing of chromosomes? Surely, this is<mask> the unique-ness of the new human is created? It's been a<mask><mask> I studied it,<mask> I'm pretty sure that<mask> soon<mask> the sperm attaches to the egg membrane, a calcium cascade occurs across the egg, originating at this point, and the orientation of this cascade determines the orientation of the cell cleavage.<mask>, the process of cell division is kicked off at a very early stage, earlier than that at which you say a human life begins. [NEWLINE] [NEWLINE] I'm not sure that was an idea in his argument, and the fact that he then assumed the position which you took meant that your argument didn't challenge his view.</s>
Label encoding: <s> [STARTQ] [Life begins] at cell division [ENDQ] [NEWLINE] So you would agree with medication that could stop this cell division in some way, functioning as a form of morning after pill? [NEWLINE] [NEWLINE] [STARTQ] This argument is similar to saying " If you shoot an arrow at a target, and each time it moves, it only moves half the distance to the target, then it never reaches the target" isn't it? [ENDQ] [NEWLINE] I don't see how that relates to this situation, but then I've never studied logic, so I'm not an expert, though it sounds rather like [Zeno's paradoxes]( [URL] %27s_paradoxes). I think it's more like a gray-scale fade from one end to the other: we're trying to draw a line between white and black, viable human life and non-viable human life, when actually none exists. [NEWLINE] [NEWLINE] Why do you say cell division and not the pairing of chromosomes? Surely, this is when the unique-ness of the new human is created? It's been a while since I studied it, but I'm pretty sure that as soon as the sperm attaches to the egg membrane, a calcium cascade occurs across the egg, originating at this point, and the orientation of this cascade determines the orientation of the cell cleavage. So, the process of cell division is kicked off at a very early stage, earlier than that at which you say a human life begins. [NEWLINE] [NEWLINE] I'm not sure that was an idea in his argument, and the fact that he then assumed the position which you took meant that your argument didn't challenge his view.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0900, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] :<mask>, then, will take away these feelings of emptiness and inadequacy? [ENDQ] [NEWLINE] You can't use another person to fill the hole in yourself. [NEWLINE] [NEWLINE] A woman isn't obliged to like you, just<mask> you're not obliged to like any individual woman. [NEWLINE] [NEWLINE] Become a better you and these feelings will slowly fade.  You'll become a better person. [NEWLINE] [NEWLINE] You're 19, you might be a single virgin for the rest of you life.  You can't control<mask> /<mask> you'll meet *her*.  It could be tomorrow, it could be never. [NEWLINE] [NEWLINE] Being miserable won't help you meet *her*.  Being miserable won't help you feel better. [NEWLINE] [NEWLINE] You need to find value in yourself.  You need to create value in yourself. [NEWLINE] [NEWLINE] You are empty and inadequate, adding another person won't fix that. <mask><mask>, it'll probably make it worse and make you both more miserable. [NEWLINE] [NEWLINE] __________ [NEWLINE] [NEWLINE] It's not *her* fault.  It's yours. [NEWLINE] [NEWLINE] Fix yourself, it might not get you a GF,<mask> you'll be happier for it. [NEWLINE] ______ [NEWLINE] [NEWLINE] Women aren't scary mythical beasts.  They (generally) don't want to hurt you.  They're (generally) pretty decent normal people, just like you. [NEWLINE] [NEWLINE] They're just<mask> shy and awkward<mask> you are.  They feel just<mask> alone and terrified. [NEWLINE] [NEWLINE] They're normal people, it's just<mask> hard for them.  It's just different than your (our) experience.</s>
Label encoding: <s> [STARTQ] : What, then, will take away these feelings of emptiness and inadequacy? [ENDQ] [NEWLINE] You can't use another person to fill the hole in yourself. [NEWLINE] [NEWLINE] A woman isn't obliged to like you, just as you're not obliged to like any individual woman. [NEWLINE] [NEWLINE] Become a better you and these feelings will slowly fade.  You'll become a better person. [NEWLINE] [NEWLINE] You're 19, you might be a single virgin for the rest of you life.  You can't control if / when you'll meet *her*.  It could be tomorrow, it could be never. [NEWLINE] [NEWLINE] Being miserable won't help you meet *her*.  Being miserable won't help you feel better. [NEWLINE] [NEWLINE] You need to find value in yourself.  You need to create value in yourself. [NEWLINE] [NEWLINE] You are empty and inadequate, adding another person won't fix that.  In fact, it'll probably make it worse and make you both more miserable. [NEWLINE] [NEWLINE] __________ [NEWLINE] [NEWLINE] It's not *her* fault.  It's yours. [NEWLINE] [NEWLINE] Fix yourself, it might not get you a GF, but you'll be happier for it. [NEWLINE] ______ [NEWLINE] [NEWLINE] Women aren't scary mythical beasts.  They (generally) don't want to hurt you.  They're (generally) pretty decent normal people, just like you. [NEWLINE] [NEWLINE] They're just as shy and awkward as you are.  They feel just as alone and terrified. [NEWLINE] [NEWLINE] They're normal people, it's just as hard for them.  It's just different than your (our) experience.</s>
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm curious<mask> you're from now.<mask> you aren't happy with a law, you work on getting it changed. It's the reason marijuana legalisation and gay marriage are big issues,<mask> people want the laws changed to allow them. Laws aren't just a fact of the universe you have to accept<mask> they aren't suited to the current state of society. You change them. Plenty of laws have been unethical in the past and been changed<mask><mask><mask>, this is just another one on the list. [NEWLINE] [NEWLINE] <mask>, I didn't mean to imply the couple was making a choice. A couple who is "forced" to abort<mask> they can't afford it is the same<mask> an individual who is "forced" to abort<mask> they can't afford it. Technically they could both have the child regardless,<mask> it'd be irresponsible. You can't say a woman is forced to abort by her partner<mask> he isn't willing to pay child support any more than you can say the couple is forced to abort by the state<mask> it isn't willing to pay child support.<mask> it's solely down to you whether or not you have a child, it's not someone else's responsibility to pay for it. They aren't being told to do something under pain of suffering, they're being told<mask> they're going to make a decision by themselves, they can be responsible for it by themselves. [NEWLINE] [NEWLINE] Edit: That said, I'd be totally behind the male who got a "financial abortion" from<mask> being obligated to pay for any "biological abortion" that occured<mask><mask><mask>.</s>
Label encoding: <s>I'm curious where you're from now. If you aren't happy with a law, you work on getting it changed. It's the reason marijuana legalisation and gay marriage are big issues, because people want the laws changed to allow them. Laws aren't just a fact of the universe you have to accept if they aren't suited to the current state of society. You change them. Plenty of laws have been unethical in the past and been changed as a result, this is just another one on the list. [NEWLINE] [NEWLINE] Also, I didn't mean to imply the couple was making a choice. A couple who is "forced" to abort because they can't afford it is the same as an individual who is "forced" to abort because they can't afford it. Technically they could both have the child regardless, but it'd be irresponsible. You can't say a woman is forced to abort by her partner because he isn't willing to pay child support any more than you can say the couple is forced to abort by the state because it isn't willing to pay child support. When it's solely down to you whether or not you have a child, it's not someone else's responsibility to pay for it. They aren't being told to do something under pain of suffering, they're being told if they're going to make a decision by themselves, they can be responsible for it by themselves. [NEWLINE] [NEWLINE] Edit: That said, I'd be totally behind the male who got a "financial abortion" from also being obligated to pay for any "biological abortion" that occured as a result.</s>
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This could well be the reason I seek music with a deeper meaning and some substance and not all the shallow pop music getting cranked out today. [ENDQ] [NEWLINE] <mask><mask> this is exactly<mask> your problem lies. You're not wrong in saying there's a lot of shallow (and just plain bad) pop music being cranked out today. It's just that bad pop music has been around for<mask><mask><mask> music has been distributed to other people for money. Particularly in the 50s, 60s, and onwards.<mask> you turned on the radio back then, I would venture to guess that you'd find quite a lot of crap that never stood the test of time (<mask> you wouldn't even recognize the music). [NEWLINE] [NEWLINE] The reason it feels more prevalent today is... well<mask> it *is* more prevalent. There are far more artists making commercial music today<mask> there are more avenues to distribute it to you. The internet is the perfect medium for sharing more niche music - and<mask> there is more variety and more volume to choose from. And<mask> more crap to wade through. [NEWLINE] [NEWLINE] <mask> that<mask> means there's more good music - it's just split into different fragments and subgenres. Decades ago, you would probably have to live in the right town to even be exposed to a particular niche of music (most underground or local music scenes didn't really have the means to spread beyond a small geographic location - and the radio certainly wasn't playing their music). Now, you can go to YouTube and find all kinds of bands that never would have seen the light of day 40 years ago.</s>
Label encoding: <s> [STARTQ] This could well be the reason I seek music with a deeper meaning and some substance and not all the shallow pop music getting cranked out today. [ENDQ] [NEWLINE] I think this is exactly where your problem lies. You're not wrong in saying there's a lot of shallow (and just plain bad) pop music being cranked out today. It's just that bad pop music has been around for as long as music has been distributed to other people for money. Particularly in the 50s, 60s, and onwards. If you turned on the radio back then, I would venture to guess that you'd find quite a lot of crap that never stood the test of time ( so you wouldn't even recognize the music). [NEWLINE] [NEWLINE] The reason it feels more prevalent today is... well because it *is* more prevalent. There are far more artists making commercial music today because there are more avenues to distribute it to you. The internet is the perfect medium for sharing more niche music - and so there is more variety and more volume to choose from. And thus more crap to wade through. [NEWLINE] [NEWLINE] But that also means there's more good music - it's just split into different fragments and subgenres. Decades ago, you would probably have to live in the right town to even be exposed to a particular niche of music (most underground or local music scenes didn't really have the means to spread beyond a small geographic location - and the radio certainly wasn't playing their music). Now, you can go to YouTube and find all kinds of bands that never would have seen the light of day 40 years ago.</s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Glad to help! That article was really helpful in cementing<mask> I personally think<mask> hopefully you find it useful. [NEWLINE] [NEWLINE] <mask> for the two differences you mentioned,<mask><mask> in the first way the violinist is supposed to be similar to an infant<mask> neither of them chose the situation (my take is that the choice to attach the violinist was made after they became unconscious).<mask>,<mask> it's the Society that chose to attach the violinist, then the question is,<mask>'s the analogue of the Society in the pregnancy scenario? In pregnancy, the Society is pretty much just "chance", in that there isn't any actual intentional thing that happens to cause you to be pregnant, just that it happens to happen. [NEWLINE] [NEWLINE] <mask><mask><mask> the second difference,<mask><mask> this is a really good question, one that I don't think the article I linked really addresses.<mask> you could think of it this way:<mask> two people decide to have sex, and they both agree that they do not want to have a pregnancy (and even demonstrate this by taking various birth control precautions), then are they consenting to a pregnancy?<mask> in, can their joint expression outweigh any implicit consent to pregnancy contained within the act of sex? Personally,<mask><mask> it can, just from<mask><mask><mask> about consent-- I don't think it would be very meaningful<mask> "avoiding trying to do something" could count<mask> "consenting" in some cases,<mask> that seems to mean you can consent to something even<mask> you don't want it to happen.<mask> I can imagine that others might disagree on this.</s>
Label encoding: <s>Glad to help! That article was really helpful in cementing what I personally think so hopefully you find it useful. [NEWLINE] [NEWLINE] As for the two differences you mentioned, I think in the first way the violinist is supposed to be similar to an infant as neither of them chose the situation (my take is that the choice to attach the violinist was made after they became unconscious). So, if it's the Society that chose to attach the violinist, then the question is, what's the analogue of the Society in the pregnancy scenario? In pregnancy, the Society is pretty much just "chance", in that there isn't any actual intentional thing that happens to cause you to be pregnant, just that it happens to happen. [NEWLINE] [NEWLINE] As far as the second difference, I think this is a really good question, one that I don't think the article I linked really addresses. However you could think of it this way: if two people decide to have sex, and they both agree that they do not want to have a pregnancy (and even demonstrate this by taking various birth control precautions), then are they consenting to a pregnancy? As in, can their joint expression outweigh any implicit consent to pregnancy contained within the act of sex? Personally, I think it can, just from how I think about consent-- I don't think it would be very meaningful if "avoiding trying to do something" could count as "consenting" in some cases, since that seems to mean you can consent to something even if you don't want it to happen. But I can imagine that others might disagree on this.</s>
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My point was that<mask><mask><mask> there are bad people, and you want to build up a system that relies<mask> heavily on responsibility, you won't get around being authoritarian,<mask> your system will rely heavily on people not stepping out of line, and that goes extremely against personal freedom. [NEWLINE] [NEWLINE] Like I said, you need only good people and practically unlimited resources, then such a system makes sense<mask> the authoritarianism part isn't needed. Otherwise, not<mask> much. [NEWLINE] [NEWLINE] <mask> still, the system we have today is an almost perfect balance already (not quite perfect,<mask> close, at least here in Switzerland). I say social liberalism is the best way to go,<mask> then you have the social part,<mask> you<mask> have a just, meritocratic distribution mechanism. I say one should be able to own<mask> much<mask> one earned,<mask> there still needs a necessary base amount of goods that you can own, even<mask> you earned nothing<mask>, and that thing that assures this is called welfare. [NEWLINE] [NEWLINE] Seriously, the system we have today is really good, and includes a meritocratic distribution mechanism.<mask> the meritocraty-part is not the case, it doesn't work of course,<mask> that's<mask> the state needs to make sure it all stays mostly meritocratic and make regulations<mask>,<mask> rampant, uncontrolled capitalism is horrible too. [NEWLINE] [NEWLINE] Either way, the solution will be that everything stays the same, except that things are even more meritocratic and just, and that we won't have to work anymore and can instead focus on the arts, sciences, sports and leisures.</s>
Label encoding: <s>My point was that as long as there are bad people, and you want to build up a system that relies so heavily on responsibility, you won't get around being authoritarian, because your system will rely heavily on people not stepping out of line, and that goes extremely against personal freedom. [NEWLINE] [NEWLINE] Like I said, you need only good people and practically unlimited resources, then such a system makes sense because the authoritarianism part isn't needed. Otherwise, not so much. [NEWLINE] [NEWLINE] But still, the system we have today is an almost perfect balance already (not quite perfect, but close, at least here in Switzerland). I say social liberalism is the best way to go, because then you have the social part, but you also have a just, meritocratic distribution mechanism. I say one should be able to own as much as one earned, but there still needs a necessary base amount of goods that you can own, even if you earned nothing yet, and that thing that assures this is called welfare. [NEWLINE] [NEWLINE] Seriously, the system we have today is really good, and includes a meritocratic distribution mechanism. If the meritocraty-part is not the case, it doesn't work of course, but that's why the state needs to make sure it all stays mostly meritocratic and make regulations accordingly, because rampant, uncontrolled capitalism is horrible too. [NEWLINE] [NEWLINE] Either way, the solution will be that everything stays the same, except that things are even more meritocratic and just, and that we won't have to work anymore and can instead focus on the arts, sciences, sports and leisures.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I thought about this for awhile, and I first thought that one major difference between the two subreddits is that TiA doesn't ban anybody with a different opinion like SRS does.<mask><mask> I thought about it some more I realized that<mask> SRS is<mask> unpopular and TiA is<mask> popular that TiA doesn't need any mod enforcement to make the sub a circlejerk. I remember seeing someone try to support Affirmative Action on the sub and they got hammered with downvotes even<mask> they were being completely reasonable. [ENDQ] [NEWLINE] That may be true,<mask> there's still a significant difference there. The way it is now the TiA community still has the potential to do better, for example to be more accepting of differing or unpopular opinions. Whereas the similiar bad attitudes we can see in SRS are built into the whole structure of that community;<mask> the subs are run,<mask> they're connected,<mask> they're introduced to new users (like the casual bigotry built right in to the official SRS faq). It's like a few bad apples in a pile can potentially be picked out, or just not eaten,<mask> make apple-sauce with that and now the bad is inextricable from the rest; its now all bad. [NEWLINE] [NEWLINE] The SRS community is irreversibly poisoned, is<mask> I'm saying. Whereas in the TiA subs you'll find some attitudes and views you agree with and some you very much do not,<mask> that's ok<mask> agreement with specific views is not enforced there and it's not meant to be.</s>
Label encoding: <s> [STARTQ] I thought about this for awhile, and I first thought that one major difference between the two subreddits is that TiA doesn't ban anybody with a different opinion like SRS does. But when I thought about it some more I realized that since SRS is so unpopular and TiA is so popular that TiA doesn't need any mod enforcement to make the sub a circlejerk. I remember seeing someone try to support Affirmative Action on the sub and they got hammered with downvotes even when they were being completely reasonable. [ENDQ] [NEWLINE] That may be true, but there's still a significant difference there. The way it is now the TiA community still has the potential to do better, for example to be more accepting of differing or unpopular opinions. Whereas the similiar bad attitudes we can see in SRS are built into the whole structure of that community; how the subs are run, how they're connected, how they're introduced to new users (like the casual bigotry built right in to the official SRS faq). It's like a few bad apples in a pile can potentially be picked out, or just not eaten, but make apple-sauce with that and now the bad is inextricable from the rest; its now all bad. [NEWLINE] [NEWLINE] The SRS community is irreversibly poisoned, is what I'm saying. Whereas in the TiA subs you'll find some attitudes and views you agree with and some you very much do not, but that's ok because agreement with specific views is not enforced there and it's not meant to be.</s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But from the point of view of someone who is convinced of their religion, the analogy is correct. I would agree with you that from an objective standpoint on the issue, the analogies are not accurate.<mask> from the theist's viewpoint, they are. [NEWLINE] [NEWLINE] From an objective standpoint, you could fix the analogies by making a small change - that the person who saw the bomb / saw the thing go into the drink may have just thought they saw those things. They are known to suffer from hallucinations and delusions, and are highly paranoid about such things.<mask> it is highly likely, in our scenario, that the person thinks they saw a bomb / the drink being drugged,<mask> in reality, neither such thing happened. [NEWLINE] [NEWLINE] Does that change anything? From their viewpoint, they are still convinced of<mask> they saw. They still see it<mask> vitally important to warn others about it. From their perspective, it is real. [NEWLINE] [NEWLINE] I don't think trying to stop someone from starting a car you think is wired to a bomb is rude. I don't think it's rude<mask> there really is a bomb, and I don't think it is rude<mask> there is not a bomb,<mask><mask><mask> the person truly believes that one is there. I don't think you can say "Well, it turned out you were wrong,<mask> actually you were retroactively rude all along". [NEWLINE] [NEWLINE] (Again, do keep in mind that I am an atheist - I'm not arguing for any gods, I'm arguing in favor of seeing things from such a person's perspective)</s>
Label encoding: <s>But from the point of view of someone who is convinced of their religion, the analogy is correct. I would agree with you that from an objective standpoint on the issue, the analogies are not accurate. But from the theist's viewpoint, they are. [NEWLINE] [NEWLINE] From an objective standpoint, you could fix the analogies by making a small change - that the person who saw the bomb / saw the thing go into the drink may have just thought they saw those things. They are known to suffer from hallucinations and delusions, and are highly paranoid about such things. So it is highly likely, in our scenario, that the person thinks they saw a bomb / the drink being drugged, while in reality, neither such thing happened. [NEWLINE] [NEWLINE] Does that change anything? From their viewpoint, they are still convinced of what they saw. They still see it as vitally important to warn others about it. From their perspective, it is real. [NEWLINE] [NEWLINE] I don't think trying to stop someone from starting a car you think is wired to a bomb is rude. I don't think it's rude if there really is a bomb, and I don't think it is rude if there is not a bomb, so long as the person truly believes that one is there. I don't think you can say "Well, it turned out you were wrong, so actually you were retroactively rude all along". [NEWLINE] [NEWLINE] (Again, do keep in mind that I am an atheist - I'm not arguing for any gods, I'm arguing in favor of seeing things from such a person's perspective)</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It has often been pointed out that in a world<mask> many nations are armed with nuclear weapons, war has the potential of destroying the human race.  War is no longer a harmless pursuit by which aggressive men fight it out to see who is more powerful (much like a football game,<mask> more extreme).  It is, or can become utterly deadly to everyone, including those who fight and those who don't fight.  Even without the use of nuclear weapons, it is routine for civilian populations to be bombed.  War is ridiculously cruel and causes immense suffering to innocent people (<mask> well<mask> to guilty people). <mask><mask> you tell me that world peace is not feasible, my reaction is that it is really war that is not feasible.  The human race faces a choice, of either becoming mature enough to solve international disagreements in a non-violent manner, or its own destruction.  I would<mask> like to point out that even without killing people directly, war has many indirect consequences.  War is very environmentally destructive, and it consumes resources which otherwise would be available to deal with environmental problems. <mask> a nation is at war, global warming is simply not an issue. <mask>, an eventually catastrophic run-away greenhouse effect can easily result simply from our inability to address environmental issues<mask> we are busy killing each other.  War is far more dangerous and impractical now, in the 21st century, than it has ever been before.  We cannot afford it. <mask> world peace truly violates human nature<mask> you claim, we can either transcend our nature, or become extinct.</s>
Label encoding: <s>It has often been pointed out that in a world where many nations are armed with nuclear weapons, war has the potential of destroying the human race.  War is no longer a harmless pursuit by which aggressive men fight it out to see who is more powerful (much like a football game, but more extreme).  It is, or can become utterly deadly to everyone, including those who fight and those who don't fight.  Even without the use of nuclear weapons, it is routine for civilian populations to be bombed.  War is ridiculously cruel and causes immense suffering to innocent people ( as well as to guilty people).  So when you tell me that world peace is not feasible, my reaction is that it is really war that is not feasible.  The human race faces a choice, of either becoming mature enough to solve international disagreements in a non-violent manner, or its own destruction.  I would also like to point out that even without killing people directly, war has many indirect consequences.  War is very environmentally destructive, and it consumes resources which otherwise would be available to deal with environmental problems.  When a nation is at war, global warming is simply not an issue.  So, an eventually catastrophic run-away greenhouse effect can easily result simply from our inability to address environmental issues while we are busy killing each other.  War is far more dangerous and impractical now, in the 21st century, than it has ever been before.  We cannot afford it.  If world peace truly violates human nature as you claim, we can either transcend our nature, or become extinct.</s>
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Except I'm pro choice<mask> I can recognize the difference between a woman having an abortion and a man deciding to leave his child. [ENDQ] [NEWLINE] And that difference (assuming you're referring to one that is relevant to this conversation and not to the numerous differences that are not relevant) is<mask> exactly? [NEWLINE] [NEWLINE] You're conflating consent to sex with consent to being a parent. This is the same thing that pro-lifers do in order to claim that a woman has no right to an abortion<mask> "she knew<mask> she was doing." Watch this: [NEWLINE] [NEWLINE] [STARTQ] The girl could have made the guy wear a condom and could have made sure she was on birth control. She intentionally or non-intentionally let the man ejaculate inside of her.<mask> she didn't want a baby she should have gone through the check list.<mask> you're arguing is assuming someone is dumb enough to get knocked up<mask> they're lack of care well then they should be force to take care of the child or pay child support<mask> they fucked up. [ENDQ] [NEWLINE] Notice<mask> it's exactly<mask> you said,<mask> switched a bit? Notice<mask> it's a fundamentally flawed pro-life argument? This is<mask> I mean. [NEWLINE] [NEWLINE] (And before you say that the woman doesn't always get to choose<mask> she's impregnated, men don't always get to choose<mask> a woman pokes holes in a condom, lies, drugs and rapes him, etc. Non-consensual or deceitful sex can occur on both sides<mask> well and isn't fundamental to the discussion<mask><mask> )</s>
Label encoding: <s> [STARTQ] Except I'm pro choice but I can recognize the difference between a woman having an abortion and a man deciding to leave his child. [ENDQ] [NEWLINE] And that difference (assuming you're referring to one that is relevant to this conversation and not to the numerous differences that are not relevant) is what exactly? [NEWLINE] [NEWLINE] You're conflating consent to sex with consent to being a parent. This is the same thing that pro-lifers do in order to claim that a woman has no right to an abortion because "she knew what she was doing." Watch this: [NEWLINE] [NEWLINE] [STARTQ] The girl could have made the guy wear a condom and could have made sure she was on birth control. She intentionally or non-intentionally let the man ejaculate inside of her. If she didn't want a baby she should have gone through the check list. What you're arguing is assuming someone is dumb enough to get knocked up because they're lack of care well then they should be force to take care of the child or pay child support since they fucked up. [ENDQ] [NEWLINE] Notice how it's exactly what you said, but switched a bit? Notice how it's a fundamentally flawed pro-life argument? This is what I mean. [NEWLINE] [NEWLINE] (And before you say that the woman doesn't always get to choose when she's impregnated, men don't always get to choose when a woman pokes holes in a condom, lies, drugs and rapes him, etc. Non-consensual or deceitful sex can occur on both sides as well and isn't fundamental to the discussion IMO )</s>
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is most of the trauma simply caused by peoples' reactions, rather than the crime itself? [ENDQ] [NEWLINE] It's both. The crime creates the first trauma. People's reactions to that crime recreate that trauma, plus multiply trauma on top of it. [NEWLINE] [NEWLINE] For example, imagine telling the survivor of an attempted homicide that they need to try strangulation again, with someone they love? It can feel wonderful, after all,<mask> the risks associated with it. [NEWLINE] [NEWLINE] Alternatively, imagine telling anyone who nearly died in a car crash, that wanting to drive again was proof they alone were responsible for the accident? Or that there was no accident at all? [NEWLINE] [NEWLINE] Both metaphors apply to the experience of being a rape victim. [NEWLINE] [NEWLINE] Then there's the way in which society in general handles sex. In some places, it's a taboo, which means it's everywhere,<mask> you're not allowed to speak of it. [NEWLINE] [NEWLINE] Imagine<mask> instead of sex, it was spiders. Imagine a world<mask> people hide spiders underneath their clothes. You can see them moving, sometimes. Sometimes, they want you to see them moving. [NEWLINE] [NEWLINE] Imagine the shadows of spiders sold perfume, or fashion, or cars. Imagine<mask> there were webs all over billboards and magazine stands, all to sell things to people who like spiders.<mask> you're not allowed to mention it.<mask> you do, it must mean you're obsessed with spiders. And<mask> you're obsessed with spiders, people can't be blamed<mask> their spiders bite you... [NEWLINE] [NEWLINE] Have any of these metaphors helped? </s>
Label encoding: <s> [STARTQ] Is most of the trauma simply caused by peoples' reactions, rather than the crime itself? [ENDQ] [NEWLINE] It's both. The crime creates the first trauma. People's reactions to that crime recreate that trauma, plus multiply trauma on top of it. [NEWLINE] [NEWLINE] For example, imagine telling the survivor of an attempted homicide that they need to try strangulation again, with someone they love? It can feel wonderful, after all, despite the risks associated with it. [NEWLINE] [NEWLINE] Alternatively, imagine telling anyone who nearly died in a car crash, that wanting to drive again was proof they alone were responsible for the accident? Or that there was no accident at all? [NEWLINE] [NEWLINE] Both metaphors apply to the experience of being a rape victim. [NEWLINE] [NEWLINE] Then there's the way in which society in general handles sex. In some places, it's a taboo, which means it's everywhere, but you're not allowed to speak of it. [NEWLINE] [NEWLINE] Imagine if instead of sex, it was spiders. Imagine a world where people hide spiders underneath their clothes. You can see them moving, sometimes. Sometimes, they want you to see them moving. [NEWLINE] [NEWLINE] Imagine the shadows of spiders sold perfume, or fashion, or cars. Imagine if there were webs all over billboards and magazine stands, all to sell things to people who like spiders. But you're not allowed to mention it. If you do, it must mean you're obsessed with spiders. And if you're obsessed with spiders, people can't be blamed if their spiders bite you... [NEWLINE] [NEWLINE] Have any of these metaphors helped? </s>
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am a huge fan of cycling and try to do it whenever I can (I wish I worked in my city). <mask>, the reason cyclists should follow the same rules<mask> cars is primarily for the overall safety of the cyclist. [NEWLINE] [NEWLINE] [NEWLINE] I believe there are a great deal of cyclists who are actively aware and very good at navigating traffic patterns. <mask> just<mask> with drivers, not all cyclists are actively aware, especially those places with prominent bike share programs.  Factor in drivers in metric ton vehicles.  Many times they are not very aware, not actively looking, and actually lack a great deal of field of view to their surroundings especially<mask> they're tourists visiting the city and are not used to driving around with bikes. <mask> a car collides with a car in the city, they are rarely traveling at a velocity that causes serious injury to each other or the persons inside. <mask>, a bicycle moving at 15-20 mph,<mask> even clipped, can have a devastating effect on the rider.  By filtering you run the risk of getting "doored", turned into, or hitting a pedestrian you may not have seen past the vehicle. [NEWLINE] [NEWLINE] Trust me, I know the frustration of having to come to full stops. <mask> traffic laws exist to create predictability and rules of conduct to make drivers safer.  By not following the expected rules of the road, your unpredictability creates a significant statistical increase in chances of an accident, which the negative results of are greatly increased due to the lack of protections a bike has versus a car. </s>
Label encoding: <s>I am a huge fan of cycling and try to do it whenever I can (I wish I worked in my city).  However, the reason cyclists should follow the same rules as cars is primarily for the overall safety of the cyclist. [NEWLINE] [NEWLINE] [NEWLINE] I believe there are a great deal of cyclists who are actively aware and very good at navigating traffic patterns.  But just as with drivers, not all cyclists are actively aware, especially those places with prominent bike share programs.  Factor in drivers in metric ton vehicles.  Many times they are not very aware, not actively looking, and actually lack a great deal of field of view to their surroundings especially if they're tourists visiting the city and are not used to driving around with bikes.  If a car collides with a car in the city, they are rarely traveling at a velocity that causes serious injury to each other or the persons inside.  However, a bicycle moving at 15-20 mph, if even clipped, can have a devastating effect on the rider.  By filtering you run the risk of getting "doored", turned into, or hitting a pedestrian you may not have seen past the vehicle. [NEWLINE] [NEWLINE] Trust me, I know the frustration of having to come to full stops.  But traffic laws exist to create predictability and rules of conduct to make drivers safer.  By not following the expected rules of the road, your unpredictability creates a significant statistical increase in chances of an accident, which the negative results of are greatly increased due to the lack of protections a bike has versus a car. </s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0990, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] People seems to think that racial profiling has no use. [ENDQ] [NEWLINE] Well,<mask> it technically matches your claim, I don't think "there is exactly 1 scenario<mask> racial profiling works" is<mask> you meant by "racial profiling works". [NEWLINE] [NEWLINE] There is certainly *a* use for racial profiling...<mask> it is a minority use.  The general situations that people think of racial profiling<mask> a tool (airports searches, police crime investigations), it almost certainly would net a negative effect. [NEWLINE] [NEWLINE] [STARTQ] Racial profiling will remain<mask> one of the assessment factor for law enforcement<mask> it is useful,<mask><mask> it is wrong. [ENDQ] [NEWLINE] I don't care about the morals of anything in law...<mask> think about this.  The aggregate job of law enforcement (like the rest of the criminal justice branch) is reduction of crime.  Should some stupid action drastically reduce the crime rate, there is an onus on the law enforcement world to take that action. [NEWLINE] [NEWLINE] And racial profiling seems like that, doesn't it...<mask><mask><mask> it's the opposite.  Racial profiling causes two effects that worsens the crime situation<mask> a whole: [NEWLINE] [NEWLINE] 1. A general mistrust of law enforcement by the minorities.  This means INNOCENTS are less likely to talk to police, and criminals are less likely to turn themselves in... the police are going to screw them bad<mask> they're black. [NEWLINE] [NEWLINE] 2. An overall increase of crime rates.  Being treated unfairly by an authority figure makes you less likely to obey that authority figure.</s>
Label encoding: <s> [STARTQ] People seems to think that racial profiling has no use. [ENDQ] [NEWLINE] Well, while it technically matches your claim, I don't think "there is exactly 1 scenario where racial profiling works" is what you meant by "racial profiling works". [NEWLINE] [NEWLINE] There is certainly *a* use for racial profiling... but it is a minority use.  The general situations that people think of racial profiling as a tool (airports searches, police crime investigations), it almost certainly would net a negative effect. [NEWLINE] [NEWLINE] [STARTQ] Racial profiling will remain as one of the assessment factor for law enforcement because it is useful, even though it is wrong. [ENDQ] [NEWLINE] I don't care about the morals of anything in law... But think about this.  The aggregate job of law enforcement (like the rest of the criminal justice branch) is reduction of crime.  Should some stupid action drastically reduce the crime rate, there is an onus on the law enforcement world to take that action. [NEWLINE] [NEWLINE] And racial profiling seems like that, doesn't it... but I think it's the opposite.  Racial profiling causes two effects that worsens the crime situation as a whole: [NEWLINE] [NEWLINE] 1. A general mistrust of law enforcement by the minorities.  This means INNOCENTS are less likely to talk to police, and criminals are less likely to turn themselves in... the police are going to screw them bad because they're black. [NEWLINE] [NEWLINE] 2. An overall increase of crime rates.  Being treated unfairly by an authority figure makes you less likely to obey that authority figure.</s>
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"A second example of dogmatic ethics gone wrong for lack of knowledge is homophobia. The basal reasoning is much the same<mask> for opposition to artificial contraception: sex not intended for reproduction must be an aberration and a sin.<mask> an abundance of evidence points to the opposite. Committed homosexuality, with the preference appearing in childhood, is heritable. This means the trait is not always fixed,<mask> part of the greater likelihood of a person's developing into a homosexual is prescribed by genes that differ from those that lead to heterosexuality. It has further turned out that heredity-influenced homosexuality occurs in populations worldwide too frequently to be due to mutations alone. Population geneticists use a rule of thumb to account for abundance at this level:<mask> a trait cannot be due solely to random mutations, and<mask> it lowers or eliminates reproduction in those who have it, then the trait must be favored by natural selection working on a target of some other kind. For example, a low dose of homosexual-tending genes may give competitive advantages to a practicing heterosexual. Or, homosexuality may give advantages to the group by special talents, unusual qualities of personality, and the specialized roles and professions it generates. There is abundant evidence that such is the case in both preliterate and modern societies. Either way, societies are mistaken to disapprove of homosexuality<mask> gays have different sexual preferences and reproduce less. Their presence should be valued instead for<mask> they contribute constructively to human diversity. A society that condemns homosexuality harms itself." [NEWLINE] [NEWLINE] EO Wilson, Harvard Professor Emeritus </s>
Label encoding: <s>"A second example of dogmatic ethics gone wrong for lack of knowledge is homophobia. The basal reasoning is much the same as for opposition to artificial contraception: sex not intended for reproduction must be an aberration and a sin. But an abundance of evidence points to the opposite. Committed homosexuality, with the preference appearing in childhood, is heritable. This means the trait is not always fixed, but part of the greater likelihood of a person's developing into a homosexual is prescribed by genes that differ from those that lead to heterosexuality. It has further turned out that heredity-influenced homosexuality occurs in populations worldwide too frequently to be due to mutations alone. Population geneticists use a rule of thumb to account for abundance at this level: if a trait cannot be due solely to random mutations, and yet it lowers or eliminates reproduction in those who have it, then the trait must be favored by natural selection working on a target of some other kind. For example, a low dose of homosexual-tending genes may give competitive advantages to a practicing heterosexual. Or, homosexuality may give advantages to the group by special talents, unusual qualities of personality, and the specialized roles and professions it generates. There is abundant evidence that such is the case in both preliterate and modern societies. Either way, societies are mistaken to disapprove of homosexuality because gays have different sexual preferences and reproduce less. Their presence should be valued instead for what they contribute constructively to human diversity. A society that condemns homosexuality harms itself." [NEWLINE] [NEWLINE] EO Wilson, Harvard Professor Emeritus </s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] All of the circumstances that cause a circumstancial depression can actually cause the imbalance in the brain chemistry, can it not? [ENDQ] [NEWLINE] Can it?  I can't find anything that suggests that it can... [NEWLINE] [NEWLINE] [STARTQ] The antidepressants will then be useful to help you deal with said circumstances.<mask> you do not help these people (with medication or not), they risk just spiraling further into depression.<mask> you catch it<mask> it is not too bad, you can prevent them from getting really ill. [ENDQ] [NEWLINE] I'll humor you<mask> you haven't responded to the first part<mask> :  Even<mask> these circumstances can cause a chemical imbalance, that would mean we have to rethink<mask> we deal with antidepressants.  There would be no excuse for prescribing antidepressants without therapy<mask> the goal would have to be normalizing people and getting them off the drugs, not keep them on them for life<mask><mask> depression were a chronic disease. [NEWLINE] [NEWLINE] [STARTQ] <mask> this imbalance is not present in all cases of clinical depression, should we just not medically treat the people who show no imbalance,<mask> have every symptom of depression? [ENDQ] [NEWLINE] <mask><mask><mask> alternative treatments have been ruled out, I don't see a problem with trialing antidepressants and monitoring patients for side-effects under these circumstances. <mask> you brought up, we know very little about<mask> the brain works; I would put this in the "people who might really need it" category,<mask> with great caution. <mask><mask> that the symptoms of depression alone aren't enough to properly determine a need for antidepressants.</s>
Label encoding: <s> [STARTQ] All of the circumstances that cause a circumstancial depression can actually cause the imbalance in the brain chemistry, can it not? [ENDQ] [NEWLINE] Can it?  I can't find anything that suggests that it can... [NEWLINE] [NEWLINE] [STARTQ] The antidepressants will then be useful to help you deal with said circumstances. If you do not help these people (with medication or not), they risk just spiraling further into depression. If you catch it while it is not too bad, you can prevent them from getting really ill. [ENDQ] [NEWLINE] I'll humor you as you haven't responded to the first part yet :  Even if these circumstances can cause a chemical imbalance, that would mean we have to rethink how we deal with antidepressants.  There would be no excuse for prescribing antidepressants without therapy as the goal would have to be normalizing people and getting them off the drugs, not keep them on them for life as if depression were a chronic disease. [NEWLINE] [NEWLINE] [STARTQ] If this imbalance is not present in all cases of clinical depression, should we just not medically treat the people who show no imbalance, but have every symptom of depression? [ENDQ] [NEWLINE] As long as alternative treatments have been ruled out, I don't see a problem with trialing antidepressants and monitoring patients for side-effects under these circumstances.  As you brought up, we know very little about how the brain works; I would put this in the "people who might really need it" category, but with great caution.  I think that the symptoms of depression alone aren't enough to properly determine a need for antidepressants.</s>
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] And damage does not always cause inconvenience.<mask> I want to junk my car<mask> they only offer me 200 bucks for it, I can instead roll my car into the middle of the road around a turn and make more money off of the driver who comes around a bit too fast and totals my vehicle. [ENDQ] [NEWLINE] You may note that this rarely occurs. <mask><mask><mask> that<mask> the cost to repair the car exceed the value of the car, insurance only pays the value of the car. <mask><mask> your car was only worth $200, insurance is only going to pay $200. [NEWLINE] [NEWLINE] Likewise, your example of the "lose a bumper here, a bumper there" doesn't seem to occur much.  First, reckless behavior and/or frequent claims raise your premiums.  Second, you can't really pick and choose<mask> damage will be done to your car. [NEWLINE] [NEWLINE] Now you may worry about insurance fraud<mask> you claim something was broken in the accident<mask> it was already broken.  This is unchanged with your proposed rule change,<mask> I can<mask> easily make this claim and get a "free" repair of the past damage<mask> to make this claim and get a "free" tv (or insulin or textbooks or whatever my best use of the money is). [NEWLINE] [NEWLINE] Basically, I don't think people are putting their vehicles in harm's way under the current system.  All the current system does is reduce inefficiency slightly<mask> it turns out that repairing my car isn't actually my best use for the money.  </s>
Label encoding: <s> [STARTQ] And damage does not always cause inconvenience. If I want to junk my car but they only offer me 200 bucks for it, I can instead roll my car into the middle of the road around a turn and make more money off of the driver who comes around a bit too fast and totals my vehicle. [ENDQ] [NEWLINE] You may note that this rarely occurs.  The reason is that if the cost to repair the car exceed the value of the car, insurance only pays the value of the car.  So if your car was only worth $200, insurance is only going to pay $200. [NEWLINE] [NEWLINE] Likewise, your example of the "lose a bumper here, a bumper there" doesn't seem to occur much.  First, reckless behavior and/or frequent claims raise your premiums.  Second, you can't really pick and choose what damage will be done to your car. [NEWLINE] [NEWLINE] Now you may worry about insurance fraud where you claim something was broken in the accident when it was already broken.  This is unchanged with your proposed rule change, since I can as easily make this claim and get a "free" repair of the past damage as to make this claim and get a "free" tv (or insulin or textbooks or whatever my best use of the money is). [NEWLINE] [NEWLINE] Basically, I don't think people are putting their vehicles in harm's way under the current system.  All the current system does is reduce inefficiency slightly when it turns out that repairing my car isn't actually my best use for the money.  </s>
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You're focused on the question of whether a person needs to take scripture literally to be properly religious,<mask> this isn't pertinent; the pertinent question is,<mask> possibly acceptable purpose could it serve for a deity to include demands to persecute homosexuals, stone women, and slaughter infidels in it's holy book<mask> that isn't the jist of the intended message? [ENDQ] [NEWLINE] And the answer that Christians give to that changes from denomination to denomination.  There are a number of answers. [NEWLINE] [NEWLINE] The answer Christians give that I find most compelling is that people did it.  Humans have long used religion to justify their abuse of other humans.  It is entirely possible that some things in the Bible are an artifact of that, mixed in with the divine truth is a large helping of human pettiness. [NEWLINE] [NEWLINE] <mask> there are other answers you may find more satisfying (or less). [NEWLINE] [NEWLINE] [STARTQ] Once you realize the Earth is round you don't cling to calling yourself a Flat-Earther who merely has adopted a non-literal interpretation of the theory. [ENDQ] [NEWLINE] Physicists still call themselves physicists without believing a word of Aristotle's physics - I bet most of them haven't even read it.  They still call themselves physicists without believing large parts of<mask> Newton had to say too.  Psychologists don't believe most of<mask> Freud wrote. [NEWLINE] [NEWLINE] The understanding of the source material changes over time, for all fields, for all attempts to understand the reality of the world. <mask> should religion be an exception?</s>
Label encoding: <s> [STARTQ] You're focused on the question of whether a person needs to take scripture literally to be properly religious, but this isn't pertinent; the pertinent question is, what possibly acceptable purpose could it serve for a deity to include demands to persecute homosexuals, stone women, and slaughter infidels in it's holy book if that isn't the jist of the intended message? [ENDQ] [NEWLINE] And the answer that Christians give to that changes from denomination to denomination.  There are a number of answers. [NEWLINE] [NEWLINE] The answer Christians give that I find most compelling is that people did it.  Humans have long used religion to justify their abuse of other humans.  It is entirely possible that some things in the Bible are an artifact of that, mixed in with the divine truth is a large helping of human pettiness. [NEWLINE] [NEWLINE] But there are other answers you may find more satisfying (or less). [NEWLINE] [NEWLINE] [STARTQ] Once you realize the Earth is round you don't cling to calling yourself a Flat-Earther who merely has adopted a non-literal interpretation of the theory. [ENDQ] [NEWLINE] Physicists still call themselves physicists without believing a word of Aristotle's physics - I bet most of them haven't even read it.  They still call themselves physicists without believing large parts of what Newton had to say too.  Psychologists don't believe most of what Freud wrote. [NEWLINE] [NEWLINE] The understanding of the source material changes over time, for all fields, for all attempts to understand the reality of the world.  Why should religion be an exception?</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>the problem lies a lot in the persons perception of<mask> is funny, interesting and well dressed (I leave out successful and athletic,<mask> the majority of "nice guys" are pretty young nerdy and socially awkward. and<mask> these things are generally out of the picture.) and<mask> you look at the media that these guys learn their technique from. they ARE being<mask> they believe to be funny, interesting and well dressed. [NEWLINE] [NEWLINE] and<mask><mask> I may not completely agree with it. the red pill subreddit has this story in the sidebar. [URL] Which really goes in depth explaining many of the issues that these "nice guys" face. [NEWLINE] [STARTQ] My issue was i always believed i was not handsome, rugged or built well enough to attract initial attention. I had poor self image. All the advice to the contrary, telling me I WAS OK<mask> I WAS allowed me to abdicate my responsibility to start working on that issue. It led me to believe people should like me for who i am, not<mask> my exterior present [ENDQ] [NEWLINE] and [NEWLINE] [NEWLINE] [STARTQ] I got to have the pleasure of defending women from the barbs and negs of my player friends only to watch these same women i defended end up going home to sleep with them. My brain simply could not comprehend<mask> the fuck was going on.<mask> the fuck is wrong with these women? Oh Wait! I’m not allowed to question that. [ENDQ] [NEWLINE] and I have to say, that these experiences resonate deeply with my own during my time<mask> a<mask> called "nice guy" </s>
Label encoding: <s>the problem lies a lot in the persons perception of what is funny, interesting and well dressed (I leave out successful and athletic, because the majority of "nice guys" are pretty young nerdy and socially awkward. and so these things are generally out of the picture.) and if you look at the media that these guys learn their technique from. they ARE being what they believe to be funny, interesting and well dressed. [NEWLINE] [NEWLINE] and even though I may not completely agree with it. the red pill subreddit has this story in the sidebar. [URL] Which really goes in depth explaining many of the issues that these "nice guys" face. [NEWLINE] [STARTQ] My issue was i always believed i was not handsome, rugged or built well enough to attract initial attention. I had poor self image. All the advice to the contrary, telling me I WAS OK AS I WAS allowed me to abdicate my responsibility to start working on that issue. It led me to believe people should like me for who i am, not what my exterior present [ENDQ] [NEWLINE] and [NEWLINE] [NEWLINE] [STARTQ] I got to have the pleasure of defending women from the barbs and negs of my player friends only to watch these same women i defended end up going home to sleep with them. My brain simply could not comprehend what the fuck was going on. What the fuck is wrong with these women? Oh Wait! I’m not allowed to question that. [ENDQ] [NEWLINE] and I have to say, that these experiences resonate deeply with my own during my time as a so called "nice guy" </s>
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You don't need to study philosophy to become a capable scientist. [ENDQ] [NEWLINE] That may be true,<mask><mask> you want to place that science in context and ensure that its understood properly, you're going to need philosophy that's more sophisticated than "<mask><mask>,<mask> I am." <mask> are the necessary criteria to satisfy before we claim to know something? <mask> does that knowing actually mean?  Do we know the truth of an untestable theory that is nevertheless consistent with everything we observe about the universe? [NEWLINE] [NEWLINE] You can dismiss that study by saying "I have standard X that functions well enough for me!",<mask> that's just you declining to make the effort and condoning your own ignorance.  Others can make that effort for you and provide you with the necessary arguments and sets of principles that underpin your work. [NEWLINE] [NEWLINE] Compartmentalization of labor, FTW. [NEWLINE] [NEWLINE] [STARTQ] You shouldn't need to study philosophy to cultivate a reasonable set of moral principles [ENDQ] [NEWLINE] You may not need to study philosophy to establish a set of moral principles *that satisfy you*,<mask> that's obvious. <mask> you want a set of morals that are consistent and rational, you're going to need to subject them to scrutiny that is essentially embodied in moral philosophy.  You can do this on your own, or you can study the collected academic reflections on the subject compiled over a couple thousand years.  In the same way, you can study astronomy by naming stars and constellations on your own or study established work and add to it.</s>
Label encoding: <s> [STARTQ] You don't need to study philosophy to become a capable scientist. [ENDQ] [NEWLINE] That may be true, but if you want to place that science in context and ensure that its understood properly, you're going to need philosophy that's more sophisticated than " I think, therefore I am."  What are the necessary criteria to satisfy before we claim to know something?  What does that knowing actually mean?  Do we know the truth of an untestable theory that is nevertheless consistent with everything we observe about the universe? [NEWLINE] [NEWLINE] You can dismiss that study by saying "I have standard X that functions well enough for me!", but that's just you declining to make the effort and condoning your own ignorance.  Others can make that effort for you and provide you with the necessary arguments and sets of principles that underpin your work. [NEWLINE] [NEWLINE] Compartmentalization of labor, FTW. [NEWLINE] [NEWLINE] [STARTQ] You shouldn't need to study philosophy to cultivate a reasonable set of moral principles [ENDQ] [NEWLINE] You may not need to study philosophy to establish a set of moral principles *that satisfy you*, but that's obvious.  If you want a set of morals that are consistent and rational, you're going to need to subject them to scrutiny that is essentially embodied in moral philosophy.  You can do this on your own, or you can study the collected academic reflections on the subject compiled over a couple thousand years.  In the same way, you can study astronomy by naming stars and constellations on your own or study established work and add to it.</s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We should make welfare much harder to get<mask> it is allowing large corporations to pay their employees $7 an hour<mask> they can go out and get welfare to supplement their income. I do think that it would get pretty bad for those people for a couple of years,<mask> businesses would be reluctant to raise wages,<mask><mask> we made welfare much much harder to be eligible for, businesses would be forced to pay their employees more<mask> without any kind of social programs to fall back on, people simply couldn't work for the amount they are currently being paid. [NEWLINE] [NEWLINE] <mask> I say that "Welfare" should be much harder to get, I'm talking about supplemental help, not disability or anything like that. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>We should make welfare much harder to get because it is allowing large corporations to pay their employees $7 an hour because they can go out and get welfare to supplement their income. I do think that it would get pretty bad for those people for a couple of years, as businesses would be reluctant to raise wages, but if we made welfare much much harder to be eligible for, businesses would be forced to pay their employees more because without any kind of social programs to fall back on, people simply couldn't work for the amount they are currently being paid. [NEWLINE] [NEWLINE] When I say that "Welfare" should be much harder to get, I'm talking about supplemental help, not disability or anything like that. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As someone who has used both extensively, I will be speaking mostly from personal experience. [NEWLINE] [NEWLINE] 1) Customization (or at least ease of): The Android has widgets that can be placed on the home page, the keyboard on the Android OS is more easily swappable and customizable, the Android supports emulators, and it's easier to set custom ring tones. [NEWLINE] [NEWLINE] 2) Software:<mask> novice phone users won't be able to get the most out of the Android system, the apple OS isn't<mask> versatile or easy to play around with. For example, most iPhones require syncing through iTunes, whereas the Android is a simple mount, drag and drop. The iPhone doesn't allow for the installation of custom ROMS which not only does<mask> jailbreaking does,<mask> at a much deeper level. [NEWLINE] [NEWLINE] 3) Variability: Android phones are not just limited to the Galaxy, there are dozens of different devices to suit you, whereas iPhones force you to cater to it's design rather than the other way around.<mask>, Google Play has actually caught up with the iTunes store in terms of apps and there are more free apps in the former. On top of that, iPhones don't vary greatly from generation from generation, whereas Android's do. [NEWLINE] [NEWLINE] 4) Design: Personal preference,<mask> I prefer the design and feel of the Galaxy, it has a bigger screen, more powerful hardware, and is lighter than their Apple counterpart. [NEWLINE] [NEWLINE] I'm sure people can add more,<mask> those are my reasons for my preference.</s>
Label encoding: <s>As someone who has used both extensively, I will be speaking mostly from personal experience. [NEWLINE] [NEWLINE] 1) Customization (or at least ease of): The Android has widgets that can be placed on the home page, the keyboard on the Android OS is more easily swappable and customizable, the Android supports emulators, and it's easier to set custom ring tones. [NEWLINE] [NEWLINE] 2) Software: Though novice phone users won't be able to get the most out of the Android system, the apple OS isn't as versatile or easy to play around with. For example, most iPhones require syncing through iTunes, whereas the Android is a simple mount, drag and drop. The iPhone doesn't allow for the installation of custom ROMS which not only does what jailbreaking does, but at a much deeper level. [NEWLINE] [NEWLINE] 3) Variability: Android phones are not just limited to the Galaxy, there are dozens of different devices to suit you, whereas iPhones force you to cater to it's design rather than the other way around. Also, Google Play has actually caught up with the iTunes store in terms of apps and there are more free apps in the former. On top of that, iPhones don't vary greatly from generation from generation, whereas Android's do. [NEWLINE] [NEWLINE] 4) Design: Personal preference, but I prefer the design and feel of the Galaxy, it has a bigger screen, more powerful hardware, and is lighter than their Apple counterpart. [NEWLINE] [NEWLINE] I'm sure people can add more, but those are my reasons for my preference.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like to be brief,<mask> it does specify 500 characters. Basically, to give an example: [NEWLINE] [NEWLINE] *Maybe* you're talking about the *preferred gender pronouns*<mask> you *actually* hold a *sexist view*. [NEWLINE] [NEWLINE] I'm not quite sure<mask> it irritates me<mask> much,<mask> whenever I see a comment like that, particularly<mask> things like feminism or gender politics are discussed on here, it turns my stomach and makes me think the person writing it has swallowed a bunch of asterisks. [NEWLINE] [NEWLINE] I'm open to this view changing, it's quite a benign one. It just rubs me up the wrong way and *I'd love to change it*. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I like to be brief, but it does specify 500 characters. Basically, to give an example: [NEWLINE] [NEWLINE] *Maybe* you're talking about the *preferred gender pronouns* because you *actually* hold a *sexist view*. [NEWLINE] [NEWLINE] I'm not quite sure why it irritates me so much, but whenever I see a comment like that, particularly when things like feminism or gender politics are discussed on here, it turns my stomach and makes me think the person writing it has swallowed a bunch of asterisks. [NEWLINE] [NEWLINE] I'm open to this view changing, it's quite a benign one. It just rubs me up the wrong way and *I'd love to change it*. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Now I know that the enforcement argument is usually a weak one,<mask><mask><mask> someone didn't know that they were pregnant?<mask><mask> they were not planning on carrying the pregnancy to term? There would be no way to make those people exempt. [ENDQ] [NEWLINE] My response to enforcing a law like that is<mask> is it any difference from other laws? The police have to be present to witness the crime, and they write the citation. Same for speeding or other traffic violations, MIPs and other banned-substances, etc. Obviously<mask> a woman is not visibly pregnant, you can't cite her for it. It would be a judgement call on the cop (after stating this, I can definitely see overweight women being a problem in this regard). I know we would not catch everyone,<mask> even have the law on record and occasionally enforced can be a deterrant in and of itself. [NEWLINE] [NEWLINE] [STARTQ] Even with a fine, not everyone can afford to pay a fine, and jail time is the alternative<mask> you ca't afford it.<mask> you know that this is infringing on someone's rights over their own body and making decisions that are best for them then that alone should be a good enough reason. [ENDQ] [NEWLINE] I can only speak for my own district,<mask><mask> I live, they usually won't throw people in jail for fine, the fine just stays on your record indefinitely and could prevent you from obtaining legal documents you may need such<mask> an I.D. or registration, etc. [NEWLINE] [NEWLINE] edit: addressing the last bit</s>
Label encoding: <s> [STARTQ] Now I know that the enforcement argument is usually a weak one, but what if someone didn't know that they were pregnant? What if they were not planning on carrying the pregnancy to term? There would be no way to make those people exempt. [ENDQ] [NEWLINE] My response to enforcing a law like that is why is it any difference from other laws? The police have to be present to witness the crime, and they write the citation. Same for speeding or other traffic violations, MIPs and other banned-substances, etc. Obviously if a woman is not visibly pregnant, you can't cite her for it. It would be a judgement call on the cop (after stating this, I can definitely see overweight women being a problem in this regard). I know we would not catch everyone, but even have the law on record and occasionally enforced can be a deterrant in and of itself. [NEWLINE] [NEWLINE] [STARTQ] Even with a fine, not everyone can afford to pay a fine, and jail time is the alternative if you ca't afford it. If you know that this is infringing on someone's rights over their own body and making decisions that are best for them then that alone should be a good enough reason. [ENDQ] [NEWLINE] I can only speak for my own district, but where I live, they usually won't throw people in jail for fine, the fine just stays on your record indefinitely and could prevent you from obtaining legal documents you may need such as an I.D. or registration, etc. [NEWLINE] [NEWLINE] edit: addressing the last bit</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My apologies, I only now saw this response. [NEWLINE] [NEWLINE] Thanks for sharing the link.<mask> I'm interested in seeing the results of the study, I smell a bit of confirmation bias present in a study that's called "scofflaw biking".<mask> we'll see<mask> comes. [NEWLINE] [NEWLINE] [STARTQ] 1) Cars are much more readily identifiable<mask> they have license plates,<mask> bicycles (especially<mask> the proliferation of CitiBikes) are much more anonymous. And<mask> we know, anonymity doesn't do good things to behavior.... [ENDQ] [NEWLINE] This may turn out to be a real issue in the future,<mask> based on case studies that we have at present (eurozone cities with much higher cycle use per capita) it doesn't look like it will be a major issue.<mask> again, speculation. [NEWLINE] [NEWLINE] [STARTQ] 2) Bicycles are trying to navigate around an infrastructure that is not designed for them. They are neither fish nor fowl, neither a motor vehicle nor pedestrian. [ENDQ] [NEWLINE] This is true.<mask> to say the solution is that they should leave until we build bike roads is silly. And<mask> anyone who has tried to navigate streets on a bike<mask><mask> they are a car knows, it's not a practical solution, either. Until the infrastructure is built, we need laws that protect cyclists and benefit all traffic. My suggestion is that this law,<mask> it's written (not<mask> it's misinterpreted by many people) is a great solution to help protect cyclists and ease any congestion that the cyclists themselves may create.</s>
Label encoding: <s>My apologies, I only now saw this response. [NEWLINE] [NEWLINE] Thanks for sharing the link. While I'm interested in seeing the results of the study, I smell a bit of confirmation bias present in a study that's called "scofflaw biking". But we'll see what comes. [NEWLINE] [NEWLINE] [STARTQ] 1) Cars are much more readily identifiable because they have license plates, but bicycles (especially since the proliferation of CitiBikes) are much more anonymous. And as we know, anonymity doesn't do good things to behavior.... [ENDQ] [NEWLINE] This may turn out to be a real issue in the future, but based on case studies that we have at present (eurozone cities with much higher cycle use per capita) it doesn't look like it will be a major issue. But again, speculation. [NEWLINE] [NEWLINE] [STARTQ] 2) Bicycles are trying to navigate around an infrastructure that is not designed for them. They are neither fish nor fowl, neither a motor vehicle nor pedestrian. [ENDQ] [NEWLINE] This is true. But to say the solution is that they should leave until we build bike roads is silly. And as anyone who has tried to navigate streets on a bike as if they are a car knows, it's not a practical solution, either. Until the infrastructure is built, we need laws that protect cyclists and benefit all traffic. My suggestion is that this law, as it's written (not as it's misinterpreted by many people) is a great solution to help protect cyclists and ease any congestion that the cyclists themselves may create.</s>
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you kidding me?! Black face was, is, and always will be about stereotyping an entire race to one idiotic archetype. [NEWLINE] [NEWLINE] Drag shows started out similar to blackface, I'll grant you that,<mask> it quickly became a way for homosexual and transgender men to be openly transgender/homosexual without fear of police retribution during the era of "decency laws"<mask> the heterosexual norm was enforced by fucking HANGINGS. [NEWLINE] [NEWLINE] It remained a way for homosexual men to be openly homosexual without fear of repercussions for decades. Hell,<mask> not for drag shows some states out west wouldn't have ever removed their bigoted laws<mask> early<mask> they did. [NEWLINE] [NEWLINE] Don't you dare act like they're the same, that's just a small minded interpretation of an institution that kept homosexual men and women ALIVE for decades. Many shows<mask> the only places for gay and trans people to safely be themselves. AND THE SHOWS CONTINUE TO BE. I've seen closeted men and women cry over the feeling of acceptance they received<mask> back stage at a drag show. I can firmly say I've never heard of an African American comfortable at a black face show. [NEWLINE] [NEWLINE] Frankly, it's irritating little shits like you who make snap judgements based on limited experience that piss me off the most. [NEWLINE] [NEWLINE] YES. Some bigots use drag shows<mask> proof of their bigotry,<mask> these shows have done infinitely more for the LGBT community over the last century than your little Tumblr based community has in its' existence.</s>
Label encoding: <s>Are you kidding me?! Black face was, is, and always will be about stereotyping an entire race to one idiotic archetype. [NEWLINE] [NEWLINE] Drag shows started out similar to blackface, I'll grant you that, but it quickly became a way for homosexual and transgender men to be openly transgender/homosexual without fear of police retribution during the era of "decency laws" when the heterosexual norm was enforced by fucking HANGINGS. [NEWLINE] [NEWLINE] It remained a way for homosexual men to be openly homosexual without fear of repercussions for decades. Hell, if not for drag shows some states out west wouldn't have ever removed their bigoted laws as early as they did. [NEWLINE] [NEWLINE] Don't you dare act like they're the same, that's just a small minded interpretation of an institution that kept homosexual men and women ALIVE for decades. Many shows where the only places for gay and trans people to safely be themselves. AND THE SHOWS CONTINUE TO BE. I've seen closeted men and women cry over the feeling of acceptance they received while back stage at a drag show. I can firmly say I've never heard of an African American comfortable at a black face show. [NEWLINE] [NEWLINE] Frankly, it's irritating little shits like you who make snap judgements based on limited experience that piss me off the most. [NEWLINE] [NEWLINE] YES. Some bigots use drag shows as proof of their bigotry, but these shows have done infinitely more for the LGBT community over the last century than your little Tumblr based community has in its' existence.</s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is /r/changemyview, not /r/changemyviewtothis<mask> I will ask you to consider that you might actually be in need of a different kind of change in perspective. [NEWLINE] The problem isn't that you have a lack of belief in God, it's that you percieve this to be somehow detrimental to your well-being. Your view is that a world *with* God gives you purpose, a feeling of being loved, safety etc. [NEWLINE] [NEWLINE] The idea of an infinite godless universe might seem bleak and harrowing, leaving you in a despondent state of mind.<mask><mask> should this be? Sure the delusion of God might supply you with a feeling of destiny and alleviate mortal concerns,<mask> it<mask> imposes an infinite amount of rules and boundaries to your life. [NEWLINE] [NEWLINE] The universe for an atheist is an oyster of opportunity. The universe wont stop you from doing *anything* (that doesn't break the laws of physics). Technology can bring about the paradise that the bible promises x10^100. The universe doesn't judge you or condemn you to hell for X, Y and Z. The universe didn't decide who you are to be, only you can choose that. [NEWLINE] [NEWLINE] You need to start thinking that perhaps being a mature being in 2015 isn't all that bad. Civillization has exited the cradle and started exploring the world. We still got a long way to go,<mask> at least we have exposed that there is no monster under our bed.</s>
Label encoding: <s>This is /r/changemyview, not /r/changemyviewtothis so I will ask you to consider that you might actually be in need of a different kind of change in perspective. [NEWLINE] The problem isn't that you have a lack of belief in God, it's that you percieve this to be somehow detrimental to your well-being. Your view is that a world *with* God gives you purpose, a feeling of being loved, safety etc. [NEWLINE] [NEWLINE] The idea of an infinite godless universe might seem bleak and harrowing, leaving you in a despondent state of mind. But why should this be? Sure the delusion of God might supply you with a feeling of destiny and alleviate mortal concerns, but it also imposes an infinite amount of rules and boundaries to your life. [NEWLINE] [NEWLINE] The universe for an atheist is an oyster of opportunity. The universe wont stop you from doing *anything* (that doesn't break the laws of physics). Technology can bring about the paradise that the bible promises x10^100. The universe doesn't judge you or condemn you to hell for X, Y and Z. The universe didn't decide who you are to be, only you can choose that. [NEWLINE] [NEWLINE] You need to start thinking that perhaps being a mature being in 2015 isn't all that bad. Civillization has exited the cradle and started exploring the world. We still got a long way to go, but at least we have exposed that there is no monster under our bed.</s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know I'm late to the party,<mask> I thought I'd give my two cents. [NEWLINE] [NEWLINE] My real issue with solely giving it to charities is that many of them are geared towards women and children. This makes sense, of course,<mask> they are the most victimized groups,<mask><mask> many former homeless people can attest, it is not very easy to get accommodations being a man on the street. [NEWLINE] [NEWLINE] A few years ago I would have agreed with your mentality, and I still would had it not been for  a little story I have about my great-grandmother my aunt told me. Back in the 1980s my great-grandmother was living in an apartment in NYC. There was a homeless man she saw on her daily route, and she gave him a small amount of money every day, about $2 usually. She didn't think much of it,<mask> the homeless man was always grateful. [NEWLINE] [NEWLINE] Two decades later, my great-grandmother was in her late 90s and being escorted by my aunt on the subway. A man came up to her in a suit and said to my aunt, "I wanted to thank this woman." He told my aunt the story, and<mask> thanks to that money he saved up, bought an apartment and found a job. Now he was a family, all thanks to my great-grandma. [NEWLINE] [NEWLINE] I don't know<mask> anyone's going to see this story,<mask> fuck it it's one of my favorite stories to tell and it certainly changed me. [NEWLINE] </s>
Label encoding: <s>I know I'm late to the party, but I thought I'd give my two cents. [NEWLINE] [NEWLINE] My real issue with solely giving it to charities is that many of them are geared towards women and children. This makes sense, of course, as they are the most victimized groups, but as many former homeless people can attest, it is not very easy to get accommodations being a man on the street. [NEWLINE] [NEWLINE] A few years ago I would have agreed with your mentality, and I still would had it not been for  a little story I have about my great-grandmother my aunt told me. Back in the 1980s my great-grandmother was living in an apartment in NYC. There was a homeless man she saw on her daily route, and she gave him a small amount of money every day, about $2 usually. She didn't think much of it, but the homeless man was always grateful. [NEWLINE] [NEWLINE] Two decades later, my great-grandmother was in her late 90s and being escorted by my aunt on the subway. A man came up to her in a suit and said to my aunt, "I wanted to thank this woman." He told my aunt the story, and how thanks to that money he saved up, bought an apartment and found a job. Now he was a family, all thanks to my great-grandma. [NEWLINE] [NEWLINE] I don't know if anyone's going to see this story, but fuck it it's one of my favorite stories to tell and it certainly changed me. [NEWLINE] </s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I recently got into a debate with my sister about the real world application of Philosophy<mask> a field of study.<mask> I started off believing that it has very tangible applications in the career world, I began becoming less and less sure<mask> the discussion progressed. Her argument is that it is not a field of study that you can find a career in unless it is pursuing a career<mask> a philosopher or other academic field of study<mask> I believed there were more real world applications for a career in philosophy<mask><mask> I did<mask><mask> it is a beneficial course to take<mask> wanting to go to Law School to become a lawyer, she<mask> stated it was not necessary to pursue Philosophy<mask> a field of study to get into Law School. She thinks it's for argumentative nuts like me who just want to sit around and debate about the world without actually affecting it.<mask> I feel like the core focus of Philosophy is not necessarily learning theories,<mask><mask> to present a theory and being able to debate for that theory (not a scientific theory,<mask> philosophical theory &lt;--- please understand the difference before posting). I have a hard time believing that philosophy is that useless of a subject<mask><mask> the more my sister brings these points the more I'm becoming convinced. I'm about to believe Philosophy is a useless field of study with no career path except in academia, CMV! [NEWLINE] [NEWLINE] EDIT: I've been convinced it is useful.<mask> subjects is it NECESSARY to specialize in Philosophy<mask> a career path instead of the previously mentioned plus computer science?</s>
Label encoding: <s>I recently got into a debate with my sister about the real world application of Philosophy as a field of study. While I started off believing that it has very tangible applications in the career world, I began becoming less and less sure as the discussion progressed. Her argument is that it is not a field of study that you can find a career in unless it is pursuing a career as a philosopher or other academic field of study while I believed there were more real world applications for a career in philosophy but while I did argue that it is a beneficial course to take if wanting to go to Law School to become a lawyer, she also stated it was not necessary to pursue Philosophy as a field of study to get into Law School. She thinks it's for argumentative nuts like me who just want to sit around and debate about the world without actually affecting it. While I feel like the core focus of Philosophy is not necessarily learning theories, but how to present a theory and being able to debate for that theory (not a scientific theory, but philosophical theory &lt;--- please understand the difference before posting). I have a hard time believing that philosophy is that useless of a subject since but the more my sister brings these points the more I'm becoming convinced. I'm about to believe Philosophy is a useless field of study with no career path except in academia, CMV! [NEWLINE] [NEWLINE] EDIT: I've been convinced it is useful. What subjects is it NECESSARY to specialize in Philosophy as a career path instead of the previously mentioned plus computer science?</s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not about being communist, it's about whether or not it really serves a positive purpose.<mask>'s wrong with people buying something at an agreed upon price,<mask><mask> the profit? People are willing to spend<mask> they feel a product or service is worth. Sometimes that means the profit margins are huge and sometimes it means they are small. For example, you can easily pay $5 for a bottle of water at a sporting event that costed the stadium a dime to purchase. You pay for the convenience. In financial terms, the seller doesn't have to spend any more money, in this instance, to sell convenience,<mask> the buyer is willing to pay for it.<mask> not allow that to happen? [NEWLINE] [NEWLINE] It can be incredibly difficult to determine<mask> much a product costed a seller. Sometimes the costs are in research and development. For example, drugs and medicine often cost very little to manufacture.<mask>, the research and development of these products takes long and costs a lot.<mask><mask><mask>, we experience huge markups beyond the manufacturing cost. [NEWLINE] [NEWLINE] <mask> keep in mind, many of the<mask> called "evil" corporations that you would want to protect society from actually function off relatively small profit margins. For example, stores like Warmart receive a lot of hate,<mask> they have incredibly small profit margins. [NEWLINE] [NEWLINE] It really all comes down to<mask> people are willing to spend.<mask> I want to purchase something for $10<mask><mask> it only costed 5 cents to produce,<mask> not allow me?</s>
Label encoding: <s>It's not about being communist, it's about whether or not it really serves a positive purpose. What's wrong with people buying something at an agreed upon price, regardless of the profit? People are willing to spend what they feel a product or service is worth. Sometimes that means the profit margins are huge and sometimes it means they are small. For example, you can easily pay $5 for a bottle of water at a sporting event that costed the stadium a dime to purchase. You pay for the convenience. In financial terms, the seller doesn't have to spend any more money, in this instance, to sell convenience, but the buyer is willing to pay for it. Why not allow that to happen? [NEWLINE] [NEWLINE] It can be incredibly difficult to determine how much a product costed a seller. Sometimes the costs are in research and development. For example, drugs and medicine often cost very little to manufacture. However, the research and development of these products takes long and costs a lot. As a result, we experience huge markups beyond the manufacturing cost. [NEWLINE] [NEWLINE] Also keep in mind, many of the so called "evil" corporations that you would want to protect society from actually function off relatively small profit margins. For example, stores like Warmart receive a lot of hate, but they have incredibly small profit margins. [NEWLINE] [NEWLINE] It really all comes down to what people are willing to spend. If I want to purchase something for $10 even though it only costed 5 cents to produce, why not allow me?</s>
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Not necessarily, or in most cases. Agnosticism argues that it is not possible at a given moment in time to know absolutely,<mask> then we don't know anything absolutely.<mask>, that that's okay, we'll work with<mask> we have. [ENDQ] [STARTQ] &gt;That's kind of a flimsy sort of agnosticism. Just saying "well, we're not ABSOLUTELY SURE" is something that<mask><mask> pretty much any really honest person, religious or atheist, will admit.<mask> that's all agnosticism means, it's a label that can be applied to pretty much anyone. [ENDQ] [NEWLINE] well yes, essentially,<mask> is flimsy a bad thing?<mask> on the question of god it can<mask> be applied to the standards of evidence argument presented earlier. That is, there is insufficient evidence to significantly attribute "god" more reason for existence than anything else. That's<mask> it's not insignificant to say absolute knowledge seems impossible at a given time. You can have the position<mask> still making significant statements. [NEWLINE] [NEWLINE] [STARTQ] This is<mask> you'd have to demonstrate,<mask> I haven't<mask> come across any supposed evidence of a God that isn't equally or more explainable by some other cause. [ENDQ] [STARTQ] &gt;The first tack he takes at it is simple enough; the apparent existence of a universal morality. [ENDQ] [NEWLINE] Moral realism is actually a very open question in philosophy.<mask>,<mask> would a universal morality necessarily imply the existence of a god?</s>
Label encoding: <s> [STARTQ] Not necessarily, or in most cases. Agnosticism argues that it is not possible at a given moment in time to know absolutely, but then we don't know anything absolutely. Moreover, that that's okay, we'll work with what we have. [ENDQ] [STARTQ] &gt;That's kind of a flimsy sort of agnosticism. Just saying "well, we're not ABSOLUTELY SURE" is something that I think pretty much any really honest person, religious or atheist, will admit. If that's all agnosticism means, it's a label that can be applied to pretty much anyone. [ENDQ] [NEWLINE] well yes, essentially, why is flimsy a bad thing? but on the question of god it can also be applied to the standards of evidence argument presented earlier. That is, there is insufficient evidence to significantly attribute "god" more reason for existence than anything else. That's why it's not insignificant to say absolute knowledge seems impossible at a given time. You can have the position while still making significant statements. [NEWLINE] [NEWLINE] [STARTQ] This is what you'd have to demonstrate, because I haven't yet come across any supposed evidence of a God that isn't equally or more explainable by some other cause. [ENDQ] [STARTQ] &gt;The first tack he takes at it is simple enough; the apparent existence of a universal morality. [ENDQ] [NEWLINE] Moral realism is actually a very open question in philosophy. Moreover, how would a universal morality necessarily imply the existence of a god?</s>
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To expand on that: It is still a social expectation that men value women's lives more than their own. There is a reason "women and children first" is an idea, there is a reason a "gentleman" walks on the left side of a woman on the sidewalk, there is a reason news reports read "106 dead, including 5 women". To use a recent example, there's a reason the kidnapping of 200 girls in Nigeria makes headline news,<mask> the kidnapping of boys is relatively routine. In our society, women are worth more than men. [NEWLINE] [NEWLINE] A lot of it stems from the view that women are relatively weak and fragile, and conversely men are tough and durable. In a society<mask> just about everyone genuinely believes those things, it makes sense to value women's lives more. I don't agree with it,<mask> it does possess a kind of logic. Today, no one believes that women are any weaker than men,<mask> at least with regards to men, everyone still acts by the old standards. [NEWLINE] [NEWLINE] [STARTQ] Is this not the goal of feminism? Providing choice and self determination based on the individual and not the gender? [ENDQ] [NEWLINE] It is one of the stated goals of feminism,<mask> at least from<mask> I've seen personally there's a hidden caveat "For women first, for men maybe<mask> we get around to it". Feminism,<mask> befits the name, focuses on womens issues.<mask> the MRM tries to get change on Men's issues. </s><pad>
Label encoding: <s>To expand on that: It is still a social expectation that men value women's lives more than their own. There is a reason "women and children first" is an idea, there is a reason a "gentleman" walks on the left side of a woman on the sidewalk, there is a reason news reports read "106 dead, including 5 women". To use a recent example, there's a reason the kidnapping of 200 girls in Nigeria makes headline news, while the kidnapping of boys is relatively routine. In our society, women are worth more than men. [NEWLINE] [NEWLINE] A lot of it stems from the view that women are relatively weak and fragile, and conversely men are tough and durable. In a society where just about everyone genuinely believes those things, it makes sense to value women's lives more. I don't agree with it, but it does possess a kind of logic. Today, no one believes that women are any weaker than men, but at least with regards to men, everyone still acts by the old standards. [NEWLINE] [NEWLINE] [STARTQ] Is this not the goal of feminism? Providing choice and self determination based on the individual and not the gender? [ENDQ] [NEWLINE] It is one of the stated goals of feminism, but at least from what I've seen personally there's a hidden caveat "For women first, for men maybe if we get around to it". Feminism, as befits the name, focuses on womens issues. So the MRM tries to get change on Men's issues. </s><pad>
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, I was hoping you'd just stop<mask> I suppose your desire to look superior overrode whatever common sense you have. By the way, it is entirely justified to say that Columbus discovered America<mask> he was the first European to arrive, claim it for a nation and establish and govern permanent settlements. [NEWLINE] [NEWLINE] Just to go through your replies to my comments. The first,<mask> I mentioned, implies that the South is uniquely justified in its expansion and it ignored my point that the South had repeatedly tried to impose Slavery, the job of collecting runaways and several wars upon the North. [NEWLINE] [NEWLINE] The second explicitly states that the arrival of the Republican Party in 1856 led to the destruction of the Democratic Party (presumably in the North). This ignores the tremendous North/South split that occurred within the Democrats over the issue of Kansas/Nebraska and popular sovereignty.<mask> that the Democrats won the 1856 election. Oh, and that the Whig Party collapsed over the issue of Slavery after the 1852 election. You then go on to be ignorant of the slave states within the Union, Maryland was occupied at the beginning of the War to protect Washington, Kentucky saw some fighting between two rival Governments and Missouri faced constant conflict from within. [NEWLINE] [NEWLINE] The final comment above provides the nonsensical word "oppression"<mask> the cause of the Civil War. You have thrown away a hundred years of academic consensus and replaced it with a word you yourself think is awful. I don't know<mask> to say. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Ok, I was hoping you'd just stop but I suppose your desire to look superior overrode whatever common sense you have. By the way, it is entirely justified to say that Columbus discovered America since he was the first European to arrive, claim it for a nation and establish and govern permanent settlements. [NEWLINE] [NEWLINE] Just to go through your replies to my comments. The first, as I mentioned, implies that the South is uniquely justified in its expansion and it ignored my point that the South had repeatedly tried to impose Slavery, the job of collecting runaways and several wars upon the North. [NEWLINE] [NEWLINE] The second explicitly states that the arrival of the Republican Party in 1856 led to the destruction of the Democratic Party (presumably in the North). This ignores the tremendous North/South split that occurred within the Democrats over the issue of Kansas/Nebraska and popular sovereignty. Also that the Democrats won the 1856 election. Oh, and that the Whig Party collapsed over the issue of Slavery after the 1852 election. You then go on to be ignorant of the slave states within the Union, Maryland was occupied at the beginning of the War to protect Washington, Kentucky saw some fighting between two rival Governments and Missouri faced constant conflict from within. [NEWLINE] [NEWLINE] The final comment above provides the nonsensical word "oppression" as the cause of the Civil War. You have thrown away a hundred years of academic consensus and replaced it with a word you yourself think is awful. I don't know what to say. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, climate change is scary. Let's talk about some much larger, ultimately scarier issues than climate change that you haven't heard about<mask> much<mask> they haven't been<mask> politicized<mask>. [NEWLINE] [NEWLINE] <mask> about the global depletion of our water resources ( [URL] /)? [NEWLINE] Climate change will do just that: change the climate. It will make some areas more agriculturally prosperous and some less<mask>. There are predictions out there for whether globally it will become better or worse<mask> in general that is still somewhat of an unknown in the research world<mask><mask><mask> I know. The depletion of our aquifers and water resources<mask> have very known, very serious consequences. [NEWLINE] [NEWLINE] Another one,<mask> about the rapid depletion of top soil ( [URL] /)? Quoting the article, "A rough calculation of current rates of soil degradation suggests we have about 60 years of topsoil left." Well fuck. That's really, really bad.<mask> we don't invent/conserve our way out of that then it isn't really going to matter for us humans<mask> the climate does change over the next 100 years. [NEWLINE] [NEWLINE] The lack of attention these issues receive and the great amount issues like climate change gets is one thing that makes people who are knowledgeable about agriculture (country folk) scoff at climate change<mask> just some hot-button political issue that only exists to divide people on party lines. Obviously climate change is a big issue. I would argue,<mask>, that these other issues are much, much more dire.</s>
Label encoding: <s>Yes, climate change is scary. Let's talk about some much larger, ultimately scarier issues than climate change that you haven't heard about as much because they haven't been as politicized though. [NEWLINE] [NEWLINE] How about the global depletion of our water resources ( [URL] /)? [NEWLINE] Climate change will do just that: change the climate. It will make some areas more agriculturally prosperous and some less so. There are predictions out there for whether globally it will become better or worse but in general that is still somewhat of an unknown in the research world as far as I know. The depletion of our aquifers and water resources though have very known, very serious consequences. [NEWLINE] [NEWLINE] Another one, how about the rapid depletion of top soil ( [URL] /)? Quoting the article, "A rough calculation of current rates of soil degradation suggests we have about 60 years of topsoil left." Well fuck. That's really, really bad. If we don't invent/conserve our way out of that then it isn't really going to matter for us humans if the climate does change over the next 100 years. [NEWLINE] [NEWLINE] The lack of attention these issues receive and the great amount issues like climate change gets is one thing that makes people who are knowledgeable about agriculture (country folk) scoff at climate change as just some hot-button political issue that only exists to divide people on party lines. Obviously climate change is a big issue. I would argue, however, that these other issues are much, much more dire.</s>
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When people make contracts they are agreeing to a meaning. One of the great tasks of contract writing is finding language that is precise enough to be unambiguous to that meaning<mask> ambiguous enough to not define things you would like to be flexible. [NEWLINE] [NEWLINE] I rent my basement to a tenant. I need to be unambiguous with my lease in defining the leased space, the price, the due date, the late fee, the included appliances, the form of payment, pets, etc. I can<mask> choose to be ambiguous in some places, like "keep noise to a reasonable level." Otherwise,<mask> I just say rent is due monthly, there are 30 days of ambiguity. [NEWLINE] [NEWLINE] <mask> a contract is in effect, it's perfectly fine to modify it<mask> the parties agree to it. That's<mask> an amendment is for.<mask><mask><mask><mask>, to will the existing contract to take on a new meaning<mask> times have changed is dishonest. [NEWLINE] [NEWLINE] <mask> you object to being under a contract that you never signed and wouldn't sign<mask> given the option today, then just say that. I don't think that the meetings and decisions that pasty, powdered-wigged men made hundreds of years ago should have any bearing on my life just<mask> I was born between a set of imaginary lines they drew up. No need to perform hermanutical gymnastics to an old document<mask> you could instead just oppose the document itself, or just,<mask> best<mask> you are able, live<mask><mask> it does not exist.</s>
Label encoding: <s>When people make contracts they are agreeing to a meaning. One of the great tasks of contract writing is finding language that is precise enough to be unambiguous to that meaning but ambiguous enough to not define things you would like to be flexible. [NEWLINE] [NEWLINE] I rent my basement to a tenant. I need to be unambiguous with my lease in defining the leased space, the price, the due date, the late fee, the included appliances, the form of payment, pets, etc. I can also choose to be ambiguous in some places, like "keep noise to a reasonable level." Otherwise, if I just say rent is due monthly, there are 30 days of ambiguity. [NEWLINE] [NEWLINE] If a contract is in effect, it's perfectly fine to modify it if the parties agree to it. That's what an amendment is for. On the other hand, to will the existing contract to take on a new meaning because times have changed is dishonest. [NEWLINE] [NEWLINE] If you object to being under a contract that you never signed and wouldn't sign if given the option today, then just say that. I don't think that the meetings and decisions that pasty, powdered-wigged men made hundreds of years ago should have any bearing on my life just because I was born between a set of imaginary lines they drew up. No need to perform hermanutical gymnastics to an old document if you could instead just oppose the document itself, or just, as best as you are able, live as if it does not exist.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Pre-licensing statistics will try to get information on the incidence of side effects and the severity.<mask> they might find for instance, 1 in 10,000 normal (i.e. without health conditions which would preclude vaccine use) people have side effects like a headache. And 1 in a 100,000 will have serious side effects like vertigo.<mask> the serious side effects are e.g. heart failure or liver damage, the vaccine won't go to market. [NEWLINE] [NEWLINE] It is then up to the pharmaceutical company to decide whether to put the product up for licensing  and the licensing board to have the final say - is the potential risk worth the benefit. They then follow the drug after it's been released to see<mask> there are any effects (say in a particular ethnic group who might not have have been represented in the clinical trials) that haven't been noted. [NEWLINE] [NEWLINE] Now you were sort of at no more at risk than anyone else and more at risk. Think of it like a peanut allergy - you are more at risk of reacting to peanuts,<mask> within the population in general your risk of having a peanut allergy is no greater than anyone else's. And there's no way to tell until you eat a peanut. [NEWLINE] [NEWLINE] Carrying the metaphor to your next point - not wanting your friends to get this vaccination is like not wanting them to eat peanuts<mask> you had an allergic reaction. Yes, peanuts are harmful to you<mask> their risk of having a peanut allergy is still incredibly low.</s>
Label encoding: <s>Pre-licensing statistics will try to get information on the incidence of side effects and the severity. So they might find for instance, 1 in 10,000 normal (i.e. without health conditions which would preclude vaccine use) people have side effects like a headache. And 1 in a 100,000 will have serious side effects like vertigo. If the serious side effects are e.g. heart failure or liver damage, the vaccine won't go to market. [NEWLINE] [NEWLINE] It is then up to the pharmaceutical company to decide whether to put the product up for licensing  and the licensing board to have the final say - is the potential risk worth the benefit. They then follow the drug after it's been released to see if there are any effects (say in a particular ethnic group who might not have have been represented in the clinical trials) that haven't been noted. [NEWLINE] [NEWLINE] Now you were sort of at no more at risk than anyone else and more at risk. Think of it like a peanut allergy - you are more at risk of reacting to peanuts, but within the population in general your risk of having a peanut allergy is no greater than anyone else's. And there's no way to tell until you eat a peanut. [NEWLINE] [NEWLINE] Carrying the metaphor to your next point - not wanting your friends to get this vaccination is like not wanting them to eat peanuts because you had an allergic reaction. Yes, peanuts are harmful to you but their risk of having a peanut allergy is still incredibly low.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do we really know<mask> internet will be in 2020? Some areas could have progressed<mask> others stagnate. Rural areas will always be behind the urban areas. Mobile &amp; non-mobile internet are different. [ENDQ] [NEWLINE] We can look at the past and compare it to now to extrapolate. Without the net neutrality laws and other things to protect competition and consumers we can see the trend that ISPs are following we can tell that instead of building out their networks ISPs will begin implementing data caps. That mobile carriers will continue with their data caps. We can take the mobile carriers and ISPs at their word that without the net neutrality rules stopping them, they will seek out and make deals to give a *fast lane* to those who will pay them, necessarily slowing down those who do not. It's not difficult to see this trend and it's not difficult to tell that unless something is done this is<mask> is going to happen. [NEWLINE] [NEWLINE] [STARTQ] Look at coverage maps of the providers and take a look at the mountain states. There are small towns there with populations under 1000. [ENDQ] [NEWLINE] Exactly, I'd<mask><mask> we should incentivize providing coverage to these areas. [NEWLINE] [NEWLINE] [STARTQ] The issue isn't speed itself<mask> companies playing favorites and trying to force you to use/not use something. [ENDQ] [NEWLINE]...yes, that's my point. The "speed" point is merely the mechanism that ISPs would be using to do this and I believe that it's wrong. That we need net neutrality to prevent it.</s>
Label encoding: <s> [STARTQ] Do we really know how internet will be in 2020? Some areas could have progressed while others stagnate. Rural areas will always be behind the urban areas. Mobile &amp; non-mobile internet are different. [ENDQ] [NEWLINE] We can look at the past and compare it to now to extrapolate. Without the net neutrality laws and other things to protect competition and consumers we can see the trend that ISPs are following we can tell that instead of building out their networks ISPs will begin implementing data caps. That mobile carriers will continue with their data caps. We can take the mobile carriers and ISPs at their word that without the net neutrality rules stopping them, they will seek out and make deals to give a *fast lane* to those who will pay them, necessarily slowing down those who do not. It's not difficult to see this trend and it's not difficult to tell that unless something is done this is what is going to happen. [NEWLINE] [NEWLINE] [STARTQ] Look at coverage maps of the providers and take a look at the mountain states. There are small towns there with populations under 1000. [ENDQ] [NEWLINE] Exactly, I'd argue that we should incentivize providing coverage to these areas. [NEWLINE] [NEWLINE] [STARTQ] The issue isn't speed itself but companies playing favorites and trying to force you to use/not use something. [ENDQ] [NEWLINE]...yes, that's my point. The "speed" point is merely the mechanism that ISPs would be using to do this and I believe that it's wrong. That we need net neutrality to prevent it.</s>
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am<mask> an atheist,<mask> I'm not suggesting we go and worship any kinds of Gods. [NEWLINE] [NEWLINE] <mask> an anecdote; I watch watching the making of the TV series "Rome", and one of the actors was talking about<mask> people behaved back in pre-imperial Rome. He talks about the Gods that the would worship, and<mask> they were constantly being unjust, killing and raping each other. He asked,<mask> those all-powerful beings up there are acting like that,<mask> can we expect to be any better down here? [NEWLINE] [NEWLINE] <mask> lets apply that to the Christian God. Is it possible to have an all-perfect being? I don't think<mask>, personally.<mask>,<mask> we can imagine that it's possible, and hold that being up to the highest standard, then we circumvent the problem of comparing ourselves against unjust Roman Gods, and using that<mask> an excuse to do the vile things that they would do. [NEWLINE] [NEWLINE] <mask> I don't necessarily disagree that in Christian texts, God is more evil (IIRC, God kills over a million people whereas Lucifer only kills about 10),<mask><mask> the point is that people hold themselves against a standard<mask> perfect<mask> possible in order to curb their own bad behaviour. [NEWLINE] [NEWLINE] Of course,<mask> we can see in reality, this doesn't always work necessarily. We all have our own ideas of<mask> God is, and on top of that, I don't believe you need to hold yourself against a divine standard to act morally.</s>
Label encoding: <s>I am also an atheist, so I'm not suggesting we go and worship any kinds of Gods. [NEWLINE] [NEWLINE] As an anecdote; I watch watching the making of the TV series "Rome", and one of the actors was talking about how people behaved back in pre-imperial Rome. He talks about the Gods that the would worship, and how they were constantly being unjust, killing and raping each other. He asked, if those all-powerful beings up there are acting like that, how can we expect to be any better down here? [NEWLINE] [NEWLINE] So lets apply that to the Christian God. Is it possible to have an all-perfect being? I don't think so, personally. But, if we can imagine that it's possible, and hold that being up to the highest standard, then we circumvent the problem of comparing ourselves against unjust Roman Gods, and using that as an excuse to do the vile things that they would do. [NEWLINE] [NEWLINE] Though I don't necessarily disagree that in Christian texts, God is more evil (IIRC, God kills over a million people whereas Lucifer only kills about 10), I think the point is that people hold themselves against a standard as perfect as possible in order to curb their own bad behaviour. [NEWLINE] [NEWLINE] Of course, as we can see in reality, this doesn't always work necessarily. We all have our own ideas of what God is, and on top of that, I don't believe you need to hold yourself against a divine standard to act morally.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's the thing, it's organized play of something that is intended for spectators. It's like the difference between exercising with weights and Olympic weightlifting. Exercising with weights is not a sport,<mask> add a clear set of rules, competition, and spectators and you have a sport. Yes, you don't need spectators every time,<mask> there does ultimately need to be someone watching somewhere and at some time for it to be a sport. [NEWLINE] [NEWLINE] The thing is that there are already intramural sports in colleges. By taking away the official leagues you are simply destroying a layer of competition that would be recreated by professional leagues in those sports<mask> that exists, resulting in competition against instead of collaboration with education. Those sports which don't have professional leagues would lose scholarship and formal equipment funding, which makes participating in any sport much more difficult and would prevent many worthy student-athletes from actually going to college<mask> without absurd amounts of debt. [NEWLINE] [NEWLINE] <mask> we have now is a compromise solution. It's no ideal, and was never meant to be ideal. It's just trying to balance<mask> is best for big schools and small schools, big sports and small sports, and the needs of athletes with those of the institutions and the general populace. There are many short ends of many sticks to be passed around. Choosing not to participate just means that there are an equal proportion of long ends and short ends of the stick in the world, not simply fewer short ends of the stick.</s>
Label encoding: <s>That's the thing, it's organized play of something that is intended for spectators. It's like the difference between exercising with weights and Olympic weightlifting. Exercising with weights is not a sport, but add a clear set of rules, competition, and spectators and you have a sport. Yes, you don't need spectators every time, but there does ultimately need to be someone watching somewhere and at some time for it to be a sport. [NEWLINE] [NEWLINE] The thing is that there are already intramural sports in colleges. By taking away the official leagues you are simply destroying a layer of competition that would be recreated by professional leagues in those sports where that exists, resulting in competition against instead of collaboration with education. Those sports which don't have professional leagues would lose scholarship and formal equipment funding, which makes participating in any sport much more difficult and would prevent many worthy student-athletes from actually going to college if without absurd amounts of debt. [NEWLINE] [NEWLINE] What we have now is a compromise solution. It's no ideal, and was never meant to be ideal. It's just trying to balance what is best for big schools and small schools, big sports and small sports, and the needs of athletes with those of the institutions and the general populace. There are many short ends of many sticks to be passed around. Choosing not to participate just means that there are an equal proportion of long ends and short ends of the stick in the world, not simply fewer short ends of the stick.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm an engineer in the United States who regularly has to work with both Metric and US customary units (which obviously share a lot of units with Imperial).  And<mask> I say I work with both, I mean it.  Just yesterday I was working on a balance spec and the data I received from somebody had the units of gram-inches.  I didn't even think twice about the fact grams is metric and inches in imperial.  I just converted it to the units I needed and moved on. [NEWLINE] [NEWLINE] Obviously imperial measurements aren't useless, otherwise millions of people wouldn't continue to use them every day.  They make sense to a lot of people and that has to count for something.  In other words, millions of people find imperial measurements more practical than metric. [NEWLINE] [NEWLINE] For me, I find Lines per square inch to be much more intuitive than Tesla<mask> I'm dealing with flux density.  <mask> that's just me... I'd never say that one is better than the other.  They're just... different.  (In electromagnetics, there are actually 3 systems of measurement that I need to convert between on a regular basis.) [NEWLINE] [NEWLINE] <mask> honestly it all comes down to<mask> you are used to. <mask> you are used to meters and centimeters, Celsius and kilograms, you are going to find those more practical. <mask><mask> you are used to feet and inches, Fahrenheit and pounds, you are going to find those more practical. [NEWLINE] </s>
Label encoding: <s>I'm an engineer in the United States who regularly has to work with both Metric and US customary units (which obviously share a lot of units with Imperial).  And when I say I work with both, I mean it.  Just yesterday I was working on a balance spec and the data I received from somebody had the units of gram-inches.  I didn't even think twice about the fact grams is metric and inches in imperial.  I just converted it to the units I needed and moved on. [NEWLINE] [NEWLINE] Obviously imperial measurements aren't useless, otherwise millions of people wouldn't continue to use them every day.  They make sense to a lot of people and that has to count for something.  In other words, millions of people find imperial measurements more practical than metric. [NEWLINE] [NEWLINE] For me, I find Lines per square inch to be much more intuitive than Tesla when I'm dealing with flux density.   But that's just me... I'd never say that one is better than the other.  They're just... different.  (In electromagnetics, there are actually 3 systems of measurement that I need to convert between on a regular basis.) [NEWLINE] [NEWLINE] But honestly it all comes down to what you are used to.  If you are used to meters and centimeters, Celsius and kilograms, you are going to find those more practical.  But if you are used to feet and inches, Fahrenheit and pounds, you are going to find those more practical. [NEWLINE] </s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First, fundamentalism is always going to end badly.<mask> one accepts beliefs such<mask> those, they must reject other perspectives and refuse to ask a lot of questions (questions that,<mask> asked by everyone, would solve many problems).<mask> they deny them,<mask> you have, the relationship between the fundamentalists and the deniers is jaggedly severred. It's not simply "I go my way, you go your way" –<mask> it is like your case (<mask> in, dealing with family), it can lead to problems in the development of the denier's worldview. Many atheists define themselves largely by their opposition to religion. Even<mask> this is not the case, there is inevitably emotional and/or psychological pain involved. [NEWLINE] [NEWLINE] <mask> for whether one should teach their children their own beliefs, this will happen inevitably<mask><mask> whether it is intentional or not. A child is socialized largely by their parents and their parents were socialized by their parents;<mask> we are subjective beings, we have to choose<mask> to teach our children and there is only<mask> much we can choose from, underlying our limitations. [NEWLINE] [NEWLINE] It is not that it is wrong to teach one's child their own worldview – I'd say that the real problem is in<mask> one does<mask>.<mask> one socializes their child in a way that denies any other perspectives, then they are setting them up for a serious shock for<mask> they go out into the world and realize that everyone is not like them.  </s>
Label encoding: <s>First, fundamentalism is always going to end badly. If one accepts beliefs such as those, they must reject other perspectives and refuse to ask a lot of questions (questions that, if asked by everyone, would solve many problems). If they deny them, as you have, the relationship between the fundamentalists and the deniers is jaggedly severred. It's not simply "I go my way, you go your way" – when it is like your case ( as in, dealing with family), it can lead to problems in the development of the denier's worldview. Many atheists define themselves largely by their opposition to religion. Even if this is not the case, there is inevitably emotional and/or psychological pain involved. [NEWLINE] [NEWLINE] As for whether one should teach their children their own beliefs, this will happen inevitably regardless of whether it is intentional or not. A child is socialized largely by their parents and their parents were socialized by their parents; because we are subjective beings, we have to choose what to teach our children and there is only so much we can choose from, underlying our limitations. [NEWLINE] [NEWLINE] It is not that it is wrong to teach one's child their own worldview – I'd say that the real problem is in how one does so. If one socializes their child in a way that denies any other perspectives, then they are setting them up for a serious shock for when they go out into the world and realize that everyone is not like them.  </s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't believe that regular citizens have anything to worry about. It would be logistically impossible for any government officials to scrawl through your history and know everything about each citizen; they just have this information at hand in case they need catch a criminal. [NEWLINE] [NEWLINE] Even<mask> a self proclaimed "liberal" I don't see<mask> the issue is. This information is going to be collected anyway, logged by every website you visit or telephone provider. I'd rather let the government have access to it in the chances that they can stop a crime. I know it's not the best way of crime fighting<mask> even stopping one crime is better than nothing. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I don't believe that regular citizens have anything to worry about. It would be logistically impossible for any government officials to scrawl through your history and know everything about each citizen; they just have this information at hand in case they need catch a criminal. [NEWLINE] [NEWLINE] Even as a self proclaimed "liberal" I don't see where the issue is. This information is going to be collected anyway, logged by every website you visit or telephone provider. I'd rather let the government have access to it in the chances that they can stop a crime. I know it's not the best way of crime fighting but even stopping one crime is better than nothing. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Perhaps you can say -<mask> drive at all in those conditions? Well -<mask> you can now<mask> you couldn't before. [ENDQ] [NEWLINE] Ah that's<mask> I love about the direction technology is going now today. We're beyond doing things<mask> it's practical. Now we do things "<mask> we can". God forbid someone makes a decision to drive in these shitty conditions using a self-driving car and the computer system fails. I mean, the technology is flawless right? It's not like it was created by humans or anything right? [NEWLINE] [NEWLINE] [STARTQ] <mask> you have an autodrive car - you did consent,<mask><mask> someone else has an autodrive car that gets into an accident -<mask> is it any different than another non self driving car that gets into an accident. [ENDQ] [NEWLINE] I would've consented only<mask> I made the choice to purchase one.<mask> I still have my manually driven car and the government makes it<mask> that everyone has to drive a self-driving car (like you and<mask> many else on this thread seem to suggest) then my use of a self-driving car is about<mask> consensual<mask> rape (not that the two are anything alike,<mask> its the best example I could come up with on the spot).<mask><mask> I still had the option to drive my manual car then yeah, I'm all for others using the technology. Especially people like yourself who clearly aren't confident enough in their driving skills to be able to drive without causing an accident. </s>
Label encoding: <s> [STARTQ] Perhaps you can say - why drive at all in those conditions? Well - because you can now when you couldn't before. [ENDQ] [NEWLINE] Ah that's what I love about the direction technology is going now today. We're beyond doing things because it's practical. Now we do things " because we can". God forbid someone makes a decision to drive in these shitty conditions using a self-driving car and the computer system fails. I mean, the technology is flawless right? It's not like it was created by humans or anything right? [NEWLINE] [NEWLINE] [STARTQ] If you have an autodrive car - you did consent, but if someone else has an autodrive car that gets into an accident - how is it any different than another non self driving car that gets into an accident. [ENDQ] [NEWLINE] I would've consented only if I made the choice to purchase one. If I still have my manually driven car and the government makes it so that everyone has to drive a self-driving car (like you and so many else on this thread seem to suggest) then my use of a self-driving car is about as consensual as rape (not that the two are anything alike, but its the best example I could come up with on the spot). If however I still had the option to drive my manual car then yeah, I'm all for others using the technology. Especially people like yourself who clearly aren't confident enough in their driving skills to be able to drive without causing an accident. </s>
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>These things must exist<mask> that man can choose to make good choices (or bad ones)<mask> man could not do evil, he would be a machine. Disaese exists to challenge humans<mask> that they have to learn and adapt.<mask>'s a story without conflict? [NEWLINE] [NEWLINE] [STARTQ] No reason to test<mask> God is all knowing [ENDQ] [NEWLINE] Humans don't have that power,<mask> God allows people to be born and to choose or reject him. In order to know something, it must be based off of prior experience. Whether or not this applies to God is debatable. [NEWLINE] Humans have their personality through genes (which is<mask> God would be involved) their early environment (human choice) and their own experiences. Man is free, his choices shape his world, and other humans shape the world<mask> well. Most<mask> not all of this worlds suffering can be avoided with morals, cooperation and science. God, by giving man free will and a bountiful planet in which to grow, has given  freedom. You are free to make the choices.<mask> you will be judged<mask><mask> all your deeds, and this is were good and evil meet. Those who have sown evil will reap it in the world to come. God would be worthy of praise and admiration simply by creating the universe. Think about<mask> complex and grand nature is,<mask> amazing the smallest insect is, the glory of stars. God should be praised for being the artist who created a true masterpiece. </s>
Label encoding: <s>These things must exist so that man can choose to make good choices (or bad ones) If man could not do evil, he would be a machine. Disaese exists to challenge humans so that they have to learn and adapt. What's a story without conflict? [NEWLINE] [NEWLINE] [STARTQ] No reason to test because God is all knowing [ENDQ] [NEWLINE] Humans don't have that power, but God allows people to be born and to choose or reject him. In order to know something, it must be based off of prior experience. Whether or not this applies to God is debatable. [NEWLINE] Humans have their personality through genes (which is where God would be involved) their early environment (human choice) and their own experiences. Man is free, his choices shape his world, and other humans shape the world as well. Most if not all of this worlds suffering can be avoided with morals, cooperation and science. God, by giving man free will and a bountiful planet in which to grow, has given  freedom. You are free to make the choices. But you will be judged according to all your deeds, and this is were good and evil meet. Those who have sown evil will reap it in the world to come. God would be worthy of praise and admiration simply by creating the universe. Think about how complex and grand nature is, how amazing the smallest insect is, the glory of stars. God should be praised for being the artist who created a true masterpiece. </s>
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's an amazing quote by David Foster Wallace about suicide: [NEWLINE] [NEWLINE] "The<mask> -called ‘psychotically depressed’ person who tries to kill herself doesn't do<mask> out of quote ‘hopelessness’ or any abstract conviction that life’s assets and debits do not square. And surely not<mask> death seems suddenly appealing. [NEWLINE] [NEWLINE] The person in whom Its invisible agony reaches a certain unendurable level will kill herself the same way a trapped person will eventually jump from the window of a burning high-rise. [NEWLINE] [NEWLINE] Make no mistake about people who leap from burning windows. Their terror of falling from a great height is still just<mask> great<mask> it would be for you or me standing speculatively at the same window just checking out the view; i.e. the fear of falling remains a constant. [NEWLINE] [NEWLINE] The variable here is the other terror, the fire’s flames:<mask> the flames get close enough, falling to death becomes the slightly less terrible of two terrors. It’s not desiring the fall; it’s terror of the flames. And<mask> nobody down on the sidewalk, looking up and yelling ‘Don’t!’ and ‘Hang on!’, can understand the jump. Not really. You’d have to have personally been trapped and felt flames to really understand a terror way beyond falling." [NEWLINE] [NEWLINE] He committed suicide by hanging himself.</s>
Label encoding: <s>There's an amazing quote by David Foster Wallace about suicide: [NEWLINE] [NEWLINE] "The so -called ‘psychotically depressed’ person who tries to kill herself doesn't do so out of quote ‘hopelessness’ or any abstract conviction that life’s assets and debits do not square. And surely not because death seems suddenly appealing. [NEWLINE] [NEWLINE] The person in whom Its invisible agony reaches a certain unendurable level will kill herself the same way a trapped person will eventually jump from the window of a burning high-rise. [NEWLINE] [NEWLINE] Make no mistake about people who leap from burning windows. Their terror of falling from a great height is still just as great as it would be for you or me standing speculatively at the same window just checking out the view; i.e. the fear of falling remains a constant. [NEWLINE] [NEWLINE] The variable here is the other terror, the fire’s flames: when the flames get close enough, falling to death becomes the slightly less terrible of two terrors. It’s not desiring the fall; it’s terror of the flames. And yet nobody down on the sidewalk, looking up and yelling ‘Don’t!’ and ‘Hang on!’, can understand the jump. Not really. You’d have to have personally been trapped and felt flames to really understand a terror way beyond falling." [NEWLINE] [NEWLINE] He committed suicide by hanging himself.</s>
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Lets switch that example around, you're saying that unless you're a multibillionaire you will never really the best you can be<mask> the money that you are giving away right now is not significant at all. Giving away half your income<mask> your working at minimum wage is much more meaningful than giving away half your income<mask> a multibillionaire simply<mask> giving away half your fortune<mask> a multibillionaire doesn't affect your quality of life that much<mask> giving away half your money<mask> person working a minimum wage affects your quality of life a lot. [NEWLINE] [NEWLINE] Now, am i saying that<mask> i based morality on the consequences of my actions that Bill Gates didn't do more good? No i'm not.<mask> I see morality<mask> character and I believe it isn't the amount that you give its the meaning that you associate with that amount. [NEWLINE] [NEWLINE] Imagine being very poor and still giving away some of the very little amount of money you earn to charity. Is that person in your eyes not a good **person** even<mask> he hasn't contributed all that much in terms of creating good consequences. [NEWLINE] [NEWLINE] Second of all none of this addresses the fact that i'm talking about a ideal person here. The question that needs to be answered is would a Bill Gates who gave away all of his fortune and left enough to sustain him and his family comfortably rather than one who gave away most of his fortune and still retained his 66,000 square foot home a better person?</s><pad>
Label encoding: <s>Lets switch that example around, you're saying that unless you're a multibillionaire you will never really the best you can be as the money that you are giving away right now is not significant at all. Giving away half your income when your working at minimum wage is much more meaningful than giving away half your income as a multibillionaire simply because giving away half your fortune as a multibillionaire doesn't affect your quality of life that much while giving away half your money as person working a minimum wage affects your quality of life a lot. [NEWLINE] [NEWLINE] Now, am i saying that if i based morality on the consequences of my actions that Bill Gates didn't do more good? No i'm not. But I see morality as character and I believe it isn't the amount that you give its the meaning that you associate with that amount. [NEWLINE] [NEWLINE] Imagine being very poor and still giving away some of the very little amount of money you earn to charity. Is that person in your eyes not a good **person** even if he hasn't contributed all that much in terms of creating good consequences. [NEWLINE] [NEWLINE] Second of all none of this addresses the fact that i'm talking about a ideal person here. The question that needs to be answered is would a Bill Gates who gave away all of his fortune and left enough to sustain him and his family comfortably rather than one who gave away most of his fortune and still retained his 66,000 square foot home a better person?</s><pad>
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What you're describing is jealousy.  It's coming from a different context and set of values than it would<mask> you were raised in America,<mask> it's basically the same thing. [NEWLINE] [NEWLINE] <mask> I'm going to tell you exactly<mask> I'd tell an American guy who had this kind of worry.<mask> you're worried about other guys looking at your girlfriend, it's probably<mask> on some level, you're afraid that someone's going to "steal her" from you - that is, that she'll see some other guy that she likes better and decide to be with him instead. There are two possibilities here: [NEWLINE] [NEWLINE] 1. You're right, and she is going to leave you for some other guy. In that case, she's obviously not committed to being in a monogamous relationship wit you, and<mask> that's the only kind of relationship you want to have then it was never going to work out with her anyway. [NEWLINE] [NEWLINE] 2. You're wrong. Your jealousy is completely unfounded, and your hypothetical girlfriend really does want to be with you and no one else, no matter<mask> many other guys look at her. The *best* thing that can happen is that you'll make yourself miserable for no reason.<mask><mask><mask><mask>,<mask> your hypothetical girlfriend feels like she has to change the way she dresses and acts around her male friends to keep you happy, she might eventually decide that whatever else she liked about you in the beginning isn't worth it.</s>
Label encoding: <s>What you're describing is jealousy.  It's coming from a different context and set of values than it would if you were raised in America, but it's basically the same thing. [NEWLINE] [NEWLINE] So I'm going to tell you exactly what I'd tell an American guy who had this kind of worry. If you're worried about other guys looking at your girlfriend, it's probably because on some level, you're afraid that someone's going to "steal her" from you - that is, that she'll see some other guy that she likes better and decide to be with him instead. There are two possibilities here: [NEWLINE] [NEWLINE] 1. You're right, and she is going to leave you for some other guy. In that case, she's obviously not committed to being in a monogamous relationship wit you, and if that's the only kind of relationship you want to have then it was never going to work out with her anyway. [NEWLINE] [NEWLINE] 2. You're wrong. Your jealousy is completely unfounded, and your hypothetical girlfriend really does want to be with you and no one else, no matter how many other guys look at her. The *best* thing that can happen is that you'll make yourself miserable for no reason. On the other hand, if your hypothetical girlfriend feels like she has to change the way she dresses and acts around her male friends to keep you happy, she might eventually decide that whatever else she liked about you in the beginning isn't worth it.</s>
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>i know that's not<mask> you were saying. i took your logic, and took it to the extreme. my point was that kind of thinking can lay the groundwork for<mask> i mentioned. [NEWLINE] [NEWLINE] <mask><mask> we had the ability to know from birth our child's genes? let's say we knew<mask> their IQ would be,<mask> tall, attractive, athletic, etc. they are,<mask> are their talents/lack there of? [NEWLINE] [NEWLINE] <mask><mask> you knew your child would grow up to be a complete failure, they won't have the intelligence to survive in the workplace/life,<mask> they went on to be a drug addict or something. is it ok to euthanize them? [NEWLINE] [NEWLINE] all extreme examples,<mask> my point is<mask> do you draw the line?<mask> is it ok to euthanize a child/<mask> is it not?<mask> they are terminally ill yes?<mask><mask> they are mentally challenged, say they have down's syndrome? is it ok to euthanize them then?<mask> do you draw the line?<mask> is it ok/not ok? [NEWLINE] [NEWLINE] it would be like tim tebow,<mask> the doctors said his mom wouldn't survive the pregnancy/<mask><mask> they<mask> said he could have issues from the birth. or the movie gattaca,<mask> ethan hawke's character, an "invalid" faces challenges from society. my argument is that kind of thinking paves the way for eugenics.</s>
Label encoding: <s>i know that's not what you were saying. i took your logic, and took it to the extreme. my point was that kind of thinking can lay the groundwork for what i mentioned. [NEWLINE] [NEWLINE] what if we had the ability to know from birth our child's genes? let's say we knew what their IQ would be, how tall, attractive, athletic, etc. they are, what are their talents/lack there of? [NEWLINE] [NEWLINE] what if you knew your child would grow up to be a complete failure, they won't have the intelligence to survive in the workplace/life, so they went on to be a drug addict or something. is it ok to euthanize them? [NEWLINE] [NEWLINE] all extreme examples, but my point is where do you draw the line? when is it ok to euthanize a child/ when is it not? if they are terminally ill yes? what if they are mentally challenged, say they have down's syndrome? is it ok to euthanize them then? where do you draw the line? when is it ok/not ok? [NEWLINE] [NEWLINE] it would be like tim tebow, how the doctors said his mom wouldn't survive the pregnancy/ i think they also said he could have issues from the birth. or the movie gattaca, how ethan hawke's character, an "invalid" faces challenges from society. my argument is that kind of thinking paves the way for eugenics.</s>
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0936, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well we<mask> spend our limited rescources on reddit, smartphones and expensive coffee. People are willing to pay for all of that. Others are willing to pay for pandas.<mask> it makes them feel good isn't that good enough? [NEWLINE] [NEWLINE] Your argument fits everything depending on your perspective. Rescourses can't be spend perfectly<mask> you can't universally quantify the importance of their effect. Is reddit worth all the electricity? We are killing our planet with our power consumption. Do you really need a 400€ smart phone? Wouldn't a 100€ one do just<mask> well? You could save a few lives with that kind of money or at least prolong them. [NEWLINE] [NEWLINE] It's obvious for things that aren't the bare minimum<mask> you could do the same thing for essentials too:<mask> is it any good that you personally stay alive? Couldn't you sacrifice your live for a higher cause?<mask>'s your life worth anyway? [NEWLINE] [NEWLINE] You see that arguing from a moral and universal perspective is fruitless. There's no universal way to determine the value of things, ideas or beings. Saying that x is more important than y is not a universal statement. It only makes sense for the individual. You can say "pandas aren't<mask> important<mask> [...], to me". This means that nobody can determine<mask> something is important in general.<mask> people like keeping pandas around<mask> they like them that's a good reason to do<mask>. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Well we also spend our limited rescources on reddit, smartphones and expensive coffee. People are willing to pay for all of that. Others are willing to pay for pandas. If it makes them feel good isn't that good enough? [NEWLINE] [NEWLINE] Your argument fits everything depending on your perspective. Rescourses can't be spend perfectly if you can't universally quantify the importance of their effect. Is reddit worth all the electricity? We are killing our planet with our power consumption. Do you really need a 400€ smart phone? Wouldn't a 100€ one do just as well? You could save a few lives with that kind of money or at least prolong them. [NEWLINE] [NEWLINE] It's obvious for things that aren't the bare minimum but you could do the same thing for essentials too: How is it any good that you personally stay alive? Couldn't you sacrifice your live for a higher cause? What's your life worth anyway? [NEWLINE] [NEWLINE] You see that arguing from a moral and universal perspective is fruitless. There's no universal way to determine the value of things, ideas or beings. Saying that x is more important than y is not a universal statement. It only makes sense for the individual. You can say "pandas aren't as important as [...], to me". This means that nobody can determine if something is important in general. If people like keeping pandas around because they like them that's a good reason to do so. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most of the points I was going to bring up have been said already: [NEWLINE] [NEWLINE] 1. You own the comic and can read it over again, unlike the movie. Most comics readers place high value on ownership of the object, either<mask> they can read it dozens of times throughout their lives or<mask> they simply value the object and being a collector of a rare or special piece of art. [NEWLINE] [NEWLINE] 2. You're not taking your time with the comic, apparently. Checking out comics criticism might help you appreciate<mask> comics fans are doing<mask> reading, and<mask> they enjoy it<mask> much. [NEWLINE] [NEWLINE] 3. The amount of time, energy and creativity that multiple writers and artists at the top of their game put into each comic is alone worth the price on an economic level, like a hand-made or hand-stitched item that might be priced higher than something factory built. [NEWLINE] [NEWLINE] <mask> I haven't seen anyone point out is<mask> often we spend $3 or more on something that is consumed in far less time than 10-15 minutes. Food may be a necessity,<mask> most food we eat is far more expensive than necessary<mask> we enjoy it more. Rather than eat rice and beans and bread at home, we go out and spend $8-10 at a restaurant. Even most things at a fast food restaurant will be more than $3-4, and will be consumed within minutes. Do you ever buy beer? Coffee at a coffee shop? [NEWLINE] </s>
Label encoding: <s>Most of the points I was going to bring up have been said already: [NEWLINE] [NEWLINE] 1. You own the comic and can read it over again, unlike the movie. Most comics readers place high value on ownership of the object, either because they can read it dozens of times throughout their lives or because they simply value the object and being a collector of a rare or special piece of art. [NEWLINE] [NEWLINE] 2. You're not taking your time with the comic, apparently. Checking out comics criticism might help you appreciate what comics fans are doing while reading, and why they enjoy it so much. [NEWLINE] [NEWLINE] 3. The amount of time, energy and creativity that multiple writers and artists at the top of their game put into each comic is alone worth the price on an economic level, like a hand-made or hand-stitched item that might be priced higher than something factory built. [NEWLINE] [NEWLINE] What I haven't seen anyone point out is how often we spend $3 or more on something that is consumed in far less time than 10-15 minutes. Food may be a necessity, but most food we eat is far more expensive than necessary because we enjoy it more. Rather than eat rice and beans and bread at home, we go out and spend $8-10 at a restaurant. Even most things at a fast food restaurant will be more than $3-4, and will be consumed within minutes. Do you ever buy beer? Coffee at a coffee shop? [NEWLINE] </s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Woof.<mask> to start with this one. [NEWLINE] [NEWLINE] <mask> people say that certain rights are inalienable, they're not saying that it's impossible for a state to take them away. No one on this Earth is unaware of the fact that states can, *and frequently have*, taken human rights away from their populace. [NEWLINE] [NEWLINE] <mask> they're saying is that those rights are a requirement of a morally-justifiable society. They are the yardstick by which a state's fitness to govern should be measured against. Those rights are things that are regarded<mask> universally advantageous to people's well-being, and are<mask> mandatory in a society that purports to be beneficial to its members. [NEWLINE] [NEWLINE] Perhaps this is the part you've forgotten: organized society, and by extension political states, exist for one primary reason above all, and that's to improve the human condition. The only reason we form civilizations is to collectively improve the lives of the humans participating in them.<mask><mask> human rights are threatened by the state, that is a perversion of the entire concept of civil society. That's<mask> people mean<mask> they refer to human rights<mask> "inalienable" or "natural" - they're inherent to the objectives of society itself. [NEWLINE] [NEWLINE] And no, our rights are not created by the state. The state is created by people, who were born with those rights, and the state should be dismantled<mask> it begins to infringe on those rights.</s>
Label encoding: <s>Woof. Where to start with this one. [NEWLINE] [NEWLINE] When people say that certain rights are inalienable, they're not saying that it's impossible for a state to take them away. No one on this Earth is unaware of the fact that states can, *and frequently have*, taken human rights away from their populace. [NEWLINE] [NEWLINE] What they're saying is that those rights are a requirement of a morally-justifiable society. They are the yardstick by which a state's fitness to govern should be measured against. Those rights are things that are regarded as universally advantageous to people's well-being, and are therefore mandatory in a society that purports to be beneficial to its members. [NEWLINE] [NEWLINE] Perhaps this is the part you've forgotten: organized society, and by extension political states, exist for one primary reason above all, and that's to improve the human condition. The only reason we form civilizations is to collectively improve the lives of the humans participating in them. So when human rights are threatened by the state, that is a perversion of the entire concept of civil society. That's what people mean when they refer to human rights as "inalienable" or "natural" - they're inherent to the objectives of society itself. [NEWLINE] [NEWLINE] And no, our rights are not created by the state. The state is created by people, who were born with those rights, and the state should be dismantled if it begins to infringe on those rights.</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Religion is about peace...Those killing aren't practicing. [ENDQ] [NEWLINE] Kill People Who Don't Listen to Priests [NEWLINE] [NEWLINE] Anyone arrogant enough to reject the verdict of the judge or of the priest who represents the LORD your God must be put to death.  Such evil must be purged from Israel.  (Deuteronomy 17:12 NLT) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Kill Witches [NEWLINE] [NEWLINE] You should not let a sorceress live.  (Exodus 22:17 NAB) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Kill Homosexuals [NEWLINE] "<mask> a man lies with a male<mask> with a women, both of them shall be put to death for their abominable deed; they have forfeited their lives."  (Leviticus 20:13 NAB) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Kill Fortunetellers [NEWLINE] [NEWLINE] A man or a woman who acts<mask> a medium or fortuneteller shall be put to death by stoning; they have no one<mask> themselves to blame for their death.  (Leviticus 20:27 NAB) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Death for Hitting Dad [NEWLINE] [NEWLINE] Whoever strikes his father or mother shall be put to death.  (Exodus 21:15 NAB) [NEWLINE] [NEWLINE] [NEWLINE] Or should I go on? Religion does preach hate and murder and slavery. Sure, people pick and choose<mask> verses to listen to and follow<mask> they use reason.<mask> to say religion is all about peace is naive.</s>
Label encoding: <s> [STARTQ] Religion is about peace...Those killing aren't practicing. [ENDQ] [NEWLINE] Kill People Who Don't Listen to Priests [NEWLINE] [NEWLINE] Anyone arrogant enough to reject the verdict of the judge or of the priest who represents the LORD your God must be put to death.  Such evil must be purged from Israel.  (Deuteronomy 17:12 NLT) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Kill Witches [NEWLINE] [NEWLINE] You should not let a sorceress live.  (Exodus 22:17 NAB) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Kill Homosexuals [NEWLINE] " If a man lies with a male as with a women, both of them shall be put to death for their abominable deed; they have forfeited their lives."  (Leviticus 20:13 NAB) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Kill Fortunetellers [NEWLINE] [NEWLINE] A man or a woman who acts as a medium or fortuneteller shall be put to death by stoning; they have no one but themselves to blame for their death.  (Leviticus 20:27 NAB) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Death for Hitting Dad [NEWLINE] [NEWLINE] Whoever strikes his father or mother shall be put to death.  (Exodus 21:15 NAB) [NEWLINE] [NEWLINE] [NEWLINE] Or should I go on? Religion does preach hate and murder and slavery. Sure, people pick and choose what verses to listen to and follow because they use reason. But to say religion is all about peace is naive.</s>
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I eat meat,<mask> have never gone hunting.<mask> someone were to challenge me to being a vegetarian, or go hunting. I would go hunting. We are top of the food chain, and we are intelligent enough to understand and track ecosystem patterns.<mask> we accidentally (or intentionally) remove a predator from one of those systems, it is on us to balance the system, or else the entire thing collapses. I like eating meat, and I understand the need to maintain an ecosystem.<mask> would I go hunting to kill two birds with one stone? Yes. Would I try to do it in<mask> a humane way<mask> possible? Yes. Would I feel bad? No, it is a part of life. We keep dogs and cats<mask> pets, and they are carnivores. Except, we have the forethought and understanding to<mask> we ideally will only hunt<mask> we need to, or do it in a way that it does not throw off or destroy any environment. [NEWLINE] [NEWLINE] Hunting for sport and trophies, I can't exactly attest to. Other than the fact that we are top of the food chain, and sometimes population control or taking out an exiled male from a group can be beneficial. At the end of the day, I eat meat, I enjoy eating meat, and<mask> I had to go out and do it myself,<mask><mask><mask> it was done logically, I could totally see myself going out and doing it every once in a<mask>. </s>
Label encoding: <s>I eat meat, but have never gone hunting. If someone were to challenge me to being a vegetarian, or go hunting. I would go hunting. We are top of the food chain, and we are intelligent enough to understand and track ecosystem patterns. If we accidentally (or intentionally) remove a predator from one of those systems, it is on us to balance the system, or else the entire thing collapses. I like eating meat, and I understand the need to maintain an ecosystem. So would I go hunting to kill two birds with one stone? Yes. Would I try to do it in as a humane way as possible? Yes. Would I feel bad? No, it is a part of life. We keep dogs and cats as pets, and they are carnivores. Except, we have the forethought and understanding to where we ideally will only hunt when we need to, or do it in a way that it does not throw off or destroy any environment. [NEWLINE] [NEWLINE] Hunting for sport and trophies, I can't exactly attest to. Other than the fact that we are top of the food chain, and sometimes population control or taking out an exiled male from a group can be beneficial. At the end of the day, I eat meat, I enjoy eating meat, and if I had to go out and do it myself, as long as it was done logically, I could totally see myself going out and doing it every once in a while. </s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The difference is that true inductive reasoning takes into account all forms of evidence, whereas naive appeals to authority take into account only the authority, and no other measure of plausibility, prior belief, or other aspects of evidence. [NEWLINE] [NEWLINE] An appeal to authority,<mask> that authority has performed well in the past, is perfectly valid<mask> an instance of inductive reasoning.<mask>, one must not naively ignore the other evidence. Let's take a case<mask> a trusted geologist at some point says "This geological structure is proof that dragons are real,<mask> you can see by the patterns the wing scrapes have made in the cliff face, and the portions that have been blackened by dragonfire." You take<mask> that geologist says into account, sure. They've performed well in the past.<mask> : [NEWLINE] [NEWLINE] a) There are a lot of other potential causes for that particular pattern in the geological structure, such<mask> fraud, other known natural [NEWLINE] [NEWLINE] b) Those other potential causes have a higher prior weight than the [NEWLINE] [NEWLINE] Basically, appeals to *well performing* authorities is a case of inductive reasoning, and not vice versa, and like all inductive reasoning other evidence must be taken into account. A stone may fall with an acceleration of about 9.8 m/s^2 each time I drop it,<mask> I don't expect it to fall at that rate<mask> I drop the same rock on the surface of the moon. </s>
Label encoding: <s>The difference is that true inductive reasoning takes into account all forms of evidence, whereas naive appeals to authority take into account only the authority, and no other measure of plausibility, prior belief, or other aspects of evidence. [NEWLINE] [NEWLINE] An appeal to authority, when that authority has performed well in the past, is perfectly valid as an instance of inductive reasoning. However, one must not naively ignore the other evidence. Let's take a case where a trusted geologist at some point says "This geological structure is proof that dragons are real, as you can see by the patterns the wing scrapes have made in the cliff face, and the portions that have been blackened by dragonfire." You take what that geologist says into account, sure. They've performed well in the past. However : [NEWLINE] [NEWLINE] a) There are a lot of other potential causes for that particular pattern in the geological structure, such as fraud, other known natural [NEWLINE] [NEWLINE] b) Those other potential causes have a higher prior weight than the [NEWLINE] [NEWLINE] Basically, appeals to *well performing* authorities is a case of inductive reasoning, and not vice versa, and like all inductive reasoning other evidence must be taken into account. A stone may fall with an acceleration of about 9.8 m/s^2 each time I drop it, but I don't expect it to fall at that rate if I drop the same rock on the surface of the moon. </s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Look, it can be made up.<mask> history is 100% empirical: stuff happened, and it does not necessarily have to make sense. Biology is a combination of empiricism and logic. It is damned hard to back up everything by empirical evidence (we don't really want to experiment brutally on people like the nazis did, for example),<mask> we can understand the general logic of things and figure some things out with a medium level of certainty. [NEWLINE] [NEWLINE] <mask> it comes to these topics, the logic of evolution, unequal parental investment, sexual selection, primate and caveman male female roles, hormones (testosterone, estrogen), we can figure a lot from them. [NEWLINE] [NEWLINE] Funny thing is, this kind of biology really lines up with the culture of 1950's - or 1850's or 1550's or -250's. In other word with the traditinal "folk wisdom" of masculinity and femininity. [NEWLINE] [NEWLINE] <mask> historical folk wisdom and biological reasoning lines up, that is not a bad Bayesian evidence. [NEWLINE] [NEWLINE] The other option is to go full Left-retard and describe human history<mask> a prolonged fight between oppressor and oppressed and assume the historical folk wisdom is just the ideology of oppressors. That I am surely not going to buy, that never made any sense, in no form, once I learned a bit of history I realized the past was generally not<mask> dark<mask> it is painted by modern progressives...</s>
Label encoding: <s>Look, it can be made up. But history is 100% empirical: stuff happened, and it does not necessarily have to make sense. Biology is a combination of empiricism and logic. It is damned hard to back up everything by empirical evidence (we don't really want to experiment brutally on people like the nazis did, for example), but we can understand the general logic of things and figure some things out with a medium level of certainty. [NEWLINE] [NEWLINE] When it comes to these topics, the logic of evolution, unequal parental investment, sexual selection, primate and caveman male female roles, hormones (testosterone, estrogen), we can figure a lot from them. [NEWLINE] [NEWLINE] Funny thing is, this kind of biology really lines up with the culture of 1950's - or 1850's or 1550's or -250's. In other word with the traditinal "folk wisdom" of masculinity and femininity. [NEWLINE] [NEWLINE] When historical folk wisdom and biological reasoning lines up, that is not a bad Bayesian evidence. [NEWLINE] [NEWLINE] The other option is to go full Left-retard and describe human history as a prolonged fight between oppressor and oppressed and assume the historical folk wisdom is just the ideology of oppressors. That I am surely not going to buy, that never made any sense, in no form, once I learned a bit of history I realized the past was generally not as dark as it is painted by modern progressives...</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The entire process of breeding the stock is faster<mask> you're backing it up with billions of dollars and a state of the art army of scientists, etc. Monsanto has the capability to produce a given number of generations in a fraction of the time that it would have taken our ancestors to breed out the same number of generations of different heirloom varieties. The one version (GMO/Monsanto)<mask> serves to limit genetic variety, and the other (traditional plant husbandry) serves to increase it. They're opposing trends; traditional agriculture has worked out for us<mask> both humankind and the plants they were growing, not to mention the ecosystem, had time to adjust. Monsanto wants to reverse that trend, ostensibly in the name of helping people<mask> we all know that the dollar amount is the bottom line. Even those of us who will eat the golden rice out of necessity [eta: will know it]. [NEWLINE] [NEWLINE] It takes an awful lot of trust in an entity like Monsanto to not be at least a tiny bit wary. The potential negative implications are huge, and to my mind they far outweigh the benefits. [NEWLINE] [NEWLINE] eta2:<mask> you want a proposal, my take on it is to put more money into local, permaculture style agriculture. Our infrastructure is crumbling anyway; might<mask> well put the money we'd need to fix it into use teaching people<mask> to form local economies again.<mask> now I'm dreaming out loud. :V</s>
Label encoding: <s>The entire process of breeding the stock is faster when you're backing it up with billions of dollars and a state of the art army of scientists, etc. Monsanto has the capability to produce a given number of generations in a fraction of the time that it would have taken our ancestors to breed out the same number of generations of different heirloom varieties. The one version (GMO/Monsanto) also serves to limit genetic variety, and the other (traditional plant husbandry) serves to increase it. They're opposing trends; traditional agriculture has worked out for us because both humankind and the plants they were growing, not to mention the ecosystem, had time to adjust. Monsanto wants to reverse that trend, ostensibly in the name of helping people but we all know that the dollar amount is the bottom line. Even those of us who will eat the golden rice out of necessity [eta: will know it]. [NEWLINE] [NEWLINE] It takes an awful lot of trust in an entity like Monsanto to not be at least a tiny bit wary. The potential negative implications are huge, and to my mind they far outweigh the benefits. [NEWLINE] [NEWLINE] eta2: If you want a proposal, my take on it is to put more money into local, permaculture style agriculture. Our infrastructure is crumbling anyway; might as well put the money we'd need to fix it into use teaching people how to form local economies again. But now I'm dreaming out loud. :V</s>
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> 3D is a perfect extension of pure-CG films [ENDQ] [NEWLINE] Not for all of them.<mask> I went to re-image one of the Blender Open Movies I had to choose between "Elephants Dream" and "Big Buck Bunny". My choice was "Elephants Dream" then,<mask> its format and visuals stroke me<mask> far better suited for 3D than BBB. And the stunning results confirmed this. [NEWLINE] [NEWLINE] Interestingly enough, many of the methods I developed to improve the experience really did pay off. I got incredible response from a lot of people who told me, that usually they get nausea in 3D movies, even<mask> the 3D effect is only weak,<mask> in my version of "Elephants Dream 3D"<mask> the strong 3D they could watch it comfortably without getting simulator/3D sickness. [NEWLINE] [NEWLINE] [STARTQ] <mask> it's something that can be added to the film without too much effort [ENDQ] [NEWLINE] Having done single handedly a complete re-imaged of "Elephants Dream" into steroscopic 3D myself I can tell you, that it's not a simple click of a button. Many things require special care to work in 3D. For example 2D matte paintings must be turned into 3D counterparts. Yes<mask> in CGI matte paintings are used. And you have to do carefull stereoscopic direction to match the stereoscopy with the scene and the action.</s>
Label encoding: <s> [STARTQ] I think 3D is a perfect extension of pure-CG films [ENDQ] [NEWLINE] Not for all of them. When I went to re-image one of the Blender Open Movies I had to choose between "Elephants Dream" and "Big Buck Bunny". My choice was "Elephants Dream" then, because its format and visuals stroke me as far better suited for 3D than BBB. And the stunning results confirmed this. [NEWLINE] [NEWLINE] Interestingly enough, many of the methods I developed to improve the experience really did pay off. I got incredible response from a lot of people who told me, that usually they get nausea in 3D movies, even if the 3D effect is only weak, but in my version of "Elephants Dream 3D" despite the strong 3D they could watch it comfortably without getting simulator/3D sickness. [NEWLINE] [NEWLINE] [STARTQ] as it's something that can be added to the film without too much effort [ENDQ] [NEWLINE] Having done single handedly a complete re-imaged of "Elephants Dream" into steroscopic 3D myself I can tell you, that it's not a simple click of a button. Many things require special care to work in 3D. For example 2D matte paintings must be turned into 3D counterparts. Yes also in CGI matte paintings are used. And you have to do carefull stereoscopic direction to match the stereoscopy with the scene and the action.</s>
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, lets set aside the idea of religion completely for a moment, and look at logistics. [NEWLINE] [NEWLINE] In this example, I am (1) married, (2) a parent to a young child (not old enough to be left home alone), (3) low-income, and (4) a practicing member of a religious faith. [NEWLINE] [NEWLINE] Now, in this scenario, my spouse and I attend a religious service. This requires, including transit time, 1.5 to two hours on Sunday.<mask> do I, who can't afford to pay for childcare every weekend, do with my small child,<mask> they shouldn't be attending services? [NEWLINE] [NEWLINE] Now that we are past that simple logistical question, we get to another question:<mask> do you handle the fact that, for many people, religion and morality are permanently entwined? [NEWLINE] [NEWLINE] <mask> it would be easy for a Hindu parent to not cook beef for their children, or for a Muslim parent to maintain a *halal* kitchen,<mask> do you recommend they say<mask> they ask<mask> they can't have bacon cheeseburgers? [NEWLINE] [NEWLINE] <mask> the parents practice a pre-meal prayer, which is a very common practice, does the child wait in another room or just cover their ears? [NEWLINE] [NEWLINE] <mask> do you tell a nine year old<mask> they ask<mask> you wear a crucifix? Or for that matter, any other religious jewelry or decor in a household?</s>
Label encoding: <s>So, lets set aside the idea of religion completely for a moment, and look at logistics. [NEWLINE] [NEWLINE] In this example, I am (1) married, (2) a parent to a young child (not old enough to be left home alone), (3) low-income, and (4) a practicing member of a religious faith. [NEWLINE] [NEWLINE] Now, in this scenario, my spouse and I attend a religious service. This requires, including transit time, 1.5 to two hours on Sunday. What do I, who can't afford to pay for childcare every weekend, do with my small child, since they shouldn't be attending services? [NEWLINE] [NEWLINE] Now that we are past that simple logistical question, we get to another question: how do you handle the fact that, for many people, religion and morality are permanently entwined? [NEWLINE] [NEWLINE] While it would be easy for a Hindu parent to not cook beef for their children, or for a Muslim parent to maintain a *halal* kitchen, what do you recommend they say when they ask why they can't have bacon cheeseburgers? [NEWLINE] [NEWLINE] If the parents practice a pre-meal prayer, which is a very common practice, does the child wait in another room or just cover their ears? [NEWLINE] [NEWLINE] What do you tell a nine year old when they ask why you wear a crucifix? Or for that matter, any other religious jewelry or decor in a household?</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Two songs in which I enjoy the fade out are Killswitch Engage's songs Arms of Sorrow and My Curse. My own personal interpretation of the fade out is a purposeful meaning of incompleteness that makes the ubiquitous feeling of sorrow in the former and the longing for reciprocated love seem eternal. [NEWLINE] [NEWLINE] In the case of Arms of Sorrow, the song captures the desperation one my feel<mask> depressed and desire to no longer feel this way; the fade out at the end of the song gives the impression that there is no guaranteed end to this emotional suffering. In the case of My Curse, the song is almost reminiscent of Romance era love songs in its lyrics, professing a deep, enduring love for someone and describing in depth the pain that comes not being near the object of one's affection. The fade out at the end of the song complements the chorus ("There is love, burning to find you. Will you wait for me?") by providing an implication that this person may have to wait forever to be in his/her love interest's presence again. [NEWLINE] [NEWLINE] I actually agree with your view that generally a fade out at the end of a song is lazy and in no stretch of the imagination can be interpreted artistically (first song that came to my mind was Let Me Blow Ya Mind by Eve featuring Gwen Stefani). I just like that this CMV is an opportunity for some of us share music!</s>
Label encoding: <s>Two songs in which I enjoy the fade out are Killswitch Engage's songs Arms of Sorrow and My Curse. My own personal interpretation of the fade out is a purposeful meaning of incompleteness that makes the ubiquitous feeling of sorrow in the former and the longing for reciprocated love seem eternal. [NEWLINE] [NEWLINE] In the case of Arms of Sorrow, the song captures the desperation one my feel when depressed and desire to no longer feel this way; the fade out at the end of the song gives the impression that there is no guaranteed end to this emotional suffering. In the case of My Curse, the song is almost reminiscent of Romance era love songs in its lyrics, professing a deep, enduring love for someone and describing in depth the pain that comes not being near the object of one's affection. The fade out at the end of the song complements the chorus ("There is love, burning to find you. Will you wait for me?") by providing an implication that this person may have to wait forever to be in his/her love interest's presence again. [NEWLINE] [NEWLINE] I actually agree with your view that generally a fade out at the end of a song is lazy and in no stretch of the imagination can be interpreted artistically (first song that came to my mind was Let Me Blow Ya Mind by Eve featuring Gwen Stefani). I just like that this CMV is an opportunity for some of us share music!</s>
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Easy... you're only allowed to state facts during the News. "Republicans cause a government shutdown" is clearly a biased statement. It places blame with republicans which is arguable. That should be designated<mask> an opinion. [NEWLINE] [NEWLINE] I just happened to catch a blurb of NPR news<mask> on the road earlier today. It went something like, "The government has now been shutdown except for critical services. President Obama stated in a news conference that blah blah, it's House Republicans' fault. &lt;sound clip of Obama quote [STARTQ] House Majority Leader John Boehner stated that Senate Democrats were unwilling to come to the table, blah blah. &lt;sound clip of Boehner quote&gt;" [ENDQ] [NEWLINE] And there you go. Only facts... the govt has been shutdown, Obama said this, Boehner said this. No color commentary... just facts. This is<mask> happened. That's it. [NEWLINE] [NEWLINE] Yes, it is still possible to put a slant on stories<mask> only reporting facts. I'm not saying it isn't.<mask> it's FARRRR better than<mask> we have to slush through now. [NEWLINE] [NEWLINE] EDIT: an additional thought. A big reason<mask> I feel your previous statement is on shaky ground is that<mask> you say 'who decides<mask> is a fact,' you are implying that facts are subjective. They aren't. That's<mask> they are called facts. They are objectively, veritably true.</s>
Label encoding: <s>Easy... you're only allowed to state facts during the News. "Republicans cause a government shutdown" is clearly a biased statement. It places blame with republicans which is arguable. That should be designated as an opinion. [NEWLINE] [NEWLINE] I just happened to catch a blurb of NPR news while on the road earlier today. It went something like, "The government has now been shutdown except for critical services. President Obama stated in a news conference that blah blah, it's House Republicans' fault. &lt;sound clip of Obama quote [STARTQ] House Majority Leader John Boehner stated that Senate Democrats were unwilling to come to the table, blah blah. &lt;sound clip of Boehner quote&gt;" [ENDQ] [NEWLINE] And there you go. Only facts... the govt has been shutdown, Obama said this, Boehner said this. No color commentary... just facts. This is what happened. That's it. [NEWLINE] [NEWLINE] Yes, it is still possible to put a slant on stories while only reporting facts. I'm not saying it isn't. But it's FARRRR better than what we have to slush through now. [NEWLINE] [NEWLINE] EDIT: an additional thought. A big reason why I feel your previous statement is on shaky ground is that when you say 'who decides what is a fact,' you are implying that facts are subjective. They aren't. That's why they are called facts. They are objectively, veritably true.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Like I said, any album or single. Released between 2000 and 2005 is likely from an established act from the 90s. I suppose it is possible for a band to form in a year and release the culture changing album that same year<mask> I imagine it is quite rare. Usually you have at least a few years of building a catalog of songs, performing and practicing together and slowly breaking in and getting an audience. Unless you go on an American idol type show or are a boy band put together by a producer it is doubtful you'll be able to form and mass release music within a couple years. [NEWLINE] [NEWLINE] I'm not saying bands formed in 200-2005 don't count. I'm saying those band probably didn't produce their culture changing work until closer to 2005. Anyway nirvana was to me a unique case. Can you name any other band that<mask> significantly and overwhelmingly changed mass culture<mask> quickly and obviously? Maybe the Beatles? Most of the op suggestions aren't<mask> broad<mask> nirvana. Bob Marley? Completely limited to his genre in terms of appeal and influence.  Hendrix? Affected mostly his genre and guitar nuts.  None of the other acts fundamentally transformed the culture in the way nirvana did which is<mask> nirvana seems like an exception rather than a rule. Most influential artists have much smaller and more subtle impact over a long career rather than an overnight and ubiquitous cultural shift. </s>
Label encoding: <s>Like I said, any album or single. Released between 2000 and 2005 is likely from an established act from the 90s. I suppose it is possible for a band to form in a year and release the culture changing album that same year but I imagine it is quite rare. Usually you have at least a few years of building a catalog of songs, performing and practicing together and slowly breaking in and getting an audience. Unless you go on an American idol type show or are a boy band put together by a producer it is doubtful you'll be able to form and mass release music within a couple years. [NEWLINE] [NEWLINE] I'm not saying bands formed in 200-2005 don't count. I'm saying those band probably didn't produce their culture changing work until closer to 2005. Anyway nirvana was to me a unique case. Can you name any other band that so significantly and overwhelmingly changed mass culture so quickly and obviously? Maybe the Beatles? Most of the op suggestions aren't as broad as nirvana. Bob Marley? Completely limited to his genre in terms of appeal and influence.  Hendrix? Affected mostly his genre and guitar nuts.  None of the other acts fundamentally transformed the culture in the way nirvana did which is why nirvana seems like an exception rather than a rule. Most influential artists have much smaller and more subtle impact over a long career rather than an overnight and ubiquitous cultural shift. </s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0905, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you are incorrect in the following ways: [NEWLINE] [NEWLINE] * Comparing yourself to others by observation is of limited use<mask> your sample size is much too small to be of use and vulnerable to location and demographic bias. For example,<mask> you wanted to be the best runner, comparing yourself to people you work or go to school with would only show<mask> you are better or worse than a very small subset of the overall population. [NEWLINE] [NEWLINE] *<mask> your goal is to improve, you would be better off comparing your objective statistics to a large sample size to determine<mask> you compare. In the running example above, it would be more useful to use your race time (an objective statistic) and compare it to the posted race times of large public races (for example, the Boston Marathon) This would give you a much better indication<mask> to your performance in the real world. [NEWLINE] [NEWLINE] *<mask> your goal is continuous improvement, you would be better off comparing your objective measurements against your own personal history. Again, using the running example, looking at your race times for the last ten races would show you<mask> you are improving or not.<mask>, you could take into consideration your efforts and see<mask> the return on your invested effort was for individual races. [NEWLINE] [NEWLINE] Finally, these methods would allow you to work with people with similar goals in a non-competitive way<mask> your performance against any one individual is insignificant<mask> compared to a large population.</s>
Label encoding: <s>I think you are incorrect in the following ways: [NEWLINE] [NEWLINE] * Comparing yourself to others by observation is of limited use because your sample size is much too small to be of use and vulnerable to location and demographic bias. For example, if you wanted to be the best runner, comparing yourself to people you work or go to school with would only show if you are better or worse than a very small subset of the overall population. [NEWLINE] [NEWLINE] * If your goal is to improve, you would be better off comparing your objective statistics to a large sample size to determine how you compare. In the running example above, it would be more useful to use your race time (an objective statistic) and compare it to the posted race times of large public races (for example, the Boston Marathon) This would give you a much better indication as to your performance in the real world. [NEWLINE] [NEWLINE] * If your goal is continuous improvement, you would be better off comparing your objective measurements against your own personal history. Again, using the running example, looking at your race times for the last ten races would show you if you are improving or not. Also, you could take into consideration your efforts and see what the return on your invested effort was for individual races. [NEWLINE] [NEWLINE] Finally, these methods would allow you to work with people with similar goals in a non-competitive way as your performance against any one individual is insignificant when compared to a large population.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Two issues with your argument that I see right off the bat. [NEWLINE] [NEWLINE] Gun owners are not required to be knowledgeable about weapons in order to own one.  Not by law at least, <mask><mask><mask> you pass the background check you can buy a gun.  Now the seller may ask you your skill level and will likely suggest you take a safety course<mask> you are new to owning a firearm.  There is a high level of respect for guns and safety in the gun owning community and they are all happy to share that knowledge with others.  My point being, the majority of gun safety and gun knowledge is voluntary in nature.  You do see classes required for concealed carry permit holders, and they will cover general safety,<mask> they are mostly intended for teaching the legalities of it and other pertinent topics.   This does not apply to simple ownership.  I'm a bit surprised by your perception of our gun laws to be honest. [NEWLINE] [NEWLINE] Every kid in America has to pass the constitution test,  which is the equivalent of the exam you sound like you are proposing in order to vote, in order to pass eighth grade.  Every new citizen must pass this or a similar test<mask> well.  I might be mistaken<mask><mask><mask> theirs is more in depth. [NEWLINE] [NEWLINE] <mask> everyone who is eligible to vote has passed a standard test covering<mask> our government works at some point in their life.  </s>
Label encoding: <s>Two issues with your argument that I see right off the bat. [NEWLINE] [NEWLINE] Gun owners are not required to be knowledgeable about weapons in order to own one.  Not by law at least,  as long as you pass the background check you can buy a gun.  Now the seller may ask you your skill level and will likely suggest you take a safety course if you are new to owning a firearm.  There is a high level of respect for guns and safety in the gun owning community and they are all happy to share that knowledge with others.  My point being, the majority of gun safety and gun knowledge is voluntary in nature.  You do see classes required for concealed carry permit holders, and they will cover general safety, but they are mostly intended for teaching the legalities of it and other pertinent topics.   This does not apply to simple ownership.  I'm a bit surprised by your perception of our gun laws to be honest. [NEWLINE] [NEWLINE] Every kid in America has to pass the constitution test,  which is the equivalent of the exam you sound like you are proposing in order to vote, in order to pass eighth grade.  Every new citizen must pass this or a similar test as well.  I might be mistaken but I think theirs is more in depth. [NEWLINE] [NEWLINE] So everyone who is eligible to vote has passed a standard test covering how our government works at some point in their life.  </s>
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0964, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was at a McDonald's the other day<mask> a 20-something year old woman became engaged in a conversation with one of the employees. I don't know<mask> the conversation was about,<mask><mask> I heard was that the woman became "offended"<mask> the McDonald's employee used a blanket statement, which lead the woman to filing a complaint against the employee. [NEWLINE] [NEWLINE] The woman argued that it was her "right" to file a complaint<mask> she was offended. All I could think was "ya, it may be your 'right',<mask> that doesn't make it *right*." [NEWLINE] [NEWLINE] <mask> you,<mask> an individual - or even<mask> a group for that matter - take offense on something, that's on *you* to work out. No person should be entitled to reparations<mask> they feel *offended*<mask> we are all offended by different things, and are not always offended by the same thing. [NEWLINE] [NEWLINE] I feel like I'm living in a world full of small children, many of which are trapped in adult bodies. [NEWLINE] [NEWLINE] A person is not an asshole<mask> they continue to say something that offends you, *you* are just overly sensitive and manipulative, trying to control the world around you with your feelings, rarely making an effort to seek ways to develop your own strength of character and thicken your skin. [NEWLINE] [NEWLINE] "Freedom of speech" until our feelings get hurt, right?</s>
Label encoding: <s>I was at a McDonald's the other day when a 20-something year old woman became engaged in a conversation with one of the employees. I don't know what the conversation was about, but what I heard was that the woman became "offended" when the McDonald's employee used a blanket statement, which lead the woman to filing a complaint against the employee. [NEWLINE] [NEWLINE] The woman argued that it was her "right" to file a complaint because she was offended. All I could think was "ya, it may be your 'right', but that doesn't make it *right*." [NEWLINE] [NEWLINE] If you, as an individual - or even as a group for that matter - take offense on something, that's on *you* to work out. No person should be entitled to reparations because they feel *offended* because we are all offended by different things, and are not always offended by the same thing. [NEWLINE] [NEWLINE] I feel like I'm living in a world full of small children, many of which are trapped in adult bodies. [NEWLINE] [NEWLINE] A person is not an asshole because they continue to say something that offends you, *you* are just overly sensitive and manipulative, trying to control the world around you with your feelings, rarely making an effort to seek ways to develop your own strength of character and thicken your skin. [NEWLINE] [NEWLINE] "Freedom of speech" until our feelings get hurt, right?</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not quite, I'm insisting that people have the right to feel their own emotions, whether those emotions are logical or not. I don't see anything wrong with that. Many people have a natural reaction to victimization,<mask><mask><mask> they're victimized. It hurts to be cheated on,<mask><mask> it's entirely possible that nothing physical has been done to you. It feels terrible to be robbed, even<mask> they're just physical objects. Rape is worse than those<mask> sex is one of,<mask> not the, most defining properties of humanity, and it carries with it a very hefty emotional connection. Tainting sex (by classically conditioning someone to associate it with fear, disgust, and anger) is fundamentally changing another person's core being. That is something more than being punched in the face. [NEWLINE] [NEWLINE] <mask> I'm robbed in an alley, I may associate that alley with danger. I might get skittish<mask> people are walking near me. I may even be intimidated by the outdoors.<mask> none of those can even hold a candle to<mask> important and pervasive sex and sexuality are to our society (not to mention the fact that being raped can cause all of those,<mask> well!). [NEWLINE] [NEWLINE] Addendum: I'm not trying to suggest that the most common type of sexual assault is by a stranger in an alley. I know it isn't. It was simply the example that I used for this situation.</s>
Label encoding: <s>Not quite, I'm insisting that people have the right to feel their own emotions, whether those emotions are logical or not. I don't see anything wrong with that. Many people have a natural reaction to victimization, regardless of how they're victimized. It hurts to be cheated on, even though it's entirely possible that nothing physical has been done to you. It feels terrible to be robbed, even if they're just physical objects. Rape is worse than those because sex is one of, if not the, most defining properties of humanity, and it carries with it a very hefty emotional connection. Tainting sex (by classically conditioning someone to associate it with fear, disgust, and anger) is fundamentally changing another person's core being. That is something more than being punched in the face. [NEWLINE] [NEWLINE] If I'm robbed in an alley, I may associate that alley with danger. I might get skittish when people are walking near me. I may even be intimidated by the outdoors. But none of those can even hold a candle to how important and pervasive sex and sexuality are to our society (not to mention the fact that being raped can cause all of those, as well!). [NEWLINE] [NEWLINE] Addendum: I'm not trying to suggest that the most common type of sexual assault is by a stranger in an alley. I know it isn't. It was simply the example that I used for this situation.</s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>dueling literacy tests [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] this is pretty hard (<mask> not impossible) [NEWLINE] [NEWLINE] for this one it depends<mask> score you need to get and most importantly<mask> it is enforced: the jim crow south didn't keep black voting down by simply creating hard tests, the whole system conspired to block the right to vote. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] "hey, good thing about those literacy tests, now the blacks can't vote." No. They were probably thinking, " [ENDQ] [NEWLINE] here's the problem: this is pretty much 100% wrong. The south really was that racist. [NEWLINE] [NEWLINE] [NEWLINE] <mask> i find the website you cite unintentionally amusing: the facts are true<mask> it's<mask> very clearly a national poll. 44% of whites in mississippi didn't approve of the civil rights movement. *and i don't actually see the "equal shot" stat there.<mask> anything the article supports the opposite view. [NEWLINE] [NEWLINE] <mask> jim crow south was in existence for more than it's very end (1960s). [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] That's the danger of assuming stuff's colorblind<mask> it's not. [ENDQ] [NEWLINE] <mask> my initial comment indicated<mask><mask> you can make a race/income argument akin to the "racism<mask> poor people don't have id cards"<mask> that's a wildely wildly wildly...wildly wildly different thing than the ways the south actively blocked minority votes. </s>
Label encoding: <s>dueling literacy tests [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] this is pretty hard ( though not impossible) [NEWLINE] [NEWLINE] for this one it depends what score you need to get and most importantly how it is enforced: the jim crow south didn't keep black voting down by simply creating hard tests, the whole system conspired to block the right to vote. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] "hey, good thing about those literacy tests, now the blacks can't vote." No. They were probably thinking, " [ENDQ] [NEWLINE] here's the problem: this is pretty much 100% wrong. The south really was that racist. [NEWLINE] [NEWLINE] [NEWLINE] also i find the website you cite unintentionally amusing: the facts are true but it's also very clearly a national poll. 44% of whites in mississippi didn't approve of the civil rights movement. *and i don't actually see the "equal shot" stat there. if anything the article supports the opposite view. [NEWLINE] [NEWLINE] also jim crow south was in existence for more than it's very end (1960s). [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] That's the danger of assuming stuff's colorblind when it's not. [ENDQ] [NEWLINE] as my initial comment indicated i think you can make a race/income argument akin to the "racism because poor people don't have id cards" but that's a wildely wildly wildly...wildly wildly different thing than the ways the south actively blocked minority votes. </s>
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Since you mentioned your background<mask> a Christian I'll approach this from a Christian point of view. [NEWLINE] [NEWLINE] 1. Per Christianity, humans were created to worship God and have a relationship with him.  This is the purpose for which we were created,<mask><mask> you grant the premise of a Christian God, then you<mask> have to accept that humanity cannot ever fulfill their purpose without God. [NEWLINE] [NEWLINE] 2. The Bible never says that Christians will go to heaven, merely that we will go to the 'kingdom of heaven' (this is an unfortunate misconception perpetuated by widespread biblical illiteracy).  The afterlife for Christians consists of living on a reformed earth similar in many respects to our own earth ruled by King Jesus.  There will still be professions, differing levels of authority, even money, and we will have an eternity to develop our skills/creative passions. [NEWLINE] [NEWLINE] For instance,<mask> you like to paint, you will have an eternity to perfect your painting skills. <mask> you like to cook, you will have the option to practice making omelettes ad infinitum. <mask> you like animals, you will be free to find and rear them. <mask> you like science you will have the ability to study the reformed universe. <mask> you are talented at business you can perfect your craft through commerce.  Presumably doctors, lawyers, and pastors will all be out of a job<mask>.</s>
Label encoding: <s>Since you mentioned your background as a Christian I'll approach this from a Christian point of view. [NEWLINE] [NEWLINE] 1. Per Christianity, humans were created to worship God and have a relationship with him.  This is the purpose for which we were created, so if you grant the premise of a Christian God, then you also have to accept that humanity cannot ever fulfill their purpose without God. [NEWLINE] [NEWLINE] 2. The Bible never says that Christians will go to heaven, merely that we will go to the 'kingdom of heaven' (this is an unfortunate misconception perpetuated by widespread biblical illiteracy).  The afterlife for Christians consists of living on a reformed earth similar in many respects to our own earth ruled by King Jesus.  There will still be professions, differing levels of authority, even money, and we will have an eternity to develop our skills/creative passions. [NEWLINE] [NEWLINE] For instance, if you like to paint, you will have an eternity to perfect your painting skills.  If you like to cook, you will have the option to practice making omelettes ad infinitum.  If you like animals, you will be free to find and rear them.  If you like science you will have the ability to study the reformed universe.  If you are talented at business you can perfect your craft through commerce.  Presumably doctors, lawyers, and pastors will all be out of a job however.</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] This is a really interesting point, and<mask> I didn't agree with all of<mask> you're saying, I'm coming around to the notion that<mask> we want to call characteristic of any medium has much less to do with its "essence" and much more to do with the stylistic use made of it. For instance,<mask><mask> you're right that there's no reason *in principle* that an episode of television *must* be the basic unit of a series; and we can imagine a universe in which the page divisions of a book *are* clearly constructed in a way similar to a TV show's episodes. I don't think it's arbitrary that the episode is typically used<mask> a "conceptual unit"<mask> a page is typically used<mask> a "delivery unit" (to use your terms); the fact that a week separates episodes' airing (and that even<mask> binge-watching we must go through the motions of setting a new episode up) makes it seem "natural" that an episode should be one unit.<mask><mask> you say,<mask> this non-episodic mode of television became the norm, we probably would no longer have that expectation. I maintain that one structure may still just in itself be better than the other,<mask> it is not a question of more structure vs. less, or paying attention to structure vs. ignoring it.</s>
Label encoding: <s>&amp;#8710; [NEWLINE] This is a really interesting point, and though I didn't agree with all of what you're saying, I'm coming around to the notion that what we want to call characteristic of any medium has much less to do with its "essence" and much more to do with the stylistic use made of it. For instance, I think you're right that there's no reason *in principle* that an episode of television *must* be the basic unit of a series; and we can imagine a universe in which the page divisions of a book *are* clearly constructed in a way similar to a TV show's episodes. I don't think it's arbitrary that the episode is typically used as a "conceptual unit" while a page is typically used as a "delivery unit" (to use your terms); the fact that a week separates episodes' airing (and that even when binge-watching we must go through the motions of setting a new episode up) makes it seem "natural" that an episode should be one unit. But as you say, if this non-episodic mode of television became the norm, we probably would no longer have that expectation. I maintain that one structure may still just in itself be better than the other, but it is not a question of more structure vs. less, or paying attention to structure vs. ignoring it.</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Aren't these ~~condirictery~~ contradictory? [ENDQ] [NEWLINE] Capitalism optimises things,<mask> it optimises things to be profitable.<mask> profitability isn't perfectly aligned with<mask>'s best for society, then society gets shafted to the degree of the misalignment. [NEWLINE] [NEWLINE] For example,<mask> externalities like CFC's aren't taxed/banned (and<mask> *not* made unprofitable compared to alternatives), we get a massive hole in the ozone layer, and everyone close to either pole starts getting sunburned a whole lot more. [NEWLINE] [NEWLINE]...In my experience, the profits from the education business are misaligned with<mask>'s actually best in society - for example, perhaps you're going to be a whole lot *more* profitable<mask> you optimise your classes for rationalising ideas that are politically acceptable, or for exploiting the flaws of scientific journals to get<mask> many papers published<mask> possible;<mask> would be best for society is<mask> the classes you're taught are optimised to teach you<mask> to actually come to the correct conclusion. [NEWLINE] [NEWLINE] That said, I'm not a big fan of technocracy. I don't really think that OP has thought it entirely<mask> ; OP probably focused on the *results*, and not *<mask> it would happen*. Say<mask> you like about capitalism,<mask> it at least has a quite solid rhetoric on ***<mask> *** it gets its results.</s>
Label encoding: <s> [STARTQ] Aren't these ~~condirictery~~ contradictory? [ENDQ] [NEWLINE] Capitalism optimises things, but it optimises things to be profitable. If profitability isn't perfectly aligned with what's best for society, then society gets shafted to the degree of the misalignment. [NEWLINE] [NEWLINE] For example, when externalities like CFC's aren't taxed/banned (and thus *not* made unprofitable compared to alternatives), we get a massive hole in the ozone layer, and everyone close to either pole starts getting sunburned a whole lot more. [NEWLINE] [NEWLINE]...In my experience, the profits from the education business are misaligned with what's actually best in society - for example, perhaps you're going to be a whole lot *more* profitable if you optimise your classes for rationalising ideas that are politically acceptable, or for exploiting the flaws of scientific journals to get as many papers published as possible; what would be best for society is if the classes you're taught are optimised to teach you how to actually come to the correct conclusion. [NEWLINE] [NEWLINE] That said, I'm not a big fan of technocracy. I don't really think that OP has thought it entirely though ; OP probably focused on the *results*, and not * how it would happen*. Say what you like about capitalism, but it at least has a quite solid rhetoric on *** how *** it gets its results.</s>
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So are you basically saying that conservative Christians don't really care about making the world a better place, only checking off another notch on the "<mask> many fights have we won" list? I mean<mask> they REALLY want to make the world better, shouldn't they take in the issues that are more important and effect more people,<mask><mask><mask> difficult it will be to win? [NEWLINE] [NEWLINE] <mask>,<mask><mask> the idea that divorce is more accepted then gay marriage is false.<mask><mask> it definitely depends on who you are speaking with and I don't even agree that the media portrays it that way at this point.<mask> a gay person were married and in a strong relationship, the media wouldn't portray it poorly.<mask> Newt Gingrich? The media kind of rips him apart<mask> of his divorces. I actually think in many places - and certainly in the mainstream media - it would be easier to get support to ban divorce then SSM.<mask> we really do all live in a bubble of our own world and this just shows<mask> out of touch with reality conservative Christians are,<mask> your reasoning is correct. Personally,<mask><mask> they are just homophobic hypocrites (most of them)<mask> those few that DO follow the logic you are suggesting, are just delusional. Gay marriage is NOT an easier fight to fight from their side at this point in time then divorce would be.<mask> anything, it's a harder sell.</s>
Label encoding: <s>So are you basically saying that conservative Christians don't really care about making the world a better place, only checking off another notch on the " how many fights have we won" list? I mean if they REALLY want to make the world better, shouldn't they take in the issues that are more important and effect more people, regardless of how difficult it will be to win? [NEWLINE] [NEWLINE] Also, I think the idea that divorce is more accepted then gay marriage is false. I think it definitely depends on who you are speaking with and I don't even agree that the media portrays it that way at this point. If a gay person were married and in a strong relationship, the media wouldn't portray it poorly. But Newt Gingrich? The media kind of rips him apart because of his divorces. I actually think in many places - and certainly in the mainstream media - it would be easier to get support to ban divorce then SSM. But we really do all live in a bubble of our own world and this just shows how out of touch with reality conservative Christians are, if your reasoning is correct. Personally, I think they are just homophobic hypocrites (most of them) but those few that DO follow the logic you are suggesting, are just delusional. Gay marriage is NOT an easier fight to fight from their side at this point in time then divorce would be. If anything, it's a harder sell.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> can these subs possibly be adding to the Reddit community? [ENDQ] [NEWLINE] Freedom. The truth is, everyone knows that hate exists, whether it's on Reddit or not. We can chase it around and try to censor it,<mask> it will always be in vein. Nobody's experience on Reddit is actually being detracted<mask> of the existence of these subreddits. [NEWLINE] [NEWLINE] <mask><mask><mask> we did try to censor them all? First we'd have to deal with a bunch of grey lines,<mask> even ignoring that,<mask> would the effect be? The people would simply move to a different site or forum and spread their hate there. All the same people who visited those communities before would keep visiting, and all the people who didn't... still wouldn't. [NEWLINE] [NEWLINE] <mask><mask> is there to gain? Will removing the subreddits make Reddit a better place? Well, it will take out the hate-subreddits that nobody ever sees anyway,<mask> it will have one much more important consequence: It will take the freedom of speech out of the website. Barring illegal content, everything is allowed on Reddit within the confines of each subreddit's rules.<mask> you burst the hate bubbles, they will simply spill out into all of the other subreddits. Is it not better to let these people have their place to vent? To keep it separate from the rest? To preserve the value of freedom of speech that we value<mask> much?</s><pad>
Label encoding: <s> [STARTQ] What can these subs possibly be adding to the Reddit community? [ENDQ] [NEWLINE] Freedom. The truth is, everyone knows that hate exists, whether it's on Reddit or not. We can chase it around and try to censor it, but it will always be in vein. Nobody's experience on Reddit is actually being detracted because of the existence of these subreddits. [NEWLINE] [NEWLINE] So what if we did try to censor them all? First we'd have to deal with a bunch of grey lines, but even ignoring that, what would the effect be? The people would simply move to a different site or forum and spread their hate there. All the same people who visited those communities before would keep visiting, and all the people who didn't... still wouldn't. [NEWLINE] [NEWLINE] So what is there to gain? Will removing the subreddits make Reddit a better place? Well, it will take out the hate-subreddits that nobody ever sees anyway, but it will have one much more important consequence: It will take the freedom of speech out of the website. Barring illegal content, everything is allowed on Reddit within the confines of each subreddit's rules. If you burst the hate bubbles, they will simply spill out into all of the other subreddits. Is it not better to let these people have their place to vent? To keep it separate from the rest? To preserve the value of freedom of speech that we value so much?</s><pad>
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ITT: OP is pretty much correct. [NEWLINE] [NEWLINE] [NEWLINE] Longer answer: the music scene has changed forever.  With the complete corporatization of almost all radio, "charts," and "awards,"<mask> is popular and influential today is very rarely significant.  99% of the time it's bland homogenized lab-created garbage. <mask> it isn't, it rarely is more than a slightly more independent shadow of this process.  Except in very rare circumstances a song or artist won't get play unless it conforms to very rigid guidelines, in some ways far stricter than those in the past. [NEWLINE] [NEWLINE] [NEWLINE] <mask> the internet has gone a long way to alleviate this, it's<mask> been somewhat of a fracturing and atomizing force.  Nothing is quite<mask> ubiquitous due to this nature. [NEWLINE] [NEWLINE] [NEWLINE] <mask> due to music in some ways becoming a science, a lot of the low-hanging fruit has already been claimed.  Just like in other art forms, the longer they exist, the harder it becomes to have breakout and revolutionary impacts. [NEWLINE] [NEWLINE] [NEWLINE] I will say I cannot think of an artist that has come up for the first time in the last 15 years and had the effect of some of those you reference. <mask> I will say that musical innovation is far from dead.  Especially in certain metal sub-genres and avant-garde areas. </s>
Label encoding: <s>ITT: OP is pretty much correct. [NEWLINE] [NEWLINE] [NEWLINE] Longer answer: the music scene has changed forever.  With the complete corporatization of almost all radio, "charts," and "awards," what is popular and influential today is very rarely significant.  99% of the time it's bland homogenized lab-created garbage.  When it isn't, it rarely is more than a slightly more independent shadow of this process.  Except in very rare circumstances a song or artist won't get play unless it conforms to very rigid guidelines, in some ways far stricter than those in the past. [NEWLINE] [NEWLINE] [NEWLINE] While the internet has gone a long way to alleviate this, it's also been somewhat of a fracturing and atomizing force.  Nothing is quite as ubiquitous due to this nature. [NEWLINE] [NEWLINE] [NEWLINE] Additionally due to music in some ways becoming a science, a lot of the low-hanging fruit has already been claimed.  Just like in other art forms, the longer they exist, the harder it becomes to have breakout and revolutionary impacts. [NEWLINE] [NEWLINE] [NEWLINE] I will say I cannot think of an artist that has come up for the first time in the last 15 years and had the effect of some of those you reference.  However I will say that musical innovation is far from dead.  Especially in certain metal sub-genres and avant-garde areas. </s>
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Something a little more like<mask> they have in the UK might be a direction to move toward: [NEWLINE] [NEWLINE] "Today only a small proportion of officers are authorised to use firearms. Latest Home Office figures show there were just 6,653 officers authorised to use firearms in England and Wales - about 5% of the total number.  None of which implies, of course, that the British police are somehow gun-free.  Each police force has its own firearms unit. Police armed response vehicles have been deployed<mask> 1991." [NEWLINE] [NEWLINE] (Source: [URL] ) [NEWLINE] [NEWLINE] <mask> in America, it might be different,<mask> of the high volume of legal and illegal weapons on the street,<mask> that doesn't mean one can't imagine a reduction in guns for the average cop (who would<mask> be trained to engage with the community more). [NEWLINE] [NEWLINE] And<mask><mask>, in the case of someone firing a weapon in the street, we want cops to be able to effect an arrest / stop that perp. <mask> in most of those scenarios, there are<mask> MANY more interventions that should have happened BEFORE the person went on a rampage (or whatever). [NEWLINE] [NEWLINE] And you can imagine<mask>, a well-trained, lightly armed cop could stop and just chat with a person of interest, and really communicate with them, versus the climate of having an armed officer just strolling around, not really interacting with the public.</s>
Label encoding: <s>Something a little more like what they have in the UK might be a direction to move toward: [NEWLINE] [NEWLINE] "Today only a small proportion of officers are authorised to use firearms. Latest Home Office figures show there were just 6,653 officers authorised to use firearms in England and Wales - about 5% of the total number.  None of which implies, of course, that the British police are somehow gun-free.  Each police force has its own firearms unit. Police armed response vehicles have been deployed since 1991." [NEWLINE] [NEWLINE] (Source: [URL] ) [NEWLINE] [NEWLINE] But in America, it might be different, because of the high volume of legal and illegal weapons on the street, but that doesn't mean one can't imagine a reduction in guns for the average cop (who would also be trained to engage with the community more). [NEWLINE] [NEWLINE] And I agree, in the case of someone firing a weapon in the street, we want cops to be able to effect an arrest / stop that perp.  But in most of those scenarios, there are also MANY more interventions that should have happened BEFORE the person went on a rampage (or whatever). [NEWLINE] [NEWLINE] And you can imagine where, a well-trained, lightly armed cop could stop and just chat with a person of interest, and really communicate with them, versus the climate of having an armed officer just strolling around, not really interacting with the public.</s>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Should we create federal legislature to make it illegal for employers to do that? [ENDQ] [NEWLINE] <mask> I talk about all these things like privilege, it is never to imply that there should be legislation to force people into accepting minorities, we are all aware of the downfalls of that. I only think it is important for people to admit that this bias exists in society,<mask> that is the only way I see of this bias disappearing.<mask> white employers understand that they have a bias towards "white" names, they might take a second to double check the stack they've thrown out and rethink the reasons they did that for. Personally, I know I have biases and prejudices, everyone does,<mask> that doesn't make them right.<mask><mask> the conversation about privilege has to happen<mask> that people admit to themselves, even<mask> to nobody else, that they have biases and prejudices, and work to counteract them. [NEWLINE] [NEWLINE] [STARTQ] Anywho,<mask> I cannot change doesn't bother me [ENDQ] [NEWLINE] By admitting to yourself you have biases (for or against any race or people) you can change your own behaviour towards others, to create a more positive experience for them. By having the conversation about privilege and societal bias (not being a tumblrina SJW, just having the conversation thatyou and I are having) you are able to plant the seed in others' heads, and lead to them reconsidering their own biases.</s>
Label encoding: <s> [STARTQ] Should we create federal legislature to make it illegal for employers to do that? [ENDQ] [NEWLINE] When I talk about all these things like privilege, it is never to imply that there should be legislation to force people into accepting minorities, we are all aware of the downfalls of that. I only think it is important for people to admit that this bias exists in society, as that is the only way I see of this bias disappearing. If white employers understand that they have a bias towards "white" names, they might take a second to double check the stack they've thrown out and rethink the reasons they did that for. Personally, I know I have biases and prejudices, everyone does, but that doesn't make them right. I think the conversation about privilege has to happen so that people admit to themselves, even if to nobody else, that they have biases and prejudices, and work to counteract them. [NEWLINE] [NEWLINE] [STARTQ] Anywho, what I cannot change doesn't bother me [ENDQ] [NEWLINE] By admitting to yourself you have biases (for or against any race or people) you can change your own behaviour towards others, to create a more positive experience for them. By having the conversation about privilege and societal bias (not being a tumblrina SJW, just having the conversation thatyou and I are having) you are able to plant the seed in others' heads, and lead to them reconsidering their own biases.</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, that's the joke.  She's supposed to look white. [NEWLINE] [NEWLINE] <mask>, there are very few actual natives playing native American roles in Hollywood.  It used to be a lucrative career for native Americans,<mask> the market is<mask> saturated with non natives that someone who could pass<mask> native may not be at all and still get the part.  You might think this means Latino's and other culturally invested people,<mask> this is not the case.  A white man with a big nose and a decent tan will do.  The famous "crying Indian" commercial,<mask> the guy cries after trash is thrown on the street was Italian, for instance. [NEWLINE] [NEWLINE] <mask> the problem becomes which devil would you rather deal with?  Hiring people you know don't belong, or hiring people that look enough like native Americans that you can't tell the difference? [NEWLINE] [NEWLINE] Finally, it's not necessary to make an issue is race out of this. There's a diverse cultural tradition, with a complex history of varied peoples and a huge background to choose from, and all of that means exactly dick to Hollywood.  White people can play Indians, just like Australians can play Americans.  It's all about who looks and sounds the best for the part, which is finely crafted to be appealing or engaging on camera, and only gives a passing look at anything from the real world. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>First off, that's the joke.  She's supposed to look white. [NEWLINE] [NEWLINE] Secondly, there are very few actual natives playing native American roles in Hollywood.  It used to be a lucrative career for native Americans, but the market is so saturated with non natives that someone who could pass as native may not be at all and still get the part.  You might think this means Latino's and other culturally invested people, but this is not the case.  A white man with a big nose and a decent tan will do.  The famous "crying Indian" commercial, where the guy cries after trash is thrown on the street was Italian, for instance. [NEWLINE] [NEWLINE] So the problem becomes which devil would you rather deal with?  Hiring people you know don't belong, or hiring people that look enough like native Americans that you can't tell the difference? [NEWLINE] [NEWLINE] Finally, it's not necessary to make an issue is race out of this. There's a diverse cultural tradition, with a complex history of varied peoples and a huge background to choose from, and all of that means exactly dick to Hollywood.  White people can play Indians, just like Australians can play Americans.  It's all about who looks and sounds the best for the part, which is finely crafted to be appealing or engaging on camera, and only gives a passing look at anything from the real world. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Amen to that. Many many people claim to be egalitarians and aren't. Many people claim not to be racist<mask> are. Many people claim not to be sexist<mask> are. That's exactly the point. It's unfortunate,<mask> that's<mask> merely saying "I'm an egalitarian" or "I'm not racist" or "I'm not sexist" isn't actually terribly useful to anyone<mask> a descriptor,<mask> its<mask> many (<mask> not most) non-egalitarians, racists, and sexists will *<mask> * say publicly. [NEWLINE] [NEWLINE] <mask><mask>, and maybe this is a better response, saying "I'm an egalitarian" is merely a description of your goals,<mask> gives no indication of<mask> to achieve them or the current state of the world. An honest (<mask><mask><mask> incorrect) self-proclaimed egalitarian might believe that equal rights for men and women is the desired outcome.<mask> he or she might<mask> believe that we have already met that goal. I doubt a feminist would share that outlook.<mask> in that sense, the term feminism is more useful than "egalitarian"<mask> it combines a *goal* with a current diagnosis of our progress towards that goal. Namely that we haven't achieved it<mask> and women  in particular need advocacy. This is a core part of feminism and is not embedded in any way whatsoever in the term "egalitarian".</s>
Label encoding: <s>Amen to that. Many many people claim to be egalitarians and aren't. Many people claim not to be racist but are. Many people claim not to be sexist but are. That's exactly the point. It's unfortunate, but that's why merely saying "I'm an egalitarian" or "I'm not racist" or "I'm not sexist" isn't actually terribly useful to anyone as a descriptor, as its what many ( if not most) non-egalitarians, racists, and sexists will * also * say publicly. [NEWLINE] [NEWLINE] In addition, and maybe this is a better response, saying "I'm an egalitarian" is merely a description of your goals, but gives no indication of how to achieve them or the current state of the world. An honest ( but IMO incorrect) self-proclaimed egalitarian might believe that equal rights for men and women is the desired outcome. But he or she might also believe that we have already met that goal. I doubt a feminist would share that outlook. So in that sense, the term feminism is more useful than "egalitarian" because it combines a *goal* with a current diagnosis of our progress towards that goal. Namely that we haven't achieved it yet and women  in particular need advocacy. This is a core part of feminism and is not embedded in any way whatsoever in the term "egalitarian".</s>
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Good on you for mentioning the edit.<mask> it has a slight edge over one other flagship. [NEWLINE] [NEWLINE] Will that 'edge' manifest itself with software or multi-tasking? Will there ever be aps that run *better* on the 6 than on a G3 or Moto X? Or is it the functional equivalent of a 4-core chip-set hobbled by a 32 bit operating system (a gimmick)?<mask>, lets be honest. By the time software is developed exclusively for the I 6 &amp; above, the phone will be at least a year out of date. And you'll be only a few months out from writing this about the I-7. [NEWLINE] [NEWLINE] Will the battery retain its charge after 25 chargings'? 250? Has the battery even *existed on this earth* long enough to test? [NEWLINE] [NEWLINE] A lot of these 'hardware compromises' through the years are compromises<mask> $ was seen better spent on other, more tangible features. [NEWLINE] [NEWLINE] <mask> you were arguing laptops....sure. Hard to beat a Mac Book.<mask> there is *<mask> much* more to phones than simple "hardware". [NEWLINE] [NEWLINE] Without value, durability, customization, and a whole host of other things consumers care about...I just think its a moot point.<mask> you're buying the phone for the hardware alone, you don't understand<mask> makes a phone "good". </s>
Label encoding: <s>Good on you for mentioning the edit. So it has a slight edge over one other flagship. [NEWLINE] [NEWLINE] Will that 'edge' manifest itself with software or multi-tasking? Will there ever be aps that run *better* on the 6 than on a G3 or Moto X? Or is it the functional equivalent of a 4-core chip-set hobbled by a 32 bit operating system (a gimmick)? Because, lets be honest. By the time software is developed exclusively for the I 6 &amp; above, the phone will be at least a year out of date. And you'll be only a few months out from writing this about the I-7. [NEWLINE] [NEWLINE] Will the battery retain its charge after 25 chargings'? 250? Has the battery even *existed on this earth* long enough to test? [NEWLINE] [NEWLINE] A lot of these 'hardware compromises' through the years are compromises because $ was seen better spent on other, more tangible features. [NEWLINE] [NEWLINE] If you were arguing laptops....sure. Hard to beat a Mac Book. But there is * so much* more to phones than simple "hardware". [NEWLINE] [NEWLINE] Without value, durability, customization, and a whole host of other things consumers care about...I just think its a moot point. If you're buying the phone for the hardware alone, you don't understand what makes a phone "good". </s>
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Well,<mask> the group has to splinter into smaller groups,<mask> not do it along the lines of individual issues rather than along the lines of the people affected? [ENDQ] [NEWLINE] They tend to splinter into smaller groups to try and affect the one thing that they're passionate about. Like environmentalists who fight Japanese whalers. [NEWLINE] [NEWLINE] [STARTQ] Rape is something that affects people of genders, ages and ethnicities,<mask><mask> about an anti-rape advocacy group? Or an anti-domestic violence advocacy group?<mask> discussing racism, we know that forming groups advocating "white rights" or "black rights" would only add to the racism in the world. [ENDQ] [NEWLINE] I don't know anyone who is anti-rape or anti-domestic violence<mask> ignores men. Feminism is very aware that men are abused in this system (especially prison), and want to do something about it. [NEWLINE] [NEWLINE] And no, groups that advocate for a specific demographic do not add to the division. [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask> do we even need group mentality to tackle issues like this in the first place. Can't people just individually do the right thing without having to belong to a special group? [ENDQ] [NEWLINE] <mask> people care more about the groups they belong to.<mask> do you even mean by "do the right thing" here? There are people who think all of feminism's claims are complete and total nonsense.</s>
Label encoding: <s> [STARTQ] Well, if the group has to splinter into smaller groups, why not do it along the lines of individual issues rather than along the lines of the people affected? [ENDQ] [NEWLINE] They tend to splinter into smaller groups to try and affect the one thing that they're passionate about. Like environmentalists who fight Japanese whalers. [NEWLINE] [NEWLINE] [STARTQ] Rape is something that affects people of genders, ages and ethnicities, so how about an anti-rape advocacy group? Or an anti-domestic violence advocacy group? When discussing racism, we know that forming groups advocating "white rights" or "black rights" would only add to the racism in the world. [ENDQ] [NEWLINE] I don't know anyone who is anti-rape or anti-domestic violence but ignores men. Feminism is very aware that men are abused in this system (especially prison), and want to do something about it. [NEWLINE] [NEWLINE] And no, groups that advocate for a specific demographic do not add to the division. [NEWLINE] [NEWLINE] [STARTQ] Besides, why do we even need group mentality to tackle issues like this in the first place. Can't people just individually do the right thing without having to belong to a special group? [ENDQ] [NEWLINE] Because people care more about the groups they belong to. What do you even mean by "do the right thing" here? There are people who think all of feminism's claims are complete and total nonsense.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I<mask> prefaced half of<mask> you quoted me saying with being on a radical stance. Obviously that's not<mask> I truly feel,<mask> I put<mask> I truly feel<mask> my first line on the comment. [NEWLINE] [NEWLINE] <mask><mask> with you that<mask> I was a cop I would want to wear a camera for my own personal protection in times<mask> I'm accused of something I didn't do. Again I stated those opinions were just<mask> *some* people think. You can't just think about your own views<mask> thinking about something this big (the police). You have to think about all the factors in order to have a real conversation about it,<mask><mask> the hypothetical about the job story I mentioned was inaccurate you know people will have inaccurate impressions about anything. Look at net neutrality. A lot of big names still don't even understand<mask> the hell it is<mask><mask> all the information is at literally at their fingertips to help them fully understand it.<mask> you handle those people and your reactions will shape whether people like you enough to get behind your opinions on a subject.<mask> you're just going to call someones opinion stupid / dumb right off the bat then you're not going to build supporters. Instead you will just add to the list of people against you<mask> of<mask> you handled the situation. [NEWLINE] [NEWLINE] That obviously wasn't my intention with my previous comments,<mask> it's an afterthought.</s>
Label encoding: <s>I also prefaced half of what you quoted me saying with being on a radical stance. Obviously that's not how I truly feel, because I put how I truly feel as my first line on the comment. [NEWLINE] [NEWLINE] I agree with you that if I was a cop I would want to wear a camera for my own personal protection in times where I'm accused of something I didn't do. Again I stated those opinions were just what *some* people think. You can't just think about your own views when thinking about something this big (the police). You have to think about all the factors in order to have a real conversation about it, because while the hypothetical about the job story I mentioned was inaccurate you know people will have inaccurate impressions about anything. Look at net neutrality. A lot of big names still don't even understand what the hell it is even though all the information is at literally at their fingertips to help them fully understand it. How you handle those people and your reactions will shape whether people like you enough to get behind your opinions on a subject. If you're just going to call someones opinion stupid / dumb right off the bat then you're not going to build supporters. Instead you will just add to the list of people against you because of how you handled the situation. [NEWLINE] [NEWLINE] That obviously wasn't my intention with my previous comments, but it's an afterthought.</s>
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Let me ask you:<mask> many house parties have you been too<mask> you think "that was absolutely brilliant; I'll remember that for a long time" and everyone was drinking sensibly? Very few I imagine.<mask><mask><mask><mask> the parties that you most remember best are the ones<mask> someone is doing something unusual and hilarious due to the amount of alcohol they have had. [ENDQ] [NEWLINE] Doesn't that only apply to young people who just started drinking and haven't seen drunk people doing stupid shit before? I've never heard an adult describe a party<mask> "absolutely brilliant"<mask> someone got drunk and hurt themself. [NEWLINE] [NEWLINE] [STARTQ] Now, I had this conversation with a friend who argued the opposite bringing up the point that the person or people who drink too much may eventually become a burden to the whole party<mask> they're sick. I,<mask>, think that they would not be that much of a burden<mask> you can give them some water and leave them alone, plus they've probably spiced the party up with their antics for two hours beforehand. [ENDQ] [NEWLINE] Of all the parties I have ever been to,<mask> someone drinks far too much they are far more likely to end up vomitting on people/smashing their head on the toilet/breaking down crying than they are to do something really cool and memorable. Sure the latter happens sometimes,<mask> not nearly<mask> often<mask> the former.</s>
Label encoding: <s> [STARTQ] Let me ask you: how many house parties have you been too where you think "that was absolutely brilliant; I'll remember that for a long time" and everyone was drinking sensibly? Very few I imagine. On the other hand the parties that you most remember best are the ones where someone is doing something unusual and hilarious due to the amount of alcohol they have had. [ENDQ] [NEWLINE] Doesn't that only apply to young people who just started drinking and haven't seen drunk people doing stupid shit before? I've never heard an adult describe a party as "absolutely brilliant" because someone got drunk and hurt themself. [NEWLINE] [NEWLINE] [STARTQ] Now, I had this conversation with a friend who argued the opposite bringing up the point that the person or people who drink too much may eventually become a burden to the whole party if they're sick. I, however, think that they would not be that much of a burden as you can give them some water and leave them alone, plus they've probably spiced the party up with their antics for two hours beforehand. [ENDQ] [NEWLINE] Of all the parties I have ever been to, when someone drinks far too much they are far more likely to end up vomitting on people/smashing their head on the toilet/breaking down crying than they are to do something really cool and memorable. Sure the latter happens sometimes, but not nearly as often as the former.</s>
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Wouldn't you say that the ability to take a computer and put it in your pocket is a huge improvement over the desktop computer? [ENDQ] [NEWLINE] I see that<mask> being the inevitable result of mundane improvements in chip efficiencies, battery and display technology. They change the way we interact with computers due to linear improvements in hardware, and<mask> I could describe it<mask> "DOS in your pocket." One app at a time. All functions imprisoned by the app. Oh, and by the way, the spellchecker is built into the OS, now. Yay. [NEWLINE] [NEWLINE] [STARTQ] The fact that a computer can now access a large fraction of all human knowledge nearly-instantaneously? [ENDQ] [NEWLINE] That part is only getting interesting<mask> the "BBS++" of the Web now has Google. We're still just taking ancient ideas and making them bigger. The Internet is still not much more than a larger, prettier Compuserv. [NEWLINE] [NEWLINE] [STARTQ] To the point that I can go shopping, see an interesting vegetable in the store, remember that I happened to bring a computer with me, look up<mask> the vegetable is, and find a recipe that incorporates it without ever leaving the grocery store. [ENDQ] [NEWLINE] That's not much more than an improvement over Prodigy or AOL (before AOL became a clone of Huffington Post, back<mask> you used to log into it through a GeoWorks client).</s>
Label encoding: <s> [STARTQ] Wouldn't you say that the ability to take a computer and put it in your pocket is a huge improvement over the desktop computer? [ENDQ] [NEWLINE] I see that as being the inevitable result of mundane improvements in chip efficiencies, battery and display technology. They change the way we interact with computers due to linear improvements in hardware, and yet I could describe it as "DOS in your pocket." One app at a time. All functions imprisoned by the app. Oh, and by the way, the spellchecker is built into the OS, now. Yay. [NEWLINE] [NEWLINE] [STARTQ] The fact that a computer can now access a large fraction of all human knowledge nearly-instantaneously? [ENDQ] [NEWLINE] That part is only getting interesting because the "BBS++" of the Web now has Google. We're still just taking ancient ideas and making them bigger. The Internet is still not much more than a larger, prettier Compuserv. [NEWLINE] [NEWLINE] [STARTQ] To the point that I can go shopping, see an interesting vegetable in the store, remember that I happened to bring a computer with me, look up what the vegetable is, and find a recipe that incorporates it without ever leaving the grocery store. [ENDQ] [NEWLINE] That's not much more than an improvement over Prodigy or AOL (before AOL became a clone of Huffington Post, back when you used to log into it through a GeoWorks client).</s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It doesn't act<mask> a crime deterrent,<mask> would it? Crimes will be committed either way [ENDQ] [NEWLINE] <mask> you're a criminal, would you try to rob a person with a pistol holstered at his/her hip?<mask><mask> you walked into a convenience store and saw a patron or worker had a gun on them? It's ridiculous to think that *at least* the majority of criminals wouldn't say "screw this" and either go home or go find an easier place to rob. [NEWLINE] [NEWLINE] [STARTQ] Open carrying draws attention, creates a public sense of fear not safety...The only reason I can attribute to people open carrying is attention seeking and maybe a little bit of a power trip. [ENDQ] [NEWLINE] The intent of open carry isn't to demand fear or attention. It's to prevent violent action (<mask> in the examples stated above). Maybe that person lives in a bad neighborhood and wants to send the message to the thugs he walks by that messing with him is more trouble than it's worth. Certainly every person that notices someone's open firearm gets a good look at the person's face<mask> they'd have to be stupid to commit a crime<mask> the cops would get a great description of him. [NEWLINE] [NEWLINE] I understand that seeing open carry makes you uncomfortable.<mask><mask> you aren't against concealed carry then maybe the problem isn't the lawful citizen open carrying. It's your discomfort.</s>
Label encoding: <s> [STARTQ] It doesn't act as a crime deterrent, why would it? Crimes will be committed either way [ENDQ] [NEWLINE] If you're a criminal, would you try to rob a person with a pistol holstered at his/her hip? What if you walked into a convenience store and saw a patron or worker had a gun on them? It's ridiculous to think that *at least* the majority of criminals wouldn't say "screw this" and either go home or go find an easier place to rob. [NEWLINE] [NEWLINE] [STARTQ] Open carrying draws attention, creates a public sense of fear not safety...The only reason I can attribute to people open carrying is attention seeking and maybe a little bit of a power trip. [ENDQ] [NEWLINE] The intent of open carry isn't to demand fear or attention. It's to prevent violent action ( as in the examples stated above). Maybe that person lives in a bad neighborhood and wants to send the message to the thugs he walks by that messing with him is more trouble than it's worth. Certainly every person that notices someone's open firearm gets a good look at the person's face so they'd have to be stupid to commit a crime because the cops would get a great description of him. [NEWLINE] [NEWLINE] I understand that seeing open carry makes you uncomfortable. But if you aren't against concealed carry then maybe the problem isn't the lawful citizen open carrying. It's your discomfort.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The study you read about spanking very young children<mask> the "best form of discipline", would you give teachers the same authority considering it's been documented<mask> the best? [ENDQ] [NEWLINE] Depends on the type of environment. Many older Catholic Schools<mask> well<mask> British boarding schools use slaps on wrist<mask> forms of punishment.<mask> the institution is a private one<mask> the job of a teacher is closely connected to the child's well-being, then yes. [NEWLINE] [NEWLINE] <mask>, a teacher with a fixed salary in a public school cannot be trusted to keep the well-being of a child<mask> their primary concern. The key here is that the rod should not be an expression of personal anger<mask> rather a disciplinary method. [NEWLINE] [NEWLINE] <mask> you are against parents spanking their kids mildly then<mask> argument do you have in favor of ANY other disciplinary method? Grounding is equated to prison, doing chores is equated to slavery.<mask> disciplinary method do you approve of that cannot fall under the same argument? [NEWLINE] [NEWLINE] <mask> you want adult punishments like that of a boss-employee, is it okay<mask> I "terminate" my child's employment for bad behavior and ask him to clear his room and go out of the house and fend for himself? Or is it okay to ask my child to pay up for property damage and attach a debt to him, the same way it works for adults?</s>
Label encoding: <s> [STARTQ] The study you read about spanking very young children as the "best form of discipline", would you give teachers the same authority considering it's been documented as the best? [ENDQ] [NEWLINE] Depends on the type of environment. Many older Catholic Schools as well as British boarding schools use slaps on wrist as forms of punishment. If the institution is a private one where the job of a teacher is closely connected to the child's well-being, then yes. [NEWLINE] [NEWLINE] However, a teacher with a fixed salary in a public school cannot be trusted to keep the well-being of a child as their primary concern. The key here is that the rod should not be an expression of personal anger but rather a disciplinary method. [NEWLINE] [NEWLINE] If you are against parents spanking their kids mildly then what argument do you have in favor of ANY other disciplinary method? Grounding is equated to prison, doing chores is equated to slavery. What disciplinary method do you approve of that cannot fall under the same argument? [NEWLINE] [NEWLINE] If you want adult punishments like that of a boss-employee, is it okay if I "terminate" my child's employment for bad behavior and ask him to clear his room and go out of the house and fend for himself? Or is it okay to ask my child to pay up for property damage and attach a debt to him, the same way it works for adults?</s>
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This right here is the crux of the matter. [NEWLINE] [NEWLINE] The military is to protect the nation or to destroy enemies.   Trade ships/caravans might have military escorts to get through lands with bandits and that protects lives of citizens and their livelihoods,<mask> trade is critical to the nation. [NEWLINE] [NEWLINE] Now that we have evolved into a internationally interconnected world, the definition of "protecting the nation" is fuzzy.  It's easy to send the navy to protect ships from the Somalian pirates,<mask> instead of expanding the military to cover the entire trade route, at<mask> point does it make more sense to bomb the pirate base? [NEWLINE] That violates Somalian sovereignty,<mask> they have a weak government,<mask> do we go in and help them get stable and have them deal with it internally? <mask> piracy is due to huge fisherman unemployment resulting from generations of overfishing and we can't put more fish in the ocean in a timely fashion,<mask><mask> do? [NEWLINE] [NEWLINE] The entire mindset of the US use of military force is that the world is a dangerous and complex place and someone has to be the "big bad" to keep everyone in line until we get our shit together for world peace,<mask> better us than someone else. [NEWLINE] [NEWLINE] <mask><mask><mask> with your sentiment and believe military actions should be performed through the UN security council<mask> much<mask> possible.  </s>
Label encoding: <s>This right here is the crux of the matter. [NEWLINE] [NEWLINE] The military is to protect the nation or to destroy enemies.   Trade ships/caravans might have military escorts to get through lands with bandits and that protects lives of citizens and their livelihoods, as trade is critical to the nation. [NEWLINE] [NEWLINE] Now that we have evolved into a internationally interconnected world, the definition of "protecting the nation" is fuzzy.  It's easy to send the navy to protect ships from the Somalian pirates, but instead of expanding the military to cover the entire trade route, at what point does it make more sense to bomb the pirate base? [NEWLINE] That violates Somalian sovereignty, but they have a weak government, so do we go in and help them get stable and have them deal with it internally?  But piracy is due to huge fisherman unemployment resulting from generations of overfishing and we can't put more fish in the ocean in a timely fashion, so what do? [NEWLINE] [NEWLINE] The entire mindset of the US use of military force is that the world is a dangerous and complex place and someone has to be the "big bad" to keep everyone in line until we get our shit together for world peace, so better us than someone else. [NEWLINE] [NEWLINE] Although I agree with your sentiment and believe military actions should be performed through the UN security council as much as possible.  </s>
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to disagree with you on<mask> and<mask> child labor ended. It didn't end<mask> we moved to an industrial economy.<mask> anything, it got worse. Reports from the era stated that nearly a fifth of all children in the country were employed. Textile factories, for example, preferred child laborers<mask> their small hands could work between the moving parts of the mills more easily than that of adults. And<mask> the early Twentieth Century did see many attempts to end child labor, it wasn't until the Fair Labor Standards Act of 1938 that it was nationally abolished. You may recognize 1938<mask> being some 70 or<mask> years after America became fully industrialized. And those groups that were trying to abolish child labor? Ethics were absolutely their concern. They were<mask> trying to pass compulsory education laws,<mask> your statement on<mask> the industrial economy freed children to pursue education is<mask> false. [NEWLINE] [NEWLINE] Have a source. Will post more<mask> it's not four in the morning anymore. [NEWLINE] [New York Times piece on child labor, published 1904]( [URL].pdf?AWSAccessKeyId=AKIAJBTN455PTTBQQNRQ&amp;Expires=1404032445&amp;Signature=%2F0AISX%2BvDUUYMId0%2FlXWPHj4M2s%3D) [NEWLINE] </s>
Label encoding: <s>I have to disagree with you on why and when child labor ended. It didn't end because we moved to an industrial economy. If anything, it got worse. Reports from the era stated that nearly a fifth of all children in the country were employed. Textile factories, for example, preferred child laborers because their small hands could work between the moving parts of the mills more easily than that of adults. And while the early Twentieth Century did see many attempts to end child labor, it wasn't until the Fair Labor Standards Act of 1938 that it was nationally abolished. You may recognize 1938 as being some 70 or so years after America became fully industrialized. And those groups that were trying to abolish child labor? Ethics were absolutely their concern. They were also trying to pass compulsory education laws, so your statement on how the industrial economy freed children to pursue education is also false. [NEWLINE] [NEWLINE] Have a source. Will post more when it's not four in the morning anymore. [NEWLINE] [New York Times piece on child labor, published 1904]( [URL].pdf?AWSAccessKeyId=AKIAJBTN455PTTBQQNRQ&amp;Expires=1404032445&amp;Signature=%2F0AISX%2BvDUUYMId0%2FlXWPHj4M2s%3D) [NEWLINE] </s>
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To your first rebuttal, I was not listing reasons that the redskins should change their name,<mask> burdens that the name brings to the organization. I assume (maybe I'm wrong) that the Detroit Lions don't have to constantly make press releases and hire lawyers to draft appeals etc etc about their mascot. I'm saying that with a different name, problems that the team currently has would go away. I'm not saying that they should have to change it. That isn't<mask> my question is about. The question is,<mask> is the downside to just changing the name? [NEWLINE] [NEWLINE] To your second point, I am not wrong, or at least I don't think you showed me<mask> I was. The Redskins have lost their trademark before in 1999. It took 4 years for the appeals case. The trademark is valid for that entire period. Even<mask> the Redskins eventually lose their appeal this time, it won't be until at least 2017. It seems to me that<mask> they changed their name, they would move a ton of product, and get a ton of PR. [NEWLINE] [NEWLINE] I read your article. Good read and I love Rick Reily,<mask> again I'm not asking whether people should be forcing the name change. I'm asking<mask>, other then nostalgia and not wanting to appear<mask><mask> you can be told<mask> to do, should the Redskins resist changing their name.</s>
Label encoding: <s>To your first rebuttal, I was not listing reasons that the redskins should change their name, but burdens that the name brings to the organization. I assume (maybe I'm wrong) that the Detroit Lions don't have to constantly make press releases and hire lawyers to draft appeals etc etc about their mascot. I'm saying that with a different name, problems that the team currently has would go away. I'm not saying that they should have to change it. That isn't what my question is about. The question is, What is the downside to just changing the name? [NEWLINE] [NEWLINE] To your second point, I am not wrong, or at least I don't think you showed me why I was. The Redskins have lost their trademark before in 1999. It took 4 years for the appeals case. The trademark is valid for that entire period. Even if the Redskins eventually lose their appeal this time, it won't be until at least 2017. It seems to me that if they changed their name, they would move a ton of product, and get a ton of PR. [NEWLINE] [NEWLINE] I read your article. Good read and I love Rick Reily, but again I'm not asking whether people should be forcing the name change. I'm asking why, other then nostalgia and not wanting to appear as if you can be told what to do, should the Redskins resist changing their name.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In the case of physical/sexual abuse and neglect, children SHOULD be taken away and can be legally. That leaves indoctrination. [NEWLINE] [NEWLINE] 1) Losing your parents and being raised by another family is traumatic. Furthermore,<mask> do you plan to put them? In the foster care system? In the United States the foster care system comes with issues of abuse (not necessarily indoctrination - more physical, sexual, etc.) too. We are allowed to raise our children<mask> the alternatives aren't<mask> pretty<mask> you'd imagine. [NEWLINE] [NEWLINE] 2) Who gets to decide<mask> type of principles get to be counted<mask> indoctrination? You are giving that power to the state.<mask> happens<mask> an administration gets elected that has far more radical ideals than you do, and believes that YOUR beliefs are dangerous? [NEWLINE] [NEWLINE] 3) People can still have children even after you take them away. Is your plan to, in essence, "neuter" those who don't meet the criteria set out by the state?<mask><mask> you can start to see, perhaps,<mask> you might be handing over an incredible power to the government. And<mask> you leave these people still capable of having children, whose to say they won't continue to produce more, which means more kids that you need to find a home for. Which leads me to... [NEWLINE] [NEWLINE] 4) Most foster care systems are already very full. </s>
Label encoding: <s>In the case of physical/sexual abuse and neglect, children SHOULD be taken away and can be legally. That leaves indoctrination. [NEWLINE] [NEWLINE] 1) Losing your parents and being raised by another family is traumatic. Furthermore, where do you plan to put them? In the foster care system? In the United States the foster care system comes with issues of abuse (not necessarily indoctrination - more physical, sexual, etc.) too. We are allowed to raise our children because the alternatives aren't as pretty as you'd imagine. [NEWLINE] [NEWLINE] 2) Who gets to decide what type of principles get to be counted as indoctrination? You are giving that power to the state. What happens when an administration gets elected that has far more radical ideals than you do, and believes that YOUR beliefs are dangerous? [NEWLINE] [NEWLINE] 3) People can still have children even after you take them away. Is your plan to, in essence, "neuter" those who don't meet the criteria set out by the state? I think you can start to see, perhaps, where you might be handing over an incredible power to the government. And if you leave these people still capable of having children, whose to say they won't continue to produce more, which means more kids that you need to find a home for. Which leads me to... [NEWLINE] [NEWLINE] 4) Most foster care systems are already very full. </s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is more behind school than you might imagine. The school is interessted in making their pupils into functioning members of society, they don't teach you math<mask> you can solve these very specific equations you learn during highschool, they want you to understand the greater logic behind it. The same goes for sport, the school isn't interessted in teaching you<mask> to play soccer just for the sake of it or<mask> they want to make you into a star soccer player,<mask> it has other benefits too: [NEWLINE] [NEWLINE] - People who are doing sport are generaly fitter than those who don't, that doesn't only decrease obesity<mask><mask> makes for better students overall [NEWLINE] - People who attend sports regularly without being forced to (like they are<mask> attending schools) will learn to form habits and stick to them<mask> they are benefitial,<mask><mask> its uncomfortable from time to time (For example<mask> its raining outside, you really aren't in the mood for it or something like that) you are not giving up<mask> easily [NEWLINE] - You learn things like teamwork or the ability to make quick decisions or overall skills that are usefull<mask> playing sports,<mask> are<mask> usefull somewhere else [NEWLINE] - The social factors is<mask> pretty huge, the school doesn't want 1000 students who are foreign to one another, its generally better for everyone<mask> students are binding, sport is excellent for that</s>
Label encoding: <s>There is more behind school than you might imagine. The school is interessted in making their pupils into functioning members of society, they don't teach you math so you can solve these very specific equations you learn during highschool, they want you to understand the greater logic behind it. The same goes for sport, the school isn't interessted in teaching you how to play soccer just for the sake of it or because they want to make you into a star soccer player, but it has other benefits too: [NEWLINE] [NEWLINE] - People who are doing sport are generaly fitter than those who don't, that doesn't only decrease obesity but also makes for better students overall [NEWLINE] - People who attend sports regularly without being forced to (like they are when attending schools) will learn to form habits and stick to them when they are benefitial, even though its uncomfortable from time to time (For example when its raining outside, you really aren't in the mood for it or something like that) you are not giving up as easily [NEWLINE] - You learn things like teamwork or the ability to make quick decisions or overall skills that are usefull when playing sports, but are also usefull somewhere else [NEWLINE] - The social factors is also pretty huge, the school doesn't want 1000 students who are foreign to one another, its generally better for everyone when students are binding, sport is excellent for that</s>
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Without knowing anything about your neighbour, you'd rather him having a gun than weed? [ENDQ] [NEWLINE] Yes, without the shadow of a doubt. Your average neighbor is at least a decent person, and isn't going to go around shooting people. Plus, I would feel safer knowing that<mask> something went bad then somebody nearby would be both armed and on the side of keeping the neighborhood safe. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> he's suffering from depression? Or has a cheating wife? Still all good? [ENDQ] [NEWLINE] Still all good. Having problems and becoming mentally deranged and shooting up something are completely different issues. Most people have problems, and many are depressed. I don't worry they'll take a knife or baseball bat to me, and I don't see it being any different<mask> they have a gun. [NEWLINE] [NEWLINE] [STARTQ] You are very misinformed about weed, btw. [ENDQ] [NEWLINE] Perhaps. I have never used it and plan on keeping it that way. Here's<mask> I do know: [NEWLINE] [NEWLINE] * Weed is illegal, at least<mask> I live [NEWLINE] [NEWLINE] * Weed, in some manner, changes<mask> your brain processes information in at least some minor way. [NEWLINE] [NEWLINE] <mask>, someone smoking weed is someone who has no respect for the law, being an active criminal, and to some degree isn't thinking clearly. That,<mask><mask><mask>, is much more dangerous than a gun.</s>
Label encoding: <s> [STARTQ] Without knowing anything about your neighbour, you'd rather him having a gun than weed? [ENDQ] [NEWLINE] Yes, without the shadow of a doubt. Your average neighbor is at least a decent person, and isn't going to go around shooting people. Plus, I would feel safer knowing that if something went bad then somebody nearby would be both armed and on the side of keeping the neighborhood safe. [NEWLINE] [NEWLINE] [STARTQ] What if he's suffering from depression? Or has a cheating wife? Still all good? [ENDQ] [NEWLINE] Still all good. Having problems and becoming mentally deranged and shooting up something are completely different issues. Most people have problems, and many are depressed. I don't worry they'll take a knife or baseball bat to me, and I don't see it being any different if they have a gun. [NEWLINE] [NEWLINE] [STARTQ] You are very misinformed about weed, btw. [ENDQ] [NEWLINE] Perhaps. I have never used it and plan on keeping it that way. Here's what I do know: [NEWLINE] [NEWLINE] * Weed is illegal, at least where I live [NEWLINE] [NEWLINE] * Weed, in some manner, changes how your brain processes information in at least some minor way. [NEWLINE] [NEWLINE] Thus, someone smoking weed is someone who has no respect for the law, being an active criminal, and to some degree isn't thinking clearly. That, in my opinion, is much more dangerous than a gun.</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're not entirely wrong<mask>,<mask><mask> their skin colour does not affect them personally,<mask> people's skin colour does define is<mask> kind of people they are with. Whether they will admit to it or not there are cultures based on skin colour, religion, country of origins, interests etc., some people do not succumb to these cultures,<mask> many do, here in england it's not uncommon to see in school many of the black people hang out together, many poles (especially the poles), gamers, people of the same economic status etc., now it is not like 90% of people do this,<mask> closer to 30% I'd say, I was part of a more multi-cultural group,<mask> we had pretty much everyone hang out with everyone,<mask> there were those that would hang out only with people that were similar to them in a sense. Heck, in college this black guy I've been hanging out told me, and I will never forget this,<mask> his friends said he's acting 'too white' in context of him furthering his education, unfortunately these cultures are reinforced from both sides. Fortunately this might not be<mask> much of a factor. Unfortunately it might. [NEWLINE] [NEWLINE] I like your question, it's interesting, it's thought provoking, I don't like<mask> people straw man your intentions, don't feel bad, some people want to be offended.</s>
Label encoding: <s>You're not entirely wrong though, even though their skin colour does not affect them personally, what people's skin colour does define is what kind of people they are with. Whether they will admit to it or not there are cultures based on skin colour, religion, country of origins, interests etc., some people do not succumb to these cultures, but many do, here in england it's not uncommon to see in school many of the black people hang out together, many poles (especially the poles), gamers, people of the same economic status etc., now it is not like 90% of people do this, but closer to 30% I'd say, I was part of a more multi-cultural group, where we had pretty much everyone hang out with everyone, but there were those that would hang out only with people that were similar to them in a sense. Heck, in college this black guy I've been hanging out told me, and I will never forget this, how his friends said he's acting 'too white' in context of him furthering his education, unfortunately these cultures are reinforced from both sides. Fortunately this might not be as much of a factor. Unfortunately it might. [NEWLINE] [NEWLINE] I like your question, it's interesting, it's thought provoking, I don't like how people straw man your intentions, don't feel bad, some people want to be offended.</s>
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "Discussion between developers" doesn't mean anything<mask> the developers don't act on it. [ENDQ] [NEWLINE] That's my point, they *do* act on it. It's precisely<mask> of western developers awareness and discussion of these issues that it's better about gender dynamics overall. We have games like the Borderlands series which are fantastic at being inclusive. Or Bioshock Infinity, The Last of Us, Mass Effect, Portal, and many others which allow you to play<mask> a woman or prominently feature women<mask> major characters. We have more in terms of strong, interesting, or dynamic female characters in games<mask><mask><mask> of these discussions happening. [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask> millions of sales, yes, Tomb Raider was considered a flop.. [ENDQ] [NEWLINE] <mask><mask> this is misleading. It was a flop financially<mask> of the amount of money they sunk into it making it, not<mask> the sales were poor in general, or poor relative to the rest of the series. It was critically acclaimed, topped charts, broke franchise records, and was second in sales only to Bioshock Infinity. Being a non-sexist portrayal of Croft didn't hurt the game at all. It's only failure was the executives overambitious goals for the game. The game showed that players were totally fine playing an interpretation of Lara Craft that wasn't completely over-sexualized. </s>
Label encoding: <s> [STARTQ] "Discussion between developers" doesn't mean anything if the developers don't act on it. [ENDQ] [NEWLINE] That's my point, they *do* act on it. It's precisely because of western developers awareness and discussion of these issues that it's better about gender dynamics overall. We have games like the Borderlands series which are fantastic at being inclusive. Or Bioshock Infinity, The Last of Us, Mass Effect, Portal, and many others which allow you to play as a woman or prominently feature women as major characters. We have more in terms of strong, interesting, or dynamic female characters in games as a result of these discussions happening. [NEWLINE] [NEWLINE] [STARTQ] Also, despite millions of sales, yes, Tomb Raider was considered a flop.. [ENDQ] [NEWLINE] I think this is misleading. It was a flop financially because of the amount of money they sunk into it making it, not because the sales were poor in general, or poor relative to the rest of the series. It was critically acclaimed, topped charts, broke franchise records, and was second in sales only to Bioshock Infinity. Being a non-sexist portrayal of Croft didn't hurt the game at all. It's only failure was the executives overambitious goals for the game. The game showed that players were totally fine playing an interpretation of Lara Craft that wasn't completely over-sexualized. </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] love your Comment and you bring up good points and have a very nice structure. Here are my rebuttals: [NEWLINE] [NEWLINE] 1. There are tests being used already in adoption agencies, and no one seems to be saying those are too subjective. I would say that their criteria is doing fine,<mask> we would use it. [NEWLINE] 2. This is true,<mask> it's obvious that the people who abuse the system are the same people who don't deserve children. I'm not promising perfection, only something that might improve things :) [NEWLINE] 3. The hope is that you are checked out beforehand, and<mask> you didn't test<mask> responsible enough you're not going to do it. Obviously that is a flaw<mask> most people are not that honorable,<mask> I do<mask> I can.<mask> I understand that taking children away can be hard to swallow, it could be for that child's benefit to not be with those parents. It's implied that<mask> you're a responsible parent, you won't have a problem with taking this test to prove to everyone that you're fine, [NEWLINE] 4. This is an unavoidable fact, and it is a very good point that I have not seen before. Delta for you [NEWLINE] 5.<mask> mentioned previously, I do not promise perfection! Nearly no laws have completely eliminated all of its intended targets. I only hope this will improve things [NEWLINE] [NEWLINE] </s>
Label encoding: <s>∆ [NEWLINE] love your Comment and you bring up good points and have a very nice structure. Here are my rebuttals: [NEWLINE] [NEWLINE] 1. There are tests being used already in adoption agencies, and no one seems to be saying those are too subjective. I would say that their criteria is doing fine, so we would use it. [NEWLINE] 2. This is true, but it's obvious that the people who abuse the system are the same people who don't deserve children. I'm not promising perfection, only something that might improve things :) [NEWLINE] 3. The hope is that you are checked out beforehand, and if you didn't test as responsible enough you're not going to do it. Obviously that is a flaw as most people are not that honorable, but I do what I can. While I understand that taking children away can be hard to swallow, it could be for that child's benefit to not be with those parents. It's implied that if you're a responsible parent, you won't have a problem with taking this test to prove to everyone that you're fine, [NEWLINE] 4. This is an unavoidable fact, and it is a very good point that I have not seen before. Delta for you [NEWLINE] 5. As mentioned previously, I do not promise perfection! Nearly no laws have completely eliminated all of its intended targets. I only hope this will improve things [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> you live in the US or in the EU or whatever.<mask><mask><mask> you miss something very simple. There is value in learning a language that you can use, there is less value in learning a language that you won't use. [NEWLINE] [NEWLINE] In the case of an European, (I'm taking the Dutch<mask> an example) [NEWLINE] Germany is our biggest trading partner, there is value in learning German for economic reasons. Germany is bordering to us, increasing the amount of times we meet, there is value in learning German for social reasons, like vacations or friendships. which isn't there for Asian languages,<mask> the people aren't vacationing that often in Asian countries<mask> they do in our neighbouring countries. People that we meet in Europe often can speak an european language, even the ones that have origins in Asia,<mask> it is easier to just keep to learning European languages<mask> you are planning to stay inside of Europe. [NEWLINE] [NEWLINE] One of the best ways to learn a new language is immersion. It is easier to do<mask> in Europe<mask> an European. We have English, French and German<mask> language options on TV for example. And<mask> TV isn't enough we can just travel a few hours and go to a country<mask> the language is spoken.<mask> there is barely any chance to immerse yourself in Asian languages the same way. [NEWLINE] </s><pad>
Label encoding: <s>I don't know if you live in the US or in the EU or whatever. But I think you miss something very simple. There is value in learning a language that you can use, there is less value in learning a language that you won't use. [NEWLINE] [NEWLINE] In the case of an European, (I'm taking the Dutch as an example) [NEWLINE] Germany is our biggest trading partner, there is value in learning German for economic reasons. Germany is bordering to us, increasing the amount of times we meet, there is value in learning German for social reasons, like vacations or friendships. which isn't there for Asian languages, because the people aren't vacationing that often in Asian countries as they do in our neighbouring countries. People that we meet in Europe often can speak an european language, even the ones that have origins in Asia, so it is easier to just keep to learning European languages if you are planning to stay inside of Europe. [NEWLINE] [NEWLINE] One of the best ways to learn a new language is immersion. It is easier to do so in Europe as an European. We have English, French and German as language options on TV for example. And if TV isn't enough we can just travel a few hours and go to a country where the language is spoken. While there is barely any chance to immerse yourself in Asian languages the same way. [NEWLINE] </s><pad>
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm saying that I knew there were risks<mask> I took the job. [ENDQ] [NEWLINE] Right, and even<mask> you took every precaution, something might happen.  Well that risk is just inherent to being a woman - they can't make a cost benefit analysis on that<mask>. [NEWLINE] [NEWLINE] [STARTQ] I'm saying, "yes, that is unfair,<mask> that's<mask> you should wear your seatbelt [ENDQ] [NEWLINE] And the problem I have with this is that: [NEWLINE] [NEWLINE] Everyone should wear their seatbelt. <mask> you really want to have this talk with kids, tell ALL kids not to be drunk in public -<mask><mask> is clear, any number of ills can befall you<mask> you are. <mask> the problem, I'd say, is that it doesn't really address the vast number of rapes, which are acquaintance rapes.  I mean, sure, you don't go to bars anymore,<mask> do you drink with your friends?  Have an 'eccentric' family member?  Go on dates? [NEWLINE] [NEWLINE] Any of those situations,<mask> *I* personally don't have "I could be raped" on my mind could a be a situation that ends up with a woman being raped.  The inclination to tell a rape victim to be more careful next time is in almost every situation unnecessary, and in many situations, not even particularly applicable. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I'm saying that I knew there were risks when I took the job. [ENDQ] [NEWLINE] Right, and even if you took every precaution, something might happen.  Well that risk is just inherent to being a woman - they can't make a cost benefit analysis on that though. [NEWLINE] [NEWLINE] [STARTQ] I'm saying, "yes, that is unfair, but that's why you should wear your seatbelt [ENDQ] [NEWLINE] And the problem I have with this is that: [NEWLINE] [NEWLINE] Everyone should wear their seatbelt.  If you really want to have this talk with kids, tell ALL kids not to be drunk in public - because as is clear, any number of ills can befall you if you are.  But the problem, I'd say, is that it doesn't really address the vast number of rapes, which are acquaintance rapes.  I mean, sure, you don't go to bars anymore, but do you drink with your friends?  Have an 'eccentric' family member?  Go on dates? [NEWLINE] [NEWLINE] Any of those situations, where *I* personally don't have "I could be raped" on my mind could a be a situation that ends up with a woman being raped.  The inclination to tell a rape victim to be more careful next time is in almost every situation unnecessary, and in many situations, not even particularly applicable. [NEWLINE] </s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A few Caveats: The rich old man had no heirs, No will/testament, absolutely no living person in his lineage except him, he died a lonely bachelor and had no security, dog, etc. I just feel that the poor family who could never afford a car would benefit far greater from it than that old selfish man! [ENDQ] [NEWLINE] In this incredibly specific, rather unrealistic set of circumstances, sure, I wouldn't hold it against them.<mask><mask> likely is this situation? [NEWLINE] [NEWLINE] Let's go with a more likely scenario and say a poor person stole from a living rich person.<mask><mask> that rich person was going to invest that money in a company and employ people?<mask><mask> he was going to give it to charity?<mask><mask> he was going to use it to create a foundation to end homelessness and poverty? And<mask><mask> the poor person was going to spend it on drugs? [NEWLINE] [NEWLINE] <mask> wealth is being transferred from rich to poor through means of theft and crime, there is no accountability to make sure that money gets used for anything of worth or will be put back into the economy in meaningful way. [NEWLINE] [NEWLINE] I'll be the first person to admit that inequality of wealth is a huge problem in our society. I always advocate for legal means of leveling the playing field,<mask> I would never try and justify stealing to accomplish those goals. </s>
Label encoding: <s> [STARTQ] A few Caveats: The rich old man had no heirs, No will/testament, absolutely no living person in his lineage except him, he died a lonely bachelor and had no security, dog, etc. I just feel that the poor family who could never afford a car would benefit far greater from it than that old selfish man! [ENDQ] [NEWLINE] In this incredibly specific, rather unrealistic set of circumstances, sure, I wouldn't hold it against them. But how likely is this situation? [NEWLINE] [NEWLINE] Let's go with a more likely scenario and say a poor person stole from a living rich person. What if that rich person was going to invest that money in a company and employ people? What if he was going to give it to charity? What if he was going to use it to create a foundation to end homelessness and poverty? And what if the poor person was going to spend it on drugs? [NEWLINE] [NEWLINE] If wealth is being transferred from rich to poor through means of theft and crime, there is no accountability to make sure that money gets used for anything of worth or will be put back into the economy in meaningful way. [NEWLINE] [NEWLINE] I'll be the first person to admit that inequality of wealth is a huge problem in our society. I always advocate for legal means of leveling the playing field, but I would never try and justify stealing to accomplish those goals. </s>
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In other words, the farmer increased his labor costs by 10%, and then had to increase his margins by 10%<mask> to keep the same level of profit. Ditto baker, distributor, et al. [ENDQ] [NEWLINE] Sure. And the end result is the the price of the product goes up by 10%. [NEWLINE] [NEWLINE] Not 10% compounded for each person involved. 10%, period. [NEWLINE] [NEWLINE] [STARTQ] This is really dangerous,<mask> in the case that costs increase the same amount<mask> minimum wage -- i.e. minimum wage workers earn 10% more and goods cost 10% more -- then they have no additional buying power, i.e. they have no more real wages than they did before, and everyone else's real wages have decreased, meaning everyone is worse off. [ENDQ] [NEWLINE] You contradict yourself here. [NEWLINE] [NEWLINE] <mask> everyone is a minimum wage worker, then yes, the price of a loaf of bread will go up by 10%...<mask> then there is no "everyone else". [NEWLINE] [NEWLINE] <mask> *not* everyone is a minimum wage worker, then the price of a loaf of bread will go up by less than 10%, and minimum wage workers will be better off. [NEWLINE] [NEWLINE] You can't have it both ways - either the price of bread goes up by less than the minimum wage increase, or everyone in the entire world is already making minimum wage.</s><pad><pad>
Label encoding: <s> [STARTQ] In other words, the farmer increased his labor costs by 10%, and then had to increase his margins by 10% also to keep the same level of profit. Ditto baker, distributor, et al. [ENDQ] [NEWLINE] Sure. And the end result is the the price of the product goes up by 10%. [NEWLINE] [NEWLINE] Not 10% compounded for each person involved. 10%, period. [NEWLINE] [NEWLINE] [STARTQ] This is really dangerous, because in the case that costs increase the same amount as minimum wage -- i.e. minimum wage workers earn 10% more and goods cost 10% more -- then they have no additional buying power, i.e. they have no more real wages than they did before, and everyone else's real wages have decreased, meaning everyone is worse off. [ENDQ] [NEWLINE] You contradict yourself here. [NEWLINE] [NEWLINE] If everyone is a minimum wage worker, then yes, the price of a loaf of bread will go up by 10%... but then there is no "everyone else". [NEWLINE] [NEWLINE] If *not* everyone is a minimum wage worker, then the price of a loaf of bread will go up by less than 10%, and minimum wage workers will be better off. [NEWLINE] [NEWLINE] You can't have it both ways - either the price of bread goes up by less than the minimum wage increase, or everyone in the entire world is already making minimum wage.</s><pad><pad>
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> saying stuff like 'that's emotional bullshit' is probably really hurtful to someone who was deeply affected like some people who already posted in this thread. Most men may be fine after incidences like this, some aren't. You should read some of these people's posts<mask> man, they're eye opening. [ENDQ] [NEWLINE] You're right. I just get<mask> tired of Reddit circle jerking that video pretending that is<mask> everyone would normally react. I need to control my reactions to these things. [NEWLINE] [NEWLINE] <mask><mask> really bugs me is this actor is talking about statutory rape and not actual rape. Notice that not once in the video did he ever says no to the adult. He even said he physically enjoyed it.<mask> I'm not saying that some boys couldn't react poorly. Some would,<mask> it isn't standard(maybe it is this way<mask> of society,<mask> the average teenage boy still doesn't feel taken advantage of). It just spreads the propaganda that teenage boys wouldn't enjoy sex with a hot adult girl (and vise versa), and they are actually being taken advantage of. It continues this idea that is okay to send adults to jail for years, let them actually get raped in jail multiple times, and put them on the sex offenders list for the rest of their life, all<mask> of something that in all likelihood the boy actually enjoyed.</s>
Label encoding: <s> [STARTQ] Although saying stuff like 'that's emotional bullshit' is probably really hurtful to someone who was deeply affected like some people who already posted in this thread. Most men may be fine after incidences like this, some aren't. You should read some of these people's posts though man, they're eye opening. [ENDQ] [NEWLINE] You're right. I just get so tired of Reddit circle jerking that video pretending that is how everyone would normally react. I need to control my reactions to these things. [NEWLINE] [NEWLINE] What also really bugs me is this actor is talking about statutory rape and not actual rape. Notice that not once in the video did he ever says no to the adult. He even said he physically enjoyed it. While I'm not saying that some boys couldn't react poorly. Some would, but it isn't standard(maybe it is this way because of society, but the average teenage boy still doesn't feel taken advantage of). It just spreads the propaganda that teenage boys wouldn't enjoy sex with a hot adult girl (and vise versa), and they are actually being taken advantage of. It continues this idea that is okay to send adults to jail for years, let them actually get raped in jail multiple times, and put them on the sex offenders list for the rest of their life, all because of something that in all likelihood the boy actually enjoyed.</s>
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I had to look up<mask> SJWs are, Lol! Yeah, the first time I heard this term I sort of cringed, for two reasons. One, it harkens back to simply calling people "colored" which was actually preferable at the time to "negro" or that other famous n word...<mask> much<mask> that the term survives<mask> part of the NAACP acronym. And<mask>, it still would have racially negative connotations<mask> used today. Two, it just sounds like<mask> another misguided attempt by uber-liberal academia, the SJWs, to decide for all the opressed non-white people<mask> they should be called. It doesn't seem like an organic self label, or identity,  at all. And<mask>... [NEWLINE] [NEWLINE] All the terms have problems. Remember "African-American?" Aside from being American-centric, I don't know a lot of black people who identify<mask> African. Or use that term to self describe, like, ever. (aside from maybe a college application. ) The inherent problem here is that any word, or set of words, are going to be problematic, <mask> language is always going to be used in this context to draw distinction. To seperate, with simple labels, based on largely superficial factors. There's really no way to do that in a way that makes people happy. </s>
Label encoding: <s>I had to look up what SJWs are, Lol! Yeah, the first time I heard this term I sort of cringed, for two reasons. One, it harkens back to simply calling people "colored" which was actually preferable at the time to "negro" or that other famous n word... so much so that the term survives as part of the NAACP acronym. And yet, it still would have racially negative connotations if used today. Two, it just sounds like yet another misguided attempt by uber-liberal academia, the SJWs, to decide for all the opressed non-white people what they should be called. It doesn't seem like an organic self label, or identity,  at all. And yet... [NEWLINE] [NEWLINE] All the terms have problems. Remember "African-American?" Aside from being American-centric, I don't know a lot of black people who identify as African. Or use that term to self describe, like, ever. (aside from maybe a college application. ) The inherent problem here is that any word, or set of words, are going to be problematic,  because language is always going to be used in this context to draw distinction. To seperate, with simple labels, based on largely superficial factors. There's really no way to do that in a way that makes people happy. </s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who said I am denying anyone's rights to start a religion based on whatever he/she believes in? Better question<mask>,<mask> would I or anyone on the internet not call anyone out for anything?<mask> did I remotely suggested that<mask> a guy said is less true compared to some other guys who said it many years ago? Perhaps they're all bullshit, who knows.<mask> I'm arguing for is the fact that Scientology lacks the value in the historical context offered in Islam and Christianity, not to mention other factors that are notably mentioned in William James' The Varieties of Religious Experience, which contains religious testimonies that belongs to their own subject of study in the academics under the fields of religious studies, philosophy and psychology etc. We have many many reason to believe it (Scientology) is bullshit over the odds that it may be true, and<mask><mask><mask> it is more irrational to believe than not to believe.<mask>,<mask> I know a guy who is a sci-fi writer and said "You don't get rich writing science fiction.<mask> you want to get rich, you start a religion", and that he really did started a religion of his own, then I am compelled to believe it's straight up bullshit even<mask> it isn't. That's the simple truth - we don't know all about Islam and Christianity,<mask> we know too much enough about Scientology.</s>
Label encoding: <s>Who said I am denying anyone's rights to start a religion based on whatever he/she believes in? Better question yet, why would I or anyone on the internet not call anyone out for anything? When did I remotely suggested that what a guy said is less true compared to some other guys who said it many years ago? Perhaps they're all bullshit, who knows. What I'm arguing for is the fact that Scientology lacks the value in the historical context offered in Islam and Christianity, not to mention other factors that are notably mentioned in William James' The Varieties of Religious Experience, which contains religious testimonies that belongs to their own subject of study in the academics under the fields of religious studies, philosophy and psychology etc. We have many many reason to believe it (Scientology) is bullshit over the odds that it may be true, and for this reason it is more irrational to believe than not to believe. Moreover, if I know a guy who is a sci-fi writer and said "You don't get rich writing science fiction. If you want to get rich, you start a religion", and that he really did started a religion of his own, then I am compelled to believe it's straight up bullshit even if it isn't. That's the simple truth - we don't know all about Islam and Christianity, but we know too much enough about Scientology.</s>
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A key issue (from smooshie's link): [NEWLINE] [NEWLINE] [STARTQ] AAPIs represent over 30 countries and ethnic groups that speak over 100 different languages. [ENDQ] [NEWLINE] You can rest assured that no one in Asia lumps everyone on the continent into a monolithic group.  Koreans don't see themselves<mask> having anything to do with the Chinese, much less pacific islanders. [NEWLINE] [NEWLINE] To your point about "privilege/oppression" [NEWLINE] [NEWLINE] I read an editorial a<mask> back (have<mask> looked for it<mask> I can't find it) written for a school newspaper by an economics student whose parents are from India.  He tells a story of an argument he had with a whatever studies student about the gender pay gap,<mask> he made all the standard arguments you'd hear from an economist, and that at the end he was told to check his white privilege (which apparently was pretty amusing to him). [NEWLINE] [NEWLINE] I thought it was funny for a different reason, which is that the accusation made perfect sense to me.  Once I was talking to my mom about my friend's daughter and she asked "he's Indian right?"  And I said no, his parents are Indian he's American.  And to the extent white privilege is more than dressed up Marxist drivel, he has it, or at least some fraction of it.  </s>
Label encoding: <s>A key issue (from smooshie's link): [NEWLINE] [NEWLINE] [STARTQ] AAPIs represent over 30 countries and ethnic groups that speak over 100 different languages. [ENDQ] [NEWLINE] You can rest assured that no one in Asia lumps everyone on the continent into a monolithic group.  Koreans don't see themselves as having anything to do with the Chinese, much less pacific islanders. [NEWLINE] [NEWLINE] To your point about "privilege/oppression" [NEWLINE] [NEWLINE] I read an editorial a while back (have since looked for it but I can't find it) written for a school newspaper by an economics student whose parents are from India.  He tells a story of an argument he had with a whatever studies student about the gender pay gap, how he made all the standard arguments you'd hear from an economist, and that at the end he was told to check his white privilege (which apparently was pretty amusing to him). [NEWLINE] [NEWLINE] I thought it was funny for a different reason, which is that the accusation made perfect sense to me.  Once I was talking to my mom about my friend's daughter and she asked "he's Indian right?"  And I said no, his parents are Indian he's American.  And to the extent white privilege is more than dressed up Marxist drivel, he has it, or at least some fraction of it.  </s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Private prisons control their employment and training of staff. This can lead to lower levels of security, bribery and general reduction in safety in the prison environment. [NEWLINE] [NEWLINE] Prisoners become wage incentives. The right thing that needs to be done right now is focus on rehabilitation and helping people who have been given a criminal sentence to resettle back into society, which prevents short term sentence prisoners from reoffending and coming back into prison. This isn't in the interests of private prisons (less prisoners = less money) and<mask> private prisons do not serve the best interests of the public. [NEWLINE] [NEWLINE] Most private prisons have incentives to have a high turnover of prisoners and resettle them.<mask> they do adhere to these schemes (which is not common - see above), they often cut corners to meet targets. [NEWLINE] [NEWLINE] Essentially, the prisoners become numbers. The best interests of the human beings who are inside the prison are not held in high esteem. [NEWLINE] [NEWLINE] Just a heads up - my belief is not that the majority of prisoners are dangerous. I'm a firm believer of resettlement and desistance and thoroughly against private prisons.<mask> parts of the prison are privatised, that's not<mask> bad (i.e. cleaning services, healthcare, mental health care)<mask> they are often the reasons for failures.<mask> full privatisation is not a good thing long term.</s>
Label encoding: <s>Private prisons control their employment and training of staff. This can lead to lower levels of security, bribery and general reduction in safety in the prison environment. [NEWLINE] [NEWLINE] Prisoners become wage incentives. The right thing that needs to be done right now is focus on rehabilitation and helping people who have been given a criminal sentence to resettle back into society, which prevents short term sentence prisoners from reoffending and coming back into prison. This isn't in the interests of private prisons (less prisoners = less money) and so private prisons do not serve the best interests of the public. [NEWLINE] [NEWLINE] Most private prisons have incentives to have a high turnover of prisoners and resettle them. When they do adhere to these schemes (which is not common - see above), they often cut corners to meet targets. [NEWLINE] [NEWLINE] Essentially, the prisoners become numbers. The best interests of the human beings who are inside the prison are not held in high esteem. [NEWLINE] [NEWLINE] Just a heads up - my belief is not that the majority of prisoners are dangerous. I'm a firm believer of resettlement and desistance and thoroughly against private prisons. If parts of the prison are privatised, that's not so bad (i.e. cleaning services, healthcare, mental health care) as they are often the reasons for failures. But full privatisation is not a good thing long term.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Growing up with the internet<mask> many here did, I got my first music circa 2000 with Napster. I moved to Morpheus and KaZaA and Limewire<mask> I went from 28.8kbps modem to 56kbps and beyond (God, I would've murdered someone for a cable line in those early days). [NEWLINE] [NEWLINE] <mask> my download speeds increased,<mask> did the speed of my collecting. I went from making mix CDs to filling data CDs with mp3s to taking up significant portions of my hard drive with music. The thousands and thousands of hours spent on my computer in the basement were all spent listening to my music, and hearing certain songs will bring me back there in a flash. [NEWLINE] [NEWLINE] These days,<mask>, I only hear these songs incidentally on<mask> are now, lamentably, throwback stations - after neglecting to back up 30GB+ of music<mask> dutifully<mask> I should have, I discovered to my chagrin that my "archive" external hard drive had died somewhere in the 10 months between copying the music over and the untimely demise of my computer at the time. My dad still has all his vinyl from the 70s,<mask>, and even the wear on the jackets tells a thousand stories. [NEWLINE] [NEWLINE] Laptops burn, hard drives die.<mask> vinyl is forever.</s>
Label encoding: <s>Growing up with the internet as many here did, I got my first music circa 2000 with Napster. I moved to Morpheus and KaZaA and Limewire as I went from 28.8kbps modem to 56kbps and beyond (God, I would've murdered someone for a cable line in those early days). [NEWLINE] [NEWLINE] As my download speeds increased, so did the speed of my collecting. I went from making mix CDs to filling data CDs with mp3s to taking up significant portions of my hard drive with music. The thousands and thousands of hours spent on my computer in the basement were all spent listening to my music, and hearing certain songs will bring me back there in a flash. [NEWLINE] [NEWLINE] These days, though, I only hear these songs incidentally on what are now, lamentably, throwback stations - after neglecting to back up 30GB+ of music as dutifully as I should have, I discovered to my chagrin that my "archive" external hard drive had died somewhere in the 10 months between copying the music over and the untimely demise of my computer at the time. My dad still has all his vinyl from the 70s, though, and even the wear on the jackets tells a thousand stories. [NEWLINE] [NEWLINE] Laptops burn, hard drives die. But vinyl is forever.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] People will certainly expect him to more,<mask> once the chip is in hand it is much more reliable to examine it's actual color than try and guess at it by which bag it came out of. [ENDQ] [NEWLINE] no reasonable person would claim ethnicity is the only predictor of behavior. it is simply the most readily available and easiest one to use,<mask> most people's cognitive pathways default to it. [NEWLINE] [NEWLINE] [STARTQ] Unless we are arguing that black people are inherently more likely to commit crimes, independent of their upbringing. [ENDQ] [NEWLINE] based on data that i have seen relating parental income (<mask> a proxy for upbringing) to criminality, this unfortunately seems to be the case. [NEWLINE] [NEWLINE] [STARTQ] No need to look to ancient history<mask> the present has plenty of obstacles to the black community's success at large. [ENDQ] [NEWLINE] you are making a causal argument. **to do<mask> you have to establish a temporal relationship between cause and effect.**<mask> the effect you are claiming precedes your cause, the argument falls apart. you quite literally need to educate yourself on the relevant history. [NEWLINE] [NEWLINE] again, this is all ancillary to the original point, that differences in behavior exists, and people will notice them and react in predictable ways to protect themselves. those that don't like this need to change the behavior to change the perception, not object to people noticing reality. </s><pad>
Label encoding: <s> [STARTQ] People will certainly expect him to more, but once the chip is in hand it is much more reliable to examine it's actual color than try and guess at it by which bag it came out of. [ENDQ] [NEWLINE] no reasonable person would claim ethnicity is the only predictor of behavior. it is simply the most readily available and easiest one to use, so most people's cognitive pathways default to it. [NEWLINE] [NEWLINE] [STARTQ] Unless we are arguing that black people are inherently more likely to commit crimes, independent of their upbringing. [ENDQ] [NEWLINE] based on data that i have seen relating parental income ( as a proxy for upbringing) to criminality, this unfortunately seems to be the case. [NEWLINE] [NEWLINE] [STARTQ] No need to look to ancient history when the present has plenty of obstacles to the black community's success at large. [ENDQ] [NEWLINE] you are making a causal argument. **to do so you have to establish a temporal relationship between cause and effect.** if the effect you are claiming precedes your cause, the argument falls apart. you quite literally need to educate yourself on the relevant history. [NEWLINE] [NEWLINE] again, this is all ancillary to the original point, that differences in behavior exists, and people will notice them and react in predictable ways to protect themselves. those that don't like this need to change the behavior to change the perception, not object to people noticing reality. </s><pad>
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Velociraptors ran at 60mph.<mask> just "outrun" wouldn't work. [NEWLINE] [NEWLINE] You think you could beat one on agility? Think again. We enter the territory of the [square/cube law]( [URL] ). That's the reason spiders climb walls, ants lift many times their weight, cats fall mostly unharmed, fleas jump a meter, and whales asphyxiate on land. Being smaller makes you *more agile*, all else being equal. Ironically, bigger TV or movie-like raptors would be easier to defeat. [NEWLINE] [NEWLINE] <mask> "all else" isn't equal, is it?<mask> on top of that, velociraptors had hollow bones, that is, less hard mineral and more bone marrow. This made them *even more agile*, carrying less inertia. [NEWLINE] [NEWLINE] Your only hope would be to catch one from behind. Tough luck,<mask> they had their eyes on the sides of their heads, giving them huge peripheral vision. [NEWLINE] [NEWLINE] <mask> there's one more thing. Even<mask> you *could* beat a velociraptor one-on-one, there's a very important fact. [NEWLINE] [NEWLINE] They. Never. Hunt. Alone. [NEWLINE] [NEWLINE] PS: And before you could blink, you would have a nasty cut done by their hook-like thumb claw.</s>
Label encoding: <s>Velociraptors ran at 60mph. So just "outrun" wouldn't work. [NEWLINE] [NEWLINE] You think you could beat one on agility? Think again. We enter the territory of the [square/cube law]( [URL] ). That's the reason spiders climb walls, ants lift many times their weight, cats fall mostly unharmed, fleas jump a meter, and whales asphyxiate on land. Being smaller makes you *more agile*, all else being equal. Ironically, bigger TV or movie-like raptors would be easier to defeat. [NEWLINE] [NEWLINE] But "all else" isn't equal, is it? Because on top of that, velociraptors had hollow bones, that is, less hard mineral and more bone marrow. This made them *even more agile*, carrying less inertia. [NEWLINE] [NEWLINE] Your only hope would be to catch one from behind. Tough luck, because they had their eyes on the sides of their heads, giving them huge peripheral vision. [NEWLINE] [NEWLINE] But there's one more thing. Even if you *could* beat a velociraptor one-on-one, there's a very important fact. [NEWLINE] [NEWLINE] They. Never. Hunt. Alone. [NEWLINE] [NEWLINE] PS: And before you could blink, you would have a nasty cut done by their hook-like thumb claw.</s>
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I wasn't citing the first study specifically,<mask> I don't believe they ever expected it to work at keeping the weight off [ENDQ] [NEWLINE] Yea I'll give them the benefit of the doubt on that one, it just made it hard for me to take it seriously after reading that. [NEWLINE] [NEWLINE] [STARTQ] I'm a naturally thin person who doesn't have to do any of those things to maintain my weight,<mask> they seem very burdensome to me. [ENDQ] [NEWLINE] I can understand<mask> it seems like a daunting task. I've been eating to gain/lose weight (depending on<mask> I'm cutting or bulking) for a couple years now and<mask> at first it's difficult after about a month it becomes second nature. At a certain point it's no longer a diet, it's just the way you eat. And I don't deprive myself of anything. I eat basically garbage on the weekends<mask> all in moderation. [NEWLINE] [NEWLINE] Not to mention [NEWLINE] [NEWLINE] [STARTQ] Avoiding all sugar, bread, pasta, and junk food [ENDQ] [NEWLINE] This isn't even necessary.<mask> you're interested look up "<mask> It Fits Your Macros" or IIFYM. Essentially a style of diet<mask><mask><mask><mask> you are consuming the right amount of calories and macro nutrients, you eat whatever you want. [NEWLINE] [NEWLINE] <mask> I appreciate the good discussion.</s>
Label encoding: <s> [STARTQ] I wasn't citing the first study specifically, but I don't believe they ever expected it to work at keeping the weight off [ENDQ] [NEWLINE] Yea I'll give them the benefit of the doubt on that one, it just made it hard for me to take it seriously after reading that. [NEWLINE] [NEWLINE] [STARTQ] I'm a naturally thin person who doesn't have to do any of those things to maintain my weight, so they seem very burdensome to me. [ENDQ] [NEWLINE] I can understand how it seems like a daunting task. I've been eating to gain/lose weight (depending on if I'm cutting or bulking) for a couple years now and while at first it's difficult after about a month it becomes second nature. At a certain point it's no longer a diet, it's just the way you eat. And I don't deprive myself of anything. I eat basically garbage on the weekends but all in moderation. [NEWLINE] [NEWLINE] Not to mention [NEWLINE] [NEWLINE] [STARTQ] Avoiding all sugar, bread, pasta, and junk food [ENDQ] [NEWLINE] This isn't even necessary. If you're interested look up " If It Fits Your Macros" or IIFYM. Essentially a style of diet where as long as you are consuming the right amount of calories and macro nutrients, you eat whatever you want. [NEWLINE] [NEWLINE] But I appreciate the good discussion.</s>
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1)<mask> wouldn't using your own social skills without the use of alcohol help you be more of a social person in the future? One who was able to break down awkward social walls without the use of alcohol? Alcohol seems<mask> unnecessary<mask> you consider it from my point of view. [NEWLINE] [NEWLINE] 2)<mask><mask> our thoughts and our actions define who we are. Our thoughts define who we are the ourselves, our actions define who we are to others, and both put together define who we are. Alcohol changes who we are to others,<mask> it temporarily changes who we are. [NEWLINE] [NEWLINE] 3)<mask><mask> this is something I've never going to change<mask><mask> on. Being drunk just seems<mask> wrong, I see no possible positive effects from it in the grand scheme of things. [NEWLINE] [NEWLINE] 4)<mask><mask>. Alcohol has actual effects on organs in your body<mask> you don't see your liver actually freaking out from reading a book. And skydiving is very physical<mask> in the sense that it invokes a physical response. Alcohol changes you physically by the use of a substance. Skydiving is your OWN body reacting to a certain stimulus (in this case falling from the sky). By that logic, you could<mask><mask> my argument was<mask> saying walking was wrong<mask> it was physical which clearly is not<mask> I'm trying to convey.</s>
Label encoding: <s>1) But wouldn't using your own social skills without the use of alcohol help you be more of a social person in the future? One who was able to break down awkward social walls without the use of alcohol? Alcohol seems so unnecessary when you consider it from my point of view. [NEWLINE] [NEWLINE] 2) I think our thoughts and our actions define who we are. Our thoughts define who we are the ourselves, our actions define who we are to others, and both put together define who we are. Alcohol changes who we are to others, therefore it temporarily changes who we are. [NEWLINE] [NEWLINE] 3) I think this is something I've never going to change my opinion on. Being drunk just seems so wrong, I see no possible positive effects from it in the grand scheme of things. [NEWLINE] [NEWLINE] 4) I disagree. Alcohol has actual effects on organs in your body while you don't see your liver actually freaking out from reading a book. And skydiving is very physical but in the sense that it invokes a physical response. Alcohol changes you physically by the use of a substance. Skydiving is your OWN body reacting to a certain stimulus (in this case falling from the sky). By that logic, you could argue that my argument was also saying walking was wrong because it was physical which clearly is not what I'm trying to convey.</s>
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] &gt;You don't have to meet women to find a romantic relationship, and you don't have to worry about being labelled a 'creep' or'scary'<mask> you do it wrong. [ENDQ] [NEWLINE] [STARTQ] Er gay men can and have been frequently labelled<mask> creeps to be fair [ENDQ] [NEWLINE] <mask> creeps? I don't really think<mask>. Certainly gay men have faced difficult stereotypes,<mask> not really this one. [NEWLINE] [NEWLINE] There are a lot of feelings and anxieties that go into the male side of forming a relationship, and people are often crushed<mask> they don't work out. You can imagine that being called a creep by the object of your affections - which is a pretty common way of viewing guys with poor social skills - is a huge blow to a lot of men. [NEWLINE] [NEWLINE] [STARTQ] &gt;<mask><mask><mask><mask> you probably do know<mask> it's like to have people look at you sideways<mask> you interact with small children. For a long time it was considered 'normal' to treat gay men<mask> a threat to small children, particularly little boys. [ENDQ] [NEWLINE] [STARTQ] Maybe its myself not realising<mask> I haven't really seen that happen to me. [ENDQ] [NEWLINE] You have the gay rights movement to thank for that. In many languages, they still use the word for pedophile<mask> a slur for gay man. </s>
Label encoding: <s> [STARTQ] &gt;You don't have to meet women to find a romantic relationship, and you don't have to worry about being labelled a 'creep' or'scary' if you do it wrong. [ENDQ] [NEWLINE] [STARTQ] Er gay men can and have been frequently labelled as creeps to be fair [ENDQ] [NEWLINE] As creeps? I don't really think so. Certainly gay men have faced difficult stereotypes, but not really this one. [NEWLINE] [NEWLINE] There are a lot of feelings and anxieties that go into the male side of forming a relationship, and people are often crushed when they don't work out. You can imagine that being called a creep by the object of your affections - which is a pretty common way of viewing guys with poor social skills - is a huge blow to a lot of men. [NEWLINE] [NEWLINE] [STARTQ] &gt; On the other hand you probably do know what it's like to have people look at you sideways when you interact with small children. For a long time it was considered 'normal' to treat gay men as a threat to small children, particularly little boys. [ENDQ] [NEWLINE] [STARTQ] Maybe its myself not realising but I haven't really seen that happen to me. [ENDQ] [NEWLINE] You have the gay rights movement to thank for that. In many languages, they still use the word for pedophile as a slur for gay man. </s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yup, the AWB had a sunset clause and<mask> it's ten years was up it was determined that it had no positive effect and was not renewed. [NEWLINE] [NEWLINE] Not that surprising<mask> it was mostly aimed at cosmetic features on a class of guns(semi auto rifles) that are only used in maybe 1% of murders in the USA. All rifles combined only made up 2.5% of murders in 2011, and stuff like the AR15 is a smaller subset of that. [NEWLINE] [NEWLINE] There is nothing really setting stuff like the [AR15]( [URL].jpg)(affected by AWB) and the [Mini14]( [URL].JPG)(not affected by AWB) apart other than their looks. Both are semi auto and shoot the same caliber and have the same capabilities. [NEWLINE] [NEWLINE] <mask><mask> it was based on a illogical and ineffective attack on<mask> scary certain guns look, and that the weapons it targeted are nearly statistically insignificant in violent crime in the USA it was always destined to be a useless law, limiting<mask> law abiding people could obtain and use, without any safety benefits.Not sure<mask> pro gun control people still like it. [NEWLINE] [NEWLINE] Aside from a coincidence in timing, there isnt really any reason to believe that the AWB was responsible for the drop, or even could theoretically limit homicides in the USA.</s>
Label encoding: <s>Yup, the AWB had a sunset clause and when it's ten years was up it was determined that it had no positive effect and was not renewed. [NEWLINE] [NEWLINE] Not that surprising as it was mostly aimed at cosmetic features on a class of guns(semi auto rifles) that are only used in maybe 1% of murders in the USA. All rifles combined only made up 2.5% of murders in 2011, and stuff like the AR15 is a smaller subset of that. [NEWLINE] [NEWLINE] There is nothing really setting stuff like the [AR15]( [URL].jpg)(affected by AWB) and the [Mini14]( [URL].JPG)(not affected by AWB) apart other than their looks. Both are semi auto and shoot the same caliber and have the same capabilities. [NEWLINE] [NEWLINE] Given that it was based on a illogical and ineffective attack on how scary certain guns look, and that the weapons it targeted are nearly statistically insignificant in violent crime in the USA it was always destined to be a useless law, limiting what law abiding people could obtain and use, without any safety benefits.Not sure why pro gun control people still like it. [NEWLINE] [NEWLINE] Aside from a coincidence in timing, there isnt really any reason to believe that the AWB was responsible for the drop, or even could theoretically limit homicides in the USA.</s>
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0819, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are things in linguistics called phonemes. Just<mask> a language may appear to have similar sounds, doesn't mean that they have the same phonemes. Close approximations,<mask> staying in the target language' mode, is appropriate. Switching between modes based on the word is ridiculous. For instance, English doesn't roll Rs.<mask><mask>, my wife has been trying and just physically cannot do<mask>. In English we have our phonemes and Spanish has theirs. [NEWLINE] [NEWLINE] <mask> I pronounce "quesadilla," I do not ignore the pronunciation of the ll<mask> y<mask> we have that.<mask>, we do not have the soft d phoneme<mask> it's going to get a hard D sound. To expect English speakers to do s is like expecting Germans to pronounce "th"<mask> it is I. English *<mask> * speaking German using an English loan word. [NEWLINE] [NEWLINE] Conversely, those who speak English fluently and have no pronunciation issues who insist on switching modes to pronounce "Latino"<mask> they would in Spanish, even<mask> they are perfectly capable of sating it the proper way in English...is just rude and disrespectful. Likewise it would be disrespectful for me to speak Spanish and pronounce iPod<mask> "eye pod" even<mask> I know<mask> to pronounce it properly for their understanding.</s>
Label encoding: <s>There are things in linguistics called phonemes. Just because a language may appear to have similar sounds, doesn't mean that they have the same phonemes. Close approximations, while staying in the target language' mode, is appropriate. Switching between modes based on the word is ridiculous. For instance, English doesn't roll Rs. In fact, my wife has been trying and just physically cannot do so. In English we have our phonemes and Spanish has theirs. [NEWLINE] [NEWLINE] When I pronounce "quesadilla," I do not ignore the pronunciation of the ll as y since we have that. However, we do not have the soft d phoneme so it's going to get a hard D sound. To expect English speakers to do s is like expecting Germans to pronounce "th" as it is I. English * while * speaking German using an English loan word. [NEWLINE] [NEWLINE] Conversely, those who speak English fluently and have no pronunciation issues who insist on switching modes to pronounce "Latino" as they would in Spanish, even if they are perfectly capable of sating it the proper way in English...is just rude and disrespectful. Likewise it would be disrespectful for me to speak Spanish and pronounce iPod as "eye pod" even if I know how to pronounce it properly for their understanding.</s>
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1.) Virginia did have a lot of slaves,<mask> Virginia was in the confederacy.  Nice try. [NEWLINE] [NEWLINE] 2.) Connecticut outlawed the slave trade before ratifying the constitution.  By 1820, there were less than 100 slaves in Connecticut.  Delaware has less than 5000 slaves<mask> early<mask> 1810. [NEWLINE] [NEWLINE] 3.) New York had a similar period of gradual emancipation with all slaves freed by 1827.  New Jersey was a worse offender, freeing children of slaves born after 1804,<mask> not freeing older slaves until 1846. [NEWLINE] [NEWLINE] 6.) In 1860, there were over 3.5 million slaves in the South, and less than 0.5 million in the border states. [NEWLINE] [NEWLINE] 7.) The border states of Maryland, Kentucky, and Missouri had the most slaves of the Northern states (first Maryland, then Kentucky),<mask> always had a significant amount less than the Southern states.  Maryland and Missouri outlawed slavery before the war was over.   Consider these graphs for visualization: [Slaves by state in 1790]( [URL].gif), [Slaves by state in 1860]( [URL].gif).  You'll notice Maryland's slave population decreasing. [NEWLINE] [NEWLINE] 8.) By any metric, it is completely disingenuous to say there was "almost<mask> much slavery" in the North.</s>
Label encoding: <s>1.) Virginia did have a lot of slaves, but Virginia was in the confederacy.  Nice try. [NEWLINE] [NEWLINE] 2.) Connecticut outlawed the slave trade before ratifying the constitution.  By 1820, there were less than 100 slaves in Connecticut.  Delaware has less than 5000 slaves as early as 1810. [NEWLINE] [NEWLINE] 3.) New York had a similar period of gradual emancipation with all slaves freed by 1827.  New Jersey was a worse offender, freeing children of slaves born after 1804, but not freeing older slaves until 1846. [NEWLINE] [NEWLINE] 6.) In 1860, there were over 3.5 million slaves in the South, and less than 0.5 million in the border states. [NEWLINE] [NEWLINE] 7.) The border states of Maryland, Kentucky, and Missouri had the most slaves of the Northern states (first Maryland, then Kentucky), but always had a significant amount less than the Southern states.  Maryland and Missouri outlawed slavery before the war was over.   Consider these graphs for visualization: [Slaves by state in 1790]( [URL].gif), [Slaves by state in 1860]( [URL].gif).  You'll notice Maryland's slave population decreasing. [NEWLINE] [NEWLINE] 8.) By any metric, it is completely disingenuous to say there was "almost as much slavery" in the North.</s>
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Insults are a valuable tool in debates and are often misused. Insulting someone during a discussion can be used<mask> a great eye opener and will make the debate more personal and significant to them. [NEWLINE] [NEWLINE] Have you even witnessed someone say something like "Well<mask> you're too stupid to just google it and see that you're clearly wrong I am not going to continue this conversation." It is a LOT more effective than just informing them there is counter evidence to their claim on google,<mask> they can't ignore it now that they have to look it up themselves to "prove you wrong". [NEWLINE] [NEWLINE] There is an aspect of social pressure to change the minds of those that are truly ignorant and get trapped into group think. I.E. bigots that only believe such things<mask> race supremacy<mask> their social circle agrees on it,<mask> lack of evidence or logic. [NEWLINE] [NEWLINE] Point out<mask> they are in the minority and<mask> many cool people are anti-racism and<mask> bad many racists are. Emphasize<mask> only a truly stupid person could ever be a racist and<mask> everyone laughs at and pities them. And maybe just maybe they'll be convinced to leave their group and join yours (not for the right reasons,<mask> some people are just content with being ignorant and it's not worth the effort) </s>
Label encoding: <s>Insults are a valuable tool in debates and are often misused. Insulting someone during a discussion can be used as a great eye opener and will make the debate more personal and significant to them. [NEWLINE] [NEWLINE] Have you even witnessed someone say something like "Well if you're too stupid to just google it and see that you're clearly wrong I am not going to continue this conversation." It is a LOT more effective than just informing them there is counter evidence to their claim on google, because they can't ignore it now that they have to look it up themselves to "prove you wrong". [NEWLINE] [NEWLINE] There is an aspect of social pressure to change the minds of those that are truly ignorant and get trapped into group think. I.E. bigots that only believe such things as race supremacy because their social circle agrees on it, despite lack of evidence or logic. [NEWLINE] [NEWLINE] Point out how they are in the minority and how many cool people are anti-racism and how bad many racists are. Emphasize how only a truly stupid person could ever be a racist and how everyone laughs at and pities them. And maybe just maybe they'll be convinced to leave their group and join yours (not for the right reasons, but some people are just content with being ignorant and it's not worth the effort) </s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Are all animal lives of equal value?<mask> the answer is yes, then all animals should be treated equally. Zoos offer preferential treatment to certain animals,<mask> they are not offering equal treatment to all animals, and are unethical. [ENDQ] [NEWLINE] I say yes,<mask><mask><mask> all zoos didn't hold animals who cant be properly accommodated by zoos - see an earlier post I responded to - and offered preferential treatment to all animals, not just the cool/cute ones? [NEWLINE] [NEWLINE] [STARTQ] *<mask> zoos = humans manipulating nature. We do this all the time,<mask> generally for a practical purpose: e.g. living space, food, clothing, etc... [ENDQ] * Considering that zoos are open to the public for general pleasure, and are only found in cities, they serve a purpose<mask> entertainment. Is using an animal for entertainment ethical? [NEWLINE] [NEWLINE] These two lines of reasoning seem to conflict. Your second point seems to imply that you shouldn't use animals period (similar to Kant's imperative to treat people<mask> ends not means),<mask> you state that animal use is justified<mask> for a "practical" purpose.<mask> defines a practical purpose? [NEWLINE] [NEWLINE] Similarly,<mask><mask> entertainment does ultimately achieve a higher goal, such<mask><mask> citizens having visited the zoo, are more willing to support environmental causes?</s>
Label encoding: <s> [STARTQ] Are all animal lives of equal value? If the answer is yes, then all animals should be treated equally. Zoos offer preferential treatment to certain animals, therefore they are not offering equal treatment to all animals, and are unethical. [ENDQ] [NEWLINE] I say yes, but what if all zoos didn't hold animals who cant be properly accommodated by zoos - see an earlier post I responded to - and offered preferential treatment to all animals, not just the cool/cute ones? [NEWLINE] [NEWLINE] [STARTQ] * So zoos = humans manipulating nature. We do this all the time, but generally for a practical purpose: e.g. living space, food, clothing, etc... [ENDQ] * Considering that zoos are open to the public for general pleasure, and are only found in cities, they serve a purpose as entertainment. Is using an animal for entertainment ethical? [NEWLINE] [NEWLINE] These two lines of reasoning seem to conflict. Your second point seems to imply that you shouldn't use animals period (similar to Kant's imperative to treat people as ends not means), but you state that animal use is justified when for a "practical" purpose. What defines a practical purpose? [NEWLINE] [NEWLINE] Similarly, what if entertainment does ultimately achieve a higher goal, such as when citizens having visited the zoo, are more willing to support environmental causes?</s>
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not trying to attack you here,<mask> do you not see<mask> you've written in the OP<mask> extremely sexist? You seem to be operating from a place of,'men don't feel things<mask> deeply<mask> women do' and 'women never want sex/men always want sex' and 'all women are defenseless and need extra protection' and 'weak men deserve punishment for being weak' and stuff like that [NEWLINE] [NEWLINE] [STARTQ] Now<mask> for gay men, that's different and I can sympathize with them more, especially<mask> they were the ones who were penetrated.<mask><mask> I feel that way with gays, I still don't think of rape of gay men<mask> bad<mask> rape of women. [ENDQ] [NEWLINE] And this... do you not realize straight men get raped by men? Or that a woman could penetratively rape a man with some device? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> of a strong (and sometimes old) man who clearly has power of a much weaker woman forcing her to do things that she finds derogatory [ENDQ] [NEWLINE] Who do you think is more injured, a scrawny 18 year old male getting anally raped by a man, or a kickboxing champion 30 year female getting anally raped by a man?<mask> it's still the woman,<mask> you said int he quote there can't be your issue.</s><pad>
Label encoding: <s>Not trying to attack you here, but do you not see what you've written in the OP as extremely sexist? You seem to be operating from a place of,'men don't feel things as deeply as women do' and 'women never want sex/men always want sex' and 'all women are defenseless and need extra protection' and 'weak men deserve punishment for being weak' and stuff like that [NEWLINE] [NEWLINE] [STARTQ] Now as for gay men, that's different and I can sympathize with them more, especially if they were the ones who were penetrated. Even though I feel that way with gays, I still don't think of rape of gay men as bad as rape of women. [ENDQ] [NEWLINE] And this... do you not realize straight men get raped by men? Or that a woman could penetratively rape a man with some device? [NEWLINE] [NEWLINE] [STARTQ] I think of a strong (and sometimes old) man who clearly has power of a much weaker woman forcing her to do things that she finds derogatory [ENDQ] [NEWLINE] Who do you think is more injured, a scrawny 18 year old male getting anally raped by a man, or a kickboxing champion 30 year female getting anally raped by a man? If it's still the woman, what you said int he quote there can't be your issue.</s><pad>
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What you're saying is fundamentally true.  The issue obviously comes in degrees of difficulty, not in absolute inability. [NEWLINE] [NEWLINE] <mask> lets posit the worst-off fat person possible,<mask> first we'll talk about their body<mask> they were skinny.  Lets say this person is a woman, weighs 120 lbs and has a [Basal metabolic rate]( [URL] #Causes_of_individual_differences_in_BMR) of 1000 calories.  This person has tried hard<mask> finds it very difficult to build any muscle through exercise.  They are<mask> generally weak and unable to sustain exercise for long. And<mask> their metabolic rate never exceeds 1500 calories in a day, even after they are exhausted from exercise. [NEWLINE] [NEWLINE] This person is<mask> cursed with a strong appetite.  They are hungry throughout the day unless they eat more food than their body currently requires.  Lets say they are generally hungry unless they eat about 2500 calories a day. [NEWLINE] [NEWLINE] <mask><mask> the formula from the wiki page,<mask> we assume their metabolic rate goes up linearly with weight, then to get to a metabolic rate of 2500 calories they'd have to be at a weight of 385 lbs. [NEWLINE] [NEWLINE] Statistically there are going to be some people like this.  Can you really say that for this person its just eating too much?</s>
Label encoding: <s>What you're saying is fundamentally true.  The issue obviously comes in degrees of difficulty, not in absolute inability. [NEWLINE] [NEWLINE] So lets posit the worst-off fat person possible, but first we'll talk about their body while they were skinny.  Lets say this person is a woman, weighs 120 lbs and has a [Basal metabolic rate]( [URL] #Causes_of_individual_differences_in_BMR) of 1000 calories.  This person has tried hard but finds it very difficult to build any muscle through exercise.  They are also generally weak and unable to sustain exercise for long. And so their metabolic rate never exceeds 1500 calories in a day, even after they are exhausted from exercise. [NEWLINE] [NEWLINE] This person is also cursed with a strong appetite.  They are hungry throughout the day unless they eat more food than their body currently requires.  Lets say they are generally hungry unless they eat about 2500 calories a day. [NEWLINE] [NEWLINE] According to the formula from the wiki page, if we assume their metabolic rate goes up linearly with weight, then to get to a metabolic rate of 2500 calories they'd have to be at a weight of 385 lbs. [NEWLINE] [NEWLINE] Statistically there are going to be some people like this.  Can you really say that for this person its just eating too much?</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No one's trying to get anything banned. That's not nessecary anyway. Organic food production has a lower output per production unit, and in turn commands a higher price, actually a price that is higher than<mask> would be justified by the decreased output,<mask> consumers in rich countries are willing to pay a premium for a sticker on their produce that says "Organic" or "Bio". [NEWLINE] [NEWLINE] And here's<mask> the trouble starts. Due to the added premium, a decrease in food output suddenly means an increase in revenue and gain<mask> of the "Organic" label premium I mentioned. Which creates a competition with regular food production, raising the price for non-organic food<mask> well. [NEWLINE] [NEWLINE] At the same time, the available areas for food production are reduced by those taken up by organic farming which produces food locals often can't afford,<mask> takes up space previously used for regular crops and regular farming, again driving the price for non-roganic food up due to increased competition for land area. [NEWLINE] [NEWLINE] Both effects together cause an increase in prices of regular, non-organic food that is wholly the result of an increase in organic food production. And<mask> rich countries in europe and north america can easily stomach the slight increase in prices for regular food, poorer countries can't that easily.</s>
Label encoding: <s>No one's trying to get anything banned. That's not nessecary anyway. Organic food production has a lower output per production unit, and in turn commands a higher price, actually a price that is higher than what would be justified by the decreased output, because consumers in rich countries are willing to pay a premium for a sticker on their produce that says "Organic" or "Bio". [NEWLINE] [NEWLINE] And here's where the trouble starts. Due to the added premium, a decrease in food output suddenly means an increase in revenue and gain because of the "Organic" label premium I mentioned. Which creates a competition with regular food production, raising the price for non-organic food as well. [NEWLINE] [NEWLINE] At the same time, the available areas for food production are reduced by those taken up by organic farming which produces food locals often can't afford, but takes up space previously used for regular crops and regular farming, again driving the price for non-roganic food up due to increased competition for land area. [NEWLINE] [NEWLINE] Both effects together cause an increase in prices of regular, non-organic food that is wholly the result of an increase in organic food production. And while rich countries in europe and north america can easily stomach the slight increase in prices for regular food, poorer countries can't that easily.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0909, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm an indie dev myself, and<mask> I don't have time to go into detail, keep in mind these indie devs are often a team of 1 - 4 people trying desperately to put something, anything out, to get a step closer to fufilling their dream. <mask> they make a bunch of crappy games that flood the market, that's no big deal, major companies do this all the time (even the retro theme - see Megaman 9/10 and Ducktales). <mask> the game sucks, we ignore it and move on. <mask> in the meantime, these indie devs are gaining experience, and growing into productive members of the game industry. <mask>, keep in mind that major game designers usually require at least one shipped game on your resume to even consider you. [NEWLINE] [NEWLINE] <mask> maybe these guys aren't<mask> good at making creative games,<mask> maybe they have a fucking phenomenal artist.  He's still got that crappy dungeon crawler on his resume, and<mask> he markets himself right, his talents can be put to great use<mask> he's picked up by a more creative, talented team.  This helps the game industry greatly,<mask> this amazing artist would never have been found<mask> he hadn't worked on that crappy dungeon crawler and put his talents out there on the market.</s>
Label encoding: <s>I'm an indie dev myself, and while I don't have time to go into detail, keep in mind these indie devs are often a team of 1 - 4 people trying desperately to put something, anything out, to get a step closer to fufilling their dream.  If they make a bunch of crappy games that flood the market, that's no big deal, major companies do this all the time (even the retro theme - see Megaman 9/10 and Ducktales).  If the game sucks, we ignore it and move on.  But in the meantime, these indie devs are gaining experience, and growing into productive members of the game industry.  Also, keep in mind that major game designers usually require at least one shipped game on your resume to even consider you. [NEWLINE] [NEWLINE] So maybe these guys aren't so good at making creative games, but maybe they have a fucking phenomenal artist.  He's still got that crappy dungeon crawler on his resume, and if he markets himself right, his talents can be put to great use when he's picked up by a more creative, talented team.  This helps the game industry greatly, because this amazing artist would never have been found if he hadn't worked on that crappy dungeon crawler and put his talents out there on the market.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What say you to the counter-argument that<mask><mask><mask> two people are consenting adults, it's nobody's business<mask> they do? [NEWLINE] [NEWLINE] They may overlap often,<mask> incest and child molestation and/or rape are distinct. The same slippery slope-type argument can be and is or was often made for the proscription of sexual acts that I'm fairly sure you do not object to. [NEWLINE] [NEWLINE] The potential for birth defects argument might seem nice,<mask> that doesn't hold up either.<mask> two people were unrelated,<mask> know their children had a 50% chance of major disabilities and a 100% of being a carrier for them, should it be illegal for them to have sex? Should they be imprisoned<mask> they have children? In either case, incest and having children through incest are different things. [NEWLINE] [NEWLINE] <mask> for a losing a family member, I'm not sure<mask> or<mask> that is anyone's business<mask> theirs. People who have romantic relationships with friends often lose that friend,<mask> I've never heard that conduct described<mask> morally wrong. Is there any other context<mask> interpersonal relationships are society's business? This seems like a post-hoc rationalization for revulsion. A lot of things can screw up families, that doesn't necessarily make them "wrong" or mean they should be illegal.</s>
Label encoding: <s>What say you to the counter-argument that as long as two people are consenting adults, it's nobody's business what they do? [NEWLINE] [NEWLINE] They may overlap often, but incest and child molestation and/or rape are distinct. The same slippery slope-type argument can be and is or was often made for the proscription of sexual acts that I'm fairly sure you do not object to. [NEWLINE] [NEWLINE] The potential for birth defects argument might seem nice, but that doesn't hold up either. If two people were unrelated, but know their children had a 50% chance of major disabilities and a 100% of being a carrier for them, should it be illegal for them to have sex? Should they be imprisoned if they have children? In either case, incest and having children through incest are different things. [NEWLINE] [NEWLINE] As for a losing a family member, I'm not sure how or why that is anyone's business but theirs. People who have romantic relationships with friends often lose that friend, but I've never heard that conduct described as morally wrong. Is there any other context where interpersonal relationships are society's business? This seems like a post-hoc rationalization for revulsion. A lot of things can screw up families, that doesn't necessarily make them "wrong" or mean they should be illegal.</s>
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Lets look at another example. [NEWLINE] [NEWLINE] In Harry Potter, everyone (readers that is) hated Voldemort. He was a bad dude. Then in book 5 we are introduced to Professor Umbridge. People *hated* her. It was more than hate. There was a sense of loathing, a visceral emotional reaction that Voldemort just didn't evoke. [NEWLINE] [NEWLINE] Voldemort was a cartoon villain. He was bad for the sake of being bad. I mean, sure he had his reasons,<mask> mostly you always knew<mask> he was going to do: be evil. It's just who he was.<mask> Umbrige was something different, something *familiar*. Everyone who's ever gone to school and had a bad teacher or principle or other authority figure had a much stronger emotional reaction to her. [NEWLINE] [NEWLINE] In one case we just have someone who's cartoonishly evil at times.<mask><mask><mask><mask>, we have a flavor of evil that's much more familiar and intimate and something we can relate to our lives and understand, and that's<mask> makes Umbridge such a good character and a good villian. [NEWLINE] [NEWLINE] Now I'm not trying to say Scar is like Umbridge - all I'm saying here is that the evil you understand can be much worse than the one you don't. </s>
Label encoding: <s>Lets look at another example. [NEWLINE] [NEWLINE] In Harry Potter, everyone (readers that is) hated Voldemort. He was a bad dude. Then in book 5 we are introduced to Professor Umbridge. People *hated* her. It was more than hate. There was a sense of loathing, a visceral emotional reaction that Voldemort just didn't evoke. [NEWLINE] [NEWLINE] Voldemort was a cartoon villain. He was bad for the sake of being bad. I mean, sure he had his reasons, but mostly you always knew what he was going to do: be evil. It's just who he was. But Umbrige was something different, something *familiar*. Everyone who's ever gone to school and had a bad teacher or principle or other authority figure had a much stronger emotional reaction to her. [NEWLINE] [NEWLINE] In one case we just have someone who's cartoonishly evil at times. On the other hand, we have a flavor of evil that's much more familiar and intimate and something we can relate to our lives and understand, and that's what makes Umbridge such a good character and a good villian. [NEWLINE] [NEWLINE] Now I'm not trying to say Scar is like Umbridge - all I'm saying here is that the evil you understand can be much worse than the one you don't. </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There is evidence to suggest the non existence in a god,<mask> nothing says it straight out.<mask> you aren't very religious, then you can be persuaded ito not believe in god.<mask> typically the people who want to debate are very religious, close minded people. [ENDQ] [NEWLINE] Don't you think you're paring down your claim more and more here? It looks like you started with'religious people who believe in God shouldn't be debated<mask> God is unfalsifiable and you can't argue God's existence' to 'actually you can argue against and for God's existence, there's room for debate here,<mask> some people are really resistant to evidence and won't change their minds no matter<mask> '. [NEWLINE] [NEWLINE] At which point your problem isn't religions, or God, or religious people who believe in God,<mask> 'unreasonable people who won't change their minds'. And yeah, there's no point in arguing with them -<mask> religion and God has nothing to do with it. It's like trying to argue with a hypothetical atheist who argues that no one could ever convince him of God's existence and any given argument or evidence or sign he receives he'll explain away. Yeah, he's unreasonable,<mask> argue God with him -<mask> 'belief in God' has nothing to do with it.</s>
Label encoding: <s> [STARTQ] There is evidence to suggest the non existence in a god, but nothing says it straight out. If you aren't very religious, then you can be persuaded ito not believe in god. But typically the people who want to debate are very religious, close minded people. [ENDQ] [NEWLINE] Don't you think you're paring down your claim more and more here? It looks like you started with'religious people who believe in God shouldn't be debated because God is unfalsifiable and you can't argue God's existence' to 'actually you can argue against and for God's existence, there's room for debate here, but some people are really resistant to evidence and won't change their minds no matter what '. [NEWLINE] [NEWLINE] At which point your problem isn't religions, or God, or religious people who believe in God, but 'unreasonable people who won't change their minds'. And yeah, there's no point in arguing with them - but religion and God has nothing to do with it. It's like trying to argue with a hypothetical atheist who argues that no one could ever convince him of God's existence and any given argument or evidence or sign he receives he'll explain away. Yeah, he's unreasonable, why argue God with him - but 'belief in God' has nothing to do with it.</s>
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you're not qualified to do other work [ENDQ] [NEWLINE] push a broom [NEWLINE] [NEWLINE] [STARTQ] you need to be available in your office<mask> someone needs you [ENDQ] [NEWLINE] organize your email, get a head start on the next job, find a way to improve your efficiency, etc [NEWLINE] [NEWLINE] [STARTQ] (among other tasks) [ENDQ] [NEWLINE] <mask> you have other tasks, then you aren't done your job,<mask> you shouldn't be on reddit anyway [NEWLINE] [NEWLINE] [STARTQ] you are bothering people who are by messing up their organization or making mistakes [ENDQ] [NEWLINE] I didn't say go fuck with other peoples shit, I said go help other people.  everything you described is the opposite of help [NEWLINE] [NEWLINE] [STARTQ] or you aren't covered by insurance<mask> you do other work than that you've been hired for [ENDQ] [NEWLINE] doesn't sound like the type of job<mask> you will be browsing reddit,<mask><mask> it is please see response #1 - push a broom.  Or maybe ask your manager<mask> there is something else that needs to be done (depending, of course,<mask> we are talking about 5 minutes downtime or 5 hours downtime) [NEWLINE] [NEWLINE] There is almost always something that can be done which is more productive for the business. <mask> you can't find it, instead of reading reddit maybe you should read a book on thinking outside of the box</s>
Label encoding: <s> [STARTQ] you're not qualified to do other work [ENDQ] [NEWLINE] push a broom [NEWLINE] [NEWLINE] [STARTQ] you need to be available in your office when someone needs you [ENDQ] [NEWLINE] organize your email, get a head start on the next job, find a way to improve your efficiency, etc [NEWLINE] [NEWLINE] [STARTQ] (among other tasks) [ENDQ] [NEWLINE] if you have other tasks, then you aren't done your job, so you shouldn't be on reddit anyway [NEWLINE] [NEWLINE] [STARTQ] you are bothering people who are by messing up their organization or making mistakes [ENDQ] [NEWLINE] I didn't say go fuck with other peoples shit, I said go help other people.  everything you described is the opposite of help [NEWLINE] [NEWLINE] [STARTQ] or you aren't covered by insurance if you do other work than that you've been hired for [ENDQ] [NEWLINE] doesn't sound like the type of job where you will be browsing reddit, but if it is please see response #1 - push a broom.  Or maybe ask your manager if there is something else that needs to be done (depending, of course, if we are talking about 5 minutes downtime or 5 hours downtime) [NEWLINE] [NEWLINE] There is almost always something that can be done which is more productive for the business.  If you can't find it, instead of reading reddit maybe you should read a book on thinking outside of the box</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That research is not inherent,<mask> it analyzes nurture rather than nature. It supports the thesis that there are culturally ingrained concepts of beauty,<mask> it doesn't prove that these concepts are necessarily universal or intrinsically biological.<mask>, might it be true that an overweight person with a very attractive face would have an easier time finding a mate than an extremely fit person with the "proper" shoulder-hip ratio or waist-hip ratio who has a particularly unattractive face? I'm not sure.<mask><mask>,<mask><mask> there's no way to say empirically or objectively which would be more attractive,<mask> different people might value different things<mask> push comes to shove. [NEWLINE] [NEWLINE] Furthermore,<mask><mask> you might want to read up on historical concepts of beauty. In medieval and renaissance Europe, for example, being overweight by modern standards was considered a mark of beauty and an attractive feature in a mate for both men and women,<mask> it demonstrated that person's social status<mask> they could afford an excess of food and the lack of need to exert oneself physically. Being thin and/or muscular, by contrast, was seen<mask> less attractive,<mask> it was the mark of a laborer who could not afford an excess of food. This seems to demonstrate that concepts of beauty adapt to the unique needs and values of particular cultures.</s>
Label encoding: <s>That research is not inherent, as it analyzes nurture rather than nature. It supports the thesis that there are culturally ingrained concepts of beauty, but it doesn't prove that these concepts are necessarily universal or intrinsically biological. Moreover, might it be true that an overweight person with a very attractive face would have an easier time finding a mate than an extremely fit person with the "proper" shoulder-hip ratio or waist-hip ratio who has a particularly unattractive face? I'm not sure. In fact, I think there's no way to say empirically or objectively which would be more attractive, as different people might value different things when push comes to shove. [NEWLINE] [NEWLINE] Furthermore, I think you might want to read up on historical concepts of beauty. In medieval and renaissance Europe, for example, being overweight by modern standards was considered a mark of beauty and an attractive feature in a mate for both men and women, as it demonstrated that person's social status since they could afford an excess of food and the lack of need to exert oneself physically. Being thin and/or muscular, by contrast, was seen as less attractive, because it was the mark of a laborer who could not afford an excess of food. This seems to demonstrate that concepts of beauty adapt to the unique needs and values of particular cultures.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, I'm American too<mask> I get<mask> you're coming from. [NEWLINE] [NEWLINE] I just don't really know<mask> to compare the Cyprus story to in American sports. Your tennis analogy is a start. They are such an unknown team in European soccer<mask> they really came out of nowhere. It was pretty incredible that they made it to the group stage.<mask> then to make it through that (going up against 3 superior teams), and then the first elimination round was absolutely unheard of. Then they came up against the superpower that is Real Madrid and it was over.<mask> the fact that a team of their caliber even had the chance to compete against Real Madrid was pretty fun to watch. [NEWLINE] [NEWLINE] <mask><mask> you could really deem any underdog success story<mask> a fluke.<mask> do you draw the line? Needless to say, soccer does have it's underdog stories they just come in a different way than American sports. I really love the idea of the FA cup system. Every country has their version of it for soccer...It's like having a ncaa tournament<mask> with pro teams taking place alongside the regular season. Every professional and even semi professional teams enter it. It's an underdog story waiting to happen. [NEWLINE] [NEWLINE] edit: Another great underdog story I forgot about was Greece winning the European Championship in 2004.</s>
Label encoding: <s>No, I'm American too so I get where you're coming from. [NEWLINE] [NEWLINE] I just don't really know what to compare the Cyprus story to in American sports. Your tennis analogy is a start. They are such an unknown team in European soccer so they really came out of nowhere. It was pretty incredible that they made it to the group stage. But then to make it through that (going up against 3 superior teams), and then the first elimination round was absolutely unheard of. Then they came up against the superpower that is Real Madrid and it was over. But the fact that a team of their caliber even had the chance to compete against Real Madrid was pretty fun to watch. [NEWLINE] [NEWLINE] I think you could really deem any underdog success story as a fluke. Where do you draw the line? Needless to say, soccer does have it's underdog stories they just come in a different way than American sports. I really love the idea of the FA cup system. Every country has their version of it for soccer...It's like having a ncaa tournament but with pro teams taking place alongside the regular season. Every professional and even semi professional teams enter it. It's an underdog story waiting to happen. [NEWLINE] [NEWLINE] edit: Another great underdog story I forgot about was Greece winning the European Championship in 2004.</s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The major thing you're missing, and really the driver of a huge amount of economic change over the past 40 or<mask> years, is the opening of India and China into the world market.  Especially China. [NEWLINE] [NEWLINE] China was closed off from the world economy and completely stagnant prior to the 1970s.  [Massive famines]( [URL] ), [social unrest]( [URL] ), etc. <mask> China has entered the world economy, it has added a *huge* supply of labor.  That competition bids down the price that laborers can get in the western world including the USA. [NEWLINE] [NEWLINE] It has<mask> resulted in the largest reduction in human poverty ever. <mask> its probably a good thing overall. [NEWLINE] [NEWLINE] <mask> workers in China get richer, they'll start to be less competitive with US workers, and it is likely US workers will be able to get higher wages. [NEWLINE] [NEWLINE] <mask> to your specific student loan situation.  That's the result of a poor decision on your part combined with a poorly structured government program. <mask> time goes on, it is likely social norms about student debt will change, and this sort of thing become a bit less common, and<mask> hopefully loan programs will be reformed.  Certainly you wouldn't let your kids get in the same kind of debt you did, for instance.</s>
Label encoding: <s>The major thing you're missing, and really the driver of a huge amount of economic change over the past 40 or so years, is the opening of India and China into the world market.  Especially China. [NEWLINE] [NEWLINE] China was closed off from the world economy and completely stagnant prior to the 1970s.  [Massive famines]( [URL] ), [social unrest]( [URL] ), etc.  As China has entered the world economy, it has added a *huge* supply of labor.  That competition bids down the price that laborers can get in the western world including the USA. [NEWLINE] [NEWLINE] It has also resulted in the largest reduction in human poverty ever.  So its probably a good thing overall. [NEWLINE] [NEWLINE] As workers in China get richer, they'll start to be less competitive with US workers, and it is likely US workers will be able to get higher wages. [NEWLINE] [NEWLINE] As to your specific student loan situation.  That's the result of a poor decision on your part combined with a poorly structured government program.  As time goes on, it is likely social norms about student debt will change, and this sort of thing become a bit less common, and also hopefully loan programs will be reformed.  Certainly you wouldn't let your kids get in the same kind of debt you did, for instance.</s>
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your first point: My fault. The "we" was SRS. [NEWLINE] [NEWLINE] Your second point: My comment gave no insight on my position on this<mask> no idea<mask> you mentioned it.<mask> to clarify I am in agreement with you. [NEWLINE] [NEWLINE] Your third point: SRS seems to think reddit dislikes them<mask> they point out sexism/racism. It empowers them and it should.<mask> you are doing something good and people hate you for it, it makes your committment that much stronger.<mask> that is not<mask> reddit dislikes them even<mask> SRS wants to think<mask>. Reddit dislikes them<mask> of 1) people go to the subreddit looking to debate/give a different view point on a particular point and are faced with extreme closemindness and a brick wall. They are upset that the subreddit refuses to have a fair discussion. They don't realize its a circle-jerk subreddit and get mad about it. That is like me going to your house uninvited and getting mad<mask> of<mask> you make for dinner. 2) They point out sexism by using sexism(or reverse sexism<mask> there is such a thing). People think its hypocritical. I call it satire. Point is SRS or any subreddit can talk about<mask> they want and<mask> they want.</s>
Label encoding: <s>Your first point: My fault. The "we" was SRS. [NEWLINE] [NEWLINE] Your second point: My comment gave no insight on my position on this so no idea why you mentioned it. But to clarify I am in agreement with you. [NEWLINE] [NEWLINE] Your third point: SRS seems to think reddit dislikes them because they point out sexism/racism. It empowers them and it should. If you are doing something good and people hate you for it, it makes your committment that much stronger. But that is not why reddit dislikes them even if SRS wants to think so. Reddit dislikes them because of 1) people go to the subreddit looking to debate/give a different view point on a particular point and are faced with extreme closemindness and a brick wall. They are upset that the subreddit refuses to have a fair discussion. They don't realize its a circle-jerk subreddit and get mad about it. That is like me going to your house uninvited and getting mad because of what you make for dinner. 2) They point out sexism by using sexism(or reverse sexism if there is such a thing). People think its hypocritical. I call it satire. Point is SRS or any subreddit can talk about what they want and how they want.</s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> overall that<mask> it were possible, that maybe this could be an option, considering that I can see<mask> this could only encourage kids to learn from their mistakes on tests rather than accepting a low grade and moving on. [ENDQ] [NEWLINE] Thank you for that. [NEWLINE] [NEWLINE] That being said, it's funny you mention a calc teacher. I took calc in  high school, and this is the class that made me think this would be a good idea. I failed my first of 3 tests. Without retakes, I would have been doomed to a pretty low grade. It was partly due to laziness. I'll admit that.<mask>,<mask> my teacher allowed multiple retakes for each test, I was motivated to accept nothing less than a complete understanding of the material.<mask><mask> I'm not a math major, I ended up taking higher level calc courses<mask> I gained confidence with the material. [NEWLINE] [NEWLINE] I don't think it's a forgone conclusion to use the bell curve in K-12.<mask>, I don't think it's burdensome on the students.<mask> is burdensome is to have to catch up anyways,<mask> lessons in a class often build on one another, knowing that you can't regain a portion of your grade no matter<mask> hard you try. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I think overall that if it were possible, that maybe this could be an option, considering that I can see how this could only encourage kids to learn from their mistakes on tests rather than accepting a low grade and moving on. [ENDQ] [NEWLINE] Thank you for that. [NEWLINE] [NEWLINE] That being said, it's funny you mention a calc teacher. I took calc in  high school, and this is the class that made me think this would be a good idea. I failed my first of 3 tests. Without retakes, I would have been doomed to a pretty low grade. It was partly due to laziness. I'll admit that. However, because my teacher allowed multiple retakes for each test, I was motivated to accept nothing less than a complete understanding of the material. Even though I'm not a math major, I ended up taking higher level calc courses because I gained confidence with the material. [NEWLINE] [NEWLINE] I don't think it's a forgone conclusion to use the bell curve in K-12. Also, I don't think it's burdensome on the students. What is burdensome is to have to catch up anyways, because lessons in a class often build on one another, knowing that you can't regain a portion of your grade no matter how hard you try. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, sorry then. I guess it's sort of a modern spiritual animism :-). [NEWLINE] [NEWLINE] <mask> the science tells us is that plants do have complex processes and signals and immune systems. Like our immune systems and some of our growth hormone signalling systems. Or<mask> we get darker in response to sun. They are beautiful and complex,<mask> there is no evidence of consciousness. I'd say the onions "screaming" is more analogous to the theory that our sweat smells bad to dispell predators or<mask> we get fevers to help fight bacteria. Complex, interesting, not ethically all that relevant. Our brain dead bodies would carry out most of the complex processes that our bodies do today,<mask><mask> my brain dies, my family has my permission to kill the rest of me. [NEWLINE] [NEWLINE] <mask> you do find those complex systems morally important, that's fine with me. Sweet even. I just would request that you don't let it interfere with your ability to recognize wrongs that are committed.<mask> my family has my brain dead body killed, that's ok with me.<mask> you sneak up behind me and kill me today - that's very deeply wrong. Even<mask> I never have a chance to notice.<mask> I value my life! And I'm alive to value my body!</s>
Label encoding: <s>Ok, sorry then. I guess it's sort of a modern spiritual animism :-). [NEWLINE] [NEWLINE] What the science tells us is that plants do have complex processes and signals and immune systems. Like our immune systems and some of our growth hormone signalling systems. Or how we get darker in response to sun. They are beautiful and complex, but there is no evidence of consciousness. I'd say the onions "screaming" is more analogous to the theory that our sweat smells bad to dispell predators or how we get fevers to help fight bacteria. Complex, interesting, not ethically all that relevant. Our brain dead bodies would carry out most of the complex processes that our bodies do today, but if my brain dies, my family has my permission to kill the rest of me. [NEWLINE] [NEWLINE] If you do find those complex systems morally important, that's fine with me. Sweet even. I just would request that you don't let it interfere with your ability to recognize wrongs that are committed. If my family has my brain dead body killed, that's ok with me. If you sneak up behind me and kill me today - that's very deeply wrong. Even if I never have a chance to notice. Because I value my life! And I'm alive to value my body!</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> hardware has plateaued and we don't need to upgrade that<mask> often<mask> macs are worse<mask> you can no longer upgrade hardware? [ENDQ] [NEWLINE] No<mask> comparing a laptop purchased after the hardware has plateaued to hardware purchased before that plateau is not a valid comparison. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] I'd<mask><mask><mask> apple is releasing free OS upgrades yearly on the old hardware it will be a better long term investment than a windows PC that may not need a hardware upgrade<mask> require a pricey software upgrade. [ENDQ] [NEWLINE] [NEWLINE] OSX updates are free,<mask> aren't pushed out to all models. <mask> you are running an old piece of hardware you won't get the OSX update, and<mask> your computer is no longer supported and you run the risk of malware and exploits. [NEWLINE] [NEWLINE] Windows is paid (<mask> this looks like it is going to be changing) upgrades,<mask> MS runs a much longer support cycle for their older OSs. [NEWLINE] [NEWLINE] You can run a Windows PC for 10 years<mask> getting software maintenance, Apple runs a shorter 5 year cycle, and<mask> you are out you must purchase a new PC or run the risk of being vulnerable to exploits. <mask> your software ever does run out of the support cycle you can upgrade your software without having to purchase new hardware. [NEWLINE] </s>
Label encoding: <s> [STARTQ] So hardware has plateaued and we don't need to upgrade that as often therefore macs are worse because you can no longer upgrade hardware? [ENDQ] [NEWLINE] No but comparing a laptop purchased after the hardware has plateaued to hardware purchased before that plateau is not a valid comparison. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] I'd argue that since apple is releasing free OS upgrades yearly on the old hardware it will be a better long term investment than a windows PC that may not need a hardware upgrade but require a pricey software upgrade. [ENDQ] [NEWLINE] [NEWLINE] OSX updates are free, but aren't pushed out to all models.  If you are running an old piece of hardware you won't get the OSX update, and thus your computer is no longer supported and you run the risk of malware and exploits. [NEWLINE] [NEWLINE] Windows is paid ( though this looks like it is going to be changing) upgrades, but MS runs a much longer support cycle for their older OSs. [NEWLINE] [NEWLINE] You can run a Windows PC for 10 years while getting software maintenance, Apple runs a shorter 5 year cycle, and when you are out you must purchase a new PC or run the risk of being vulnerable to exploits.  If your software ever does run out of the support cycle you can upgrade your software without having to purchase new hardware. [NEWLINE] </s>
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have been struggling with this<mask> one part of my brain recognizes that people are not<mask> fortunate<mask> I am, don't have the resources or the physical/emotional/social support to sustain themselves and<mask>not, and the other part is deeply mistrustful of what these people have done to end up<mask> they are. [ENDQ] [NEWLINE] An old Cherokee chief was teaching his grandson about life... [NEWLINE] [NEWLINE] "A fight is going on inside me," he said to the boy. [NEWLINE] "It is a terrible fight and it is between two wolves. [NEWLINE] [NEWLINE] "One is evil - he is anger, envy, sorrow, regret, greed, arrogance, self-pity, guilt, resentment, inferiority, lies, false pride, superiority, self-doubt, and ego. [NEWLINE] [NEWLINE] "The other is good - he is joy, peace, love, hope, serenity, humility, kindness, benevolence, empathy, generosity, truth, compassion, and faith. [NEWLINE] [NEWLINE] "This same fight is going on inside you - and inside every other person, too." [NEWLINE] [NEWLINE] The grandson thought about it for a minute and then asked his grandfather, [NEWLINE] "Which wolf will win?" [NEWLINE] [NEWLINE] The old chief simply replied, [NEWLINE] "The one you feed." [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I have been struggling with this because one part of my brain recognizes that people are not as fortunate as I am, don't have the resources or the physical/emotional/social support to sustain themselves and whatnot, and the other part is deeply mistrustful of what these people have done to end up where they are. [ENDQ] [NEWLINE] An old Cherokee chief was teaching his grandson about life... [NEWLINE] [NEWLINE] "A fight is going on inside me," he said to the boy. [NEWLINE] "It is a terrible fight and it is between two wolves. [NEWLINE] [NEWLINE] "One is evil - he is anger, envy, sorrow, regret, greed, arrogance, self-pity, guilt, resentment, inferiority, lies, false pride, superiority, self-doubt, and ego. [NEWLINE] [NEWLINE] "The other is good - he is joy, peace, love, hope, serenity, humility, kindness, benevolence, empathy, generosity, truth, compassion, and faith. [NEWLINE] [NEWLINE] "This same fight is going on inside you - and inside every other person, too." [NEWLINE] [NEWLINE] The grandson thought about it for a minute and then asked his grandfather, [NEWLINE] "Which wolf will win?" [NEWLINE] [NEWLINE] The old chief simply replied, [NEWLINE] "The one you feed." [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Alright you definitely got me on the idea that piracy isn't theft. You've changed my view congrats.<mask> I'm still going to argue with you on the fact that digital information is a finite resource. You can make an incredibly large amount of copies,<mask> eventually you will run out of space. There might be trillions of gigabits of digital information,<mask> there is a finite amount. It is still my stand with my view that certain ideas can be owned. Things like songs, mathematical and scientific formulas, names. The creators own the ideas and have the right to chose whether they can be copied or not. [NEWLINE] [NEWLINE] <mask> I don't really seeing<mask> your saying about the labor and value. Labor definitely definitely goes into the value of product<mask> you are measuring it by price, and not by utility. A shirt in produced in America compared to one made in Bangladesh costs more<mask> the labor cost more. In the words of Adam Smith, "The value of any commodity,... to the person who possesses it, and who means not to use or consume it himself,<mask> to exchange it for other commodities, is equal to the quantity of labour which it enables him to purchase or command. Labour,<mask>, is the real measure of the exchangeable value of all commodities"</s><pad>
Label encoding: <s>Alright you definitely got me on the idea that piracy isn't theft. You've changed my view congrats. But I'm still going to argue with you on the fact that digital information is a finite resource. You can make an incredibly large amount of copies, but eventually you will run out of space. There might be trillions of gigabits of digital information, but there is a finite amount. It is still my stand with my view that certain ideas can be owned. Things like songs, mathematical and scientific formulas, names. The creators own the ideas and have the right to chose whether they can be copied or not. [NEWLINE] [NEWLINE] Also I don't really seeing what your saying about the labor and value. Labor definitely definitely goes into the value of product if you are measuring it by price, and not by utility. A shirt in produced in America compared to one made in Bangladesh costs more because the labor cost more. In the words of Adam Smith, "The value of any commodity,... to the person who possesses it, and who means not to use or consume it himself, but to exchange it for other commodities, is equal to the quantity of labour which it enables him to purchase or command. Labour, therefore, is the real measure of the exchangeable value of all commodities"</s><pad>
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is very similar-you are around people in every situation. You have a camaraderie inherent within your group. Even<mask> you hate your workmate, you will save their life<mask> risking your own. You'll push people past their comfort zone, to the point<mask> they are crying in pain<mask> laughing at the shear insanity of<mask> is happening. [NEWLINE] [NEWLINE] UTI's? Guys get those too. Along with STI's. And extremely high levels of mental illness. Arthritis too. Females,<mask><mask><mask> I know, are actually more healthy due to better concentration on their health by our societies. Men don't tend to talk about their health<mask> much, to their detriment. [NEWLINE] [NEWLINE] Female servicemen tend to take contraceptive pills in the field or employ other methods of preventing pregnancy, much like males. From my experience they are still very useful right up until they give birth, especially<mask> officers etc.<mask><mask><mask> I know, females<mask> have higher pain thresholds, which would be incredibly useful during combat. [NEWLINE] [NEWLINE] Have you any experience of the military? There is a lot more to the front line than just the front line soldiers-they can't do much without huge support. That's<mask> the majority of military resources are actually based/expended. </s>
Label encoding: <s>It is very similar-you are around people in every situation. You have a camaraderie inherent within your group. Even if you hate your workmate, you will save their life while risking your own. You'll push people past their comfort zone, to the point where they are crying in pain while laughing at the shear insanity of what is happening. [NEWLINE] [NEWLINE] UTI's? Guys get those too. Along with STI's. And extremely high levels of mental illness. Arthritis too. Females, as far as I know, are actually more healthy due to better concentration on their health by our societies. Men don't tend to talk about their health as much, to their detriment. [NEWLINE] [NEWLINE] Female servicemen tend to take contraceptive pills in the field or employ other methods of preventing pregnancy, much like males. From my experience they are still very useful right up until they give birth, especially as officers etc. As far as I know, females also have higher pain thresholds, which would be incredibly useful during combat. [NEWLINE] [NEWLINE] Have you any experience of the military? There is a lot more to the front line than just the front line soldiers-they can't do much without huge support. That's where the majority of military resources are actually based/expended. </s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] (without fighting about it, talking or asking about it) [ENDQ] [NEWLINE] You didn't specifry this in your OP or at any other time. I have been<mask><mask> talking and arguing have been taking place - and have mentioned such in a previous post,<mask> you failed to correct me. This was not a foundation of your argument.<mask>, it remains that even without talking or asking, it is possible for a woman to decide that this is the last straw in her relationship - for example,<mask> her husband has previously demonstrated other controlling behaviours. [NEWLINE] [NEWLINE] At the same time, your premise assumes that decisions such<mask> ending a relationship are always made with a level head and some rationality. It is entirely possible for a woman to react irrationally to such a request due to hurt feelings, and<mask> end the relationship, without having ever cheated. People don't always act rationally. They do sometimes act in haste, without thinking things through - especially<mask> they've got pregnancy hormones flooding through them messing up their thought processes.<mask> again, here is another scenario in which a faithful wife might end a relationship over a request for a paternity test. [NEWLINE] [NEWLINE] [STARTQ] looks like you have very suspect motives [ENDQ] [NEWLINE] <mask> something looks like doesn't mean that is<mask> it is.</s>
Label encoding: <s> [STARTQ] (without fighting about it, talking or asking about it) [ENDQ] [NEWLINE] You didn't specifry this in your OP or at any other time. I have been assuming that talking and arguing have been taking place - and have mentioned such in a previous post, where you failed to correct me. This was not a foundation of your argument. However, it remains that even without talking or asking, it is possible for a woman to decide that this is the last straw in her relationship - for example, if her husband has previously demonstrated other controlling behaviours. [NEWLINE] [NEWLINE] At the same time, your premise assumes that decisions such as ending a relationship are always made with a level head and some rationality. It is entirely possible for a woman to react irrationally to such a request due to hurt feelings, and so end the relationship, without having ever cheated. People don't always act rationally. They do sometimes act in haste, without thinking things through - especially when they've got pregnancy hormones flooding through them messing up their thought processes. But again, here is another scenario in which a faithful wife might end a relationship over a request for a paternity test. [NEWLINE] [NEWLINE] [STARTQ] looks like you have very suspect motives [ENDQ] [NEWLINE] What something looks like doesn't mean that is what it is.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It doesn't change anything<mask> it doesn't matter. The only one harping on this is you. [ENDQ] [NEWLINE] Of course it matters. They *pay* for the whole website, making the whole thing their property.<mask> are people entitled to these resources? [NEWLINE] [NEWLINE] [STARTQ] I tolerate neonazis, racists, and other "haters"<mask> one day there may be a worthwhile cause. [ENDQ] [NEWLINE] <mask> this day ever comes, I'll be happy to revise my position. Until then, nothing of value was lost and nobody's prerogative were infringed upon. Again, I see no problem with "tolerating" anyone,<mask><mask> I do<mask> every day.<mask>, that's not<mask> you're asking for at all. You're asking for financial support and direct association. Two things that have *nothing* at all to do with a group's right to free speech. You want them to finance and associate with all speech indiscriminately, something you're obviously not asking of yourself or anyone else.<mask>? [NEWLINE] [NEWLINE] You can write banners. You can hang banners on your property. You cannot *demand* I pay for these banners or *demand* I hang them on *my* property.<mask> is the distinction<mask> thoroughly lost on you? </s>
Label encoding: <s> [STARTQ] It doesn't change anything because it doesn't matter. The only one harping on this is you. [ENDQ] [NEWLINE] Of course it matters. They *pay* for the whole website, making the whole thing their property. How are people entitled to these resources? [NEWLINE] [NEWLINE] [STARTQ] I tolerate neonazis, racists, and other "haters" because one day there may be a worthwhile cause. [ENDQ] [NEWLINE] If this day ever comes, I'll be happy to revise my position. Until then, nothing of value was lost and nobody's prerogative were infringed upon. Again, I see no problem with "tolerating" anyone, in fact I do so every day. However, that's not what you're asking for at all. You're asking for financial support and direct association. Two things that have *nothing* at all to do with a group's right to free speech. You want them to finance and associate with all speech indiscriminately, something you're obviously not asking of yourself or anyone else. Why? [NEWLINE] [NEWLINE] You can write banners. You can hang banners on your property. You cannot *demand* I pay for these banners or *demand* I hang them on *my* property. How is the distinction so thoroughly lost on you? </s>
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> you are cherrypicking data to ratify a position you already presupposed. [ENDQ] [NEWLINE] You're welcome to think that,<mask> you're wrong.  I took *all* the data I could find regarding national murder rates and national gun ownership rates, and presented *all* of it. <mask> you want to disregard the graphs that remove outliers, that's fine,<mask> the international data *still holds.* [NEWLINE] [NEWLINE] [STARTQ] <mask> do you mean "above average"? [ENDQ] [NEWLINE] Seriously? Look at the chart.  There are 39 data points for europe.  The median datapoint for gun ownership is 12:100. [NEWLINE] [NEWLINE] And actually, now that I'm not just eyeballing it (<mask> I assumed the median gun ownership was around 20:100), I can conclusively say that 80% of the top quartile of murder rates in Europe are in the bottom two quartiles of firearms ownership rate, and 70% of the bottom quartile of murder rates are in the top two quartiles of firearms ownership rates (classifying the median gun ownership datapoint<mask> part of the 2nd quartile). <mask>'s more, 60% of the bottom quartile of murder rates are in the *top* quartile of firearms ownership.</s>
Label encoding: <s> [STARTQ] I think you are cherrypicking data to ratify a position you already presupposed. [ENDQ] [NEWLINE] You're welcome to think that, but you're wrong.  I took *all* the data I could find regarding national murder rates and national gun ownership rates, and presented *all* of it.  If you want to disregard the graphs that remove outliers, that's fine, but the international data *still holds.* [NEWLINE] [NEWLINE] [STARTQ] What do you mean "above average"? [ENDQ] [NEWLINE] Seriously? Look at the chart.  There are 39 data points for europe.  The median datapoint for gun ownership is 12:100. [NEWLINE] [NEWLINE] And actually, now that I'm not just eyeballing it ( where I assumed the median gun ownership was around 20:100), I can conclusively say that 80% of the top quartile of murder rates in Europe are in the bottom two quartiles of firearms ownership rate, and 70% of the bottom quartile of murder rates are in the top two quartiles of firearms ownership rates (classifying the median gun ownership datapoint as part of the 2nd quartile).  What's more, 60% of the bottom quartile of murder rates are in the *top* quartile of firearms ownership.</s>
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You have touched on an interesting part of this topic, that I feel has been missed out. Essentially it is the abhorrently low conviction rate of rapists.<mask>, this is<mask> you are making the mistake, it is not<mask> society favors the rapist. It is due to that rape is almost exclusively done in private without witness and evidence is hard to obtain. I can only speak of the Scottish legal system<mask> there has been great efforts to try negate this. Judges have found ways around<mask> was considered foundation points of law to accommodate and convict, who they believe to be, rapists. Legislation has been put up to protect rape victims from question (which is required in an adversarial system). The problem is not societies view causing this low conviction rate. It is that we believe, rightly, that crimes must be proven beyond a reasonable doubt. A he said she said scenario which most rape cases are is almost impossible to convict on. Just a side note it is important for people to understand that the vast majority of rape cases are acquaintance rape. In other words the victim and the perpetrator know each other (i should note here this does not make it worse or better in nay respect<mask> it is worth noting) this idea of a man in the bushes is vastly mistaken.</s>
Label encoding: <s>You have touched on an interesting part of this topic, that I feel has been missed out. Essentially it is the abhorrently low conviction rate of rapists. However, this is where you are making the mistake, it is not because society favors the rapist. It is due to that rape is almost exclusively done in private without witness and evidence is hard to obtain. I can only speak of the Scottish legal system but there has been great efforts to try negate this. Judges have found ways around what was considered foundation points of law to accommodate and convict, who they believe to be, rapists. Legislation has been put up to protect rape victims from question (which is required in an adversarial system). The problem is not societies view causing this low conviction rate. It is that we believe, rightly, that crimes must be proven beyond a reasonable doubt. A he said she said scenario which most rape cases are is almost impossible to convict on. Just a side note it is important for people to understand that the vast majority of rape cases are acquaintance rape. In other words the victim and the perpetrator know each other (i should note here this does not make it worse or better in nay respect but it is worth noting) this idea of a man in the bushes is vastly mistaken.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am married. Very happily<mask>. I have one baby and he is our entire world. We don't want another for many reasons.<mask><mask><mask> your argument my choice is to abstain from sex with my husband, wear condoms with the man I've been with for over 7 years or risk it?<mask> should we not be allowed to make the responsible decision to use birth control to ensure we don't have children? [NEWLINE] [NEWLINE] It is much cheaper and better for society<mask> reasonable choices are provided. Unwanted pregnancies lead to tough decisions. Children that families can't afford<mask> they have to turn to the government for help. Children they don't want<mask> they are unloved and neglected which more than likely will cause problems for them and those around them later in life (whether it leads them to a life of crime, they become abusive or whatever). There are many who are put up for adoption and grow up in the foster system which is terrible and not near enough are adopted out or they are aborted. [NEWLINE] [NEWLINE] All of those things can be prevented by allowing all women access to birth control. In a society plagued by poverty and over population<mask><mask> this is the only logical choice. It's easy to preach abstinence<mask> unrealistic and proven to be ineffective. </s>
Label encoding: <s>I am married. Very happily so. I have one baby and he is our entire world. We don't want another for many reasons. So according to your argument my choice is to abstain from sex with my husband, wear condoms with the man I've been with for over 7 years or risk it? Why should we not be allowed to make the responsible decision to use birth control to ensure we don't have children? [NEWLINE] [NEWLINE] It is much cheaper and better for society when reasonable choices are provided. Unwanted pregnancies lead to tough decisions. Children that families can't afford so they have to turn to the government for help. Children they don't want so they are unloved and neglected which more than likely will cause problems for them and those around them later in life (whether it leads them to a life of crime, they become abusive or whatever). There are many who are put up for adoption and grow up in the foster system which is terrible and not near enough are adopted out or they are aborted. [NEWLINE] [NEWLINE] All of those things can be prevented by allowing all women access to birth control. In a society plagued by poverty and over population I think this is the only logical choice. It's easy to preach abstinence but unrealistic and proven to be ineffective. </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One thing you should know is that virtually every Studio Ghibli film revolves around a coming-of-age plot. The plot throws challenges and quick change at immature characters that are unprepared for it, they must adapt and must be brave and show<mask>'s inside, their internal struggle is often emphasized. [NEWLINE] [NEWLINE] <mask> every character, good or bad, has motives and development. Yubaba the bathouse owner is not evil just<mask>, she's got a business to run. The witch of the waste is heartbroken and proud, which leads to wrath. Lady Eboshi cared about her town, employing the marginalized lepers, starting nothing short of an industrial revolution, albeit forgetting the damage to the environment. [NEWLINE] [NEWLINE] The braveness of characters is<mask> rarely your macho western protagonist willing to face and slice million monsters and then self-immolate. The braveness is often displayed by vulnerable characters that have much to lose and don't have the skills or the strength to chop their way to victory, they must be cunning, firm, and rely on their (often newly made) friends<mask> they want to succeed. [NEWLINE] [NEWLINE] All this combined with the art and the music, like you mentioned, makes just about every Studio Ghibli film a masterpiece.</s>
Label encoding: <s>One thing you should know is that virtually every Studio Ghibli film revolves around a coming-of-age plot. The plot throws challenges and quick change at immature characters that are unprepared for it, they must adapt and must be brave and show what's inside, their internal struggle is often emphasized. [NEWLINE] [NEWLINE] Also every character, good or bad, has motives and development. Yubaba the bathouse owner is not evil just because, she's got a business to run. The witch of the waste is heartbroken and proud, which leads to wrath. Lady Eboshi cared about her town, employing the marginalized lepers, starting nothing short of an industrial revolution, albeit forgetting the damage to the environment. [NEWLINE] [NEWLINE] The braveness of characters is also rarely your macho western protagonist willing to face and slice million monsters and then self-immolate. The braveness is often displayed by vulnerable characters that have much to lose and don't have the skills or the strength to chop their way to victory, they must be cunning, firm, and rely on their (often newly made) friends if they want to succeed. [NEWLINE] [NEWLINE] All this combined with the art and the music, like you mentioned, makes just about every Studio Ghibli film a masterpiece.</s>
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do they have fiat on<mask> the rest of the world calls the city? [ENDQ] [NEWLINE] They don't. They've just expressed<mask> they *would like* the city to be called. People go along with it<mask> the name of the city is largely unimportant to them (<mask> not to the Indians) and calling it the traditional name is the nice thing to do. It's the same reason we use Myanmar instead of Burma. I have no emotional tie to the name Burma and the people there want to use Myanmar *<mask><mask> not*. [NEWLINE] [NEWLINE] [STARTQ] Like...<mask> is the name in french? In greek? Did they change<mask> the name of the city is called in Iroquois? [ENDQ] [NEWLINE] Again. They didn't change anything. They *can't* change anything. There are no English laws that dictate<mask> the name of something is.<mask> a critical mass of people decide to call Mumbai, nmhunateville then that's<mask> we would call it. [NEWLINE] [NEWLINE] The reason the transition from Bombay to Mumbai was accepted by the broader population is<mask> you kind of look like a jerk<mask> you continue to use a colonial name (for literally no real reason) that the formerly colonized people have stated *they don't like*. [NEWLINE] </s>
Label encoding: <s> [STARTQ] How do they have fiat on how the rest of the world calls the city? [ENDQ] [NEWLINE] They don't. They've just expressed what they *would like* the city to be called. People go along with it because the name of the city is largely unimportant to them ( but not to the Indians) and calling it the traditional name is the nice thing to do. It's the same reason we use Myanmar instead of Burma. I have no emotional tie to the name Burma and the people there want to use Myanmar * so why not*. [NEWLINE] [NEWLINE] [STARTQ] Like... What is the name in french? In greek? Did they change how the name of the city is called in Iroquois? [ENDQ] [NEWLINE] Again. They didn't change anything. They *can't* change anything. There are no English laws that dictate what the name of something is. If a critical mass of people decide to call Mumbai, nmhunateville then that's what we would call it. [NEWLINE] [NEWLINE] The reason the transition from Bombay to Mumbai was accepted by the broader population is because you kind of look like a jerk if you continue to use a colonial name (for literally no real reason) that the formerly colonized people have stated *they don't like*. [NEWLINE] </s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] True, we don't need to eat meat,<mask> it's more difficult (and expensive) to have a balanced vegetarian diet. [ENDQ] [NEWLINE] Whether something is difficult does not really affect whether it is immoral.  Instead it affects whether I will do it anyway, fully recognizing that I'm doing something wrong.  Sometimes I litter<mask> otherwise I would have to carry my trash for like a mile, and I recognize that it's wrong.  I'm not going to try to convince anyone that littering is morally acceptable just<mask> it's hard not to. [NEWLINE] [NEWLINE] [STARTQ] <mask> do you feel the need to compare "killing pigs for food" with just "killing dogs for fun"?<mask> not rather "killing pigs for food/fun", or "killing pigs/dogs for food"? [ENDQ] [NEWLINE] <mask> I don't eat dogs and I'm not 100% sure I'm opposed to killing pigs for fun.  I know for a fact that I am opposed to killing dogs for fun<mask>. [NEWLINE] [NEWLINE] [STARTQ] Can a dog care about his own death? [ENDQ] [NEWLINE] I don't know for sure,<mask> my guess would be that dogs want to live. <mask> someone were to convince me otherwise, I might be willing to change my view based on that.</s>
Label encoding: <s> [STARTQ] True, we don't need to eat meat, but it's more difficult (and expensive) to have a balanced vegetarian diet. [ENDQ] [NEWLINE] Whether something is difficult does not really affect whether it is immoral.  Instead it affects whether I will do it anyway, fully recognizing that I'm doing something wrong.  Sometimes I litter because otherwise I would have to carry my trash for like a mile, and I recognize that it's wrong.  I'm not going to try to convince anyone that littering is morally acceptable just because it's hard not to. [NEWLINE] [NEWLINE] [STARTQ] Why do you feel the need to compare "killing pigs for food" with just "killing dogs for fun"? Why not rather "killing pigs for food/fun", or "killing pigs/dogs for food"? [ENDQ] [NEWLINE] Because I don't eat dogs and I'm not 100% sure I'm opposed to killing pigs for fun.  I know for a fact that I am opposed to killing dogs for fun though. [NEWLINE] [NEWLINE] [STARTQ] Can a dog care about his own death? [ENDQ] [NEWLINE] I don't know for sure, but my guess would be that dogs want to live.  If someone were to convince me otherwise, I might be willing to change my view based on that.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Reading print media makes us more educated,<mask> opposed to the digital media we revel in today. [ENDQ] [NEWLINE] <mask> would that be true?  Even accepting that the medium is involved, are we really arguing over whether 50 Shades of Gray makes someone more educated than Shindler's List?  Even<mask> you *do* want to make that argument,<mask> are you defining "educated"?  Dimes to dollars I'm both more informed and more educated than most of the population, and I rarely read anything in print. [NEWLINE] [NEWLINE] [STARTQ] Playing violent video games and watching violent films makes us more violent. [ENDQ] [NEWLINE] There is no evidence to suggest that.  And there is no significant correlation between the rise of film or video games (the the violent versions of them) and increased violence in society. [NEWLINE] [NEWLINE] [STARTQ] Watching television makes us less social. [ENDQ] [NEWLINE] <mask> would it? <mask> it's a solitary activity? [NEWLINE] [NEWLINE] [STARTQ] Popcorn brain can be defined<mask> a brain<mask> accustomed to the constant stimulation of electronic multitasking that it becomes unfit for life offline,<mask> things pop at a slower pace. [ENDQ] [NEWLINE] Maybe,<mask> would that be a bad thing?  Someone able to, and accustomed to, thinking on multiple things at once.</s>
Label encoding: <s> [STARTQ] Reading print media makes us more educated, as opposed to the digital media we revel in today. [ENDQ] [NEWLINE] Why would that be true?  Even accepting that the medium is involved, are we really arguing over whether 50 Shades of Gray makes someone more educated than Shindler's List?  Even if you *do* want to make that argument, how are you defining "educated"?  Dimes to dollars I'm both more informed and more educated than most of the population, and I rarely read anything in print. [NEWLINE] [NEWLINE] [STARTQ] Playing violent video games and watching violent films makes us more violent. [ENDQ] [NEWLINE] There is no evidence to suggest that.  And there is no significant correlation between the rise of film or video games (the the violent versions of them) and increased violence in society. [NEWLINE] [NEWLINE] [STARTQ] Watching television makes us less social. [ENDQ] [NEWLINE] Why would it?  Because it's a solitary activity? [NEWLINE] [NEWLINE] [STARTQ] Popcorn brain can be defined as a brain so accustomed to the constant stimulation of electronic multitasking that it becomes unfit for life offline, where things pop at a slower pace. [ENDQ] [NEWLINE] Maybe, but would that be a bad thing?  Someone able to, and accustomed to, thinking on multiple things at once.</s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The war on terror has made terrorists hate us MORE,... [ENDQ] [NEWLINE] Perhaps they hate us more<mask> they kill us less. In the years leading up to 9/11 the number of terrorist attacks against Americans were slowly growing and obviously peaked at 9/11.<mask> then we've basically had no foreign terrorist attacks on our soil.<mask> they may hate us more<mask> they aren't attacking us on our homeland. [NEWLINE] [NEWLINE] You're absolutely right that from a purely objective viewpoint terrorism rates low on the scale of concerns even with 9/11<mask> that isn't<mask> people work. It's not called objectiveconcernism it's called terrorism. Terrorism incites terror.<mask> emotional people terror affects us. [NEWLINE] [NEWLINE] I don't know<mask> old you were in 2001<mask> those who were around know a country that was gripped with fear. Irrational, sure<mask> like I said we aren't rational creatures.<mask> the war did protect our freedom. Our freedom to have reduced fear. [NEWLINE] [NEWLINE] I'm with you that it is annoying and I was, and still am, frustrated with the entirely irrational fear of terrorism<mask><mask> a rational person I have to accept that most people see it irrationally and we have to deal with that and not just ignore it.</s>
Label encoding: <s> [STARTQ] The war on terror has made terrorists hate us MORE,... [ENDQ] [NEWLINE] Perhaps they hate us more but they kill us less. In the years leading up to 9/11 the number of terrorist attacks against Americans were slowly growing and obviously peaked at 9/11. Since then we've basically had no foreign terrorist attacks on our soil. So they may hate us more but they aren't attacking us on our homeland. [NEWLINE] [NEWLINE] You're absolutely right that from a purely objective viewpoint terrorism rates low on the scale of concerns even with 9/11 but that isn't how people work. It's not called objectiveconcernism it's called terrorism. Terrorism incites terror. As emotional people terror affects us. [NEWLINE] [NEWLINE] I don't know how old you were in 2001 but those who were around know a country that was gripped with fear. Irrational, sure but like I said we aren't rational creatures. So the war did protect our freedom. Our freedom to have reduced fear. [NEWLINE] [NEWLINE] I'm with you that it is annoying and I was, and still am, frustrated with the entirely irrational fear of terrorism but as a rational person I have to accept that most people see it irrationally and we have to deal with that and not just ignore it.</s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The point that the views [ENDQ] [NEWLINE] [NEWLINE] [STARTQ] &gt;Patriarchy exists [ENDQ] [NEWLINE] [STARTQ] and [ENDQ] [NEWLINE] [STARTQ] &gt;Women who don't believe that patriarchy exist are justified in that belief [ENDQ] [NEWLINE] [STARTQ] Are mutually exclusive. [ENDQ] [NEWLINE] That's not necessarily true. Typically,<mask> discussing whether or not a person's belief is justified, the justification is relative to evidence that that person has access to.<mask> it were impossible to be justified in holding a belief that wasn't true, justification would be an unnecessary concept<mask> it would simply be the case that subject A is justified in believing belief B<mask>f belief B is true, in which case a justified belief and a true belief would be the same thing.<mask>, instead we usually say that subject A is justified in believing belief B<mask> the evidence E that A is privy to logically leads to the conclusion that B is true. There may be other evidence that leads to the conclusion that B is false,<mask><mask> A is not privy to this evidence, they're still justified in believing B. [NEWLINE] [NEWLINE] Even if there is a patriarchy, there can be women whose evidence does not lead to the conclusion that one exists. These women are justified in their belief that there is no patriarchy.</s>
Label encoding: <s> [STARTQ] The point that the views [ENDQ] [NEWLINE] [NEWLINE] [STARTQ] &gt;Patriarchy exists [ENDQ] [NEWLINE] [STARTQ] and [ENDQ] [NEWLINE] [STARTQ] &gt;Women who don't believe that patriarchy exist are justified in that belief [ENDQ] [NEWLINE] [STARTQ] Are mutually exclusive. [ENDQ] [NEWLINE] That's not necessarily true. Typically, when discussing whether or not a person's belief is justified, the justification is relative to evidence that that person has access to. If it were impossible to be justified in holding a belief that wasn't true, justification would be an unnecessary concept because it would simply be the case that subject A is justified in believing belief B iff belief B is true, in which case a justified belief and a true belief would be the same thing. However, instead we usually say that subject A is justified in believing belief B if the evidence E that A is privy to logically leads to the conclusion that B is true. There may be other evidence that leads to the conclusion that B is false, but if A is not privy to this evidence, they're still justified in believing B. [NEWLINE] [NEWLINE] Even if there is a patriarchy, there can be women whose evidence does not lead to the conclusion that one exists. These women are justified in their belief that there is no patriarchy.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What kind of socialising are you doing? Going to social events is good for your social skills, yes,<mask><mask> you spend 4 years doing that then you haven't networked. The two groups of people in college who I've seen do exceptionally well in terms of their prospects are 1. those who are heavily involved in the Students' Union/clubs and societies/organising of major student events on campus, and 2. those who build fantastic rapport with their lecturers and the staff of their department (which absolutely requires going to class). The first group tend to have a lot of impressive extra-curricular stuff on their CV/resume that helps them get jobs outside of college, and the second group tend to be a lot more successful getting internships and references from their department.<mask><mask>, people who have left college benefit more from networking<mask> they know people in the same field<mask> them, which<mask> requires going to class to meet people studying the same things<mask> you.<mask> you're doing neither of those things, then you're just having fun with your friends, which isn't a bad thing<mask> you're getting decent grades,<mask> isn't going to help you in terms of job prospects more than doing well academically.</s>
Label encoding: <s>What kind of socialising are you doing? Going to social events is good for your social skills, yes, but if you spend 4 years doing that then you haven't networked. The two groups of people in college who I've seen do exceptionally well in terms of their prospects are 1. those who are heavily involved in the Students' Union/clubs and societies/organising of major student events on campus, and 2. those who build fantastic rapport with their lecturers and the staff of their department (which absolutely requires going to class). The first group tend to have a lot of impressive extra-curricular stuff on their CV/resume that helps them get jobs outside of college, and the second group tend to be a lot more successful getting internships and references from their department. In addition, people who have left college benefit more from networking if they know people in the same field as them, which also requires going to class to meet people studying the same things as you. If you're doing neither of those things, then you're just having fun with your friends, which isn't a bad thing if you're getting decent grades, but isn't going to help you in terms of job prospects more than doing well academically.</s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While it's probably true that filters can reproduce the musical experience of vinyl (<mask><mask> others have said, there are other experiences there that people may be seeking), I would say that there's a kind of camaraderie with the artist that only comes from experiencing the music the way that the artist intended it to be experienced. [NEWLINE] [NEWLINE] Perhaps that's "pretentious",<mask> it seems to me that you lose something<mask> you,<mask> the listener, change the music. Now, you may similarly *gain* something by doing this,<mask> it's not a valueless transaction. [NEWLINE] [NEWLINE] Now, some people prefer the sound of vinyl. Whether this is "right" or not is a purely subjective question. [NEWLINE] [NEWLINE] Many people seem to prefer movies shot on film,<mask><mask><mask> the undeniable fact that video/digital reproduces reality more accurately and with higher fidelity than all<mask> the most absurd film formats (and even then, color reproduction is better). [NEWLINE] [NEWLINE] <mask> someone likes the sound of vinyl, then it's not bad of them, in any of the directions you complain about, to seek out music that was composed and recorded for vinyl. They want to hear that sound, and they want to hear the sound that the artist intended.</s>
Label encoding: <s>While it's probably true that filters can reproduce the musical experience of vinyl ( though as others have said, there are other experiences there that people may be seeking), I would say that there's a kind of camaraderie with the artist that only comes from experiencing the music the way that the artist intended it to be experienced. [NEWLINE] [NEWLINE] Perhaps that's "pretentious", but it seems to me that you lose something when you, as the listener, change the music. Now, you may similarly *gain* something by doing this, but it's not a valueless transaction. [NEWLINE] [NEWLINE] Now, some people prefer the sound of vinyl. Whether this is "right" or not is a purely subjective question. [NEWLINE] [NEWLINE] Many people seem to prefer movies shot on film, in spite of the undeniable fact that video/digital reproduces reality more accurately and with higher fidelity than all but the most absurd film formats (and even then, color reproduction is better). [NEWLINE] [NEWLINE] If someone likes the sound of vinyl, then it's not bad of them, in any of the directions you complain about, to seek out music that was composed and recorded for vinyl. They want to hear that sound, and they want to hear the sound that the artist intended.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My example is too specific? Okay,<mask> about this: Architecture. Do you think of architecture<mask> a lucrative, not to mention useful profession? Would you expect architects with graduate degrees to have an easier time finding jobs than people with degrees in the arts? [Me too,<mask> we should both think again]( [URL].final.update1.pdf) (pdf). See the chart on page 7. [NEWLINE] [NEWLINE] An alarming 13.9% unemployment rate among recent college architects, compared to 11.1% in the arts (2010) and a relatively respectable 9.4% in the humanities. Hell, the humanities are within percentage point range of computer science grads, social scientists, law &amp; public policy people, etc. [NEWLINE] [NEWLINE] <mask> degrees win the employment challenge? Education and Health graduate degrees are tied at 1.9%,<mask> doctors and teachers are the most employable of all. Business school graduates have almost double that unemployment rate at 4.4%! Engineers with PhDs at 3.4%! [NEWLINE] [NEWLINE] <mask> for earnings, arts are at the bottom,<mask> they're in good company with science, agriculture, teachers, communications and journalism, law, recreation and psychology each within a few thousand dollars per year. </s>
Label encoding: <s>My example is too specific? Okay, how about this: Architecture. Do you think of architecture as a lucrative, not to mention useful profession? Would you expect architects with graduate degrees to have an easier time finding jobs than people with degrees in the arts? [Me too, but we should both think again]( [URL].final.update1.pdf) (pdf). See the chart on page 7. [NEWLINE] [NEWLINE] An alarming 13.9% unemployment rate among recent college architects, compared to 11.1% in the arts (2010) and a relatively respectable 9.4% in the humanities. Hell, the humanities are within percentage point range of computer science grads, social scientists, law &amp; public policy people, etc. [NEWLINE] [NEWLINE] What degrees win the employment challenge? Education and Health graduate degrees are tied at 1.9%, so doctors and teachers are the most employable of all. Business school graduates have almost double that unemployment rate at 4.4%! Engineers with PhDs at 3.4%! [NEWLINE] [NEWLINE] As for earnings, arts are at the bottom, but they're in good company with science, agriculture, teachers, communications and journalism, law, recreation and psychology each within a few thousand dollars per year. </s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, remember that we're talking about a 17 year-old woman. Not an 8 year old. [ENDQ] [NEWLINE] No, we're talking about puberty and ethics of sex with humans of young ages. Your reason for it *not* being unethical to have sex with a child (postpubescent) was that it didn't physically harm them.<mask> you didn't explain<mask> it's unethical to have sex with an 8 year old<mask> done in a way to avoid physical harm. Your argument is meaningless<mask> it specifically only applies to 17 year olds,<mask> then you would have to explain<mask> physical damage is a moral qualifier for one age range and not the other. [NEWLINE] [NEWLINE] [STARTQ] <mask> about Native American cultures<mask> this was common?<mask> about the Mosuo? [ENDQ] [NEWLINE] <mask> studies were done there, I'm sure sexual encounters with adults at a young age wouldn't cause psychological damage. [NEWLINE] [NEWLINE] <mask> we don't live in those societies.<mask> we lived 3000 years ago, 11 year old girls living in polygamous relationships would be psychologically undamaged<mask> they're in the same boat<mask> most other young girls.<mask><mask><mask>? Is the psychological damage proven today somehow invalidated by the lack of damage in history and in other societies?</s><pad>
Label encoding: <s> [STARTQ] However, remember that we're talking about a 17 year-old woman. Not an 8 year old. [ENDQ] [NEWLINE] No, we're talking about puberty and ethics of sex with humans of young ages. Your reason for it *not* being unethical to have sex with a child (postpubescent) was that it didn't physically harm them. But you didn't explain why it's unethical to have sex with an 8 year old when done in a way to avoid physical harm. Your argument is meaningless if it specifically only applies to 17 year olds, because then you would have to explain why physical damage is a moral qualifier for one age range and not the other. [NEWLINE] [NEWLINE] [STARTQ] What about Native American cultures where this was common? How about the Mosuo? [ENDQ] [NEWLINE] If studies were done there, I'm sure sexual encounters with adults at a young age wouldn't cause psychological damage. [NEWLINE] [NEWLINE] But we don't live in those societies. If we lived 3000 years ago, 11 year old girls living in polygamous relationships would be psychologically undamaged since they're in the same boat as most other young girls. But so what? Is the psychological damage proven today somehow invalidated by the lack of damage in history and in other societies?</s><pad>
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem is that "Date" has multiple definitions. It can mean two people who only just met and are now going to go on "a date," or it can mean me and my boyfriend of seven years who "are dating." [NEWLINE] [NEWLINE] It probably is morally wrong to date someone only for their looks for the long term.<mask> being sexually attracted to one another is part of dating, and<mask> that's<mask> motivates you to go on your first few dates, then that's okay.<mask> you go on more dates than that, then more substances<mask> just appearance should be driving you to do<mask>.<mask> for the first date or two, just appearance is okay<mask> a motivator. The first few dates exist *<mask> that* you can get to know someone's other traits in the first place.<mask> you don't know the person well enough<mask>, you *can't* be attracted to their other traits. Appearance is the only trait you can tell right away without getting to know someone; that's<mask> it's usually the first motivator. That's okay. [NEWLINE] [NEWLINE] And in both different situations above, you'd describe the two people<mask> "dating."<mask> that's<mask> the morals around the word are tricky. </s>
Label encoding: <s>The problem is that "Date" has multiple definitions. It can mean two people who only just met and are now going to go on "a date," or it can mean me and my boyfriend of seven years who "are dating." [NEWLINE] [NEWLINE] It probably is morally wrong to date someone only for their looks for the long term. But being sexually attracted to one another is part of dating, and if that's what motivates you to go on your first few dates, then that's okay. If you go on more dates than that, then more substances besides just appearance should be driving you to do so. But for the first date or two, just appearance is okay as a motivator. The first few dates exist * so that* you can get to know someone's other traits in the first place. If you don't know the person well enough yet, you *can't* be attracted to their other traits. Appearance is the only trait you can tell right away without getting to know someone; that's why it's usually the first motivator. That's okay. [NEWLINE] [NEWLINE] And in both different situations above, you'd describe the two people as "dating." So that's why the morals around the word are tricky. </s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, it's going to be hard to change your view on something<mask> subjective<mask> finding fireworks boring.  I personally think they're mildly entertaining - sometimes it's just nice to look at something pretty, whether it's art or flowers or a bonfire.  And watching things blow up has always been a popular pastime... I mean, just look at all the Youtube videos of shit getting blown to pieces.  It's entertaining in in small doses.  Most of the fireworks shows I've seen last for about 15 minutes,<mask> yeah, your friend's hour and a half extavanganza probably got pretty boring. [NEWLINE] [NEWLINE] Fireworks do have a long history (back to the 7th century) of being used for celebrations in many cultures,<mask><mask> there's nothing that ties them to the concept of independence, they are tied to celebration in general. [NEWLINE] [NEWLINE] 40% of injuries come from small, "safe" fireworks like sparklers and firecrackers, not from backyard displays.  And 50% of injuries come from intentional misuse, aka pure dumbfuckery.  There were about 11 fireworks-related deaths in 2014...<mask> other summer activities like swimming, grilling, etc. cause many deaths<mask> well.</s>
Label encoding: <s>Well, it's going to be hard to change your view on something as subjective as finding fireworks boring.  I personally think they're mildly entertaining - sometimes it's just nice to look at something pretty, whether it's art or flowers or a bonfire.  And watching things blow up has always been a popular pastime... I mean, just look at all the Youtube videos of shit getting blown to pieces.  It's entertaining in in small doses.  Most of the fireworks shows I've seen last for about 15 minutes, so yeah, your friend's hour and a half extavanganza probably got pretty boring. [NEWLINE] [NEWLINE] Fireworks do have a long history (back to the 7th century) of being used for celebrations in many cultures, so while there's nothing that ties them to the concept of independence, they are tied to celebration in general. [NEWLINE] [NEWLINE] 40% of injuries come from small, "safe" fireworks like sparklers and firecrackers, not from backyard displays.  And 50% of injuries come from intentional misuse, aka pure dumbfuckery.  There were about 11 fireworks-related deaths in 2014... but other summer activities like swimming, grilling, etc. cause many deaths as well.</s>
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand that a very small percentage of decisions to have a child might be made for selfless reasons. Maybe your partner really wants a kid and you don't, for example, and you do it to make him or her happy.<mask> beyond doing it to please *another person*,<mask> possible selfless reason could there be to have a child? I am not saying there aren't any,<mask> I generally have a pretty good imagination, and I can't think of one,<mask> I want to know<mask> it is you have in mind. [NEWLINE] [NEWLINE] EDIT: I rescind my statement. I just realized that EVERYTHING we do is either to please ourselves or another person (barring depression and the like, which might make us do things that do neither, like eating a donut that isn't even enjoyable).<mask> that was a stupid question I asked up there. Nevermind. [NEWLINE] [NEWLINE]...<mask> obviously the vast, vast majority of decisions to have children are selfish. Of course, the vast, vast majority of ALL decisions are selfish. It's like Louie CK says: every day I keep driving my car instead of selling it and giving the money to starving people, I'm killing them. I don't mind,<mask>.</s>
Label encoding: <s>I understand that a very small percentage of decisions to have a child might be made for selfless reasons. Maybe your partner really wants a kid and you don't, for example, and you do it to make him or her happy. But beyond doing it to please *another person*, what possible selfless reason could there be to have a child? I am not saying there aren't any, but I generally have a pretty good imagination, and I can't think of one, so I want to know what it is you have in mind. [NEWLINE] [NEWLINE] EDIT: I rescind my statement. I just realized that EVERYTHING we do is either to please ourselves or another person (barring depression and the like, which might make us do things that do neither, like eating a donut that isn't even enjoyable). So that was a stupid question I asked up there. Nevermind. [NEWLINE] [NEWLINE]... but obviously the vast, vast majority of decisions to have children are selfish. Of course, the vast, vast majority of ALL decisions are selfish. It's like Louie CK says: every day I keep driving my car instead of selling it and giving the money to starving people, I'm killing them. I don't mind, though.</s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think too many women fear rape too much. I know it would be horrible,<mask> at least you'd be alive. You can get better. [NEWLINE] [NEWLINE] I hate seeing guys chase women. Women should be chasing us. We're the ones making the money, and have the possessions. I don't chase, have never chased and never will.<mask>, women still find me attractive. Imagine that, sex isn't the most important thing in the world. Guys of the world, you need to realize that. Maybe then women will wake up and figure out that we will not cater to irrational, and illogical behavior. [NEWLINE] [NEWLINE] I'm a 17 year old straight and atheist boy, and I don't respect woman that post nudes of themselves,<mask><mask> they have no self respect and none of them have ever had a stable relationship. A couple months ago, my dad and I went to a car show in Norway, we walked past 5 beautiful girls in bikinis washing cars in about 0 degrees Celsius, and I said to my dad, ''look at those beautiful girls with no self respect, such a shame'', and I walked away.<mask> on the flip side, I watch porn and I enjoy it -.-</s>
Label encoding: <s>I think too many women fear rape too much. I know it would be horrible, but at least you'd be alive. You can get better. [NEWLINE] [NEWLINE] I hate seeing guys chase women. Women should be chasing us. We're the ones making the money, and have the possessions. I don't chase, have never chased and never will. Yet, women still find me attractive. Imagine that, sex isn't the most important thing in the world. Guys of the world, you need to realize that. Maybe then women will wake up and figure out that we will not cater to irrational, and illogical behavior. [NEWLINE] [NEWLINE] I'm a 17 year old straight and atheist boy, and I don't respect woman that post nudes of themselves, I think they have no self respect and none of them have ever had a stable relationship. A couple months ago, my dad and I went to a car show in Norway, we walked past 5 beautiful girls in bikinis washing cars in about 0 degrees Celsius, and I said to my dad, ''look at those beautiful girls with no self respect, such a shame'', and I walked away. but on the flip side, I watch porn and I enjoy it -.-</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm pretty sure 'doing oral' falls into the spectrum of having sex. [NEWLINE] Regardless,<mask><mask> with the poster above: in a way, there is a responsibility (on both sides!) to communicate. It's been said here that it might be awkward for the girl to make clear she does not want sex,<mask> I am sure there are MANY ways she could make it clear. For example,<mask> they go home together, she could say: Hey, I really like you<mask> I hope you don't expect something to happen, I'm not in the mood/whatever. [NEWLINE] Hopefully, this should be understood<mask> the guy has a brain.<mask> he FORCES himself onto her (after she said<mask> ), that would be closer to rape than a scenario<mask> she gets naked, doesn't communicate<mask> she wants and 'lets it happen'. At least that's<mask><mask>. I know real life situations aren't always<mask> black and white clean cut like these thought models,<mask> I wanted to demonstrate<mask><mask><mask> a girl getting naked in bed with a guy is<mask> sending mixed signals. The same goes for a guy getting naked in bed with a girl without clearly stating he does not want anything to happen. </s>
Label encoding: <s>I'm pretty sure 'doing oral' falls into the spectrum of having sex. [NEWLINE] Regardless, I agree with the poster above: in a way, there is a responsibility (on both sides!) to communicate. It's been said here that it might be awkward for the girl to make clear she does not want sex, but I am sure there are MANY ways she could make it clear. For example, if they go home together, she could say: Hey, I really like you but I hope you don't expect something to happen, I'm not in the mood/whatever. [NEWLINE] Hopefully, this should be understood if the guy has a brain. If he FORCES himself onto her (after she said so ), that would be closer to rape than a scenario where she gets naked, doesn't communicate what she wants and 'lets it happen'. At least that's my opinion. I know real life situations aren't always as black and white clean cut like these thought models, but I wanted to demonstrate why I agree a girl getting naked in bed with a guy is indeed sending mixed signals. The same goes for a guy getting naked in bed with a girl without clearly stating he does not want anything to happen. </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think much of these arguments stem from having different definitions of a soul. Is your soul your personality?<mask> you act, think, and feel? Or is it an indefinable, intangible form of consciousness? Our connection with the universe,<mask> to speak. Whether or not you believe in a god, the universe exists<mask> an all powerful being. It is ever expanding, it is not entirely understood by humans, and I honestly don't think it ever will be. It is our creator and one day it will be our destroyer. I am of the belief that our souls are our connection to the universe. I can't prove it, and I really don't have to. That is the beauty of faith. [NEWLINE] Will my soul live on after I die? I don't know, maybe it moves on to my next form of existence, like reincarnation. Maybe it is released into the universe to join with its creator. Maybe it will disintegrate with my body. All I know is that it is there. I feel it every time I look at the stars. [NEWLINE] I realize<mask> lame all this sounds, and that it probably won't change your view, it's just<mask> I feel about the subject. </s>
Label encoding: <s>I think much of these arguments stem from having different definitions of a soul. Is your soul your personality? How you act, think, and feel? Or is it an indefinable, intangible form of consciousness? Our connection with the universe, so to speak. Whether or not you believe in a god, the universe exists as an all powerful being. It is ever expanding, it is not entirely understood by humans, and I honestly don't think it ever will be. It is our creator and one day it will be our destroyer. I am of the belief that our souls are our connection to the universe. I can't prove it, and I really don't have to. That is the beauty of faith. [NEWLINE] Will my soul live on after I die? I don't know, maybe it moves on to my next form of existence, like reincarnation. Maybe it is released into the universe to join with its creator. Maybe it will disintegrate with my body. All I know is that it is there. I feel it every time I look at the stars. [NEWLINE] I realize how lame all this sounds, and that it probably won't change your view, it's just how I feel about the subject. </s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know using genders in the example was just to make it<mask> clear<mask> possible. [NEWLINE] [NEWLINE] [STARTQ] I do,<mask>, stand by the idea that a contracted, legal commitment is not the same<mask> a social "community". The government can't technically force me to pay off my neighbor's debts. They can come after me for my wife's debt. The government can't force me to let my neighbor see my child just<mask> my neighbor has helped raise them for 10 years. The government can quite literally throw me in jail<mask> I don't let my wife see my child. [ENDQ] [NEWLINE] I completely agree with you.<mask> nobody's forcing these people to get married. They can choose to live together in a 'community' without getting married<mask> that's<mask> they would like to do.<mask><mask> polygamous marriage would encourage a more community-centred approach to raising children<mask><mask> you rightly point out, there are legal distinctions between the two. [NEWLINE] [NEWLINE] [STARTQ] <mask> again, I'm not really interested in arguing anymore than I was just trying to toss an idea out there. [ENDQ] [NEWLINE] Oh well, thank you very much for putting forward your ideas anyway! They were very interesting to hear and certainly provided some food for thought!</s>
Label encoding: <s>I know using genders in the example was just to make it as clear as possible. [NEWLINE] [NEWLINE] [STARTQ] I do, however, stand by the idea that a contracted, legal commitment is not the same as a social "community". The government can't technically force me to pay off my neighbor's debts. They can come after me for my wife's debt. The government can't force me to let my neighbor see my child just because my neighbor has helped raise them for 10 years. The government can quite literally throw me in jail if I don't let my wife see my child. [ENDQ] [NEWLINE] I completely agree with you. But nobody's forcing these people to get married. They can choose to live together in a 'community' without getting married if that's what they would like to do. I think polygamous marriage would encourage a more community-centred approach to raising children but as you rightly point out, there are legal distinctions between the two. [NEWLINE] [NEWLINE] [STARTQ] So again, I'm not really interested in arguing anymore than I was just trying to toss an idea out there. [ENDQ] [NEWLINE] Oh well, thank you very much for putting forward your ideas anyway! They were very interesting to hear and certainly provided some food for thought!</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Abortion is not always an available option for women. [ENDQ] [NEWLINE] The number of women for whom abortion is not a *medical* option is very small.  The availability of abortion<mask> a *practical* option (due to cost or location issues, for example) is an issue of execution rather than morality or legality. [NEWLINE] [NEWLINE] The solution to issues with abortion availability is not arbitrarily depriving all men of reproductive autonomy, it's fixing issues related to abortion availability. [NEWLINE] [NEWLINE] [STARTQ] And even<mask> it is, abortion affects a woman in ways that it does not and can not affect men. [ENDQ] [NEWLINE] Absolutely,<mask><mask> does this have to do with the argument or point at hand?  Fundamentally, the existence of abortion means that women do *not* necessarily have to carry a pregnancy to term, let alone give birth and then take legal/financial responsibility for the resulting child. [NEWLINE] [NEWLINE] Which are closer, abortion and legal/paper/financial abortion or abortion and forced legal/biological parenthood?  It isn't really a complicated question, and<mask><mask><mask> legal fairness (and justice, for that matter) are concerned it makes sense for us to opt for the more analogous of the two.</s>
Label encoding: <s> [STARTQ] Abortion is not always an available option for women. [ENDQ] [NEWLINE] The number of women for whom abortion is not a *medical* option is very small.  The availability of abortion as a *practical* option (due to cost or location issues, for example) is an issue of execution rather than morality or legality. [NEWLINE] [NEWLINE] The solution to issues with abortion availability is not arbitrarily depriving all men of reproductive autonomy, it's fixing issues related to abortion availability. [NEWLINE] [NEWLINE] [STARTQ] And even when it is, abortion affects a woman in ways that it does not and can not affect men. [ENDQ] [NEWLINE] Absolutely, but what does this have to do with the argument or point at hand?  Fundamentally, the existence of abortion means that women do *not* necessarily have to carry a pregnancy to term, let alone give birth and then take legal/financial responsibility for the resulting child. [NEWLINE] [NEWLINE] Which are closer, abortion and legal/paper/financial abortion or abortion and forced legal/biological parenthood?  It isn't really a complicated question, and as far as legal fairness (and justice, for that matter) are concerned it makes sense for us to opt for the more analogous of the two.</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Drug policies typically target poor minority communities. [ENDQ] [NEWLINE] Don't use drugs.  Should drugs like weed be legal everywhere? Probably.<mask> they aren't,<mask><mask> you just don't use, sell, etc. you should be okay. [NEWLINE] [NEWLINE] [STARTQ] A white man with a criminal record is more likely to get the job than a.black man with no record<mask> the same qualifications. [ENDQ] [NEWLINE] Probably.<mask> honestly just saying "criminal record" is pretty vague, and unless an employer really needs position filled, they will skip over both a black and white person with "criminal records" [NEWLINE] [NEWLINE] [STARTQ] A person with an<mask>ian sounding name is about half<mask> likely to get an interview<mask> a white sounding name with the same qualifications. [ENDQ] [NEWLINE] Perhaps.  Unless that employer needs to make a "diversity hire." Then maybe the Asian has the upper hand. I<mask> don't like that as an example at all<mask> it is just in a vacuum.<mask> often are two people applying for the same job going to have the exact same qualifications and just happen to be different races. [NEWLINE] [NEWLINE] Can't really argue about any of the other examples you point out<mask> I would tend to agree with them.  </s>
Label encoding: <s> [STARTQ] Drug policies typically target poor minority communities. [ENDQ] [NEWLINE] Don't use drugs.  Should drugs like weed be legal everywhere? Probably. But they aren't, so if you just don't use, sell, etc. you should be okay. [NEWLINE] [NEWLINE] [STARTQ] A white man with a criminal record is more likely to get the job than a.black man with no record but the same qualifications. [ENDQ] [NEWLINE] Probably. But honestly just saying "criminal record" is pretty vague, and unless an employer really needs position filled, they will skip over both a black and white person with "criminal records" [NEWLINE] [NEWLINE] [STARTQ] A person with an asian sounding name is about half as likely to get an interview as a white sounding name with the same qualifications. [ENDQ] [NEWLINE] Perhaps.  Unless that employer needs to make a "diversity hire." Then maybe the Asian has the upper hand. I also don't like that as an example at all because it is just in a vacuum. How often are two people applying for the same job going to have the exact same qualifications and just happen to be different races. [NEWLINE] [NEWLINE] Can't really argue about any of the other examples you point out because I would tend to agree with them.  </s>
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with this CMV is that it ignores the realities of the situation and creates a false dichotomy i.e. either we blame the companies for selling diabetes inducing food or blame the consumer who made the poor purchase decision. [NEWLINE] [NEWLINE] The real answer is blame Congress.  Its no secret that the US taxpayers subsidize the corn and soy industries to the tune of billions of dollars a year ( [URL] ). [NEWLINE] [NEWLINE] <mask> effect does this have on our producing/purchasing habits?  First, it incentivizes the producers to produce more of the foods that are killing us (high fructose corn syrup).  Second, it creates absurd results such<mask> the cost of a bottle of soda being, in many instances, cheaper than a bottle of water. [NEWLINE] [NEWLINE] The take away is don't blame companies (they are just out to profit) and don't blame consumers (they just buy<mask>'s cheap and tastes good).  Blame the people who are creating the bad incentives through mass subsidization of harmful foods. [NEWLINE] [NEWLINE] <mask> we were to say, remove subsidies for corn and soy and instead subsidized vegetables and fruit this entire problem would solve itself through the market.  Supply and demand.  Economics 101.</s>
Label encoding: <s>The problem with this CMV is that it ignores the realities of the situation and creates a false dichotomy i.e. either we blame the companies for selling diabetes inducing food or blame the consumer who made the poor purchase decision. [NEWLINE] [NEWLINE] The real answer is blame Congress.  Its no secret that the US taxpayers subsidize the corn and soy industries to the tune of billions of dollars a year ( [URL] ). [NEWLINE] [NEWLINE] What effect does this have on our producing/purchasing habits?  First, it incentivizes the producers to produce more of the foods that are killing us (high fructose corn syrup).  Second, it creates absurd results such as the cost of a bottle of soda being, in many instances, cheaper than a bottle of water. [NEWLINE] [NEWLINE] The take away is don't blame companies (they are just out to profit) and don't blame consumers (they just buy what's cheap and tastes good).  Blame the people who are creating the bad incentives through mass subsidization of harmful foods. [NEWLINE] [NEWLINE] If we were to say, remove subsidies for corn and soy and instead subsidized vegetables and fruit this entire problem would solve itself through the market.  Supply and demand.  Economics 101.</s>
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>well i see that you've changed your view,<mask> I must say I was somewhat refreshed to see this post. I recall a discussion with a friend of mine who is now into pain pills and who has an uncle who struggles with cocaine and crack. I come from a different side of the whole drug thing being that both my parents and a lot of my family struggle with alcohol dependence. I told him that alcohol was a hard drug, and he lost it. He goes on and on about<mask> crack is a hard drug, heroin is a hard drug. It kills people, he says. Yes,<mask><mask> does alcohol. It destroys your body, your mind, and your relationships<mask> you can't handle it. Of course it's not crack,<mask> it will fuck shit up regardless. It is a hard drug in that it can kill you in one night or over 40 years. It's poison. [NEWLINE] [NEWLINE] That being said I still get fucked up now and then cause I'd feel pretty dumb sitting in a hospital bed dying from nothing. [NEWLINE] [NEWLINE] I just wanted to tell you I appreciate your view. It will change somewhat, and already has to a degree,<mask> it is refreshing to hear nonetheless.</s>
Label encoding: <s>well i see that you've changed your view, but I must say I was somewhat refreshed to see this post. I recall a discussion with a friend of mine who is now into pain pills and who has an uncle who struggles with cocaine and crack. I come from a different side of the whole drug thing being that both my parents and a lot of my family struggle with alcohol dependence. I told him that alcohol was a hard drug, and he lost it. He goes on and on about how crack is a hard drug, heroin is a hard drug. It kills people, he says. Yes, but so does alcohol. It destroys your body, your mind, and your relationships if you can't handle it. Of course it's not crack, but it will fuck shit up regardless. It is a hard drug in that it can kill you in one night or over 40 years. It's poison. [NEWLINE] [NEWLINE] That being said I still get fucked up now and then cause I'd feel pretty dumb sitting in a hospital bed dying from nothing. [NEWLINE] [NEWLINE] I just wanted to tell you I appreciate your view. It will change somewhat, and already has to a degree, but it is refreshing to hear nonetheless.</s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While we're talking about<mask> screwed up the OP's position about a father's right to court-ordered visitation is, can we just acknowledge that<mask> the father may or may not be a good guy, his ex-wife is clearly being a huge asshole.  She makes about $250k/year consistently, and she was able to secure a job in New Jersey making [STARTQ] $400k per year. <mask><mask>, the husband is in a commission-only job that has given him $100k/year and $30k/year depending on the year. [ENDQ] [NEWLINE] <mask> mom makes $250k *at least* and dad makes between $30k and $100k.  She petitions to take the 4 kids with her 1000 miles away, acts like it's no big deal and that dad can get all the same visitation by flying to New Jersey twice a month and staying in a hotel or in an "in-law suite" with his ex-wife and her family. [NEWLINE] [NEWLINE] And, to put the fucking cherry on top,  *she asks the court to make him pay her MORE child support!*" <mask> a peach she is!  No wonder he drinks.</s>
Label encoding: <s>While we're talking about how screwed up the OP's position about a father's right to court-ordered visitation is, can we just acknowledge that while the father may or may not be a good guy, his ex-wife is clearly being a huge asshole.  She makes about $250k/year consistently, and she was able to secure a job in New Jersey making [STARTQ] $400k per year.  In contrast, the husband is in a commission-only job that has given him $100k/year and $30k/year depending on the year. [ENDQ] [NEWLINE] So mom makes $250k *at least* and dad makes between $30k and $100k.  She petitions to take the 4 kids with her 1000 miles away, acts like it's no big deal and that dad can get all the same visitation by flying to New Jersey twice a month and staying in a hotel or in an "in-law suite" with his ex-wife and her family. [NEWLINE] [NEWLINE] And, to put the fucking cherry on top,  *she asks the court to make him pay her MORE child support!*"  What a peach she is!  No wonder he drinks.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can we make room here by introducing the notion of *implied consent?*  It seems like people can be beholden to contracts even without their consent, given they chose to perform certain actions. [NEWLINE] [NEWLINE] <mask> I walk into 7/11 and eat a candy bar off the shelf, I am implicitly agreeing to pay for it, even<mask> I don't explicitly say<mask>. <mask> I then try to leave without paying, they will rightfully take the dollar from me by force. [NEWLINE] [NEWLINE] [STARTQ] Wolff makes the distinction that democracy is simply majoritarianism. Many would assume that democracy is a just system,<mask> it affords the most freedoms to the greatest number of people in a population,<mask> all that basically means that it is a better alternative to other forms of government. Just<mask> democracy violates the fewest rights in comparison to other systems does not make it fully just. It merely makes it the least unjust of all the different forms of government currently possible. [ENDQ] [NEWLINE] <mask> it can be shown that an absence of government would be *even more* unjust, then we could conclude that democracy is not just the least unjust form of government,<mask> the least unjust structure of society in general.</s>
Label encoding: <s>Can we make room here by introducing the notion of *implied consent?*  It seems like people can be beholden to contracts even without their consent, given they chose to perform certain actions. [NEWLINE] [NEWLINE] If I walk into 7/11 and eat a candy bar off the shelf, I am implicitly agreeing to pay for it, even if I don't explicitly say so.  If I then try to leave without paying, they will rightfully take the dollar from me by force. [NEWLINE] [NEWLINE] [STARTQ] Wolff makes the distinction that democracy is simply majoritarianism. Many would assume that democracy is a just system, because it affords the most freedoms to the greatest number of people in a population, but all that basically means that it is a better alternative to other forms of government. Just because democracy violates the fewest rights in comparison to other systems does not make it fully just. It merely makes it the least unjust of all the different forms of government currently possible. [ENDQ] [NEWLINE] If it can be shown that an absence of government would be *even more* unjust, then we could conclude that democracy is not just the least unjust form of government, but the least unjust structure of society in general.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your first example completely mischaracterizes<mask> op is getting at.  Your example is pretty blatant victim blaming and that's one of the things he's arguing against. [NEWLINE] [NEWLINE] [STARTQ] The correct response is: "Well, that's unfortunate. That's a really unsafe area. The news/the police/the community should do more to ensure peoples awareness and safety in that area." [ENDQ] [NEWLINE] This is not the correct response at all<mask> you completely fail to address the issue that Mary actually did make a risky decision by walking through that area or whatever. A much better response (and I believe a closer representation of<mask> op is getting at) would look more like this: "Well, that's unfortunate. That's a really unsafe area. The news/the police/the community should do more to ensure peoples awareness and safety in that area.  For the time being until the problem is fixed<mask>, it's probably best for you to avoid that area at night,<mask> walking through it alone is an unnecessary risk.". Note that this is very similar to<mask> you claim the correct response is,<mask> it<mask> attempts to point out something Mary can do in the future to avoid this happening again.</s>
Label encoding: <s>Your first example completely mischaracterizes what op is getting at.  Your example is pretty blatant victim blaming and that's one of the things he's arguing against. [NEWLINE] [NEWLINE] [STARTQ] The correct response is: "Well, that's unfortunate. That's a really unsafe area. The news/the police/the community should do more to ensure peoples awareness and safety in that area." [ENDQ] [NEWLINE] This is not the correct response at all because you completely fail to address the issue that Mary actually did make a risky decision by walking through that area or whatever. A much better response (and I believe a closer representation of what op is getting at) would look more like this: "Well, that's unfortunate. That's a really unsafe area. The news/the police/the community should do more to ensure peoples awareness and safety in that area.  For the time being until the problem is fixed though, it's probably best for you to avoid that area at night, as walking through it alone is an unnecessary risk.". Note that this is very similar to what you claim the correct response is, but it also attempts to point out something Mary can do in the future to avoid this happening again.</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In our system of justice, sentence length has no relationship with required certainty of guilt. [ENDQ] [NEWLINE] To your first point, death is permanent. There is no going back after you kill someone. Could you accept that our confidence level of guilt is built in to the severity of the punishment? The reason that the death penalty has *mandatory* appeals is<mask> that we can increase our confidence. [NEWLINE] [NEWLINE] [STARTQ] You can't get 50 years in prison back. You might say that death is worse than 50 years in prison,<mask> I'm not<mask> sure. [ENDQ] [NEWLINE] You can't get 50 years back,<mask> you get a hell of a lot of money to use on the remainder of your life. I imagine 50 years is on the high end,<mask> losing 20 or<mask> years of your life vs all of it is a huge difference. [NEWLINE] [NEWLINE] [STARTQ] <mask> really<mask><mask><mask> there's any reasonable prospect that someone who's been ruled guilty might later be found not guilty, that's a reasonable doubt and they should be set totally and unconditionally free. [ENDQ] [NEWLINE] The jury rules on evidence available at the time. You can't predict any future revelations, and rule not guilty based on that.</s>
Label encoding: <s> [STARTQ] In our system of justice, sentence length has no relationship with required certainty of guilt. [ENDQ] [NEWLINE] To your first point, death is permanent. There is no going back after you kill someone. Could you accept that our confidence level of guilt is built in to the severity of the punishment? The reason that the death penalty has *mandatory* appeals is so that we can increase our confidence. [NEWLINE] [NEWLINE] [STARTQ] You can't get 50 years in prison back. You might say that death is worse than 50 years in prison, but I'm not so sure. [ENDQ] [NEWLINE] You can't get 50 years back, but you get a hell of a lot of money to use on the remainder of your life. I imagine 50 years is on the high end, so losing 20 or so years of your life vs all of it is a huge difference. [NEWLINE] [NEWLINE] [STARTQ] but really I think if there's any reasonable prospect that someone who's been ruled guilty might later be found not guilty, that's a reasonable doubt and they should be set totally and unconditionally free. [ENDQ] [NEWLINE] The jury rules on evidence available at the time. You can't predict any future revelations, and rule not guilty based on that.</s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry it took me<mask> long. I don't even know<mask> Ishould be responding at this point,<mask> I will anyway. [NEWLINE] [NEWLINE] [NEWLINE] I've heard of it, yeah! Nihilists get a bad rap,<mask> honestly,<mask><mask> it's the most soundly logical philosophical movement there ever was. I honestly can't see the point in doing things the way everyone else seems to. [NEWLINE] [NEWLINE] [NEWLINE] For example, I'm interested in politics,<mask><mask> reading political news and watching debates I can only sustain my interest for a short<mask>. After a<mask> of candidates passionately discussing policies I start to rhink, "God, this is<mask> futile. They act like it's the most important thing in the universe. Of course university students should be able to receive their degree at a fair price. Of course. Of course birthright citizenship should be a given and not even brought nto question! Of course. Those things matter.<mask> really, who cares? At the end of the day we're all just gonna die anyway." [NEWLINE] [NEWLINE] [NEWLINE] Fun, isn't it? It makes me which I could believe in some fairy tale God just to appease my need for purpose.</s>
Label encoding: <s>I'm sorry it took me so long. I don't even know if Ishould be responding at this point, but I will anyway. [NEWLINE] [NEWLINE] [NEWLINE] I've heard of it, yeah! Nihilists get a bad rap, but honestly, I think it's the most soundly logical philosophical movement there ever was. I honestly can't see the point in doing things the way everyone else seems to. [NEWLINE] [NEWLINE] [NEWLINE] For example, I'm interested in politics, but while reading political news and watching debates I can only sustain my interest for a short while. After a while of candidates passionately discussing policies I start to rhink, "God, this is so futile. They act like it's the most important thing in the universe. Of course university students should be able to receive their degree at a fair price. Of course. Of course birthright citizenship should be a given and not even brought nto question! Of course. Those things matter. But really, who cares? At the end of the day we're all just gonna die anyway." [NEWLINE] [NEWLINE] [NEWLINE] Fun, isn't it? It makes me which I could believe in some fairy tale God just to appease my need for purpose.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Incestuous couples are more likely to have deformed children,<mask> marriage is not about reproduction [ENDQ] [NEWLINE] One of the main functions of marriage is to indicate which of a person's children is *legitimate* (that is, which inherit) and which not—and that applies equally well to homosexual people<mask> to heterosexual ones. [NEWLINE] [NEWLINE] [STARTQ] there is no legal […] argument against allowing incestuous marriage that hasn't been thrown away in pursuit of gay marriage. [ENDQ] [NEWLINE] <mask> in a secular democratic society, such<mask> the one I live in and I'l bet the one that you live in too, the laws are whatever we collectively choose them to be. [NEWLINE] [NEWLINE] In the UK it is illegal for very close relatives (including adoptive parents) to have sex, married or no. The rules about who can and cannot marry [are different, more complicated and change more over time]( [URL] #England_and_Wales).  It's my understanding that the reasons for these rules is on the one hand eugenic and on the other to sanction various kinds of relationship which are very likely to be abusive or exploitative. Gay marriage doesn't seem to have any impact on these reasons.</s>
Label encoding: <s> [STARTQ] Incestuous couples are more likely to have deformed children, but marriage is not about reproduction [ENDQ] [NEWLINE] One of the main functions of marriage is to indicate which of a person's children is *legitimate* (that is, which inherit) and which not—and that applies equally well to homosexual people as to heterosexual ones. [NEWLINE] [NEWLINE] [STARTQ] there is no legal […] argument against allowing incestuous marriage that hasn't been thrown away in pursuit of gay marriage. [ENDQ] [NEWLINE] But in a secular democratic society, such as the one I live in and I'l bet the one that you live in too, the laws are whatever we collectively choose them to be. [NEWLINE] [NEWLINE] In the UK it is illegal for very close relatives (including adoptive parents) to have sex, married or no. The rules about who can and cannot marry [are different, more complicated and change more over time]( [URL] #England_and_Wales).  It's my understanding that the reasons for these rules is on the one hand eugenic and on the other to sanction various kinds of relationship which are very likely to be abusive or exploitative. Gay marriage doesn't seem to have any impact on these reasons.</s>
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> you would rather have a system that instead of providing those who need assistance with assistance would instead go with a shotgun approach of simply giving everybody more money? [NEWLINE] [NEWLINE] <mask> you increase minimum wage to support a single family, people who are single and live at home with their parents would be extremely well off. Instead of that<mask> not simply have a system which provides extra assistance<mask> it's needed? [NEWLINE] [NEWLINE] [STARTQ] It skews the costs of goods [ENDQ] [NEWLINE] Err, no. Having an artificially high minimum wage skews the cost of goods. [NEWLINE] [NEWLINE] [STARTQ] We<mask> a society have made a choice that we will do<mask> we can to ensure our citizens won't starve or be homeless without major malfeasance on their end. [ENDQ] [NEWLINE] And<mask> do you think the purpose of existing aid is for? [NEWLINE] [NEWLINE] [STARTQ] So<mask> a company pushes the wages down to a point<mask> their employee receives tax funded benefits the cost of that good is a lie. [ENDQ] [NEWLINE] It's not like companies all get together and say "let's pay these unskilled workers this artificially low wage!" The rate, in the absence is just the market rate which is determined by the supply and demand of labor. [NEWLINE] </s>
Label encoding: <s>So you would rather have a system that instead of providing those who need assistance with assistance would instead go with a shotgun approach of simply giving everybody more money? [NEWLINE] [NEWLINE] If you increase minimum wage to support a single family, people who are single and live at home with their parents would be extremely well off. Instead of that why not simply have a system which provides extra assistance where it's needed? [NEWLINE] [NEWLINE] [STARTQ] It skews the costs of goods [ENDQ] [NEWLINE] Err, no. Having an artificially high minimum wage skews the cost of goods. [NEWLINE] [NEWLINE] [STARTQ] We as a society have made a choice that we will do what we can to ensure our citizens won't starve or be homeless without major malfeasance on their end. [ENDQ] [NEWLINE] And what do you think the purpose of existing aid is for? [NEWLINE] [NEWLINE] [STARTQ] So when a company pushes the wages down to a point where their employee receives tax funded benefits the cost of that good is a lie. [ENDQ] [NEWLINE] It's not like companies all get together and say "let's pay these unskilled workers this artificially low wage!" The rate, in the absence is just the market rate which is determined by the supply and demand of labor. [NEWLINE] </s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right.  A global scale collapse that leaves enough people left alive to matter will likely not happen. <mask> that's not<mask> all of us are preparing for.  Take me, for example.  I live in Florida near the atlantic ocean, and we get yearly threats of hurricanes, sometimes several.  Last year one of the projected paths for Hurricane Sandy swept right up through the area I live in.  We've had many storms pass us by, sometimes big ones.  And sometimes we get hit by one.  A couple years ago we get slammed by Tropical Storm Fay.  The right weather combinations could lead to a scenario like that,<mask> with a much larger storm.  You may feel differently,<mask><mask><mask> that it's rather dumb for someone in my position *not* to be prepared for such an occurrence.  Even a couple jugs of fresh water could mean a difference between life and death in a hurricane scenario. [NEWLINE] [NEWLINE] TL;DR You're mostly right.  A global collapse will likely never happen. <mask> localized collapses can and do happen and it's wise to be prepared<mask> you live in an area that could be affected.</s>
Label encoding: <s>You're right.  A global scale collapse that leaves enough people left alive to matter will likely not happen.  But that's not what all of us are preparing for.  Take me, for example.  I live in Florida near the atlantic ocean, and we get yearly threats of hurricanes, sometimes several.  Last year one of the projected paths for Hurricane Sandy swept right up through the area I live in.  We've had many storms pass us by, sometimes big ones.  And sometimes we get hit by one.  A couple years ago we get slammed by Tropical Storm Fay.  The right weather combinations could lead to a scenario like that, but with a much larger storm.  You may feel differently, but I think that it's rather dumb for someone in my position *not* to be prepared for such an occurrence.  Even a couple jugs of fresh water could mean a difference between life and death in a hurricane scenario. [NEWLINE] [NEWLINE] TL;DR You're mostly right.  A global collapse will likely never happen.  But localized collapses can and do happen and it's wise to be prepared if you live in an area that could be affected.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Cashing in a 401k early brings with it it a big tax penalty [NEWLINE] [NEWLINE] Anything you buy to store in your home is now a security risk.   Will you<mask> be installing a high quality safe?  And increasing your homeowners insurance<mask> you disclose to them you have tens of thousands of dollars worth of gold bars in your closet?   Maybe add an alarm service?   That can be expensive. [NEWLINE] [NEWLINE] <mask> the financial system really is wiped out to the point<mask> all banks (even FDIC insured ones) fail, that implies the central government has collapsed.  <mask> things are that bad,  you might want to spent some of that money on doomsday prepping: weapons, ammo, food,  a disaster shelter.   (Have you ever watched the Doomsday Preppers show? )   Is that really<mask> youre expecting will happen?   Do you truly believe this is probable to happen in your lifetime? [NEWLINE] [NEWLINE] Most people don't think civilization will end in out lifetime. <mask> you want to put some of your money somewhere it can't vanish in a crash,  just open a savings account in a FDIC insured bank.</s>
Label encoding: <s>Cashing in a 401k early brings with it it a big tax penalty [NEWLINE] [NEWLINE] Anything you buy to store in your home is now a security risk.   Will you also be installing a high quality safe?  And increasing your homeowners insurance when you disclose to them you have tens of thousands of dollars worth of gold bars in your closet?   Maybe add an alarm service?   That can be expensive. [NEWLINE] [NEWLINE] If the financial system really is wiped out to the point where all banks (even FDIC insured ones) fail, that implies the central government has collapsed.   If things are that bad,  you might want to spent some of that money on doomsday prepping: weapons, ammo, food,  a disaster shelter.   (Have you ever watched the Doomsday Preppers show? )   Is that really what youre expecting will happen?   Do you truly believe this is probable to happen in your lifetime? [NEWLINE] [NEWLINE] Most people don't think civilization will end in out lifetime.  If you want to put some of your money somewhere it can't vanish in a crash,  just open a savings account in a FDIC insured bank.</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll use an example I saw recently in this particular debate. Say one of my siblings gets in a car accident and needed some of my blood to live, they would die<mask> they didn't get it, and I'm the only person who can give them this specific type of blood.<mask><mask> the process is quick and painless, no one can force me to give blood to save another human's life. That is<mask>'s called "bodily autonomy", someone's right to control their body is important and cannot be infringed upon. Even corpses have bodily autonomy, we can't take perfectly good organs not being used unless the person (now deceased) had given their consent before their death. Your "abortion is wrong<mask> a fetus is a human being argument, and the woman must carry that baby whether she wants to or not" argument infringes on that woman's bodily autonomy, that woman would have to undergo a difficult, lengthy, and life changing process for 9 months<mask> a fetus is a human being, therefor it would be immoral to kill it. That's asking a pregnant woman to give up more bodily autonomy than a person gives up<mask> they **die**.</s>
Label encoding: <s>I'll use an example I saw recently in this particular debate. Say one of my siblings gets in a car accident and needed some of my blood to live, they would die if they didn't get it, and I'm the only person who can give them this specific type of blood. Even though the process is quick and painless, no one can force me to give blood to save another human's life. That is what's called "bodily autonomy", someone's right to control their body is important and cannot be infringed upon. Even corpses have bodily autonomy, we can't take perfectly good organs not being used unless the person (now deceased) had given their consent before their death. Your "abortion is wrong because a fetus is a human being argument, and the woman must carry that baby whether she wants to or not" argument infringes on that woman's bodily autonomy, that woman would have to undergo a difficult, lengthy, and life changing process for 9 months because a fetus is a human being, therefor it would be immoral to kill it. That's asking a pregnant woman to give up more bodily autonomy than a person gives up when they **die**.</s>
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, I can understand<mask> trying to sustain the environment can seem irrelevant and pointless. [NEWLINE] [NEWLINE] Especially<mask> one expects that he wouldn't see any significant, readily observable benefits in his lifetime. He *personally* has nothing to gain from helping the human race survive longer - whenever it is humans will be going extinct, its usually assumed that we're far the fuck away from it. [NEWLINE] [NEWLINE] <mask>, I wouldn't say it means you can't justify motivation for things. There's a lot to care about,<mask> our inevitable end<mask> a species. You can be motivated towards a career<mask> you'll be able to see the fruits of your efforts. There's people out there being passionate and changing the world n shit. You could be motivated by life threatening circumstances. You could be motivated by your emotions and loved ones.  Things that give you a sense of accomplishment. You can be motivated by social pressure. Validation. Novel concepts. Ingenious ideas. [NEWLINE] [NEWLINE] OP your post kind of reminds me of this quote I saw: [NEWLINE] [NEWLINE] [STARTQ] "The true meaning of life is to plant trees, under whose shade you do not expect to sit." [ENDQ] -Nelson Henderson</s>
Label encoding: <s>Well, I can understand how trying to sustain the environment can seem irrelevant and pointless. [NEWLINE] [NEWLINE] Especially if one expects that he wouldn't see any significant, readily observable benefits in his lifetime. He *personally* has nothing to gain from helping the human race survive longer - whenever it is humans will be going extinct, its usually assumed that we're far the fuck away from it. [NEWLINE] [NEWLINE] However, I wouldn't say it means you can't justify motivation for things. There's a lot to care about, despite our inevitable end as a species. You can be motivated towards a career where you'll be able to see the fruits of your efforts. There's people out there being passionate and changing the world n shit. You could be motivated by life threatening circumstances. You could be motivated by your emotions and loved ones.  Things that give you a sense of accomplishment. You can be motivated by social pressure. Validation. Novel concepts. Ingenious ideas. [NEWLINE] [NEWLINE] OP your post kind of reminds me of this quote I saw: [NEWLINE] [NEWLINE] [STARTQ] "The true meaning of life is to plant trees, under whose shade you do not expect to sit." [ENDQ] -Nelson Henderson</s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Human nature isn't lion nature,<mask>. There are certain patterns that groups of humans tend to follow (just like lions and ants and wolves have their own). They aren't entirely universal or set in stone,<mask> individual humans and cultures can go against them out of necessity, desire, or chance, etc. (human nature and culture are nothing<mask> not flexible),<mask> that doesn't mean they don't generally exist. [NEWLINE] [NEWLINE] Like I doubt there are many cultures<mask> murdering your fellow denizen is acceptable. <mask> that is "universally" immoral, then we could make the claim that in general a human has the natural right *not to be murdered for no reason in his neighborhood*. [NEWLINE] [NEWLINE] <mask><mask> we can use the same reasoning to make some other basic assumptions about<mask> a human is naturally entitled to within his society. <mask>,  I say within his society, <mask> human nature is fundamentally social. Sure,  two passing wanderers in a desert don't owe each other respect or rights, <mask> that isn't the natural setting of human life.  That's an unnatural or at least an unusual setting in which unusual responses should be expected. </s>
Label encoding: <s>Human nature isn't lion nature, however. There are certain patterns that groups of humans tend to follow (just like lions and ants and wolves have their own). They aren't entirely universal or set in stone, as individual humans and cultures can go against them out of necessity, desire, or chance, etc. (human nature and culture are nothing if not flexible), but that doesn't mean they don't generally exist. [NEWLINE] [NEWLINE] Like I doubt there are many cultures where murdering your fellow denizen is acceptable.  If that is "universally" immoral, then we could make the claim that in general a human has the natural right *not to be murdered for no reason in his neighborhood*. [NEWLINE] [NEWLINE] I think we can use the same reasoning to make some other basic assumptions about what a human is naturally entitled to within his society.  Also,  I say within his society,  because human nature is fundamentally social. Sure,  two passing wanderers in a desert don't owe each other respect or rights,  but that isn't the natural setting of human life.  That's an unnatural or at least an unusual setting in which unusual responses should be expected. </s>
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you're a lady and you're pregnant, fucking more guys won't get you more pregnant and spread your genes any more.<mask><mask>, it's in your interest to get pregnant and keep that one male around to share in child-rearing responsibility. [NEWLINE] [NEWLINE] Men<mask><mask><mask><mask> can have sex with a ton of women and theoretically get them all pregnant,<mask> it's to their advantage to attempt to do<mask><mask> they want to pass on their genes. On this flip side of this coin, human women may be unwilling to mate with a male who has demonstrated that he has no intention participating in child rearing,<mask> that is a VERY bad outcome for a pregnant woman. [NEWLINE] [NEWLINE] <mask>,<mask> the social consequences of the male's non-monogamy effectively prevent him from being able to act non-monogamously, the scales tip in favor of him mating with a partner and more or less sticking with her, producing many offspring, and helping to ensure that they survive until reproductive age,<mask> it doesn't mean much<mask> you impregnate 100 women<mask> none of them can keep the offspring alive long enough for them to further propagate your genes.</s>
Label encoding: <s>If you're a lady and you're pregnant, fucking more guys won't get you more pregnant and spread your genes any more. In fact, it's in your interest to get pregnant and keep that one male around to share in child-rearing responsibility. [NEWLINE] [NEWLINE] Men on the other hand can have sex with a ton of women and theoretically get them all pregnant, so it's to their advantage to attempt to do so if they want to pass on their genes. On this flip side of this coin, human women may be unwilling to mate with a male who has demonstrated that he has no intention participating in child rearing, because that is a VERY bad outcome for a pregnant woman. [NEWLINE] [NEWLINE] Thus, if the social consequences of the male's non-monogamy effectively prevent him from being able to act non-monogamously, the scales tip in favor of him mating with a partner and more or less sticking with her, producing many offspring, and helping to ensure that they survive until reproductive age, because it doesn't mean much if you impregnate 100 women but none of them can keep the offspring alive long enough for them to further propagate your genes.</s>
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My point was that not all cats are good at ndoor only cats. [ENDQ] [NEWLINE] This is not a response to my question. You said your cat would be "a million times better off" in circumstances other than<mask> you can provide it.<mask> don't you give your cat to someone who can give it "a million times better" life? [NEWLINE] [NEWLINE] [STARTQ] i meant that quality of life can mean more than risks. [ENDQ] [NEWLINE] This is more anthropomorphism. [NEWLINE] [NEWLINE] [STARTQ] <mask> is wrong with taking them in<mask> you can. [ENDQ] [NEWLINE] It is very odd to say to any cat you see, "hello, I am your new master. You are coming home with me." [NEWLINE] [NEWLINE] [STARTQ] Our cats were fed outside<mask> strays ould eat it to [ENDQ] [NEWLINE] Your house is a nuisance. I am glad nothing like that exists in my neighborhood. [NEWLINE] [NEWLINE] [STARTQ] <mask> one seemed friendly or at least pregnant [ENDQ] [NEWLINE] Oh, good job. We need more stray cats. [NEWLINE] [NEWLINE] [STARTQ] I find that a cat not being miserable to be worth the risk of it getting ran over. [ENDQ] [NEWLINE] Claiming that a cat can feel miserable is anthropomorphism. </s>
Label encoding: <s> [STARTQ] My point was that not all cats are good at ndoor only cats. [ENDQ] [NEWLINE] This is not a response to my question. You said your cat would be "a million times better off" in circumstances other than what you can provide it. Why don't you give your cat to someone who can give it "a million times better" life? [NEWLINE] [NEWLINE] [STARTQ] i meant that quality of life can mean more than risks. [ENDQ] [NEWLINE] This is more anthropomorphism. [NEWLINE] [NEWLINE] [STARTQ] What is wrong with taking them in if you can. [ENDQ] [NEWLINE] It is very odd to say to any cat you see, "hello, I am your new master. You are coming home with me." [NEWLINE] [NEWLINE] [STARTQ] Our cats were fed outside so strays ould eat it to [ENDQ] [NEWLINE] Your house is a nuisance. I am glad nothing like that exists in my neighborhood. [NEWLINE] [NEWLINE] [STARTQ] when one seemed friendly or at least pregnant [ENDQ] [NEWLINE] Oh, good job. We need more stray cats. [NEWLINE] [NEWLINE] [STARTQ] I find that a cat not being miserable to be worth the risk of it getting ran over. [ENDQ] [NEWLINE] Claiming that a cat can feel miserable is anthropomorphism. </s>
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That being said, it is highly unlikely that the aliens would bother expending energy to inhabit a planet<mask> objectively hospitable<mask> Earth without wanting to preserve at least some core integrity of the planet.<mask> it is possible that they wouldn't care about any existing life, it is likely that they would at least make an attempt to preserve the ability to harbor life.<mask> it's not unreasonable to estimate that any massively nuclear or irradiating, climate altering (unless it was a terraforming operation), or water destroying weapon would not be used.<mask> they just used something innocuous like mass-ordinance weapon, it would easily take out population centers,<mask> it might not be suitable for wide swath destruction. In order to completely eradicate all life with impunity, you would probably have to damage the habitability of the planet. Damaging the habitability might not be an issue for them,<mask> it's certainly reasonable to<mask><mask> it would be an issue. [NEWLINE] [NEWLINE] [NEWLINE] We didn't just nuke Iraq into a pile of glass<mask> (obvious PR/ethics/intl relations reasons)<mask><mask><mask> it would make no sense to irradiate the oil fields. </s>
Label encoding: <s>That being said, it is highly unlikely that the aliens would bother expending energy to inhabit a planet as objectively hospitable as Earth without wanting to preserve at least some core integrity of the planet. While it is possible that they wouldn't care about any existing life, it is likely that they would at least make an attempt to preserve the ability to harbor life. So it's not unreasonable to estimate that any massively nuclear or irradiating, climate altering (unless it was a terraforming operation), or water destroying weapon would not be used. If they just used something innocuous like mass-ordinance weapon, it would easily take out population centers, but it might not be suitable for wide swath destruction. In order to completely eradicate all life with impunity, you would probably have to damage the habitability of the planet. Damaging the habitability might not be an issue for them, but it's certainly reasonable to argue that it would be an issue. [NEWLINE] [NEWLINE] [NEWLINE] We didn't just nuke Iraq into a pile of glass because (obvious PR/ethics/intl relations reasons) but also because it would make no sense to irradiate the oil fields. </s>
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a few problems I can see. I will use some of my experience in China to talk about them. [NEWLINE] [NEWLINE] <mask> you movement is just the take over  of factory farms you're not going to win in the court of public opinion. You're going to be people destroying people's businesses. that won't go over<mask> well. You might think you're acting<mask> a liberator,<mask> to others you will be just vandals. [NEWLINE] [NEWLINE] [NEWLINE] <mask> for the Chinese, the issue isn't<mask> much cost it is more of meat's tie to certain dishes that have been made for thousands of years. Meat is<mask> important that China even has a strategic pork reserve. You stop the influx of meat and there will be rioting in the streets. The Chinese starved generations ago. The restriction of food,<mask><mask> a cause you see<mask> noble, won't go over well. [NEWLINE] [NEWLINE] <mask> there were full scale meat restrictions it would be like America in the 1920's. Underground restaurants would pop up overnight. Hundreds of them. It is unrealistic to expect a culture that had had thousands of years of cultural ties with meat dishes to abandon them over night. [NEWLINE] </s>
Label encoding: <s>There are a few problems I can see. I will use some of my experience in China to talk about them. [NEWLINE] [NEWLINE] If you movement is just the take over  of factory farms you're not going to win in the court of public opinion. You're going to be people destroying people's businesses. that won't go over so well. You might think you're acting as a liberator, but to others you will be just vandals. [NEWLINE] [NEWLINE] [NEWLINE] As for the Chinese, the issue isn't so much cost it is more of meat's tie to certain dishes that have been made for thousands of years. Meat is so important that China even has a strategic pork reserve. You stop the influx of meat and there will be rioting in the streets. The Chinese starved generations ago. The restriction of food, regardless of a cause you see as noble, won't go over well. [NEWLINE] [NEWLINE] If there were full scale meat restrictions it would be like America in the 1920's. Underground restaurants would pop up overnight. Hundreds of them. It is unrealistic to expect a culture that had had thousands of years of cultural ties with meat dishes to abandon them over night. [NEWLINE] </s>
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Then<mask>, in the resources that I have provided for you, is it<mask> very predictable<mask> gender roles they will align with based on simply biological testing? [NEWLINE] [NEWLINE] And<mask><mask>, in the resources I have provided you, has it been the case that<mask> children raised the *opposite* to<mask> would be expected based on biological indicators of gender and sex do the overwhelming majority of their own volition come to the realization themselves that they have been raised opposite to their preferred gender? [NEWLINE] [NEWLINE] In these cases of opposite rearing nearly. every. single. environmental/social. indicator. seems to reinforce that these intersex children should be a certain gender. And<mask> they still will deny it. [NEWLINE] [NEWLINE] <mask> do they feel discomfort performing the gender they were raised to perform a vast majority of the time? This is absolutely at odds with your view that the comfort emerges from the societal expectations of the gender.<mask> that were true, these children would conform and enjoy performing the societal expectations of the gender they were re-assigned to, raised<mask>, TOLD they were -<mask> they don't. [NEWLINE] [NEWLINE] <mask><mask> you will like [*this video*]( [URL] ).</s>
Label encoding: <s>Then why, in the resources that I have provided for you, is it so very predictable what gender roles they will align with based on simply biological testing? [NEWLINE] [NEWLINE] And also why, in the resources I have provided you, has it been the case that when children raised the *opposite* to what would be expected based on biological indicators of gender and sex do the overwhelming majority of their own volition come to the realization themselves that they have been raised opposite to their preferred gender? [NEWLINE] [NEWLINE] In these cases of opposite rearing nearly. every. single. environmental/social. indicator. seems to reinforce that these intersex children should be a certain gender. And yet they still will deny it. [NEWLINE] [NEWLINE] Why do they feel discomfort performing the gender they were raised to perform a vast majority of the time? This is absolutely at odds with your view that the comfort emerges from the societal expectations of the gender. If that were true, these children would conform and enjoy performing the societal expectations of the gender they were re-assigned to, raised as, TOLD they were - but they don't. [NEWLINE] [NEWLINE] I think you will like [*this video*]( [URL] ).</s>
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] By leader do you only the president or prime minister should possess this privilege.<mask><mask>, we are dealing with a very small sample size<mask> we have to be willing to accept that there aren't many examples. [ENDQ] [NEWLINE] I would say any elected official acting in an executive capacity.  Not sure<mask> I feel about cabinet ministers in a parliamentary system. [NEWLINE] [NEWLINE] <mask><mask> we have a small sample size to work with for examples. [NEWLINE] [NEWLINE] [STARTQ] Depends on the system in place. In the US the prosecution of presidents would lead to the same party maintaining power.<mask> it wouldn't necessarily have to transfer between adversaries. [ENDQ] [NEWLINE] A one party state is not a democracy.  That's exactly the sort of result I'm worried about. <mask> you can't transition to another party, you don't have democracy.  Power *has* to be able to transfer to an adversary to be meaningful. [NEWLINE] [NEWLINE] [STARTQ] Egypt wasn't exactly stable to begin with. [ENDQ] [NEWLINE] They were stable,<mask> undemocratic.  Their experiment with democracy failed<mask><mask> in part due to the prosecutions of Mubarak and Morsi.  Now they're back to military rule.</s>
Label encoding: <s> [STARTQ] By leader do you only the president or prime minister should possess this privilege. If so, we are dealing with a very small sample size so we have to be willing to accept that there aren't many examples. [ENDQ] [NEWLINE] I would say any elected official acting in an executive capacity.  Not sure how I feel about cabinet ministers in a parliamentary system. [NEWLINE] [NEWLINE] I agree we have a small sample size to work with for examples. [NEWLINE] [NEWLINE] [STARTQ] Depends on the system in place. In the US the prosecution of presidents would lead to the same party maintaining power. So it wouldn't necessarily have to transfer between adversaries. [ENDQ] [NEWLINE] A one party state is not a democracy.  That's exactly the sort of result I'm worried about.  If you can't transition to another party, you don't have democracy.  Power *has* to be able to transfer to an adversary to be meaningful. [NEWLINE] [NEWLINE] [STARTQ] Egypt wasn't exactly stable to begin with. [ENDQ] [NEWLINE] They were stable, but undemocratic.  Their experiment with democracy failed I think in part due to the prosecutions of Mubarak and Morsi.  Now they're back to military rule.</s>
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you can honestly prove me that everything in the realm of general education is crucial enough that it can't be cut for the sake of me learning<mask> to survive outside of school, then my view will be changed. [ENDQ] [NEWLINE] That can't really be done,<mask>,<mask> it's entirely subjective.  Maybe you don't think it's important to have music classes,<mask><mask><mask> it's critical. [NEWLINE] [NEWLINE] The fact is, you'd have to sacrifice multiple general education classes across a few years in order to fit in these life skills, many of which would certainly be rushed through.  The most oft-requested class is personal finance.  To properly cover it, you'd need at least 2 semesters of it to cover budgeting, taxes, 401(k). Which classes get the ax, especially<mask> literally every kid has to take it? [NEWLINE] [NEWLINE] The way I've seen it is this: the school is responsible for teaching me about a multitude of subjects (math, science, english, history)<mask> that I can decide<mask> I'm interested in for college, and my parents are responsible for teaching me<mask> to function in society.</s>
Label encoding: <s> [STARTQ] If you can honestly prove me that everything in the realm of general education is crucial enough that it can't be cut for the sake of me learning how to survive outside of school, then my view will be changed. [ENDQ] [NEWLINE] That can't really be done, though, because it's entirely subjective.  Maybe you don't think it's important to have music classes, but I think it's critical. [NEWLINE] [NEWLINE] The fact is, you'd have to sacrifice multiple general education classes across a few years in order to fit in these life skills, many of which would certainly be rushed through.  The most oft-requested class is personal finance.  To properly cover it, you'd need at least 2 semesters of it to cover budgeting, taxes, 401(k). Which classes get the ax, especially since literally every kid has to take it? [NEWLINE] [NEWLINE] The way I've seen it is this: the school is responsible for teaching me about a multitude of subjects (math, science, english, history) so that I can decide what I'm interested in for college, and my parents are responsible for teaching me how to function in society.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is a common fallacy for those who have grown up in a society<mask> people respect each others rights and property to assume this is somehow universal. That humans have evolved from the brutes of ages past, learned to be civilized, and altered the human condition to reflect prosperity through cooperation and peace. [NEWLINE] [NEWLINE] The truth is that this kind of society only exists under the constant threat of force. A certain segment of the population will always take<mask> they want<mask> they can. Societies are endless fountains of cooperation<mask> much<mask> self sustaining threats. "Do<mask> we consider right or we will kill you and your family." [NEWLINE] [NEWLINE] That is<mask> the US Military ultimately provides. The force behind that threat, the only thing that makes a proclamation like "people have the right to trade things they own" into a reality. [NEWLINE] [NEWLINE] Take away that threat, take away the US Navy, and those who want to take will take until someone else lifts the sword to stop them. [NEWLINE] [NEWLINE] The question then becomes: who is more worthy to wield the sword that hangs over civilization?  Is there a more trustworthy organization than the US? Who?<mask>? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>It is a common fallacy for those who have grown up in a society where people respect each others rights and property to assume this is somehow universal. That humans have evolved from the brutes of ages past, learned to be civilized, and altered the human condition to reflect prosperity through cooperation and peace. [NEWLINE] [NEWLINE] The truth is that this kind of society only exists under the constant threat of force. A certain segment of the population will always take what they want when they can. Societies are endless fountains of cooperation as much as self sustaining threats. "Do what we consider right or we will kill you and your family." [NEWLINE] [NEWLINE] That is what the US Military ultimately provides. The force behind that threat, the only thing that makes a proclamation like "people have the right to trade things they own" into a reality. [NEWLINE] [NEWLINE] Take away that threat, take away the US Navy, and those who want to take will take until someone else lifts the sword to stop them. [NEWLINE] [NEWLINE] The question then becomes: who is more worthy to wield the sword that hangs over civilization?  Is there a more trustworthy organization than the US? Who? Why? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What about his life improves with an extra 14.85M? Perhaps his dream was to own a private island. Perhaps he has 14 kids he wants to put through Harvard. It's not particularly useful for us to muse about other people's circumstances, or motivations. Nor is it really our business. There is no objective standard. [NEWLINE] [NEWLINE] It's<mask> not a zero-sum game. The scientist who cures cancer hasn't stolen research potential from anyone else; he has expanded the pool of human knowledge. Similarly, economic contributions do not steal growth from other places; they expand the overall pot. Granted, there will be times<mask> individuals prosper at the expense of others,<mask> on aggregate, economic activity and technological advance make us all wealthier - there is more to go around.<mask> high concentrations of wealth, the global poverty level halved in the last 20 years. [NEWLINE] [NEWLINE] EDIT: I'd<mask> dispute the "getting lucky" comment, and respectfully suggest that is indicates an inherent bias in your argument - that you are unwilling to comprehend that someone can work hard and create far more value than the average person (certainly outstripping your arbitrary salary cap.)</s>
Label encoding: <s>What about his life improves with an extra 14.85M? Perhaps his dream was to own a private island. Perhaps he has 14 kids he wants to put through Harvard. It's not particularly useful for us to muse about other people's circumstances, or motivations. Nor is it really our business. There is no objective standard. [NEWLINE] [NEWLINE] It's also not a zero-sum game. The scientist who cures cancer hasn't stolen research potential from anyone else; he has expanded the pool of human knowledge. Similarly, economic contributions do not steal growth from other places; they expand the overall pot. Granted, there will be times where individuals prosper at the expense of others, but on aggregate, economic activity and technological advance make us all wealthier - there is more to go around. Despite high concentrations of wealth, the global poverty level halved in the last 20 years. [NEWLINE] [NEWLINE] EDIT: I'd also dispute the "getting lucky" comment, and respectfully suggest that is indicates an inherent bias in your argument - that you are unwilling to comprehend that someone can work hard and create far more value than the average person (certainly outstripping your arbitrary salary cap.)</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>(Disclaimer:  I'm not a conservative.) [NEWLINE] [NEWLINE] I don't believe right wing views are inherently selfish.  The best description I've heard of left vs. right approaches to social programs is<mask> follows: [NEWLINE] [NEWLINE] * People on the left are afraid that someone, somewhere, is not getting help they *need*. [NEWLINE] * People on the right are afraid that someone, somewhere, is getting help they don't *deserve*. [NEWLINE] [NEWLINE] Generally we compromise by having assistance programs which are means-tested,<mask> that waste is kept to a minimum. [NEWLINE] [NEWLINE] The conservative approach is not inherently selfish - it's about eliminating wasteful spending<mask> that those resources can be applied more productively (either publicly or privately). [NEWLINE] [NEWLINE] I don't know about specifics of UK politics,<mask> in the US there's<mask> an element of tribalism that enters into the question of who deserves to be helped.  For a lot of conservatives (not all conservatives,<mask> enough that politicians can appeal to the sentiment), people who don't conform to the norms (including norms of race, gender and class) are perceived<mask> being less deserving of assistance than those who do.</s>
Label encoding: <s>(Disclaimer:  I'm not a conservative.) [NEWLINE] [NEWLINE] I don't believe right wing views are inherently selfish.  The best description I've heard of left vs. right approaches to social programs is as follows: [NEWLINE] [NEWLINE] * People on the left are afraid that someone, somewhere, is not getting help they *need*. [NEWLINE] * People on the right are afraid that someone, somewhere, is getting help they don't *deserve*. [NEWLINE] [NEWLINE] Generally we compromise by having assistance programs which are means-tested, so that waste is kept to a minimum. [NEWLINE] [NEWLINE] The conservative approach is not inherently selfish - it's about eliminating wasteful spending so that those resources can be applied more productively (either publicly or privately). [NEWLINE] [NEWLINE] I don't know about specifics of UK politics, but in the US there's also an element of tribalism that enters into the question of who deserves to be helped.  For a lot of conservatives (not all conservatives, but enough that politicians can appeal to the sentiment), people who don't conform to the norms (including norms of race, gender and class) are perceived as being less deserving of assistance than those who do.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The fact is Comcast can use their monopolistic and lobbying powers to stop you from starting an ISP and competing. They're doing it all the time.<mask><mask> people sometimes forget that the reason we have capitalism is not<mask> it is inherently good,<mask><mask> it results in optimal outcomes for society. People's self interest drives a superior allocation of assets<mask> compared to other models, and theoretically leads to the most efficient outcome.<mask> the outcome is not efficient, this is a market failure.<mask> companies abuse monopolistic power to drive out competitors, this does not lead to the most efficient outcome. In this situation,<mask> the outcome is clearly not efficient and a market failure exists, government needs to intervene and correct the failure. [NEWLINE] [NEWLINE] <mask><mask> my main point is that capitalism is good<mask> it is the best means to an end (efficient allocation of resources), NOT<mask> it is *inherently* good. We need to keep this in mind<mask> arguing for pure unrestrained capitalism. The system is good, and works well for the most part,<mask> it isn't perfect, and controlling elements of it is a perfectly rational thing to do. [NEWLINE] </s>
Label encoding: <s>The fact is Comcast can use their monopolistic and lobbying powers to stop you from starting an ISP and competing. They're doing it all the time. I think people sometimes forget that the reason we have capitalism is not because it is inherently good, but because it results in optimal outcomes for society. People's self interest drives a superior allocation of assets as compared to other models, and theoretically leads to the most efficient outcome. Where the outcome is not efficient, this is a market failure. When companies abuse monopolistic power to drive out competitors, this does not lead to the most efficient outcome. In this situation, where the outcome is clearly not efficient and a market failure exists, government needs to intervene and correct the failure. [NEWLINE] [NEWLINE] I think my main point is that capitalism is good because it is the best means to an end (efficient allocation of resources), NOT because it is *inherently* good. We need to keep this in mind when arguing for pure unrestrained capitalism. The system is good, and works well for the most part, but it isn't perfect, and controlling elements of it is a perfectly rational thing to do. [NEWLINE] </s>
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Vegs are not making a moral choice by choosing plant life over animal life, they are merely choosing the least anthropomorphic life available. [ENDQ] [NEWLINE] I can really only speak for myself,<mask><mask><mask> that many vegans/vegetarians aren't only concerned with death,<mask> with the exploitation and suffering inherent in eating animal products. Plants don't "suffer" by any reasonable definition,<mask> they don't have central nervous systems. [NEWLINE] [NEWLINE] <mask> *even<mask> they did*, I hope you realize that the animals that are eaten have eat something themselves. It varies a little bit from animal to animal,<mask> it takes about ten pounds of plants material to produce one pound of animal material. [NEWLINE] [NEWLINE] <mask>, just for the sake of discussion, let's say that it's equally wrong to "kill" a plant and an animal. By eating an animal, you're responsible for the deaths of all the plants it took to raise the animal until its death, plus the death of the animal itself.<mask> even<mask> you were right, vegans and vegetarians still have the ethical position. [NEWLINE] [NEWLINE] (edit: fixed typo)</s>
Label encoding: <s> [STARTQ] Vegs are not making a moral choice by choosing plant life over animal life, they are merely choosing the least anthropomorphic life available. [ENDQ] [NEWLINE] I can really only speak for myself, but I think that many vegans/vegetarians aren't only concerned with death, but with the exploitation and suffering inherent in eating animal products. Plants don't "suffer" by any reasonable definition, since they don't have central nervous systems. [NEWLINE] [NEWLINE] But *even if they did*, I hope you realize that the animals that are eaten have eat something themselves. It varies a little bit from animal to animal, but it takes about ten pounds of plants material to produce one pound of animal material. [NEWLINE] [NEWLINE] So, just for the sake of discussion, let's say that it's equally wrong to "kill" a plant and an animal. By eating an animal, you're responsible for the deaths of all the plants it took to raise the animal until its death, plus the death of the animal itself. So even if you were right, vegans and vegetarians still have the ethical position. [NEWLINE] [NEWLINE] (edit: fixed typo)</s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] singles out racism and sexism [ENDQ] [NEWLINE] <mask> they don't single out racism and sexism.  They indiscriminately hate anything that doesn't conform to radical feminist ideology -- on which you appear to be heavily misinformed.  Many SRSers themselves are *extremely* racist and sexist. [NEWLINE] [NEWLINE] For example, [NEWLINE] [NEWLINE] [One.]( [URL] ) [NEWLINE] [NEWLINE] [Two.]( [URL].png) [NEWLINE] [NEWLINE] [Three.]( [URL] ) [NEWLINE] [NEWLINE] [Four.]( [URL].jpg) [NEWLINE] [NEWLINE] And that's only the tip of the iceberg. [NEWLINE] [NEWLINE] More importantly,<mask>, the reason vote brigades are prohibited (except in the case of bestof) is that scoring in the reddit community is supposed to reflect the enjoyment of the majority of its viewers -- of which the "normal" voters are a random sample.  SRS,<mask>, is not a random sample -- it's a collection of users with a single ideology that floods certain posts on command.   They skew the distribution from "<mask> the 'average' redditor likes/finds funny/whatever" to "<mask> the average SRSer likes."</s>
Label encoding: <s> [STARTQ] singles out racism and sexism [ENDQ] [NEWLINE] But they don't single out racism and sexism.  They indiscriminately hate anything that doesn't conform to radical feminist ideology -- on which you appear to be heavily misinformed.  Many SRSers themselves are *extremely* racist and sexist. [NEWLINE] [NEWLINE] For example, [NEWLINE] [NEWLINE] [One.]( [URL] ) [NEWLINE] [NEWLINE] [Two.]( [URL].png) [NEWLINE] [NEWLINE] [Three.]( [URL] ) [NEWLINE] [NEWLINE] [Four.]( [URL].jpg) [NEWLINE] [NEWLINE] And that's only the tip of the iceberg. [NEWLINE] [NEWLINE] More importantly, however, the reason vote brigades are prohibited (except in the case of bestof) is that scoring in the reddit community is supposed to reflect the enjoyment of the majority of its viewers -- of which the "normal" voters are a random sample.  SRS, however, is not a random sample -- it's a collection of users with a single ideology that floods certain posts on command.   They skew the distribution from " what the 'average' redditor likes/finds funny/whatever" to " what the average SRSer likes."</s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What about /r/aww? Puppies and kittens (and a bunch of other animals) are adorable<mask><mask> the rest of society.<mask> about /r/conservative and /r/liberal that are both reflective of groups in the real world.<mask> about /r/funny? Most people like a good laugh and I don't think the humor there is largely composed of things the rest of the world wouldn't find humorous.<mask><mask> it is easy to cherry-pick a couple subreddits and say that they are just being contrarian for the hell of it,<mask> that doesn't mean all of reddit is counter-cultural. [NEWLINE] [NEWLINE] <mask>,<mask> is the problem with having debates and discussions? There are plenty of good productive discussions that happen on reddit too. There are<mask> plenty of circlejerking and ideological wars on pretty much every internet thread known to man.<mask> should reddit be any different. More importantly,<mask> does it mean to be too counter-cultural for its own good?  Is there something else reddit is supposed to be that is better? Will it destroy itself or is it unsustainable<mask> of the content?</s>
Label encoding: <s>What about /r/aww? Puppies and kittens (and a bunch of other animals) are adorable according to the rest of society. What about /r/conservative and /r/liberal that are both reflective of groups in the real world. What about /r/funny? Most people like a good laugh and I don't think the humor there is largely composed of things the rest of the world wouldn't find humorous. I think it is easy to cherry-pick a couple subreddits and say that they are just being contrarian for the hell of it, but that doesn't mean all of reddit is counter-cultural. [NEWLINE] [NEWLINE] Also, what is the problem with having debates and discussions? There are plenty of good productive discussions that happen on reddit too. There are also plenty of circlejerking and ideological wars on pretty much every internet thread known to man. Why should reddit be any different. More importantly, what does it mean to be too counter-cultural for its own good?  Is there something else reddit is supposed to be that is better? Will it destroy itself or is it unsustainable because of the content?</s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You haven't done anything to answer my question of "who decides<mask> the'science' is?" <mask> we need a special committee to decide<mask> the "expert opinion" is,<mask> is this different from electing policy-makers in the first place? [NEWLINE] [NEWLINE] [STARTQ] It's controlled by the chairman of the Fed who is not elected. [ENDQ] [NEWLINE] Not really.  Sure, I'm not voting for a presidential candidate<mask> of who he'll appoint to the Fed board. <mask> I vote based on things like experience dealing with economic issues, education, and things like that.  The point isn't to ensure that your choice of Fed chairmen gets appointed.  The point is to elect capable leaders who can make an informed decision in appointing the Fed chairman (appointing the Fed chairman here is a proxy for all economic issues). [NEWLINE] [NEWLINE] And I don't mean to limit decision to the economy.  There are plenty of areas<mask> I am not educated.  I don't know anything about international diplomacy or national defense policy, either.  Should the general populace with no military training be voting on defense protocols?</s>
Label encoding: <s>You haven't done anything to answer my question of "who decides what the'science' is?"  If we need a special committee to decide what the "expert opinion" is, how is this different from electing policy-makers in the first place? [NEWLINE] [NEWLINE] [STARTQ] It's controlled by the chairman of the Fed who is not elected. [ENDQ] [NEWLINE] Not really.  Sure, I'm not voting for a presidential candidate because of who he'll appoint to the Fed board.  But I vote based on things like experience dealing with economic issues, education, and things like that.  The point isn't to ensure that your choice of Fed chairmen gets appointed.  The point is to elect capable leaders who can make an informed decision in appointing the Fed chairman (appointing the Fed chairman here is a proxy for all economic issues). [NEWLINE] [NEWLINE] And I don't mean to limit decision to the economy.  There are plenty of areas where I am not educated.  I don't know anything about international diplomacy or national defense policy, either.  Should the general populace with no military training be voting on defense protocols?</s>
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Knowing that the child will face hardship is not an argument to not have kids. The child will<mask> experience much joy. [ENDQ] [NEWLINE] I responded to this already<mask> follows: [NEWLINE] [NEWLINE] [STARTQ] The presence of pain is bad. [ENDQ] [NEWLINE] [STARTQ] The presence of pleasure is good. [ENDQ] [NEWLINE] [STARTQ] <mask> far, pleasure and pain are symmetrical in their goodness and badness.<mask> they are not symmetrical with respect to their absence. More specifically: [ENDQ] [NEWLINE] [STARTQ] The absence of pain is good, even<mask> that good is not enjoyed by anyone,<mask> [ENDQ] [NEWLINE] [STARTQ] The absence of pleasure is not bad unless there is somebody (an actual somebody) who is deprived by its absence. [ENDQ] [NEWLINE] Put simply, you are not around to lament your lack of pleasure, it is not immoral for that pleasure to be denied - you don't exist in the first place to be sad about it.<mask><mask> you are brought into the world and you face suffering, that is certainly an immoral thing to be subjected to. [NEWLINE] [NEWLINE] [STARTQ] I simply disagree about the level of destruction bringing a child into the world causes to the environment. [ENDQ] [NEWLINE] [URL] </s>
Label encoding: <s> [STARTQ] Knowing that the child will face hardship is not an argument to not have kids. The child will also experience much joy. [ENDQ] [NEWLINE] I responded to this already as follows: [NEWLINE] [NEWLINE] [STARTQ] The presence of pain is bad. [ENDQ] [NEWLINE] [STARTQ] The presence of pleasure is good. [ENDQ] [NEWLINE] [STARTQ] So far, pleasure and pain are symmetrical in their goodness and badness. But they are not symmetrical with respect to their absence. More specifically: [ENDQ] [NEWLINE] [STARTQ] The absence of pain is good, even if that good is not enjoyed by anyone, but [ENDQ] [NEWLINE] [STARTQ] The absence of pleasure is not bad unless there is somebody (an actual somebody) who is deprived by its absence. [ENDQ] [NEWLINE] Put simply, you are not around to lament your lack of pleasure, it is not immoral for that pleasure to be denied - you don't exist in the first place to be sad about it. But if you are brought into the world and you face suffering, that is certainly an immoral thing to be subjected to. [NEWLINE] [NEWLINE] [STARTQ] I simply disagree about the level of destruction bringing a child into the world causes to the environment. [ENDQ] [NEWLINE] [URL] </s>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] College isn't an investment.<mask> it is, you are dealing with Bernie Madoff and his minuscule dick taking your money. Don't go into college expecting to recoup your investment,<mask> your ROI would be shit, and the date of return falls somewhere after your fiftieth birthday. [ENDQ] [NEWLINE]??????? [NEWLINE] With the cost rising every year, it most certainly is an investment, even from a time standpoint. You could spend 4-5 years at college, spending $50,000 for example, or you could spend that time working and making more or less that amount with a high school diploma. <mask>, there's a greater probability that with that college degree, you could earn more over the course of your lifetime. <mask>, it's an investment<mask> you could either earn more with it, or less. [NEWLINE] [NEWLINE] [STARTQ] <mask> the primary purpose of learning isn't to make money. It's to learn. [ENDQ] [NEWLINE] That depends on who you ask, actually. For me, college is means to an end, financially. <mask> I learn is great,<mask> not my primary goal.</s>
Label encoding: <s> [STARTQ] College isn't an investment. If it is, you are dealing with Bernie Madoff and his minuscule dick taking your money. Don't go into college expecting to recoup your investment, because your ROI would be shit, and the date of return falls somewhere after your fiftieth birthday. [ENDQ] [NEWLINE]??????? [NEWLINE] With the cost rising every year, it most certainly is an investment, even from a time standpoint. You could spend 4-5 years at college, spending $50,000 for example, or you could spend that time working and making more or less that amount with a high school diploma.  However, there's a greater probability that with that college degree, you could earn more over the course of your lifetime.  Hence, it's an investment where you could either earn more with it, or less. [NEWLINE] [NEWLINE] [STARTQ] but the primary purpose of learning isn't to make money. It's to learn. [ENDQ] [NEWLINE] That depends on who you ask, actually. For me, college is means to an end, financially.  What I learn is great, but not my primary goal.</s>
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, unless your friend is intellectually disabled, you wouldn't be hurting anyone. Go ahead and say<mask> you will. [NEWLINE] [NEWLINE] I don't think I know of anyone with special needs (part of the campaign) that's trying to go into your room and tell you to knock it off. That's just unrealistic. [NEWLINE] [NEWLINE] Would you say something "is retarded" in the presence of a member of that population? You probably wouldn't,<mask> many people would, whether they mean to or not; it just slips out sometimes. That is more negotiable, and something the campaign tries to change. [NEWLINE] [NEWLINE] <mask> for ending all insults, that's a great goal, and something we should all strive towards<mask> a community and individually. It's been a long time<mask> I've heard an insult used with malicious intent from a person with special needs,<mask>.<mask><mask> they are suspecting us neurotypical people to campaign on the larger picture? :) [NEWLINE] [NEWLINE] I'm out of the house for a bit,<mask> I would be interested in continuing this train of thought<mask> you'll give me some downtime! </s>
Label encoding: <s>Well, unless your friend is intellectually disabled, you wouldn't be hurting anyone. Go ahead and say what you will. [NEWLINE] [NEWLINE] I don't think I know of anyone with special needs (part of the campaign) that's trying to go into your room and tell you to knock it off. That's just unrealistic. [NEWLINE] [NEWLINE] Would you say something "is retarded" in the presence of a member of that population? You probably wouldn't, but many people would, whether they mean to or not; it just slips out sometimes. That is more negotiable, and something the campaign tries to change. [NEWLINE] [NEWLINE] As for ending all insults, that's a great goal, and something we should all strive towards as a community and individually. It's been a long time since I've heard an insult used with malicious intent from a person with special needs, though. I think they are suspecting us neurotypical people to campaign on the larger picture? :) [NEWLINE] [NEWLINE] I'm out of the house for a bit, but I would be interested in continuing this train of thought if you'll give me some downtime! </s>
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Reread<mask> I said mate, I said "doesn't make it right<mask> ". I said in no way that I condone it, I don't want the death penalty, I don't want the government to be able to kill its citizens. Please read it more closely next time. [NEWLINE] [NEWLINE] <mask> now to be sure I'm just gonna repeat myself. [NEWLINE] [NEWLINE] People become really irrational<mask> hurt,<mask><mask> you start putting them in a position of power (ie<mask> they have a big say in a court case about a murder of their relative/friend) they will choose the vindictive and vengeful path. They will want them killed<mask> that is<mask> they did to their loved one. [NEWLINE] [NEWLINE] <mask> you ask me people who are for the death penalty are more or less hypocrites, or just don't fear a government with a lot of power.<mask> the government get too much power, all it takes is one crazy person with good speaker skills to wreck havoc across the entire country. [NEWLINE] [NEWLINE] Now I have made<mask><mask><mask> clear<mask> possible.<mask> you misunderstand it again it is your incompetence. </s>
Label encoding: <s>Reread what I said mate, I said "doesn't make it right though ". I said in no way that I condone it, I don't want the death penalty, I don't want the government to be able to kill its citizens. Please read it more closely next time. [NEWLINE] [NEWLINE] So now to be sure I'm just gonna repeat myself. [NEWLINE] [NEWLINE] People become really irrational when hurt, so if you start putting them in a position of power (ie when they have a big say in a court case about a murder of their relative/friend) they will choose the vindictive and vengeful path. They will want them killed because that is what they did to their loved one. [NEWLINE] [NEWLINE] If you ask me people who are for the death penalty are more or less hypocrites, or just don't fear a government with a lot of power. When the government get too much power, all it takes is one crazy person with good speaker skills to wreck havoc across the entire country. [NEWLINE] [NEWLINE] Now I have made my opinion as clear as possible. If you misunderstand it again it is your incompetence. </s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is to some extent,<mask> certainly the options<mask> you're poor all tend to suck. <mask> there are certainly people who start out small businesses from just some skill and a lot of work. [NEWLINE] [NEWLINE] My viewpoint is that reducing the scope of government intervention and regulation, e.g. getting rid of licensure requirements for [tour guides]( [URL].shtml) and [interior designers]( [URL] ) will often greatly help low income people. [NEWLINE] [NEWLINE] <mask>, I'd want to see the welfare state transformed into a more cohesive policy such<mask> a [guaranteed basic income]( [URL] ) and eliminating most need-based programs, which have terrible incentive effects with respect to employment.  Especially<mask> you put multiple programs together, means tested benefits create incredibly high [implicit marginal tax rates]( [URL] /) which mean that increases in income are often entirely or nearly entirely offset by reduced benefits. [NEWLINE] [NEWLINE] I'm not an "abolish all taxes" sort of guy<mask>, I want to see changes within the existing structure of democratic states to increase people's range of choices and increase the total economic output.</s>
Label encoding: <s>It is to some extent, though certainly the options when you're poor all tend to suck.  But there are certainly people who start out small businesses from just some skill and a lot of work. [NEWLINE] [NEWLINE] My viewpoint is that reducing the scope of government intervention and regulation, e.g. getting rid of licensure requirements for [tour guides]( [URL].shtml) and [interior designers]( [URL] ) will often greatly help low income people. [NEWLINE] [NEWLINE] Additionally, I'd want to see the welfare state transformed into a more cohesive policy such as a [guaranteed basic income]( [URL] ) and eliminating most need-based programs, which have terrible incentive effects with respect to employment.  Especially when you put multiple programs together, means tested benefits create incredibly high [implicit marginal tax rates]( [URL] /) which mean that increases in income are often entirely or nearly entirely offset by reduced benefits. [NEWLINE] [NEWLINE] I'm not an "abolish all taxes" sort of guy though, I want to see changes within the existing structure of democratic states to increase people's range of choices and increase the total economic output.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you entertain the possibility of having at some point children?<mask><mask>, don't you want to make it survivable for them? [NEWLINE] [NEWLINE] No offense,<mask> do you care about anything at all,<mask> yourself? Do you have/care about your, friends?<mask> some of them have children, do you care about their ability to survive? (<mask> a due to your friends.) [NEWLINE] [NEWLINE] [STARTQ] The human race is going extinct, it's just a matter of time [ENDQ] [NEWLINE] <mask> do you know that? It may not happen...<mask>,<mask> you knew you were the tipping point of "humanity survives" vs "humanity perishes",<mask> would you choose? [NEWLINE] [NEWLINE] Finally, to make an analog in a much smaller scale.<mask> do you think of this: "You (personally) are going to eventually die. It's just a matter of time.<mask>, now that I see you are in peril of drowning in a big, turbulent river, I choose not to throw you the rope I have and save you. I leave you to drown. I don't care."? [NEWLINE] </s><pad>
Label encoding: <s>Do you entertain the possibility of having at some point children? If so, don't you want to make it survivable for them? [NEWLINE] [NEWLINE] No offense, but do you care about anything at all, besides yourself? Do you have/care about your, friends? If some of them have children, do you care about their ability to survive? ( As a due to your friends.) [NEWLINE] [NEWLINE] [STARTQ] The human race is going extinct, it's just a matter of time [ENDQ] [NEWLINE] How do you know that? It may not happen... Also, if you knew you were the tipping point of "humanity survives" vs "humanity perishes", what would you choose? [NEWLINE] [NEWLINE] Finally, to make an analog in a much smaller scale. What do you think of this: "You (personally) are going to eventually die. It's just a matter of time. So, now that I see you are in peril of drowning in a big, turbulent river, I choose not to throw you the rope I have and save you. I leave you to drown. I don't care."? [NEWLINE] </s><pad>
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Spills happen.  The restaurant has cloths and spray bottles of cleaning chemicals.  We have a used paper napkin and a little water (unless our glasses didn't get refilled). [NEWLINE] [NEWLINE] Is "nine out of ten" literally<mask> many do something truly rant-worthy?  Or is that<mask> many do some violation of your standards, no matter<mask> slight, with only one in ten being utterly heinous? [NEWLINE] [NEWLINE] <mask> you were a child, you had a child's perception of<mask> well you were behaving, and of<mask> often you got reprimanded. <mask> you were six, it seemed perfectly normal to you that you were<mask> coordinated<mask> a six-year-old. <mask> you got reprimanded quietly<mask> you spilled something, the server probably didn't hear it,<mask> you remembered the injustice of getting punished for an innocent mistake. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, maybe you worked at a restaurant<mask> the clientele included a lot of parents with very different ideas about child-rearing than your parents had. <mask><mask>, I don't think it's representative of parents everywhere.</s>
Label encoding: <s>Spills happen.  The restaurant has cloths and spray bottles of cleaning chemicals.  We have a used paper napkin and a little water (unless our glasses didn't get refilled). [NEWLINE] [NEWLINE] Is "nine out of ten" literally how many do something truly rant-worthy?  Or is that how many do some violation of your standards, no matter how slight, with only one in ten being utterly heinous? [NEWLINE] [NEWLINE] When you were a child, you had a child's perception of how well you were behaving, and of how often you got reprimanded.  When you were six, it seemed perfectly normal to you that you were as coordinated as a six-year-old.  If you got reprimanded quietly when you spilled something, the server probably didn't hear it, but you remembered the injustice of getting punished for an innocent mistake. [NEWLINE] [NEWLINE] On the other hand, maybe you worked at a restaurant where the clientele included a lot of parents with very different ideas about child-rearing than your parents had.  If so, I don't think it's representative of parents everywhere.</s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You're on "Change my view". Forcing me to make my claim less broad is certainly a step in the right direction. [ENDQ] [NEWLINE] Right,<mask><mask> the OP says that challenging one of four listed assumptions will change their view, it's natural for the replies to focus on that.<mask> one of those assumptions is clearly challenged and the OP responds with "well, that's not<mask> I meant, I really meant something else" the argument is impossible to pin down. [NEWLINE] [NEWLINE] [STARTQ] Does an alien that's smarter than us have the ability to strip our rights away? [ENDQ] [NEWLINE] Ability? Sure. [NEWLINE] [NEWLINE] [STARTQ] Should Einstein have the ability to strip your rights away? [ENDQ] [NEWLINE] I mean, he's dead<mask> I could envision a scenario<mask> Albert Einstein had the ability to imprison me for a crime I did not convict based on clearly manipulated evidence. Should people be put in power that they could abuse?<mask><mask> that is necessary and society needs to have persons in positions of power and simultaneously guard against abuses of that power.<mask> in the seven hells does this have to do with consuming animal products? [NEWLINE] </s>
Label encoding: <s> [STARTQ] You're on "Change my view". Forcing me to make my claim less broad is certainly a step in the right direction. [ENDQ] [NEWLINE] Right, so when the OP says that challenging one of four listed assumptions will change their view, it's natural for the replies to focus on that. When one of those assumptions is clearly challenged and the OP responds with "well, that's not what I meant, I really meant something else" the argument is impossible to pin down. [NEWLINE] [NEWLINE] [STARTQ] Does an alien that's smarter than us have the ability to strip our rights away? [ENDQ] [NEWLINE] Ability? Sure. [NEWLINE] [NEWLINE] [STARTQ] Should Einstein have the ability to strip your rights away? [ENDQ] [NEWLINE] I mean, he's dead but I could envision a scenario where Albert Einstein had the ability to imprison me for a crime I did not convict based on clearly manipulated evidence. Should people be put in power that they could abuse? I think that is necessary and society needs to have persons in positions of power and simultaneously guard against abuses of that power. What in the seven hells does this have to do with consuming animal products? [NEWLINE] </s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Should young adults with trust funds not be allowed to vote<mask> they don't know<mask> it is like to live pay check to pay check?<mask> about terminally ill adults? They wont experience long term consequences.<mask> about the unemployed? They don't pay income tax<mask> they don't contribute (for the sake of argument). Can we just start taking away the voting rights of everyone who has a different political ideology than you (I assume this is<mask> you are really getting at, "can the old people just die<mask> we can have some progress already!?!?")? [NEWLINE] [NEWLINE] <mask>,<mask> old people can't vote than who will represent their interests? You are making the argument that<mask> they aren't living through<mask> younger generations are living through then they can't make informed voting decisions regarding those people. Well<mask> can you expect young people who haven't experienced being old to make informed voting decisions regarding retirement? You could say that young people will vote in interests of creating the best retirements<mask> they will some day retire<mask> do you really expect people to prioritizes decades in the future over the short term?</s><pad>
Label encoding: <s>Should young adults with trust funds not be allowed to vote because they don't know what it is like to live pay check to pay check? What about terminally ill adults? They wont experience long term consequences. What about the unemployed? They don't pay income tax so they don't contribute (for the sake of argument). Can we just start taking away the voting rights of everyone who has a different political ideology than you (I assume this is what you are really getting at, "can the old people just die so we can have some progress already!?!?")? [NEWLINE] [NEWLINE] Also, if old people can't vote than who will represent their interests? You are making the argument that because they aren't living through what younger generations are living through then they can't make informed voting decisions regarding those people. Well how can you expect young people who haven't experienced being old to make informed voting decisions regarding retirement? You could say that young people will vote in interests of creating the best retirements because they will some day retire but do you really expect people to prioritizes decades in the future over the short term?</s><pad>
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, I don't think anyone "owns" your idea, not even you. [ENDQ] [STARTQ] [ENDQ] [NEWLINE] <mask> not?<mask> I own the computer I'm typing this on, and I sell it to my brother,<mask> leave the computer<mask> it is, nothing has changed about the computer. It's still in the same spot, physically identical. The difference is not tangible. Ownership is an idea. Possession of something isn't ownership, nor is usage of something. The concept of ownership exists<mask> it is useful. We apply that concept to both physical things and ideas. [NEWLINE] [NEWLINE] [STARTQ] [URL] [ENDQ] [NEWLINE] This is an argument that we need to reform patent laws, not do away with them completely. [NEWLINE] [NEWLINE] Patent laws are actually useful. Patents require companies to disclose<mask> the product works<mask> that everyone knows<mask> the invention does and<mask>,<mask> then gives that person or company exclusive rights for a period of time. That's good, now society has more knowledge that it can build on. The alternative would be far more industry trade secrets, and that means even less dissemination of knowledge. </s>
Label encoding: <s> [STARTQ] However, I don't think anyone "owns" your idea, not even you. [ENDQ] [STARTQ] [ENDQ] [NEWLINE] Why not? If I own the computer I'm typing this on, and I sell it to my brother, but leave the computer where it is, nothing has changed about the computer. It's still in the same spot, physically identical. The difference is not tangible. Ownership is an idea. Possession of something isn't ownership, nor is usage of something. The concept of ownership exists because it is useful. We apply that concept to both physical things and ideas. [NEWLINE] [NEWLINE] [STARTQ] [URL] [ENDQ] [NEWLINE] This is an argument that we need to reform patent laws, not do away with them completely. [NEWLINE] [NEWLINE] Patent laws are actually useful. Patents require companies to disclose how the product works so that everyone knows what the invention does and how, but then gives that person or company exclusive rights for a period of time. That's good, now society has more knowledge that it can build on. The alternative would be far more industry trade secrets, and that means even less dissemination of knowledge. </s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> by this logic, people who are unable to donate blood shouldn't be able to receive blood in a time of crisis<mask> well? [NEWLINE] [NEWLINE] I can't donate blood, I haven't been able to for two years. I have untreated precancerous cells in my cervix (I'm going through treatment in October. Maybe) and this disqualifies me from donating blood.<mask> should I not be able to receive blood<mask> I'm in an accident just<mask> I'm medically unable to give it? [NEWLINE] [NEWLINE] I ask the same question for organ donors. For a<mask>, I did not have good control of my Type 1 diabetes and my doctor suggested that I not become an organ donor until I get my shit together. I've done<mask>, and my diabetes shouldn't be a problem with organ donating,<mask> there is (again) the small problem of having active pre-cancerous cells inside me. I would be unable to donate anything<mask> of these. [NEWLINE] [NEWLINE] So<mask><mask> someone is medically unable to donate an organ or blood? Are we automatically exempt from getting the same treatment<mask> those who donate?</s>
Label encoding: <s>So by this logic, people who are unable to donate blood shouldn't be able to receive blood in a time of crisis as well? [NEWLINE] [NEWLINE] I can't donate blood, I haven't been able to for two years. I have untreated precancerous cells in my cervix (I'm going through treatment in October. Maybe) and this disqualifies me from donating blood. So should I not be able to receive blood if I'm in an accident just because I'm medically unable to give it? [NEWLINE] [NEWLINE] I ask the same question for organ donors. For a while, I did not have good control of my Type 1 diabetes and my doctor suggested that I not become an organ donor until I get my shit together. I've done so, and my diabetes shouldn't be a problem with organ donating, but there is (again) the small problem of having active pre-cancerous cells inside me. I would be unable to donate anything because of these. [NEWLINE] [NEWLINE] So what if someone is medically unable to donate an organ or blood? Are we automatically exempt from getting the same treatment as those who donate?</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you insult somebody by saying they're a member of another identity, you are insinuating that there's something wrong with belonging to that identity. [ENDQ] [NEWLINE] Not necessarily.  By telling people something they are not, they are removing that person's identity, and<mask><mask> that is the insult in itself without much thought<mask> to<mask> the means to remove the identity are. [NEWLINE] [NEWLINE] <mask> you call someone a douchebag we don't mean douchebags (bags to perform douches) are bad, they just are offensive to someone who is not a douchebag (<mask> douchebags are objects, no-one happens to be one). <mask> you call a straight person gay, or a man a girl, or a human a dog or pig, etc. you rarely have the target of the word in mind. [NEWLINE] [NEWLINE] The thing is,<mask> someone does segregate against gays and<mask> uses the word gay frequently<mask> an offence we have an overlap of offensive attitudes and this is to be criticized.  This doesn't mean that someone saying "that's<mask> gay" is automatically offensive.</s>
Label encoding: <s> [STARTQ] When you insult somebody by saying they're a member of another identity, you are insinuating that there's something wrong with belonging to that identity. [ENDQ] [NEWLINE] Not necessarily.  By telling people something they are not, they are removing that person's identity, and I think that is the insult in itself without much thought as to what the means to remove the identity are. [NEWLINE] [NEWLINE] When you call someone a douchebag we don't mean douchebags (bags to perform douches) are bad, they just are offensive to someone who is not a douchebag ( as douchebags are objects, no-one happens to be one).  When you call a straight person gay, or a man a girl, or a human a dog or pig, etc. you rarely have the target of the word in mind. [NEWLINE] [NEWLINE] The thing is, if someone does segregate against gays and also uses the word gay frequently as an offence we have an overlap of offensive attitudes and this is to be criticized.  This doesn't mean that someone saying "that's so gay" is automatically offensive.</s>
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can't change your view<mask><mask><mask> with it<mask> you have a few premises that you need to wonder about: [NEWLINE] [NEWLINE] [STARTQ] crimes like these are a perfect display of the human nature in it's most basic state. [ENDQ] [NEWLINE] Systematic abuse is not a perfect display of the human nature in it's most basic state. Beating each other with whatever object is nearest to us is more closer to it. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> it is natural for humans to be brutal and "barbaric" [ENDQ] [NEWLINE] The worst of the nazis, in some people's eyes, were not the ones doing the pointless beatings -<mask> the ones in charge of them. The ones who decided who was going to live today and who was going to die today, based on a whim of evilness (arguable term, see [Arendt]( [URL] #The_banality_of_evil)) [NEWLINE] [NEWLINE] -------------------- [NEWLINE] [NEWLINE] It is a natural trait of humans to do<mask> humans do. And<mask><mask><mask> humans commit pointless violent acts then those acts are a natural trait of humans. </s>
Label encoding: <s>I can't change your view because I agree with it but you have a few premises that you need to wonder about: [NEWLINE] [NEWLINE] [STARTQ] crimes like these are a perfect display of the human nature in it's most basic state. [ENDQ] [NEWLINE] Systematic abuse is not a perfect display of the human nature in it's most basic state. Beating each other with whatever object is nearest to us is more closer to it. [NEWLINE] [NEWLINE] [STARTQ] IMO it is natural for humans to be brutal and "barbaric" [ENDQ] [NEWLINE] The worst of the nazis, in some people's eyes, were not the ones doing the pointless beatings - but the ones in charge of them. The ones who decided who was going to live today and who was going to die today, based on a whim of evilness (arguable term, see [Arendt]( [URL] #The_banality_of_evil)) [NEWLINE] [NEWLINE] -------------------- [NEWLINE] [NEWLINE] It is a natural trait of humans to do what humans do. And as long as humans commit pointless violent acts then those acts are a natural trait of humans. </s>
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> I watch Dexter cutting people up into little pieces, I'm simply observing his actions.<mask><mask> I take control of a serial killer in a video game, *I'm* the one who is pressing the button prompt to saw this leg off or remove this finger.<mask> I don't press the button;<mask> I just turn off the game... that body will never be dismembered.<mask><mask> I *do* press the button... then I am directly contributing to killing that victim. [NEWLINE] [NEWLINE] I become the killer (or I at least assist the killer) instead of *watching* the killer. [NEWLINE] [NEWLINE] When I listen to Eminem rap about beating a girl, I'm listening to him doing it...<mask> I'm not actively taking part.<mask> there was a game<mask> you could beat women, I *would* be taking part by controlling my character and pressing all of the right buttons in order to make it happen. [NEWLINE] [NEWLINE] Does that clear up my meaning at all? I'm not necessarily talking about inducing feelings<mask> more about actively taking part in the violent activity that is being portrayed. </s>
Label encoding: <s>When I watch Dexter cutting people up into little pieces, I'm simply observing his actions. But when I take control of a serial killer in a video game, *I'm* the one who is pressing the button prompt to saw this leg off or remove this finger. If I don't press the button; if I just turn off the game... that body will never be dismembered. But if I *do* press the button... then I am directly contributing to killing that victim. [NEWLINE] [NEWLINE] I become the killer (or I at least assist the killer) instead of *watching* the killer. [NEWLINE] [NEWLINE] When I listen to Eminem rap about beating a girl, I'm listening to him doing it... but I'm not actively taking part. If there was a game where you could beat women, I *would* be taking part by controlling my character and pressing all of the right buttons in order to make it happen. [NEWLINE] [NEWLINE] Does that clear up my meaning at all? I'm not necessarily talking about inducing feelings but more about actively taking part in the violent activity that is being portrayed. </s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's the thing. <mask><mask> that, out of context, the sections you've quoted are a little rapey. <mask>, in the context of the rest of the guide, tofu is talking about treating women extremely well.  It's about realizing yourself<mask> a man and about learning<mask> makes women happy. <mask> you take this section at face value and follow his advice without trying to learn from the negative reactions you'd get by following it explicitly, then you are not really working in the spirit of the rest of the guide.  The point is to try things and learn from them. [NEWLINE] [NEWLINE] <mask>, given the language of the rest of the guide,<mask><mask> the "get the fuck off me you creep" business is not to be taken word for word.  It's meant to be a little funny.  You go until you get resistance and then you respectfully back off.  That's the real takeaway. [NEWLINE] [NEWLINE] This gets to my basic argument about the kickstarter thing. <mask><mask> people are reacting to this one section of the guide without looking at it in context.</s><pad>
Label encoding: <s>Here's the thing.  I agree that, out of context, the sections you've quoted are a little rapey.  However, in the context of the rest of the guide, tofu is talking about treating women extremely well.  It's about realizing yourself as a man and about learning what makes women happy.  If you take this section at face value and follow his advice without trying to learn from the negative reactions you'd get by following it explicitly, then you are not really working in the spirit of the rest of the guide.  The point is to try things and learn from them. [NEWLINE] [NEWLINE] Also, given the language of the rest of the guide, I think the "get the fuck off me you creep" business is not to be taken word for word.  It's meant to be a little funny.  You go until you get resistance and then you respectfully back off.  That's the real takeaway. [NEWLINE] [NEWLINE] This gets to my basic argument about the kickstarter thing.  I think people are reacting to this one section of the guide without looking at it in context.</s><pad>
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] and risk-taking is left to indie developers, whose ideas and properties will be co-opted, copied, or simply bought, and integrated [ENDQ] [NEWLINE] <mask>, these indie developers still exist, and their ideas are still being created. I'm not really sure<mask> you mean by the properties being co-opted/copied/whatever. I mean, take Braid for example. Did any of those things happen there? Microsoft bought Minecraft,<mask><mask><mask>? That didn't change the fact that Minecraft was created in the first place. Minecraft is still there. [NEWLINE] [NEWLINE] <mask> I had to give you a reason for optimism, its that Indie devs are still out there doing their thing, and advancing technology and new tools and APIs are making the things they can do closer and closer to<mask> AAA titles can create.<mask> the barrier to creating "AAA quality" games continues to decrease, we're getting closer and closer to the best of both worlds. Indie developers that are not beholden to any of the things you're worried about, being able to simultaneously innovate and take advantage of the latest and greatest technology.</s>
Label encoding: <s> [STARTQ] and risk-taking is left to indie developers, whose ideas and properties will be co-opted, copied, or simply bought, and integrated [ENDQ] [NEWLINE] So, these indie developers still exist, and their ideas are still being created. I'm not really sure what you mean by the properties being co-opted/copied/whatever. I mean, take Braid for example. Did any of those things happen there? Microsoft bought Minecraft, but so what? That didn't change the fact that Minecraft was created in the first place. Minecraft is still there. [NEWLINE] [NEWLINE] If I had to give you a reason for optimism, its that Indie devs are still out there doing their thing, and advancing technology and new tools and APIs are making the things they can do closer and closer to what AAA titles can create. As the barrier to creating "AAA quality" games continues to decrease, we're getting closer and closer to the best of both worlds. Indie developers that are not beholden to any of the things you're worried about, being able to simultaneously innovate and take advantage of the latest and greatest technology.</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm clearly a different person here,<mask> I'm<mask> an American football fan (I like your football too, I'm simply more well versed in American). I am a Detroit Lions fan and we vehemently hate the Green Bay Packers.<mask>, one of my best friends from my youth to today is a Packers fan. We take jabs at each other,<mask> at the end of the day, I'd do most anything for her. I'm<mask> a University of Michigan fan, and we hate Ohio State University (that was painful to type) even more than Lions fans hate Packers fans. A professor of mine who is originally from the UK has used the Michigan-Ohio State rivalry to help us Yanks get a better understanding of the Celtic-Rangers rivalry.<mask> again, a bunch of my friends are OSU fans. We go after each other on game day/week,<mask> the game doesn't affect our friendship. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the game is over, you could get a drink with the other team's supporters and hang? [ENDQ] [NEWLINE] Yes I could, and routinely have.</s>
Label encoding: <s>I'm clearly a different person here, but I'm also an American football fan (I like your football too, I'm simply more well versed in American). I am a Detroit Lions fan and we vehemently hate the Green Bay Packers. However, one of my best friends from my youth to today is a Packers fan. We take jabs at each other, but at the end of the day, I'd do most anything for her. I'm also a University of Michigan fan, and we hate Ohio State University (that was painful to type) even more than Lions fans hate Packers fans. A professor of mine who is originally from the UK has used the Michigan-Ohio State rivalry to help us Yanks get a better understanding of the Celtic-Rangers rivalry. Yet again, a bunch of my friends are OSU fans. We go after each other on game day/week, but the game doesn't affect our friendship. [NEWLINE] [NEWLINE] [STARTQ] So when the game is over, you could get a drink with the other team's supporters and hang? [ENDQ] [NEWLINE] Yes I could, and routinely have.</s>
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Taking a child away from a parent is a terrible thing to do.  The only time it should ever be considered is<mask> leaving the child with the parent is an even more terrible option. [NEWLINE] [NEWLINE] I can't imagine<mask> it would be like to have to deal with some asshole bureaucrat at the child-rearing version of the DMV just to be allowed to raise my own children.  Have you ever used a DMV?  Have you ever waited two hours in line, just to be told that the letter from the utility company confirming that they've turned on your service at your address in your name is not valid to prove your address, that you need a bill instead?  Can you imagine those knuckleheads sitting in God-like judgement over you, determining<mask> you're fit to raise your own children? [NEWLINE] [NEWLINE] Can you imagine being judged on adherence to the latest parenting fads?  In California would your kids be taken away<mask> you decided to vaccinate? <mask> you paddled your child?  In Texas would your kid be taken away<mask> you didn't paddle your child?</s>
Label encoding: <s>Taking a child away from a parent is a terrible thing to do.  The only time it should ever be considered is when leaving the child with the parent is an even more terrible option. [NEWLINE] [NEWLINE] I can't imagine what it would be like to have to deal with some asshole bureaucrat at the child-rearing version of the DMV just to be allowed to raise my own children.  Have you ever used a DMV?  Have you ever waited two hours in line, just to be told that the letter from the utility company confirming that they've turned on your service at your address in your name is not valid to prove your address, that you need a bill instead?  Can you imagine those knuckleheads sitting in God-like judgement over you, determining if you're fit to raise your own children? [NEWLINE] [NEWLINE] Can you imagine being judged on adherence to the latest parenting fads?  In California would your kids be taken away if you decided to vaccinate?  If you paddled your child?  In Texas would your kid be taken away if you didn't paddle your child?</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree, it's definitely something that needs to be considered,<mask> confirmation bias can<mask> work in the other direction. It may not be from internet articles,<mask> like you say, that isn't sensationalized,<mask><mask> your experience and general attitude is that "feminists are normally level-headed and reasonable", then you will tend to re-enforce that idea<mask> you see examples of it and discount examples to the contrary<mask> "vocal minority".<mask><mask> both sides are guilty of it,<mask> to be realistic, OP's "side" probably falls victim to it moreso.<mask>, the sheer frequency and absurdity of some examples of<mask> many call the "vocal minority" does lend some credence to the idea that it isn't just a made-up boogeyman; there are enough people out there presenting things this way that it seems unlikely that it is just a small fringe thing. Obviously there isn't any way to measure or quantify that, it is just based on personal experiences<mask> those are still valid and do indicate the true state of things to a certain degree. </s>
Label encoding: <s>I agree, it's definitely something that needs to be considered, but confirmation bias can also work in the other direction. It may not be from internet articles, because like you say, that isn't sensationalized, but if your experience and general attitude is that "feminists are normally level-headed and reasonable", then you will tend to re-enforce that idea when you see examples of it and discount examples to the contrary as "vocal minority". I think both sides are guilty of it, but to be realistic, OP's "side" probably falls victim to it moreso. However, the sheer frequency and absurdity of some examples of what many call the "vocal minority" does lend some credence to the idea that it isn't just a made-up boogeyman; there are enough people out there presenting things this way that it seems unlikely that it is just a small fringe thing. Obviously there isn't any way to measure or quantify that, it is just based on personal experiences but those are still valid and do indicate the true state of things to a certain degree. </s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you should focus more on changing your view of mental illness,<mask> "actively pursu[ing] help" is sometimes far more difficult than waking up and deciding to go see a therapist and take some happy pills<mask> you'll be okay tomorrow. [NEWLINE] [NEWLINE] Source: 12 years and counting of treatment, no end in sight, life still in shambles. I'm 24, and I've been in treatment literally half my life now. Panic attacks and flashbacks can stop even the most determined person, and the crippling depression can sap any motivation whatsoever, even<mask> you know that logically, your feelings are incorrect. [NEWLINE] [NEWLINE] Good for the people you know, and I'm not trying to detract from that,<mask> the plural of anecdote is not data. Don't think that mental illness can be cured by pushing through it all the time, or even most of the time. [NEWLINE] [NEWLINE] Random note,<mask> - is that friend in the US?<mask> disability claims often get denied the first and second times,<mask> I'm very, VERY confident that she could get aid with the help of a lawyer.</s>
Label encoding: <s>I think you should focus more on changing your view of mental illness, as "actively pursu[ing] help" is sometimes far more difficult than waking up and deciding to go see a therapist and take some happy pills so you'll be okay tomorrow. [NEWLINE] [NEWLINE] Source: 12 years and counting of treatment, no end in sight, life still in shambles. I'm 24, and I've been in treatment literally half my life now. Panic attacks and flashbacks can stop even the most determined person, and the crippling depression can sap any motivation whatsoever, even if you know that logically, your feelings are incorrect. [NEWLINE] [NEWLINE] Good for the people you know, and I'm not trying to detract from that, but the plural of anecdote is not data. Don't think that mental illness can be cured by pushing through it all the time, or even most of the time. [NEWLINE] [NEWLINE] Random note, though - is that friend in the US? Because disability claims often get denied the first and second times, but I'm very, VERY confident that she could get aid with the help of a lawyer.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your point of view seems to be that video games are useless<mask> you can develop those skills in other ways.<mask> honestly,<mask> kind of logic is that? That’s like saying swimming is useless<mask> you can get into shape using a weight machine.<mask><mask><mask> I don’t want to use a weight machine? I like swimming<mask> I can get fit and do something I love to do. [NEWLINE] [NEWLINE] <mask> you are saying that video games are essentially useless<mask> you can get those analytical skills by, say, doing a puzzle, right? Well then who is to say the puzzle isn’t the useless activity?<mask> is it video games that you label<mask> “useless”? [NEWLINE] [NEWLINE] I<mask> really think you are underestimating video games<mask> an art form. Some have really powerful stories that teach you about morality, good and evil, suffering, sacrifice, etc. There are games that have made grown men cry and literally changed the way people think.<mask> you think these are useless, then are books are movies useless too?<mask> makes them better? [NEWLINE] </s>
Label encoding: <s>Your point of view seems to be that video games are useless because you can develop those skills in other ways. But honestly, what kind of logic is that? That’s like saying swimming is useless because you can get into shape using a weight machine. But what if I don’t want to use a weight machine? I like swimming because I can get fit and do something I love to do. [NEWLINE] [NEWLINE] So you are saying that video games are essentially useless because you can get those analytical skills by, say, doing a puzzle, right? Well then who is to say the puzzle isn’t the useless activity? Why is it video games that you label as “useless”? [NEWLINE] [NEWLINE] I also really think you are underestimating video games as an art form. Some have really powerful stories that teach you about morality, good and evil, suffering, sacrifice, etc. There are games that have made grown men cry and literally changed the way people think. If you think these are useless, then are books are movies useless too? What makes them better? [NEWLINE] </s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] which is of course sort of hard to define. most subs are subs for discussion<mask><mask> i understand you right you want "discussion" to mean something harder (<mask> banning half the ideological spectrum on a discuss ideology sub would be wrong)<mask> can we draw a line that excludes sane discussion stuff without that's justified even<mask> we acknowledge the sorieties problem isn't a real objection here? [ENDQ] [NEWLINE] By "discussion" I mean ***proper*** discussion (not "circlejerk" discussion)<mask> there are multiple, differing views and opinions being discussed. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> there seems to be some use for useful "tone policing" (<mask><mask> that term is often used in a negative context) [ENDQ] [NEWLINE] In the context of the sub, the active-commenting-to-moderator ratio is<mask> close that it isn't necessary to downvote for tonal policing; anything out of line--or is borderline--can (and will) quickly be addressed by a moderator. This may not always be true in the future,<mask> at the moment it is.</s>
Label encoding: <s> [STARTQ] which is of course sort of hard to define. most subs are subs for discussion though if i understand you right you want "discussion" to mean something harder ( so banning half the ideological spectrum on a discuss ideology sub would be wrong) but can we draw a line that excludes sane discussion stuff without that's justified even if we acknowledge the sorieties problem isn't a real objection here? [ENDQ] [NEWLINE] By "discussion" I mean ***proper*** discussion (not "circlejerk" discussion) where there are multiple, differing views and opinions being discussed. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] though there seems to be some use for useful "tone policing" ( even though that term is often used in a negative context) [ENDQ] [NEWLINE] In the context of the sub, the active-commenting-to-moderator ratio is so close that it isn't necessary to downvote for tonal policing; anything out of line--or is borderline--can (and will) quickly be addressed by a moderator. This may not always be true in the future, but at the moment it is.</s>
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Well,<mask> not?<mask> do you lose exactly from calling a woman with a beard a woman? [ENDQ] [NEWLINE] Depends on the situation.  I (and most normal people) don't have a problem with it,<mask><mask> there is a person with a beard wearing a dress in the men's room (or worse, a person with a beard wearing a dress in the ladies room) then<mask><mask> we need to call a spade a spade. [NEWLINE] [NEWLINE] <mask> you want me to call you "she", that's fine,<mask> you keep to the area designated for people whose genitals match yours. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] there doesn't seem to be much agreement about which set and whether any of the gender-neutral pronouns should be promoted [ENDQ] [NEWLINE] <mask> making up words is stupid. [NEWLINE] [NEWLINE] [STARTQ] people are dismissive of them [ENDQ] [NEWLINE] See above. [NEWLINE] [NEWLINE] [STARTQ] and we seem to do well enough with "you"<mask> both a singular and a plural that I'm not too worried about<mask> we'll handle "they." [ENDQ] [NEWLINE] Nothing wrong with being non-specific at all.</s>
Label encoding: <s> [STARTQ] Well, why not? What do you lose exactly from calling a woman with a beard a woman? [ENDQ] [NEWLINE] Depends on the situation.  I (and most normal people) don't have a problem with it, but when there is a person with a beard wearing a dress in the men's room (or worse, a person with a beard wearing a dress in the ladies room) then I think we need to call a spade a spade. [NEWLINE] [NEWLINE] If you want me to call you "she", that's fine, but you keep to the area designated for people whose genitals match yours. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] there doesn't seem to be much agreement about which set and whether any of the gender-neutral pronouns should be promoted [ENDQ] [NEWLINE] Because making up words is stupid. [NEWLINE] [NEWLINE] [STARTQ] people are dismissive of them [ENDQ] [NEWLINE] See above. [NEWLINE] [NEWLINE] [STARTQ] and we seem to do well enough with "you" as both a singular and a plural that I'm not too worried about how we'll handle "they." [ENDQ] [NEWLINE] Nothing wrong with being non-specific at all.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the OP's point should be clarified<mask> follows: Republicans are either hypocritical, or dumb. Case in point: [NEWLINE] [NEWLINE] [STARTQ] Huge military spending is something a lot of people see<mask> necessary, especially in a world<mask> most of our allies are spending a lot less on the military. Take the recent troubles in Ukraine, for instance. I personally think it isn't going to amount to much and that Putin isn't going to invade Ukraine,<mask><mask> they did...<mask> would happen? France, Germany, Britain, none of them have near the military strength to credibly threaten or dissuade Russia. We do.<mask> for the other issues,<mask><mask> you're misrepresenting<mask> Republicans think the economy works. I wont' get in to the specific arguments here unless you're interested,<mask> Republicans want a better economy and they think that both smaller government and less taxes (across the board, which includes the rich) both facilitate that. You may disagree,<mask> there is no hypocrisy here. [ENDQ] [NEWLINE] Really??? You are going to enter a WWIII with Russia? Over Ukraine?</s>
Label encoding: <s>I think the OP's point should be clarified as follows: Republicans are either hypocritical, or dumb. Case in point: [NEWLINE] [NEWLINE] [STARTQ] Huge military spending is something a lot of people see as necessary, especially in a world where most of our allies are spending a lot less on the military. Take the recent troubles in Ukraine, for instance. I personally think it isn't going to amount to much and that Putin isn't going to invade Ukraine, but if they did... what would happen? France, Germany, Britain, none of them have near the military strength to credibly threaten or dissuade Russia. We do. As for the other issues, I think you're misrepresenting how Republicans think the economy works. I wont' get in to the specific arguments here unless you're interested, but Republicans want a better economy and they think that both smaller government and less taxes (across the board, which includes the rich) both facilitate that. You may disagree, but there is no hypocrisy here. [ENDQ] [NEWLINE] Really??? You are going to enter a WWIII with Russia? Over Ukraine?</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, you make an assumption they will get pregnant.<mask> others have pointed out, they may be infertile. [ENDQ] [NEWLINE] No he isn't. He is making the assumption that there is a chance they could get pregnant. [NEWLINE] He isn't saying "all women always get pregnant", he is saying "a woman could get pregnant, which a man can't,<mask> I prefer to hire men". You are misrepresenting the OPs arguments. [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] Thirdly, you do admit that there are qualifications that would allow you to hire a female employee. This means your view isn't absolute<mask> you said. [NEWLINE] [NEWLINE] Of course there are qualifications, and<mask> a person can clearly offset the loss they might incur<mask> pregnant then this would be a general plus for the business. I don't think any businessman/woman would say no to a deal which ends up working out in their favor. This doesn't change that<mask> both a man and a woman have these equal qualifications, the OP would go with the man<mask> a man doesn't have the chance to get pregnant.</s>
Label encoding: <s> [STARTQ] Firstly, you make an assumption they will get pregnant. As others have pointed out, they may be infertile. [ENDQ] [NEWLINE] No he isn't. He is making the assumption that there is a chance they could get pregnant. [NEWLINE] He isn't saying "all women always get pregnant", he is saying "a woman could get pregnant, which a man can't, hence I prefer to hire men". You are misrepresenting the OPs arguments. [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] Thirdly, you do admit that there are qualifications that would allow you to hire a female employee. This means your view isn't absolute as you said. [NEWLINE] [NEWLINE] Of course there are qualifications, and if a person can clearly offset the loss they might incur while pregnant then this would be a general plus for the business. I don't think any businessman/woman would say no to a deal which ends up working out in their favor. This doesn't change that if both a man and a woman have these equal qualifications, the OP would go with the man because a man doesn't have the chance to get pregnant.</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know some companies have "flex time", <mask> they are decidedly and probably permanently in the minority. Most of the business world would judge me pretty harshly for this behavior and many places would fire me for tardiness. [NEWLINE] [NEWLINE] <mask><mask> this is a holdover from the work schedules of farms, factories and the military. It has no bearing whatsoever on my job or my performance,<mask> most corporate types would righteously call me lazy. [NEWLINE] [NEWLINE] In trying to adapt to their schedule, I sit like a zombie through the morning hours and then feel groggy during<mask> used to be my most productive time - late at night. [NEWLINE] [NEWLINE] I am not a corporate malcontent. I like wearing suits and working in an office. I can totally accept that office politics are inevitable. I know<mask> I have to have 3 bosses. I went to business school. I am good at this.<mask> this one thing just kills me. [NEWLINE] [NEWLINE] This is obviously a significant source of friction in my life and I would love it<mask> someone could clear it up for me. CMV.</s>
Label encoding: <s>I know some companies have "flex time",  but they are decidedly and probably permanently in the minority. Most of the business world would judge me pretty harshly for this behavior and many places would fire me for tardiness. [NEWLINE] [NEWLINE] I think this is a holdover from the work schedules of farms, factories and the military. It has no bearing whatsoever on my job or my performance, yet most corporate types would righteously call me lazy. [NEWLINE] [NEWLINE] In trying to adapt to their schedule, I sit like a zombie through the morning hours and then feel groggy during what used to be my most productive time - late at night. [NEWLINE] [NEWLINE] I am not a corporate malcontent. I like wearing suits and working in an office. I can totally accept that office politics are inevitable. I know why I have to have 3 bosses. I went to business school. I am good at this. But this one thing just kills me. [NEWLINE] [NEWLINE] This is obviously a significant source of friction in my life and I would love it if someone could clear it up for me. CMV.</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Let's stick with your metaphor,<mask> you were driving drunk and speeding, and the person you're hooked up to in order to save their life is your child who was<mask> in the car. [ENDQ] [NEWLINE] The difference between that and my example is using protection or not. I'm not trying to argue<mask> abortion is moral or not, merely that it is a women's rights issue. [NEWLINE] [NEWLINE] [STARTQ] <mask> we can argue about whether or not it overrides your bodily autonomy. [ENDQ] [NEWLINE] And the fact that we are arguing about it at all proves it is a women's rights issue. [NEWLINE] [NEWLINE] I'm not saying it's not a fetus's rights issue<mask> well.<mask><mask> it's both. [NEWLINE] [NEWLINE] [STARTQ] Rape and contraception failure are interesting cases,<mask> they are minorities. Let's first let's address the strongest version of OP's claim. [ENDQ] [NEWLINE] <mask>? OP's opinion is that it is not a women's rights issue at all.<mask> I can show it is a women's right issue in at least one case, then<mask> shouldn't OP change their mind?</s>
Label encoding: <s> [STARTQ] Let's stick with your metaphor, but you were driving drunk and speeding, and the person you're hooked up to in order to save their life is your child who was also in the car. [ENDQ] [NEWLINE] The difference between that and my example is using protection or not. I'm not trying to argue if abortion is moral or not, merely that it is a women's rights issue. [NEWLINE] [NEWLINE] [STARTQ] although we can argue about whether or not it overrides your bodily autonomy. [ENDQ] [NEWLINE] And the fact that we are arguing about it at all proves it is a women's rights issue. [NEWLINE] [NEWLINE] I'm not saying it's not a fetus's rights issue as well. Imo it's both. [NEWLINE] [NEWLINE] [STARTQ] Rape and contraception failure are interesting cases, but they are minorities. Let's first let's address the strongest version of OP's claim. [ENDQ] [NEWLINE] Why? OP's opinion is that it is not a women's rights issue at all. If I can show it is a women's right issue in at least one case, then why shouldn't OP change their mind?</s>
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My whole point was that the idea of pre-determined social roles is sexist [ENDQ] [NEWLINE] Sure I got that.<mask> my point is that *sex* is itself a constrictive social construct. Basing gender around genitals is not really productive, nor is it (<mask><mask><mask> ) particularly progressive, radical, or liberal. My point was that gender = genitals is a good deal more restrictive than anything your average trans person does. Or at least equally bad. [NEWLINE] [NEWLINE] And I did see the delta,<mask> felt that it was still worthwhile to add some extra perspective into the issue, from a non-gender conforming trans woman :) I<mask> didn't think you were an ignorant bigot - this is a perspective (Trans people are sexist in their assumptions about<mask> makes them man/woman/etc) I've encountered a few times from left leaning folks.<mask> it's a pet peeve, I don't think it really makes you evil or anything like that. I'm certainly happy your view was changed and hope my post adds some useful perspective for you to chew on :)</s>
Label encoding: <s> [STARTQ] My whole point was that the idea of pre-determined social roles is sexist [ENDQ] [NEWLINE] Sure I got that. But my point is that *sex* is itself a constrictive social construct. Basing gender around genitals is not really productive, nor is it ( in my opinion ) particularly progressive, radical, or liberal. My point was that gender = genitals is a good deal more restrictive than anything your average trans person does. Or at least equally bad. [NEWLINE] [NEWLINE] And I did see the delta, but felt that it was still worthwhile to add some extra perspective into the issue, from a non-gender conforming trans woman :) I also didn't think you were an ignorant bigot - this is a perspective (Trans people are sexist in their assumptions about what makes them man/woman/etc) I've encountered a few times from left leaning folks. While it's a pet peeve, I don't think it really makes you evil or anything like that. I'm certainly happy your view was changed and hope my post adds some useful perspective for you to chew on :)</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that humans are naturally omniovores, not herbivores,<mask><mask><mask> you may be overstating the relative importance of meat in the diets (and the evolution) of chimps and early hominids. [NEWLINE] [NEWLINE] [STARTQ] All you have to do is look at our nearest ancestor (chimps) and see a wild omnivore, which is exactly<mask> we are. [ENDQ] [NEWLINE] Chimps don't eat much meat - it's only about 2-3% of their diet, and about 5-8%<mask> you add in insects. (The exact amount seems to vary a fair bit between different populations and seasons.) [NEWLINE] [NEWLINE] [STARTQ] Every adaptation we have from binocular vision to standing upright is a means of making us more effective hunters. [ENDQ] [NEWLINE] *Most* primates have binocular vision. It helps with brachiating. And there's a bunch of different hypotheses about<mask> humans ended up bipedal, plenty of which have nothing to do with hunting: [URL] #Evolution_of_human_bipedalism [NEWLINE] </s>
Label encoding: <s>I agree that humans are naturally omniovores, not herbivores, but I think you may be overstating the relative importance of meat in the diets (and the evolution) of chimps and early hominids. [NEWLINE] [NEWLINE] [STARTQ] All you have to do is look at our nearest ancestor (chimps) and see a wild omnivore, which is exactly what we are. [ENDQ] [NEWLINE] Chimps don't eat much meat - it's only about 2-3% of their diet, and about 5-8% when you add in insects. (The exact amount seems to vary a fair bit between different populations and seasons.) [NEWLINE] [NEWLINE] [STARTQ] Every adaptation we have from binocular vision to standing upright is a means of making us more effective hunters. [ENDQ] [NEWLINE] *Most* primates have binocular vision. It helps with brachiating. And there's a bunch of different hypotheses about how humans ended up bipedal, plenty of which have nothing to do with hunting: [URL] #Evolution_of_human_bipedalism [NEWLINE] </s>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are right.<mask><mask> I've tried to get across is that<mask> you express yourself you do more than simply express your feelings; you imply that you have a reason to say it. That reason won't be assumed to be malicious in many,<mask> not most, contexts.<mask><mask> you're in a situation<mask> the subject of the criticism is present, they may quite reasonably take it in a humiliating way. [NEWLINE] [NEWLINE] I mean, on a very basic level, someone who thinks that they smell will not go around saying to others that they should take a shower,<mask> they'd realise that that would be hypocritical and pointless. It wouldn't be their place.<mask>,<mask> somebody tells you that they smell, you kind of think "woah, wait a sec, you think that you don't smell?<mask> do you see yourself<mask><mask> high and mighty?" [NEWLINE] [NEWLINE] <mask> I've said many a time by now, I don't think this applies to all criticisms.<mask> I'm saying is that the context of a criticism can lend an extra dimension of humiliation.</s>
Label encoding: <s>You are right. But what I've tried to get across is that when you express yourself you do more than simply express your feelings; you imply that you have a reason to say it. That reason won't be assumed to be malicious in many, if not most, contexts. But if you're in a situation where the subject of the criticism is present, they may quite reasonably take it in a humiliating way. [NEWLINE] [NEWLINE] I mean, on a very basic level, someone who thinks that they smell will not go around saying to others that they should take a shower, as they'd realise that that would be hypocritical and pointless. It wouldn't be their place. Thus, if somebody tells you that they smell, you kind of think "woah, wait a sec, you think that you don't smell? Why do you see yourself as so high and mighty?" [NEWLINE] [NEWLINE] As I've said many a time by now, I don't think this applies to all criticisms. What I'm saying is that the context of a criticism can lend an extra dimension of humiliation.</s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not about honesty<mask> much<mask> it is 'not biting the hand that feeds you'. [NEWLINE] [NEWLINE] Think a second about the nature of our elections and<mask> much private money is needed to win them.<mask> it's true that the 'guy with the most money' doesn't always win, you still need to meet some money threshold in order to be able to run even biographical ads and get your name out there. It's not<mask> much that money has a corrupting influence, it's that money is the *only* influence. You can read 10 letters from your constituents and respond to them by hand, or you can run an ad and speaks to thousands of your constituents, and carefully craft a message by hiring communications experts to perfectly tailor your message to the people that vote for you. [NEWLINE] [NEWLINE] Money is<mask> you get elected, particularly recently. Advertising isn't cheap. Staff for messaging isn't cheap. Door Knockers are individually cheap,<mask> you have to hire a lot of them. Those are all provably effective at getting you (re)elected.</s>
Label encoding: <s>It's not about honesty so much as it is 'not biting the hand that feeds you'. [NEWLINE] [NEWLINE] Think a second about the nature of our elections and how much private money is needed to win them. While it's true that the 'guy with the most money' doesn't always win, you still need to meet some money threshold in order to be able to run even biographical ads and get your name out there. It's not so much that money has a corrupting influence, it's that money is the *only* influence. You can read 10 letters from your constituents and respond to them by hand, or you can run an ad and speaks to thousands of your constituents, and carefully craft a message by hiring communications experts to perfectly tailor your message to the people that vote for you. [NEWLINE] [NEWLINE] Money is how you get elected, particularly recently. Advertising isn't cheap. Staff for messaging isn't cheap. Door Knockers are individually cheap, but you have to hire a lot of them. Those are all provably effective at getting you (re)elected.</s>
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyone answer to this is going to be stupid.  People are going to say things like "Some people DO!" which is a typical childish argument you hear from Reddit all the time. <mask>,<mask> a scotch lover, I am going to clear all this up for you. [NEWLINE] [NEWLINE] People start drinking scotch<mask> its expensive and high class.  It taste like shit at first.  They keep drinking it<mask> they really want to be classy, and eventually they get used to the taste.  After more classiness they start to like it.  This is<mask> I am at.  I started drinking this stuff for the reasons you listed,<mask> actually ended up really liking it after a few weeks.  The reason you can't do this with other types of alcohol is<mask> those actually DO taste like shit,<mask><mask> the bad taste of scotch is just the high alcohol content. <mask><mask> you get used to the alcohol taste, the rest of it comes out, and THEN it is good. <mask> you're partially correct.</s>
Label encoding: <s>Everyone answer to this is going to be stupid.  People are going to say things like "Some people DO!" which is a typical childish argument you hear from Reddit all the time.  So, as a scotch lover, I am going to clear all this up for you. [NEWLINE] [NEWLINE] People start drinking scotch because its expensive and high class.  It taste like shit at first.  They keep drinking it because they really want to be classy, and eventually they get used to the taste.  After more classiness they start to like it.  This is where I am at.  I started drinking this stuff for the reasons you listed, but actually ended up really liking it after a few weeks.  The reason you can't do this with other types of alcohol is because those actually DO taste like shit, where as the bad taste of scotch is just the high alcohol content.  So when you get used to the alcohol taste, the rest of it comes out, and THEN it is good.  So you're partially correct.</s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The people who usually abstain will abstain [ENDQ] [NEWLINE] See, this is<mask> I definitely disagree - Democrats have a bigger problem with voter apathy than Republicans, and<mask><mask> voter apathy, by and large, is a result of not feeling like any of the options represent your interests/ideals. There are a huge number of progressives on the left that are simply apathetic towards the centrist, GDP worshiping Democratic establishment, who<mask><mask> would be very likely to feel far more engaged with Sanders<mask> the nominee. [NEWLINE] [NEWLINE] [STARTQ] the people who are swing will vote for the GOP. [ENDQ] [NEWLINE] <mask> the GOP nominates Rand Paul, I'd 100% agree with you here -<mask> the GOP nominates Rand Paul, all bets are off,<mask><mask><mask> I'm concerned -<mask> I'm *pretty* sure they won't. That being the case, describe to me the swing voter who would vote for Hillary over Jeb,<mask> would vote for Jeb over Bernie, and<mask> they would do that,<mask> I simply can't envision that person right now.</s>
Label encoding: <s> [STARTQ] The people who usually abstain will abstain [ENDQ] [NEWLINE] See, this is where I definitely disagree - Democrats have a bigger problem with voter apathy than Republicans, and I think voter apathy, by and large, is a result of not feeling like any of the options represent your interests/ideals. There are a huge number of progressives on the left that are simply apathetic towards the centrist, GDP worshiping Democratic establishment, who I think would be very likely to feel far more engaged with Sanders as the nominee. [NEWLINE] [NEWLINE] [STARTQ] the people who are swing will vote for the GOP. [ENDQ] [NEWLINE] If the GOP nominates Rand Paul, I'd 100% agree with you here - if the GOP nominates Rand Paul, all bets are off, as far as I'm concerned - but I'm *pretty* sure they won't. That being the case, describe to me the swing voter who would vote for Hillary over Jeb, but would vote for Jeb over Bernie, and why they would do that, because I simply can't envision that person right now.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think this is<mask> much of a problem anymore. Saying no to a drunk girl<mask> she is drunk pops up in movies and on TV quite regularly these days. Just to give examples the most recent I recall seeing are Superbad and The Inbetweeners. In both cases the male protagonist is lead to a bedroom by a very drunk girl that they have a crush on, and the male expresses that he would really like to<mask> won't<mask> it doesn't feel right with the girl being<mask> drunk. [NEWLINE] [NEWLINE] <mask><mask> this scenario pops up regularly enough in teen movies. It's possible not enough emphasis is put on the fact that<mask> they have sex in that scenario it is rape,<mask> it doesn't change the fact that modern media definitely portrays having sex under those conditions<mask> wrong. I feel that anyone who has sex<mask> the other person is clearly very drunk (assuming they themselves are not<mask> drunk they can't tell) knows<mask> they are doing is wrong, just<mask> much<mask> the man in the bushes knows<mask> he is doing is wrong.</s>
Label encoding: <s>I don't think this is so much of a problem anymore. Saying no to a drunk girl because she is drunk pops up in movies and on TV quite regularly these days. Just to give examples the most recent I recall seeing are Superbad and The Inbetweeners. In both cases the male protagonist is lead to a bedroom by a very drunk girl that they have a crush on, and the male expresses that he would really like to but won't because it doesn't feel right with the girl being so drunk. [NEWLINE] [NEWLINE] I think this scenario pops up regularly enough in teen movies. It's possible not enough emphasis is put on the fact that if they have sex in that scenario it is rape, but it doesn't change the fact that modern media definitely portrays having sex under those conditions as wrong. I feel that anyone who has sex where the other person is clearly very drunk (assuming they themselves are not so drunk they can't tell) knows what they are doing is wrong, just as much as the man in the bushes knows what he is doing is wrong.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, this is an answer from the UK, I really don't want 100 replies from Americans countering this with US police anecdotal evidence. Our legal and policing systems are very different and really cannot be compared. [NEWLINE] [NEWLINE] I don't understand your point,<mask> their job involves reacting to violence, they must secretly love doling it out? Every police officer I've known has been super kind, and far and away more gentle than anyone else I know. You don't get into the police force for a violent kick (anomalies aside), you do it<mask> you want to make a difference. Most people are policemen and women<mask><mask><mask><mask> violence, not<mask> of it; they'd like nothing more than to be redundant in their anti-violence role in society. [NEWLINE] [NEWLINE] You seem to have the idea that all police officers are just itching for a chance to get their hittin' stick out, barely concealing gleeful grins<mask> a drunk piece of shit starts screaming obscenities. This couldn't be further from the truth in my experience.</s>
Label encoding: <s>Firstly, this is an answer from the UK, I really don't want 100 replies from Americans countering this with US police anecdotal evidence. Our legal and policing systems are very different and really cannot be compared. [NEWLINE] [NEWLINE] I don't understand your point, because their job involves reacting to violence, they must secretly love doling it out? Every police officer I've known has been super kind, and far and away more gentle than anyone else I know. You don't get into the police force for a violent kick (anomalies aside), you do it because you want to make a difference. Most people are policemen and women in SPITE of violence, not because of it; they'd like nothing more than to be redundant in their anti-violence role in society. [NEWLINE] [NEWLINE] You seem to have the idea that all police officers are just itching for a chance to get their hittin' stick out, barely concealing gleeful grins when a drunk piece of shit starts screaming obscenities. This couldn't be further from the truth in my experience.</s>
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've received plenty of helpful answers<mask> far<mask> I have a personal anecdote to contribute.<mask> I was young and played soccer, I was bullied by my team mates. I was<mask> miserable my first year. I begged to quit, over and over.<mask> the trophy I got for participating was my first ever, and I was<mask> happy! My neighbor was much older and had several, and I would have quit soccer<mask> not for that little token at the end of the season. I was terrible, didn't have fun with my team,<mask> I did enjoy the support and was rewarded for giving it my best effort.<mask> I stuck with it. Things like that are really big deals to little kids! Learning to be a team player and be active, these are all things the trophy can reward<mask> winning. I ended up being the best goal keeper in my league,<mask> I quit at 16-17 to pursue archery. I referee and my mom coaches, and we see these trophies inspire love for the game and that's really all that matters. </s>
Label encoding: <s>You've received plenty of helpful answers so far but I have a personal anecdote to contribute. When I was young and played soccer, I was bullied by my team mates. I was so miserable my first year. I begged to quit, over and over. But the trophy I got for participating was my first ever, and I was so happy! My neighbor was much older and had several, and I would have quit soccer if not for that little token at the end of the season. I was terrible, didn't have fun with my team, but I did enjoy the support and was rewarded for giving it my best effort. So I stuck with it. Things like that are really big deals to little kids! Learning to be a team player and be active, these are all things the trophy can reward besides winning. I ended up being the best goal keeper in my league, but I quit at 16-17 to pursue archery. I referee and my mom coaches, and we see these trophies inspire love for the game and that's really all that matters. </s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In my view, it is not somebody's job to make themselves more tolerable,<mask> ours to become more tolerant. For example, one could<mask><mask> black people should stop speaking in AAVE (African-American Vernacular English)<mask> that others will treat them with more respect.<mask>,<mask><mask><mask>, there is nothing intrinsically wrong with AAVE, people who dislike<mask> it is spoken need to lighten up. [NEWLINE] [NEWLINE] Or<mask> about gay people? I've heard the argument that we should stop supporting gay pride parades<mask> it just makes gay people seem stupid and flamboyant and re-enforces stereotypes. Who gives a shit? Let them do<mask> they want. The homophobes need to learn to deal. [NEWLINE] [NEWLINE] It's not fair to ask someone to stop doing a harmless thing that is meaningful/a part of who they are for the benefit of others. The others have to learn to accept that other people speak and act a different way from them. That's part of growing up; we learn it<mask> children.</s>
Label encoding: <s>In my view, it is not somebody's job to make themselves more tolerable, but ours to become more tolerant. For example, one could argue that black people should stop speaking in AAVE (African-American Vernacular English) so that others will treat them with more respect. But, in my opinion, there is nothing intrinsically wrong with AAVE, people who dislike when it is spoken need to lighten up. [NEWLINE] [NEWLINE] Or how about gay people? I've heard the argument that we should stop supporting gay pride parades because it just makes gay people seem stupid and flamboyant and re-enforces stereotypes. Who gives a shit? Let them do what they want. The homophobes need to learn to deal. [NEWLINE] [NEWLINE] It's not fair to ask someone to stop doing a harmless thing that is meaningful/a part of who they are for the benefit of others. The others have to learn to accept that other people speak and act a different way from them. That's part of growing up; we learn it as children.</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> is the more important musical movement? People love to complain about the state of hip-hop now<mask> ignoring the masterpieces of the last 15 years that have moved the art forward, and Kanye is the most important of those artists, who do you think is more important? [NEWLINE] [NEWLINE] Did you miss Outkast? Didi you miss Kendrick Lamar? Did you miss MF Doom? Did you miss Lupe Fiasco? Did you miss Mos Def? Did you miss Dead Prez? Are you missing Chance the Rapper? Jay-Z's last 15 years? Common's last 15 years? Kanye's album's are brilliant - College Dropout, Graduation, Beautiful Dark Twisted Fantasy, Watch the Throne. [NEWLINE] [NEWLINE] Kanye's wants to do more,<mask> shouldn't he? Everyone said the same racist nonsense<mask> Puffy started Sean John and<mask> Jay-Z started Rocawear.<mask> Kanye wants to make clothes for white people and everyone calls him a narcissist and a megalomaniac. Hold on to your butts.</s>
Label encoding: <s>So what is the more important musical movement? People love to complain about the state of hip-hop now while ignoring the masterpieces of the last 15 years that have moved the art forward, and Kanye is the most important of those artists, who do you think is more important? [NEWLINE] [NEWLINE] Did you miss Outkast? Didi you miss Kendrick Lamar? Did you miss MF Doom? Did you miss Lupe Fiasco? Did you miss Mos Def? Did you miss Dead Prez? Are you missing Chance the Rapper? Jay-Z's last 15 years? Common's last 15 years? Kanye's album's are brilliant - College Dropout, Graduation, Beautiful Dark Twisted Fantasy, Watch the Throne. [NEWLINE] [NEWLINE] Kanye's wants to do more, why shouldn't he? Everyone said the same racist nonsense when Puffy started Sean John and when Jay-Z started Rocawear. So Kanye wants to make clothes for white people and everyone calls him a narcissist and a megalomaniac. Hold on to your butts.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is a distinction between academic feminism and popular feminism, and I am certainly not trying to absolutely encapsulate fifty years of academic thought, particularly<mask><mask> much of it consists of vastly differing opinions. I'm just trying to give a general overview of the beliefs of at least a part of the MRM. [NEWLINE] [NEWLINE] With reference to patriarchy theory, I have read both the term and the theory in feminist literature.<mask> you are correct in that strictly speaking it refers to men in positions of power, I have often seen this extended to "men possess social power". I was simply trying to point out the misconception that extension involves. [NEWLINE] [NEWLINE] Intersectionality originated in the intersection of of Marxist, feminist, critical race and half a dozen other critiques of social order. The point I'm making is not that intersectionality is not feminist<mask> that in a history of power, class has far more influence than gender. Nowhere do I deny that privilege exists, I'm simply criticizing the extent, magnitude and monolithic nature I have occasionally seen it ascribed. </s>
Label encoding: <s>There is a distinction between academic feminism and popular feminism, and I am certainly not trying to absolutely encapsulate fifty years of academic thought, particularly when so much of it consists of vastly differing opinions. I'm just trying to give a general overview of the beliefs of at least a part of the MRM. [NEWLINE] [NEWLINE] With reference to patriarchy theory, I have read both the term and the theory in feminist literature. While you are correct in that strictly speaking it refers to men in positions of power, I have often seen this extended to "men possess social power". I was simply trying to point out the misconception that extension involves. [NEWLINE] [NEWLINE] Intersectionality originated in the intersection of of Marxist, feminist, critical race and half a dozen other critiques of social order. The point I'm making is not that intersectionality is not feminist but that in a history of power, class has far more influence than gender. Nowhere do I deny that privilege exists, I'm simply criticizing the extent, magnitude and monolithic nature I have occasionally seen it ascribed. </s>
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So just to make my point clear. [NEWLINE] [NEWLINE] Yes I admit that Gandhi has become a propaganda tool for the state.<mask> that does not mean that he did not accomplish stuff which has a lot for us to learn. [NEWLINE] [NEWLINE] Imagine<mask> someone said "I believe Albert Einstein is overrated, CMV", and then talks about all these stories Christians tell each other("..and that boy's name was Albert Einstein")<mask> of his belief in god, then would he be right? These stupid stories do not discount<mask> Einstein did. [NEWLINE] [NEWLINE] Maybe at least you admit that Gandhi the man and Gandhi the propaganda are two different things, and Gandhi is great for different reasons. I would recommend reading him more. There's a lot of bullshit to go through(like his religiousness),<mask> its nothing like<mask> you read through<mask> child. [NEWLINE] [NEWLINE] I must warn you, you will face palm everything you'll hear Anna Hazare take his name on television. Anna Hazare has resurrected Gandhi's corpse,<mask> it does not have his soul in it.</s>
Label encoding: <s>So just to make my point clear. [NEWLINE] [NEWLINE] Yes I admit that Gandhi has become a propaganda tool for the state. But that does not mean that he did not accomplish stuff which has a lot for us to learn. [NEWLINE] [NEWLINE] Imagine if someone said "I believe Albert Einstein is overrated, CMV", and then talks about all these stories Christians tell each other("..and that boy's name was Albert Einstein") because of his belief in god, then would he be right? These stupid stories do not discount what Einstein did. [NEWLINE] [NEWLINE] Maybe at least you admit that Gandhi the man and Gandhi the propaganda are two different things, and Gandhi is great for different reasons. I would recommend reading him more. There's a lot of bullshit to go through(like his religiousness), but its nothing like what you read through as child. [NEWLINE] [NEWLINE] I must warn you, you will face palm everything you'll hear Anna Hazare take his name on television. Anna Hazare has resurrected Gandhi's corpse, but it does not have his soul in it.</s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was going through previous posts and realized I never replied back. I did look into your points. [NEWLINE] [NEWLINE] You are right about the Korean war itself not being caused by the Americans<mask> rather due to the political maneuvering of the Soviets. <mask> the seeds of the Korean war were laid by the arbitrary division by the allies of the Korean peninsula and in that respect is similar to most modern disputes.<mask> the US chose to ignore Korea in it's larger strategy and some would<mask><mask> emboldened the Soviets and China to support North Korea in its invasion.<mask> the only real fault of the US is apathy and negligence considering this was right after World War 2,<mask><mask> the US can be forgiven that. The US can<mask> not really be blamed that the Soviets chose Korea<mask> the point to target American interests and support. [NEWLINE] [NEWLINE] <mask> Korea would not have really achieved independence without American intervention.<mask><mask> with you the Korean war was not exacerbated by US intervention. Thank you for changing my view on this. [NEWLINE] [NEWLINE] &amp;#8710;</s>
Label encoding: <s>I was going through previous posts and realized I never replied back. I did look into your points. [NEWLINE] [NEWLINE] You are right about the Korean war itself not being caused by the Americans but rather due to the political maneuvering of the Soviets.  However the seeds of the Korean war were laid by the arbitrary division by the allies of the Korean peninsula and in that respect is similar to most modern disputes. However the US chose to ignore Korea in it's larger strategy and some would argue that emboldened the Soviets and China to support North Korea in its invasion. So the only real fault of the US is apathy and negligence considering this was right after World War 2, I think the US can be forgiven that. The US can also not really be blamed that the Soviets chose Korea as the point to target American interests and support. [NEWLINE] [NEWLINE] Since Korea would not have really achieved independence without American intervention. I agree with you the Korean war was not exacerbated by US intervention. Thank you for changing my view on this. [NEWLINE] [NEWLINE] &amp;#8710;</s>
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for prompting this discussion. [NEWLINE] [NEWLINE] Would it change your view<mask> I asked you to put yourself in the mind of someone who enjoys the loud exhaust? [NEWLINE] [NEWLINE] From their point of view, they are in the minority of drivers on the road. It's easy to rally the opinion you suggest<mask> most either don't like the exhaust sound or are indifferent. [NEWLINE] [NEWLINE] <mask> attack the liberty of those who enjoy life in this way? Is it<mask> they are showing off and we have to teach them humility? Perhaps. Are our police force the ones to do<mask>? I suggest no. [NEWLINE] [NEWLINE] <mask><mask> we should turn our focus to better educating our society in areas of critical thinking, logic and mental health. This would lead to individual streets and their neighbors conversing constructively with those having the muffler-less, aka loud, motorcycles. This is a better solution than blanket laws. These laws are irrespective and frankly insulting to those riders who enjoy muffler-less, aka loud, exhaust without malice towards others. [NEWLINE] [NEWLINE] Thanks again. </s>
Label encoding: <s>Thanks for prompting this discussion. [NEWLINE] [NEWLINE] Would it change your view if I asked you to put yourself in the mind of someone who enjoys the loud exhaust? [NEWLINE] [NEWLINE] From their point of view, they are in the minority of drivers on the road. It's easy to rally the opinion you suggest as most either don't like the exhaust sound or are indifferent. [NEWLINE] [NEWLINE] Why attack the liberty of those who enjoy life in this way? Is it because they are showing off and we have to teach them humility? Perhaps. Are our police force the ones to do so? I suggest no. [NEWLINE] [NEWLINE] I think we should turn our focus to better educating our society in areas of critical thinking, logic and mental health. This would lead to individual streets and their neighbors conversing constructively with those having the muffler-less, aka loud, motorcycles. This is a better solution than blanket laws. These laws are irrespective and frankly insulting to those riders who enjoy muffler-less, aka loud, exhaust without malice towards others. [NEWLINE] [NEWLINE] Thanks again. </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Are there not basic principles in society that we can hold everyone to? [ENDQ] [NEWLINE] Ironically, that is the point of the BLM movement.  They want the rule of law to be applied to all people,<mask><mask> race. [NEWLINE] [NEWLINE] I work in a profession<mask> I travel a lot.  Years back, just after 9/11, there was an Arab man that I worked with who would always leave for the airport an hour before the rest of the team.  He had to - he knew that he would be "randomly" selected for additional screening by TSA. [NEWLINE] [NEWLINE] <mask> a caucasian, we only have a vague understanding of<mask> it is like to be a part of a discriminated class. <mask> a cop pulls me over, I don't worry about getting shot - my African American friends do.  The goal of the BLM movement is to get this fact in the collective attention and try to get _equal_ treatment by law enforcement. <mask> that takes small acts of civil disobedience, then many consider that worth it.</s>
Label encoding: <s> [STARTQ] Are there not basic principles in society that we can hold everyone to? [ENDQ] [NEWLINE] Ironically, that is the point of the BLM movement.  They want the rule of law to be applied to all people, regardless of race. [NEWLINE] [NEWLINE] I work in a profession where I travel a lot.  Years back, just after 9/11, there was an Arab man that I worked with who would always leave for the airport an hour before the rest of the team.  He had to - he knew that he would be "randomly" selected for additional screening by TSA. [NEWLINE] [NEWLINE] As a caucasian, we only have a vague understanding of what it is like to be a part of a discriminated class.  If a cop pulls me over, I don't worry about getting shot - my African American friends do.  The goal of the BLM movement is to get this fact in the collective attention and try to get _equal_ treatment by law enforcement.  If that takes small acts of civil disobedience, then many consider that worth it.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I put zero blame on civilians.<mask><mask><mask>, I have no idea<mask> the gun laws in Gotham are. Who knows<mask> civilians are carrying guns?<mask>, they are not in the business of crime fighting. [ENDQ] [NEWLINE] Batman is (in the eyes of the law) a civilian,<mask> that doesn't quite parse. [NEWLINE] [NEWLINE] [STARTQ] I see<mask> you're saying about the cops,<mask> cops have to follow orders. [ENDQ] [NEWLINE] The police can put out an order of shoot on sight.  They have not,<mask> the fact that he is a known mass murderer, are they not then culpable? [NEWLINE] [NEWLINE] [STARTQ] Unfortunately, they are trapped in a comic book/TV/movie universe in which only the main protagonists can actually hit anything they aim at. [ENDQ] [NEWLINE] I have to disregard this<mask> it is changing the goal posts.  We have to (for better or worse) impose the rules of the "real world" on the inhabitants of this fictionalized universe, otherwise we have no rules by which to run, no? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I put zero blame on civilians. First of all, I have no idea what the gun laws in Gotham are. Who knows if civilians are carrying guns? Secondly, they are not in the business of crime fighting. [ENDQ] [NEWLINE] Batman is (in the eyes of the law) a civilian, so that doesn't quite parse. [NEWLINE] [NEWLINE] [STARTQ] I see what you're saying about the cops, but cops have to follow orders. [ENDQ] [NEWLINE] The police can put out an order of shoot on sight.  They have not, despite the fact that he is a known mass murderer, are they not then culpable? [NEWLINE] [NEWLINE] [STARTQ] Unfortunately, they are trapped in a comic book/TV/movie universe in which only the main protagonists can actually hit anything they aim at. [ENDQ] [NEWLINE] I have to disregard this as it is changing the goal posts.  We have to (for better or worse) impose the rules of the "real world" on the inhabitants of this fictionalized universe, otherwise we have no rules by which to run, no? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, cheer-leading is a very physical and skilled activity, and in<mask> being falls into the general loose definition of "sport". It is certainly more of a sport than, say, darts, which is very often classed<mask> a sport. [NEWLINE] [NEWLINE] You mention winning by objective means; cheer-leading teams can be scored based on performance, which includes things like style, innovation, choreographic effort and pulling off difficult routines flawlessly. This enables competitions and tournaments to take place with the end result being a ranking, which determines winners etc. [NEWLINE] [NEWLINE] <mask> to conclude: cheer-leading is active, team-oriented and challenging. It is<mask> not luck-based, and<mask> requires a fair amount of developed skill. Teams can be scored and can<mask> compete against other teams in an objective manner, conforming to your own definition of sports. [NEWLINE] <mask> there is no one specific strict definition for<mask> qualifies<mask> a sport, something like the abovedescribed should definitely qualify based on<mask> we typically think of a sport<mask> being.</s>
Label encoding: <s>Firstly, cheer-leading is a very physical and skilled activity, and in so being falls into the general loose definition of "sport". It is certainly more of a sport than, say, darts, which is very often classed as a sport. [NEWLINE] [NEWLINE] You mention winning by objective means; cheer-leading teams can be scored based on performance, which includes things like style, innovation, choreographic effort and pulling off difficult routines flawlessly. This enables competitions and tournaments to take place with the end result being a ranking, which determines winners etc. [NEWLINE] [NEWLINE] So to conclude: cheer-leading is active, team-oriented and challenging. It is also not luck-based, and therefore requires a fair amount of developed skill. Teams can be scored and can therefore compete against other teams in an objective manner, conforming to your own definition of sports. [NEWLINE] Since there is no one specific strict definition for what qualifies as a sport, something like the abovedescribed should definitely qualify based on what we typically think of a sport as being.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask> you're asking is the equivalent of, "<mask> buy your son or daughter a new car for their birthday<mask> you can gift them thousands of dollars in cash instead?"<mask><mask> you're ignoring the human tradition and culture of valuing seemingly concrete and tangible gifts over practical ones. [ENDQ] [NEWLINE] Yes, that's pretty much it. To me, cash is inherently a "better" gift<mask> it's<mask> fungible. You could give me thousands of dollars<mask> a cash gift, and not only would I be able to still buy a car with it, I would<mask> be able to buy one of a make and colour I prefer. [NEWLINE] [NEWLINE] And in a context like commerce<mask> you don't really have a personal relationship, I'd rather have the cash than the concrete, tangible gift that's exactly<mask> I wanted, that only someone who knows me well could have chosen for me. I don't think I'm ignoring the tradition of gift exchange: I simply don't think it is (or: should be) relevant in commerce.</s>
Label encoding: <s> [STARTQ] I think what you're asking is the equivalent of, " why buy your son or daughter a new car for their birthday when you can gift them thousands of dollars in cash instead?" I think you're ignoring the human tradition and culture of valuing seemingly concrete and tangible gifts over practical ones. [ENDQ] [NEWLINE] Yes, that's pretty much it. To me, cash is inherently a "better" gift because it's so fungible. You could give me thousands of dollars as a cash gift, and not only would I be able to still buy a car with it, I would also be able to buy one of a make and colour I prefer. [NEWLINE] [NEWLINE] And in a context like commerce where you don't really have a personal relationship, I'd rather have the cash than the concrete, tangible gift that's exactly what I wanted, that only someone who knows me well could have chosen for me. I don't think I'm ignoring the tradition of gift exchange: I simply don't think it is (or: should be) relevant in commerce.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This. For a lot of the people of the armed forces, it's just the career path they chose. A lot of people join<mask> they know they will be able to take care of themselves and their family, and serving their counrty in the process is a nice rewarding bonus.<mask><mask> the majority of them would never ask for war. You think they want to leave their loved ones and risk not coming home, often times for wars they can't justify or know the reason for? My husband is in the US Air Force and he works 8 hours a day calibrating machinery. Would he ever ask for war? No. Does he have any say whatso ever in participating in a war? No. Neither does anyone he knows. Maybe some people just like being apart of something beneficial to others. Maybe they like to be around jets<mask> they are fascinating pieces of machinery, not<mask> they can kill people. You really should just talk to or read about some people that are actually in a military rather than assume they are all terrorists. </s>
Label encoding: <s>This. For a lot of the people of the armed forces, it's just the career path they chose. A lot of people join because they know they will be able to take care of themselves and their family, and serving their counrty in the process is a nice rewarding bonus. I think the majority of them would never ask for war. You think they want to leave their loved ones and risk not coming home, often times for wars they can't justify or know the reason for? My husband is in the US Air Force and he works 8 hours a day calibrating machinery. Would he ever ask for war? No. Does he have any say whatso ever in participating in a war? No. Neither does anyone he knows. Maybe some people just like being apart of something beneficial to others. Maybe they like to be around jets because they are fascinating pieces of machinery, not because they can kill people. You really should just talk to or read about some people that are actually in a military rather than assume they are all terrorists. </s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be fair...Turkey hasn't overthrown anyone<mask>, and a lot of people have been killed by the government. [NEWLINE] [NEWLINE] And Egypt...there was a lot of chaos, a lot of people died...and nothing happened until the military got involved, supposedly on the side of the protesters. The military was who overthrew that government. The people with guns. Did the protests motivate the military to act? Undoubtedly.<mask> it was still the military that got shit done.<mask> the military in egypt had been on the side of the corrupt government, it would probably have ended up like Syria or Libya,<mask> the only thing stopping the military from rolling over the opposition to the government was, of course, people with guns willing to become'rebels' and fight that government. [NEWLINE] [NEWLINE] And<mask><mask> you've fallen prey to some misinformation yourself. There is no culture of mass shootings. The number of mass murders in the US has stayed pretty constant at around [25 per year]( [URL] ) for at least the last hundred years. </s>
Label encoding: <s>To be fair...Turkey hasn't overthrown anyone yet, and a lot of people have been killed by the government. [NEWLINE] [NEWLINE] And Egypt...there was a lot of chaos, a lot of people died...and nothing happened until the military got involved, supposedly on the side of the protesters. The military was who overthrew that government. The people with guns. Did the protests motivate the military to act? Undoubtedly. But it was still the military that got shit done. If the military in egypt had been on the side of the corrupt government, it would probably have ended up like Syria or Libya, where the only thing stopping the military from rolling over the opposition to the government was, of course, people with guns willing to become'rebels' and fight that government. [NEWLINE] [NEWLINE] And I think you've fallen prey to some misinformation yourself. There is no culture of mass shootings. The number of mass murders in the US has stayed pretty constant at around [25 per year]( [URL] ) for at least the last hundred years. </s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But my past has nothing to do with who I am now. It is over with and done with. I changed and moved on. [NEWLINE] [NEWLINE] Assaulting women didn't make me a better person. It didn't teach me anything. I learned from my therapist that it was wrong and that I was hurting these women,<mask> I stopped. [NEWLINE] [NEWLINE] <mask> take a risk<mask> it isn't necessary?<mask> affect my relationship negatively<mask> it's not important or necessary to our relationship?<mask> she knew, she would be very upset.<mask><mask> should I make her upset<mask> it isn't necessary? [NEWLINE] [NEWLINE] She wants to know and understand all of me,<mask> that part of my past isn't part of me anymore<mask> I changed and moved on.<mask> in essence, I'm being 100% honest by sharing myself with her. I will let her make the decision based on myself and nothing<mask> myself. Not based on a certain past event that doesn't make up who I am,<mask> I changed and moved on from that. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>But my past has nothing to do with who I am now. It is over with and done with. I changed and moved on. [NEWLINE] [NEWLINE] Assaulting women didn't make me a better person. It didn't teach me anything. I learned from my therapist that it was wrong and that I was hurting these women, so I stopped. [NEWLINE] [NEWLINE] Why take a risk if it isn't necessary? Why affect my relationship negatively when it's not important or necessary to our relationship? If she knew, she would be very upset. So why should I make her upset if it isn't necessary? [NEWLINE] [NEWLINE] She wants to know and understand all of me, but that part of my past isn't part of me anymore because I changed and moved on. So in essence, I'm being 100% honest by sharing myself with her. I will let her make the decision based on myself and nothing but myself. Not based on a certain past event that doesn't make up who I am, because I changed and moved on from that. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Every-day users are more likely on average to lose motivation and become content with short term stimulus compared to achieving long-term goals. [ENDQ] [NEWLINE] anecdotal. source? i know several business owners who smoke daily. [NEWLINE] [NEWLINE] [STARTQ] Heavy use<mask> tends to lead to a certain amount of dependency [ENDQ] [NEWLINE] source? everything i have ever seen says that there is zero biological addictive properties in marijuana and thc. Addiction occurs<mask> the substance in question goes from being a component in serotonin production to the singular causal source of serotonin production. [NEWLINE] [NEWLINE] marijuana does not do this. [NEWLINE] [NEWLINE] marijuana makes people happy, it may not come with your philosophical grounding<mask> its still happiness. a brain is a bran. a brain soaked in endorphin<mask> of weed or a brain soaked in endorphin<mask> of winning a marathon are both happy. It is a value judgement to say that one is superior. [NEWLINE] I would say that this is a fallacy of believing that your happiness is some<mask> superior to others happiness<mask> its yours. </s>
Label encoding: <s> [STARTQ] Every-day users are more likely on average to lose motivation and become content with short term stimulus compared to achieving long-term goals. [ENDQ] [NEWLINE] anecdotal. source? i know several business owners who smoke daily. [NEWLINE] [NEWLINE] [STARTQ] Heavy use also tends to lead to a certain amount of dependency [ENDQ] [NEWLINE] source? everything i have ever seen says that there is zero biological addictive properties in marijuana and thc. Addiction occurs when the substance in question goes from being a component in serotonin production to the singular causal source of serotonin production. [NEWLINE] [NEWLINE] marijuana does not do this. [NEWLINE] [NEWLINE] marijuana makes people happy, it may not come with your philosophical grounding but its still happiness. a brain is a bran. a brain soaked in endorphin because of weed or a brain soaked in endorphin because of winning a marathon are both happy. It is a value judgement to say that one is superior. [NEWLINE] I would say that this is a fallacy of believing that your happiness is some how superior to others happiness because its yours. </s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think we are a long way off having a 'casino-industrial complex' like the US and China have<mask> there is entire cities whose economies are built upon gambling revenues. At the same time<mask><mask> we lag behind more progressive nations like Finland<mask> they have 1 Euro maximum bets, 50 euro maximum wins and no chairs at the pokies. Anecdotally I'd say these three measures make it physically tiring and extremely time consuming to be a problem gambler in Finland<mask> the fact they have pokies in the supermarket. You just cant gamble fast enough to lose or win a significant amount of money and the revenues are petty meaning there isnt enough money for a casino industrial complex to form. [NEWLINE] [NEWLINE] Australia is sort of in the middle, we dont have a big problem with gambling<mask> we<mask> arent moving forward with the kind of policies that evidence proves could crush the smallish problem we do have. Australian anti-gambling policy lacks teeth<mask> no one cares<mask> its not really a big problem anyway. [NEWLINE] </s>
Label encoding: <s>I think we are a long way off having a 'casino-industrial complex' like the US and China have where there is entire cities whose economies are built upon gambling revenues. At the same time I think we lag behind more progressive nations like Finland where they have 1 Euro maximum bets, 50 euro maximum wins and no chairs at the pokies. Anecdotally I'd say these three measures make it physically tiring and extremely time consuming to be a problem gambler in Finland despite the fact they have pokies in the supermarket. You just cant gamble fast enough to lose or win a significant amount of money and the revenues are petty meaning there isnt enough money for a casino industrial complex to form. [NEWLINE] [NEWLINE] Australia is sort of in the middle, we dont have a big problem with gambling but we also arent moving forward with the kind of policies that evidence proves could crush the smallish problem we do have. Australian anti-gambling policy lacks teeth but no one cares because its not really a big problem anyway. [NEWLINE] </s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>hmm... well, I was brought up that chopsticks were the correct way to eat certain food. Thats<mask> i've always ate sushi or Chinese food. Thats<mask> I eat it at home<mask> no one is there to watch,<mask> I doubt that I'm doing it to show off. [NEWLINE] [NEWLINE] Generally,<mask><mask> of chopsticks<mask> just another utensil. Sure, there are different ways that food could be eaten,<mask> the same holds true to most food. Some people drink soup out of a bowl, some people use a spoon - is the person using the spoon showing off? Arguably, it would be easier to pick up the bowl and drink it,<mask> people tend to do<mask> is habit for them. [NEWLINE] [NEWLINE] <mask>, the "industrial fast food" is dependent on<mask> you are buying it.<mask> you are buying it from a fast-food restaurant, then yes, it is.<mask> you can't make that statement across the board<mask> food varies<mask> much from restaurant to restaurant. </s>
Label encoding: <s>hmm... well, I was brought up that chopsticks were the correct way to eat certain food. Thats how i've always ate sushi or Chinese food. Thats how I eat it at home when no one is there to watch, so I doubt that I'm doing it to show off. [NEWLINE] [NEWLINE] Generally, I think of chopsticks as just another utensil. Sure, there are different ways that food could be eaten, but the same holds true to most food. Some people drink soup out of a bowl, some people use a spoon - is the person using the spoon showing off? Arguably, it would be easier to pick up the bowl and drink it, but people tend to do what is habit for them. [NEWLINE] [NEWLINE] Also, the "industrial fast food" is dependent on where you are buying it. If you are buying it from a fast-food restaurant, then yes, it is. But you can't make that statement across the board because food varies so much from restaurant to restaurant. </s>
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Young people voting and us not having more policies generated on the onions of young people isn't supported by the premise that<mask> some young people vote and we don't have many policies generated based on their opinions means your vote doesn't count. [NEWLINE] None of that matches up at all. [NEWLINE] Apathy in general is a coping mechanism,<mask><mask> coping mechanisms mean by definition we are lacking a clearer solution most people want to get involved in. That means apathy is not good, it's not even a solution, it's the lack of a solution. [NEWLINE] <mask> /u/Neckbeard_The_Great pointed out the best solution would be everyone becoming educated on the issues<mask> the direction our country is headed given the exact balance of laws we have on the books becomes more well known<mask> elected officials have a more challenging job<mask> they don't get to only respond towards changing times<mask> the majority of 40+ people want things to go alone. [NEWLINE] Apathy is a bad thing<mask> it implies it's a solution and it's not.</s>
Label encoding: <s>Young people voting and us not having more policies generated on the onions of young people isn't supported by the premise that since some young people vote and we don't have many policies generated based on their opinions means your vote doesn't count. [NEWLINE] None of that matches up at all. [NEWLINE] Apathy in general is a coping mechanism, so since coping mechanisms mean by definition we are lacking a clearer solution most people want to get involved in. That means apathy is not good, it's not even a solution, it's the lack of a solution. [NEWLINE] As /u/Neckbeard_The_Great pointed out the best solution would be everyone becoming educated on the issues so the direction our country is headed given the exact balance of laws we have on the books becomes more well known so elected officials have a more challenging job because they don't get to only respond towards changing times however the majority of 40+ people want things to go alone. [NEWLINE] Apathy is a bad thing because it implies it's a solution and it's not.</s>
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry,<mask> this is a rather naïve view, unfortunately. [NEWLINE] [NEWLINE] Most people who think Google and Facebook can be a problem in the future couldn't care less about e-mails, search engines and number of employees. It's all in the *amount* of data handled, and in<mask> *detailed* it is. Your entire life is essentially filtered through Android. [NEWLINE] [NEWLINE] <mask><mask>, some academics (notable, not fringe, mind you), think the world is shifting towards a new form of governing<mask> of companies like Google and Facebook. They're simply far to effective at collecting and handling data in comparison with governments. Google and Facebook knows<mask> you need, the government doesn't. [NEWLINE] [NEWLINE] This is seen<mask> a democratic problem,<mask> Google and Facebook can, in theory, shift focus to a very individualistic mode of thinking, in effect turning democratically elected systems into less viable solutions to the problem of governing. [NEWLINE] [NEWLINE] You're stuck 50 years in the past. Money isn't power, information is.</s>
Label encoding: <s>I'm sorry, but this is a rather naïve view, unfortunately. [NEWLINE] [NEWLINE] Most people who think Google and Facebook can be a problem in the future couldn't care less about e-mails, search engines and number of employees. It's all in the *amount* of data handled, and in how *detailed* it is. Your entire life is essentially filtered through Android. [NEWLINE] [NEWLINE] In fact, some academics (notable, not fringe, mind you), think the world is shifting towards a new form of governing because of companies like Google and Facebook. They're simply far to effective at collecting and handling data in comparison with governments. Google and Facebook knows what you need, the government doesn't. [NEWLINE] [NEWLINE] This is seen as a democratic problem, because Google and Facebook can, in theory, shift focus to a very individualistic mode of thinking, in effect turning democratically elected systems into less viable solutions to the problem of governing. [NEWLINE] [NEWLINE] You're stuck 50 years in the past. Money isn't power, information is.</s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In context, this verse reads: [NEWLINE] [NEWLINE] [STARTQ] **23 Now before faith came, we were held captive under the law, imprisoned until the coming faith would be revealed. 24<mask> then, the law was our guardian until Christ came, in order that we might be justified by faith. 25<mask> now that faith has come, we are no longer under a guardian,** 26 for in Christ Jesus you are all sons of God, through faith. 27 For<mask> many of you<mask> were baptized into Christ have put on Christ. 28 There is neither Jew nor Greek, there is neither slave[g] nor free, there is no male and female, for you are all one in Christ Jesus. 29 And<mask> you are Christ's, then you are Abraham's offspring, heirs<mask><mask> promise. [ENDQ] [NEWLINE] This paragraph talks about freedom from legalism, not societal rules. Jewish law addresses sexuality,<mask> it doesn't really go into trans-genderism. This verse,<mask>, is not entirely relevant to this specific topic.</s>
Label encoding: <s>In context, this verse reads: [NEWLINE] [NEWLINE] [STARTQ] **23 Now before faith came, we were held captive under the law, imprisoned until the coming faith would be revealed. 24 So then, the law was our guardian until Christ came, in order that we might be justified by faith. 25 But now that faith has come, we are no longer under a guardian,** 26 for in Christ Jesus you are all sons of God, through faith. 27 For as many of you as were baptized into Christ have put on Christ. 28 There is neither Jew nor Greek, there is neither slave[g] nor free, there is no male and female, for you are all one in Christ Jesus. 29 And if you are Christ's, then you are Abraham's offspring, heirs according to promise. [ENDQ] [NEWLINE] This paragraph talks about freedom from legalism, not societal rules. Jewish law addresses sexuality, but it doesn't really go into trans-genderism. This verse, therefore, is not entirely relevant to this specific topic.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That's<mask> you end up with protesters being gassed needlessly [ENDQ] [NEWLINE] You can't use gas in war. The gas they used is designed to look scary, it doesn't do much more than make you tear up and make your nose runny. [NEWLINE] [NEWLINE] [STARTQ] potrolmen aiming automatic weapons at citizens for the purpose of intimidation. [ENDQ] [NEWLINE] That rarely ever happens<mask> ever. Any cop that would do that wouldn't have a job for much longer. Yes cops get fired [NEWLINE] [NEWLINE] [STARTQ] About 80% of SWAT raids aren't even for arrests, they're for executing search warrants. [ENDQ] [NEWLINE] Yes that is<mask> you get evidence on dangerous people. This usually leads to arrest. You break down the door and the guy is bagging a kilo of cocaine and you arrest him or her [NEWLINE] [NEWLINE] [NEWLINE] <mask>, the gangsters have big guns,<mask> shouldn't the cops,<mask> anyone is driving a tank or having a machine gun I prefer it be the cops over the gangsters with m-16s </s>
Label encoding: <s> [STARTQ] That's how you end up with protesters being gassed needlessly [ENDQ] [NEWLINE] You can't use gas in war. The gas they used is designed to look scary, it doesn't do much more than make you tear up and make your nose runny. [NEWLINE] [NEWLINE] [STARTQ] potrolmen aiming automatic weapons at citizens for the purpose of intimidation. [ENDQ] [NEWLINE] That rarely ever happens if ever. Any cop that would do that wouldn't have a job for much longer. Yes cops get fired [NEWLINE] [NEWLINE] [STARTQ] About 80% of SWAT raids aren't even for arrests, they're for executing search warrants. [ENDQ] [NEWLINE] Yes that is how you get evidence on dangerous people. This usually leads to arrest. You break down the door and the guy is bagging a kilo of cocaine and you arrest him or her [NEWLINE] [NEWLINE] [NEWLINE] Lastly, the gangsters have big guns, why shouldn't the cops, if anyone is driving a tank or having a machine gun I prefer it be the cops over the gangsters with m-16s </s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I didn't think his characters were<mask> contrived<mask> you seem to think they are. I know<mask> I was a teenager I went around saying stuff like "some infinities are bigger than other infinities" or arguing with other kids about whether 0.999... is really exactly equal to 1 or not. [NEWLINE] [NEWLINE] <mask>, you know, at least the main character of Looking for Alaska is clearly supposed to be painfully awkward. I don't know about his other books<mask><mask> they're<mask> similar<mask> you claim I wouldn't be surprised<mask> his other main characters were<mask> supposed to be painfully awkward. [NEWLINE] [NEWLINE] [STARTQ] You can have complex metaphors without one of your characters saying, 'by the way,<mask> I'm doing right now is a metaphor and a recurring motif throughout the novel kbye'. [ENDQ] [NEWLINE] John Green has said that just<mask> one of his characters thinks something is a metaphor doesn't mean it's actually a metaphor, or at least that it's not always the metaphor they want it to be.</s>
Label encoding: <s>I didn't think his characters were as contrived as you seem to think they are. I know when I was a teenager I went around saying stuff like "some infinities are bigger than other infinities" or arguing with other kids about whether 0.999... is really exactly equal to 1 or not. [NEWLINE] [NEWLINE] Also, you know, at least the main character of Looking for Alaska is clearly supposed to be painfully awkward. I don't know about his other books but if they're as similar as you claim I wouldn't be surprised if his other main characters were also supposed to be painfully awkward. [NEWLINE] [NEWLINE] [STARTQ] You can have complex metaphors without one of your characters saying, 'by the way, what I'm doing right now is a metaphor and a recurring motif throughout the novel kbye'. [ENDQ] [NEWLINE] John Green has said that just because one of his characters thinks something is a metaphor doesn't mean it's actually a metaphor, or at least that it's not always the metaphor they want it to be.</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's funny that none of the rebuttals here are refuting the holes in the story. They are all just questioning the motive for the government to either perpetuate the attacks or allow the attacks to happen. The insider trading is interesting,<mask> I bet it is more a symptom than the reason. Someone in the know figured they would quietly make some money on the whole ordeal. Simple. The state needed a BIG attack that they could convince the public was perpetrated by terrorists that live in caves. I don't see any issue with motive. States do expensive, destructive, evil things every day. [NEWLINE] [NEWLINE] <mask><mask><mask> people saying there is no way they could have kept it secret, I'd point to the Snowden revelations. These are highly secret spy programs that have existed for decades, involve thousands of people, and span through international governments. We would still be oblivious<mask> it wasn't for Snowden. Sure, the conspiracy theorists pieced together many things that he revealed,<mask> without hard evidence, few normal people were convinced. </s>
Label encoding: <s>It's funny that none of the rebuttals here are refuting the holes in the story. They are all just questioning the motive for the government to either perpetuate the attacks or allow the attacks to happen. The insider trading is interesting, but I bet it is more a symptom than the reason. Someone in the know figured they would quietly make some money on the whole ordeal. Simple. The state needed a BIG attack that they could convince the public was perpetrated by terrorists that live in caves. I don't see any issue with motive. States do expensive, destructive, evil things every day. [NEWLINE] [NEWLINE] As far as people saying there is no way they could have kept it secret, I'd point to the Snowden revelations. These are highly secret spy programs that have existed for decades, involve thousands of people, and span through international governments. We would still be oblivious if it wasn't for Snowden. Sure, the conspiracy theorists pieced together many things that he revealed, but without hard evidence, few normal people were convinced. </s>
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a nonsensical strawman point, the sort of thing I'd expect from a rant against "SJWs" or whoever it is people hate this week. [NEWLINE] [NEWLINE] I'm one of the people who would mention the stats,<mask><mask> proof of problems such<mask> poverty and institutional racism etc. (the type of person /u/Doppleganger07 mentioned), and people do that all the time and it is easy to recognise, and always recognised, on this website that the person doing that obviously isn't being racist. Usually<mask> they're countering actual racists who are deliberately misusing statistics to promote racism. [NEWLINE] [NEWLINE] You're creating a non-existent problem; people who mention crime stats involving race are not going to get banned universally,<mask> the difference between the following people... [NEWLINE] [NEWLINE] A. Doing<mask> I and others do which is have a grown up conversation and not use cherry picked stats to hate black people [NEWLINE] [NEWLINE] B. Being an intellectually dishonest racist idiot. [NEWLINE] [NEWLINE]...is usually very obvious. </s>
Label encoding: <s>This is a nonsensical strawman point, the sort of thing I'd expect from a rant against "SJWs" or whoever it is people hate this week. [NEWLINE] [NEWLINE] I'm one of the people who would mention the stats, but as proof of problems such as poverty and institutional racism etc. (the type of person /u/Doppleganger07 mentioned), and people do that all the time and it is easy to recognise, and always recognised, on this website that the person doing that obviously isn't being racist. Usually because they're countering actual racists who are deliberately misusing statistics to promote racism. [NEWLINE] [NEWLINE] You're creating a non-existent problem; people who mention crime stats involving race are not going to get banned universally, because the difference between the following people... [NEWLINE] [NEWLINE] A. Doing what I and others do which is have a grown up conversation and not use cherry picked stats to hate black people [NEWLINE] [NEWLINE] B. Being an intellectually dishonest racist idiot. [NEWLINE] [NEWLINE]...is usually very obvious. </s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll keep this succinct to try and suppress your mental gymnastics. [NEWLINE] [NEWLINE] Cancer - a one off incident that can affect anyone at any time, people will do anything to fix it<mask> it happens to them. [NEWLINE] [NEWLINE] Overeating - people over a long period of time observing themselves getting larger and larger without taking any corrective action until it becomes an immediate life threat or is needed for other life saving surgery. [NEWLINE] [NEWLINE] Living healthy doesn't need to be expensive,<mask><mask> more education is required,<mask><mask> you have fatty foods you need less of them to survive. Poor, fat people piss me off the most with their "woe is me, I can barely afford 5 times my recommended calorie intake". Here's an idea,<mask> about you eat less, then you will have more money and less fat! [NEWLINE] [NEWLINE] I've been well outside my recommended bmi,<mask> I thought about<mask> I was eating, hey presto I'm the right size again 6 months later. It's not difficult. </s>
Label encoding: <s>I'll keep this succinct to try and suppress your mental gymnastics. [NEWLINE] [NEWLINE] Cancer - a one off incident that can affect anyone at any time, people will do anything to fix it if it happens to them. [NEWLINE] [NEWLINE] Overeating - people over a long period of time observing themselves getting larger and larger without taking any corrective action until it becomes an immediate life threat or is needed for other life saving surgery. [NEWLINE] [NEWLINE] Living healthy doesn't need to be expensive, I agree more education is required, but if you have fatty foods you need less of them to survive. Poor, fat people piss me off the most with their "woe is me, I can barely afford 5 times my recommended calorie intake". Here's an idea, how about you eat less, then you will have more money and less fat! [NEWLINE] [NEWLINE] I've been well outside my recommended bmi, so I thought about what I was eating, hey presto I'm the right size again 6 months later. It's not difficult. </s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Believe it or not a restaurant is a business. Businesses love getting information from their customers, it means they can improve the quality of their service/product and indirectly increase their profits. [NEWLINE] [NEWLINE] There is always a risk someone might take offense<mask> you can minimize it to negligible by learning to word constructive criticism. [NEWLINE] [NEWLINE] 1: People respond better to positive confrontations. Sandwich the criticism between 2 compliments. [NEWLINE] [NEWLINE] [STARTQ] I like the rice, too much salt in the soup, loved the pudding. [ENDQ] [NEWLINE] 2: It is more difficult for people to disagree with facts than opinions. Simply state the facts that led to your opinion. [NEWLINE] [NEWLINE] [STARTQ] Ok it says the soup is 5% salt, you can taste salt at 1%.<mask>... [ENDQ] [NEWLINE] 3: Acknowledge that you don't know everything and you're just throwing it out there. [NEWLINE] [NEWLINE] &gt; I don't know whether the ingredients of the soup have salt in them or anything or salt was added<mask> it was a bit salty.</s>
Label encoding: <s>Believe it or not a restaurant is a business. Businesses love getting information from their customers, it means they can improve the quality of their service/product and indirectly increase their profits. [NEWLINE] [NEWLINE] There is always a risk someone might take offense but you can minimize it to negligible by learning to word constructive criticism. [NEWLINE] [NEWLINE] 1: People respond better to positive confrontations. Sandwich the criticism between 2 compliments. [NEWLINE] [NEWLINE] [STARTQ] I like the rice, too much salt in the soup, loved the pudding. [ENDQ] [NEWLINE] 2: It is more difficult for people to disagree with facts than opinions. Simply state the facts that led to your opinion. [NEWLINE] [NEWLINE] [STARTQ] Ok it says the soup is 5% salt, you can taste salt at 1%. So... [ENDQ] [NEWLINE] 3: Acknowledge that you don't know everything and you're just throwing it out there. [NEWLINE] [NEWLINE] &gt; I don't know whether the ingredients of the soup have salt in them or anything or salt was added but it was a bit salty.</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's a less nefarious and more convincing (to me at least) explanation of those "political" 5-4 decisions: [NEWLINE] [NEWLINE] Democrats appoint judges who truly believe in a jurisprudential theory that leads to "Democratic" outcomes. Republicans do the opposite. This helps both parties get<mask> they want from the Court and helps give their arguments legitimacy (a judge who simply threw out any theory to support his side would be more susceptible to exposure/criticism). [NEWLINE] [NEWLINE] It's not a coincidence that all of the originalists on the Court were appointed by the GOP --<mask> originalist theory supports an old vision of America that's more libertarian, less regulatory, and less centralized than the state we have today. Those judges were appointed<mask> they believe in a theory that leads to "Republican" outcomes -- not<mask> they were<mask> Republican that they are willing adopt any theory to further the cause. [NEWLINE] [NEWLINE] I'm sure I could have phrased that more clearly,<mask> do you see the gist of the argument?</s>
Label encoding: <s>Here's a less nefarious and more convincing (to me at least) explanation of those "political" 5-4 decisions: [NEWLINE] [NEWLINE] Democrats appoint judges who truly believe in a jurisprudential theory that leads to "Democratic" outcomes. Republicans do the opposite. This helps both parties get what they want from the Court and helps give their arguments legitimacy (a judge who simply threw out any theory to support his side would be more susceptible to exposure/criticism). [NEWLINE] [NEWLINE] It's not a coincidence that all of the originalists on the Court were appointed by the GOP -- because originalist theory supports an old vision of America that's more libertarian, less regulatory, and less centralized than the state we have today. Those judges were appointed because they believe in a theory that leads to "Republican" outcomes -- not because they were so Republican that they are willing adopt any theory to further the cause. [NEWLINE] [NEWLINE] I'm sure I could have phrased that more clearly, but do you see the gist of the argument?</s>
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The labor is intrinsic to the product, or else there wouldn't be a product. I'm referring to the labor the person who made it put into it, which exists in the product, and is paid for in the final price. [NEWLINE] [NEWLINE] That's there's no labor to copy it just means there's no added labor to that from the creator that exists in the product already. [NEWLINE] [NEWLINE] <mask> there was no value in the art, or the labor of it, then the creators would not be paid for their crafting the product. [NEWLINE] [NEWLINE] P=M+L,<mask> M is physical materials, P is the product itself or,<mask> you'd rather, the set price, and L is the labor. Take away the material through the process of copying, and you have P = -M + L, or P-M=L. The labor is still left over,<mask> it can't be removed from the product.<mask> you take away L, too, then you don't get the product.</s>
Label encoding: <s>The labor is intrinsic to the product, or else there wouldn't be a product. I'm referring to the labor the person who made it put into it, which exists in the product, and is paid for in the final price. [NEWLINE] [NEWLINE] That's there's no labor to copy it just means there's no added labor to that from the creator that exists in the product already. [NEWLINE] [NEWLINE] If there was no value in the art, or the labor of it, then the creators would not be paid for their crafting the product. [NEWLINE] [NEWLINE] P=M+L, where M is physical materials, P is the product itself or, if you'd rather, the set price, and L is the labor. Take away the material through the process of copying, and you have P = -M + L, or P-M=L. The labor is still left over, so it can't be removed from the product. If you take away L, too, then you don't get the product.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Anthropomorphism is assigning human characteristics to nonhuman entities. [ENDQ] [NEWLINE] Did you know that cats are nonhumans? [NEWLINE] [NEWLINE] [STARTQ] It is not comparing the treatment of humans with the treatment of nonhuman entities. [ENDQ] [NEWLINE] It certainly is<mask> people are making assumptions about a cat's being "happy" under various circumstances. [NEWLINE] [NEWLINE] [STARTQ] Taken to its extreme, this would be like calling the assertion "You wouldn't want to be set on fire,<mask> you shouldn't set your dog on fire" anthropomorphism. [ENDQ] [NEWLINE] Except that argument by analogy is almost always fallacious. "You wouldn't want to eat off the floor or walk naked in public,<mask> you shouldn't do that to your dog." [NEWLINE] [NEWLINE] The anthropomorphism here is speaking of cats being happy. There is no evidence that cats can experience the emotion of happiness. Even<mask> this could be established, you have to make the argument that a cat's happiness should be prioritized to the exclusion of all other considerations.</s>
Label encoding: <s> [STARTQ] Anthropomorphism is assigning human characteristics to nonhuman entities. [ENDQ] [NEWLINE] Did you know that cats are nonhumans? [NEWLINE] [NEWLINE] [STARTQ] It is not comparing the treatment of humans with the treatment of nonhuman entities. [ENDQ] [NEWLINE] It certainly is when people are making assumptions about a cat's being "happy" under various circumstances. [NEWLINE] [NEWLINE] [STARTQ] Taken to its extreme, this would be like calling the assertion "You wouldn't want to be set on fire, so you shouldn't set your dog on fire" anthropomorphism. [ENDQ] [NEWLINE] Except that argument by analogy is almost always fallacious. "You wouldn't want to eat off the floor or walk naked in public, so you shouldn't do that to your dog." [NEWLINE] [NEWLINE] The anthropomorphism here is speaking of cats being happy. There is no evidence that cats can experience the emotion of happiness. Even if this could be established, you have to make the argument that a cat's happiness should be prioritized to the exclusion of all other considerations.</s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> you liked this person and genuinely wanted to be with them, something this stupid wouldn't bother you. [ENDQ] [NEWLINE] This simply isn't true.  These things can bug you AND you still want to be with them. <mask>, you compromise and either talk to them about it or just learn to ignore.  No, neither of you deserve the relationship or deserve the change from the other,<mask> that doesn't mean that you can't choose to make the change. [NEWLINE] [NEWLINE] <mask> i read your posts I see a dude with a mental check list that he runs down.  And<mask> the girl doesn't meet every criteria than he just leaves. No questions asked, no talking about anything, just leaves.  And<mask> that's okay and your decision, it doesn't mean it is good. [NEWLINE] [NEWLINE] There are lot of things that people are allowed to do/capable of doing. <mask> just becasue you can do something doesn't mean it is the best thing to do.</s>
Label encoding: <s> [STARTQ] Because if you liked this person and genuinely wanted to be with them, something this stupid wouldn't bother you. [ENDQ] [NEWLINE] This simply isn't true.  These things can bug you AND you still want to be with them.  So, you compromise and either talk to them about it or just learn to ignore.  No, neither of you deserve the relationship or deserve the change from the other, but that doesn't mean that you can't choose to make the change. [NEWLINE] [NEWLINE] When i read your posts I see a dude with a mental check list that he runs down.  And if the girl doesn't meet every criteria than he just leaves. No questions asked, no talking about anything, just leaves.  And while that's okay and your decision, it doesn't mean it is good. [NEWLINE] [NEWLINE] There are lot of things that people are allowed to do/capable of doing.  But just becasue you can do something doesn't mean it is the best thing to do.</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't believe that a velociraptor's brain is equipped to deal with confrontations that involve more than simple pursuit or evasion. I believe they are fundamentally incapable of dealing with a dynamic combat event, and<mask> could be relatively easily subdued by strategic fighting and basic wrestling. [ENDQ] [NEWLINE] <mask> a predator, velocirapotors would fight in dynamic situations to kill their prey. I challenge you to fight a grey wolf, a predator that would be fairly similar in size to the type potrayed in jurassic park. I highly doubt you could beat a wolf or even a german shepard in a fight unarmed,<mask> it is far stronger than you nad will bite you to death before you could choke it. you have obviously never fought a wild animal before,<mask> they will roll, twist, bite and kick to get away and kill you.<mask> you can defeat a common wolf or even a coyote, than you might consider. you are woefully underestimating wild animals</s>
Label encoding: <s> [STARTQ] I don't believe that a velociraptor's brain is equipped to deal with confrontations that involve more than simple pursuit or evasion. I believe they are fundamentally incapable of dealing with a dynamic combat event, and thus could be relatively easily subdued by strategic fighting and basic wrestling. [ENDQ] [NEWLINE] As a predator, velocirapotors would fight in dynamic situations to kill their prey. I challenge you to fight a grey wolf, a predator that would be fairly similar in size to the type potrayed in jurassic park. I highly doubt you could beat a wolf or even a german shepard in a fight unarmed, as it is far stronger than you nad will bite you to death before you could choke it. you have obviously never fought a wild animal before, as they will roll, twist, bite and kick to get away and kill you. if you can defeat a common wolf or even a coyote, than you might consider. you are woefully underestimating wild animals</s>
Loss: tensor(0.1245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I used to tell women I wasn't interested in them<mask> the first date or two didn't pan out.  At the time I was meeting women online, and that made it hard to determine<mask> there was any real-life chemistry.  In most cases, the women outright lied on their profiles. <mask> I found out the "real" person, things ended, and I would tell them I just didn't think it was going to work out. [NEWLINE] [NEWLINE] I was amazed<mask> upset they got.  I will always remember this one woman was yelling at me, insulting me, calling me every name under the book.  She drove over to my house to tell me off, nearly a 45 minute drive.  After having stuff like this happen to me 3 times in a row I simply stopped trying to be open and honest.  Some people don't deserve it, and I don't feel like I should be treated<mask> poorly. [NEWLINE] [NEWLINE] <mask> I did learn was to stop meeting women online.</s>
Label encoding: <s>I used to tell women I wasn't interested in them if the first date or two didn't pan out.  At the time I was meeting women online, and that made it hard to determine if there was any real-life chemistry.  In most cases, the women outright lied on their profiles.  When I found out the "real" person, things ended, and I would tell them I just didn't think it was going to work out. [NEWLINE] [NEWLINE] I was amazed how upset they got.  I will always remember this one woman was yelling at me, insulting me, calling me every name under the book.  She drove over to my house to tell me off, nearly a 45 minute drive.  After having stuff like this happen to me 3 times in a row I simply stopped trying to be open and honest.  Some people don't deserve it, and I don't feel like I should be treated so poorly. [NEWLINE] [NEWLINE] What I did learn was to stop meeting women online.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> the best analogies for my case is the fat guy that keeps going to the gym,<mask> can't drop a pound,<mask> he always "rewards himself with a burger / fries / cake" [ENDQ] [NEWLINE] Tell me something,<mask> does he buy those things?<mask> happens to the short-order cook<mask> the fat guy can't afford a burger?<mask><mask> there's 10 fat guys that can no longer afford to eat there? [NEWLINE] [NEWLINE] Welfare money doesn't stop<mask> it gets to the recipient's pocket, *it gets spent*. The best way to get out of poverty is to get a job, right?<mask><mask> does one go about getting a job<mask> there are no customers to support employers? [NEWLINE] [NEWLINE] There is no economic argument to be made here<mask> people with no money are excluded from the economy. It doesn't matter<mask> they are spending their money on,<mask><mask><mask> they are spending it. Welfare isn't a drain on the economy, it props it up.</s>
Label encoding: <s> [STARTQ] I think the best analogies for my case is the fat guy that keeps going to the gym, but can't drop a pound, because he always "rewards himself with a burger / fries / cake" [ENDQ] [NEWLINE] Tell me something, where does he buy those things? What happens to the short-order cook when the fat guy can't afford a burger? What if there's 10 fat guys that can no longer afford to eat there? [NEWLINE] [NEWLINE] Welfare money doesn't stop when it gets to the recipient's pocket, *it gets spent*. The best way to get out of poverty is to get a job, right? But how does one go about getting a job when there are no customers to support employers? [NEWLINE] [NEWLINE] There is no economic argument to be made here because people with no money are excluded from the economy. It doesn't matter what they are spending their money on, so long as they are spending it. Welfare isn't a drain on the economy, it props it up.</s>
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I've chosen to decide that God is<mask> above us and unknowable that we cannot know or understand his reasoning. We just have to trust that he know<mask>'s best. [ENDQ] [NEWLINE] <mask> this is the case,<mask> is it<mask> far fetched that the bible isn't literally true<mask>'spiritually' true, meaning<mask> literally true<mask> the writer can comprehend,<mask><mask> the writer is receiving knowledge from such God. [NEWLINE] [NEWLINE] Stating that the bible is literally true is it's own flavor of hubris, I've always thought.  Literal truth would imply that God,<mask> divinely communicating knowledge to his vessels, did<mask> in a way that was totally coherent.  Is God usually like that, perfectly coherent all the time?  Does it make sense that he would be?  Would it, at least, be *plausible* that one of the writers of the bible was simply dealing with jumbled visions of a profound truth, and used the words they thought were best?</s>
Label encoding: <s> [STARTQ] I've chosen to decide that God is so above us and unknowable that we cannot know or understand his reasoning. We just have to trust that he know what's best. [ENDQ] [NEWLINE] If this is the case, why is it so far fetched that the bible isn't literally true but'spiritually' true, meaning as literally true as the writer can comprehend, given that the writer is receiving knowledge from such God. [NEWLINE] [NEWLINE] Stating that the bible is literally true is it's own flavor of hubris, I've always thought.  Literal truth would imply that God, when divinely communicating knowledge to his vessels, did so in a way that was totally coherent.  Is God usually like that, perfectly coherent all the time?  Does it make sense that he would be?  Would it, at least, be *plausible* that one of the writers of the bible was simply dealing with jumbled visions of a profound truth, and used the words they thought were best?</s>
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason I reacted to the comment is that<mask><mask> that it, and comments like it, are used to smother and invalidate discussion of false accusations of rape.  And<mask><mask> that's a dangerous road.  It's like the 'check your privilege' meme.  It's designed to silence views, not based on their validity,<mask> on the validity of the speaker.  I interpreted her comment<mask> 'thought policing'.  It wasn't addressing anything I'd said.  Only the things she assumed I must be thinking based on her projection onto<mask> I hadn't said. [NEWLINE] [NEWLINE] Not sure<mask> that's making sense.  I don't tend to invoke the ire of SJWs often (I hate that term too! I just couldn't think of a better one right now)<mask> there's lots of lower hanging fruit out there. <mask><mask> i see those kinds of comments<mask><mask> the things that I said here. <mask> I guess it was time to say them.</s>
Label encoding: <s>The reason I reacted to the comment is that I think that it, and comments like it, are used to smother and invalidate discussion of false accusations of rape.  And I think that's a dangerous road.  It's like the 'check your privilege' meme.  It's designed to silence views, not based on their validity, but on the validity of the speaker.  I interpreted her comment as 'thought policing'.  It wasn't addressing anything I'd said.  Only the things she assumed I must be thinking based on her projection onto what I hadn't said. [NEWLINE] [NEWLINE] Not sure if that's making sense.  I don't tend to invoke the ire of SJWs often (I hate that term too! I just couldn't think of a better one right now) as there's lots of lower hanging fruit out there.  But when i see those kinds of comments i think the things that I said here.  So I guess it was time to say them.</s>
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Have you considered that liberals get their news from sources other than TV news networks? Do you know<mask> conservative newspapers get more readers than liberal newspapers, or<mask> conservative websites get more hits than liberal websites, or<mask> conservative radio shows get more listeners than liberal radio shows? I don't know the answers to these questions,<mask> you can assume that liberals get their news from internet sources more than TV or radio sources<mask> liberals tend to be younger and younger people tend to get their news more from the internet than anywhere else,<mask> older people (who tend to be more conservative) might get their news from TV or radio sources<mask> this is<mask> they have always had for their news sources.<mask>, are the ratings of all other news stations added up more or less than Fox News,<mask> liberals might simply watch different news programs<mask> conservatives might just stick to Fox News,<mask> you have one group of people using one source<mask> another group of people use different sources, meaning the one source will have more viewers.</s>
Label encoding: <s>Have you considered that liberals get their news from sources other than TV news networks? Do you know if conservative newspapers get more readers than liberal newspapers, or if conservative websites get more hits than liberal websites, or if conservative radio shows get more listeners than liberal radio shows? I don't know the answers to these questions, but you can assume that liberals get their news from internet sources more than TV or radio sources because liberals tend to be younger and younger people tend to get their news more from the internet than anywhere else, while older people (who tend to be more conservative) might get their news from TV or radio sources since this is what they have always had for their news sources. Also, are the ratings of all other news stations added up more or less than Fox News, because liberals might simply watch different news programs while conservatives might just stick to Fox News, so you have one group of people using one source while another group of people use different sources, meaning the one source will have more viewers.</s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The police are not law enforcers,<mask> soldiers hired to capture and punish people who break the law. <mask> their only job is to capture and punish, they are going to use whatever means that is necessary to achieve that end. [NEWLINE] [NEWLINE] <mask> they only care about getting a conviction in order to fulfill capture, whatever they do is going to be slanted toward that end. Maybe it's time to change their job back to law enforcement: causing people to stay within the law, instead of using them<mask> capturers.  This would eliminate the conflict of interest that led to their lying to get a conviction in the first place. [NEWLINE] [NEWLINE] Questioning someone with malicious intent is an attack upon them that will often cause them damage.<mask>,<mask> we have is a state-sponsored machine designed to abuse anyone accused of a crime.  That's the real reason people don't trust the police.  I don't think asking them not to lie is going to change that. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>The police are not law enforcers, but soldiers hired to capture and punish people who break the law.  Since their only job is to capture and punish, they are going to use whatever means that is necessary to achieve that end. [NEWLINE] [NEWLINE] If they only care about getting a conviction in order to fulfill capture, whatever they do is going to be slanted toward that end. Maybe it's time to change their job back to law enforcement: causing people to stay within the law, instead of using them as capturers.  This would eliminate the conflict of interest that led to their lying to get a conviction in the first place. [NEWLINE] [NEWLINE] Questioning someone with malicious intent is an attack upon them that will often cause them damage. So, what we have is a state-sponsored machine designed to abuse anyone accused of a crime.  That's the real reason people don't trust the police.  I don't think asking them not to lie is going to change that. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's pretty interesting, after I finish up responding to my inbox<mask><mask> I'll take the next hour or<mask> reading through that. [NEWLINE] [NEWLINE] A lot of people here are mentioning canned responses like it's a bad thing.<mask> someone asks the same question that has been asked and answered a million times, I don't see<mask> it couldn't be answered through a canned response. [NEWLINE] [NEWLINE] <mask><mask> it comes to choosing the questions they want to answer, I feel like this is mutually beneficial for the majority of the people. [NEWLINE] [NEWLINE] It's in the best interest of the candidate to acquire<mask> many votes<mask> possible.<mask>, they would have a better incentive to want to answer the questions which are relatable to the majority of the population. [NEWLINE] [NEWLINE] It would be impossible to get to everyone's questions through this method,<mask><mask> a candidate can answer questions that the majority of the population holds, I feel like they've done the most good they could with the resources they have.</s>
Label encoding: <s>That's pretty interesting, after I finish up responding to my inbox I think I'll take the next hour or so reading through that. [NEWLINE] [NEWLINE] A lot of people here are mentioning canned responses like it's a bad thing. If someone asks the same question that has been asked and answered a million times, I don't see why it couldn't be answered through a canned response. [NEWLINE] [NEWLINE] Also when it comes to choosing the questions they want to answer, I feel like this is mutually beneficial for the majority of the people. [NEWLINE] [NEWLINE] It's in the best interest of the candidate to acquire as many votes as possible. Therefore, they would have a better incentive to want to answer the questions which are relatable to the majority of the population. [NEWLINE] [NEWLINE] It would be impossible to get to everyone's questions through this method, but if a candidate can answer questions that the majority of the population holds, I feel like they've done the most good they could with the resources they have.</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The concept of god cannot be proved/disproved. The concept of the Abrahamic god contradicts itself MANY MANY times and is easily logically "disproved". [NEWLINE] [NEWLINE] This is just one of hundreds of ways of to do<mask>. [NEWLINE] [NEWLINE] <mask> this is a nice logic game to know, most believers would just shrug it off with a "God works in mysterious ways". The whole idea of faith (especially in the Abrahamic faith) is about holding to an idea against all logic and evidence. [NEWLINE] [NEWLINE] Buddhism is different from a lot of faith systems, it tries to be logically consistent, it even encourages you to challenge your own beliefs, something which is a taboo in most faith systems. [NEWLINE] [NEWLINE] <mask><mask> I completely agree with you about this argument being a good logical pit-fall, it will not "disprove" faith,<mask> it is usually an anti-logical concept for which logical proof or disproof just don't apply.</s>
Label encoding: <s>The concept of god cannot be proved/disproved. The concept of the Abrahamic god contradicts itself MANY MANY times and is easily logically "disproved". [NEWLINE] [NEWLINE] This is just one of hundreds of ways of to do so. [NEWLINE] [NEWLINE] While this is a nice logic game to know, most believers would just shrug it off with a "God works in mysterious ways". The whole idea of faith (especially in the Abrahamic faith) is about holding to an idea against all logic and evidence. [NEWLINE] [NEWLINE] Buddhism is different from a lot of faith systems, it tries to be logically consistent, it even encourages you to challenge your own beliefs, something which is a taboo in most faith systems. [NEWLINE] [NEWLINE] So while I completely agree with you about this argument being a good logical pit-fall, it will not "disprove" faith, as it is usually an anti-logical concept for which logical proof or disproof just don't apply.</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The whole point is that Israel isn't being singled out. [ENDQ] [NEWLINE] Israel's foreign policy needs to be understood in the context that it's almost entirely surrounded by countries that don't just wish to damage Israel for resources or profit,<mask> wish to wipe Israel off the map for no purpose other than itself. No other country on the entire face of the planet comes under such criticism for it's own self defence. [NEWLINE] [NEWLINE] It leads to the conclusion that perhaps many of the people who criticize Israel's foreign policy actually want Israel to bend over and let itself be fucked. They certainly aren't suggesting any new method for them to defend themselves without being wiped out. [NEWLINE] [NEWLINE] [STARTQ] You seem to be starting from the premise that any criticism of Israel is based in antisemitism. [ENDQ] [NEWLINE] I'm certainly not. In another comment here, I addressed that their participation in the Suez Crisis is a good example of something awful Israel did that other countries in the Middle East haven't done.</s>
Label encoding: <s> [STARTQ] The whole point is that Israel isn't being singled out. [ENDQ] [NEWLINE] Israel's foreign policy needs to be understood in the context that it's almost entirely surrounded by countries that don't just wish to damage Israel for resources or profit, but wish to wipe Israel off the map for no purpose other than itself. No other country on the entire face of the planet comes under such criticism for it's own self defence. [NEWLINE] [NEWLINE] It leads to the conclusion that perhaps many of the people who criticize Israel's foreign policy actually want Israel to bend over and let itself be fucked. They certainly aren't suggesting any new method for them to defend themselves without being wiped out. [NEWLINE] [NEWLINE] [STARTQ] You seem to be starting from the premise that any criticism of Israel is based in antisemitism. [ENDQ] [NEWLINE] I'm certainly not. In another comment here, I addressed that their participation in the Suez Crisis is a good example of something awful Israel did that other countries in the Middle East haven't done.</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think we can all agree then that the two methods are not mutually exclusive.  Sure, educate everyone on<mask> rape is and<mask> it is wrong. <mask> at the same time, don't skimp on educating women on the steps they can take to avoid being raped.  Just<mask> most of today's rape prevention education would not prevent the majority of rape cases does not mean it is valueless. Maybe we can be more effective in<mask> we educate women, focusing more on<mask> they can do to prevent the more common form of rape (rape by an acquaintance). [NEWLINE] [NEWLINE] To make this point, I sometimes use the example of a pedestrian at a crosswalk. The pedestrian has the right of way to cross the street. Without looking for traffic, he steps into the street and gets hit by a car. [NEWLINE] [NEWLINE] Did the pedestrian have the right of way? Yes.<mask> that doesn't mean you don't have to take some personable responsibility for your own safety.</s>
Label encoding: <s>I think we can all agree then that the two methods are not mutually exclusive.  Sure, educate everyone on what rape is and why it is wrong.  But at the same time, don't skimp on educating women on the steps they can take to avoid being raped.  Just because most of today's rape prevention education would not prevent the majority of rape cases does not mean it is valueless. Maybe we can be more effective in how we educate women, focusing more on what they can do to prevent the more common form of rape (rape by an acquaintance). [NEWLINE] [NEWLINE] To make this point, I sometimes use the example of a pedestrian at a crosswalk. The pedestrian has the right of way to cross the street. Without looking for traffic, he steps into the street and gets hit by a car. [NEWLINE] [NEWLINE] Did the pedestrian have the right of way? Yes. But that doesn't mean you don't have to take some personable responsibility for your own safety.</s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do you think you'd want to stay in the middle class? You still get more money<mask> you earn more. People will try to get around tax no matter<mask> low the rate. Do you think a guy who earns 2 million net in a year will stop trying to wrought the tax system just<mask> he gets 2.5 million now? [NEWLINE] [NEWLINE] On a more moral perspective, money is just a social construct for value. Consider<mask> you are really taking from people<mask> you tax them.<mask> you tax a poor person, you may be taking away their ability to eat nutritiously, or for a middle class person you may be taking away their ability to get their kids a decent education.<mask> you tax a rich person, you are taking away their second annual holiday. In other words, the first 20000 you earn in a year is worth a lot more than each subsequent 20000.<mask> increasing tax brackets is consistent with whatever money is worth to the person you are taxing.</s>
Label encoding: <s>Why do you think you'd want to stay in the middle class? You still get more money if you earn more. People will try to get around tax no matter how low the rate. Do you think a guy who earns 2 million net in a year will stop trying to wrought the tax system just because he gets 2.5 million now? [NEWLINE] [NEWLINE] On a more moral perspective, money is just a social construct for value. Consider what you are really taking from people when you tax them. When you tax a poor person, you may be taking away their ability to eat nutritiously, or for a middle class person you may be taking away their ability to get their kids a decent education. When you tax a rich person, you are taking away their second annual holiday. In other words, the first 20000 you earn in a year is worth a lot more than each subsequent 20000. Therefore increasing tax brackets is consistent with whatever money is worth to the person you are taxing.</s>
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I mean<mask> I say "consistent" is that they are tailored to both males and females equally.<mask> you can't tell a girl she can't wear a tanktop and let a guy wear a muscle shirt.<mask> you don't let girls wear earings (for whatever reason), you can't let guys do it.<mask> you don't let girls wear shorts that go up to their groin, you can't let guys wear something equivalent. [NEWLINE] [NEWLINE] Of course it's not fair in your example. That's not exactly the topic we're on,<mask>. Sure men and women are different - physically and mentally. Physical testing is a completely different topic than dress codes. [NEWLINE] [NEWLINE] Obviously you wouldn't tell a boy "no cleavage"<mask> he doesn't have the same anatomy<mask> a girl. To try to help you understand<mask> I'm saying: you would tell that boy that he can't wear a cut off t-shirt that exposed his pecs.</s>
Label encoding: <s>What I mean when I say "consistent" is that they are tailored to both males and females equally. So you can't tell a girl she can't wear a tanktop and let a guy wear a muscle shirt. If you don't let girls wear earings (for whatever reason), you can't let guys do it. If you don't let girls wear shorts that go up to their groin, you can't let guys wear something equivalent. [NEWLINE] [NEWLINE] Of course it's not fair in your example. That's not exactly the topic we're on, though. Sure men and women are different - physically and mentally. Physical testing is a completely different topic than dress codes. [NEWLINE] [NEWLINE] Obviously you wouldn't tell a boy "no cleavage" since he doesn't have the same anatomy as a girl. To try to help you understand what I'm saying: you would tell that boy that he can't wear a cut off t-shirt that exposed his pecs.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How to get wages to rise in 3rd world countries without hurting 1st world economies has been a question that has plagued economists for a life time. <mask> you just cut off countries like Malaysia from American business<mask> wages are too low, that doesn't help Malaysians.  All you did was take employment out of their country and plunge them deeper into poverty. <mask> Malaysians try to raise wages they will see most of the companies go to other countries.  Putting factories in Vietnam or Kenya or somewhere else. [NEWLINE] [NEWLINE] I don't know the solution.  I'm not sure anyone does.  Or<mask> they do, there are 3 other genius economists who think that they are wrong. [NEWLINE] [NEWLINE] <mask>, most politicians think about<mask> is best for their country, not<mask> is best for the world's citizens. <mask> low business costs in Malaysia means cheaper goods in America, that's good for Americans economically. <mask> most politicians are ok with this exploitation. </s>
Label encoding: <s>How to get wages to rise in 3rd world countries without hurting 1st world economies has been a question that has plagued economists for a life time.  If you just cut off countries like Malaysia from American business because wages are too low, that doesn't help Malaysians.  All you did was take employment out of their country and plunge them deeper into poverty.  If Malaysians try to raise wages they will see most of the companies go to other countries.  Putting factories in Vietnam or Kenya or somewhere else. [NEWLINE] [NEWLINE] I don't know the solution.  I'm not sure anyone does.  Or if they do, there are 3 other genius economists who think that they are wrong. [NEWLINE] [NEWLINE] Additionally, most politicians think about what is best for their country, not what is best for the world's citizens.  If low business costs in Malaysia means cheaper goods in America, that's good for Americans economically.  So most politicians are ok with this exploitation. </s>
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I<mask> think that the trending reaction has been rather knee-jerk. [NEWLINE] [NEWLINE] [STARTQ] I check myself in order to dress not too revealing<mask> well. [ENDQ] [NEWLINE] [STARTQ] I just want to come to a thought process that changes<mask> I feel about this<mask> that it doesn't bother me anymore. [ENDQ] [NEWLINE] From my moderately modest female perspective pop culture feels superficial and hypersexualized.<mask> identifying<mask> counter culture, I still struggle with vanity and self worth being tied to appearance.<mask><mask> this is a huge for women who invest in being culturally significant. [NEWLINE] [NEWLINE] In a culture that objectifies woman, it is very hard to not feel that pressure. You may be in a loving, committed relationship,<mask> that doesn't change the environment. I'm just saying that it is a lot easier for a guy to be modest and still considered sexy. You shouldn't take it<mask> a personal affront. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, I wouldn't walk around with nipple showing.</s>
Label encoding: <s>I also think that the trending reaction has been rather knee-jerk. [NEWLINE] [NEWLINE] [STARTQ] I check myself in order to dress not too revealing as well. [ENDQ] [NEWLINE] [STARTQ] I just want to come to a thought process that changes how I feel about this so that it doesn't bother me anymore. [ENDQ] [NEWLINE] From my moderately modest female perspective pop culture feels superficial and hypersexualized. Despite identifying as counter culture, I still struggle with vanity and self worth being tied to appearance. I think this is a huge for women who invest in being culturally significant. [NEWLINE] [NEWLINE] In a culture that objectifies woman, it is very hard to not feel that pressure. You may be in a loving, committed relationship, but that doesn't change the environment. I'm just saying that it is a lot easier for a guy to be modest and still considered sexy. You shouldn't take it as a personal affront. [NEWLINE] [NEWLINE] On the other hand, I wouldn't walk around with nipple showing.</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let us look at the procedure in question. [NEWLINE] [NEWLINE] It's apparently 4000USD. 4000 USD is out of the reach of most Indians,<mask> is well within the reach of most Americans. [NEWLINE] [NEWLINE] <mask> - 26000 USD is not within the reach of most Americans. To a person who has to get the operation in question, is it unethical to take the option they can afford to save their own life<mask> it *may* displace someone else's care? [NEWLINE] [NEWLINE] <mask> expensive do you think Int'l travel is? A ticket from the US- [STARTQ] India-&gt;US can cost<mask> little<mask> 1000$<mask> booked in advance. [ENDQ] [NEWLINE] You're prescribing insolvency<mask> a potential out for an expensive procedure? That is completely life changing. [NEWLINE] [NEWLINE] And insurance is very difficult to obtain for people with pre-existing conditions. [NEWLINE] [NEWLINE] <mask> does a citizen of the country in which a procedure exists have a superior claim to that procedure? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Let us look at the procedure in question. [NEWLINE] [NEWLINE] It's apparently 4000USD. 4000 USD is out of the reach of most Indians, but is well within the reach of most Americans. [NEWLINE] [NEWLINE] However - 26000 USD is not within the reach of most Americans. To a person who has to get the operation in question, is it unethical to take the option they can afford to save their own life because it *may* displace someone else's care? [NEWLINE] [NEWLINE] How expensive do you think Int'l travel is? A ticket from the US- [STARTQ] India-&gt;US can cost as little as 1000$ if booked in advance. [ENDQ] [NEWLINE] You're prescribing insolvency as a potential out for an expensive procedure? That is completely life changing. [NEWLINE] [NEWLINE] And insurance is very difficult to obtain for people with pre-existing conditions. [NEWLINE] [NEWLINE] Why does a citizen of the country in which a procedure exists have a superior claim to that procedure? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it is visually distracting and unpleasant. [ENDQ] [NEWLINE] I do not find it unpleasant or distracting... It must be me. [NEWLINE] [NEWLINE] <mask> I wore a purple dress with pink flowers on it to a US highschool (mind you, college is very different), I'm pretty sure I would be laughed at, sent to the principal, and/or avoided by other students. It's not that I want to, it's simply a statement. Most people would make fast judgments about me<mask> I did this. [NEWLINE] [NEWLINE] Cross-dressing is not accepted in society, especially in countries other than the USA. In many places it's illegal, such<mask> in the UAE and Saudi Arabia. [NEWLINE] [NEWLINE] I'm not saying all fashion is worthless, it is an art after all. I'm just saying fashion trends and high fashion cause more harm than good. Fashion *could* become an art such<mask> photography, film, and game design,<mask> popular opinion stunts its growth.</s><pad>
Label encoding: <s> [STARTQ] Because it is visually distracting and unpleasant. [ENDQ] [NEWLINE] I do not find it unpleasant or distracting... It must be me. [NEWLINE] [NEWLINE] If I wore a purple dress with pink flowers on it to a US highschool (mind you, college is very different), I'm pretty sure I would be laughed at, sent to the principal, and/or avoided by other students. It's not that I want to, it's simply a statement. Most people would make fast judgments about me if I did this. [NEWLINE] [NEWLINE] Cross-dressing is not accepted in society, especially in countries other than the USA. In many places it's illegal, such as in the UAE and Saudi Arabia. [NEWLINE] [NEWLINE] I'm not saying all fashion is worthless, it is an art after all. I'm just saying fashion trends and high fashion cause more harm than good. Fashion *could* become an art such as photography, film, and game design, but popular opinion stunts its growth.</s><pad>
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yet movies and video games, which "everyone" says turns you into a rampaging murderer, are still on the table. In a future with the technology he describes, a lot of things might very well become blasé<mask><mask><mask> of changing sensibilities, and people will be less obsessed with the minority scenario of a common technology. [NEWLINE] [NEWLINE] I know a few people who I'd want to send to Antarctica.<mask> this goes into the Good People Gone Bad scenario;<mask> they can't be reeducated, then I'd rather they be dealt with "permanently." That's more honest than plugging your ears and imagining them safely sequestered away from you in Hell. That's a classic villain mentality...and one that many self-righteous family folk share. It's not pretty,<mask> it is reality: we are uncharitable to the Other. [NEWLINE] [NEWLINE] Just remember: even ex-convicts deserve a second chance to [bake bread]( [URL] )</s>
Label encoding: <s>Yet movies and video games, which "everyone" says turns you into a rampaging murderer, are still on the table. In a future with the technology he describes, a lot of things might very well become blasé as a result of changing sensibilities, and people will be less obsessed with the minority scenario of a common technology. [NEWLINE] [NEWLINE] I know a few people who I'd want to send to Antarctica. But this goes into the Good People Gone Bad scenario; if they can't be reeducated, then I'd rather they be dealt with "permanently." That's more honest than plugging your ears and imagining them safely sequestered away from you in Hell. That's a classic villain mentality...and one that many self-righteous family folk share. It's not pretty, but it is reality: we are uncharitable to the Other. [NEWLINE] [NEWLINE] Just remember: even ex-convicts deserve a second chance to [bake bread]( [URL] )</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Science is just a philisophical paradigm like any other. Granted it may be the most genuinely reliable paradigm we humans have managed to come up with,<mask> reason is just another way of approaching the world. [NEWLINE] [NEWLINE] <mask>,<mask> I believe wholeheartedly in the ultimate reliability of the scientific process, you take a lot of science on faith. Unless you've personally run every experiment known to man and verified the results for yourself, you have absolutely no real proof that any of the claims of science are any more legitimate than religion. You have faith that the scientific community is not a bunch of hacks and charlatans, and that faith is likely based on *reason*<mask> you can see many of the results of science reproduced faithfully in things like cell phones, cars, internet, etc.<mask> you've weighed the evidence in your head and determined that science is most likely the one true belief system. [NEWLINE] [NEWLINE] <mask> at the end of the day it's just another paradigm.</s>
Label encoding: <s>Science is just a philisophical paradigm like any other. Granted it may be the most genuinely reliable paradigm we humans have managed to come up with, but reason is just another way of approaching the world. [NEWLINE] [NEWLINE] Also, while I believe wholeheartedly in the ultimate reliability of the scientific process, you take a lot of science on faith. Unless you've personally run every experiment known to man and verified the results for yourself, you have absolutely no real proof that any of the claims of science are any more legitimate than religion. You have faith that the scientific community is not a bunch of hacks and charlatans, and that faith is likely based on *reason* because you can see many of the results of science reproduced faithfully in things like cell phones, cars, internet, etc. So you've weighed the evidence in your head and determined that science is most likely the one true belief system. [NEWLINE] [NEWLINE] But at the end of the day it's just another paradigm.</s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0858, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask> the false confessions. they are extremely rare these days<mask> not almost non existent. [ENDQ] [NEWLINE] Any source for that broad sweeping conclusion about the non-existence of a serious societal plague?  I recently interviewed a police officer that worked in homicide for 10-15 years, and she told me directly that most people she interviewed confessed.  She had to spend more time sifting through the confessions to determine which were real than she did to actually get the confessions in the first place. She did not give me an exact percentage of the number of false confessions,<mask> she clearly indicated that it was a frequent occurrence. [NEWLINE] [NEWLINE] That sounds really insane, right? <mask> take an under-educated person, place them in a room with several authority figures telling them, "We have concrete evidence that you did it.  Confess and you'll get 5 years, otherwise you get 25 to life," and you have a recipe for disaster and abuse.</s>
Label encoding: <s> [STARTQ] as far as the false confessions. they are extremely rare these days if not almost non existent. [ENDQ] [NEWLINE] Any source for that broad sweeping conclusion about the non-existence of a serious societal plague?  I recently interviewed a police officer that worked in homicide for 10-15 years, and she told me directly that most people she interviewed confessed.  She had to spend more time sifting through the confessions to determine which were real than she did to actually get the confessions in the first place. She did not give me an exact percentage of the number of false confessions, but she clearly indicated that it was a frequent occurrence. [NEWLINE] [NEWLINE] That sounds really insane, right?  But take an under-educated person, place them in a room with several authority figures telling them, "We have concrete evidence that you did it.  Confess and you'll get 5 years, otherwise you get 25 to life," and you have a recipe for disaster and abuse.</s>
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> For pretty much everyone. I mean, that's really dramatic of you to say and all,<mask> not actually reflective of some sort of precarious instability. Maybe you should read up on, say, the Guilded Age? Or the Great Depression? For<mask><mask><mask> the United States continues to exist (and it will for a long time) it will face new problems and inequalities it will struggle to rectify. [NEWLINE] [NEWLINE] You can say, "The country is broken for everyone<mask> the five percenters, man!"<mask>, uh, we've been striving to form "a more perfect" union<mask> 300 years ago. We'll probably keep getting better<mask> the last 300 years are any indication. Most of us in the United States aren't starving, most of us are warm, most of us are educated, most of us are healthy, and most of us are free to do whatever we want. We have problems,<mask> I'd say we're functioning. </s>
Label encoding: <s> For pretty much everyone. I mean, that's really dramatic of you to say and all, but not actually reflective of some sort of precarious instability. Maybe you should read up on, say, the Guilded Age? Or the Great Depression? For as long as the United States continues to exist (and it will for a long time) it will face new problems and inequalities it will struggle to rectify. [NEWLINE] [NEWLINE] You can say, "The country is broken for everyone but the five percenters, man!" but, uh, we've been striving to form "a more perfect" union since 300 years ago. We'll probably keep getting better if the last 300 years are any indication. Most of us in the United States aren't starving, most of us are warm, most of us are educated, most of us are healthy, and most of us are free to do whatever we want. We have problems, but I'd say we're functioning. </s>
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1300, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] *Zeitgeist* claims that it is a documentary, it is full of shit.  Bill Maher's *Religulous* used some of the same evidence to try and<mask><mask> Jesus was a myth, he is full of shit. *Triumph of the Will* is, in a sense, a documentary.  *Super Size Me* was a groundbreaking documentary...the results of which nobody has been able to reproduce....<mask> you want to learn history, you shouldn't be trusting one source that claims to give you the full story; especially one that is deliberately packaged to entertain you. [ENDQ] [NEWLINE] Ah, you are correct here. In my emotionally-charged ranting I forgot that docos aren't fully worthy of trust either - just another perspective, like a movie. Excellent point. Your other points are good, too; I find myself unable to answer a lot of your questions. Good on you. &amp;#8710;</s><pad>
Label encoding: <s> [STARTQ] *Zeitgeist* claims that it is a documentary, it is full of shit.  Bill Maher's *Religulous* used some of the same evidence to try and argue that Jesus was a myth, he is full of shit. *Triumph of the Will* is, in a sense, a documentary.  *Super Size Me* was a groundbreaking documentary...the results of which nobody has been able to reproduce.... If you want to learn history, you shouldn't be trusting one source that claims to give you the full story; especially one that is deliberately packaged to entertain you. [ENDQ] [NEWLINE] Ah, you are correct here. In my emotionally-charged ranting I forgot that docos aren't fully worthy of trust either - just another perspective, like a movie. Excellent point. Your other points are good, too; I find myself unable to answer a lot of your questions. Good on you. &amp;#8710;</s><pad>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have 2 cancer survivors in my family, and even THEY have spent more money on insurance (roughly adjusted for inflation,<mask> this shouldn't be taken<mask> fact) than they ever would. [ENDQ] [NEWLINE] I have serious doubts about this. Do you have numbers to back this up? I mean, there's no doubt that on average, insurance companies charge more than they pay out. I mean, that's kind of a "duh" observation. They're a business not a charity. Of course they have to make a profit. The point is to have protection against something catastrophic happening (like cancer), which only happens to a smaller number of customers,<mask> could potentially ruin them financially.<mask> your family members had insurance policies that weren't cost effective for cancer, I have serious concerns about those policies, and not to kick the political hornets' nest,<mask> I hope that those are the kinds of policies that don't meet the Obamacare standards.</s>
Label encoding: <s> [STARTQ] I have 2 cancer survivors in my family, and even THEY have spent more money on insurance (roughly adjusted for inflation, so this shouldn't be taken as fact) than they ever would. [ENDQ] [NEWLINE] I have serious doubts about this. Do you have numbers to back this up? I mean, there's no doubt that on average, insurance companies charge more than they pay out. I mean, that's kind of a "duh" observation. They're a business not a charity. Of course they have to make a profit. The point is to have protection against something catastrophic happening (like cancer), which only happens to a smaller number of customers, but could potentially ruin them financially. If your family members had insurance policies that weren't cost effective for cancer, I have serious concerns about those policies, and not to kick the political hornets' nest, but I hope that those are the kinds of policies that don't meet the Obamacare standards.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> interesting that you only talk about one of the several things I mentioned. Will you be taking equal time off over their youth for sick days and extra curricular activities, even<mask> that costs you more than it does her? [ENDQ] [NEWLINE] Yes it is only fair. [NEWLINE] [NEWLINE] [STARTQ] <mask> you got a promotion or job offer that required you to move, would you expect her to quit her job and come with you? [ENDQ] [NEWLINE] I would give her the option<mask> not expect it. [NEWLINE] [NEWLINE] [STARTQ] Would you ever buy a house together, or share finances? [ENDQ] [NEWLINE] Possibly, that remains to be seen. [NEWLINE] [NEWLINE] [STARTQ] And<mask> you are equals,<mask> would there not be an equal sharing of assets, at least the ones earned during the time you've been committed partners? [ENDQ] [NEWLINE] There would and should be. Keyphrase, "the ones earned during the time you've been committed partners". This is<mask> the prenup would come in.</s>
Label encoding: <s> [STARTQ] How interesting that you only talk about one of the several things I mentioned. Will you be taking equal time off over their youth for sick days and extra curricular activities, even if that costs you more than it does her? [ENDQ] [NEWLINE] Yes it is only fair. [NEWLINE] [NEWLINE] [STARTQ] If you got a promotion or job offer that required you to move, would you expect her to quit her job and come with you? [ENDQ] [NEWLINE] I would give her the option but not expect it. [NEWLINE] [NEWLINE] [STARTQ] Would you ever buy a house together, or share finances? [ENDQ] [NEWLINE] Possibly, that remains to be seen. [NEWLINE] [NEWLINE] [STARTQ] And if you are equals, why would there not be an equal sharing of assets, at least the ones earned during the time you've been committed partners? [ENDQ] [NEWLINE] There would and should be. Keyphrase, "the ones earned during the time you've been committed partners". This is where the prenup would come in.</s>
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> the market is saturated with shit<mask> of profit motivation.<mask> once a tomato was a tomato... our options are now mutant tomatoes for average price, and classic tomatoes which were previously average price now being chic and expensive by comparison. The status quo is created by profit motivation---<mask> making quality less accessible to the common man [ENDQ] [NEWLINE] Heirloom tomatoes were never cheap. Families used to spend dramatically more of their income on food. The CPI assumes a basket of spending that's roughly 30% on food,<mask> the modern reality is that it's closer to half of that: [URL].jpg [NEWLINE] [NEWLINE] This dramatic reduction in the real cost of food is due to huge, huge improvements in productivity - breeding efficient, rather than tasty, crops is a big part of that (it's<mask> due to improved fertilizers and pesticides, increasing mechanization, increased atmospheric CO2, genetically modified crops, and better logistics and planning). [NEWLINE] [NEWLINE] source: [URL] /</s><pad><pad>
Label encoding: <s> [STARTQ] But the market is saturated with shit because of profit motivation. Where once a tomato was a tomato... our options are now mutant tomatoes for average price, and classic tomatoes which were previously average price now being chic and expensive by comparison. The status quo is created by profit motivation--- thus making quality less accessible to the common man [ENDQ] [NEWLINE] Heirloom tomatoes were never cheap. Families used to spend dramatically more of their income on food. The CPI assumes a basket of spending that's roughly 30% on food, when the modern reality is that it's closer to half of that: [URL].jpg [NEWLINE] [NEWLINE] This dramatic reduction in the real cost of food is due to huge, huge improvements in productivity - breeding efficient, rather than tasty, crops is a big part of that (it's also due to improved fertilizers and pesticides, increasing mechanization, increased atmospheric CO2, genetically modified crops, and better logistics and planning). [NEWLINE] [NEWLINE] source: [URL] /</s><pad><pad>
Loss: tensor(0.1155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, this was a discussion of practicality, not morality.  Remember, the first comment in this thread started with "Ignoring for a moment whether or not a test is a good idea"<mask> I don't see<mask> whether or not it's humane has any bearing.  To my mind the only thing inhumane about it is the fact that you'd be trying to control reproduction. <mask> that's your concern then you should be arguing against OP's larger point, not this particular method of implementing it. <mask> it isn't, would you care to explain your meaning better?  "That's a terrible idea" isn't a very persuasive argument on its own right.  And<mask><mask><mask> inhumane it's really just a slight extension of things we already do to humans or other animals<mask> calling it inhumane seems a bit hypocritical unless you<mask> happen to want to dissolve child protective services and the pet and livestock industries.</s>
Label encoding: <s>First of all, this was a discussion of practicality, not morality.  Remember, the first comment in this thread started with "Ignoring for a moment whether or not a test is a good idea" so I don't see how whether or not it's humane has any bearing.  To my mind the only thing inhumane about it is the fact that you'd be trying to control reproduction.  If that's your concern then you should be arguing against OP's larger point, not this particular method of implementing it.  If it isn't, would you care to explain your meaning better?  "That's a terrible idea" isn't a very persuasive argument on its own right.  And as far as inhumane it's really just a slight extension of things we already do to humans or other animals so calling it inhumane seems a bit hypocritical unless you also happen to want to dissolve child protective services and the pet and livestock industries.</s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like your comment, and<mask><mask> that they draw from a larger pot that includes bluegrass,<mask><mask>'s the substance?<mask>'s the purity? [NEWLINE] [NEWLINE] I've hesitated until now to make this point,<mask><mask> a musician, I have a priori and a postieriori knowledge that this music takes little effort to write/orchestrate, and is certainly presented in the most dishonest way to its listeners. [NEWLINE] [NEWLINE] <mask> we live in a world with any loyalty, there are true bluegrass, country, folk, pop, and rock fans, who'd dismiss this band in a heartbeat<mask> a systematic combination of all these things, perfectly executed, with the soul of the actual music stripped right out of it. [NEWLINE] [NEWLINE] It's not easy to define,<mask> I couldn't more fervently believe that's the truth. Please, don't get me started on<mask> modern top 40 country stands compared to its roots. </s>
Label encoding: <s>I like your comment, and I agree that they draw from a larger pot that includes bluegrass, but where's the substance? Where's the purity? [NEWLINE] [NEWLINE] I've hesitated until now to make this point, but as a musician, I have a priori and a postieriori knowledge that this music takes little effort to write/orchestrate, and is certainly presented in the most dishonest way to its listeners. [NEWLINE] [NEWLINE] If we live in a world with any loyalty, there are true bluegrass, country, folk, pop, and rock fans, who'd dismiss this band in a heartbeat as a systematic combination of all these things, perfectly executed, with the soul of the actual music stripped right out of it. [NEWLINE] [NEWLINE] It's not easy to define, but I couldn't more fervently believe that's the truth. Please, don't get me started on where modern top 40 country stands compared to its roots. </s>
Loss: tensor(0.1090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand<mask> you're saying, and<mask><mask> that most of his situations are kinda fuzzy.<mask> with men it's a little different.<mask> he had forcefully pinned a girl down to the bar and kissed her, he'd have gone to jail *at least* for a night to "cool off."<mask> he had pestered some one night stand into having sex with him after she said no, she could claim abuse (a la Julian Assange) or at least make an angry tumblr rant that gets a few hundred thousand hits. [NEWLINE] [NEWLINE] We're still stuck on penis power logic. Having the ability to penetrate gives you the agency in the situation. That's<mask> some dom women like pegging. We,<mask> a culture, inherently see the "male" figure<mask> having the ability to choose,<mask> the female figure has to be chased.<mask> possible exaggeration, this would be completely different and more sympathetic<mask> it were a woman posting.</s>
Label encoding: <s>I understand what you're saying, and I agree that most of his situations are kinda fuzzy. But with men it's a little different. If he had forcefully pinned a girl down to the bar and kissed her, he'd have gone to jail *at least* for a night to "cool off." If he had pestered some one night stand into having sex with him after she said no, she could claim abuse (a la Julian Assange) or at least make an angry tumblr rant that gets a few hundred thousand hits. [NEWLINE] [NEWLINE] We're still stuck on penis power logic. Having the ability to penetrate gives you the agency in the situation. That's why some dom women like pegging. We, as a culture, inherently see the "male" figure as having the ability to choose, while the female figure has to be chased. Despite possible exaggeration, this would be completely different and more sympathetic if it were a woman posting.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would partially agree that funding certain disciplines can be suspect,<mask> disagree that they shouldn't be carried out at universities. [NEWLINE] [NEWLINE] Having many of the social sciences, for example, carry out their research in the company of mathematicians and statisticians has, through osmosis, leant a lot more scientific rigor to a ton of humanistic undertaking. I'm definitely for using statistical modeling in sociology or economics, and more than that, believe progress in those fields contributes to the everyday lives of countless humans. [NEWLINE] [NEWLINE] To boot, having a theologist at the same university<mask> a number of historians and natural scientists is absolutely instrumental in raising the bar of<mask> the theologist can and cannot say, in good conscience. With scientists having a stake in<mask> the university publishes and proclaims about the world, they're that much more likely to engage the theologist's arguments and,<mask> you and I probably assume, neutralize them. </s>
Label encoding: <s>I would partially agree that funding certain disciplines can be suspect, but disagree that they shouldn't be carried out at universities. [NEWLINE] [NEWLINE] Having many of the social sciences, for example, carry out their research in the company of mathematicians and statisticians has, through osmosis, leant a lot more scientific rigor to a ton of humanistic undertaking. I'm definitely for using statistical modeling in sociology or economics, and more than that, believe progress in those fields contributes to the everyday lives of countless humans. [NEWLINE] [NEWLINE] To boot, having a theologist at the same university as a number of historians and natural scientists is absolutely instrumental in raising the bar of what the theologist can and cannot say, in good conscience. With scientists having a stake in what the university publishes and proclaims about the world, they're that much more likely to engage the theologist's arguments and, as you and I probably assume, neutralize them. </s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>first off, calling someone blind or deaf *is* offensive to blind and deaf people. you keep saying that it's acceptable and no one cares,<mask> blind and deaf people do care. [NEWLINE] [NEWLINE] here's the rational i've heard: calling someone retarded is commenting on their ability and not their behaviour. you're saying that<mask> they are being thoughtless or harmful, their natural ability to function is in question,<mask> really you're trying to say they're not using their actual ability.<mask>, it's at the least, inaccurate. [NEWLINE] [NEWLINE] <mask> more than that, by calling bad things'retarded', you're furthering a false stereotype that mentally disabled people are Other, untrustworthy, dangerous, and/or useless. many disabled people are perfectly capable of kindness, rational thinking, and ingenuity.<mask> by calling people and things retarded, you are complicit in furthering a society that marginalizes and oppresses disabled people. </s>
Label encoding: <s>first off, calling someone blind or deaf *is* offensive to blind and deaf people. you keep saying that it's acceptable and no one cares, but blind and deaf people do care. [NEWLINE] [NEWLINE] here's the rational i've heard: calling someone retarded is commenting on their ability and not their behaviour. you're saying that because they are being thoughtless or harmful, their natural ability to function is in question, when really you're trying to say they're not using their actual ability. so, it's at the least, inaccurate. [NEWLINE] [NEWLINE] but more than that, by calling bad things'retarded', you're furthering a false stereotype that mentally disabled people are Other, untrustworthy, dangerous, and/or useless. many disabled people are perfectly capable of kindness, rational thinking, and ingenuity. but by calling people and things retarded, you are complicit in furthering a society that marginalizes and oppresses disabled people. </s>
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Just to be clear, I believe that no one should be able to pay for school (except for extra curricular things like excursions). [ENDQ] [NEWLINE] <mask>? I'm doing nothing to harm you or your kids and I'm even benefiting your kids by still paying my taxes<mask> removing the state's obligation to educate my child.<mask> should I be forbidden to help my child just<mask> I can't help hem all. [NEWLINE] [NEWLINE] [STARTQ] <mask> all schools were free to attend then eventually, I predict all schools would become much more equal, meaning that everyone would receive a much more equal education, which could eventually close the gap between rich and poor rather than widen it the way it does now. [ENDQ] [NEWLINE] The schools would have he same amount of money with less tuition,<mask> likely everyone would suffer. It seems like an odd ethic to support making the current system worse just<mask> everyone is equal. You might consider reading “Harrison Bergeron."</s>
Label encoding: <s> [STARTQ] Just to be clear, I believe that no one should be able to pay for school (except for extra curricular things like excursions). [ENDQ] [NEWLINE] Why? I'm doing nothing to harm you or your kids and I'm even benefiting your kids by still paying my taxes while removing the state's obligation to educate my child. Why should I be forbidden to help my child just because I can't help hem all. [NEWLINE] [NEWLINE] [STARTQ] If all schools were free to attend then eventually, I predict all schools would become much more equal, meaning that everyone would receive a much more equal education, which could eventually close the gap between rich and poor rather than widen it the way it does now. [ENDQ] [NEWLINE] The schools would have he same amount of money with less tuition, so likely everyone would suffer. It seems like an odd ethic to support making the current system worse just so everyone is equal. You might consider reading “Harrison Bergeron."</s>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can I add a comment? I'm catholic. Raised in a very religious family (my mother is Opus Dei), and we always believed in evolution. We never felt it was 100% literally true. Thomas Aquinas certainly was a fantastic theologian and even he didn't take evolution<mask> strictly true: [URL] #Creation [NEWLINE] [NEWLINE] Let me ask you this: in the bible God created the rainbow<mask> a sign he wouldn't flood the world again. Do you really feel that God changed<mask> light and water work after the flood?<mask> did it work beforehand? [NEWLINE] [NEWLINE] Could he have made the world in 6 literal 24 hour periods? Sure. He could have done it in the blink of an eye,<mask> he didn't. All evidence we have points towards him creating it in billions of years. God gave us the ability to observe the world.<mask> would he then lie to us about it using the tools he created?</s>
Label encoding: <s>Can I add a comment? I'm catholic. Raised in a very religious family (my mother is Opus Dei), and we always believed in evolution. We never felt it was 100% literally true. Thomas Aquinas certainly was a fantastic theologian and even he didn't take evolution as strictly true: [URL] #Creation [NEWLINE] [NEWLINE] Let me ask you this: in the bible God created the rainbow as a sign he wouldn't flood the world again. Do you really feel that God changed how light and water work after the flood? How did it work beforehand? [NEWLINE] [NEWLINE] Could he have made the world in 6 literal 24 hour periods? Sure. He could have done it in the blink of an eye, but he didn't. All evidence we have points towards him creating it in billions of years. God gave us the ability to observe the world. Why would he then lie to us about it using the tools he created?</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1: The point of eating meat is not just enjoyment. True, we don't *need* to eat meat,<mask> it's more difficult (and expensive) to have a balanced vegetarian diet. [NEWLINE] [NEWLINE] 4:<mask> do you feel the need to compare "killing pigs for food" with just "killing dogs for fun"?<mask> not rather "killing pigs for food/fun", or "killing pigs/dogs for food"? You seem to be stacking the cultural taboo against killing dogs, and the somewhat-more-universal opposition to killing just for fun. The two issues, are better understood separately. [NEWLINE] [NEWLINE] 5: Can a dog care about his own death? I don't mean "pain" or "dizziness", I mean *death*.<mask> a being *can't* care about its own death to begin with, I don't think it should be candidate for an *intrinsic* right to life.</s>
Label encoding: <s>1: The point of eating meat is not just enjoyment. True, we don't *need* to eat meat, but it's more difficult (and expensive) to have a balanced vegetarian diet. [NEWLINE] [NEWLINE] 4: Why do you feel the need to compare "killing pigs for food" with just "killing dogs for fun"? Why not rather "killing pigs for food/fun", or "killing pigs/dogs for food"? You seem to be stacking the cultural taboo against killing dogs, and the somewhat-more-universal opposition to killing just for fun. The two issues, are better understood separately. [NEWLINE] [NEWLINE] 5: Can a dog care about his own death? I don't mean "pain" or "dizziness", I mean *death*. If a being *can't* care about its own death to begin with, I don't think it should be candidate for an *intrinsic* right to life.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Women now have the choice to work either in or out of the home. [ENDQ] [NEWLINE] <mask> it may still be more socially acceptable for women to not work outside the home, the reality is that, for most women in the United States, there is no choice,<mask> dual income households have, for most families, become necessary to maintain a base standard of living. Certainly it would be nice to see more societal support for men who have the means to devote themselves full-time to childcare and who choose to do<mask>.<mask> households<mask> either partner has this opportunity are rare, and,<mask><mask> women are still less represented in the highest levels of the workforce, there are likely far fewer women who are able to support their partners on their single income than the reverse. I'm not saying stigma against stay-at-home dads isn't a problem;<mask> it's a problem that only a relatively small and otherwise privileged demographic group faces. </s>
Label encoding: <s> [STARTQ] Women now have the choice to work either in or out of the home. [ENDQ] [NEWLINE] While it may still be more socially acceptable for women to not work outside the home, the reality is that, for most women in the United States, there is no choice, as dual income households have, for most families, become necessary to maintain a base standard of living. Certainly it would be nice to see more societal support for men who have the means to devote themselves full-time to childcare and who choose to do so. But households where either partner has this opportunity are rare, and, given that women are still less represented in the highest levels of the workforce, there are likely far fewer women who are able to support their partners on their single income than the reverse. I'm not saying stigma against stay-at-home dads isn't a problem; but it's a problem that only a relatively small and otherwise privileged demographic group faces. </s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point,<mask><mask>, is balance. [NEWLINE] [NEWLINE] I have written before that I believe that excessive veneration of the military is dangerous,<mask> it leads us towards a fascist society.  I still think that. [NEWLINE] [NEWLINE] At the same time,<mask>, I no longer support or endorse negative generalisations about the troops, either.  Some of them genuinely are decent and well-meaning people, who enlisted in good faith, and got seriously screwed over by the government,<mask> well<mask> having a lot of horrible, traumatic experiences. [NEWLINE] [NEWLINE] <mask><mask> I don't think that every day should be Memorial Day, I don't think there's anything wrong with extending soldiers a certain amount of compassion, either.  They have been very poorly treated by the American government, which basically uses them and then throws them away<mask> it is finished with them. [NEWLINE] [NEWLINE] Don't worship them,<mask> be kind to them.  They need that.</s>
Label encoding: <s>The point, I think, is balance. [NEWLINE] [NEWLINE] I have written before that I believe that excessive veneration of the military is dangerous, because it leads us towards a fascist society.  I still think that. [NEWLINE] [NEWLINE] At the same time, however, I no longer support or endorse negative generalisations about the troops, either.  Some of them genuinely are decent and well-meaning people, who enlisted in good faith, and got seriously screwed over by the government, as well as having a lot of horrible, traumatic experiences. [NEWLINE] [NEWLINE] So while I don't think that every day should be Memorial Day, I don't think there's anything wrong with extending soldiers a certain amount of compassion, either.  They have been very poorly treated by the American government, which basically uses them and then throws them away when it is finished with them. [NEWLINE] [NEWLINE] Don't worship them, but be kind to them.  They need that.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A right is something you can take and defend. You have the "right" to do anything that is within your capability. Often times your capability is far diminished from the capability of the state, and in many cases the state 'guarantees' specific rights which is a way of saying that the collective capability of the state ensures your access to that right. [NEWLINE] [NEWLINE] <mask> rights exist. [NEWLINE] [NEWLINE] Interestingly in the US the only "right" that is mentioned in the Articles of the Constitution was the right to Habeus corpus, and it is only mentioned in context to the states right to suspend it during times of emergency (Art. 1, Sec. 9). [NEWLINE] [NEWLINE] The Bill of Rights was written after the constitution was ratified and in western legal theory all of those "rights" stem form Habeus corpus. Basically the state (in the US) has the right to suspend all of your rights<mask> /<mask> necessary. </s>
Label encoding: <s>A right is something you can take and defend. You have the "right" to do anything that is within your capability. Often times your capability is far diminished from the capability of the state, and in many cases the state 'guarantees' specific rights which is a way of saying that the collective capability of the state ensures your access to that right. [NEWLINE] [NEWLINE] Therefore rights exist. [NEWLINE] [NEWLINE] Interestingly in the US the only "right" that is mentioned in the Articles of the Constitution was the right to Habeus corpus, and it is only mentioned in context to the states right to suspend it during times of emergency (Art. 1, Sec. 9). [NEWLINE] [NEWLINE] The Bill of Rights was written after the constitution was ratified and in western legal theory all of those "rights" stem form Habeus corpus. Basically the state (in the US) has the right to suspend all of your rights if / when necessary. </s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that to expect realism is an unrealistic expectancy. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] <mask> comparing character's of a TV show to oneself is a natural human behaviour which is hard to break<mask><mask> not broken will almost always result in a bad viewer experience. The key is to understand that the characters each have their own individual identity that is different from you or anything that you know or experience. And the show in itself has it's own reality which is different to the reality of the real world. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The question of whether or not Breaking Bad is realistic to the real world is a hard one, realistic by who's perception? I guess the only people who truly know the answer to this question are drug dealers and career criminals which I am glad to say I do not have any acquaintances with....<mask>, even then, one career criminal's  perception of realism may be different to that of another's and may result in conflicting ideas.</s>
Label encoding: <s>I think that to expect realism is an unrealistic expectancy. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Also comparing character's of a TV show to oneself is a natural human behaviour which is hard to break but if not broken will almost always result in a bad viewer experience. The key is to understand that the characters each have their own individual identity that is different from you or anything that you know or experience. And the show in itself has it's own reality which is different to the reality of the real world. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The question of whether or not Breaking Bad is realistic to the real world is a hard one, realistic by who's perception? I guess the only people who truly know the answer to this question are drug dealers and career criminals which I am glad to say I do not have any acquaintances with.... but, even then, one career criminal's  perception of realism may be different to that of another's and may result in conflicting ideas.</s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Porn is not anywhere near like having sex in real life.<mask> you are unable to perform in the bedroom due to porn related issues, yes it is bad.<mask> porn destroys relationships and affects your social life, yes it is bad. Maybe you should just take a look at /r/nofap and actually see some of the things discussed there before making such a quick judgement. [ENDQ] [NEWLINE] I'm not doubting that it has that effect on some people, just<mask> I don't doubt that some people are effected negatively by alcohol.  I'm saying that doesn't make in inherently bad.  Inherently. [NEWLINE] [NEWLINE] <mask>, the inability to moderate is basically the definition of addiction.  Saying some people are addicted to something<mask> not affected by it is pretty absurd.  Perhaps you mean they "manage" it. [NEWLINE] [NEWLINE] I'm sure you'll forgive me for ceasing to read any further.</s>
Label encoding: <s> [STARTQ] Porn is not anywhere near like having sex in real life. If you are unable to perform in the bedroom due to porn related issues, yes it is bad. If porn destroys relationships and affects your social life, yes it is bad. Maybe you should just take a look at /r/nofap and actually see some of the things discussed there before making such a quick judgement. [ENDQ] [NEWLINE] I'm not doubting that it has that effect on some people, just as I don't doubt that some people are effected negatively by alcohol.  I'm saying that doesn't make in inherently bad.  Inherently. [NEWLINE] [NEWLINE] Additionally, the inability to moderate is basically the definition of addiction.  Saying some people are addicted to something but not affected by it is pretty absurd.  Perhaps you mean they "manage" it. [NEWLINE] [NEWLINE] I'm sure you'll forgive me for ceasing to read any further.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Practicing an art is one of the most mentally stimulating things a person can do. I know there are studies showing that students that are involved in a music program on average do better in school than those who are not.   Being involved in the arts someway or another can greatly benefit your mental health, and often the best way to get involved is through a school group. [NEWLINE] [NEWLINE] Someone else made the point "Our society cannot function without a balance of both." in respect to artists and STEM/ other workers. It's my perspective that this "balance" applies on a personal level too, that everyone needs the sort of mental stimulation that the arts can provide. The best way to get this stimulation is to either practice an art yourself or to be able to analyze, dissect, and appreciate artistic media. To be able to do either of those requires some background or instruction that can be learned<mask> in High-school.  </s>
Label encoding: <s>Practicing an art is one of the most mentally stimulating things a person can do. I know there are studies showing that students that are involved in a music program on average do better in school than those who are not.   Being involved in the arts someway or another can greatly benefit your mental health, and often the best way to get involved is through a school group. [NEWLINE] [NEWLINE] Someone else made the point "Our society cannot function without a balance of both." in respect to artists and STEM/ other workers. It's my perspective that this "balance" applies on a personal level too, that everyone needs the sort of mental stimulation that the arts can provide. The best way to get this stimulation is to either practice an art yourself or to be able to analyze, dissect, and appreciate artistic media. To be able to do either of those requires some background or instruction that can be learned while in High-school.  </s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] some laws do just seem ridiculous [ENDQ] [NEWLINE] <mask> a german: can you tell me<mask> laws? [NEWLINE] [NEWLINE] Edit: /u/flatcap_monty : well<mask><mask> that the EU loves to control every shithole all over europe.<mask> arguing this way wont do much for a brit<mask> the streets are literally covered with cameras. [NEWLINE] [NEWLINE] <mask><mask><mask><mask> many EU laws are pretty sensible ( at least the ones we see in action)<mask><mask> some are pretty shitty<mask>. [NEWLINE] Borderpatrol around Italy and Greece is not a Borderpatrol its more like a Murderpatrol. [NEWLINE] [NEWLINE] We definitely have areas of improvement.<mask> the EU is just about 20 years old ( you can make a point that it started in the 50`), Noone expected the USA to make sensible laws 20 years after its foundation. ^^they ^^still ^^fail ^^at ^^that.</s>
Label encoding: <s> [STARTQ] some laws do just seem ridiculous [ENDQ] [NEWLINE] as a german: can you tell me what laws? [NEWLINE] [NEWLINE] Edit: /u/flatcap_monty : well I agree that the EU loves to control every shithole all over europe. But arguing this way wont do much for a brit where the streets are literally covered with cameras. [NEWLINE] [NEWLINE] So while I think many EU laws are pretty sensible ( at least the ones we see in action) I think some are pretty shitty indeed. [NEWLINE] Borderpatrol around Italy and Greece is not a Borderpatrol its more like a Murderpatrol. [NEWLINE] [NEWLINE] We definitely have areas of improvement. But the EU is just about 20 years old ( you can make a point that it started in the 50`), Noone expected the USA to make sensible laws 20 years after its foundation. ^^they ^^still ^^fail ^^at ^^that.</s>
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I personally don't like using the word;<mask> I do use things like white trash..<mask><mask> I said black trash in reference to people who live in the ghetto and act shitty it's racist. [NEWLINE] [NEWLINE] <mask>,<mask> you say "trashy black people" instead of "black trash" all of a sudden it's not viewed<mask> racist. [NEWLINE] [NEWLINE] <mask> I want to refer to someone insultingly I generally try to stick to phrases that don't use a color/race to describe them.<mask> whats funny is I've seen people<mask> afraid of offending they won't even use the color of someone to point them out in the crowd: [NEWLINE] [NEWLINE] "That guy with the blue hat. White shirt..." [NEWLINE] [NEWLINE] 'Yeah, I don't see him.' [NEWLINE] [NEWLINE] "Shorts? He's got yellow tennis shoes<mask><mask>." [NEWLINE] [NEWLINE] 'Oh, the only black guy?' [NEWLINE] [NEWLINE] "Uh, yes."</s>
Label encoding: <s>I personally don't like using the word; but I do use things like white trash.. But if I said black trash in reference to people who live in the ghetto and act shitty it's racist. [NEWLINE] [NEWLINE] However, if you say "trashy black people" instead of "black trash" all of a sudden it's not viewed as racist. [NEWLINE] [NEWLINE] If I want to refer to someone insultingly I generally try to stick to phrases that don't use a color/race to describe them. But whats funny is I've seen people so afraid of offending they won't even use the color of someone to point them out in the crowd: [NEWLINE] [NEWLINE] "That guy with the blue hat. White shirt..." [NEWLINE] [NEWLINE] 'Yeah, I don't see him.' [NEWLINE] [NEWLINE] "Shorts? He's got yellow tennis shoes I think." [NEWLINE] [NEWLINE] 'Oh, the only black guy?' [NEWLINE] [NEWLINE] "Uh, yes."</s>
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My source is common sense. [ENDQ] [NEWLINE] [STARTQ] I don't even know<mask> you're trying to say at this point. [ENDQ] [NEWLINE] /sigh [NEWLINE] [NEWLINE] I'm trying to get you to answer this question: [NEWLINE] [NEWLINE] **<mask> is the APFT a "proxy" for, other than its stated purpose of muscular strength and endurance, and<mask> evidence do you have to back that up?** [NEWLINE] [NEWLINE] You can say this is a'stamp of approval' for being healthy all you want,<mask> there is nothing to back that up. **Army recruits are already given a physical and a medical check-up,<mask> isn't that sufficient<mask> strength is irrelevant?** [NEWLINE] [NEWLINE] This is CMV, not SRS. '<mask> feelz' is not an appropriate response.<mask> you have no evidence or sources, that's fine. We can just leave it at that, and with my source I win :)</s>
Label encoding: <s> [STARTQ] My source is common sense. [ENDQ] [NEWLINE] [STARTQ] I don't even know what you're trying to say at this point. [ENDQ] [NEWLINE] /sigh [NEWLINE] [NEWLINE] I'm trying to get you to answer this question: [NEWLINE] [NEWLINE] ** What is the APFT a "proxy" for, other than its stated purpose of muscular strength and endurance, and what evidence do you have to back that up?** [NEWLINE] [NEWLINE] You can say this is a'stamp of approval' for being healthy all you want, but there is nothing to back that up. **Army recruits are already given a physical and a medical check-up, why isn't that sufficient if strength is irrelevant?** [NEWLINE] [NEWLINE] This is CMV, not SRS.'Because feelz' is not an appropriate response. If you have no evidence or sources, that's fine. We can just leave it at that, and with my source I win :)</s>
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Based on<mask> statistic is your supposition? [ENDQ] [NEWLINE] I don't have a specific statistic to cite here,<mask><mask><mask> the concept of the [Cycle of Poverty]( [URL] ) is pretty well known.<mask>,<mask><mask> the "value negative" population would likely be far smaller than 46%,<mask> we pay taxes in many more ways than just income tax. [NEWLINE] [NEWLINE] [STARTQ] <mask> 46% are value negative and that number is neither decreasing markedly nor increasing markedly, it stands to reason that they produce nearly equal amounts of value positive and negative children,<mask><mask> value negative people have more children than value positive. [ENDQ] [NEWLINE] This doesn't follow at all. The value-negative population could hypothetically be 100% likely to produce children who are value-negative,<mask><mask><mask> the value-positive population was equally likely to produce children who are value-positive, without any changes in the relative proportions of each group.</s>
Label encoding: <s> [STARTQ] Based on what statistic is your supposition? [ENDQ] [NEWLINE] I don't have a specific statistic to cite here, but I think the concept of the [Cycle of Poverty]( [URL] ) is pretty well known. Additionally, I think the "value negative" population would likely be far smaller than 46%, since we pay taxes in many more ways than just income tax. [NEWLINE] [NEWLINE] [STARTQ] If 46% are value negative and that number is neither decreasing markedly nor increasing markedly, it stands to reason that they produce nearly equal amounts of value positive and negative children, given that value negative people have more children than value positive. [ENDQ] [NEWLINE] This doesn't follow at all. The value-negative population could hypothetically be 100% likely to produce children who are value-negative, so long as the value-positive population was equally likely to produce children who are value-positive, without any changes in the relative proportions of each group.</s>
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0984, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] he'll at least be less coercive [ENDQ] [NEWLINE] I don't think you know<mask> the word "coercive" means.  It means to get something by force, or by threatening violence.  Franco did none of these things.  The fact that he is famous doesn't make him asking someone<mask> they want to have sex a threat. [NEWLINE] [NEWLINE] <mask> Franco were a famous mobster, then I could see your point. <mask> a famous actor - there should be no expectation that "Should I get a hotel room?" should be taken<mask> "I will hurt you<mask> you don't have sex with me". [NEWLINE] [NEWLINE] The problem with misusing words like this, other than just being wrong from a syntactical perspective, is that it diminishes the act of actually coercing people into having sex, or punishes too severely those who never committed the act - except by your misuse of language.</s>
Label encoding: <s> [STARTQ] he'll at least be less coercive [ENDQ] [NEWLINE] I don't think you know what the word "coercive" means.  It means to get something by force, or by threatening violence.  Franco did none of these things.  The fact that he is famous doesn't make him asking someone if they want to have sex a threat. [NEWLINE] [NEWLINE] If Franco were a famous mobster, then I could see your point.  As a famous actor - there should be no expectation that "Should I get a hotel room?" should be taken as "I will hurt you if you don't have sex with me". [NEWLINE] [NEWLINE] The problem with misusing words like this, other than just being wrong from a syntactical perspective, is that it diminishes the act of actually coercing people into having sex, or punishes too severely those who never committed the act - except by your misuse of language.</s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My CMV is mainly against the lax attitude we have towards protecting the respect and eliminating stereotypes of 'foreign' cultures and religions in the west, and for the elevation of the swastika to a neutral symbol<mask> that it doesn't create hatred. [ENDQ] [NEWLINE] Towards that end, consider<mask> educating people actually means. The only place "we" really have the ability to educate people is in schools. The American public school system lags behind much of the rest of the developed world in terms of actually teaching material the children. On top of its failure to teach core course material, even things<mask> simple<mask> not bullying gay people are considered controversial in parts of the country. [NEWLINE] [NEWLINE] <mask>, yes, people should know that the swastika symbol means different things to different people.<mask>, is it really worth pushing to teach children that, or are we better served teaching them that evolution really is a thing?</s>
Label encoding: <s> [STARTQ] My CMV is mainly against the lax attitude we have towards protecting the respect and eliminating stereotypes of 'foreign' cultures and religions in the west, and for the elevation of the swastika to a neutral symbol so that it doesn't create hatred. [ENDQ] [NEWLINE] Towards that end, consider what educating people actually means. The only place "we" really have the ability to educate people is in schools. The American public school system lags behind much of the rest of the developed world in terms of actually teaching material the children. On top of its failure to teach core course material, even things as simple as not bullying gay people are considered controversial in parts of the country. [NEWLINE] [NEWLINE] So, yes, people should know that the swastika symbol means different things to different people. But, is it really worth pushing to teach children that, or are we better served teaching them that evolution really is a thing?</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's<mask> nonsensical to say that I didn't buy the data<mask> I have a copy of the data that I can copy and manipulate without the control or consent of the person selling me the data. [ENDQ] [NEWLINE] It isn't nonsensical<mask> you consider that human civilization is more than just the universe and its physical laws. You're taking a ludicrously reductionist perspective which ignores societal context - a context which is needed to explain<mask> the thing came to be in the first place. [NEWLINE] [NEWLINE] The rights we give content creators encourage them to produce their work and make it available to the general public. Yes, there will always be some who will make art without demand for anything else.<mask>, such art is only a small amount of the total made, and it leaves out much of our civilizations' greatest works. Furthermore,<mask><mask><mask> art has existed, it has enjoyed very close ties to compensation for that art.</s><pad>
Label encoding: <s> [STARTQ] It's also nonsensical to say that I didn't buy the data when I have a copy of the data that I can copy and manipulate without the control or consent of the person selling me the data. [ENDQ] [NEWLINE] It isn't nonsensical when you consider that human civilization is more than just the universe and its physical laws. You're taking a ludicrously reductionist perspective which ignores societal context - a context which is needed to explain how the thing came to be in the first place. [NEWLINE] [NEWLINE] The rights we give content creators encourage them to produce their work and make it available to the general public. Yes, there will always be some who will make art without demand for anything else. However, such art is only a small amount of the total made, and it leaves out much of our civilizations' greatest works. Furthermore, as long as art has existed, it has enjoyed very close ties to compensation for that art.</s><pad>
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> women can serve in the military, I don't see<mask> they need doors opened and objects lifted for them. [ENDQ] [NEWLINE] Ok, well women don't *need* those things. Plus, opening a door is generally a common courtesy, not one that is only expected to be done by men. I don't really see your argument here about lifting things. Just<mask> some women can serve in the military doesn't mean all women are strong enough to lift certain objects. On average, men will be stronger,<mask><mask> they help. I mean,<mask> for some reason you needed to get into a small space (let's say an attic to retreive something), would you call a woman or a man? You'd probably call a woman<mask> that's an advantage she has over a man. No one is going to complain that that's something men need, it's just a nice thing to do.</s>
Label encoding: <s> [STARTQ] If women can serve in the military, I don't see why they need doors opened and objects lifted for them. [ENDQ] [NEWLINE] Ok, well women don't *need* those things. Plus, opening a door is generally a common courtesy, not one that is only expected to be done by men. I don't really see your argument here about lifting things. Just because some women can serve in the military doesn't mean all women are strong enough to lift certain objects. On average, men will be stronger, hence why they help. I mean, if for some reason you needed to get into a small space (let's say an attic to retreive something), would you call a woman or a man? You'd probably call a woman because that's an advantage she has over a man. No one is going to complain that that's something men need, it's just a nice thing to do.</s>
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0801, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My problem with such scenarios is that they have to work<mask> hard to present a dichotomy that in the real world would simply be a false dichotomy. [NEWLINE] [NEWLINE] You would not know the trolley would not stop<mask> it hit you. [NEWLINE] You would not know that the trolley would stop in time<mask> you threw the other person off the bridge. [NEWLINE] You would not know that the other 5 people would get off the tracks in time or not. [NEWLINE] You would not know that the driver did not see the 5 people. [NEWLINE] [NEWLINE] Etc. [NEWLINE] [NEWLINE] To me, any "problem" that requires one go to such lengths to redefine reality in order to "reason" about it is simply clearly not material to any real world discussion and is by definition not a valid candidate for a rational discourse,<mask> to be a reasonable discussions the content must be grounded in some way with a shared reality. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>My problem with such scenarios is that they have to work so hard to present a dichotomy that in the real world would simply be a false dichotomy. [NEWLINE] [NEWLINE] You would not know the trolley would not stop if it hit you. [NEWLINE] You would not know that the trolley would stop in time if you threw the other person off the bridge. [NEWLINE] You would not know that the other 5 people would get off the tracks in time or not. [NEWLINE] You would not know that the driver did not see the 5 people. [NEWLINE] [NEWLINE] Etc. [NEWLINE] [NEWLINE] To me, any "problem" that requires one go to such lengths to redefine reality in order to "reason" about it is simply clearly not material to any real world discussion and is by definition not a valid candidate for a rational discourse, as to be a reasonable discussions the content must be grounded in some way with a shared reality. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The U.S. has terrible maternity leave (usually unpaid),<mask> shes forced to go back to work ASAP,<mask> postpartum recovery. Daycare is similarly expensive,<mask> most of her salary goes toward that. It takes a fantastic salary to be able to handle both on your own. [NEWLINE] [NEWLINE] Extortion is a better word for it. I discussed this in another comment,<mask> I'll summarize here. In live donor organ donations, there are mandatory waiting periods and counseling. There are tons of restrictions on who can donate under<mask> conditions, and any financial incentive is illegal. The reason for this is that a medical procedure requires clear and unpressured consent. [NEWLINE] [NEWLINE] The dad is not being extorted anymore than a dog is extorting his owner for kibble. A decision was made (to get the dog/to have sex with this person) and a responsibility was taken on. </s>
Label encoding: <s>The U.S. has terrible maternity leave (usually unpaid), so shes forced to go back to work ASAP, despite postpartum recovery. Daycare is similarly expensive, so most of her salary goes toward that. It takes a fantastic salary to be able to handle both on your own. [NEWLINE] [NEWLINE] Extortion is a better word for it. I discussed this in another comment, but I'll summarize here. In live donor organ donations, there are mandatory waiting periods and counseling. There are tons of restrictions on who can donate under what conditions, and any financial incentive is illegal. The reason for this is that a medical procedure requires clear and unpressured consent. [NEWLINE] [NEWLINE] The dad is not being extorted anymore than a dog is extorting his owner for kibble. A decision was made (to get the dog/to have sex with this person) and a responsibility was taken on. </s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's very true that the loyalists would be defending their home<mask> the rebels would<mask> likely see it that way. It depends<mask> caused the revolution. [NEWLINE] [NEWLINE] 300 million was a bit of a silly number<mask><mask>.<mask> still<mask><mask> the numbers would be in the tens of millions, and that's still insanely huge compared to the relatively small 1 million strong US army. And bear in mind the US army relies on supply infrastructure within the US that would get trashed by the rebels- in a way they would have much better targets than the army would. At the end of the day deploying the entire army would be hugely costly, whereas the population have plenty of guns lying around to fight with. [NEWLINE] [NEWLINE] Of course these are all hypotheticals, and I'm not claiming that I know for sure that a revolution would fail. I believe it mainly depends on<mask> the cause of the conflict was.</s>
Label encoding: <s>It's very true that the loyalists would be defending their home but the rebels would also likely see it that way. It depends what caused the revolution. [NEWLINE] [NEWLINE] 300 million was a bit of a silly number I agree. But still I think the numbers would be in the tens of millions, and that's still insanely huge compared to the relatively small 1 million strong US army. And bear in mind the US army relies on supply infrastructure within the US that would get trashed by the rebels- in a way they would have much better targets than the army would. At the end of the day deploying the entire army would be hugely costly, whereas the population have plenty of guns lying around to fight with. [NEWLINE] [NEWLINE] Of course these are all hypotheticals, and I'm not claiming that I know for sure that a revolution would fail. I believe it mainly depends on what the cause of the conflict was.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Frankly, you're correct, it isn't about the power any more. Cheap chips are very fast, you don't *need* the cutting edge to do high end computations any more. [NEWLINE] [NEWLINE] A mac is a simple choice that's guaranteed to be good, and look good. Yes, you can buy a more powerful machine for the money<mask> you know<mask> you're looking for, and can arguably buy prettier machines<mask> well. [NEWLINE] [NEWLINE] None of that discounts that a mac is almost never a "bad" choice for someone, whereas PC shopping is full of pitfalls and shoddy machines, even at comparable price points. [NEWLINE] [NEWLINE] In laptops, everything that high end is expensive. In workstations, everything costs a lot. In desktops, apple build more expensive machines than the sum of their costs,<mask> it's the things aside from the components that make people want them. </s>
Label encoding: <s>Frankly, you're correct, it isn't about the power any more. Cheap chips are very fast, you don't *need* the cutting edge to do high end computations any more. [NEWLINE] [NEWLINE] A mac is a simple choice that's guaranteed to be good, and look good. Yes, you can buy a more powerful machine for the money if you know what you're looking for, and can arguably buy prettier machines as well. [NEWLINE] [NEWLINE] None of that discounts that a mac is almost never a "bad" choice for someone, whereas PC shopping is full of pitfalls and shoddy machines, even at comparable price points. [NEWLINE] [NEWLINE] In laptops, everything that high end is expensive. In workstations, everything costs a lot. In desktops, apple build more expensive machines than the sum of their costs, but it's the things aside from the components that make people want them. </s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's an interesting take on it,<mask> honest fiction (a novel) is<mask> good for that. [NEWLINE] [NEWLINE] I<mask> found that the more I read the "news", the harder it becomes to see through it. I find myself starting to believe<mask> I read and take position on issues, get worried about stuff that I have no control over and may very well be fiction, and lose sight of<mask> is really important to me. [NEWLINE] [NEWLINE] Some poeple say that the truth can be gleaned<mask> you spend the time to read many different sources and keep in mind each source's bias and interests. I doubt that really works<mask> it will surely make you appear more informed to others (the one-eyed man in the land of the blind),<mask> in any case that's a full-time job and I have better things to do with my time than trying to read the signals in the noise.</s>
Label encoding: <s>That's an interesting take on it, but honest fiction (a novel) is also good for that. [NEWLINE] [NEWLINE] I also found that the more I read the "news", the harder it becomes to see through it. I find myself starting to believe what I read and take position on issues, get worried about stuff that I have no control over and may very well be fiction, and lose sight of what is really important to me. [NEWLINE] [NEWLINE] Some poeple say that the truth can be gleaned if you spend the time to read many different sources and keep in mind each source's bias and interests. I doubt that really works although it will surely make you appear more informed to others (the one-eyed man in the land of the blind), but in any case that's a full-time job and I have better things to do with my time than trying to read the signals in the noise.</s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> I've posted to someone else's comment, [definition of Lady.]( [URL] ) I believe that<mask> a man said he would like to be treated<mask> a gentleman, that he is not implying that he would like to be treated a certain way *<mask> * he is male,<mask><mask> he would like that respect. [NEWLINE] [NEWLINE] [STARTQ] <mask> there is no room for sarcasm, farce, or satire? Got it. [ENDQ] [NEWLINE] As yes, you're sarcasm is taken loud and clear. Perhaps I wrote without thinking. I meant to say that any work that has the intention seriously by others that is sexist makes it's creator sexist. For example, Jonathon Swift is not really a cannibal<mask> he is obviously using satire.<mask> MTV's Super Sweet Sixteen showing bratty girls and having us laugh at them is anti-feminist, thereby making it's creator anti-feminist.</s>
Label encoding: <s>As I've posted to someone else's comment, [definition of Lady.]( [URL] ) I believe that if a man said he would like to be treated as a gentleman, that he is not implying that he would like to be treated a certain way * because * he is male, but because he would like that respect. [NEWLINE] [NEWLINE] [STARTQ] So there is no room for sarcasm, farce, or satire? Got it. [ENDQ] [NEWLINE] As yes, you're sarcasm is taken loud and clear. Perhaps I wrote without thinking. I meant to say that any work that has the intention seriously by others that is sexist makes it's creator sexist. For example, Jonathon Swift is not really a cannibal because he is obviously using satire. However MTV's Super Sweet Sixteen showing bratty girls and having us laugh at them is anti-feminist, thereby making it's creator anti-feminist.</s>
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do you have a problem<mask> I raise my children to be feminists, or liberal?<mask> I raise them to value hard work and equality? [ENDQ] [NEWLINE] No, I have not witnessed toxic social behaviour due to these values. [NEWLINE] [NEWLINE] I have problems with religious fundamentalism, and<mask><mask> increasing the age at which you adopt religion formally could solve the problem. [NEWLINE] [NEWLINE] [STARTQ] I will have an undue influence on my children no matter<mask> hard I try to teach them neutrally. [ENDQ] [NEWLINE] That is fine,<mask><mask> you don't have a huge manipulative machine to assist you, the child will<mask> have other input and be able to temper their beliefs<mask> they reach the age. [NEWLINE] [NEWLINE] [STARTQ] demonstrate that being raised religiously is harmful. [ENDQ] [NEWLINE] For purposes of this discussion it's enough<mask> we agree religious fundamentalism (muslim extremists, creationists, etc.) is harmful.  Do we?</s>
Label encoding: <s> [STARTQ] Do you have a problem if I raise my children to be feminists, or liberal? If I raise them to value hard work and equality? [ENDQ] [NEWLINE] No, I have not witnessed toxic social behaviour due to these values. [NEWLINE] [NEWLINE] I have problems with religious fundamentalism, and I think increasing the age at which you adopt religion formally could solve the problem. [NEWLINE] [NEWLINE] [STARTQ] I will have an undue influence on my children no matter how hard I try to teach them neutrally. [ENDQ] [NEWLINE] That is fine, but if you don't have a huge manipulative machine to assist you, the child will also have other input and be able to temper their beliefs when they reach the age. [NEWLINE] [NEWLINE] [STARTQ] demonstrate that being raised religiously is harmful. [ENDQ] [NEWLINE] For purposes of this discussion it's enough if we agree religious fundamentalism (muslim extremists, creationists, etc.) is harmful.  Do we?</s>
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry for having a broken link - it's fixed now.<mask><mask> reading the actual study will clear up some of your misconceptions about it. [NEWLINE] [NEWLINE] [STARTQ] Hmm...<mask><mask> is it that men find men funny<mask> not women? Women find everyone funny? [ENDQ] [NEWLINE] Men find both men and women funny,<mask><mask><mask> they don't know the gender.<mask> do women.<mask> both men and women don't know the gender, they're likely to assume it was a man, and<mask> they did know the gender, they're likely to misremember that it was a man. [NEWLINE] [NEWLINE] [STARTQ] <mask> women are more tuned into men being funny... [ENDQ] [NEWLINE] This is not the case. The paper notes that the bias towards men is largely coming from men. Men are more tuned in to men being funny, not the other way around. The "mating" hypothesis is not a valid one. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Sorry for having a broken link - it's fixed now. I think reading the actual study will clear up some of your misconceptions about it. [NEWLINE] [NEWLINE] [STARTQ] Hmm... so why is it that men find men funny but not women? Women find everyone funny? [ENDQ] [NEWLINE] Men find both men and women funny, as long as they don't know the gender. So do women. When both men and women don't know the gender, they're likely to assume it was a man, and when they did know the gender, they're likely to misremember that it was a man. [NEWLINE] [NEWLINE] [STARTQ] If women are more tuned into men being funny... [ENDQ] [NEWLINE] This is not the case. The paper notes that the bias towards men is largely coming from men. Men are more tuned in to men being funny, not the other way around. The "mating" hypothesis is not a valid one. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From the sidebar: [NEWLINE] [NEWLINE] [STARTQ] Change My View: For people who have an opinion on something<mask> accept that they may be wrong or want help changing their view. [ENDQ] [NEWLINE] It's not that the poster must *want* their view changed. My wording was imprecise. They could<mask> want to have their perspective on their view changed in such a way that their initial view isn't really the same one. [NEWLINE] [NEWLINE] Really they just need a motive<mask> to<mask> they are making the post. Are they uncertain of the view/<mask> they came to it and want some input that would possibly change it? Do they think it's a rancid view to hold and want it to be changed through persuasion and rational argument? [NEWLINE] [NEWLINE] <mask><mask> the second sense of "Change my view" that you gave is more akin to soapboxing than to the expressed purpose of this sub<mask> stated in the sidebar.</s>
Label encoding: <s>From the sidebar: [NEWLINE] [NEWLINE] [STARTQ] Change My View: For people who have an opinion on something but accept that they may be wrong or want help changing their view. [ENDQ] [NEWLINE] It's not that the poster must *want* their view changed. My wording was imprecise. They could also want to have their perspective on their view changed in such a way that their initial view isn't really the same one. [NEWLINE] [NEWLINE] Really they just need a motive as to why they are making the post. Are they uncertain of the view/ how they came to it and want some input that would possibly change it? Do they think it's a rancid view to hold and want it to be changed through persuasion and rational argument? [NEWLINE] [NEWLINE] I think the second sense of "Change my view" that you gave is more akin to soapboxing than to the expressed purpose of this sub as stated in the sidebar.</s>
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your post is extraordinarily misleading. FTA that picture is from: [NEWLINE] [NEWLINE] [STARTQ] The Family and Medical Leave Act of 1993, signed into law during President Bill Clinton's first term, guaranteed maternity leave to many new mothers across the nation. It mandated a maximum of 12 weeks unpaid leave to mothers for the purpose of attending to a newborn or newly adopted child. [ENDQ] [NEWLINE] Most states don't *augment* these provisions,<mask> due to the federal law you can still get maternity leave in all states<mask><mask><mask> you're covered by the restrictions in the federal law. To be explicit: [NEWLINE] [NEWLINE] &gt;<mask>, the act did not attain universal coverage<mask> it included several limiting stipulations. In order to receive maternity leave, employees must work in a firm of 50 or more employees, maintain employment with the same business for 12 months and have accumulated at least 1,250 working hours over those 12 months.</s>
Label encoding: <s>Your post is extraordinarily misleading. FTA that picture is from: [NEWLINE] [NEWLINE] [STARTQ] The Family and Medical Leave Act of 1993, signed into law during President Bill Clinton's first term, guaranteed maternity leave to many new mothers across the nation. It mandated a maximum of 12 weeks unpaid leave to mothers for the purpose of attending to a newborn or newly adopted child. [ENDQ] [NEWLINE] Most states don't *augment* these provisions, but due to the federal law you can still get maternity leave in all states as long as you're covered by the restrictions in the federal law. To be explicit: [NEWLINE] [NEWLINE] &gt; However, the act did not attain universal coverage as it included several limiting stipulations. In order to receive maternity leave, employees must work in a firm of 50 or more employees, maintain employment with the same business for 12 months and have accumulated at least 1,250 working hours over those 12 months.</s>
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Welfare in and of itself is not a solution to anything, at least not a long-term one. It doesn't create anything, it just transfers who spends the money (after all, money will eventually be spent, whether it's now or in the future due to saving is irrelevant - and<mask> it's spent, that will stimulate the economy).<mask> anything saving is better than welfare payments<mask> saving accumulates interest over time, making the value of the money saved worth more in the future than it is in the present. [NEWLINE] [NEWLINE] I'm currently getting my masters in economics, and<mask> I'd rather not go through the time of citing every source I've come across I can assure you than just transferring wealth does nothing to stimulate productivity (and in turn economic growth).<mask> tax dollars were spent more on infrastructure or investments than transfer payments, the economy would grow and EVERYONE would be better off overall</s>
Label encoding: <s>Welfare in and of itself is not a solution to anything, at least not a long-term one. It doesn't create anything, it just transfers who spends the money (after all, money will eventually be spent, whether it's now or in the future due to saving is irrelevant - and when it's spent, that will stimulate the economy). If anything saving is better than welfare payments because saving accumulates interest over time, making the value of the money saved worth more in the future than it is in the present. [NEWLINE] [NEWLINE] I'm currently getting my masters in economics, and although I'd rather not go through the time of citing every source I've come across I can assure you than just transferring wealth does nothing to stimulate productivity (and in turn economic growth). If tax dollars were spent more on infrastructure or investments than transfer payments, the economy would grow and EVERYONE would be better off overall</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's relevant<mask> by not being specific, it's not violating the Lemon v Kurtzman "Lemon Law". [NEWLINE] [NEWLINE] [URL]._Kurtzman [NEWLINE] [NEWLINE] **1. It must have a secular legislative purpose.** "God" is considered a secular entity<mask> referred to in the general sense and has been used throughout history to bring the American people together against threats. [NEWLINE] [NEWLINE] **2. It must not have a primary effect of advancing or inhibiting religion.** By having the motto on money it's not advancing Christianity. It's not advancing anything. It's certainly not inhibiting atheism or Hinduism. [NEWLINE] [NEWLINE] **3. It must not result in "excessive government entanglement" with religion.** There's no entanglement. The words on the money don't mean we have to worship God or that the church has control over the government. [NEWLINE] [NEWLINE] It's safe.</s>
Label encoding: <s>It's relevant because by not being specific, it's not violating the Lemon v Kurtzman "Lemon Law". [NEWLINE] [NEWLINE] [URL]._Kurtzman [NEWLINE] [NEWLINE] **1. It must have a secular legislative purpose.** "God" is considered a secular entity when referred to in the general sense and has been used throughout history to bring the American people together against threats. [NEWLINE] [NEWLINE] **2. It must not have a primary effect of advancing or inhibiting religion.** By having the motto on money it's not advancing Christianity. It's not advancing anything. It's certainly not inhibiting atheism or Hinduism. [NEWLINE] [NEWLINE] **3. It must not result in "excessive government entanglement" with religion.** There's no entanglement. The words on the money don't mean we have to worship God or that the church has control over the government. [NEWLINE] [NEWLINE] It's safe.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do you think is the strongest argument against this perspective? [ENDQ] [NEWLINE] 1 is not provable scientifically.  You are either trying to justify the Scientific method by using the scientific method (which would be highly unscientific) or else you are choosing to have faith in it, and then claiming your faith-system is better than everybody else's faith systems. [NEWLINE] [NEWLINE] A third option (which may just be a different way to restate the first option) would be to admit that using the Scientific Method to justify the Scientific method is circular,<mask> then make a case of special pleading that it is the "best" circularity, based on the results.  This would be a bad idea<mask> [NEWLINE] [NEWLINE] a) special pleading is not persuasive [NEWLINE] b) you have already said you like your mother,<mask> in your case, you can't claim that your philosophy produces better results than hers anyway.</s>
Label encoding: <s> [STARTQ] what do you think is the strongest argument against this perspective? [ENDQ] [NEWLINE] 1 is not provable scientifically.  You are either trying to justify the Scientific method by using the scientific method (which would be highly unscientific) or else you are choosing to have faith in it, and then claiming your faith-system is better than everybody else's faith systems. [NEWLINE] [NEWLINE] A third option (which may just be a different way to restate the first option) would be to admit that using the Scientific Method to justify the Scientific method is circular, but then make a case of special pleading that it is the "best" circularity, based on the results.  This would be a bad idea because [NEWLINE] [NEWLINE] a) special pleading is not persuasive [NEWLINE] b) you have already said you like your mother, so in your case, you can't claim that your philosophy produces better results than hers anyway.</s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask> the physical world goes, without us identifying "floatiness" of boats, it doesnt exist. It is something we characterised in the world around us, in reality, composed of those interactions. [ENDQ] [NEWLINE] Excellent observation! A lot of<mask> -called properties that we think that things have aren't properties at all - much (everything?) has to do with interactions of the surrounding environment, from people to chemicals. [NEWLINE] [NEWLINE] Regarding the hum of the machine quote,<mask><mask> it's cute,<mask> I'm not sure<mask> seriously to take it. It sounds like the idea is that consciousness is produced<mask> a byproduct of the machine, and really isn't important to its function. Perhaps that's true... it seems like consciousness is an integral part of thought,<mask><mask> neurons can fire and wire together without it, I suppose it really can be an epiphenomenon.</s>
Label encoding: <s> [STARTQ] As far as the physical world goes, without us identifying "floatiness" of boats, it doesnt exist. It is something we characterised in the world around us, in reality, composed of those interactions. [ENDQ] [NEWLINE] Excellent observation! A lot of so -called properties that we think that things have aren't properties at all - much (everything?) has to do with interactions of the surrounding environment, from people to chemicals. [NEWLINE] [NEWLINE] Regarding the hum of the machine quote, I think it's cute, though I'm not sure how seriously to take it. It sounds like the idea is that consciousness is produced as a byproduct of the machine, and really isn't important to its function. Perhaps that's true... it seems like consciousness is an integral part of thought, but if neurons can fire and wire together without it, I suppose it really can be an epiphenomenon.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, that is not actually a plea in the sense of insanity being a plea and it is not at all applicable to all situations of rage. I'm just saying that a parents reaction to their child being abused is very intense and not unreasonable at all.<mask> I said above, "<mask><mask> it is natural for a parent to react<mask> strongly to their child being abused,<mask><mask> it is human nature for parents to turn violent<mask> their offspring are in danger." A parent-child relationship is very unique in that parents are meant to protect their children. This type of circumstance and accompanying rage is not the same<mask> "just getting mad" or "just being caught up in the moment." My point is that this type of reaction to a child being abused (or another loved one) is more intense and<mask><mask> that to a certain degree, actions steaming from it should be excused. [NEWLINE] </s>
Label encoding: <s>No, that is not actually a plea in the sense of insanity being a plea and it is not at all applicable to all situations of rage. I'm just saying that a parents reaction to their child being abused is very intense and not unreasonable at all. As I said above, " I think it is natural for a parent to react so strongly to their child being abused, I think it is human nature for parents to turn violent if their offspring are in danger." A parent-child relationship is very unique in that parents are meant to protect their children. This type of circumstance and accompanying rage is not the same as "just getting mad" or "just being caught up in the moment." My point is that this type of reaction to a child being abused (or another loved one) is more intense and I think that to a certain degree, actions steaming from it should be excused. [NEWLINE] </s>
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To me checking your privilege can be<mask> simple<mask> not complaining about eating at the same place twice. [NEWLINE] [NEWLINE] I bring that up specifically<mask> almost everyone at my job does that and I don't understand<mask>. To me that is a perfect opportunity to check your privilege. To understand that you have it<mask> well<mask> you are that you can pick<mask> you want to eat, and<mask> you want you can pick a new place every single day, and you can eat for over an hour before you come back to your air conditioned job and your cushy chair. [NEWLINE] [NEWLINE] There are people who can't eat at all,<mask> that's<mask> I check my privilege. Be positive, think about<mask> well off you are. Doesn't mean you should stop striving for better things in your life,<mask> mundane things like food, clothing and brand gear should not have any significant impact on your daily decisions. </s>
Label encoding: <s>To me checking your privilege can be as simple as not complaining about eating at the same place twice. [NEWLINE] [NEWLINE] I bring that up specifically because almost everyone at my job does that and I don't understand why. To me that is a perfect opportunity to check your privilege. To understand that you have it so well where you are that you can pick where you want to eat, and if you want you can pick a new place every single day, and you can eat for over an hour before you come back to your air conditioned job and your cushy chair. [NEWLINE] [NEWLINE] There are people who can't eat at all, so that's how I check my privilege. Be positive, think about how well off you are. Doesn't mean you should stop striving for better things in your life, but mundane things like food, clothing and brand gear should not have any significant impact on your daily decisions. </s>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A few things, volunteers are often compensated in some way<mask> they are outright not paid. Free vehicle registration and property tax reductions are fairly common [NEWLINE] [NEWLINE] Volunteers are often required to pass the same certifications<mask> a full time department. Saving money on staffing, extra money can be used to buy better equipment. The volunteer station I worked at had the highest conversion rate (no pulse or vfib to breathing with a pulse) in the province and we were completely volunteer [NEWLINE] [NEWLINE] You might be misinterpreting the joking part. A lot of firefighters and emergency response personnel in general joke about calls and victims<mask> a way to cope. They see terrible things all the time and joking about it is a way to get it off their chest without diving into the meat of the trauma in front of their friends. I know it's off putting<mask> you would really need to experience it to understand </s>
Label encoding: <s>A few things, volunteers are often compensated in some way if they are outright not paid. Free vehicle registration and property tax reductions are fairly common [NEWLINE] [NEWLINE] Volunteers are often required to pass the same certifications as a full time department. Saving money on staffing, extra money can be used to buy better equipment. The volunteer station I worked at had the highest conversion rate (no pulse or vfib to breathing with a pulse) in the province and we were completely volunteer [NEWLINE] [NEWLINE] You might be misinterpreting the joking part. A lot of firefighters and emergency response personnel in general joke about calls and victims as a way to cope. They see terrible things all the time and joking about it is a way to get it off their chest without diving into the meat of the trauma in front of their friends. I know it's off putting but you would really need to experience it to understand </s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When you read a really good poem, it often brings you to a conclusion without *telling* you. Arguing for essays over poems is comparing apples and oranges. They're totally different mediums that have completely different purposes. [NEWLINE] [NEWLINE] Poems do not usually have one meaning. Every reader brings their own experiences to the poem, and this allows them to relate in completely different ways. [NEWLINE] [NEWLINE] For example, [this poem, the First Elegy]( [URL] ) is one of my favorites. There are obvious meanings and ones that each person draws. It speaks to me particularly. [NEWLINE] [NEWLINE] <mask>, saying that poems aren't<mask> valid<mask> essays or other mediums is like saying abstract art isn't<mask> good<mask> realism<mask> it doesn't paint<mask> complete a picture.<mask> there is something inherently special about a subject that doesn't give you everything, and makes you draw your own conclusions. </s>
Label encoding: <s>When you read a really good poem, it often brings you to a conclusion without *telling* you. Arguing for essays over poems is comparing apples and oranges. They're totally different mediums that have completely different purposes. [NEWLINE] [NEWLINE] Poems do not usually have one meaning. Every reader brings their own experiences to the poem, and this allows them to relate in completely different ways. [NEWLINE] [NEWLINE] For example, [this poem, the First Elegy]( [URL] ) is one of my favorites. There are obvious meanings and ones that each person draws. It speaks to me particularly. [NEWLINE] [NEWLINE] Additionally, saying that poems aren't as valid as essays or other mediums is like saying abstract art isn't as good as realism because it doesn't paint as complete a picture. But there is something inherently special about a subject that doesn't give you everything, and makes you draw your own conclusions. </s>
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask> say it's unfair to the future partners of people who have had incestuous sexual relationships. [NEWLINE] [NEWLINE] I assume most of the time someone would never even learn their partner had ever engaged in incest, and would have to obliviously socialize with their partner and their partner's sibling/cousin/parent/etc. either unaware of or disgruntled by the sexual lines that have been crossed. [NEWLINE] [NEWLINE] Maybe<mask> the stigma was removed this wouldn't be<mask> much of an issue,<mask> even<mask> it's at the very least like hanging out with your girlfriend and one of her ex-boyfriends, whom she very likely still has a magnitude of feelings for, for numerous reasons. [NEWLINE] [NEWLINE] <mask><mask> reddit is way too liberal sometimes. I'm a very open minded atheist<mask><mask><mask> incest either in the open or<mask> a dirty secret can be very damaging for many involved.</s>
Label encoding: <s>I would also say it's unfair to the future partners of people who have had incestuous sexual relationships. [NEWLINE] [NEWLINE] I assume most of the time someone would never even learn their partner had ever engaged in incest, and would have to obliviously socialize with their partner and their partner's sibling/cousin/parent/etc. either unaware of or disgruntled by the sexual lines that have been crossed. [NEWLINE] [NEWLINE] Maybe if the stigma was removed this wouldn't be as much of an issue, but even so it's at the very least like hanging out with your girlfriend and one of her ex-boyfriends, whom she very likely still has a magnitude of feelings for, for numerous reasons. [NEWLINE] [NEWLINE] I think reddit is way too liberal sometimes. I'm a very open minded atheist but I think incest either in the open or as a dirty secret can be very damaging for many involved.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A very skillfully rendered picture of a flower,<mask> real to life and formally beautiful<mask> I can find,  does nothing to change my perspective or progress art in any way. It's idea: flowers are beautiful. [NEWLINE] [NEWLINE] Duchamp places ready made galleries in a gallery. He doesn't not invest tim in their making,<mask> their presence inspire people to intellectually reflect on<mask> qualifies<mask> art,<mask> it means to be and artist or a maker, and<mask> cultural bias dictates<mask> is considered valuable and<mask> is considered trash and<mask> people can experience art. [NEWLINE] [NEWLINE] <mask><mask> to label approach Duchamp's work they same way you would a child's drawing misses the real conceptual power of his work. And to praise that flower painting over him you create climate<mask> art is nothing more than decoration, and  communicating ideas or exploring human life is secondary to making pretty objects. </s>
Label encoding: <s>A very skillfully rendered picture of a flower, as real to life and formally beautiful as I can find,  does nothing to change my perspective or progress art in any way. It's idea: flowers are beautiful. [NEWLINE] [NEWLINE] Duchamp places ready made galleries in a gallery. He doesn't not invest tim in their making, but their presence inspire people to intellectually reflect on what qualifies as art, what it means to be and artist or a maker, and how cultural bias dictates what is considered valuable and what is considered trash and where people can experience art. [NEWLINE] [NEWLINE] I think to label approach Duchamp's work they same way you would a child's drawing misses the real conceptual power of his work. And to praise that flower painting over him you create climate where art is nothing more than decoration, and  communicating ideas or exploring human life is secondary to making pretty objects. </s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not taking sides, or advocating that someone take sides. I'm just saying insults should be avoided<mask> you hope to convince the other person of anything, for reasons I outlined in my OP.<mask><mask> it is generally immoral to insult someone.<mask> that doesn't mean<mask><mask><mask> you say is automatically wrong. [NEWLINE] [NEWLINE] In short, it's not<mask><mask><mask> it's immoral, it's<mask>, in the vast majority of cases, *it's not an effective tactic and makes you look bad*, even<mask> your argument is completely sound and correct. It simply makes the other person less likely to listen to you, and gives them validation in their own mind<mask> it makes you look like a jackass who can't address the substance of their view. Whether it's true or not, it gives them reason not to listen to you, and that's the damage it does.</s>
Label encoding: <s>I'm not taking sides, or advocating that someone take sides. I'm just saying insults should be avoided if you hope to convince the other person of anything, for reasons I outlined in my OP. I think it is generally immoral to insult someone. But that doesn't mean I think what you say is automatically wrong. [NEWLINE] [NEWLINE] In short, it's not because I think it's immoral, it's because, in the vast majority of cases, *it's not an effective tactic and makes you look bad*, even if your argument is completely sound and correct. It simply makes the other person less likely to listen to you, and gives them validation in their own mind since it makes you look like a jackass who can't address the substance of their view. Whether it's true or not, it gives them reason not to listen to you, and that's the damage it does.</s>
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Parliament was not representative or democratic, restrained rights, and imposed judicial systems that did not view all people equally. [ENDQ] [NEWLINE] The actual language of the document does not mention representative/democratic governments. It does not mention inequality. It's only language for defining<mask> types of governments should be abolished is: [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> a long train of abuses and usurpations, pursuing invariably the same Object evinces a design to reduce them under absolute Despotism, it is their right, it is their duty, to throw off such Government [ENDQ] [NEWLINE] In<mask> way does revolution against democratically elected leaders on the grounds that they are abusing their power and ignoring the rights of citizens "go[ing] against all the Constitution stands for"<mask> the same group of people who wrote the constitution have stated explicitly that any government that abuses it's power should be destroyed and replaced? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Parliament was not representative or democratic, restrained rights, and imposed judicial systems that did not view all people equally. [ENDQ] [NEWLINE] The actual language of the document does not mention representative/democratic governments. It does not mention inequality. It's only language for defining what types of governments should be abolished is: [NEWLINE] [NEWLINE] [STARTQ] But when a long train of abuses and usurpations, pursuing invariably the same Object evinces a design to reduce them under absolute Despotism, it is their right, it is their duty, to throw off such Government [ENDQ] [NEWLINE] In what way does revolution against democratically elected leaders on the grounds that they are abusing their power and ignoring the rights of citizens "go[ing] against all the Constitution stands for" when the same group of people who wrote the constitution have stated explicitly that any government that abuses it's power should be destroyed and replaced? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You believe that the pain of a family raising a heavily disabled child, along with the deterioration of the quality of life for any other children in the family, is clearly not justification for murder.<mask>, you do think that a reduction procedure to allow a fetus to live, is justified murder. [NEWLINE] [NEWLINE] I don't know<mask> you are coming to your conclusion of a value of life. Murdering a fetus to save a fetus is acceptable,<mask> you net 1 "life."<mask> a lifetime of hardship for a child, its family, and the community is not greater than 1 "life," no matter<mask> painful, unproductive, or unwanted that "life" is.<mask> do you figure this?<mask> is it possible to compare these to each other? And<mask> you're going to make a law about it, don't you have to have a clear understanding of the value?</s>
Label encoding: <s>You believe that the pain of a family raising a heavily disabled child, along with the deterioration of the quality of life for any other children in the family, is clearly not justification for murder. Yet, you do think that a reduction procedure to allow a fetus to live, is justified murder. [NEWLINE] [NEWLINE] I don't know how you are coming to your conclusion of a value of life. Murdering a fetus to save a fetus is acceptable, because you net 1 "life." But a lifetime of hardship for a child, its family, and the community is not greater than 1 "life," no matter how painful, unproductive, or unwanted that "life" is. How do you figure this? How is it possible to compare these to each other? And if you're going to make a law about it, don't you have to have a clear understanding of the value?</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>More information is a good thing, yes. Being educated about the world is important, and the way that is achieved by having a free flow of information, a lot of it. The spread of information is a lot about<mask> you perceive it<mask> well. You have access to all kinds of information.<mask> you choose to be more affected by information you think is negative, then you will think negatively, and vice versa for positive information. [NEWLINE] [NEWLINE] Advancing technology does not necessarily have to be a good thing,<mask> it is essential. Stagnated societies crumble very quickly (see [the fall of the Roman Empire]( [URL] )) and innovation leads to improvements in<mask> we do anything. Luddites would want us to be back in "the good old days"<mask> we can't go back to the past. We just have to keep the human race going.  </s>
Label encoding: <s>More information is a good thing, yes. Being educated about the world is important, and the way that is achieved by having a free flow of information, a lot of it. The spread of information is a lot about how you perceive it as well. You have access to all kinds of information. If you choose to be more affected by information you think is negative, then you will think negatively, and vice versa for positive information. [NEWLINE] [NEWLINE] Advancing technology does not necessarily have to be a good thing, but it is essential. Stagnated societies crumble very quickly (see [the fall of the Roman Empire]( [URL] )) and innovation leads to improvements in how we do anything. Luddites would want us to be back in "the good old days" but we can't go back to the past. We just have to keep the human race going.  </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Its highly unlikely that culture is tied to race. [ENDQ] [NEWLINE] Uh...<mask><mask> we may be misunderstanding each other somehow,<mask> this statement would never hold true otherwise. [NEWLINE] [NEWLINE] [STARTQ] The point is that<mask> the premise that there is still unconcious discrimination in hiring is true then affirmative action doesn't unfairly prevent anyone from getting a job, it allows minorities who would otherwise be denied jobs due to discrimination to get those jobs. [ENDQ] [NEWLINE] I don't know too much about affirmative action for jobs, I'll admit: I'm talking about colleges,<mask> there are only a given amount of spots in the entering class, and several are taken up by people who do not deserve to be there<mask> much<mask> others, who got in<mask> of their race. It is much more difficult to get into a college<mask> you are Asian or Indian than<mask> you are Black or Hispanic. </s>
Label encoding: <s> [STARTQ] Its highly unlikely that culture is tied to race. [ENDQ] [NEWLINE] Uh... I think we may be misunderstanding each other somehow, because this statement would never hold true otherwise. [NEWLINE] [NEWLINE] [STARTQ] The point is that if the premise that there is still unconcious discrimination in hiring is true then affirmative action doesn't unfairly prevent anyone from getting a job, it allows minorities who would otherwise be denied jobs due to discrimination to get those jobs. [ENDQ] [NEWLINE] I don't know too much about affirmative action for jobs, I'll admit: I'm talking about colleges, where there are only a given amount of spots in the entering class, and several are taken up by people who do not deserve to be there as much as others, who got in because of their race. It is much more difficult to get into a college if you are Asian or Indian than if you are Black or Hispanic. </s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Depending on the nature of the case, it is possible that legal fees will be paid by the losing party. That very much varies by statute<mask>. [NEWLINE] [NEWLINE] <mask>, even in cases<mask> this is true, the person bringing the claim will have to pay their own legal fees<mask> they lose, which is a pretty strong disincentive to bring frivolous lawsuits.<mask> in most cases that go to trial, both parties will legitimately believe they are right, and they have chosen to take the matter to court to prove it at risk of losing. The role of the court is to resolve their disagreement by making a decision<mask> an authority that is binding. [NEWLINE] [NEWLINE] Bottom line: It is unlikely that people will risk bringing baseless lawsuits<mask> they will bear the costs of their own legal fees<mask> they lose. Cases will generally only go to trial<mask> there is a legitimate question of responsibility.</s>
Label encoding: <s>Depending on the nature of the case, it is possible that legal fees will be paid by the losing party. That very much varies by statute however. [NEWLINE] [NEWLINE] Also, even in cases where this is true, the person bringing the claim will have to pay their own legal fees if they lose, which is a pretty strong disincentive to bring frivolous lawsuits. Therefore in most cases that go to trial, both parties will legitimately believe they are right, and they have chosen to take the matter to court to prove it at risk of losing. The role of the court is to resolve their disagreement by making a decision as an authority that is binding. [NEWLINE] [NEWLINE] Bottom line: It is unlikely that people will risk bringing baseless lawsuits when they will bear the costs of their own legal fees when they lose. Cases will generally only go to trial if there is a legitimate question of responsibility.</s>
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Alright, I ran a more complete test and you are correct in that the prices increase more then 7%. <mask>, the number of steps doesn't change the overall increase in prices. [NEWLINE] Compare this 6 step model to this 12 step model. [NEWLINE] The percentage changes<mask> you change the cost of labor remain constant<mask><mask><mask> many steps you have in the process.  From tinkering with the spreadsheet, the effectiveness in changes in minimum wage seem more dependent on the labor:overhead ratio of the businesses involved, and the cost of raw materials.  Note that in this model, I assumed businesses would keep their profit margins static, instead of their profit per basket of goods.  This increases the final cost<mask> compared to keeping profits static,<mask> in the original post example. [NEWLINE] 6 step: [URL] ;output=html [NEWLINE] 12 step: [URL] ;output=html</s>
Label encoding: <s>Alright, I ran a more complete test and you are correct in that the prices increase more then 7%.  However, the number of steps doesn't change the overall increase in prices. [NEWLINE] Compare this 6 step model to this 12 step model. [NEWLINE] The percentage changes when you change the cost of labor remain constant regardless of how many steps you have in the process.  From tinkering with the spreadsheet, the effectiveness in changes in minimum wage seem more dependent on the labor:overhead ratio of the businesses involved, and the cost of raw materials.  Note that in this model, I assumed businesses would keep their profit margins static, instead of their profit per basket of goods.  This increases the final cost as compared to keeping profits static, as in the original post example. [NEWLINE] 6 step: [URL] ;output=html [NEWLINE] 12 step: [URL] ;output=html</s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] From videos I have seen of dogs, minks or monkeys etc being allowed out of a cage and seeing grass and seeing the sky for the first time of their lives, I have realised that they too have feelings and emotions and caging a human or other animal up for their whole lives is torture and should be totally illegal. [ENDQ] [NEWLINE] And then days later they starve to death<mask> they aren't fed, or they are easily predated<mask> they do not know that wild animals are trying to eat them. [NEWLINE] [NEWLINE] <mask> the animals are exotics, you are<mask> possibly wreaking ecological havoc on the native animal population.  See [exotic reptiles in the everglades]( [URL] ) for more information. [NEWLINE] [NEWLINE] You are not considering the other factors and issues of releasing animals into a habitat, and<mask><mask> it is the glaring flaw in your argument.</s>
Label encoding: <s> [STARTQ] From videos I have seen of dogs, minks or monkeys etc being allowed out of a cage and seeing grass and seeing the sky for the first time of their lives, I have realised that they too have feelings and emotions and caging a human or other animal up for their whole lives is torture and should be totally illegal. [ENDQ] [NEWLINE] And then days later they starve to death because they aren't fed, or they are easily predated because they do not know that wild animals are trying to eat them. [NEWLINE] [NEWLINE] If the animals are exotics, you are also possibly wreaking ecological havoc on the native animal population.  See [exotic reptiles in the everglades]( [URL] ) for more information. [NEWLINE] [NEWLINE] You are not considering the other factors and issues of releasing animals into a habitat, and I think it is the glaring flaw in your argument.</s>
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe that Michael Brown's behavior was affected having just committed a crime and being confronted by a police officer moments later. [ENDQ] [NEWLINE] This is exactly the problem with the release of this sort of information. You assume that it could offer an explanation<mask> to<mask> Brown could have provoked a physical confrontation with a police officer. The logic there relies on a HUGE assumption. Wouldn't it be equally likely<mask> not more probable to assume that a person would be LESS likely to physically provoke a police officer<mask> he just committed a crime, especially<mask> the police officer said nothing of a robbery and called them out only on jaywalking? [NEWLINE] [NEWLINE] Unless you're in a courtroom, a person's view of<mask> is "relevant" is wholly subjective. The release of this information is likely an attempt to exploit those subjective components of public opinion about criminals and criminal acts. </s>
Label encoding: <s> [STARTQ] I believe that Michael Brown's behavior was affected having just committed a crime and being confronted by a police officer moments later. [ENDQ] [NEWLINE] This is exactly the problem with the release of this sort of information. You assume that it could offer an explanation as to why Brown could have provoked a physical confrontation with a police officer. The logic there relies on a HUGE assumption. Wouldn't it be equally likely if not more probable to assume that a person would be LESS likely to physically provoke a police officer since he just committed a crime, especially if the police officer said nothing of a robbery and called them out only on jaywalking? [NEWLINE] [NEWLINE] Unless you're in a courtroom, a person's view of what is "relevant" is wholly subjective. The release of this information is likely an attempt to exploit those subjective components of public opinion about criminals and criminal acts. </s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This gets to be endless.  Who chooses the experts who choose the experts? [NEWLINE] [NEWLINE] Sure, you can have a panel of experts with different views,<mask> then, the decision is made by whichever opinion has more experts, which is more a factor of who chooses the experts than objective scientific assessment of the situation. [NEWLINE] [NEWLINE] <mask> will a statistician help you decide priorities?  Sure,<mask> you're going to decide which is more likely to have a return on your investment in terms of potential product development, at least in theory,<mask> some of it is that we want to KNOW our world.  There isn't an equation or a statistic to tell you whether it's better to understand subatomic physics or astronomy.  It's about the human quest for knowledge. [NEWLINE] [NEWLINE] Who participates in the referendum? <mask> are their qualifications? Who selects them? </s>
Label encoding: <s>This gets to be endless.  Who chooses the experts who choose the experts? [NEWLINE] [NEWLINE] Sure, you can have a panel of experts with different views, but then, the decision is made by whichever opinion has more experts, which is more a factor of who chooses the experts than objective scientific assessment of the situation. [NEWLINE] [NEWLINE] How will a statistician help you decide priorities?  Sure, if you're going to decide which is more likely to have a return on your investment in terms of potential product development, at least in theory, but some of it is that we want to KNOW our world.  There isn't an equation or a statistic to tell you whether it's better to understand subatomic physics or astronomy.  It's about the human quest for knowledge. [NEWLINE] [NEWLINE] Who participates in the referendum?  What are their qualifications? Who selects them? </s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hey, come on. I am interested in truth and would like to learn from you. I'm a scientist by education and training and this kind of debate is not my forte. Especially<mask> it becomes emotionally charged. Let's stay civil. [NEWLINE] [NEWLINE] I didn't deny<mask> you said about korea. Regarding Israel, it couldn't be<mask> it is today without the US,<mask><mask><mask> it was founded. Libya: Gaddaffi oppressed all political dissent against himself and had a terrible record on human rights. [NEWLINE] [NEWLINE] I am not attacking any of your positions on the terrible things the US has done in its self-promotion,<mask> I take issue with having a one-sided view. [NEWLINE] [NEWLINE] <mask><mask> I said, this is all<mask> the point. I'd like to hear a coherent impartial view:<mask> is democracy contradicted by capitalism?</s><pad><pad>
Label encoding: <s>Hey, come on. I am interested in truth and would like to learn from you. I'm a scientist by education and training and this kind of debate is not my forte. Especially when it becomes emotionally charged. Let's stay civil. [NEWLINE] [NEWLINE] I didn't deny what you said about korea. Regarding Israel, it couldn't be what it is today without the US, regardless of how it was founded. Libya: Gaddaffi oppressed all political dissent against himself and had a terrible record on human rights. [NEWLINE] [NEWLINE] I am not attacking any of your positions on the terrible things the US has done in its self-promotion, but I take issue with having a one-sided view. [NEWLINE] [NEWLINE] But as I said, this is all besides the point. I'd like to hear a coherent impartial view: Why is democracy contradicted by capitalism?</s><pad><pad>
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Addiction affects communities,<mask> opposed to just the individual.  Major addictions impair functionality, which then impairs the individual from many jobs.  Drugs, legal or regulated, cost money.  The idea here is that drug addiction leads to increased theft and/or violence.  I don't think the current system we have is appropriate for decreasing drug abuse,<mask> my argument is in reference to the idea that drug abuse only harms the user. [NEWLINE] [NEWLINE] Legalizing prostitution,<mask><mask><mask><mask>, has been shown to [increase illegal forms,]( [URL].ashx?id=14222&amp;p=0) such<mask> child prostitution and sex trafficking.  It's<mask> shown to be [correlated with increased violence]( [URL].ashx?id=8843&amp;p=0) in areas<mask> prostitution is legal.</s>
Label encoding: <s>Addiction affects communities, as opposed to just the individual.  Major addictions impair functionality, which then impairs the individual from many jobs.  Drugs, legal or regulated, cost money.  The idea here is that drug addiction leads to increased theft and/or violence.  I don't think the current system we have is appropriate for decreasing drug abuse, but my argument is in reference to the idea that drug abuse only harms the user. [NEWLINE] [NEWLINE] Legalizing prostitution, on the other hand, has been shown to [increase illegal forms,]( [URL].ashx?id=14222&amp;p=0) such as child prostitution and sex trafficking.  It's also shown to be [correlated with increased violence]( [URL].ashx?id=8843&amp;p=0) in areas where prostitution is legal.</s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] With more old people needing pensions and healthcare and with less young people to finance that [ENDQ] [NEWLINE] That's a problem with central planning, not birth rates. Of course,<mask> you assume that central planning is necessary, I can see<mask> you'd blame birth rates for its failure. In a free market, people pay for<mask> they use, not<mask> other people use,<mask> birth rates don't matter. [NEWLINE] [NEWLINE] [STARTQ] The other one that I heard being touted is that very soon, we will be living in a post scarcity society, and that population decrease will not have any consequences,<mask> I don't believe that it is feasible. [ENDQ] [NEWLINE] [NEWLINE] The rate at which we will reach something like a post-scarcity society will be slowed the more central planning there is<mask> it results in wealth being allocated away from productive people, which slows technological progress. </s>
Label encoding: <s> [STARTQ] With more old people needing pensions and healthcare and with less young people to finance that [ENDQ] [NEWLINE] That's a problem with central planning, not birth rates. Of course, if you assume that central planning is necessary, I can see why you'd blame birth rates for its failure. In a free market, people pay for what they use, not what other people use, so birth rates don't matter. [NEWLINE] [NEWLINE] [STARTQ] The other one that I heard being touted is that very soon, we will be living in a post scarcity society, and that population decrease will not have any consequences, but I don't believe that it is feasible. [ENDQ] [NEWLINE] [NEWLINE] The rate at which we will reach something like a post-scarcity society will be slowed the more central planning there is because it results in wealth being allocated away from productive people, which slows technological progress. </s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sexual relations do not "severely scar" young lives. Please stop regurgitating primitive, harmful, and empirically unsubstantiated mythology on the Internet. [NEWLINE] [NEWLINE] [STARTQ] A Dutch study published in 1987 found that a sample of boys in paedophilic relationships felt positively about them. And a major<mask> still controversial 1998-2000 meta-study suggests –<mask> J Michael Bailey of Northwestern University, Chicago, says – that such relationships, entered into voluntarily, are "nearly uncorrelated with undesirable outcomes". [ENDQ] [NEWLINE] [STARTQ] Most people find that idea impossible.<mask> writing last year in the peer-reviewed Archives of Sexual Behaviour, Bailey said that<mask> he<mask> found the notion "disturbing", he was forced to recognise that **"persuasive evidence for the harmfulness of paedophilic relationships does not<mask> exist".** [ENDQ] [NEWLINE] [URL] /</s>
Label encoding: <s>Sexual relations do not "severely scar" young lives. Please stop regurgitating primitive, harmful, and empirically unsubstantiated mythology on the Internet. [NEWLINE] [NEWLINE] [STARTQ] A Dutch study published in 1987 found that a sample of boys in paedophilic relationships felt positively about them. And a major if still controversial 1998-2000 meta-study suggests – as J Michael Bailey of Northwestern University, Chicago, says – that such relationships, entered into voluntarily, are "nearly uncorrelated with undesirable outcomes". [ENDQ] [NEWLINE] [STARTQ] Most people find that idea impossible. But writing last year in the peer-reviewed Archives of Sexual Behaviour, Bailey said that while he also found the notion "disturbing", he was forced to recognise that **"persuasive evidence for the harmfulness of paedophilic relationships does not yet exist".** [ENDQ] [NEWLINE] [URL] /</s>
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I'm outside and I want to check my email/look up directions/etc. on a laptop [ENDQ] [NEWLINE] <mask> is checking email outside *absolutely necessary* at that given moment? Does the fate and integrity of your job record rely on your instant email response? Is it not possible to wait until you return to a desk setting to check your email? [NEWLINE] [NEWLINE] A GPS can<mask> look up directions. [NEWLINE] [NEWLINE] [STARTQ] <mask>, it would suck a lot more for me to lose my $1000 dollar laptop with around $400 in software than it would to lose my couple-hundred-dollar smartphone. [ENDQ] [NEWLINE] It seems easier to lose a smartphone than a laptop, given<mask> small they are and are more accessible to one handed carelessness. Given the same amount of pampering, surely the smartphone is more prone to theft, loss, etc.</s>
Label encoding: <s> [STARTQ] If I'm outside and I want to check my email/look up directions/etc. on a laptop [ENDQ] [NEWLINE] But is checking email outside *absolutely necessary* at that given moment? Does the fate and integrity of your job record rely on your instant email response? Is it not possible to wait until you return to a desk setting to check your email? [NEWLINE] [NEWLINE] A GPS can also look up directions. [NEWLINE] [NEWLINE] [STARTQ] Also, it would suck a lot more for me to lose my $1000 dollar laptop with around $400 in software than it would to lose my couple-hundred-dollar smartphone. [ENDQ] [NEWLINE] It seems easier to lose a smartphone than a laptop, given how small they are and are more accessible to one handed carelessness. Given the same amount of pampering, surely the smartphone is more prone to theft, loss, etc.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You cannot believe WTC7<mask> a piece of evidence without implicating the FDNY is same way, shape, or form; which, considering the circumstances, is a rather scummy thing to do. At roughly 2:00 PM EST on 9/11, FDNY Chief Daniel Nigro created a collapse zone around the severely damaged WTC7. Nigro says it was he, and nobody else, that determined a collapse zone must be established to protect lives from being lost. He knew the building would collapse due to its structural integrity being lost. Whether they want to acknowledge it or not, "truthers" are implying that Nigro, and quite possibly, the rest of the FDNY, were complicit with the attacks. [NEWLINE] [NEWLINE] Furthermore, many in the AE911Truth group propagate that WTC7 fell at free fall speed, which is not true.</s>
Label encoding: <s>You cannot believe WTC7 as a piece of evidence without implicating the FDNY is same way, shape, or form; which, considering the circumstances, is a rather scummy thing to do. At roughly 2:00 PM EST on 9/11, FDNY Chief Daniel Nigro created a collapse zone around the severely damaged WTC7. Nigro says it was he, and nobody else, that determined a collapse zone must be established to protect lives from being lost. He knew the building would collapse due to its structural integrity being lost. Whether they want to acknowledge it or not, "truthers" are implying that Nigro, and quite possibly, the rest of the FDNY, were complicit with the attacks. [NEWLINE] [NEWLINE] Furthermore, many in the AE911Truth group propagate that WTC7 fell at free fall speed, which is not true.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Another Korean War would set back the peninsula to pre-1900's. I can't believe<mask> naive you are.  I have no idea<mask> you think a unified Korea would return to prominence.  Your logic is extremely flawed.  By the time Korea recovered from the war enough to focus on economy Vietnam,  Indonesia,  Malaysia,  Thailand, and even Bangladesh would have surpassed it. You would long dead of old age before Korea would get their legs back under them. The growth of South Korea was practically a miracle that will and cannot be repeated. Recovery would be a slow,  painful, and would not result in a stronger unified Korea. Grow up and look at reality instead of your fantastical view inside your head. Your post makes me utterly shocked and ashamed that another Korean American would think war would be good in the long run.  </s>
Label encoding: <s>Another Korean War would set back the peninsula to pre-1900's. I can't believe how naive you are.  I have no idea why you think a unified Korea would return to prominence.  Your logic is extremely flawed.  By the time Korea recovered from the war enough to focus on economy Vietnam,  Indonesia,  Malaysia,  Thailand, and even Bangladesh would have surpassed it. You would long dead of old age before Korea would get their legs back under them. The growth of South Korea was practically a miracle that will and cannot be repeated. Recovery would be a slow,  painful, and would not result in a stronger unified Korea. Grow up and look at reality instead of your fantastical view inside your head. Your post makes me utterly shocked and ashamed that another Korean American would think war would be good in the long run.  </s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0961, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Perhaps I'm an exception,<mask> I'd<mask><mask> for me, smoking pot has helped direct me to some thoughts that have actually accelerated my ambition in nearly all aspects of life. I realize that every human mind is different,<mask> I can't help<mask> think others have shared these types of experience with me. Regardless, you don't know<mask> types of self destructive thoughts plague each and every human. For many, cannabis can ease the demons, restore balance back to the mind. Its definitely not for everyone,<mask> it provides enough benefits that each adult should be able to decide for themselves whether or not they consume it. [NEWLINE] [NEWLINE] <mask>, I feel pretty strongly that the massive efficiencies of our current(america's) educational system closes more doors for people than smoking pot does. [NEWLINE] [NEWLINE] (I hope<mask><mask> gives you some amount of insight)</s>
Label encoding: <s>Perhaps I'm an exception, but I'd argue that for me, smoking pot has helped direct me to some thoughts that have actually accelerated my ambition in nearly all aspects of life. I realize that every human mind is different, but I can't help but think others have shared these types of experience with me. Regardless, you don't know what types of self destructive thoughts plague each and every human. For many, cannabis can ease the demons, restore balance back to the mind. Its definitely not for everyone, but it provides enough benefits that each adult should be able to decide for themselves whether or not they consume it. [NEWLINE] [NEWLINE] Additionally, I feel pretty strongly that the massive efficiencies of our current(america's) educational system closes more doors for people than smoking pot does. [NEWLINE] [NEWLINE] (I hope my opinion gives you some amount of insight)</s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Luvz2Spooje, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Luvz2Spooje+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry Luvz2Spooje, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Luvz2Spooje+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wow your buddy's job sucks. <mask><mask> that they should tell them that they need to hire certain groups of people based solely off of the criteria of their race, sex etc.  I mean don't not hire the qualified black woman etc who sought out an interview and looks like a good fit. <mask><mask> don't just going seeking a black woman etc for a position simply to satisfy a diversity hire some firm said  you needed.   That's pointless. [NEWLINE] [NEWLINE] <mask><mask> diversity hiring is a bad practice and is<mask> bad<mask> affirmative action.  I<mask> thinks it cheapens the minority candidate and sort of marginalizes them like we only need you here to fill a position<mask> our HR economist said we need more people like you,<mask> opposed to hiring them<mask> they can do the job well and came to you seeing work.</s>
Label encoding: <s>Wow your buddy's job sucks.  I disagree that they should tell them that they need to hire certain groups of people based solely off of the criteria of their race, sex etc.  I mean don't not hire the qualified black woman etc who sought out an interview and looks like a good fit.  But also don't just going seeking a black woman etc for a position simply to satisfy a diversity hire some firm said  you needed.   That's pointless. [NEWLINE] [NEWLINE] I think diversity hiring is a bad practice and is as bad as affirmative action.  I also thinks it cheapens the minority candidate and sort of marginalizes them like we only need you here to fill a position because our HR economist said we need more people like you, as opposed to hiring them because they can do the job well and came to you seeing work.</s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> there is no proof that the act of carrying out the measurement wasn't predetermined<mask> well. [ENDQ] [NEWLINE] Even<mask> the measurement was predetermined the state, that wave function collapses to on measurement, is necessarily random. [NEWLINE] [NEWLINE] [STARTQ] Just<mask> we aren't capable of tracing the history of an event mathematically doesn't mean it was random. [ENDQ] [NEWLINE] bells inequality and the experiments related to it show that there is no one history of the event<mask> its not measured.  Basically the measured probabilities of a couple events occurring must satisfy bells inequality<mask> they actually do have a single traceable history after a previous measurement. <mask> the probabilities we have experimentally found do break this inequality. [NEWLINE] [NEWLINE] Anyways are you convinced that there is no true randomness?<mask><mask><mask>? <mask> do you intuitively reject the possibility of true randomness?</s>
Label encoding: <s> [STARTQ] But there is no proof that the act of carrying out the measurement wasn't predetermined as well. [ENDQ] [NEWLINE] Even if the measurement was predetermined the state, that wave function collapses to on measurement, is necessarily random. [NEWLINE] [NEWLINE] [STARTQ] Just because we aren't capable of tracing the history of an event mathematically doesn't mean it was random. [ENDQ] [NEWLINE] bells inequality and the experiments related to it show that there is no one history of the event when its not measured.  Basically the measured probabilities of a couple events occurring must satisfy bells inequality if they actually do have a single traceable history after a previous measurement.  But the probabilities we have experimentally found do break this inequality. [NEWLINE] [NEWLINE] Anyways are you convinced that there is no true randomness? If so how?  Why do you intuitively reject the possibility of true randomness?</s>
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the only morally justifiable system. [ENDQ] [NEWLINE] The problem with this statement, and maybe it has already been pointed out, is that the way universal healthcare is enforced, at the very core makes it immoral<mask> it is implemented by the use of force. The only way for healthcare to be moral is<mask> it is completely voluntary system. Of course in the US, it's not moral the way it is currently run<mask> we are forced to have health insurance. The most moral place I've seen is this free market surgery center in Oklahoma. Their costs are low<mask> they don't take insurance at all, all their prices are posted online, and they don't have any overhead costs. No hospital administration fees, they don't over charge for supplies, etc.<mask> you want to look at something that is truly moral, this is it. [URL] /</s>
Label encoding: <s> [STARTQ] the only morally justifiable system. [ENDQ] [NEWLINE] The problem with this statement, and maybe it has already been pointed out, is that the way universal healthcare is enforced, at the very core makes it immoral because it is implemented by the use of force. The only way for healthcare to be moral is if it is completely voluntary system. Of course in the US, it's not moral the way it is currently run because we are forced to have health insurance. The most moral place I've seen is this free market surgery center in Oklahoma. Their costs are low because they don't take insurance at all, all their prices are posted online, and they don't have any overhead costs. No hospital administration fees, they don't over charge for supplies, etc. If you want to look at something that is truly moral, this is it. [URL] /</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0768, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ok,<mask> rights aren't something we do have; they're an ideal situation that we're trying to have [ENDQ] [NEWLINE] No, we do have the rights.  They are sometimes violated.  An "inalienable" right is not one which is literally impossible to violate, it is one which is always immoral to violate. [NEWLINE] [NEWLINE] [STARTQ] Well, my argument is that the entire model of "rights" is flawed; Of course no-one should be raped. No-one should rape. Everybody has the responsibility to not rape. [ENDQ] [NEWLINE] Every human might have that responsibility on an individual basis.<mask> the government's responsibilities are whatever we vote them to be. <mask><mask> we say that "even<mask> we vote to make rape a punishment, it would still be wrong" is to say you have a right not to be raped.</s>
Label encoding: <s> [STARTQ] Ok, so rights aren't something we do have; they're an ideal situation that we're trying to have [ENDQ] [NEWLINE] No, we do have the rights.  They are sometimes violated.  An "inalienable" right is not one which is literally impossible to violate, it is one which is always immoral to violate. [NEWLINE] [NEWLINE] [STARTQ] Well, my argument is that the entire model of "rights" is flawed; Of course no-one should be raped. No-one should rape. Everybody has the responsibility to not rape. [ENDQ] [NEWLINE] Every human might have that responsibility on an individual basis. But the government's responsibilities are whatever we vote them to be.  So if we say that "even if we vote to make rape a punishment, it would still be wrong" is to say you have a right not to be raped.</s>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I was just following orders" is a horrible excuse. [ENDQ] [NEWLINE] You're right. *We shouldn't be making excuses*. [NEWLINE] [NEWLINE] [STARTQ] Your ass is grass, regardless<mask> you decided to do it or some asshole ordered you to. [ENDQ] [NEWLINE] Number one, this is bullshit. Basically no one got punished for My Lai. [NEWLINE] [NEWLINE] Second, we shouldn't be punishing our army for *fighting a war*. "War crimes" are just bullshit we make up in order to execute foreign generals. Things that would be considered crimes happen, and should happen in war. It's bullshit to put irons on our military and limit them to doing only a specific set of military procedures. Vae Victis. [NEWLINE] [NEWLINE] [STARTQ] The military is about honor, courage, commitment. [ENDQ] [NEWLINE] The military is about projecting force internally and externally. </s>
Label encoding: <s> [STARTQ] I was just following orders" is a horrible excuse. [ENDQ] [NEWLINE] You're right. *We shouldn't be making excuses*. [NEWLINE] [NEWLINE] [STARTQ] Your ass is grass, regardless if you decided to do it or some asshole ordered you to. [ENDQ] [NEWLINE] Number one, this is bullshit. Basically no one got punished for My Lai. [NEWLINE] [NEWLINE] Second, we shouldn't be punishing our army for *fighting a war*. "War crimes" are just bullshit we make up in order to execute foreign generals. Things that would be considered crimes happen, and should happen in war. It's bullshit to put irons on our military and limit them to doing only a specific set of military procedures. Vae Victis. [NEWLINE] [NEWLINE] [STARTQ] The military is about honor, courage, commitment. [ENDQ] [NEWLINE] The military is about projecting force internally and externally. </s>
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1) Heat. [ENDQ] [NEWLINE] Plenty of reasons to cut a sandwich before you are ready to eat it. An example would be a social gathering<mask> the sandwiches wouldn't be immediately eaten. You wouldn't want to cut each of them<mask> you are ready to eat, you would cut them all up at the beginning<mask> someone can just grab them<mask> they feel like eating. [NEWLINE] [NEWLINE] [STARTQ] 2) Excess toppings. [ENDQ] [NEWLINE] That would completely change the toppings to bread ratio. Maybe you want less bread and more toppings. [NEWLINE] [NEWLINE] [STARTQ] 3) Messy meat styles. [ENDQ] [NEWLINE] More rustic bread slices<mask> work perfectly fine with pulled pork.<mask> even with traditional bread, you still ignored the other examples of pickle slices or shredded lettuce. And there are of course more options like egg or tuna salad. </s><pad>
Label encoding: <s> [STARTQ] 1) Heat. [ENDQ] [NEWLINE] Plenty of reasons to cut a sandwich before you are ready to eat it. An example would be a social gathering where the sandwiches wouldn't be immediately eaten. You wouldn't want to cut each of them as you are ready to eat, you would cut them all up at the beginning so someone can just grab them when they feel like eating. [NEWLINE] [NEWLINE] [STARTQ] 2) Excess toppings. [ENDQ] [NEWLINE] That would completely change the toppings to bread ratio. Maybe you want less bread and more toppings. [NEWLINE] [NEWLINE] [STARTQ] 3) Messy meat styles. [ENDQ] [NEWLINE] More rustic bread slices also work perfectly fine with pulled pork. But even with traditional bread, you still ignored the other examples of pickle slices or shredded lettuce. And there are of course more options like egg or tuna salad. </s><pad>
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0900, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because any attraction that I had previously had to math had always been<mask> you could follow a procedure to arrive at the right answer. There was this certain process you do, and<mask> you just follow the process correctly, you *will* get the right answer. It was black and white,<mask> opposed to creative and subjective like other subjects. That's<mask> I liked about it,<mask><mask> I got to the point<mask> we had to be inventing processes to solve the problems, I guess that kind of took away the magic for me of<mask> I'd always thought math was. [NEWLINE] [NEWLINE] TL;DR: I wanted math to be a simple path that you walk along and arrive at the answer by simply doing the process correctly. Finding out that it wasn't always like that and sometimes creativity and guesswork was needed irked me.</s>
Label encoding: <s>Because any attraction that I had previously had to math had always been how you could follow a procedure to arrive at the right answer. There was this certain process you do, and if you just follow the process correctly, you *will* get the right answer. It was black and white, as opposed to creative and subjective like other subjects. That's what I liked about it, but when I got to the point where we had to be inventing processes to solve the problems, I guess that kind of took away the magic for me of what I'd always thought math was. [NEWLINE] [NEWLINE] TL;DR: I wanted math to be a simple path that you walk along and arrive at the answer by simply doing the process correctly. Finding out that it wasn't always like that and sometimes creativity and guesswork was needed irked me.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As some have inferred, yes I am under 30. I accept that it may be a bit of a non-exposure bias,<mask> it is a self-confirming one,<mask> I have never had an urge to suffer through it until it's tolerable. [NEWLINE] [NEWLINE] From the multitude of responses, I gather that to appreciate it, I either have to be born with a taste for it or just have to suck it up and drink it until it's good? [NEWLINE] [NEWLINE] I honestly do hope it gets better<mask> I age.<mask> it is, I suppose I just have to accept at face value that with experience comes appreciation. [NEWLINE] [NEWLINE] I suppose my view has been changed,<mask> I wouldn't say by much. Granted, there is only<mask> much an internet debate can do for a subjective experience. [NEWLINE] [NEWLINE] ∆</s>
Label encoding: <s>As some have inferred, yes I am under 30. I accept that it may be a bit of a non-exposure bias, but it is a self-confirming one, as I have never had an urge to suffer through it until it's tolerable. [NEWLINE] [NEWLINE] From the multitude of responses, I gather that to appreciate it, I either have to be born with a taste for it or just have to suck it up and drink it until it's good? [NEWLINE] [NEWLINE] I honestly do hope it gets better as I age. As it is, I suppose I just have to accept at face value that with experience comes appreciation. [NEWLINE] [NEWLINE] I suppose my view has been changed, but I wouldn't say by much. Granted, there is only so much an internet debate can do for a subjective experience. [NEWLINE] [NEWLINE] ∆</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not justifying it- I'm explaining it with basic psychology. I'm not saying we shouldn't attempt to find a perfect solution<mask> at the moment that is the best solution we have- your metaphor with the doctor isn't far off<mask> it's more like "shit you've got cancer, we can't cure it<mask> we can give you chemotherapy, it's not perfect<mask> it's all we have". [NEWLINE] [NEWLINE] Sorry<mask> you are straight up wrong- murder and pedophilia are not products of evolution in the same way<mask> in-group vs out-group mentality is-<mask> can be evidenced by numerous pieces of research: [URL] those who don't murder<mask> instead have strong bonds to other out-groups perform better from an evolutionary perspective,<mask> time goes on this will continue to be true, it just takes time. </s>
Label encoding: <s>I'm not justifying it- I'm explaining it with basic psychology. I'm not saying we shouldn't attempt to find a perfect solution but at the moment that is the best solution we have- your metaphor with the doctor isn't far off but it's more like "shit you've got cancer, we can't cure it but we can give you chemotherapy, it's not perfect but it's all we have". [NEWLINE] [NEWLINE] Sorry but you are straight up wrong- murder and pedophilia are not products of evolution in the same way as in-group vs out-group mentality is- as can be evidenced by numerous pieces of research: [URL] those who don't murder but instead have strong bonds to other out-groups perform better from an evolutionary perspective, as time goes on this will continue to be true, it just takes time. </s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry IndignantChubbs, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=IndignantChubbs+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry IndignantChubbs, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=IndignantChubbs+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for the delta. [NEWLINE] [NEWLINE] Personally I wouldn't put on sunscreen on myself or my child every time we went outside (I don't actually have a child, only a theoretical one),<mask> it is an annoyance,<mask><mask> I'm outside doing work, hiking, or doing whatever outdoor activities out in the sun, I will put sunscreen on. It takes no more than a minute and lasts for several hours, and it conveniently almost eliminates any risk of sunburn (<mask><mask><mask> I reapply<mask> needed,<mask> reapplication is usually not necessary unless I'm outside all day).<mask><mask> simply never having to deal with sunburns is reason enough for me to wear sun screen. [NEWLINE] [NEWLINE] Edit:<mask> you can now get spray on sun screen which is 10x more convenient. It takes about 20 seconds to apply.</s>
Label encoding: <s>Thanks for the delta. [NEWLINE] [NEWLINE] Personally I wouldn't put on sunscreen on myself or my child every time we went outside (I don't actually have a child, only a theoretical one), as it is an annoyance, but if I'm outside doing work, hiking, or doing whatever outdoor activities out in the sun, I will put sunscreen on. It takes no more than a minute and lasts for several hours, and it conveniently almost eliminates any risk of sunburn ( so long as I reapply as needed, but reapplication is usually not necessary unless I'm outside all day). I think simply never having to deal with sunburns is reason enough for me to wear sun screen. [NEWLINE] [NEWLINE] Edit: also you can now get spray on sun screen which is 10x more convenient. It takes about 20 seconds to apply.</s>
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Lets just pretend that there was a murder. With attorney client privelege in place, the murderer safely goes to his attorney and confesses. The attorney, knowing this guy is screwed, talks to the police and negotiates leniency<mask> the guy turns himself in and confesses. The guy does<mask> and he goes to jail. [NEWLINE] [NEWLINE] Now we get rid of attorney client privledge The murderer doesn't talk to anyone and the police have to track him down and hope it doesn't result in a bloody shootout. [NEWLINE] [NEWLINE] [STARTQ] Less than 5 percent of criminal cases go to trial; most result in plea bargains. [URL] #sthash.AuLTgZ3L.dpuf [ENDQ] [NEWLINE] I don't believe that there would be many plea bargains<mask> criminals couldn't feel safe confiding in their attorneys</s>
Label encoding: <s>Lets just pretend that there was a murder. With attorney client privelege in place, the murderer safely goes to his attorney and confesses. The attorney, knowing this guy is screwed, talks to the police and negotiates leniency if the guy turns himself in and confesses. The guy does so and he goes to jail. [NEWLINE] [NEWLINE] Now we get rid of attorney client privledge The murderer doesn't talk to anyone and the police have to track him down and hope it doesn't result in a bloody shootout. [NEWLINE] [NEWLINE] [STARTQ] Less than 5 percent of criminal cases go to trial; most result in plea bargains. [URL] #sthash.AuLTgZ3L.dpuf [ENDQ] [NEWLINE] I don't believe that there would be many plea bargains if criminals couldn't feel safe confiding in their attorneys</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem I find it that (using tournament rules) you can't do all those fancy trades you're on about. [NEWLINE] [NEWLINE] You can **only** trade property + money in single transactions (not even sure<mask> pay back X per turn is allowed). [NEWLINE] [NEWLINE] The main problem I have is that the first person to get a set wins. That is unless one player basically **sacrifices** themselves to make the other one lose. [NEWLINE] [NEWLINE] <mask> of the nature of<mask> the trading is restricted, you have to make these sacrifices (you can't "work together" very easily). [NEWLINE] [NEWLINE] I enjoy a little bit of luck in a game,<mask> monopoly seems to have way too much. In any group of people any one person has an equal chance of winning<mask><mask><mask> they don't play stupidly.</s>
Label encoding: <s>The problem I find it that (using tournament rules) you can't do all those fancy trades you're on about. [NEWLINE] [NEWLINE] You can **only** trade property + money in single transactions (not even sure if pay back X per turn is allowed). [NEWLINE] [NEWLINE] The main problem I have is that the first person to get a set wins. That is unless one player basically **sacrifices** themselves to make the other one lose. [NEWLINE] [NEWLINE] Because of the nature of how the trading is restricted, you have to make these sacrifices (you can't "work together" very easily). [NEWLINE] [NEWLINE] I enjoy a little bit of luck in a game, but monopoly seems to have way too much. In any group of people any one person has an equal chance of winning as long as they don't play stupidly.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I dislike the way they make themselves look,<mask><mask> they're being innovative and changing the way we do everything for the better. [NEWLINE] Remember<mask> they said they simply could *not* take away internet connectivity<mask> "the console was built from the ground up around it",<mask> they managed to change it once everyone voiced their opinions? [NEWLINE] Microsoft with the xbox one is simply not worth it, the price is currently ~450$ for the console, plus a year of gold is 60$.<mask> you pay around 500$ to have the ability to watch netflix (<mask> you pay 8$ for netflix<mask> well) and browse the web. New AAA games are usually around 60-70$,<mask> you're almost at 600$<mask> you want to play a AAA title on xbone, and that's just one game.</s>
Label encoding: <s>I dislike the way they make themselves look, as if they're being innovative and changing the way we do everything for the better. [NEWLINE] Remember when they said they simply could *not* take away internet connectivity because "the console was built from the ground up around it", yet they managed to change it once everyone voiced their opinions? [NEWLINE] Microsoft with the xbox one is simply not worth it, the price is currently ~450$ for the console, plus a year of gold is 60$. So you pay around 500$ to have the ability to watch netflix ( if you pay 8$ for netflix as well) and browse the web. New AAA games are usually around 60-70$, so you're almost at 600$ if you want to play a AAA title on xbone, and that's just one game.</s>
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No it doesn't. [NEWLINE] [NEWLINE] [The Jet suit is real]( [URL] ) [NEWLINE] [NEWLINE] [The Exosuit is real]( [URL] ) [NEWLINE] [NEWLINE] The thing is, just<mask> it's fantastic doesn't mean it's *fantasy* the way you mean it. [NEWLINE] [NEWLINE] Did you know a lot of the tech that first showed up in the Star Trek TV series in the sixties *now exists*? [NEWLINE] [NEWLINE] Look at the cell phone, the hypospray, the ipad, **doors that open<mask> you walk up to them**. [NEWLINE] [NEWLINE] All of that *did not* exist<mask> that show was on tv. It does now. [NEWLINE] [NEWLINE] Here's a list of ten things - [URL] [NEWLINE] [NEWLINE] The thing about fiction is that it can give form to dreams that become reality later. </s>
Label encoding: <s>No it doesn't. [NEWLINE] [NEWLINE] [The Jet suit is real]( [URL] ) [NEWLINE] [NEWLINE] [The Exosuit is real]( [URL] ) [NEWLINE] [NEWLINE] The thing is, just because it's fantastic doesn't mean it's *fantasy* the way you mean it. [NEWLINE] [NEWLINE] Did you know a lot of the tech that first showed up in the Star Trek TV series in the sixties *now exists*? [NEWLINE] [NEWLINE] Look at the cell phone, the hypospray, the ipad, **doors that open when you walk up to them**. [NEWLINE] [NEWLINE] All of that *did not* exist when that show was on tv. It does now. [NEWLINE] [NEWLINE] Here's a list of ten things - [URL] [NEWLINE] [NEWLINE] The thing about fiction is that it can give form to dreams that become reality later. </s>
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0946, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Canada is<mask> a country with a population size similar to just one of America's states. In Europe every country is bordering other countries of similar size and population that they need to be aware of<mask> America only has Canada and Mexico, countries which it can afford to mostly ignore.<mask><mask> to that I've been to quite a bit of Canada and<mask> I liked it more than the US culturally it has mimicked a great deal of its southern neighbour. Mexico is different<mask> of the language barrier<mask> one could<mask><mask> it's the biggest importer of foreign media<mask> of the large number of Hispanic immigrants that are living there.<mask> overall I personally think that main issue that the US has is that it's just<mask> large that it's too self-sufficient for its own good<mask> it comes to its culture and media. </s>
Label encoding: <s>Canada is also a country with a population size similar to just one of America's states. In Europe every country is bordering other countries of similar size and population that they need to be aware of where America only has Canada and Mexico, countries which it can afford to mostly ignore. In addition to that I've been to quite a bit of Canada and while I liked it more than the US culturally it has mimicked a great deal of its southern neighbour. Mexico is different because of the language barrier but one could argue that it's the biggest importer of foreign media because of the large number of Hispanic immigrants that are living there. But overall I personally think that main issue that the US has is that it's just so large that it's too self-sufficient for its own good when it comes to its culture and media. </s>
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0858, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> you know,<mask> you are advocating for is not Voluntarism.  Voluntarism is a sect of anarcho-capitalist theory suggesting that the only ethical system of rule is one<mask> all interactions are voluntary.  In a voluntarist society, a government may exist,<mask><mask> it doesn't infringe on the rights of those uninterested and is supported entirely via voluntary contracts,<mask> it is entirely possible that no government would exist at all. <mask> there was a government, it would have no obligation to provide services or protections to anyone who doesn't voluntarily agree to its terms,<mask> an "obligation" would be coercive. [NEWLINE] [NEWLINE] Anyways,<mask><mask> with you mostly,<mask> you should definitely read more about Voluntarism and Anarcho-capitalism.  </s>
Label encoding: <s>Just so you know, what you are advocating for is not Voluntarism.  Voluntarism is a sect of anarcho-capitalist theory suggesting that the only ethical system of rule is one where all interactions are voluntary.  In a voluntarist society, a government may exist, provided that it doesn't infringe on the rights of those uninterested and is supported entirely via voluntary contracts, but it is entirely possible that no government would exist at all.  If there was a government, it would have no obligation to provide services or protections to anyone who doesn't voluntarily agree to its terms, as an "obligation" would be coercive. [NEWLINE] [NEWLINE] Anyways, I agree with you mostly, but you should definitely read more about Voluntarism and Anarcho-capitalism.  </s>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In areas of the South<mask> Christians reign supreme (Bible Belt) there are places teaching creationism<mask> fact and evolution is the fringe idea. I would think in a more perfect world the religious backlash to the ground zero mosque wouldn't have been allowed to take place or given space on the air waves. They NEVER had a proper claim to keep them from building a mosque several blocks away. [NEWLINE] [NEWLINE] <mask>, France is messed up. They are secular<mask> frankly more anti-religion than just allowing people to worship<mask> they like. People aren't allowed to wear religious symbols or clothing. That's ridiculous. I DO think that women that wear burkas should be required to remove them in banks (just<mask> you can't wear ski masks)<mask> well<mask> for driver's licenses and<mask> you're dealing with police.</s>
Label encoding: <s>In areas of the South where Christians reign supreme (Bible Belt) there are places teaching creationism as fact and evolution is the fringe idea. I would think in a more perfect world the religious backlash to the ground zero mosque wouldn't have been allowed to take place or given space on the air waves. They NEVER had a proper claim to keep them from building a mosque several blocks away. [NEWLINE] [NEWLINE] Also, France is messed up. They are secular but frankly more anti-religion than just allowing people to worship what they like. People aren't allowed to wear religious symbols or clothing. That's ridiculous. I DO think that women that wear burkas should be required to remove them in banks (just as you can't wear ski masks) as well as for driver's licenses and when you're dealing with police.</s>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The impressing goes both ways - someone who is stuck up about typos in a resume is clearly somebody who has their priorities backwards and is not worth working for.<mask> would I want to be part of an organization that weighs inconsequential mistakes with the maximum possible severity?<mask> would I provide value to a company that inefficient? Are we going to have a 6 hour review each week of<mask> I tie my shoelaces too? [NEWLINE] [NEWLINE] The impression is arbitrary - are you<mask> shredding resumes of people who don't use aesthetically pleasing margin sizes?  MLA standards? <mask> someone can't be bothered to google<mask> is the most common &amp; formal font type, clearly they don't put even a modicum of research into things before going into a situation, and<mask> will make a poor employee.</s>
Label encoding: <s>The impressing goes both ways - someone who is stuck up about typos in a resume is clearly somebody who has their priorities backwards and is not worth working for. Why would I want to be part of an organization that weighs inconsequential mistakes with the maximum possible severity? Why would I provide value to a company that inefficient? Are we going to have a 6 hour review each week of how I tie my shoelaces too? [NEWLINE] [NEWLINE] The impression is arbitrary - are you also shredding resumes of people who don't use aesthetically pleasing margin sizes?  MLA standards?  If someone can't be bothered to google what is the most common &amp; formal font type, clearly they don't put even a modicum of research into things before going into a situation, and therefore will make a poor employee.</s>
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You, or someone in your family, are more likely to be killed by your gun either purposefully or accidentally than you are to use the gun against an attacker successfully in self-defense. [ENDQ] [NEWLINE] This is 100% false and based on a purposefully exclusionary study that excludes most forms of defensive gun use. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] <mask> you look through those stats real quick you will see that you are more likely to defend yourself with a gun than you are to die from a gun by every estimate done that includes all forms of self-defense with a gun. Some of the stats put gun self-defense higher than gun assault. [NEWLINE] [NEWLINE] The worst part about this is you haven't even posted a source to support your claim. Which makes me question your actual understanding of the stat you are referencing.</s>
Label encoding: <s> [STARTQ] You, or someone in your family, are more likely to be killed by your gun either purposefully or accidentally than you are to use the gun against an attacker successfully in self-defense. [ENDQ] [NEWLINE] This is 100% false and based on a purposefully exclusionary study that excludes most forms of defensive gun use. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] If you look through those stats real quick you will see that you are more likely to defend yourself with a gun than you are to die from a gun by every estimate done that includes all forms of self-defense with a gun. Some of the stats put gun self-defense higher than gun assault. [NEWLINE] [NEWLINE] The worst part about this is you haven't even posted a source to support your claim. Which makes me question your actual understanding of the stat you are referencing.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You can break up with her and appease your parents,<mask> you can't just not be gay. [ENDQ] [NEWLINE] That may be true,<mask> I don't see<mask> it supports<mask> you said: [NEWLINE] [NEWLINE] [STARTQ] <mask> homosexuality is a choice, it becomes less awful for me to hate you for making choices<mask><mask> with rather than something you just naturally are. [ENDQ] [NEWLINE] <mask><mask> most reasonable people would have a serious problem both with a kid getting kicked out of his home for being gay and a kid being kicked out of his home for dating someone of the wrong race. Heck, show me a video of a christian kid being kicked out of an atheist or muslim household, and I'll show you Fox News's top story for the next month,<mask> a shit-ton of people would bat an eye.</s>
Label encoding: <s> [STARTQ] You can break up with her and appease your parents, but you can't just not be gay. [ENDQ] [NEWLINE] That may be true, but I don't see how it supports what you said: [NEWLINE] [NEWLINE] [STARTQ] If homosexuality is a choice, it becomes less awful for me to hate you for making choices I disagree with rather than something you just naturally are. [ENDQ] [NEWLINE] I think most reasonable people would have a serious problem both with a kid getting kicked out of his home for being gay and a kid being kicked out of his home for dating someone of the wrong race. Heck, show me a video of a christian kid being kicked out of an atheist or muslim household, and I'll show you Fox News's top story for the next month, because a shit-ton of people would bat an eye.</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0918, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, a good rider should be aware of, and spend very little time in any blind spots of co-moving traffic -- unless traffic is heavily congested and this is impossible. Blind spots of traffic approaching from other roads should be accounted for<mask> those vehicles are visible, and a path and speed picked that gives the driver a very small window to collide with the rider, even<mask> they were actively trying to do<mask>. Obstacles that block an entire vehicle could easily mask or block noise, and you can't rely on someone hearing you in good time at a difference of 60km/h from the inside of most modern vehicles (which provide very good sound isolation). [NEWLINE] [NEWLINE] <mask>, the visual cue is much more likely to get you noticed *before* you enter said blind spot,<mask> they actually do help.</s><pad>
Label encoding: <s>Firstly, a good rider should be aware of, and spend very little time in any blind spots of co-moving traffic -- unless traffic is heavily congested and this is impossible. Blind spots of traffic approaching from other roads should be accounted for when those vehicles are visible, and a path and speed picked that gives the driver a very small window to collide with the rider, even if they were actively trying to do so. Obstacles that block an entire vehicle could easily mask or block noise, and you can't rely on someone hearing you in good time at a difference of 60km/h from the inside of most modern vehicles (which provide very good sound isolation). [NEWLINE] [NEWLINE] Secondly, the visual cue is much more likely to get you noticed *before* you enter said blind spot, so they actually do help.</s><pad>
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's the thing your argument fails at: you talk about wealth rather than the rate of resource usage. [NEWLINE] [NEWLINE] Western nations use resources at a terrifying rate, a rate which is entirely unsustainable.<mask> we were to carry on at such a rate then, yes, we would<mask> be impoverishing the non-western nations<mask> they would never get a chance to use those resources. [NEWLINE] [NEWLINE] <mask><mask> you could really benefit from reading a few texts which cover the spirit of your subject: [NEWLINE] [NEWLINE] * [The Limits to Growth]( [URL] ) - the book, not just the Wiki entry. [NEWLINE] * [Without the hot air]( [URL] /) -<mask> we're fundamentally talking about resource and energy usage. [NEWLINE] * [Steady State Economy]( [URL] ) - all non-expansionary solutions fall into this category.</s>
Label encoding: <s>Here's the thing your argument fails at: you talk about wealth rather than the rate of resource usage. [NEWLINE] [NEWLINE] Western nations use resources at a terrifying rate, a rate which is entirely unsustainable. If we were to carry on at such a rate then, yes, we would indeed be impoverishing the non-western nations because they would never get a chance to use those resources. [NEWLINE] [NEWLINE] I think you could really benefit from reading a few texts which cover the spirit of your subject: [NEWLINE] [NEWLINE] * [The Limits to Growth]( [URL] ) - the book, not just the Wiki entry. [NEWLINE] * [Without the hot air]( [URL] /) - since we're fundamentally talking about resource and energy usage. [NEWLINE] * [Steady State Economy]( [URL] ) - all non-expansionary solutions fall into this category.</s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't feel comfortable telling you<mask> many children to have,<mask><mask> you asked. [NEWLINE] [NEWLINE] <mask> the only reason preventing you for having another child is financial, I'm not convinced that's a good reason. Your child is 6 weeks old. He or she will be going to college in 18 years. You could have a new job in 18 years. It could be a much better job than your current one. It could be much worse. You never really know<mask>'s going to happen in the future. [NEWLINE] [NEWLINE] The main advantage to having more than one child that springs to my childless mind is that having a sibling dramatically improves both children's social development. They learn to share much quicker. They become used to not being the center of attention. They compete against each other, encouraging growth. </s>
Label encoding: <s>I don't feel comfortable telling you how many children to have, but since you asked. [NEWLINE] [NEWLINE] If the only reason preventing you for having another child is financial, I'm not convinced that's a good reason. Your child is 6 weeks old. He or she will be going to college in 18 years. You could have a new job in 18 years. It could be a much better job than your current one. It could be much worse. You never really know what's going to happen in the future. [NEWLINE] [NEWLINE] The main advantage to having more than one child that springs to my childless mind is that having a sibling dramatically improves both children's social development. They learn to share much quicker. They become used to not being the center of attention. They compete against each other, encouraging growth. </s>
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0871, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"You're a smarter person just from existing there" [NEWLINE] [NEWLINE] No, you're not, you're just better at dealing with New York.. This is the thing I hate the most about New York, everybody there just thinks that they are superior<mask> they understand<mask> to get around the city. New Yorkers think everybody who can't figure out<mask> to use the daunting public transportation system their first weekend there is stupid..naive...inferior etc. Guess<mask>,<mask> you were born and raised in New York, you will be equally<mask> lost the first time you visit Chicago, or L.A. or Topeka fucking Kansas. Just living somewhere does not make you a better person. New York is a great city<mask> you don't become great by moving there, you just become insufferable.</s>
Label encoding: <s>"You're a smarter person just from existing there" [NEWLINE] [NEWLINE] No, you're not, you're just better at dealing with New York.. This is the thing I hate the most about New York, everybody there just thinks that they are superior because they understand how to get around the city. New Yorkers think everybody who can't figure out how to use the daunting public transportation system their first weekend there is stupid..naive...inferior etc. Guess what, If you were born and raised in New York, you will be equally as lost the first time you visit Chicago, or L.A. or Topeka fucking Kansas. Just living somewhere does not make you a better person. New York is a great city but you don't become great by moving there, you just become insufferable.</s>
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We are only 4 comments in and you already invoke [Hitler]( [URL] %27s_law)?  Can't you come up with something better? [NEWLINE] [NEWLINE] [STARTQ] this is a website, a forum, a community one that wants to be the front page of the internet<mask> people can "think" elsewhere. [ENDQ] [NEWLINE] <mask> should one set of people move and not the other?  9gag is a clean version of Reddit. [NEWLINE] [NEWLINE] [STARTQ] No literally aiming.<mask>, /wehatecatphobes. A place to focus hate on a particular human target. [ENDQ] [NEWLINE] I don't understand this at all.  Can you clarify? [NEWLINE] [NEWLINE] And<mask> about negative and intent examples I gave that you asked for. <mask> are the intent of these and should they be banned? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>We are only 4 comments in and you already invoke [Hitler]( [URL] %27s_law)?  Can't you come up with something better? [NEWLINE] [NEWLINE] [STARTQ] this is a website, a forum, a community one that wants to be the front page of the internet so people can "think" elsewhere. [ENDQ] [NEWLINE] Why should one set of people move and not the other?  9gag is a clean version of Reddit. [NEWLINE] [NEWLINE] [STARTQ] No literally aiming. So, /wehatecatphobes. A place to focus hate on a particular human target. [ENDQ] [NEWLINE] I don't understand this at all.  Can you clarify? [NEWLINE] [NEWLINE] And what about negative and intent examples I gave that you asked for.  What are the intent of these and should they be banned? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's interesting<mask> many people don't even know the actual rules of the game. It's one of the few games<mask> house rules are more common than the actual rules. [NEWLINE] [NEWLINE] I actually didn't know the real rules until I tried playing the video game. I was surprised that you have to auction and realized that there is no money on free parking and that you really should trade to get ahead. This puts a lot of strategy in the game,<mask> I've played very few live games<mask> anyone wanted to follow the real rules. Most people I've met hate trading<mask> well (possibly<mask> we don't really haggle in America outside of garage sales and used car dealerships). It's too bad,<mask> it does make the game a lot better to play by the real rules. </s>
Label encoding: <s>It's interesting how many people don't even know the actual rules of the game. It's one of the few games where house rules are more common than the actual rules. [NEWLINE] [NEWLINE] I actually didn't know the real rules until I tried playing the video game. I was surprised that you have to auction and realized that there is no money on free parking and that you really should trade to get ahead. This puts a lot of strategy in the game, but I've played very few live games where anyone wanted to follow the real rules. Most people I've met hate trading as well (possibly because we don't really haggle in America outside of garage sales and used car dealerships). It's too bad, because it does make the game a lot better to play by the real rules. </s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There aren't necessarily many easy sources for seeds. [ENDQ] [NEWLINE] OK,<mask> i come from (France), there is really 0 problem in getting seeds of endemic varieties, i have a lot of friends who are big on the organic farming thing, and they have no shortage of traditional seeds to work with. [NEWLINE] I don't see<mask> Monsanto could "get rid of reuse of any seeds", except by offering a better, more profitable alternative. [NEWLINE] [NEWLINE] [STARTQ] The other thing is Monsanto sues farmers that don't buy from them [ENDQ] [NEWLINE] I was under the impression that this is very anecdotal, do you have any figures to share about this phenomenon please?<mask>, i don't see<mask> stops farmers from associating and protecting each other against the Big Guy, just like every industry does, all the time?</s>
Label encoding: <s> [STARTQ] There aren't necessarily many easy sources for seeds. [ENDQ] [NEWLINE] OK, where i come from (France), there is really 0 problem in getting seeds of endemic varieties, i have a lot of friends who are big on the organic farming thing, and they have no shortage of traditional seeds to work with. [NEWLINE] I don't see how Monsanto could "get rid of reuse of any seeds", except by offering a better, more profitable alternative. [NEWLINE] [NEWLINE] [STARTQ] The other thing is Monsanto sues farmers that don't buy from them [ENDQ] [NEWLINE] I was under the impression that this is very anecdotal, do you have any figures to share about this phenomenon please? Also, i don't see what stops farmers from associating and protecting each other against the Big Guy, just like every industry does, all the time?</s>
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> your beliefs are really rocked that hard by<mask> someone says/writes, maybe you should re-evaluate your beliefs instead of attacking someone for expressing their own. [ENDQ] [NEWLINE] <mask> I take offense to someone claiming my race is one of inferiority, which belief should I be reevaluating? <mask> I am offended<mask> someone claims a family member is something they are not, which of my beliefs are wrong? [NEWLINE] [NEWLINE] [STARTQ] There are<mask> many examples of the danger and horrific power of being offended. I can list them<mask> you like<mask> I don't even have to, you can think of enough on your own. [ENDQ] [NEWLINE] Or maybe you can actually establish your view for people to respond to, instead of expecting us to come up with our own argument then try to refute it. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] If your beliefs are really rocked that hard by what someone says/writes, maybe you should re-evaluate your beliefs instead of attacking someone for expressing their own. [ENDQ] [NEWLINE] If I take offense to someone claiming my race is one of inferiority, which belief should I be reevaluating?  If I am offended because someone claims a family member is something they are not, which of my beliefs are wrong? [NEWLINE] [NEWLINE] [STARTQ] There are so many examples of the danger and horrific power of being offended. I can list them if you like but I don't even have to, you can think of enough on your own. [ENDQ] [NEWLINE] Or maybe you can actually establish your view for people to respond to, instead of expecting us to come up with our own argument then try to refute it. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm a person who prefers (and in some cases, really needs) trigger warnings for things depicting or discussing rape/sexual abuse/domestic violence, just<mask> of my past experiences. In cases like mine<mask> certain images trigger anxiety, panic, self-destructive behavior, etc, trigger warnings are most certainly needed. [NEWLINE] [NEWLINE] <mask>, on tumblr I see other people my age triggering things like eye contact, food, water, and even images of other people. This hurts me personally<mask> I now see people, especially on Reddit, completely mocking the idea of a trigger warning<mask> a whole.<mask> yeah, trigger warnings should really only be reserved for those who have had traumatic experiences.<mask> now I can't even admit my triggers without people mocking me or looking at me funny, which sucks.</s>
Label encoding: <s>I'm a person who prefers (and in some cases, really needs) trigger warnings for things depicting or discussing rape/sexual abuse/domestic violence, just because of my past experiences. In cases like mine where certain images trigger anxiety, panic, self-destructive behavior, etc, trigger warnings are most certainly needed. [NEWLINE] [NEWLINE] However, on tumblr I see other people my age triggering things like eye contact, food, water, and even images of other people. This hurts me personally because I now see people, especially on Reddit, completely mocking the idea of a trigger warning as a whole. So yeah, trigger warnings should really only be reserved for those who have had traumatic experiences. Because now I can't even admit my triggers without people mocking me or looking at me funny, which sucks.</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I disagree; the problem is not that singular they already exists (and has for centuries, possibly millenia, I'll have to check), the problem is that the term "it" specifically denies personhood to the thing it is replacing. [NEWLINE] [NEWLINE] Random wild dog?  That's an it,<mask> you don't care about it having feelings, or even existing for the most part. [NEWLINE] *Your* dog?  He/She, even<mask> they've been spayed/neutered. [NEWLINE] [NEWLINE] The reason it's uncomfortable applying "it" to a being that is *known* to be a person triggers a reaction that is not dissimilar to saying something like "he was jumps;" it's close,<mask> one parameter off, doesn't agree with<mask> it's supposed to.</s>
Label encoding: <s>I disagree; the problem is not that singular they already exists (and has for centuries, possibly millenia, I'll have to check), the problem is that the term "it" specifically denies personhood to the thing it is replacing. [NEWLINE] [NEWLINE] Random wild dog?  That's an it, because you don't care about it having feelings, or even existing for the most part. [NEWLINE] *Your* dog?  He/She, even if they've been spayed/neutered. [NEWLINE] [NEWLINE] The reason it's uncomfortable applying "it" to a being that is *known* to be a person triggers a reaction that is not dissimilar to saying something like "he was jumps;" it's close, but one parameter off, doesn't agree with what it's supposed to.</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I kind of agree with this<mask> you have to remember that "gay" has had different meanings throughout history. It originally meant you were joyous then some<mask> it turned into a word to describe homosexuals. Many words that are considered to be "bad" words had different meanings at one point.<mask> you call someone a dick or a cunt you don't actually think that person is genitalia.<mask> I called someone gay<mask> I was younger, did I actually think that person was a homosexual? [NEWLINE] [NEWLINE] It is different<mask> you call someone a derogatory word to their face such<mask> faggot and queer. You are actually making an attempt to bully that person. <mask> calling someone gay jokingly hurts no one. [NEWLINE] [NEWLINE] The real problem in my eyes is that we are too easily offended.</s>
Label encoding: <s>I kind of agree with this but you have to remember that "gay" has had different meanings throughout history. It originally meant you were joyous then some how it turned into a word to describe homosexuals. Many words that are considered to be "bad" words had different meanings at one point. When you call someone a dick or a cunt you don't actually think that person is genitalia. When I called someone gay when I was younger, did I actually think that person was a homosexual? [NEWLINE] [NEWLINE] It is different when you call someone a derogatory word to their face such as faggot and queer. You are actually making an attempt to bully that person.  While calling someone gay jokingly hurts no one. [NEWLINE] [NEWLINE] The real problem in my eyes is that we are too easily offended.</s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Stress is caused by your body releasing a hormone called cortisol in issues of fight or flight, which acts on the brain the keep it alert and stave off drowsiness, and on the muscles to keep them firing and ready to go past recommended energy usage. You can imagine<mask> useful that was in the past,<mask> today is a much safer world, and allowing stress to have<mask> much influence over one's behaviour<mask> it did then is not only pointless,<mask> dangerous,<mask> your body learns to recognise situations in which cortisol was released and repeats the process, in completely safe situations, and can cause one to act inappropriately.<mask> it is part of a more complex reaction including adrenaline there are ways to control it, including sensory reduction (closing your eyes, for example) and controlled breathing.</s>
Label encoding: <s>Stress is caused by your body releasing a hormone called cortisol in issues of fight or flight, which acts on the brain the keep it alert and stave off drowsiness, and on the muscles to keep them firing and ready to go past recommended energy usage. You can imagine how useful that was in the past, but today is a much safer world, and allowing stress to have as much influence over one's behaviour as it did then is not only pointless, but dangerous, because your body learns to recognise situations in which cortisol was released and repeats the process, in completely safe situations, and can cause one to act inappropriately. As it is part of a more complex reaction including adrenaline there are ways to control it, including sensory reduction (closing your eyes, for example) and controlled breathing.</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's not like the losing team are losing their homes or livelihoods. [ENDQ] [NEWLINE] Depends on the country and the importance of the match. On the more extreme end, North Korea likely imprisoned/tortured/killed it's FIFA soccer team for losing a televised match. Less extreme, medalling and not at the Olympics can mean the difference between endorsement deals and, well, nothing to show for your decades of effort ever. [NEWLINE] [NEWLINE] <mask> to the dream thing, it's all about expectations. It's different<mask> something that's clearly in reach is snatched away. That's<mask> most times you see athletes crying it's after an exceedingly close game or match, not<mask> they went in knowing they were probably going to lose 10-0 (see Women's Fencing, etc). </s>
Label encoding: <s> [STARTQ] It's not like the losing team are losing their homes or livelihoods. [ENDQ] [NEWLINE] Depends on the country and the importance of the match. On the more extreme end, North Korea likely imprisoned/tortured/killed it's FIFA soccer team for losing a televised match. Less extreme, medalling and not at the Olympics can mean the difference between endorsement deals and, well, nothing to show for your decades of effort ever. [NEWLINE] [NEWLINE] As to the dream thing, it's all about expectations. It's different when something that's clearly in reach is snatched away. That's why most times you see athletes crying it's after an exceedingly close game or match, not when they went in knowing they were probably going to lose 10-0 (see Women's Fencing, etc). </s>
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It should be a legal issue<mask> public urination on walls and on buildings leaves permanent marks and smells. Ever walk down a piss soaked wall. Even<mask> it was not recent the smells are still there.<mask> it is your restaurant, pub, hotel, or<mask> it belongs to your cities government, you have to pay money to clean it up. There are negative connotations with the smell and marks of piss. It makes people think the area is run down or not hygienic.<mask> people were not charged than the owners of the areas that were pissed on would have to pay money every time someone felt like their place was a good place to piss or face the consequence of loosing perspective customers. Making it illegal serves<mask> a disincentive for people doing it again to these owners.</s>
Label encoding: <s>It should be a legal issue as public urination on walls and on buildings leaves permanent marks and smells. Ever walk down a piss soaked wall. Even if it was not recent the smells are still there. If it is your restaurant, pub, hotel, or if it belongs to your cities government, you have to pay money to clean it up. There are negative connotations with the smell and marks of piss. It makes people think the area is run down or not hygienic. If people were not charged than the owners of the areas that were pissed on would have to pay money every time someone felt like their place was a good place to piss or face the consequence of loosing perspective customers. Making it illegal serves as a disincentive for people doing it again to these owners.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry andero, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=andero+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry andero, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=andero+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I've never been a father,<mask> I still don't see that having any weight in the matter. [ENDQ] [NEWLINE] <mask> not? A fetus is<mask> a potential child of his father. [NEWLINE] [NEWLINE] Sure,<mask> the mother is the one carrying the fetus, she gets to have most of the choice regarding that fetus. [NEWLINE] [NEWLINE] <mask> clearly, at some point (e.g. viability of the fetus outside of the womb) a father (and the society at large) can have some interest in the well-being of this (potential) child. [NEWLINE] [NEWLINE] For example, it would be a hard choice to make abortion legal one day before the due date for no medical reason. Sure,<mask> a society we might allow such abortions,<mask><mask> should men be excluded from this decision?</s>
Label encoding: <s> [STARTQ] I've never been a father, but I still don't see that having any weight in the matter. [ENDQ] [NEWLINE] Why not? A fetus is also a potential child of his father. [NEWLINE] [NEWLINE] Sure, because the mother is the one carrying the fetus, she gets to have most of the choice regarding that fetus. [NEWLINE] [NEWLINE] But clearly, at some point (e.g. viability of the fetus outside of the womb) a father (and the society at large) can have some interest in the well-being of this (potential) child. [NEWLINE] [NEWLINE] For example, it would be a hard choice to make abortion legal one day before the due date for no medical reason. Sure, as a society we might allow such abortions, but why should men be excluded from this decision?</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sometimes, it is simply more fun to scramble and scrum! Sometimes, getting there is more important than the actual destination. [NEWLINE] [NEWLINE] For example: [NEWLINE] [NEWLINE] People enjoy the bouquet throwing tradition at the wedding. Single women line up and try to catch the bouquet in a "scram"  fashion: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Would  it be fun<mask> bridesmaids queued up, and the first one in line got the bouquet? [NEWLINE] [NEWLINE] Similarly, I would<mask><mask> in certain venues (e.g. bars) creating the atmosphere of chaos is just part of the fun, part of the experience (like in the  bouquet throw example). [NEWLINE] [NEWLINE] P.S.<mask><mask> the word "queue" has too many vowels. CMV.</s>
Label encoding: <s>Sometimes, it is simply more fun to scramble and scrum! Sometimes, getting there is more important than the actual destination. [NEWLINE] [NEWLINE] For example: [NEWLINE] [NEWLINE] People enjoy the bouquet throwing tradition at the wedding. Single women line up and try to catch the bouquet in a "scram"  fashion: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Would  it be fun if bridesmaids queued up, and the first one in line got the bouquet? [NEWLINE] [NEWLINE] Similarly, I would argue that in certain venues (e.g. bars) creating the atmosphere of chaos is just part of the fun, part of the experience (like in the  bouquet throw example). [NEWLINE] [NEWLINE] P.S. I think the word "queue" has too many vowels. CMV.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No you're not, you're a racist troll from SRS: [NEWLINE] [NEWLINE] [STARTQ] &gt;I post in SRS,<mask> not for the reasons you might think. [ENDQ] [NEWLINE] [STARTQ] &gt;I've done the math, and being 'from SRS' is literally the most efficient way to infuriate white people on the internet. Nothing works better. Now,<mask> a part time troll, this makes being an 'SRSter' a valuable characteristic. [ENDQ] [NEWLINE] [STARTQ] &gt;I<mask> like the social justice stuff,<mask> come on, everyone knows that SRS prime is ridiculous. Its only job is to piss off white people, and<mask><mask> we have succeeded. [ENDQ] [NEWLINE] [STARTQ] &gt;TL; DR Fuck you, honkey. [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>No you're not, you're a racist troll from SRS: [NEWLINE] [NEWLINE] [STARTQ] &gt;I post in SRS, but not for the reasons you might think. [ENDQ] [NEWLINE] [STARTQ] &gt;I've done the math, and being 'from SRS' is literally the most efficient way to infuriate white people on the internet. Nothing works better. Now, as a part time troll, this makes being an 'SRSter' a valuable characteristic. [ENDQ] [NEWLINE] [STARTQ] &gt;I also like the social justice stuff, but come on, everyone knows that SRS prime is ridiculous. Its only job is to piss off white people, and I think we have succeeded. [ENDQ] [NEWLINE] [STARTQ] &gt;TL; DR Fuck you, honkey. [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, the Guernica link is a bit of a stretch,<mask> historically art is full of depictions of unrealistically "perfect" people - "perfect" obviously being massively subjective, dependent upon depictor, culture, time period etc. [NEWLINE] [NEWLINE] <mask><mask> it's about personal taste.<mask> someone with tattoos, I wouldn't get any in that particular style,<mask> of course it's up to the individual. I mean, shit, even<mask> it is offensive, inappropriate, racist, sexist or whatever,<mask><mask> it's dangerous to start dictating<mask> people can do to their own body. [NEWLINE] [NEWLINE] <mask> for whether it's offensive, clearly it is to some people.<mask> that in itself is not an indicator that it's wrong. People get offended by all kinds of things.</s>
Label encoding: <s>Yeah, the Guernica link is a bit of a stretch, but historically art is full of depictions of unrealistically "perfect" people - "perfect" obviously being massively subjective, dependent upon depictor, culture, time period etc. [NEWLINE] [NEWLINE] I agree it's about personal taste. As someone with tattoos, I wouldn't get any in that particular style, but of course it's up to the individual. I mean, shit, even if it is offensive, inappropriate, racist, sexist or whatever, I think it's dangerous to start dictating what people can do to their own body. [NEWLINE] [NEWLINE] As for whether it's offensive, clearly it is to some people. But that in itself is not an indicator that it's wrong. People get offended by all kinds of things.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[A recent Freakonomics podcast]( [URL] /) looked into the ROI of learning a foreign language - the conclusion is that for an American it will only increase their salaries by something like 2% (less for Spanish<mask> it's<mask> common, more for some of the more exotic options). [NEWLINE] [NEWLINE] By contrast, the ROI for people in places like Russia to learn *English*<mask> a second language is much higher - 15-20%. [NEWLINE] [NEWLINE] <mask> you tell me: which one makes more sense? [NEWLINE] [NEWLINE] (FWIW I don't agree that English should be made the official language, nor that various government entities should stop offering translators, etc.<mask><mask> for pushing for everyone in the US to learn a second language - really doesn't seem to be worth it.)</s>
Label encoding: <s>[A recent Freakonomics podcast]( [URL] /) looked into the ROI of learning a foreign language - the conclusion is that for an American it will only increase their salaries by something like 2% (less for Spanish because it's so common, more for some of the more exotic options). [NEWLINE] [NEWLINE] By contrast, the ROI for people in places like Russia to learn *English* as a second language is much higher - 15-20%. [NEWLINE] [NEWLINE] So you tell me: which one makes more sense? [NEWLINE] [NEWLINE] (FWIW I don't agree that English should be made the official language, nor that various government entities should stop offering translators, etc. But as for pushing for everyone in the US to learn a second language - really doesn't seem to be worth it.)</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I simply am not going to be able to take a lecture on molecular biology seriously<mask> it is being given by someone who sounds like Jesse Pinkman. [ENDQ] [NEWLINE] Serious question:<mask> not? [NEWLINE] [NEWLINE] Prestige uses circular reasoning. Whatever the elite does is prestige, and it's prestige<mask> the elite does it. This is true for many societal aspects, not just language. [NEWLINE] [NEWLINE] Textbooks are written in a standard dialect<mask> they are expected to be used over a large geographical area (perhaps even the entire world), and the authors are not going to use a local dialect for a global audience. [NEWLINE] [NEWLINE] Just<mask> an anecdote, I had a maths teacher in university who had a very thick Berlin German dialect.<mask> his classes were packed,<mask> he knew his maths.</s><pad>
Label encoding: <s> [STARTQ] I simply am not going to be able to take a lecture on molecular biology seriously if it is being given by someone who sounds like Jesse Pinkman. [ENDQ] [NEWLINE] Serious question: Why not? [NEWLINE] [NEWLINE] Prestige uses circular reasoning. Whatever the elite does is prestige, and it's prestige because the elite does it. This is true for many societal aspects, not just language. [NEWLINE] [NEWLINE] Textbooks are written in a standard dialect because they are expected to be used over a large geographical area (perhaps even the entire world), and the authors are not going to use a local dialect for a global audience. [NEWLINE] [NEWLINE] Just as an anecdote, I had a maths teacher in university who had a very thick Berlin German dialect. But his classes were packed, because he knew his maths.</s><pad>
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have two things to say to that. [NEWLINE] <mask>,<mask> would you have them prove that? By definition, the supernatural is not definable by natural methods. There is no way to prove<mask> happens after you die, except in a literal, physical sense. A lot of proof is very much personal and non-provable. Someone can have a personal, supernatural experience that they will never be able to share with anyone else,<mask> you would discount it<mask> proof<mask> you couldn't see it. [NEWLINE] <mask>,<mask> should one person be able to dictate<mask> happens with someone else's body anyway?<mask> exactly should they gain the right to decide<mask> to do with the body, and who should gain said right?<mask> should someone not have control of<mask> happens with their own body?</s>
Label encoding: <s>I have two things to say to that. [NEWLINE] Firstly, how would you have them prove that? By definition, the supernatural is not definable by natural methods. There is no way to prove what happens after you die, except in a literal, physical sense. A lot of proof is very much personal and non-provable. Someone can have a personal, supernatural experience that they will never be able to share with anyone else, but you would discount it as proof because you couldn't see it. [NEWLINE] Secondly, why should one person be able to dictate what happens with someone else's body anyway? How exactly should they gain the right to decide what to do with the body, and who should gain said right? Why should someone not have control of what happens with their own body?</s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you think feminism can't reduce rape statistics, consider this. [NEWLINE] [NEWLINE] Countries<mask> women are treated<mask> second-class citizens have higher rates of rape. Makes sense, really.<mask> women are denied liberties and are put on the back burner, the men will over time come to see them<mask> "less than." It's not that the men are bad people. It's just that<mask> they are born and raised in a society<mask> they are told that they are better and more important than women and are accustomed to seeing women be mistreated, they are more likely to think rape isn't a big deal. [NEWLINE] [NEWLINE] Countries<mask> women are more or less on par with men, you don't see the same attitude. Feminism can reduce rape. It's all about education.</s><pad>
Label encoding: <s>If you think feminism can't reduce rape statistics, consider this. [NEWLINE] [NEWLINE] Countries where women are treated as second-class citizens have higher rates of rape. Makes sense, really. When women are denied liberties and are put on the back burner, the men will over time come to see them as "less than." It's not that the men are bad people. It's just that if they are born and raised in a society where they are told that they are better and more important than women and are accustomed to seeing women be mistreated, they are more likely to think rape isn't a big deal. [NEWLINE] [NEWLINE] Countries where women are more or less on par with men, you don't see the same attitude. Feminism can reduce rape. It's all about education.</s><pad>
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What alternatives could there be to loans? Right now,<mask> you want to go to college or buy a home or start your own business, you need thousands, sometimes even tends of thousands or hundreds of thousands, of dollars. The only way to get that is through a bank. We all know the pitfalls there. [NEWLINE] [NEWLINE] The other thing I'm questioning now is,<mask> do you give workers more options? I can pull up a hundred different job listings for my area,<mask> that doesn't mean I'm qualified for any of them. And I can't be qualified<mask> I don't go to school for the skills,<mask> I'm still stuck with having to pay absurd amounts of money<mask> that I MIGHT be able to get a slightly better job five years down the road.</s>
Label encoding: <s>What alternatives could there be to loans? Right now, if you want to go to college or buy a home or start your own business, you need thousands, sometimes even tends of thousands or hundreds of thousands, of dollars. The only way to get that is through a bank. We all know the pitfalls there. [NEWLINE] [NEWLINE] The other thing I'm questioning now is, how do you give workers more options? I can pull up a hundred different job listings for my area, but that doesn't mean I'm qualified for any of them. And I can't be qualified if I don't go to school for the skills, so I'm still stuck with having to pay absurd amounts of money so that I MIGHT be able to get a slightly better job five years down the road.</s>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This could mean that he could be hiding the multiplujillions in overseas accounts, safe from the IRS [ENDQ] [NEWLINE] I doubt this was the case.<mask> we take the Barks/Don Rosa timeline that firmly places stories in 1950s and Scrooge **was** friends with Teddy Rooseveldt, then it's fairly unlikely that he would be allowed to move to a tax haven. In most interpretations being patriotic (<mask> Scottish immigrant) would fit Scrooge's character well. [NEWLINE] [NEWLINE] I have seen comic book stories<mask> the mayor of Duckburg coerces Scrooge into doing something with promise of tax cuts, and some stories<mask> he actually serves<mask> city's main financial officer (not a bad choice<mask> we assume that he's not going after Mayan temples).</s>
Label encoding: <s> [STARTQ] This could mean that he could be hiding the multiplujillions in overseas accounts, safe from the IRS [ENDQ] [NEWLINE] I doubt this was the case. If we take the Barks/Don Rosa timeline that firmly places stories in 1950s and Scrooge **was** friends with Teddy Rooseveldt, then it's fairly unlikely that he would be allowed to move to a tax haven. In most interpretations being patriotic ( as Scottish immigrant) would fit Scrooge's character well. [NEWLINE] [NEWLINE] I have seen comic book stories where the mayor of Duckburg coerces Scrooge into doing something with promise of tax cuts, and some stories where he actually serves as city's main financial officer (not a bad choice if we assume that he's not going after Mayan temples).</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the word'sexual dimorphism' might sort of be<mask> you're looking for,<mask> I don't think it really applies to very much, considering the amount of variation of<mask>'s considered 'women's work' around the world. Unless you think one culture is more 'natural' than the others, you'd have to blend together all of humanity's gendered labor division and you'd come up with a bit of a mess. [NEWLINE] [NEWLINE] There's nothing 'female' about being a nurse or a teacher or a call-centre employee, for example. Nothing in biology prepares women for those roles. In some places female teachers are extremely rare. In some places farming is traditionally considered women's work and weaving is men's work. And<mask> on and<mask> forth</s>
Label encoding: <s>I think the word'sexual dimorphism' might sort of be what you're looking for, but I don't think it really applies to very much, considering the amount of variation of what's considered 'women's work' around the world. Unless you think one culture is more 'natural' than the others, you'd have to blend together all of humanity's gendered labor division and you'd come up with a bit of a mess. [NEWLINE] [NEWLINE] There's nothing 'female' about being a nurse or a teacher or a call-centre employee, for example. Nothing in biology prepares women for those roles. In some places female teachers are extremely rare. In some places farming is traditionally considered women's work and weaving is men's work. And so on and so forth</s>
Loss: tensor(0.1398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Unfortunately, terminating a pregnancy requires invasive medical procedures that it would be barbaric to give fathers the right to do unilaterally. That's a fact of life--it's non-ideal and makes the rights pitted against each other the way they are. [NEWLINE] [NEWLINE] I really hate to go all "personal responsibility" on this thread,<mask> everyone involved does have the option to stop this all before it begins. The fact that women have more rights than men later is an accident of biology, not a policy designed against men. Honestly, it only seems like men have "no rights"<mask> women and children have more pressing needs that override certain financial rights late in the game. Men only have "no rights" once they're in a situation they either willingly entered or didn't properly avoid.</s>
Label encoding: <s>Unfortunately, terminating a pregnancy requires invasive medical procedures that it would be barbaric to give fathers the right to do unilaterally. That's a fact of life--it's non-ideal and makes the rights pitted against each other the way they are. [NEWLINE] [NEWLINE] I really hate to go all "personal responsibility" on this thread, but everyone involved does have the option to stop this all before it begins. The fact that women have more rights than men later is an accident of biology, not a policy designed against men. Honestly, it only seems like men have "no rights" because women and children have more pressing needs that override certain financial rights late in the game. Men only have "no rights" once they're in a situation they either willingly entered or didn't properly avoid.</s>
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Or more accurately<mask> I told him I was wearing my leather cowboy boots to work<mask> they're marginally more safe than my tennis shoes he'd probably wonder<mask> I don't just wear my steel toe work boots, and then point out the only reason I want to wear them is<mask> they look cooler. [ENDQ] [NEWLINE] Sure;<mask> would he make you take your sweet boots off and work barefoot<mask> you weren't doing the safest thing you possibly could be? [NEWLINE] [NEWLINE] No,<mask> there is still *some* safety value in them. [NEWLINE] [NEWLINE] That's all that's going on here.  Are loud pipes *primarily* for style (or whatever other word you want to use here)?  Absolutely.  Do they *<mask> * contribute to safety?  Yes.</s>
Label encoding: <s> [STARTQ] Or more accurately if I told him I was wearing my leather cowboy boots to work because they're marginally more safe than my tennis shoes he'd probably wonder why I don't just wear my steel toe work boots, and then point out the only reason I want to wear them is because they look cooler. [ENDQ] [NEWLINE] Sure; but would he make you take your sweet boots off and work barefoot because you weren't doing the safest thing you possibly could be? [NEWLINE] [NEWLINE] No, because there is still *some* safety value in them. [NEWLINE] [NEWLINE] That's all that's going on here.  Are loud pipes *primarily* for style (or whatever other word you want to use here)?  Absolutely.  Do they * also * contribute to safety?  Yes.</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> that's life, right?  Extraterrestrial - i.e. alien - life.  Ergo, aliens. [ENDQ] [NEWLINE] I meant I'm not convinced of intelligent life -- aliens that have brains, who are able to communicate, live in their own society, etc. [NEWLINE] [NEWLINE] I'm not convinced nor unconvinced that unintelligent life exists, just that I believe it might be possible. [NEWLINE] [NEWLINE] [STARTQ] Yeah.  Maybe maybe not.  I certainly can't prove it<mask><mask> you already acknowledge the strong possibility of *life* outside of our solar system, then it's not hard to believe that somewhere some of it has evolved to meet some criteria of intelligence. [ENDQ] [NEWLINE] I acknowledge that there's a *slight* possibility. </s>
Label encoding: <s> [STARTQ] But that's life, right?  Extraterrestrial - i.e. alien - life.  Ergo, aliens. [ENDQ] [NEWLINE] I meant I'm not convinced of intelligent life -- aliens that have brains, who are able to communicate, live in their own society, etc. [NEWLINE] [NEWLINE] I'm not convinced nor unconvinced that unintelligent life exists, just that I believe it might be possible. [NEWLINE] [NEWLINE] [STARTQ] Yeah.  Maybe maybe not.  I certainly can't prove it but if you already acknowledge the strong possibility of *life* outside of our solar system, then it's not hard to believe that somewhere some of it has evolved to meet some criteria of intelligence. [ENDQ] [NEWLINE] I acknowledge that there's a *slight* possibility. </s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry, I don't quite follow<mask> you're trying to say with the last sentence? [NEWLINE] [NEWLINE] [STARTQ] <mask> you get creeped out by them and don't want to invite them to a ceremony just<mask> of who they are,<mask> you think they are just<mask> valid<mask> heterosexual relationships, except<mask> it comes to having a personal ceremony to commemorate their love for each other. They don't have a right to that. [ENDQ] [NEWLINE] <mask> I've said several times, its just their actions that they may do at my wedding that may creep me out and I'm not sure that I want that.<mask> a lot of people here may be convincing me that its time just to expose myself more to situations that scare me. Then maybe I'll reach a point of acceptance.</s>
Label encoding: <s>I'm sorry, I don't quite follow what you're trying to say with the last sentence? [NEWLINE] [NEWLINE] [STARTQ] So you get creeped out by them and don't want to invite them to a ceremony just because of who they are, but you think they are just as valid as heterosexual relationships, except when it comes to having a personal ceremony to commemorate their love for each other. They don't have a right to that. [ENDQ] [NEWLINE] AS I've said several times, its just their actions that they may do at my wedding that may creep me out and I'm not sure that I want that. But a lot of people here may be convincing me that its time just to expose myself more to situations that scare me. Then maybe I'll reach a point of acceptance.</s>
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You know, I never once thought about<mask> breaking bad taught me. Doing something illegal is wrong. Doing something illegal to support your family, still wrong. Being immoral at all is wrong even<mask> it helps people. [NEWLINE] [NEWLINE] SPOILER [<mask> Walt's hit guys kill Hank, it makes an unforgiving twist. His family, the agency, his friend and everyone else (even his former friends who founded Grey Matter) see him<mask> an awful person. He may think he's helping his family,<mask><mask> his morals mortality has punished him more than the law.](/spoiler) [NEWLINE] [NEWLINE] Sorry for poor English and grammar,<mask> you have ever tried to use the full site on a cell phone, you'd know the troubles. I tried making this quick.</s>
Label encoding: <s>You know, I never once thought about what breaking bad taught me. Doing something illegal is wrong. Doing something illegal to support your family, still wrong. Being immoral at all is wrong even when it helps people. [NEWLINE] [NEWLINE] SPOILER [ When Walt's hit guys kill Hank, it makes an unforgiving twist. His family, the agency, his friend and everyone else (even his former friends who founded Grey Matter) see him as an awful person. He may think he's helping his family, but but his morals mortality has punished him more than the law.](/spoiler) [NEWLINE] [NEWLINE] Sorry for poor English and grammar, if you have ever tried to use the full site on a cell phone, you'd know the troubles. I tried making this quick.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I wrote about<mask> this lack of competition is due to the government [ENDQ] [NEWLINE] It is<mask> definitely due in part to economic factors. Someone of my net worth would have a *hell* of a time accumulating the sort of investment it takes to run an ISP that could compete with comcast. [NEWLINE] [NEWLINE] Most arguments against the "harmlessness" of markets focus on the implausibility of expecting those that don't "win" to *actually* have an equal opportunity to success. In the mean time, they are stuck paying rent (same<mask> taxes- you can move to a new apartment, you can move to a new country), and working for an employer (who have unity control over them during their days<mask> they would otherwise be going to Harvard or whatever.)</s><pad><pad><pad>
Label encoding: <s> [STARTQ] I wrote about how this lack of competition is due to the government [ENDQ] [NEWLINE] It is also definitely due in part to economic factors. Someone of my net worth would have a *hell* of a time accumulating the sort of investment it takes to run an ISP that could compete with comcast. [NEWLINE] [NEWLINE] Most arguments against the "harmlessness" of markets focus on the implausibility of expecting those that don't "win" to *actually* have an equal opportunity to success. In the mean time, they are stuck paying rent (same as taxes- you can move to a new apartment, you can move to a new country), and working for an employer (who have unity control over them during their days when they would otherwise be going to Harvard or whatever.)</s><pad><pad><pad>
Loss: tensor(0.1913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't think gun ownership by itself is evil. [ENDQ] [NEWLINE] Didnt think you did, just curious<mask> to the difference in reactions in this thread<mask> comparing different countries. It is the exact opposite from most gun control discussions. [NEWLINE] [NEWLINE] The gun control people are pushing comparisons between individual states, and disregarding ones between other countries. Normally its the exact opposite with comparisons between states like Vermont and California being tossed out in favour of comparing the entire USA to australia. [NEWLINE] [NEWLINE] Anyways, just found that interesting. [NEWLINE] [NEWLINE] Had this comment saved from /u/totallynotatf he made in a different thread on suicides, regarding canadian statistics. He made another post in this thread to a similar effect,<mask> here you go [NEWLINE] [URL] [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] I don't think gun ownership by itself is evil. [ENDQ] [NEWLINE] Didnt think you did, just curious as to the difference in reactions in this thread when comparing different countries. It is the exact opposite from most gun control discussions. [NEWLINE] [NEWLINE] The gun control people are pushing comparisons between individual states, and disregarding ones between other countries. Normally its the exact opposite with comparisons between states like Vermont and California being tossed out in favour of comparing the entire USA to australia. [NEWLINE] [NEWLINE] Anyways, just found that interesting. [NEWLINE] [NEWLINE] Had this comment saved from /u/totallynotatf he made in a different thread on suicides, regarding canadian statistics. He made another post in this thread to a similar effect, but here you go [NEWLINE] [URL] [NEWLINE] </s><pad>
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're absolutely correct, and I apologize for expressing myself poorly. I wrote that I don't think the government should regulate marriage, and that's actually an overstatement of my feelings.<mask> I should have written is that ***I don't believe the government should regulate marriage much beyond the extent that they regulate and enforce other contracts***.  I absolutely believe that some regulation is needed. [NEWLINE] [NEWLINE] I wouldn't say that I feel the existing system simply needs to be re-calibrated<mask>. Stronger change is needed.<mask> it stands, the government is able to determine who is and is not able to be married, and both historically and presently they do this using somewhat arbitrary (and unfair) standards. I see no advantage to them having that power </s>
Label encoding: <s>You're absolutely correct, and I apologize for expressing myself poorly. I wrote that I don't think the government should regulate marriage, and that's actually an overstatement of my feelings. What I should have written is that ***I don't believe the government should regulate marriage much beyond the extent that they regulate and enforce other contracts***.  I absolutely believe that some regulation is needed. [NEWLINE] [NEWLINE] I wouldn't say that I feel the existing system simply needs to be re-calibrated though. Stronger change is needed. As it stands, the government is able to determine who is and is not able to be married, and both historically and presently they do this using somewhat arbitrary (and unfair) standards. I see no advantage to them having that power </s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Since abortion is directly and intentionally killing a pre born child I believe that it would only be morally permissible<mask> there truly was no other alternative to saving the life of the mother. There are other medical procedures that can save the life of the mother.<mask><mask><mask>, the only morally permissible thing to do is to try to save all human lives involved. For example,<mask> a fetus dies or is injured<mask> a mother needs to undergo some form of cancer therapy or a fetus dies<mask> a mother needs to remove her fallopian tube<mask><mask><mask> of an ectopic pregnancy (both attempts at saving her life) that is not the same<mask> a fetus dying from an abortion which is a direct and intentional killing. Only the former is morally permissible. The latter is murder.</s>
Label encoding: <s>Since abortion is directly and intentionally killing a pre born child I believe that it would only be morally permissible if there truly was no other alternative to saving the life of the mother. There are other medical procedures that can save the life of the mother. In my opinion, the only morally permissible thing to do is to try to save all human lives involved. For example, if a fetus dies or is injured because a mother needs to undergo some form of cancer therapy or a fetus dies because a mother needs to remove her fallopian tube as a result of an ectopic pregnancy (both attempts at saving her life) that is not the same as a fetus dying from an abortion which is a direct and intentional killing. Only the former is morally permissible. The latter is murder.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1808, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for posting to /r/changemyview! Unfortunately, your post has been removed from this subreddit. [NEWLINE] [NEWLINE] Your comment violated **Comment Rule 1: "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments."** [See the wiki page for more information.]( [URL] #wiki_rule_1) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] <mask> you would like to appeal this decision, please [message the moderators]( [URL] )! [NEWLINE] [NEWLINE] *Regards, IAmAN00bie and the mods at /r/changemyview.*</s><pad>
Label encoding: <s>Thank you for posting to /r/changemyview! Unfortunately, your post has been removed from this subreddit. [NEWLINE] [NEWLINE] Your comment violated **Comment Rule 1: "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments."** [See the wiki page for more information.]( [URL] #wiki_rule_1) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] If you would like to appeal this decision, please [message the moderators]( [URL] )! [NEWLINE] [NEWLINE] *Regards, IAmAN00bie and the mods at /r/changemyview.*</s><pad>
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Religions are made entirely of organizations. "Christianity" is just the brand name. The priest has been a social and economic class<mask> the beginning of the agricultural revolution. [NEWLINE] [NEWLINE] I used the catholics<mask> they're the best example.<mask> go find some evangelical southern Baptist mega church that holds it's Sunday meetings in sports stadiums. Christianity is the biggest industry in the US. [NEWLINE] [NEWLINE] There are over 40,000 recognized denominations of Christianity worldwide. All with their own ministers and priests and pastors. All of whom live off someone else's dime. It's a business, and an extremely effective business at that. Shit you have millionaire ministers who still manage to say with a straight face that their churches are non-profits and<mask> tax exempt. </s>
Label encoding: <s>Religions are made entirely of organizations. "Christianity" is just the brand name. The priest has been a social and economic class since the beginning of the agricultural revolution. [NEWLINE] [NEWLINE] I used the catholics because they're the best example. But go find some evangelical southern Baptist mega church that holds it's Sunday meetings in sports stadiums. Christianity is the biggest industry in the US. [NEWLINE] [NEWLINE] There are over 40,000 recognized denominations of Christianity worldwide. All with their own ministers and priests and pastors. All of whom live off someone else's dime. It's a business, and an extremely effective business at that. Shit you have millionaire ministers who still manage to say with a straight face that their churches are non-profits and thus tax exempt. </s>
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Everyone (or even most) who has an abortion is some kind of idiot? or lowlife? or drunk/drug user? o_0? [ENDQ] [NEWLINE] This is a confusing statement, I did say that most abortions are carried out<mask> people were careless and I did call those people stupid. <mask> you should better understand my view with this, I consider them careless and stupid<mask> they took a risk and<mask> it didnt work out, they decided to abort the child. [NEWLINE] [NEWLINE] That is stupid, that is careless<mask><mask><mask>. [NEWLINE] [NEWLINE] ------ [NEWLINE] [NEWLINE] For the rest of your post I was deeply confused,<mask> I realize the sky wont fall<mask> abortions even become a culture norm all over the world and everyone is having fetus snowball fights.</s>
Label encoding: <s> [STARTQ] Everyone (or even most) who has an abortion is some kind of idiot? or lowlife? or drunk/drug user? o_0? [ENDQ] [NEWLINE] This is a confusing statement, I did say that most abortions are carried out because people were careless and I did call those people stupid.  However you should better understand my view with this, I consider them careless and stupid because they took a risk and when it didnt work out, they decided to abort the child. [NEWLINE] [NEWLINE] That is stupid, that is careless in my opinion. [NEWLINE] [NEWLINE] ------ [NEWLINE] [NEWLINE] For the rest of your post I was deeply confused, as I realize the sky wont fall if abortions even become a culture norm all over the world and everyone is having fetus snowball fights.</s>
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Aye, doomsday predictions have been overflowing<mask> the dawn of time. [NEWLINE] [NEWLINE] [STARTQ] a nano-techological robot specifically designed to take over our bodies, create billions of copies of itself, and release them into the air, killing us in a way that none of our guns and bombs can stop. [ENDQ] [NEWLINE] That's obviously very disturbing<mask> from the point of view of a computer scientist, don't you think that is is merely<mask> humans do with AI that will make it<mask> dangerous? [NEWLINE] [NEWLINE] Isn't it that, at least from my point of view, which will trigger the beginning of AI becoming uncontrollable or<mask> complicated that not even the creator will be able to stop it? (through physical barriers such<mask> death, for example). </s>
Label encoding: <s>Aye, doomsday predictions have been overflowing since the dawn of time. [NEWLINE] [NEWLINE] [STARTQ] a nano-techological robot specifically designed to take over our bodies, create billions of copies of itself, and release them into the air, killing us in a way that none of our guns and bombs can stop. [ENDQ] [NEWLINE] That's obviously very disturbing but from the point of view of a computer scientist, don't you think that is is merely what humans do with AI that will make it so dangerous? [NEWLINE] [NEWLINE] Isn't it that, at least from my point of view, which will trigger the beginning of AI becoming uncontrollable or so complicated that not even the creator will be able to stop it? (through physical barriers such as death, for example). </s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My line of thinking is<mask> follows: [NEWLINE] [NEWLINE] <mask> I'm in a relationship,<mask> do I want most (other than my own well-being)? Easy, I want my partner to be happy. Now,<mask> the fact that I'm totally awesome at sex and romance (jk), I'm totally aware that sleeping with other guys will make my partner happier. [NEWLINE] [NEWLINE] <mask><mask> shouldn't I be ok with an open relationship? You suggest that one can't form a truly intimate connection in an open relationship,<mask> I don't see<mask> prohibiting your partner from doing something they enjoy will form one.<mask> anything, my partner sleeping with different people<mask> always coming back to me makes me more confident in our relationship, and makes us closer.</s>
Label encoding: <s>My line of thinking is as follows: [NEWLINE] [NEWLINE] When I'm in a relationship, what do I want most (other than my own well-being)? Easy, I want my partner to be happy. Now, despite the fact that I'm totally awesome at sex and romance (jk), I'm totally aware that sleeping with other guys will make my partner happier. [NEWLINE] [NEWLINE] So why shouldn't I be ok with an open relationship? You suggest that one can't form a truly intimate connection in an open relationship, but I don't see how prohibiting your partner from doing something they enjoy will form one. If anything, my partner sleeping with different people but always coming back to me makes me more confident in our relationship, and makes us closer.</s>
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's a pretty big generalization. There are plenty of people who don't try and convert people, just<mask> a few are annoying about it doesn't mean we should generalize all of them. There are plenty of religious individuals who feel the need to convert others, just<mask> they are annoying does not mean that they all are. [NEWLINE] [NEWLINE] There was a period of time<mask> I was avoiding grains; whenever people asked me I would tell them<mask> I never tried to force my lifestyle on them. For that period of time it made me feel good and every once in a<mask> I do it, and<mask> I am trying to say, bashing all individuals<mask> of a select few is just<mask> ignorant<mask> bashing all religious individuals<mask> of a few.</s>
Label encoding: <s>That's a pretty big generalization. There are plenty of people who don't try and convert people, just because a few are annoying about it doesn't mean we should generalize all of them. There are plenty of religious individuals who feel the need to convert others, just because they are annoying does not mean that they all are. [NEWLINE] [NEWLINE] There was a period of time where I was avoiding grains; whenever people asked me I would tell them but I never tried to force my lifestyle on them. For that period of time it made me feel good and every once in a while I do it, and as I am trying to say, bashing all individuals because of a select few is just as ignorant as bashing all religious individuals because of a few.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't honestly think I can change your view. I am catholic and I do believe in all this. I<mask> believe it isn't my place to change your view,<mask> it's<mask> you believe.<mask> I will say my interpretation of it.<mask> one good to heaven, they can experience<mask> the splendor they ever dreamed.<mask> being in heaven everyone is given a sense of joy in this, it's made to be eternal reward for a good life. Your time on heaven will be sent being given an eternal sense of joy and reward,<mask> resulting in you not feeling in the way you put. The ideal of heaven is very vague and very open to interpretation.<mask> I love the way you worded the entire thing.</s>
Label encoding: <s>I don't honestly think I can change your view. I am catholic and I do believe in all this. I also believe it isn't my place to change your view, as it's what you believe. But I will say my interpretation of it. When one good to heaven, they can experience so the splendor they ever dreamed. Also being in heaven everyone is given a sense of joy in this, it's made to be eternal reward for a good life. Your time on heaven will be sent being given an eternal sense of joy and reward, therefore resulting in you not feeling in the way you put. The ideal of heaven is very vague and very open to interpretation. Also I love the way you worded the entire thing.</s>
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Smoke inhalation is always bad for your respiratory system. Is dope<mask> bad<mask> smoking cigarettes? Probably not.<mask>,<mask> you are smoking a joint then it is probably cut with tobacco to help it burn. [NEWLINE] [NEWLINE] Brain development doesn't magically stop<mask> you reach voting age. It's more like 24yrs old. OP is 20 and gives no information about prior use. [NEWLINE] [NEWLINE] Anything you don't need costs you money.<mask> you want it, it's always your money to spend. [NEWLINE] [NEWLINE] Marijuana has a stupefying effect that many with mental illness appreciate.<mask> it exactly interacts with psychiatric medication or psychiatric illness is an unknown quantity (<mask> research on such a subject is unethical). [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [URL] </s>
Label encoding: <s>Smoke inhalation is always bad for your respiratory system. Is dope as bad as smoking cigarettes? Probably not. However, if you are smoking a joint then it is probably cut with tobacco to help it burn. [NEWLINE] [NEWLINE] Brain development doesn't magically stop when you reach voting age. It's more like 24yrs old. OP is 20 and gives no information about prior use. [NEWLINE] [NEWLINE] Anything you don't need costs you money. If you want it, it's always your money to spend. [NEWLINE] [NEWLINE] Marijuana has a stupefying effect that many with mental illness appreciate. How it exactly interacts with psychiatric medication or psychiatric illness is an unknown quantity ( as research on such a subject is unethical). [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [URL] </s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Personally,<mask><mask> a mature responsible adult should be allowed to do whatever he wants. [ENDQ] [NEWLINE] <mask> he is allowed to by your rules. [NEWLINE] [NEWLINE] [STARTQ] <mask> someone's actions truly does lower the value of the neighborhood, then I consider that to be harmful. [ENDQ] [NEWLINE] Its not<mask> it does or not,<mask> it a person believes that it does then he believes he is right to interfere. [NEWLINE] [NEWLINE] [STARTQ] The problem here lies in society judging his behavior. [ENDQ] [NEWLINE] <mask> we are not suppose to have an opinion on each other?"Your home needs a paint job" isn't allowed?  "You daughter is smart" isn't allowed?  Ignore everything about each other?  That is not realistic nor a society. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Personally, I think a mature responsible adult should be allowed to do whatever he wants. [ENDQ] [NEWLINE] But he is allowed to by your rules. [NEWLINE] [NEWLINE] [STARTQ] If someone's actions truly does lower the value of the neighborhood, then I consider that to be harmful. [ENDQ] [NEWLINE] Its not if it does or not, if it a person believes that it does then he believes he is right to interfere. [NEWLINE] [NEWLINE] [STARTQ] The problem here lies in society judging his behavior. [ENDQ] [NEWLINE] So we are not suppose to have an opinion on each other?"Your home needs a paint job" isn't allowed?  "You daughter is smart" isn't allowed?  Ignore everything about each other?  That is not realistic nor a society. [NEWLINE] </s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you register and choose to share your information with the analysts and data gatherers, you'll be counted. [NEWLINE] [NEWLINE] Then<mask> the next election comes round there will be a proportion more policy and promises directed at you.<mask> I will admit the amount you individually change the situation is minuscule,<mask> I know<mask> it comes to the next general election I want to be in the demographic they are doing there best to appeal to. [NEWLINE] [NEWLINE] That's my voting tactic, I'm a student and I can vote in either of 2 constituencies, and both seats will remain labour seats,<mask> I'd still rather be on the list of 20-somethings who voted, than the list of those who didn't.</s>
Label encoding: <s>If you register and choose to share your information with the analysts and data gatherers, you'll be counted. [NEWLINE] [NEWLINE] Then when the next election comes round there will be a proportion more policy and promises directed at you. Although I will admit the amount you individually change the situation is minuscule, however I know when it comes to the next general election I want to be in the demographic they are doing there best to appeal to. [NEWLINE] [NEWLINE] That's my voting tactic, I'm a student and I can vote in either of 2 constituencies, and both seats will remain labour seats, but I'd still rather be on the list of 20-somethings who voted, than the list of those who didn't.</s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a question: which Beatles songs have you heard, exactly? Which albums have you listened to? [NEWLINE] [NEWLINE] I'm a huge Beatles fan, and I actually don't love many of their songs that are considered<mask> great by the masses. (E.g., Let It Be, Hey Jude, Revolution, etc.) [NEWLINE] [NEWLINE] Have you ever listened to [Being For the Benefit of Mr. Kite?]( [URL] ) Or [Norwegian Wood]( [URL] )?<mask> about the [Abbey Road Medley]( [URL] ), which,<mask><mask>, is their most spectacular achievement? [NEWLINE] [NEWLINE] Try giving all of those a listen –<mask><mask> they're a lot less "generic" than the Beatles songs you may have heard.</s>
Label encoding: <s>I have a question: which Beatles songs have you heard, exactly? Which albums have you listened to? [NEWLINE] [NEWLINE] I'm a huge Beatles fan, and I actually don't love many of their songs that are considered so great by the masses. (E.g., Let It Be, Hey Jude, Revolution, etc.) [NEWLINE] [NEWLINE] Have you ever listened to [Being For the Benefit of Mr. Kite?]( [URL] ) Or [Norwegian Wood]( [URL] )? How about the [Abbey Road Medley]( [URL] ), which, IMO, is their most spectacular achievement? [NEWLINE] [NEWLINE] Try giving all of those a listen – I think they're a lot less "generic" than the Beatles songs you may have heard.</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know you aren't advocating for it,<mask> I want to point out that this would only work once you already have a stable population of sterilized folks. <mask> do you propose knowing about every pregnancy/birth and<mask> have the ability to sterilize the baby upon birth? <mask><mask> there is the same logical issue that others have said.  There will always be "illegal immigrants, people without a stable residence, people who would purposely evade the monitoring", etc... who will prevent you from having your stable population of sterilized folks.  And I would assume the people who were not sterilized would become much more "in demand" on the black market<mask> people sell their "old style" baby making skills.</s>
Label encoding: <s>I know you aren't advocating for it, but I want to point out that this would only work once you already have a stable population of sterilized folks.  How do you propose knowing about every pregnancy/birth and therefore have the ability to sterilize the baby upon birth?  I think there is the same logical issue that others have said.  There will always be "illegal immigrants, people without a stable residence, people who would purposely evade the monitoring", etc... who will prevent you from having your stable population of sterilized folks.  And I would assume the people who were not sterilized would become much more "in demand" on the black market where people sell their "old style" baby making skills.</s>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the only reason abortion is a problem is<mask> many women believe it's somehow sacred to abort even a very young fetus. [ENDQ] [NEWLINE] This believe,<mask> the believe in a soul, is not one held solely by women. [NEWLINE] [NEWLINE] 'Emotional scaring' is not the only complication or side effect that can occur. Abortion comes with medical complications and risks<mask> well.<mask> 'emotional scaring' is not depended on unrationality. You can hold the view that it's just a clump of cells and still feel affected. Is not that simple. [NEWLINE] [NEWLINE] <mask>, I don't agree with your premise that both (biological?) parents are needed or 'optimal'. Neither do most studies. </s>
Label encoding: <s> [STARTQ] the only reason abortion is a problem is because many women believe it's somehow sacred to abort even a very young fetus. [ENDQ] [NEWLINE] This believe, as the believe in a soul, is not one held solely by women. [NEWLINE] [NEWLINE] 'Emotional scaring' is not the only complication or side effect that can occur. Abortion comes with medical complications and risks as well. Also 'emotional scaring' is not depended on unrationality. You can hold the view that it's just a clump of cells and still feel affected. Is not that simple. [NEWLINE] [NEWLINE] Besides, I don't agree with your premise that both (biological?) parents are needed or 'optimal'. Neither do most studies. </s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0811, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Another thing i just noticed. [NEWLINE] [NEWLINE] the OP specifically says "More good people are killed by guns than are saved by them."  Considering the sheer volume of murder victims who are<mask> criminals...  something like 70+%  "we're talking an incredibly small number of "good people" being victimized by gun violence. [NEWLINE] [NEWLINE] Suicides are not "victims" of gun violence<mask> they're choosing to die.... [NEWLINE] [NEWLINE] <mask> we're talking 70%+ of 11,000 ish people killed each year are not by definition "good people"  <mask> we're at a  8:0.3 Ratio of documented defensive gun uses to "good persons" being killed with a gun. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Another thing i just noticed. [NEWLINE] [NEWLINE] the OP specifically says "More good people are killed by guns than are saved by them."  Considering the sheer volume of murder victims who are also criminals...  something like 70+%  "we're talking an incredibly small number of "good people" being victimized by gun violence. [NEWLINE] [NEWLINE] Suicides are not "victims" of gun violence because they're choosing to die.... [NEWLINE] [NEWLINE] so we're talking 70%+ of 11,000 ish people killed each year are not by definition "good people"   so we're at a  8:0.3 Ratio of documented defensive gun uses to "good persons" being killed with a gun. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why should need to be incentivized to work harder or faster? <mask><mask> I go home at 4pm and spend some leisure time instead of working that last hour every day. <mask> burning needs are there that<mask> I'm not working my absolute ass off, that humanity is suffering<mask><mask><mask>? [NEWLINE] [NEWLINE] <mask> yes, making more money might incentivize me to work more,<mask> my question is,<mask> do we think that this system of incentives is positive? [NEWLINE] [NEWLINE] <mask> i'm already living comfortably,<mask> should I take a salary increase<mask> my day to day responsibilities are manageable.  and<mask> they aren't manageable (people working 50-60hr weeks) then<mask> not hire a second person?</s>
Label encoding: <s>Why should need to be incentivized to work harder or faster?  What if I go home at 4pm and spend some leisure time instead of working that last hour every day.  What burning needs are there that if I'm not working my absolute ass off, that humanity is suffering as a result? [NEWLINE] [NEWLINE] So yes, making more money might incentivize me to work more, but my question is, why do we think that this system of incentives is positive? [NEWLINE] [NEWLINE] If i'm already living comfortably, why should I take a salary increase if my day to day responsibilities are manageable.  and if they aren't manageable (people working 50-60hr weeks) then why not hire a second person?</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ya it totally could. He could just<mask> easily be one of those three bullies.<mask> my second point is based definitely more on literature and personal philosophies. There's a great quote from John Green in *The Fault in Our Stars* - "We're<mask> likely to hurt the universe than we are to help it, and we're not likely to do either."<mask> think of a time someone said something to you and it stuck - even<mask> it's totally insignificant, it pervades your thought and might even change your perspective on something. It would be naïve to think you don't have a similar effect on others.<mask><mask> it's a view realistic and optimistic enough. It sure keeps me sane. </s>
Label encoding: <s>Ya it totally could. He could just as easily be one of those three bullies. Though my second point is based definitely more on literature and personal philosophies. There's a great quote from John Green in *The Fault in Our Stars* - "We're as likely to hurt the universe than we are to help it, and we're not likely to do either." But think of a time someone said something to you and it stuck - even if it's totally insignificant, it pervades your thought and might even change your perspective on something. It would be naïve to think you don't have a similar effect on others. I think it's a view realistic and optimistic enough. It sure keeps me sane. </s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm with Unhappytrombone,<mask> do you suddenly change from this [NEWLINE] [NEWLINE] [STARTQ] I thought I could never truly 100% respect him for having a belief I thought was completely wrong. [ENDQ] [NEWLINE] just<mask> his "completely wrong" belief happens to not condemn you to hell? The original objection didn't seem to be that she thought that he thought she was going to burn in hell, in the first place. [NEWLINE] [NEWLINE] I don't care<mask> happy-fuzzy a belief someone I'm with holds,<mask> I happen to think it's completely ridiculous, wrong, and unfounded, I'm going to have trouble respecting that person's belief<mask> they can't give me a good reason for it.</s>
Label encoding: <s>I'm with Unhappytrombone, how do you suddenly change from this [NEWLINE] [NEWLINE] [STARTQ] I thought I could never truly 100% respect him for having a belief I thought was completely wrong. [ENDQ] [NEWLINE] just because his "completely wrong" belief happens to not condemn you to hell? The original objection didn't seem to be that she thought that he thought she was going to burn in hell, in the first place. [NEWLINE] [NEWLINE] I don't care how happy-fuzzy a belief someone I'm with holds, if I happen to think it's completely ridiculous, wrong, and unfounded, I'm going to have trouble respecting that person's belief if they can't give me a good reason for it.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>She and I both know that I'm really working on this whole jealousy/paranoia problem.<mask> I've mentioned above, it isn't even her...it's me. She doesn't make these kinds of comments. I'm just always paranoid that she could think them. It's a pretty bad issue in my own mind. She thinks that all I need is time. [NEWLINE] [NEWLINE] It stems from a pretty bad relationship. I can elaborate<mask> necessary,<mask> let's just say that it left me in a state of extreme mistrust. [NEWLINE] [NEWLINE] I'm doing better,<mask> I still need to really get there. Thank you for the very kind words, sir. :) [NEWLINE] [NEWLINE] Oh, and: ∆</s>
Label encoding: <s>She and I both know that I'm really working on this whole jealousy/paranoia problem. As I've mentioned above, it isn't even her...it's me. She doesn't make these kinds of comments. I'm just always paranoid that she could think them. It's a pretty bad issue in my own mind. She thinks that all I need is time. [NEWLINE] [NEWLINE] It stems from a pretty bad relationship. I can elaborate if necessary, but let's just say that it left me in a state of extreme mistrust. [NEWLINE] [NEWLINE] I'm doing better, but I still need to really get there. Thank you for the very kind words, sir. :) [NEWLINE] [NEWLINE] Oh, and: ∆</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. All my friends who were mugged were mugged in the am in good neighborhoods. <mask> now,<mask> you go to work, you are a cause of your own mugging. [NEWLINE] [NEWLINE] 2. I live in a bad neighborhood<mask> it is<mask> I can afford.  I walk alone at night<mask> it is dark at 4 pm. <mask>, I have a life. <mask>, I need to be unemployed or move cities or I am part of the cause of my own mugging? [NEWLINE] [NEWLINE] 3. Being attacked by another person is nothing like gravity.  You are equating the danger of physics and engineering in skydiving with human choice such<mask> committing a crime.</s>
Label encoding: <s>1. All my friends who were mugged were mugged in the am in good neighborhoods.  So now, if you go to work, you are a cause of your own mugging. [NEWLINE] [NEWLINE] 2. I live in a bad neighborhood because it is what I can afford.  I walk alone at night because it is dark at 4 pm.  Also, I have a life.  So, I need to be unemployed or move cities or I am part of the cause of my own mugging? [NEWLINE] [NEWLINE] 3. Being attacked by another person is nothing like gravity.  You are equating the danger of physics and engineering in skydiving with human choice such as committing a crime.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why anyone would be "proud" of a government for doing<mask> it is designed and *obligated* to do is completely beyond me. Furthermore<mask> the government fails its people and the global community,<mask> you continue to have pride in said system only confuses me more. [NEWLINE] [NEWLINE] The list of governments it has undermined or removed, is terribly long(Bay of Pigs, Mogadishu). The people both abroad and locally it has used and abandoned(think about the veterans of the wars we fight, or the Al-Queda) is similarly long.<mask> anyone has pride in a government which treats people<mask> expendable is not only confusing to me, it's maddening.</s>
Label encoding: <s>Why anyone would be "proud" of a government for doing what it is designed and *obligated* to do is completely beyond me. Furthermore when the government fails its people and the global community, why you continue to have pride in said system only confuses me more. [NEWLINE] [NEWLINE] The list of governments it has undermined or removed, is terribly long(Bay of Pigs, Mogadishu). The people both abroad and locally it has used and abandoned(think about the veterans of the wars we fight, or the Al-Queda) is similarly long. Why anyone has pride in a government which treats people as expendable is not only confusing to me, it's maddening.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Forget the surgeries and physical changes. Transgender comes down to the mind: someone identifies<mask> the gender that is not that of their physical body. *That's* the definition of transgender. Some might opt for surgery or hormone therapy and<mask> on,<mask> some might not. It's<mask>'s going on in their mind that counts and the point is don't they deserve equal rights and respect? Who are you to tell them that their mind is behaving in an unnatural way? It's no different than a homophobe telling a gay man he shouldn't be sexually attracted to other men<mask> it's unnatural. It's not about the sex acts, it's about the desire within, isn't it?</s>
Label encoding: <s>Forget the surgeries and physical changes. Transgender comes down to the mind: someone identifies as the gender that is not that of their physical body. *That's* the definition of transgender. Some might opt for surgery or hormone therapy and so on, but some might not. It's what's going on in their mind that counts and the point is don't they deserve equal rights and respect? Who are you to tell them that their mind is behaving in an unnatural way? It's no different than a homophobe telling a gay man he shouldn't be sexually attracted to other men because it's unnatural. It's not about the sex acts, it's about the desire within, isn't it?</s>
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh i totally agree, it's a shame that it takes campaigns by groups such<mask> the one you are involved in to make the public aware of this. All of this controversy and bias would be eliminated in one fell swoop<mask> politicians had to declare on an official government register which is accessible by the public<mask> they get their funds from.<mask><mask><mask> i would implement legislation which mean that politicians could not vote or campaign on an issue in which they have a monetary or personal stake. An example of this being abused would be the nominee for Secretary of State Susan Rice who was promoting the Keystone XL pipeline<mask> she has stocks in TransCanada, the company applying for the permit. [NEWLINE] [NEWLINE] EDIT: Spelling</s>
Label encoding: <s>Oh i totally agree, it's a shame that it takes campaigns by groups such as the one you are involved in to make the public aware of this. All of this controversy and bias would be eliminated in one fell swoop if politicians had to declare on an official government register which is accessible by the public where they get their funds from. Further to this i would implement legislation which mean that politicians could not vote or campaign on an issue in which they have a monetary or personal stake. An example of this being abused would be the nominee for Secretary of State Susan Rice who was promoting the Keystone XL pipeline while she has stocks in TransCanada, the company applying for the permit. [NEWLINE] [NEWLINE] EDIT: Spelling</s>
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I made no claims outside of an appeal to caution. I'm glad you did some research on acid rain, bravo, good job. [NEWLINE] [NEWLINE] Your link to<mask> we should do to mitigate global warming is aimed at policy makers. Perhaps you are a policy maker,<mask> I am not. I don't see anything in your rebuttal that contradicts my initial conclusion that there is little that we can do outside of trying to do our best and not worry too much about the constant stream of fear, doom and gloom. [NEWLINE] [NEWLINE] <mask> would you suggest the average citizen do about global warming? I'm sure your rhetorical skills are sharp,<mask> I've not seen anything that refutes my main conclusion.</s><pad>
Label encoding: <s>I made no claims outside of an appeal to caution. I'm glad you did some research on acid rain, bravo, good job. [NEWLINE] [NEWLINE] Your link to what we should do to mitigate global warming is aimed at policy makers. Perhaps you are a policy maker, but I am not. I don't see anything in your rebuttal that contradicts my initial conclusion that there is little that we can do outside of trying to do our best and not worry too much about the constant stream of fear, doom and gloom. [NEWLINE] [NEWLINE] What would you suggest the average citizen do about global warming? I'm sure your rhetorical skills are sharp, but I've not seen anything that refutes my main conclusion.</s><pad>
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0725, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's completely possible a child born into poverty can climb the social ladder and it's possible for a child born in a rich family to take the social slide<mask> it's generally proven that families that come from poverty, produce poverty and families that come from wealthy/hard-working/etc will produce that.<mask> a child from a rich family ends up living of the government and not doing anything to generally be useful, then they lose their rights. Like I've said, poor people still have the ability to move up the social ladder or even stay poor and be able to reproduce. It's the ones that don't work and make no effort to work that<mask><mask> shouldn't have reproductive rights.</s>
Label encoding: <s>It's completely possible a child born into poverty can climb the social ladder and it's possible for a child born in a rich family to take the social slide but it's generally proven that families that come from poverty, produce poverty and families that come from wealthy/hard-working/etc will produce that. If a child from a rich family ends up living of the government and not doing anything to generally be useful, then they lose their rights. Like I've said, poor people still have the ability to move up the social ladder or even stay poor and be able to reproduce. It's the ones that don't work and make no effort to work that I think shouldn't have reproductive rights.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I'm trying to say, is that people should not be forced into believing and supporting gay marriage, gay/lesbian/transgender/pansexual and all that—they should be allowed to express their freedom of religion... [ENDQ] [NEWLINE] People are already free to practice their religion, they don't have to support LGBT people and they're not being forced to either.<mask> they can't do, is impose their religious beliefs on other people, they can't get their religious beliefs enshrined into law. Religious people are in no way being oppressed by two gay people getting married, they just aren't allowed to dictate the rules on who gets to get married based on their religion. [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] What I'm trying to say, is that people should not be forced into believing and supporting gay marriage, gay/lesbian/transgender/pansexual and all that—they should be allowed to express their freedom of religion... [ENDQ] [NEWLINE] People are already free to practice their religion, they don't have to support LGBT people and they're not being forced to either. What they can't do, is impose their religious beliefs on other people, they can't get their religious beliefs enshrined into law. Religious people are in no way being oppressed by two gay people getting married, they just aren't allowed to dictate the rules on who gets to get married based on their religion. [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reasons people cheat are varied,<mask> I don't understand<mask> you mean by "biological urge" [NEWLINE] [NEWLINE] Even<mask> it was some kind of urge,<mask><mask>? We are basically defined by rejecting harmful biological urges, like just raping the first female you see.<mask> is that any different? [NEWLINE] [NEWLINE] That to me just looks like the fallacy of appeal to nature. [NEWLINE] [NEWLINE] Actually, the most "biological" urge that I would suggest is to form communities. And for whatever reason, we have developed emotions that glue those communities together. And those emotions get hurt<mask> people cheat. [NEWLINE] [NEWLINE] <mask> cheating is basically destroying the emotions that hold us together<mask> communal and social animals.</s>
Label encoding: <s>The reasons people cheat are varied, but I don't understand what you mean by "biological urge" [NEWLINE] [NEWLINE] Even if it was some kind of urge, so what? We are basically defined by rejecting harmful biological urges, like just raping the first female you see. Why is that any different? [NEWLINE] [NEWLINE] That to me just looks like the fallacy of appeal to nature. [NEWLINE] [NEWLINE] Actually, the most "biological" urge that I would suggest is to form communities. And for whatever reason, we have developed emotions that glue those communities together. And those emotions get hurt when people cheat. [NEWLINE] [NEWLINE] So cheating is basically destroying the emotions that hold us together as communal and social animals.</s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Against the rules. <mask> I concur.  Pit bulls are trained to be viscous. They are not inherently aggressive, which is<mask> training them to be can really fuck them up. <mask> labs were the favorite breed to train for dog fighting, the same thing would be said about them. [NEWLINE] [NEWLINE] Having said that, Pit Bulls do have one of the strongest jaws. <mask><mask> do Rots and Dobermans<mask> we don't here people calling for banning those breeds.  I am suspicious that class comes into play here.  Pits are a poor, urban dog,<mask> the Rottweiler and Doberman and respected wealthy persons breed.  </s>
Label encoding: <s>Against the rules.  But I concur.  Pit bulls are trained to be viscous. They are not inherently aggressive, which is why training them to be can really fuck them up.  If labs were the favorite breed to train for dog fighting, the same thing would be said about them. [NEWLINE] [NEWLINE] Having said that, Pit Bulls do have one of the strongest jaws.  But so do Rots and Dobermans but we don't here people calling for banning those breeds.  I am suspicious that class comes into play here.  Pits are a poor, urban dog, but the Rottweiler and Doberman and respected wealthy persons breed.  </s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Godless-apostate, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Godless-apostate+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry Godless-apostate, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Godless-apostate+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Unconditional gets much of the time on reddit<mask> is extraordinarily poorly supported in advanced economies; its incredibly expensive (the distortionary effects from increasing taxation to pay for it would counteract its economic benefits many times over), would have a huge labor discouragement issue and would cause significant inflationary problems. Unconditional basic income in an advanced economy would eviscerate economic growth without correcting many of the problems those who support it claim. [ENDQ] [NEWLINE] I guess I've been victim to the misinformation. I always thought that an unconditional basic income, + clawback from unneeding citizens through income taxes, was the preferable system. Could you elaborate on<mask> this is a poor idea?</s>
Label encoding: <s> [STARTQ] Unconditional gets much of the time on reddit but is extraordinarily poorly supported in advanced economies; its incredibly expensive (the distortionary effects from increasing taxation to pay for it would counteract its economic benefits many times over), would have a huge labor discouragement issue and would cause significant inflationary problems. Unconditional basic income in an advanced economy would eviscerate economic growth without correcting many of the problems those who support it claim. [ENDQ] [NEWLINE] I guess I've been victim to the misinformation. I always thought that an unconditional basic income, + clawback from unneeding citizens through income taxes, was the preferable system. Could you elaborate on why this is a poor idea?</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> did some 75% of heroin-addicted Vietnam vets kick the drug<mask> they returned home? [ENDQ] [NEWLINE] There's a question<mask> to whether they were *addicted* vs self medicating based on [environmental stresses]( [URL] /) [NEWLINE] [NEWLINE] Now, before you go saying that this proves your point, I would point out that not all diseases are physical.  PTSD is (almost?) purely psychological,<mask> I don't think you can claim it isn't *really* a malady... [NEWLINE] [NEWLINE] [STARTQ] yeah, I know<mask> you mean, I drink a lot. [ENDQ] [NEWLINE] "Drink a lot"!= "Don't know<mask> to *not* drink"</s>
Label encoding: <s> [STARTQ] Why did some 75% of heroin-addicted Vietnam vets kick the drug when they returned home? [ENDQ] [NEWLINE] There's a question as to whether they were *addicted* vs self medicating based on [environmental stresses]( [URL] /) [NEWLINE] [NEWLINE] Now, before you go saying that this proves your point, I would point out that not all diseases are physical.  PTSD is (almost?) purely psychological, but I don't think you can claim it isn't *really* a malady... [NEWLINE] [NEWLINE] [STARTQ] yeah, I know what you mean, I drink a lot. [ENDQ] [NEWLINE] "Drink a lot"!= "Don't know how to *not* drink"</s>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0996, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because you can have sex without planning on having a child. Aside from the fact that rape is common, condom's break, the pill can fail, even a vasectomy can fail and still get a girl pregnant. Not to mention young girls who are uneducated about sex and might not understand the risk of pregnancy, especially due to the prevalence of "abstinence only" education, which has been found to result in a surprising number of teenagers who think that they can't get pregnant<mask> it's their first time,<mask> the girl is on top,<mask> they don't want to, and etc. [NEWLINE] [NEWLINE] Simply put. Sex is not the choice to have a child.</s>
Label encoding: <s>Because you can have sex without planning on having a child. Aside from the fact that rape is common, condom's break, the pill can fail, even a vasectomy can fail and still get a girl pregnant. Not to mention young girls who are uneducated about sex and might not understand the risk of pregnancy, especially due to the prevalence of "abstinence only" education, which has been found to result in a surprising number of teenagers who think that they can't get pregnant if it's their first time, if the girl is on top, if they don't want to, and etc. [NEWLINE] [NEWLINE] Simply put. Sex is not the choice to have a child.</s>
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0895, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you. ∆ [NEWLINE] Perhaps instead of "Modernity is evil" I ought to have led with something like: We,<mask> the beneficiaries of modernity, ought to immediately cease the valorization of western liberal norms based on the violence with which they have always concomitant. We ought to reconceptualize and revitalize our notions of rights by engaging with Indigenous epistimologies, which themselves are continually supressed within the current dominant order. [NEWLINE] [NEWLINE] In other words, we have a lot to learn from our suppressed past. We already need to change things drastically<mask> we (humans and ecosystems) are going to survive this century.    </s>
Label encoding: <s>I agree with you. ∆ [NEWLINE] Perhaps instead of "Modernity is evil" I ought to have led with something like: We, as the beneficiaries of modernity, ought to immediately cease the valorization of western liberal norms based on the violence with which they have always concomitant. We ought to reconceptualize and revitalize our notions of rights by engaging with Indigenous epistimologies, which themselves are continually supressed within the current dominant order. [NEWLINE] [NEWLINE] In other words, we have a lot to learn from our suppressed past. We already need to change things drastically if we (humans and ecosystems) are going to survive this century.    </s>
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL] [NEWLINE] [NEWLINE] A score of 50 on the ASVAB equates to a score of about 110 on standardized iq tests.  Source- my friend of 25 years was a recruiter. [NEWLINE] [NEWLINE] Even the most lowly grunt foot soldier must be capable of operating sophisticated weapon systems,  often with minimal or no formal training in improvised,  stressful combat situations.  Our advanced weapons systems and high quality personell is<mask> we can routinely engage hostiles at 10:1 Frontline numbers and expect to win. [NEWLINE] [NEWLINE] We rely extensively on Force multipliers for our military prowess...<mask> force multipliers<mask> act<mask> stupidity multipliers in the wrong hands. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [URL] [NEWLINE] [NEWLINE] A score of 50 on the ASVAB equates to a score of about 110 on standardized iq tests.  Source- my friend of 25 years was a recruiter. [NEWLINE] [NEWLINE] Even the most lowly grunt foot soldier must be capable of operating sophisticated weapon systems,  often with minimal or no formal training in improvised,  stressful combat situations.  Our advanced weapons systems and high quality personell is why we can routinely engage hostiles at 10:1 Frontline numbers and expect to win. [NEWLINE] [NEWLINE] We rely extensively on Force multipliers for our military prowess... But force multipliers also act as stupidity multipliers in the wrong hands. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah<mask> you word it that way, it actually seems pretty obvious that even without rape culture, the questioning about a traumatic experience would be pretty traumatic even without bias against you. I still think the victim should report,<mask> I'd be more understanding<mask> he/she didnt. I still don't agree that the system is mostly biased against the victim,<mask> i do agree about the topic now. I guess i was thinking to much into it. Can I give deltas or is that only for op? [NEWLINE] [NEWLINE] Edit: Did someone really downvote my post about<mask> my view was changed, without any explanation at all, in a subreddit about changing views? Really?</s>
Label encoding: <s>Yeah when you word it that way, it actually seems pretty obvious that even without rape culture, the questioning about a traumatic experience would be pretty traumatic even without bias against you. I still think the victim should report, but I'd be more understanding if he/she didnt. I still don't agree that the system is mostly biased against the victim, but i do agree about the topic now. I guess i was thinking to much into it. Can I give deltas or is that only for op? [NEWLINE] [NEWLINE] Edit: Did someone really downvote my post about how my view was changed, without any explanation at all, in a subreddit about changing views? Really?</s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Different rules for turning at a red light [ENDQ] [STARTQ] [ENDQ] [STARTQ] No stopping for Bikes [ENDQ] [NEWLINE] Those are two specific exemptions. Not laws. It's like "No parking" [NEWLINE] [NEWLINE] The speed/weight/idling laws are for the road. The road can't handle the weight. Idling is noise. Trucks obviously make more noise. [NEWLINE] [NEWLINE] All the things you are listing are minor exemptions to the laws in place. Not actual laws themselves. [NEWLINE] [NEWLINE] The rules aren't useless. They are a basis to build on. EVERYONE has to stop at a red light.<mask> you are a truck you have to stop at red lights AND train tracks.</s>
Label encoding: <s> [STARTQ] Different rules for turning at a red light [ENDQ] [STARTQ] [ENDQ] [STARTQ] No stopping for Bikes [ENDQ] [NEWLINE] Those are two specific exemptions. Not laws. It's like "No parking" [NEWLINE] [NEWLINE] The speed/weight/idling laws are for the road. The road can't handle the weight. Idling is noise. Trucks obviously make more noise. [NEWLINE] [NEWLINE] All the things you are listing are minor exemptions to the laws in place. Not actual laws themselves. [NEWLINE] [NEWLINE] The rules aren't useless. They are a basis to build on. EVERYONE has to stop at a red light. If you are a truck you have to stop at red lights AND train tracks.</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Soft_Needles, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Soft_Needles+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry Soft_Needles, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Soft_Needles+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I guess American intervention in the Middle East and South America isn't "really" capitalism then, is it? [ENDQ] [NEWLINE] I don't think<mask>.<mask> I'd take that over your 'not really communism' any day. [NEWLINE] [NEWLINE] [STARTQ] Thousands upon thousands of years of human history are on my side in this. [ENDQ] [NEWLINE] Thousands upon thousands of years of human suffering prove you correct? [NEWLINE] [NEWLINE] [STARTQ] <mask> claiming capitalism is this great and wonderful economic system is<mask> foolish<mask> doing the same with communism. [ENDQ] [NEWLINE] I would<mask><mask> a communist in this day and age is far more foolish. Again, I welcome a death toll comparison<mask> you remain unconvinced.</s>
Label encoding: <s> [STARTQ] I guess American intervention in the Middle East and South America isn't "really" capitalism then, is it? [ENDQ] [NEWLINE] I don't think so. But I'd take that over your 'not really communism' any day. [NEWLINE] [NEWLINE] [STARTQ] Thousands upon thousands of years of human history are on my side in this. [ENDQ] [NEWLINE] Thousands upon thousands of years of human suffering prove you correct? [NEWLINE] [NEWLINE] [STARTQ] But claiming capitalism is this great and wonderful economic system is as foolish as doing the same with communism. [ENDQ] [NEWLINE] I would argue that a communist in this day and age is far more foolish. Again, I welcome a death toll comparison if you remain unconvinced.</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My view has been shaped a lot about this discussion, and even now I am unsure about my original argument and the issue of levels of blindness/deafness.  The only distinction I can still see is that<mask> you are considered any level of blind or deaf, it describes specifically your ability to see or hear. <mask> someone is considered mentally retarded, it refers to an impaired development,<mask> that doesn't necessarily lead to any discernible outcome, such<mask> being able to live on your own or needing constant care.  Being blind or deaf means you can't see or hear.  That is the outcome.  Being mentally retarded is ambiguous in the outcome.</s>
Label encoding: <s>My view has been shaped a lot about this discussion, and even now I am unsure about my original argument and the issue of levels of blindness/deafness.  The only distinction I can still see is that if you are considered any level of blind or deaf, it describes specifically your ability to see or hear.  If someone is considered mentally retarded, it refers to an impaired development, but that doesn't necessarily lead to any discernible outcome, such as being able to live on your own or needing constant care.  Being blind or deaf means you can't see or hear.  That is the outcome.  Being mentally retarded is ambiguous in the outcome.</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry FraggedFoundry, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=FraggedFoundry+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry FraggedFoundry, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=FraggedFoundry+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, it is clear that you believe that it wouldn't be a problem, much like Rand Paul did<mask> he offered a similar remark. [NEWLINE] [NEWLINE] I've tried to refrain from commenting on that aspect of things,<mask> even<mask> I credit you fully with such idealism, I cannot do<mask> with others who espouse the same view.  Still, I feel that would end up being a discussion of anecdotes and not especially productive. [NEWLINE] [NEWLINE] <mask> I stick to<mask> I see<mask> the most egregious problem, that to allow discrimination on such terms, it will inevitably involve the state and its agents.  That is a line I am unwilling to cross. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Yes, it is clear that you believe that it wouldn't be a problem, much like Rand Paul did when he offered a similar remark. [NEWLINE] [NEWLINE] I've tried to refrain from commenting on that aspect of things, because even if I credit you fully with such idealism, I cannot do so with others who espouse the same view.  Still, I feel that would end up being a discussion of anecdotes and not especially productive. [NEWLINE] [NEWLINE] So I stick to what I see as the most egregious problem, that to allow discrimination on such terms, it will inevitably involve the state and its agents.  That is a line I am unwilling to cross. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's like saying a law on inveterate shoppers is gender-neutral. [ENDQ] [NEWLINE] You need to explain further<mask> this is supposed to mean. [NEWLINE] [NEWLINE] Alimony laws are gender neutral.  In cases<mask> women were making more than their husbands and especially in cases<mask> the husband stayed at home women pay alimony. [NEWLINE] [NEWLINE] <mask> you want more women to pay alimony you should be fighting for women to earn more -<mask> I'm guessing that's not<mask> you really want. [NEWLINE] [NEWLINE] [STARTQ] <mask> feminism is about seeing women<mask> people and women are not people, then who are? [ENDQ] [NEWLINE] Excuse me?  Women aren't people?</s>
Label encoding: <s> [STARTQ] It's like saying a law on inveterate shoppers is gender-neutral. [ENDQ] [NEWLINE] You need to explain further what this is supposed to mean. [NEWLINE] [NEWLINE] Alimony laws are gender neutral.  In cases where women were making more than their husbands and especially in cases where the husband stayed at home women pay alimony. [NEWLINE] [NEWLINE] If you want more women to pay alimony you should be fighting for women to earn more - but I'm guessing that's not what you really want. [NEWLINE] [NEWLINE] [STARTQ] If feminism is about seeing women as people and women are not people, then who are? [ENDQ] [NEWLINE] Excuse me?  Women aren't people?</s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Was it a revolutionary concept? Woman were extremely important religiously in Roman culture and in most other Pagan cultures. [NEWLINE] [NEWLINE] Christianity has categorically not been the single greatest force against any of those things. The invention of vaccines, the invention of fertilizer, the invention of democracy, the invention of pesticides, the invention of medicine did the things you list. In many cases Christianity slowed and hindered and still does slow and hinder. Man did those things and Religion is the least of mans creations. [NEWLINE] [NEWLINE] Christian societies? Christian societies are from the middle ages. We live in societies founded on the ideas of the enlightenment, humanist societies. The red cross are people. </s>
Label encoding: <s>Was it a revolutionary concept? Woman were extremely important religiously in Roman culture and in most other Pagan cultures. [NEWLINE] [NEWLINE] Christianity has categorically not been the single greatest force against any of those things. The invention of vaccines, the invention of fertilizer, the invention of democracy, the invention of pesticides, the invention of medicine did the things you list. In many cases Christianity slowed and hindered and still does slow and hinder. Man did those things and Religion is the least of mans creations. [NEWLINE] [NEWLINE] Christian societies? Christian societies are from the middle ages. We live in societies founded on the ideas of the enlightenment, humanist societies. The red cross are people. </s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason it get beaten to death is it shows your lack of understanding of the issue. I read a lot hate about Walmart on reddit,not unlike yours, and tends be focused on talking points rather real life experiences with the store. [NEWLINE] [NEWLINE] I am no fan of going to Walmart at peak hours<mask> i do not like dealing with the hustle and bussle that surrounds a visit to Walmart.<mask> it provideds a cheap and extremely conveint source for just about everything one could need. I use to go Walmart in college all the time<mask> I was able to in the mornings<mask> it wasn't busy. It saved me a ton of money. [NEWLINE] </s>
Label encoding: <s>The reason it get beaten to death is it shows your lack of understanding of the issue. I read a lot hate about Walmart on reddit,not unlike yours, and tends be focused on talking points rather real life experiences with the store. [NEWLINE] [NEWLINE] I am no fan of going to Walmart at peak hours because i do not like dealing with the hustle and bussle that surrounds a visit to Walmart. However it provideds a cheap and extremely conveint source for just about everything one could need. I use to go Walmart in college all the time because I was able to in the mornings when it wasn't busy. It saved me a ton of money. [NEWLINE] </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If the toilet seat is the most pressing issue in the relationship, fall on your knees and thank god for it every morning. [NEWLINE] [NEWLINE] (Seriously,<mask> you spouse says "we're buying a house, having kids, not having kids, moving to Japan," it's probably a good topic for a discussion.<mask> you spouse says "honey-bunch, I prefer not to fall into the toilet in the middle of the night, my darling love-bunny," let it go. She's probably letting go you never putting a cap on the toothpaste or leaving socks outside the laundry hamper, etc.) [NEWLINE] [NEWLINE] Source: Married 20 years.</s>
Label encoding: <s>If the toilet seat is the most pressing issue in the relationship, fall on your knees and thank god for it every morning. [NEWLINE] [NEWLINE] (Seriously, if you spouse says "we're buying a house, having kids, not having kids, moving to Japan," it's probably a good topic for a discussion. If you spouse says "honey-bunch, I prefer not to fall into the toilet in the middle of the night, my darling love-bunny," let it go. She's probably letting go you never putting a cap on the toothpaste or leaving socks outside the laundry hamper, etc.) [NEWLINE] [NEWLINE] Source: Married 20 years.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've removed this comment per Rule 1: "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question.". [See the wiki page for more information.]( [URL] #wiki_rule_1). [NEWLINE] [NEWLINE] <mask> you wish to edit your post to more directly challenge an aspect of the OP's view, go ahead and then [message the moderators]( [URL] )<mask> we can re-approve it. [NEWLINE] [NEWLINE] <mask> you still wish to argue on OP's side, then you're welcome to do<mask> in replies to other people's comments.</s><pad>
Label encoding: <s>I've removed this comment per Rule 1: "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question.". [See the wiki page for more information.]( [URL] #wiki_rule_1). [NEWLINE] [NEWLINE] If you wish to edit your post to more directly challenge an aspect of the OP's view, go ahead and then [message the moderators]( [URL] ) so we can re-approve it. [NEWLINE] [NEWLINE] If you still wish to argue on OP's side, then you're welcome to do so in replies to other people's comments.</s><pad>
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Tons of professions care about whether or not you've gone to art or music school before you're hired. [NEWLINE] [NEWLINE] [NEWLINE] You basically have to have those degrees to get most good teaching type of jobs, they matter a shit ton<mask> it comes to commissions, and most importantly the structure and pedagogy of art and music schools is not presently replaceable by the internet. Having one on one lessons and feedback with university level music and art professors can not be replaced by a textbook. Especially<mask> it comes to artistic creative expression, one on one and communal feedback that you get from professors and classmates is invaluable. That's not something you can replace easily. </s>
Label encoding: <s>Tons of professions care about whether or not you've gone to art or music school before you're hired. [NEWLINE] [NEWLINE] [NEWLINE] You basically have to have those degrees to get most good teaching type of jobs, they matter a shit ton when it comes to commissions, and most importantly the structure and pedagogy of art and music schools is not presently replaceable by the internet. Having one on one lessons and feedback with university level music and art professors can not be replaced by a textbook. Especially when it comes to artistic creative expression, one on one and communal feedback that you get from professors and classmates is invaluable. That's not something you can replace easily. </s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That is true,<mask> of the rights we have ending a life is not one of them. [NEWLINE] [NEWLINE] <mask> it comes to assisted suicide,<mask> it is ever the option it should be the very last one of them. Depending on the issue, there should almost always be another way.<mask> a person was going to make that decision for themselves, I believe it's possible that there is another less extreme solution that is possible that the person may not be aware of. Without suicide or assisted suicide<mask> an available option, it's more incentive to use or search for a more appropriate solution that can save the person's life and the condition the person may be in.</s>
Label encoding: <s>That is true, however of the rights we have ending a life is not one of them. [NEWLINE] [NEWLINE] When it comes to assisted suicide, if it is ever the option it should be the very last one of them. Depending on the issue, there should almost always be another way. If a person was going to make that decision for themselves, I believe it's possible that there is another less extreme solution that is possible that the person may not be aware of. Without suicide or assisted suicide as an available option, it's more incentive to use or search for a more appropriate solution that can save the person's life and the condition the person may be in.</s>
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's an argument that nonmonogamy is more biologically ingrained that monogamy is. Look at our closest relatives with bonobos and chimpanzees - they have tribal societies, and we used to too. There's none of the staleness that a monogamous relationship can have, the entire group takes care of children, there's a lot of benefits to that. [NEWLINE] [NEWLINE] Being open is just opening up to the idea that maybe one person isn't the end-all of relationships. Look at /r/deadbedrooms, /r/breakingmom, the divorce rate, and see<mask> there are a lot of people not happy with their current status.</s>
Label encoding: <s>There's an argument that nonmonogamy is more biologically ingrained that monogamy is. Look at our closest relatives with bonobos and chimpanzees - they have tribal societies, and we used to too. There's none of the staleness that a monogamous relationship can have, the entire group takes care of children, there's a lot of benefits to that. [NEWLINE] [NEWLINE] Being open is just opening up to the idea that maybe one person isn't the end-all of relationships. Look at /r/deadbedrooms, /r/breakingmom, the divorce rate, and see how there are a lot of people not happy with their current status.</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's....still just<mask> determined by you.  Claiming that anything less than strict adherence to the topic of a post in the comments (i.e. no tangents) detracts from the post is not necessarily something that is taken for granted on Reddit.  I've had plenty of delightful comment conversations that were somewhat tangential to the content of the post.  For the post you described, I probably wouldn't personally be too interested in a tangential comment about<mask> some character still looked hot,<mask> I'd express that by not-upvoting it, not by claiming some general rule that that commented detracted from the discussion.</s>
Label encoding: <s>That's....still just as determined by you.  Claiming that anything less than strict adherence to the topic of a post in the comments (i.e. no tangents) detracts from the post is not necessarily something that is taken for granted on Reddit.  I've had plenty of delightful comment conversations that were somewhat tangential to the content of the post.  For the post you described, I probably wouldn't personally be too interested in a tangential comment about how some character still looked hot, but I'd express that by not-upvoting it, not by claiming some general rule that that commented detracted from the discussion.</s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your issue is with the term superior or inferior.<mask> certain aspects of a culture may be more criminal, or offensive, this does not make them inherently inferior. Inferior is entirely up to the person viewing it. [NEWLINE] [NEWLINE] <mask> intrinsic numerical value would you assign to each trait of various cultures that you could weigh them against each other and come up with a white culture&gt;black culture answer? For example black culture<mask> promotes physical exercise at least in males. Maybe that is worth +5 points.<mask> that scale is entirely in the eyes of the beholder. Nothing any human does is implicitly inferior or superior than<mask> another does.</s>
Label encoding: <s>I think your issue is with the term superior or inferior. While certain aspects of a culture may be more criminal, or offensive, this does not make them inherently inferior. Inferior is entirely up to the person viewing it. [NEWLINE] [NEWLINE] What intrinsic numerical value would you assign to each trait of various cultures that you could weigh them against each other and come up with a white culture&gt;black culture answer? For example black culture also promotes physical exercise at least in males. Maybe that is worth +5 points. But that scale is entirely in the eyes of the beholder. Nothing any human does is implicitly inferior or superior than what another does.</s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>FallingSnowAngel wasn't making an assumption, they're arguing that the taboo exists to deter people from abusing power they might have over their family members. You're talking about individual cases and they're talking about things at a societal level. [NEWLINE] [NEWLINE] <mask>, the second paragraph directly addresses your point: [NEWLINE] [NEWLINE] [STARTQ] The taboo isn't supposed to judge those in loving, compassionate relationships. It's just meant to make the possibility for abuse more difficult... [ENDQ] [NEWLINE] <mask><mask> it's very similar to students who end up in relationships with their teachers. There's a taboo<mask> of the increased possibility of abuse,<mask><mask> whether or not individual relationships are abusive.</s>
Label encoding: <s>FallingSnowAngel wasn't making an assumption, they're arguing that the taboo exists to deter people from abusing power they might have over their family members. You're talking about individual cases and they're talking about things at a societal level. [NEWLINE] [NEWLINE] Also, the second paragraph directly addresses your point: [NEWLINE] [NEWLINE] [STARTQ] The taboo isn't supposed to judge those in loving, compassionate relationships. It's just meant to make the possibility for abuse more difficult... [ENDQ] [NEWLINE] I think it's very similar to students who end up in relationships with their teachers. There's a taboo because of the increased possibility of abuse, regardless of whether or not individual relationships are abusive.</s>
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0833, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's the problem. [ENDQ] [NEWLINE] Not anymore than other crimes. [NEWLINE] [NEWLINE] [STARTQ] The conviction rates of rape are<mask> impossibly low they might<mask> well be non-existent. [ENDQ] [NEWLINE] you are again repeating the nonsensical canard<mask> the researcher herself is pointing out that this is<mask> causes rapes to go unreported in the first place. [NEWLINE] [NEWLINE] [STARTQ] It's a lot lower than your claimed 60% [ENDQ] [NEWLINE] My claimed? it's from the report you're picking the 9 out 10 rapes go unreported fact from. [NEWLINE] [NEWLINE] [STARTQ] Let's scroll down even further. [ENDQ] [NEWLINE] No, first you bother to read the article from the top.</s>
Label encoding: <s> [STARTQ] There's the problem. [ENDQ] [NEWLINE] Not anymore than other crimes. [NEWLINE] [NEWLINE] [STARTQ] The conviction rates of rape are so impossibly low they might as well be non-existent. [ENDQ] [NEWLINE] you are again repeating the nonsensical canard when the researcher herself is pointing out that this is what causes rapes to go unreported in the first place. [NEWLINE] [NEWLINE] [STARTQ] It's a lot lower than your claimed 60% [ENDQ] [NEWLINE] My claimed? it's from the report you're picking the 9 out 10 rapes go unreported fact from. [NEWLINE] [NEWLINE] [STARTQ] Let's scroll down even further. [ENDQ] [NEWLINE] No, first you bother to read the article from the top.</s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'd<mask><mask> for yoga, the health benefits are rather low, and the risks are moderate (<mask> of the risk of causing/exacerbating injury with inappropriate stretches, and<mask> yoga is often promoted<mask> an alternative medicine cure-all for everything from back pain to fatigue to depression - alternative medicine can cause people to delay seeking necessary help from mainstream medicine.) [ENDQ] [NEWLINE] And this is... speculation backed by nothing?  You certainly use the word pseudoscience a lot for someone who believes their preconceived notions are reality. [NEWLINE] [NEWLINE] And yoga stances are specifically designed not to injure you, they're aren't random or "inappropriate." [NEWLINE] </s>
Label encoding: <s> [STARTQ] I'd argue that for yoga, the health benefits are rather low, and the risks are moderate ( because of the risk of causing/exacerbating injury with inappropriate stretches, and because yoga is often promoted as an alternative medicine cure-all for everything from back pain to fatigue to depression - alternative medicine can cause people to delay seeking necessary help from mainstream medicine.) [ENDQ] [NEWLINE] And this is... speculation backed by nothing?  You certainly use the word pseudoscience a lot for someone who believes their preconceived notions are reality. [NEWLINE] [NEWLINE] And yoga stances are specifically designed not to injure you, they're aren't random or "inappropriate." [NEWLINE] </s>
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That is not<mask> statistics work either...<mask> questioning a statistic we would look at observed data and check to see whether the events that occurred are significant at a certain level. We would take the sample size of the amount of people his wife knows getting immunized and test the value of 3 people getting the disease at rate of 1/1,000,000 and see<mask> it is statistically significant. It is obvious to see that it is a very rare occurrence<mask> the rate 1/1,000,000 would be wrong,<mask><mask> that the sample size is very small and i don't know<mask> would happen<mask> looking at a more national scale.</s><pad>
Label encoding: <s>That is not how statistics work either... When questioning a statistic we would look at observed data and check to see whether the events that occurred are significant at a certain level. We would take the sample size of the amount of people his wife knows getting immunized and test the value of 3 people getting the disease at rate of 1/1,000,000 and see if it is statistically significant. It is obvious to see that it is a very rare occurrence so the rate 1/1,000,000 would be wrong, but also that the sample size is very small and i don't know what would happen when looking at a more national scale.</s><pad>
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>False. False. False. This has been shown to be untrue time and time again. Car drivers who cut off motorcyclists, or make a left turn, cutting them off at intersections are unable to hear the motorcycles coming up from behind, or the pipes facing the opposite directions. [NEWLINE] [NEWLINE] "Loud Pipes Save Lives" Is BULLSHIT. [NEWLINE] [NEWLINE] Yes, I ride. And I have heard this over and over again. It's bullshit. [NEWLINE] [NEWLINE] The most significant reason for motorcycle accidents (80%) is the rider being at fault. Not cars. [NEWLINE] [NEWLINE] Stop perpetuating this bullshit. [NEWLINE] [NEWLINE] Thank You. </s>
Label encoding: <s>False. False. False. This has been shown to be untrue time and time again. Car drivers who cut off motorcyclists, or make a left turn, cutting them off at intersections are unable to hear the motorcycles coming up from behind, or the pipes facing the opposite directions. [NEWLINE] [NEWLINE] "Loud Pipes Save Lives" Is BULLSHIT. [NEWLINE] [NEWLINE] Yes, I ride. And I have heard this over and over again. It's bullshit. [NEWLINE] [NEWLINE] The most significant reason for motorcycle accidents (80%) is the rider being at fault. Not cars. [NEWLINE] [NEWLINE] Stop perpetuating this bullshit. [NEWLINE] [NEWLINE] Thank You. </s>
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0863, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The reason this lyric makes people feel uncomfortable is that it's at least reminiscent of a common date rape tactic.<mask> the girl is reluctant, just trick her into getting really drunk and then have sex with her<mask> she's too inebriated to say no. [ENDQ] [NEWLINE] Absolutely, and it's totally fair for people in 2014 to be uncomfortable and think that<mask> date rape via alcohol is<mask> prevalent in our society currently. I'm not sure that it was<mask> much back then, and my argument is that the song wasn't written in 1944 to be about it,<mask> the fact that it may make us all think of it today.</s>
Label encoding: <s> [STARTQ] The reason this lyric makes people feel uncomfortable is that it's at least reminiscent of a common date rape tactic. If the girl is reluctant, just trick her into getting really drunk and then have sex with her when she's too inebriated to say no. [ENDQ] [NEWLINE] Absolutely, and it's totally fair for people in 2014 to be uncomfortable and think that because date rape via alcohol is so prevalent in our society currently. I'm not sure that it was so much back then, and my argument is that the song wasn't written in 1944 to be about it, despite the fact that it may make us all think of it today.</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Other posters have done a great job of covering the major arguments<mask> here's a more subtle one. [NEWLINE] [NEWLINE] Let's say that you pirated a 2-hour movie and then watched it.<mask> piracy were somehow (magically) impossible,<mask> would you have done with those 2 hours instead? [NEWLINE] [NEWLINE] I'd<mask><mask><mask> you pirate a product you never would have bought, you deprive competing products of money too. Imagine<mask> much smaller your library of entertainment would be<mask> you couldn't pirate anything. Then you might need something to fill your free time and products that you currently think are slightly too expensive would move into your price range.</s>
Label encoding: <s>Other posters have done a great job of covering the major arguments so here's a more subtle one. [NEWLINE] [NEWLINE] Let's say that you pirated a 2-hour movie and then watched it. If piracy were somehow (magically) impossible, what would you have done with those 2 hours instead? [NEWLINE] [NEWLINE] I'd argue that when you pirate a product you never would have bought, you deprive competing products of money too. Imagine how much smaller your library of entertainment would be if you couldn't pirate anything. Then you might need something to fill your free time and products that you currently think are slightly too expensive would move into your price range.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd like to interject with this random datum: [NEWLINE] [NEWLINE] Crack is assumed in this argument and others to be<mask> addictive<mask> to be blinding.  There are no controls for it, crack will make you sell your home. [NEWLINE] [NEWLINE] [This study]( [URL]?_r=0) seems to contradict this concept.  In it, crack addicts are observed to behave rationally in refraining from use.   The article suggests that crack has such high failure rates for quitting<mask> it's usually the poorest folk who are addicted, and poor folk behave less rationally about money to start with. [NEWLINE] [NEWLINE] Just a point.</s>
Label encoding: <s>I'd like to interject with this random datum: [NEWLINE] [NEWLINE] Crack is assumed in this argument and others to be so addictive as to be blinding.  There are no controls for it, crack will make you sell your home. [NEWLINE] [NEWLINE] [This study]( [URL]?_r=0) seems to contradict this concept.  In it, crack addicts are observed to behave rationally in refraining from use.   The article suggests that crack has such high failure rates for quitting because it's usually the poorest folk who are addicted, and poor folk behave less rationally about money to start with. [NEWLINE] [NEWLINE] Just a point.</s>
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>60+ hour weeks aren't typical<mask> long hours are viewed<mask> a good and noble thing by most Americans.<mask> you aren't wealthy leisure is seen<mask> decadent. [NEWLINE] [NEWLINE] The 9-5, mon-fri schedule has disappeared with the proliferation of mobile devices.  The pressure to be productive is immense.  The people who put in hours late at night and on weekends are obviously favored. I'm given generous vacation time<mask> I never have a chance to use it all. [NEWLINE] [NEWLINE] <mask><mask> Europeans have a much healthier attitude towards working. OP is lucky to be<mask> he is and shouldn't leave without considering cultural differences. </s>
Label encoding: <s>60+ hour weeks aren't typical but long hours are viewed as a good and noble thing by most Americans. If you aren't wealthy leisure is seen as decadent. [NEWLINE] [NEWLINE] The 9-5, mon-fri schedule has disappeared with the proliferation of mobile devices.  The pressure to be productive is immense.  The people who put in hours late at night and on weekends are obviously favored. I'm given generous vacation time but I never have a chance to use it all. [NEWLINE] [NEWLINE] I think Europeans have a much healthier attitude towards working. OP is lucky to be where he is and shouldn't leave without considering cultural differences. </s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] <mask><mask> that the education aspect of it really does it for me. I'm not really thinking in terms of scholarships, just really it amazes me<mask> much a college will downgrade their admissions requirements just to let a certain person in. [NEWLINE] [NEWLINE] <mask> you are right in saying that those athletes still will learn, even<mask> they still aren't doing<mask> well academically in their college classes. They will be learning the art of their sports. [NEWLINE] [NEWLINE] I only focused on athletes<mask> that is<mask> I am the most familiar with, and seems to be the most popular in instances of colleges making exceptions. [NEWLINE] [NEWLINE] Good points!</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] I think that the education aspect of it really does it for me. I'm not really thinking in terms of scholarships, just really it amazes me how much a college will downgrade their admissions requirements just to let a certain person in. [NEWLINE] [NEWLINE] BUT you are right in saying that those athletes still will learn, even if they still aren't doing as well academically in their college classes. They will be learning the art of their sports. [NEWLINE] [NEWLINE] I only focused on athletes as that is what I am the most familiar with, and seems to be the most popular in instances of colleges making exceptions. [NEWLINE] [NEWLINE] Good points!</s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How the hell is talking about IQ or FSIQ ego-stroking? "Oh look at me I know<mask> IQ stands for praise me!!!" Is that<mask> you think I sound like? [NEWLINE] [NEWLINE] I'm guess<mask> I haven't made some groundbreaking discovery or fund raised for dying children in Africa I'm just some sort of useless parasite who's accomplished nothing? [NEWLINE] [NEWLINE] I'm not complaining about the "cruel system" wasting all my potential. I realize that I need to take control of my education<mask> all I'm asking for is some help. I guess you didn't have a great experience in your gifted school. [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s>How the hell is talking about IQ or FSIQ ego-stroking? "Oh look at me I know what IQ stands for praise me!!!" Is that what you think I sound like? [NEWLINE] [NEWLINE] I'm guess since I haven't made some groundbreaking discovery or fund raised for dying children in Africa I'm just some sort of useless parasite who's accomplished nothing? [NEWLINE] [NEWLINE] I'm not complaining about the "cruel system" wasting all my potential. I realize that I need to take control of my education but all I'm asking for is some help. I guess you didn't have a great experience in your gifted school. [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Brain surgeons are already readily-available at the hospital. [NEWLINE] [NEWLINE] You keep equating *specialties* within physiological health to having someone who works in psychological health.<mask><mask> that's disingenuous. [NEWLINE] [NEWLINE] There's a key difference between my suggestion and the one you're debating; a big difference between saying that there needs to be specialists in every ambulance and saying there should be ubiquitous emergency mental health services. The former would be like saying police cars should have forensic scientists, intelligence analysts, etc. The latter is more like saying police are busy enough with public safety that we *<mask> * need firefighters to serve related,<mask> distinct services.</s>
Label encoding: <s>Brain surgeons are already readily-available at the hospital. [NEWLINE] [NEWLINE] You keep equating *specialties* within physiological health to having someone who works in psychological health. I think that's disingenuous. [NEWLINE] [NEWLINE] There's a key difference between my suggestion and the one you're debating; a big difference between saying that there needs to be specialists in every ambulance and saying there should be ubiquitous emergency mental health services. The former would be like saying police cars should have forensic scientists, intelligence analysts, etc. The latter is more like saying police are busy enough with public safety that we * also * need firefighters to serve related, but distinct services.</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes I am,<mask> I'm not the person to ask about Islamic banking unless you want only the gist of it. I know very little about this complex topic: [NEWLINE] [NEWLINE] No interest [NEWLINE] [NEWLINE] Mandatory donation (2.5% of your wealth that has been untouched for 1 year(lunar)<mask> you have over a certain amount of wealth. [NEWLINE] [NEWLINE] A way for wealth distribution after the death of a spouse. [NEWLINE] [NEWLINE] Governments (Caliphates) responsibility to make sure you have food to eat. ie.<mask> you steal food<mask> you can't afford food eat it's not your fault its the governments fault.</s>
Label encoding: <s>Yes I am, but I'm not the person to ask about Islamic banking unless you want only the gist of it. I know very little about this complex topic: [NEWLINE] [NEWLINE] No interest [NEWLINE] [NEWLINE] Mandatory donation (2.5% of your wealth that has been untouched for 1 year(lunar) if you have over a certain amount of wealth. [NEWLINE] [NEWLINE] A way for wealth distribution after the death of a spouse. [NEWLINE] [NEWLINE] Governments (Caliphates) responsibility to make sure you have food to eat. ie. If you steal food because you can't afford food eat it's not your fault its the governments fault.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Since<mask> are cigarettes acceptable? We've made a ton of laws regulating who can smoke and<mask>. We've taxed the shit out of them. Cigarette use [has been decreasing for decades and has never been lower]( [URL] /). It's illegal for tobacco companies to advertise their product<mask> a number of agencies actually advertise *against* cigarettes. Can you think of any other product out there that only has commercials telling you *not* to buy it? [NEWLINE] [NEWLINE] I don't know which country you're living in<mask> it's definitely not America. Saying we accept cigarettes is not true, saying that we "support" cigarettes is ridiculous.</s>
Label encoding: <s>Since when are cigarettes acceptable? We've made a ton of laws regulating who can smoke and where. We've taxed the shit out of them. Cigarette use [has been decreasing for decades and has never been lower]( [URL] /). It's illegal for tobacco companies to advertise their product while a number of agencies actually advertise *against* cigarettes. Can you think of any other product out there that only has commercials telling you *not* to buy it? [NEWLINE] [NEWLINE] I don't know which country you're living in but it's definitely not America. Saying we accept cigarettes is not true, saying that we "support" cigarettes is ridiculous.</s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Seems to me that<mask> you truly believed<mask> you typed here, you would have suicided long ago. [NEWLINE] [NEWLINE] Yours is a very clinical, askance perspective on life. It presumes that the efforts put into living aren't worth the fruits of those efforts - that the ends don't justify the means. [NEWLINE] [NEWLINE] It is subjective, after all. Perhaps you see it that way. [NEWLINE] [NEWLINE] <mask> don't state it<mask><mask> it were a fact. "All of the rest of the bullshit" to use your terms, may well mean "Other engaging life activities to become involved in" to others. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Seems to me that if you truly believed what you typed here, you would have suicided long ago. [NEWLINE] [NEWLINE] Yours is a very clinical, askance perspective on life. It presumes that the efforts put into living aren't worth the fruits of those efforts - that the ends don't justify the means. [NEWLINE] [NEWLINE] It is subjective, after all. Perhaps you see it that way. [NEWLINE] [NEWLINE] But don't state it as if it were a fact. "All of the rest of the bullshit" to use your terms, may well mean "Other engaging life activities to become involved in" to others. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't think hardware wallets are the solution. People aren't going to keep their life savings in a hardware wallet [ENDQ] [NEWLINE] Sometimes the terminology is misleading.  The coins aren't actually in the wallet.  It might be better to say 'People aren't going to *control* their life savings with a hardware wallet'.  With a hardware wallet there is a sequence of letters and numbers that you should store somewhere secure that will give you access to your coins<mask> the wallet breaks or is stolen.  The sequence can be encrypted or copied or split in a way that requires 2 out of 3 parts to restore the sequence etc.</s>
Label encoding: <s> [STARTQ] I don't think hardware wallets are the solution. People aren't going to keep their life savings in a hardware wallet [ENDQ] [NEWLINE] Sometimes the terminology is misleading.  The coins aren't actually in the wallet.  It might be better to say 'People aren't going to *control* their life savings with a hardware wallet'.  With a hardware wallet there is a sequence of letters and numbers that you should store somewhere secure that will give you access to your coins if the wallet breaks or is stolen.  The sequence can be encrypted or copied or split in a way that requires 2 out of 3 parts to restore the sequence etc.</s>
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All art is about connecting with your audience. And don't joke about that paul mccartney shit in here, I almost had a heart attack. [NEWLINE] [NEWLINE] Kanye has a style. Just<mask> someone has a dynamic personality doesn't mean they don't have a style. Hell,<mask><mask> the development of a personality is actually<mask> makes a style. [NEWLINE] [NEWLINE] We had conscious rap before Kanye. Mos Def &amp; Talib Kwali, Common, fucking NAS. [NEWLINE] [NEWLINE] Wiz raps about weed, and the weeknd sings about getting fucked up and disrespecting the girls he sleeps with. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>All art is about connecting with your audience. And don't joke about that paul mccartney shit in here, I almost had a heart attack. [NEWLINE] [NEWLINE] Kanye has a style. Just because someone has a dynamic personality doesn't mean they don't have a style. Hell, I think the development of a personality is actually what makes a style. [NEWLINE] [NEWLINE] We had conscious rap before Kanye. Mos Def &amp; Talib Kwali, Common, fucking NAS. [NEWLINE] [NEWLINE] Wiz raps about weed, and the weeknd sings about getting fucked up and disrespecting the girls he sleeps with. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Note:** Your thread has **not** been removed. [NEWLINE] [NEWLINE] Your post's topic seems to be fairly common on this subreddit. Similar posts can be found through our [wiki page]( [URL] #link) or via the [search function]( [URL] ;amp;amp;restrict_sr=on). [NEWLINE] [NEWLINE] Regards, the mods of /r/changemyview. [NEWLINE] [NEWLINE] [NEWLINE] *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/changemyview)<mask> you have any questions or concerns.*</s>
Label encoding: <s>**Note:** Your thread has **not** been removed. [NEWLINE] [NEWLINE] Your post's topic seems to be fairly common on this subreddit. Similar posts can be found through our [wiki page]( [URL] #link) or via the [search function]( [URL] ;amp;amp;restrict_sr=on). [NEWLINE] [NEWLINE] Regards, the mods of /r/changemyview. [NEWLINE] [NEWLINE] [NEWLINE] *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/changemyview) if you have any questions or concerns.*</s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The idea of traffic laws aren't based on some Darwanism argument that people who get hit are dumb and should have know better, it's saying this is a heavy pedestrian area, and we want it to be<mask> safe<mask> possible.<mask> you killed a guy<mask> you were driving fast and he was being dumb, you would still feel bad, someone died, that sucks. You can hit the brakes and try to prevent ending a human life. It's not that hard, and it's good public policy. [NEWLINE] [NEWLINE] Tl;dr even<mask> you're dumb and jump in traffic it's better for you not to die</s>
Label encoding: <s>The idea of traffic laws aren't based on some Darwanism argument that people who get hit are dumb and should have know better, it's saying this is a heavy pedestrian area, and we want it to be as safe as possible. If you killed a guy because you were driving fast and he was being dumb, you would still feel bad, someone died, that sucks. You can hit the brakes and try to prevent ending a human life. It's not that hard, and it's good public policy. [NEWLINE] [NEWLINE] Tl;dr even if you're dumb and jump in traffic it's better for you not to die</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well it kinda fits into the narrative,<mask> you look around the internet you will find a lot of Chad/Chaz bashing. Chad was a popular name for upper middle class white kids at the time, many of them turned out to be on the douchey side. here is a reference from [Urban Dictionary] ( [URL].php?term=chad),<mask><mask> you go through the numerous name association threads chad = douchebag is pretty highly ranked on most threads.<mask> basically having douchebag chad<mask> the lead singer of a douchebag band fit the I hate nickelback narrative very well.</s>
Label encoding: <s>Well it kinda fits into the narrative, If you look around the internet you will find a lot of Chad/Chaz bashing. Chad was a popular name for upper middle class white kids at the time, many of them turned out to be on the douchey side. here is a reference from [Urban Dictionary] ( [URL].php?term=chad), also if you go through the numerous name association threads chad = douchebag is pretty highly ranked on most threads. So basically having douchebag chad as the lead singer of a douchebag band fit the I hate nickelback narrative very well.</s>
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is not true. The statistics show that the percentage of black people earning college degrees has been shooting up and up for many years.<mask> black people didn't value education,<mask> could this be true? [ENDQ] [NEWLINE] It's moving in the right direction,<mask> there's still an overwhelming portion of the black pie chart that isn't taking advantage of the educational opportunities available. [NEWLINE] [NEWLINE] It's important to remember too that not all black people live in the same areas with the same problems. A city with a higher crime rate, higher rates of gang violence, and increased segregation is going to have a harder time with education. </s>
Label encoding: <s> [STARTQ] This is not true. The statistics show that the percentage of black people earning college degrees has been shooting up and up for many years. If black people didn't value education, how could this be true? [ENDQ] [NEWLINE] It's moving in the right direction, but there's still an overwhelming portion of the black pie chart that isn't taking advantage of the educational opportunities available. [NEWLINE] [NEWLINE] It's important to remember too that not all black people live in the same areas with the same problems. A city with a higher crime rate, higher rates of gang violence, and increased segregation is going to have a harder time with education. </s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While<mask><mask>, I don't think that splitting semantic hairs to invalidate the original argument is a very effective way to influence someone's opinion. Especially<mask> e.g. "simply" is used<mask> a colloquial intensifier. [NEWLINE] [NEWLINE] I don't think the spirit of OP's statement was: [NEWLINE] [NEWLINE] [STARTQ] Any fat person can lose weight by adjusting their diet and exercise, a simple process. [ENDQ] [NEWLINE] <mask><mask> it was: [NEWLINE] [NEWLINE] [STARTQ] It's simple that any fat person can lose weight by adjusting their diet and exercise. [ENDQ] [NEWLINE] JM2C,<mask>. I obviously can't speak for OP.</s>
Label encoding: <s>While I agree, I don't think that splitting semantic hairs to invalidate the original argument is a very effective way to influence someone's opinion. Especially when e.g. "simply" is used as a colloquial intensifier. [NEWLINE] [NEWLINE] I don't think the spirit of OP's statement was: [NEWLINE] [NEWLINE] [STARTQ] Any fat person can lose weight by adjusting their diet and exercise, a simple process. [ENDQ] [NEWLINE] I think it was: [NEWLINE] [NEWLINE] [STARTQ] It's simple that any fat person can lose weight by adjusting their diet and exercise. [ENDQ] [NEWLINE] JM2C, though. I obviously can't speak for OP.</s>
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Financial services are unregulated you say? [ENDQ] [NEWLINE] The existence of regulations and regulatory agencies is meaningless *unless* they prevent the self-destructive business practices which emanate from the financial AND energy industries. <mask> your list may appear impressive and onerous, that list meant *squat*<mask> it came to preventing the Financial Crisis or BP Gulf Oil spill. [NEWLINE] [NEWLINE] The [Hedge Fund industry isn't subject to the same regulatory scrutiny<mask> the rest of the financial industry]( [URL] ) and neither is the shadow banking system that Wall Street created to circumvent both those regulations and agencies you are whining about here. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Financial services are unregulated you say? [ENDQ] [NEWLINE] The existence of regulations and regulatory agencies is meaningless *unless* they prevent the self-destructive business practices which emanate from the financial AND energy industries.  While your list may appear impressive and onerous, that list meant *squat* when it came to preventing the Financial Crisis or BP Gulf Oil spill. [NEWLINE] [NEWLINE] The [Hedge Fund industry isn't subject to the same regulatory scrutiny as the rest of the financial industry]( [URL] ) and neither is the shadow banking system that Wall Street created to circumvent both those regulations and agencies you are whining about here. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think you actually have to respect the person behind the work to be a fan. This doesn't mean you disrespect them either. I am a fan of lots of music and art,<mask> I really have no idea who made it other than their name. Fans enjoyment should not be underestimated. Thats the reason Michael Jackson is who he is, after all. [NEWLINE] [NEWLINE] You don't know<mask> the artists wishes are beyond<mask> they said publicly. You don't know the conversations between those chosen to execute their trust. For that reason,<mask><mask> we should respect the wishes of those entrusted to an artists work. </s>
Label encoding: <s>I don't think you actually have to respect the person behind the work to be a fan. This doesn't mean you disrespect them either. I am a fan of lots of music and art, but I really have no idea who made it other than their name. Fans enjoyment should not be underestimated. Thats the reason Michael Jackson is who he is, after all. [NEWLINE] [NEWLINE] You don't know what the artists wishes are beyond what they said publicly. You don't know the conversations between those chosen to execute their trust. For that reason, I think we should respect the wishes of those entrusted to an artists work. </s>
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*sigh*... [NEWLINE] -No. We're not talking about anything else<mask> the hypocrisy in regards to choice in the pro-life movement.<mask>,<mask> no one chooses to take care of that mentally I'll person, there should be no consequence. [NEWLINE] [NEWLINE] -No. You forfeit your rights to human decency<mask> you commit a crime. And fetuses are not people nor "alive". [NEWLINE] [NEWLINE] - No. We're not talking about anything else<mask> the hypocrisy in regards to choice in the pro-life movement. [NEWLINE] [NEWLINE] It's unrealistic to blanket all these issues,<mask> each one is completely different.</s>
Label encoding: <s>*sigh*... [NEWLINE] -No. We're not talking about anything else but the hypocrisy in regards to choice in the pro-life movement. However, if no one chooses to take care of that mentally I'll person, there should be no consequence. [NEWLINE] [NEWLINE] -No. You forfeit your rights to human decency when you commit a crime. And fetuses are not people nor "alive". [NEWLINE] [NEWLINE] - No. We're not talking about anything else but the hypocrisy in regards to choice in the pro-life movement. [NEWLINE] [NEWLINE] It's unrealistic to blanket all these issues, as each one is completely different.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There are ways to make it pretty difficult. It's not like they'd all be stored on a windows 2000 machine. [ENDQ] [NEWLINE] The problem is that it will always create doubt [NEWLINE] [NEWLINE] [STARTQ] People don't trust bitcoin,<mask> the actual protocol has not been breached, just servers that they're on. There are ways of making similar technologies<mask> they'd have to hack more than just 1 server. [ENDQ] [NEWLINE] Oy vey don't get me started on bitcoin<mask> let me say something, these votes don't need to be like bitcoin they can be stolen or manipulated, people don't trust computers for certain things </s>
Label encoding: <s> [STARTQ] There are ways to make it pretty difficult. It's not like they'd all be stored on a windows 2000 machine. [ENDQ] [NEWLINE] The problem is that it will always create doubt [NEWLINE] [NEWLINE] [STARTQ] People don't trust bitcoin, but the actual protocol has not been breached, just servers that they're on. There are ways of making similar technologies where they'd have to hack more than just 1 server. [ENDQ] [NEWLINE] Oy vey don't get me started on bitcoin but let me say something, these votes don't need to be like bitcoin they can be stolen or manipulated, people don't trust computers for certain things </s>
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Does no one else view this<mask> we are being protected from the idiots that cannot control themselves? [NEWLINE] [NEWLINE] Will there be irresponsible people? Yes. [NEWLINE] [NEWLINE] Is it the employers job to worry about this? No. [NEWLINE] [NEWLINE] Is it the government's job to worry about this? Yes. [NEWLINE] [NEWLINE] Birth control is a part of health care and employers have to offer it. [NEWLINE] [NEWLINE] The way I see it is that this protects a future neglected child and generation. It will make our future better, and we won't have people who cannot afford 30,000$ in medical bills crippled along side of us in society.</s>
Label encoding: <s>Does no one else view this as we are being protected from the idiots that cannot control themselves? [NEWLINE] [NEWLINE] Will there be irresponsible people? Yes. [NEWLINE] [NEWLINE] Is it the employers job to worry about this? No. [NEWLINE] [NEWLINE] Is it the government's job to worry about this? Yes. [NEWLINE] [NEWLINE] Birth control is a part of health care and employers have to offer it. [NEWLINE] [NEWLINE] The way I see it is that this protects a future neglected child and generation. It will make our future better, and we won't have people who cannot afford 30,000$ in medical bills crippled along side of us in society.</s>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I am most likely experiencing a filter bubble problem, thanks for sharing that concept, I was not aware of it previously. [ENDQ] [NEWLINE] It was quite telling that your examples were Facebook (which by design creates a filter bubble in your news feed), Reddit (<mask> the user creates their own filter bubble through subreddit subscriptions), and Tumblr (another example of the user creating their own filter bubble based on who they choose to follow). [NEWLINE] [NEWLINE] My experience from starting at the front page of dedicated news websites is that the social issues you highlighted feature far, far less prominently, in favor of a lot more economic and domestic political discussion.</s><pad>
Label encoding: <s> [STARTQ] I am most likely experiencing a filter bubble problem, thanks for sharing that concept, I was not aware of it previously. [ENDQ] [NEWLINE] It was quite telling that your examples were Facebook (which by design creates a filter bubble in your news feed), Reddit ( where the user creates their own filter bubble through subreddit subscriptions), and Tumblr (another example of the user creating their own filter bubble based on who they choose to follow). [NEWLINE] [NEWLINE] My experience from starting at the front page of dedicated news websites is that the social issues you highlighted feature far, far less prominently, in favor of a lot more economic and domestic political discussion.</s><pad>
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not lane-splitting under California law.  That is reckless driving, and an arrestable offense<mask> it should be. [NEWLINE] [NEWLINE] I am advocating lane-splitting within the realm of the recently published CHP guide to legal lane splitting: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] This specifies a top speed of the biker<mask> lane splitting of 39mph, and a max speed differential of 10mph. <mask> in 5mph traffic the biker should stay at 15mph or less. [NEWLINE] [NEWLINE] Any state considering legalizing lane splitting should<mask> back the CHP standards by statute.  California should do<mask><mask> well.</s>
Label encoding: <s>That's not lane-splitting under California law.  That is reckless driving, and an arrestable offense as it should be. [NEWLINE] [NEWLINE] I am advocating lane-splitting within the realm of the recently published CHP guide to legal lane splitting: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] This specifies a top speed of the biker when lane splitting of 39mph, and a max speed differential of 10mph.  So in 5mph traffic the biker should stay at 15mph or less. [NEWLINE] [NEWLINE] Any state considering legalizing lane splitting should also back the CHP standards by statute.  California should do so as well.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Good on you, we need more people like you :) [NEWLINE] [NEWLINE] I would join<mask> i'm stuck in uni in the UK haha. [NEWLINE] [NEWLINE] I want to become involved in Environmental Policy and Politics once i graduate from uni<mask> i find this all very interesting, shame the actions of politicians is all<mask> depressing. [NEWLINE] [NEWLINE] Somewhere in this mess i have just posted a comment to the OP about<mask> steps could be taken, please comment on it<mask> you wish i would love to hear your input. They are mainly just the basic which most people will point out<mask> there is plenty more in my head :)</s>
Label encoding: <s>Good on you, we need more people like you :) [NEWLINE] [NEWLINE] I would join but i'm stuck in uni in the UK haha. [NEWLINE] [NEWLINE] I want to become involved in Environmental Policy and Politics once i graduate from uni so i find this all very interesting, shame the actions of politicians is all so depressing. [NEWLINE] [NEWLINE] Somewhere in this mess i have just posted a comment to the OP about what steps could be taken, please comment on it if you wish i would love to hear your input. They are mainly just the basic which most people will point out but there is plenty more in my head :)</s>
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1462, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask><mask> your moral standard is flawed, that "well-being" would make a better basis than "happiness"<mask> it not only includes happiness<mask> personal development, self-actualization, connection (and contribution) to society, and more.  By focusing on mere happiness, you deprive that person of growth, you deprive the world of the things they might discover or create, and you rob them of their personal agency in the process.  From the "total well-being" standpoint, your actions would by highly immoral, essentially strapping someone down and putting them on a permanent heroine drip.</s>
Label encoding: <s>I would argue that your moral standard is flawed, that "well-being" would make a better basis than "happiness" because it not only includes happiness but personal development, self-actualization, connection (and contribution) to society, and more.  By focusing on mere happiness, you deprive that person of growth, you deprive the world of the things they might discover or create, and you rob them of their personal agency in the process.  From the "total well-being" standpoint, your actions would by highly immoral, essentially strapping someone down and putting them on a permanent heroine drip.</s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you could believe that Watson was the real genius and Holmes was actually an idiot (<mask><mask>, there was an excellent movie made with that premise) [ENDQ] [NEWLINE] Doesn't that work in my favour?<mask> it can be backed up with reference to the text it seems like a reasonable thing to suggest. [NEWLINE] [NEWLINE] [STARTQ] It sounds like you want to start a new religion. OK, make one and then get back to me with your holy text [ENDQ] [NEWLINE] Whoah,<mask> is this coming from? I don't want to start any religion, I'm just enjoying a discussion about an alternate interpretation of a text.</s>
Label encoding: <s> [STARTQ] you could believe that Watson was the real genius and Holmes was actually an idiot ( in fact, there was an excellent movie made with that premise) [ENDQ] [NEWLINE] Doesn't that work in my favour? If it can be backed up with reference to the text it seems like a reasonable thing to suggest. [NEWLINE] [NEWLINE] [STARTQ] It sounds like you want to start a new religion. OK, make one and then get back to me with your holy text [ENDQ] [NEWLINE] Whoah, where is this coming from? I don't want to start any religion, I'm just enjoying a discussion about an alternate interpretation of a text.</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0752, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL] / [NEWLINE] [NEWLINE] It's a great read. Not only is it a great analysis of *Yeezus*,<mask> it<mask> says a lot about Kanye's artistic vision<mask> he creates things. There's a lot he did on this album (and much of his previous works) that really challenged the conventions of Hip-Hop, and that alone should be worth remembering.<mask> imagine that he has been consistently experimenting and moving forward with his style release after release, and you've got something special. You may not like all of his albums,<mask> you can't say any of them aren't trying new things.</s>
Label encoding: <s> [URL] / [NEWLINE] [NEWLINE] It's a great read. Not only is it a great analysis of *Yeezus*, but it also says a lot about Kanye's artistic vision when he creates things. There's a lot he did on this album (and much of his previous works) that really challenged the conventions of Hip-Hop, and that alone should be worth remembering. But imagine that he has been consistently experimenting and moving forward with his style release after release, and you've got something special. You may not like all of his albums, but you can't say any of them aren't trying new things.</s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Continuing my outside the box concept from above,<mask> perhaps those super-churches can be seen<mask> using their charitable money to ensure that local people are employed.  Artisans who normally would not be hired to build fancy statues are now employed by the church in exchange for their services.  And<mask> I understand the concept that people in Somalia need water and that sounds like a much better use of money rather than building statues, it is reasonable to expect that a churches reach is limited and perhaps that each church should help the local community.  Hopefully there are churches in Somalia that could do the same.</s>
Label encoding: <s>Continuing my outside the box concept from above, but perhaps those super-churches can be seen as using their charitable money to ensure that local people are employed.  Artisans who normally would not be hired to build fancy statues are now employed by the church in exchange for their services.  And while I understand the concept that people in Somalia need water and that sounds like a much better use of money rather than building statues, it is reasonable to expect that a churches reach is limited and perhaps that each church should help the local community.  Hopefully there are churches in Somalia that could do the same.</s>
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[lank]( [URL] #q=check%20your%20privilege%2C%20die%20cis%20scum%2C%20shitredditsays%2C%20shitlord&amp;cmpt=q) [NEWLINE] [NEWLINE] This is a pretty good example of the context of usage or at least the association of "Check your privilege" In the end dictionaries are concerned with usage, not the scientific definitions of words. OP is<mask> arguing about the type of people that use the phrase, not the intended meaning of the phrase in a perfect world. </s>
Label encoding: <s>[lank]( [URL] #q=check%20your%20privilege%2C%20die%20cis%20scum%2C%20shitredditsays%2C%20shitlord&amp;cmpt=q) [NEWLINE] [NEWLINE] This is a pretty good example of the context of usage or at least the association of "Check your privilege" In the end dictionaries are concerned with usage, not the scientific definitions of words. OP is also arguing about the type of people that use the phrase, not the intended meaning of the phrase in a perfect world. </s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's a good point.  It's probably severely concentrated into several small areas and I've been lucky enough to experience them first hand.  I wasn't really aware of that study, and<mask> I don't dispute its claim (<mask><mask>, ∆), I will point out that it might not be a perfect representation of the South<mask> [NEWLINE] [NEWLINE] A) Many people don't have internet [NEWLINE] B) Many people wouldn't bother [NEWLINE] C) Many people (my brother included) would've been able to sign the petition,<mask> think "the government is just trying to single out rebels to kill".</s>
Label encoding: <s>That's a good point.  It's probably severely concentrated into several small areas and I've been lucky enough to experience them first hand.  I wasn't really aware of that study, and while I don't dispute its claim ( in fact, ∆), I will point out that it might not be a perfect representation of the South because [NEWLINE] [NEWLINE] A) Many people don't have internet [NEWLINE] B) Many people wouldn't bother [NEWLINE] C) Many people (my brother included) would've been able to sign the petition, but think "the government is just trying to single out rebels to kill".</s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Perhaps it's not at all about<mask> you are coming off to people physically,<mask> your attitude and demeanor<mask> you open your mouth? For instance, I've worked in retail and customer service and after seeing<mask> many different people and personalities, you learn not to judge a book by its cover. That being said, some of the most well put together people have some of the dirtiest attitudes. And<mask><mask><mask><mask><mask>, it got to the point<mask> I'd be more weary of someone who was attractive,<mask> most of the time, they were more rude and quicker to catch an attitude.</s>
Label encoding: <s>Perhaps it's not at all about how you are coming off to people physically, but your attitude and demeanor when you open your mouth? For instance, I've worked in retail and customer service and after seeing so many different people and personalities, you learn not to judge a book by its cover. That being said, some of the most well put together people have some of the dirtiest attitudes. And as a matter of fact, it got to the point where I'd be more weary of someone who was attractive, because most of the time, they were more rude and quicker to catch an attitude.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can make a causal justification for both of the at-risk insurance groups. Young men are more at risk to reckless behavior than women<mask> of testosterone. Women have reproductive healthcare costs that men simply don't have to worry about. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, there's no inherent reason for black people to do more crime. In Senegal, the police don't and can't profile black people. They might profile minority ethnic groups,<mask> with no basis apart from statistical discrimination.<mask> in Senegal, or anywhere else, women still require more healthcare and young men are still more likely to be reckless.</s>
Label encoding: <s>You can make a causal justification for both of the at-risk insurance groups. Young men are more at risk to reckless behavior than women because of testosterone. Women have reproductive healthcare costs that men simply don't have to worry about. [NEWLINE] [NEWLINE] On the other hand, there's no inherent reason for black people to do more crime. In Senegal, the police don't and can't profile black people. They might profile minority ethnic groups, but with no basis apart from statistical discrimination. But in Senegal, or anywhere else, women still require more healthcare and young men are still more likely to be reckless.</s>
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have serious concerns that not enough is being done to save the environment [ENDQ] [NEWLINE] The environment will continue on<mask><mask> our actions (barring nuclear war and a massive fallout/winter).  We,<mask><mask><mask><mask>, are just making the environment [more inhospitable for ourselves]( [URL].jpg).  The bright side of this is that climate change will likely cause huge food shortages and then tons and tons of people will die.  This will cut carbon emissions, which will eventually self correct climate change. <mask> everything is happening just<mask> it should and we are culling ourselves.</s>
Label encoding: <s> [STARTQ] I have serious concerns that not enough is being done to save the environment [ENDQ] [NEWLINE] The environment will continue on regardless of our actions (barring nuclear war and a massive fallout/winter).  We, on the other hand, are just making the environment [more inhospitable for ourselves]( [URL].jpg).  The bright side of this is that climate change will likely cause huge food shortages and then tons and tons of people will die.  This will cut carbon emissions, which will eventually self correct climate change.  So everything is happening just as it should and we are culling ourselves.</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My point is that<mask> we were able to get livable wages without a minimum wage regulation, then we wouldn't have to raise the minimum wage. The fact that we can point at the current minimum wage and have economists agree that it's too low means that without the current minimum wage regulation wages would be *even lower*, showing that we do<mask> need a minimum wage regulation. [NEWLINE] [NEWLINE] I'm not saying that all regulations are good.<mask> i'm saying that deregulation *in and of itself* is not helpful at all, and that the minimum wage is massively helpful to our economy and necessary.</s>
Label encoding: <s>My point is that if we were able to get livable wages without a minimum wage regulation, then we wouldn't have to raise the minimum wage. The fact that we can point at the current minimum wage and have economists agree that it's too low means that without the current minimum wage regulation wages would be *even lower*, showing that we do indeed need a minimum wage regulation. [NEWLINE] [NEWLINE] I'm not saying that all regulations are good. But i'm saying that deregulation *in and of itself* is not helpful at all, and that the minimum wage is massively helpful to our economy and necessary.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As with a lot or moral questions, it comes down to the proximity of the action. Flipping the lever (indirect) comes a lot more easily that pushing the fat man (direct). [NEWLINE] [NEWLINE] Similarly, people cheat more readily on tests in which they're rewarded with tokens they can exchange for money than in tests in which they're rewarded with actual money. [NEWLINE] [NEWLINE] Abortion is a continuum, and the question is<mask> to draw the line. In Western cultures, we're all pretty much in agreement that the line is at or sometime before birth. That is not universal,<mask>.</s><pad>
Label encoding: <s>As with a lot or moral questions, it comes down to the proximity of the action. Flipping the lever (indirect) comes a lot more easily that pushing the fat man (direct). [NEWLINE] [NEWLINE] Similarly, people cheat more readily on tests in which they're rewarded with tokens they can exchange for money than in tests in which they're rewarded with actual money. [NEWLINE] [NEWLINE] Abortion is a continuum, and the question is where to draw the line. In Western cultures, we're all pretty much in agreement that the line is at or sometime before birth. That is not universal, however.</s><pad>
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do people give gifts? You'll never see that robot you gave your nephew again,<mask> every time it puts a smile on their face it will be<mask> of something you did. You created their happiness, and it makes you happy.<mask> you gave away all your possessions the day before you died then the happiness they provide wouldn't be dependent on your physical presence. On a greater scale, your legacy is the same thing; the gift you give to future generations, and<mask> you know that you're leaving a great legacy then you are able to live a more happy and fulfilling life. </s>
Label encoding: <s>Why do people give gifts? You'll never see that robot you gave your nephew again, but every time it puts a smile on their face it will be because of something you did. You created their happiness, and it makes you happy. If you gave away all your possessions the day before you died then the happiness they provide wouldn't be dependent on your physical presence. On a greater scale, your legacy is the same thing; the gift you give to future generations, and if you know that you're leaving a great legacy then you are able to live a more happy and fulfilling life. </s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The way I like to see this is kinda human nature. Survival of the fittest. Eventually, animals we eat could evolve to survive and not be killed off<mask> easily.<mask> you shouldn't think it's wrong.<mask> you think that we should do our best to make bears stop eating fish? Lions stop eating gazelle? We do it<mask> it's in our nature. We are stronger and smarter than a lot of animals we eat,<mask> we use them to live. Survival of the fittest is pretty much the only thing you need to not see eating meat<mask> murder. </s>
Label encoding: <s>The way I like to see this is kinda human nature. Survival of the fittest. Eventually, animals we eat could evolve to survive and not be killed off so easily. But you shouldn't think it's wrong. So you think that we should do our best to make bears stop eating fish? Lions stop eating gazelle? We do it because it's in our nature. We are stronger and smarter than a lot of animals we eat, so we use them to live. Survival of the fittest is pretty much the only thing you need to not see eating meat as murder. </s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's funny<mask> I would be willing to bet that most consumers think they're "buying a game" and not buying a license let alone<mask> that license actually entails.  Incidentally I don't believe consumers are dumb to not know this<mask> that game advertising and sales strategies reinforce this incorrect notion<mask> it is in the interests of the game sellers for people to not fully understand their rights and obligations in these matters.  It might hurt sales<mask> people start considering<mask> they're getting for<mask> they're paying and they can always sue<mask> this ignorance ever actually hurts them at any point.</s>
Label encoding: <s>That's funny because I would be willing to bet that most consumers think they're "buying a game" and not buying a license let alone what that license actually entails.  Incidentally I don't believe consumers are dumb to not know this but that game advertising and sales strategies reinforce this incorrect notion because it is in the interests of the game sellers for people to not fully understand their rights and obligations in these matters.  It might hurt sales if people start considering what they're getting for what they're paying and they can always sue if this ignorance ever actually hurts them at any point.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0790, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can see<mask> that would be the case. Personally, I find those types of people to be annoying and generally self-centered. They often think that<mask> of their intelligence, they are somehow "above" having social graces. Maybe<mask> I could get past the annoyingness, I could appreciate the satire of Sheldon's character<mask> he makes fun of those people,<mask> I just can't do it. He is far too irritating on screen for me to find humor in his character. [NEWLINE] [NEWLINE] <mask> humor is subjective, and<mask> you find him funny, more power to you. </s>
Label encoding: <s>I can see where that would be the case. Personally, I find those types of people to be annoying and generally self-centered. They often think that because of their intelligence, they are somehow "above" having social graces. Maybe if I could get past the annoyingness, I could appreciate the satire of Sheldon's character as he makes fun of those people, but I just can't do it. He is far too irritating on screen for me to find humor in his character. [NEWLINE] [NEWLINE] But humor is subjective, and if you find him funny, more power to you. </s>
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Take homosexuality for example.<mask> everyone chose to have sex with only same-sex partners, guess<mask> long human society would last... Exactly one generation. There's nothing immoral about not reproducing...<mask> you can<mask><mask> its selfish and that's<mask> society frowns upon it. [ENDQ] [NEWLINE] [STARTQ] Polyamory<mask><mask><mask><mask> is frowned upon<mask> it creates an unstable family unit and produces children that tend to have limited access to one of their parents. [ENDQ] [NEWLINE] In this case, shouldn't voluntary celibacy and divorce<mask> be illegal? [NEWLINE] [NEWLINE] <mask> not,<mask> not?</s>
Label encoding: <s> [STARTQ] Take homosexuality for example. If everyone chose to have sex with only same-sex partners, guess how long human society would last... Exactly one generation. There's nothing immoral about not reproducing... but you can argue that its selfish and that's why society frowns upon it. [ENDQ] [NEWLINE] [STARTQ] Polyamory on the other hand is frowned upon because it creates an unstable family unit and produces children that tend to have limited access to one of their parents. [ENDQ] [NEWLINE] In this case, shouldn't voluntary celibacy and divorce also be illegal? [NEWLINE] [NEWLINE] If not, why not?</s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1 marginal cost of your child is 0: the environment is ruined<mask> of rising 3rd world wealth [NEWLINE] [NEWLINE] 2. declining birth rates create real harms for first world<mask> we force fewer productive people to provide for the lives of more and more old people. [NEWLINE] [NEWLINE] 3. "existence causes suffering" is a fine argument (buddhism)<mask> for it to hold it presupposes said suffering is the greatest net moral thing about life creation. creating a new human being is a good in itself<mask><mask> existence increases overall suffering that is balanced by the fact existence itself is good. </s>
Label encoding: <s>1 marginal cost of your child is 0: the environment is ruined because of rising 3rd world wealth [NEWLINE] [NEWLINE] 2. declining birth rates create real harms for first world as we force fewer productive people to provide for the lives of more and more old people. [NEWLINE] [NEWLINE] 3. "existence causes suffering" is a fine argument (buddhism) but for it to hold it presupposes said suffering is the greatest net moral thing about life creation. creating a new human being is a good in itself because while existence increases overall suffering that is balanced by the fact existence itself is good. </s>
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0998, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Don't tell me<mask> my argument is<mask> you misunderstand it. [NEWLINE] [NEWLINE] [STARTQ] You are saying that I should intentionally not tell my children the reasons I most strongly feel something is wrong,<mask> instead I should only tell them reasons that have been...<mask>?... pre-approved by some government committee on<mask> valid a-religious reasoning for all common childhood questions are? [ENDQ] [NEWLINE] You never answered my very first question. Are you ONLY not racist<mask> of your religion? Do you not see ANY valid reason to raise your child to not be racist other than religion? </s><pad><pad><pad><pad>
Label encoding: <s>Don't tell me what my argument is when you misunderstand it. [NEWLINE] [NEWLINE] [STARTQ] You are saying that I should intentionally not tell my children the reasons I most strongly feel something is wrong, but instead I should only tell them reasons that have been... what?... pre-approved by some government committee on what valid a-religious reasoning for all common childhood questions are? [ENDQ] [NEWLINE] You never answered my very first question. Are you ONLY not racist because of your religion? Do you not see ANY valid reason to raise your child to not be racist other than religion? </s><pad><pad><pad><pad>
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are making them conscious of the choice before the memory is erased,<mask> it really makes no difference whether or not the memory will be erased, they will still potentially have a negative feeling<mask> making the choice. [NEWLINE] [NEWLINE] That's really the same situation<mask> the guy in front of the bus,<mask> death and memory loss are equivalent<mask> the prospect of having a good feeling after is not possible. [NEWLINE] [NEWLINE] <mask> you know you won't have memory after (death, or loss of memory of the event) then the only logical motivation would be to avoid the current negative feeling. [NEWLINE] </s>
Label encoding: <s>You are making them conscious of the choice before the memory is erased, so it really makes no difference whether or not the memory will be erased, they will still potentially have a negative feeling when making the choice. [NEWLINE] [NEWLINE] That's really the same situation as the guy in front of the bus, as death and memory loss are equivalent as the prospect of having a good feeling after is not possible. [NEWLINE] [NEWLINE] If you know you won't have memory after (death, or loss of memory of the event) then the only logical motivation would be to avoid the current negative feeling. [NEWLINE] </s>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0900, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can't believe I haven't seen it like this. Such a simple thing to think about in all honesty. You get the &amp;#8710; my friend. [NEWLINE] [NEWLINE] I guess that<mask> I can see it now is that cheating isn't justified<mask> the student,<mask> they try hard, doesn't have the same level of skill/intellect the other student does,<mask> they do not deserve<mask> the other student gets. The way I've been seeing it is that we all deserve the same thing,<mask> it's quite clear now that we don't.</s>
Label encoding: <s>Can't believe I haven't seen it like this. Such a simple thing to think about in all honesty. You get the &amp;#8710; my friend. [NEWLINE] [NEWLINE] I guess that how I can see it now is that cheating isn't justified because the student, although they try hard, doesn't have the same level of skill/intellect the other student does, so they do not deserve what the other student gets. The way I've been seeing it is that we all deserve the same thing, but it's quite clear now that we don't.</s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Perhaps not<mask><mask> people are accused of robbery or murder there isn't the same "maybe he was falsely accused" response like there is with rape. Its horrible<mask> people are falsely accused<mask> it shouldn't overshadow the real cases. [NEWLINE] [NEWLINE] Unlike other crimes most rape does go unreported partially<mask> the stigma of people questioning<mask> it was legitimate or not. The extra benefit of the doubt that is given to the accused needs to end.  I believe in innocent until proven guilty<mask><mask> people do make a claim it needs to be taken more seriously than it has been. </s>
Label encoding: <s>Perhaps not but when people are accused of robbery or murder there isn't the same "maybe he was falsely accused" response like there is with rape. Its horrible when people are falsely accused but it shouldn't overshadow the real cases. [NEWLINE] [NEWLINE] Unlike other crimes most rape does go unreported partially because the stigma of people questioning if it was legitimate or not. The extra benefit of the doubt that is given to the accused needs to end.  I believe in innocent until proven guilty but when people do make a claim it needs to be taken more seriously than it has been. </s>
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They most definitely have and will continue to do<mask>. [NEWLINE] [NEWLINE] Guns are not bludgeoning objects.<mask> you're hitting them with your gun instead of shooting them, you've lost 100% of your advantage. [NEWLINE] [NEWLINE] Yes, the purpose is to stop them,<mask> I dont think you understand<mask> it sometimes takes to stop people like this.<mask> they break into your home and threaten your family, you don't take chances. You do your best to get them to leave.<mask> that doesn't work, you take whatever measures necessary to put them down.</s>
Label encoding: <s>They most definitely have and will continue to do so. [NEWLINE] [NEWLINE] Guns are not bludgeoning objects. If you're hitting them with your gun instead of shooting them, you've lost 100% of your advantage. [NEWLINE] [NEWLINE] Yes, the purpose is to stop them, but I dont think you understand what it sometimes takes to stop people like this. When they break into your home and threaten your family, you don't take chances. You do your best to get them to leave. If that doesn't work, you take whatever measures necessary to put them down.</s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that you could easily apply "victim-blaming" responses to pick-pocketing, theft, burglary, etc. and still call it victim blaming. It's just that we don't, possibly<mask> rape is a much more personal, emotional thing.<mask> naturally, there's much more of a backlash<mask> we "blame" the victims of rape scenarios than<mask> we "blame" the victims of burglary, etc. [NEWLINE] [NEWLINE] <mask>,<mask> you say QED at the end of your argument, it makes you sound like a complete tool.</s>
Label encoding: <s>I think that you could easily apply "victim-blaming" responses to pick-pocketing, theft, burglary, etc. and still call it victim blaming. It's just that we don't, possibly because rape is a much more personal, emotional thing. So naturally, there's much more of a backlash when we "blame" the victims of rape scenarios than when we "blame" the victims of burglary, etc. [NEWLINE] [NEWLINE] Also, when you say QED at the end of your argument, it makes you sound like a complete tool.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0832, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0853, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What the Israelis are doing is the same a this: [NEWLINE] [NEWLINE] Man sleeps with your wife. You want to kill him for revenge.<mask> you get in front of his house with 50 strong guys (pretty much the difference of power between Israel and Palestinians), he sees you and rushes into his house<mask> his 24 kids (based on the death ratio explained in this thread) are. You get in, then you kill everyone and wait until some nice guy says on Reddit "The bastard hid himself behind his 24 kids,<mask> any complaint is to be considered mere propaganda"</s>
Label encoding: <s>What the Israelis are doing is the same a this: [NEWLINE] [NEWLINE] Man sleeps with your wife. You want to kill him for revenge. When you get in front of his house with 50 strong guys (pretty much the difference of power between Israel and Palestinians), he sees you and rushes into his house where his 24 kids (based on the death ratio explained in this thread) are. You get in, then you kill everyone and wait until some nice guy says on Reddit "The bastard hid himself behind his 24 kids, so any complaint is to be considered mere propaganda"</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> unfortunately,<mask>'s wrong changes and<mask> do people. [ENDQ] [NEWLINE] This is absolute hogwash.<mask> people "believe" is wrong may change a bit,<mask> 1) mass extermination of a race of people will never NOT be wrong and 2)<mask> actually is wrong never changes no matter<mask> or<mask> you live. Murdering an innocent child has never been "right" and it will never be "right", no matter<mask> someone someday in some country claims it to be right. To say such a thing shows a profound misunderstanding of justice and morality.</s>
Label encoding: <s> [STARTQ] But unfortunately, what's wrong changes and so do people. [ENDQ] [NEWLINE] This is absolute hogwash. What people "believe" is wrong may change a bit, but 1) mass extermination of a race of people will never NOT be wrong and 2) what actually is wrong never changes no matter where or when you live. Murdering an innocent child has never been "right" and it will never be "right", no matter if someone someday in some country claims it to be right. To say such a thing shows a profound misunderstanding of justice and morality.</s>
Loss: tensor(0.1266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This comes up now and then and always makes me go over to /r/atheism to actually see whats going on over there. The same thing always happens. Some guy says "oh atheism is nothing<mask> hatred towards religious people",<mask> just take a look at the page 1 topic is sorta saying something hateful the rest doesn't. [NEWLINE] [NEWLINE] <mask> much can you cherrypick to get a very very bad strawman across? Do you ever go to /r/atheism or do you just listen to the pathetic circlejerk on the rest of reddit?</s>
Label encoding: <s>This comes up now and then and always makes me go over to /r/atheism to actually see whats going on over there. The same thing always happens. Some guy says "oh atheism is nothing but hatred towards religious people", but just take a look at the page 1 topic is sorta saying something hateful the rest doesn't. [NEWLINE] [NEWLINE] How much can you cherrypick to get a very very bad strawman across? Do you ever go to /r/atheism or do you just listen to the pathetic circlejerk on the rest of reddit?</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0993, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you go outside with your friends and every single one of them spark a cigarette, there is always a temptation whether you like it or not. Eventually you'll say "Sure,<mask> not" and that's<mask> you're finished. [ENDQ] [NEWLINE] That's just not true. There has never been temptation for me. I don't think peer pressure can force you to do anything you didn't want to do already, it's just the catalyst you use to rationalize your decision to bum a cigarette from a friend and jump headfirst down the rabbit hole. </s>
Label encoding: <s> [STARTQ] When you go outside with your friends and every single one of them spark a cigarette, there is always a temptation whether you like it or not. Eventually you'll say "Sure, why not" and that's when you're finished. [ENDQ] [NEWLINE] That's just not true. There has never been temptation for me. I don't think peer pressure can force you to do anything you didn't want to do already, it's just the catalyst you use to rationalize your decision to bum a cigarette from a friend and jump headfirst down the rabbit hole. </s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It is not a myth that there are a couple billion too many people on Earth. We can fix the distribution problem with immigration reform. [ENDQ] [NEWLINE] <mask> it is a myth. the entire earths population could fit into Texas and have the same population density that is in New York. Like I posted elsewhere the problem isn't overpopulation<mask> overconsumption. [NEWLINE] [NEWLINE] <mask>,<mask> do you figure we can fix the distribution problem through immigration reform? Serious question I have studied these things in for my degree and I am not sure I have heard this solution. [NEWLINE] </s>
Label encoding: <s> [STARTQ] It is not a myth that there are a couple billion too many people on Earth. We can fix the distribution problem with immigration reform. [ENDQ] [NEWLINE] But it is a myth. the entire earths population could fit into Texas and have the same population density that is in New York. Like I posted elsewhere the problem isn't overpopulation but overconsumption. [NEWLINE] [NEWLINE] Also, how do you figure we can fix the distribution problem through immigration reform? Serious question I have studied these things in for my degree and I am not sure I have heard this solution. [NEWLINE] </s>
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Question regarding US law. Can parents actually be punished for their children's crimes, fine or otherwise? And I don't mean punishment like their children being taken away (<mask> that isn't punishment<mask> child protection) or punishment for neglect (which is punishment for a different crime they themselves did). In Germany for instance parents can't actually be obliged to pay fines for child misdeeds. In theory it becomes a debt that can't be collected until the child is capable of being criminally judged (14<mask><mask>?). Of course most parents rather just pay it immediately instead.</s>
Label encoding: <s>Question regarding US law. Can parents actually be punished for their children's crimes, fine or otherwise? And I don't mean punishment like their children being taken away ( as that isn't punishment but child protection) or punishment for neglect (which is punishment for a different crime they themselves did). In Germany for instance parents can't actually be obliged to pay fines for child misdeeds. In theory it becomes a debt that can't be collected until the child is capable of being criminally judged (14 I think?). Of course most parents rather just pay it immediately instead.</s>
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The religion shouldn't have to change it's view<mask> something is made legal.<mask> their's is an unpopular opinion, they'll just get less believers anyway. [ENDQ] [NEWLINE] The problem occurs<mask> religious belief exists outside the church. Should a Christian that vehemently opposes gay marriage have to provide the services of their business for gay weddings? [NEWLINE] [NEWLINE] One of the reason churches are getting in trouble is<mask> they rent out facilities<mask><mask> they were a business, and operating a business means you can't discriminate. Do we allow discrimination, or do we respect religious belief?</s>
Label encoding: <s> [STARTQ] The religion shouldn't have to change it's view because something is made legal. If their's is an unpopular opinion, they'll just get less believers anyway. [ENDQ] [NEWLINE] The problem occurs when religious belief exists outside the church. Should a Christian that vehemently opposes gay marriage have to provide the services of their business for gay weddings? [NEWLINE] [NEWLINE] One of the reason churches are getting in trouble is because they rent out facilities as if they were a business, and operating a business means you can't discriminate. Do we allow discrimination, or do we respect religious belief?</s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We got that out of the way before having our wedding. [ENDQ] [NEWLINE] Not generalizable, obviously. Many people remain in a tight financial situation well into middle age, and some people never really get out of their financial rut. [NEWLINE] [NEWLINE] [STARTQ] <mask> you've got the money and can afford it,<mask> not throw one? [ENDQ] [NEWLINE] Some people may never have the money. There's no problem with you throwing a party,<mask> everyone interprets your party<mask> a response to a social obligation - which they then project onto others who can't afford it.</s>
Label encoding: <s> [STARTQ] We got that out of the way before having our wedding. [ENDQ] [NEWLINE] Not generalizable, obviously. Many people remain in a tight financial situation well into middle age, and some people never really get out of their financial rut. [NEWLINE] [NEWLINE] [STARTQ] If you've got the money and can afford it, why not throw one? [ENDQ] [NEWLINE] Some people may never have the money. There's no problem with you throwing a party, but everyone interprets your party as a response to a social obligation - which they then project onto others who can't afford it.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If a god such<mask> Christians worship exists, it is all-powerful and all-knowing. Its definition of good would be arbitrary: it's its universe and we're just living in it. We would be in no position to say<mask> is worthy of praise and worship and<mask> is not. We would simply be its subjects. The question of whether God is worthy or not worthy would be an irrelevancy, an open and shut case. [NEWLINE] [NEWLINE] tl;dr: God,<mask> it exists, is *by definition* worthy of our praise.</s><pad>
Label encoding: <s>If a god such as Christians worship exists, it is all-powerful and all-knowing. Its definition of good would be arbitrary: it's its universe and we're just living in it. We would be in no position to say what is worthy of praise and worship and what is not. We would simply be its subjects. The question of whether God is worthy or not worthy would be an irrelevancy, an open and shut case. [NEWLINE] [NEWLINE] tl;dr: God, if it exists, is *by definition* worthy of our praise.</s><pad>
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>At that weight you were burning [NEWLINE] [NEWLINE] [STARTQ] 2816 kcal [ENDQ] [NEWLINE] Which means (using matts numbers) you expect poor people to find ~ an extra 31.04 dollars a day(18.16 *2 - 1.76 *3) and thats at the maximum end of the scale, 500 cal is the recommended<mask> add another 9 dollars a day. [NEWLINE] [NEWLINE] Telling the poor to "just eat healthy" is terrible advice;<mask> the want to lose weight they would need to learn<mask> to do it on unhealthy food, which is harder.</s>
Label encoding: <s>At that weight you were burning [NEWLINE] [NEWLINE] [STARTQ] 2816 kcal [ENDQ] [NEWLINE] Which means (using matts numbers) you expect poor people to find ~ an extra 31.04 dollars a day(18.16 *2 - 1.76 *3) and thats at the maximum end of the scale, 500 cal is the recommended so add another 9 dollars a day. [NEWLINE] [NEWLINE] Telling the poor to "just eat healthy" is terrible advice; if the want to lose weight they would need to learn how to do it on unhealthy food, which is harder.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Yes, you can get much larger search spaces with random characters,<mask> it does become harder to remember and you're more prone to making a plaintext note somewhere. [ENDQ] [NEWLINE] Open Source Password managers. I can say that it is very irksome<mask> a website limits my passwords to 8 characters<mask> my passwords are generated for me with high entropy and each website has it's own password. [NEWLINE] [NEWLINE] Of course, the password manager then itself becomes a weak point,<mask> with strong encryption and a high entropy password, one need not be overly worried. </s>
Label encoding: <s> [STARTQ] Yes, you can get much larger search spaces with random characters, but it does become harder to remember and you're more prone to making a plaintext note somewhere. [ENDQ] [NEWLINE] Open Source Password managers. I can say that it is very irksome when a website limits my passwords to 8 characters when my passwords are generated for me with high entropy and each website has it's own password. [NEWLINE] [NEWLINE] Of course, the password manager then itself becomes a weak point, but with strong encryption and a high entropy password, one need not be overly worried. </s>
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Note that the analogy is only even slightly relevant<mask> the fetus is a person, which I have argued against in other comments. [NEWLINE] [NEWLINE] [STARTQ] A way better analogy would be<mask> you intentionally hit someone else with your car [ENDQ] [NEWLINE] This would be an analog for intentionally getting pregnant and aborting, not accidentally getting pregnant<mask> taking all precautions. [NEWLINE] [NEWLINE] [STARTQ] and now they want to share some of your organs for 9 months, after which you will get them all to yourself. [ENDQ] [NEWLINE] Which the law still would not force you to, even in that worst case.</s>
Label encoding: <s>Note that the analogy is only even slightly relevant if the fetus is a person, which I have argued against in other comments. [NEWLINE] [NEWLINE] [STARTQ] A way better analogy would be if you intentionally hit someone else with your car [ENDQ] [NEWLINE] This would be an analog for intentionally getting pregnant and aborting, not accidentally getting pregnant despite taking all precautions. [NEWLINE] [NEWLINE] [STARTQ] and now they want to share some of your organs for 9 months, after which you will get them all to yourself. [ENDQ] [NEWLINE] Which the law still would not force you to, even in that worst case.</s>
Loss: tensor(0.0858, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's some recent research that has shown that the 10,000 hours thing is not very reliable. Some people master a skill in significantly less time, and others in a lot longer time. Passion, raw ability and luck can play a major role in development of a skill. Sometimes you just need to try something and see<mask> you're good at it. [NEWLINE] [NEWLINE] At 24, unless you've got a kid and a mortgage, you've got more than enough time to develop a new skill, particularly in an area<mask> you've got passion or talent.</s>
Label encoding: <s>There's some recent research that has shown that the 10,000 hours thing is not very reliable. Some people master a skill in significantly less time, and others in a lot longer time. Passion, raw ability and luck can play a major role in development of a skill. Sometimes you just need to try something and see if you're good at it. [NEWLINE] [NEWLINE] At 24, unless you've got a kid and a mortgage, you've got more than enough time to develop a new skill, particularly in an area where you've got passion or talent.</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, they don't get to choose whether to give it for free,<mask> you do? [NEWLINE] [NEWLINE] Who said "huge profit". <mask> about any profit? <mask> you are an author or a studio-type musician you won't be able to do it full time<mask> you don't have a way to make money off of it. [NEWLINE] [NEWLINE] You're right, there will not be "no" incentive,<mask> there's going to be a lot more artists working office jobs to pay the rent<mask> you refuse to pay for<mask> you consume.</s>
Label encoding: <s>So, they don't get to choose whether to give it for free, but you do? [NEWLINE] [NEWLINE] Who said "huge profit".  How about any profit?  If you are an author or a studio-type musician you won't be able to do it full time if you don't have a way to make money off of it. [NEWLINE] [NEWLINE] You're right, there will not be "no" incentive, but there's going to be a lot more artists working office jobs to pay the rent because you refuse to pay for what you consume.</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Very true. A big part of rapping is personality and<mask> you can evoke emotion through the way you speak. For example, listening to a lot of Waka Flockas harsh, blunt, yelling/screaming delivery evokes power, rage, and scorn. Kevin Gates raps sometimes it sounds reminiscent of sobbing, and kind of evokes the feeling of this messed up, emotional, bitter individual. [NEWLINE] [NEWLINE] There is absolutely great importance to your voice and<mask> you deliver the lyrics, whether they're intelligent and deep or not. </s>
Label encoding: <s>Very true. A big part of rapping is personality and how you can evoke emotion through the way you speak. For example, listening to a lot of Waka Flockas harsh, blunt, yelling/screaming delivery evokes power, rage, and scorn. Kevin Gates raps sometimes it sounds reminiscent of sobbing, and kind of evokes the feeling of this messed up, emotional, bitter individual. [NEWLINE] [NEWLINE] There is absolutely great importance to your voice and how you deliver the lyrics, whether they're intelligent and deep or not. </s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0973, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyone would need to get rid of them in order for there to not be unbridled chaos.  Mutually assured destruction ensures that one superpower won't use their nukes<mask> of the inevitable nuclear backlash from the other superpowers.  That's a recipe for a completely fucked planet. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>,<mask> everyone got rid of their nukes, it's likely that more wars would be started between superpowers.  Without mutually assured destruction to curb conflicts, there's less risk to the entire planet<mask> a war breaks out.</s>
Label encoding: <s>Everyone would need to get rid of them in order for there to not be unbridled chaos.  Mutually assured destruction ensures that one superpower won't use their nukes because of the inevitable nuclear backlash from the other superpowers.  That's a recipe for a completely fucked planet. [NEWLINE] [NEWLINE] On the other hand, if everyone got rid of their nukes, it's likely that more wars would be started between superpowers.  Without mutually assured destruction to curb conflicts, there's less risk to the entire planet when a war breaks out.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0998, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is<mask> I understand you are saying: [NEWLINE] [NEWLINE] In the flirting process, males use humour more than females. [NEWLINE] <mask>, there are few,<mask> any, successful female comedians,<mask> they don't develop their skills. [NEWLINE] [NEWLINE] This presupposes that a) the successful male comedians honed their skills<mask> flirting and b) there are not enough exceptional females that use humour to flirt to create a successful female comedian. [NEWLINE] [NEWLINE] <mask><mask> both presuppositions are unlikely. There are better reasons to explain the lack of successful female comedians.</s>
Label encoding: <s>This is what I understand you are saying: [NEWLINE] [NEWLINE] In the flirting process, males use humour more than females. [NEWLINE] Therefore, there are few, if any, successful female comedians, because they don't develop their skills. [NEWLINE] [NEWLINE] This presupposes that a) the successful male comedians honed their skills while flirting and b) there are not enough exceptional females that use humour to flirt to create a successful female comedian. [NEWLINE] [NEWLINE] I think both presuppositions are unlikely. There are better reasons to explain the lack of successful female comedians.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Climate change is happening. [NEWLINE] [NEWLINE] It will not eradicate human civilization.  Rising sea levels may make flooding worse in coastal cities.  Weather changes may make winter storms worse.  We might pay more for food<mask> droughts worsen. [NEWLINE] [NEWLINE] <mask><mask> you're able to log into the Internet and read those reports, you very likely are part of a country that has resources to deal with it. [NEWLINE] [NEWLINE] I believe you'll be fine.  Just don't become an impoverished dweller of a tiny south Pacific Island.   </s>
Label encoding: <s>Climate change is happening. [NEWLINE] [NEWLINE] It will not eradicate human civilization.  Rising sea levels may make flooding worse in coastal cities.  Weather changes may make winter storms worse.  We might pay more for food if droughts worsen. [NEWLINE] [NEWLINE] However if you're able to log into the Internet and read those reports, you very likely are part of a country that has resources to deal with it. [NEWLINE] [NEWLINE] I believe you'll be fine.  Just don't become an impoverished dweller of a tiny south Pacific Island.   </s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you don't want to run the risk of having to work late then: [ENDQ] [STARTQ] Don't work in a restaurant. [ENDQ] [STARTQ] Don't work the evening shift. [ENDQ] [STARTQ] Work out your schedule requirements with your manager ahead of time. [ENDQ] [STARTQ] Find someone to cover for you. [ENDQ] [NEWLINE] <mask>, to summarize, "Next time, don't be poor." [NEWLINE] [NEWLINE] Not everyone has a huge array of choices in<mask> jobs and hours they work. Many people work the jobs they can and the hours they can to stay afloat. </s>
Label encoding: <s> [STARTQ] If you don't want to run the risk of having to work late then: [ENDQ] [STARTQ] Don't work in a restaurant. [ENDQ] [STARTQ] Don't work the evening shift. [ENDQ] [STARTQ] Work out your schedule requirements with your manager ahead of time. [ENDQ] [STARTQ] Find someone to cover for you. [ENDQ] [NEWLINE] So, to summarize, "Next time, don't be poor." [NEWLINE] [NEWLINE] Not everyone has a huge array of choices in what jobs and hours they work. Many people work the jobs they can and the hours they can to stay afloat. </s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0807, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1316, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Absolute truth is pretty hard to grasp. It is entirely possible that you could both be right, or, much more likely, both be wrong and get nowhere by discussing it. [NEWLINE] [NEWLINE] You<mask> have to ask: is it more important to be *right* or to be *happy*?<mask> your friend is deeply misinformed about some political issue and planning to vote in the upcoming election, it might be worth arguing about.<mask> your friend is wrong about something that doesn't directly negatively impact your life, just let it go. </s>
Label encoding: <s>Absolute truth is pretty hard to grasp. It is entirely possible that you could both be right, or, much more likely, both be wrong and get nowhere by discussing it. [NEWLINE] [NEWLINE] You also have to ask: is it more important to be *right* or to be *happy*? If your friend is deeply misinformed about some political issue and planning to vote in the upcoming election, it might be worth arguing about. If your friend is wrong about something that doesn't directly negatively impact your life, just let it go. </s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here is a source that details various causes for pay disparity. [NEWLINE] [NEWLINE] [URL] /?a=Files.Serve&amp;File_id=9118a9ef-0771-4777-9c1f-8232fe70a45c [NEWLINE] [NEWLINE] One of the things that stuck out for me on page 85: <mask> an employer can tell whether a potential applicant is male or female compared to a control group<mask> they could not tell, women were less likely to get the job than they would in the control group.</s>
Label encoding: <s>Here is a source that details various causes for pay disparity. [NEWLINE] [NEWLINE] [URL] /?a=Files.Serve&amp;File_id=9118a9ef-0771-4777-9c1f-8232fe70a45c [NEWLINE] [NEWLINE] One of the things that stuck out for me on page 85:  When an employer can tell whether a potential applicant is male or female compared to a control group where they could not tell, women were less likely to get the job than they would in the control group.</s>
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I do think being white has made life a little easier for me than<mask> I wasn't. I'm not going to pretend otherwise.<mask><mask><mask> that life is 'easier' for me<mask> a chick. [ENDQ] [NEWLINE] Very first sentence of her post.  The original claim up for debate is "white women lead easier lives than the rest of society."  Commenter is agreeing on the white part,<mask> disagreeing on the woman part,<mask> the rest of her post being about things *women* in general may face. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I do think being white has made life a little easier for me than if I wasn't. I'm not going to pretend otherwise. But I disagree that life is 'easier' for me as a chick. [ENDQ] [NEWLINE] Very first sentence of her post.  The original claim up for debate is "white women lead easier lives than the rest of society."  Commenter is agreeing on the white part, but disagreeing on the woman part, hence the rest of her post being about things *women* in general may face. [NEWLINE] </s>
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0880, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0941, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [NEWLINE] [STARTQ] Of course not. It's not that people who have been raised in poverty have no motivation, it's that many of them had no parents/education/figures in their life to teach them the things needed to become successful in society. [ENDQ] [NEWLINE] To this I will add<mask><mask> the skills taught to the children not only disillusioned them about the world they were growing into<mask><mask> were out dated from the moment those lessons about surviving were taught, given<mask> rapidly the world changes. [NEWLINE] [NEWLINE] Food for thought. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [NEWLINE] [NEWLINE] [STARTQ] Of course not. It's not that people who have been raised in poverty have no motivation, it's that many of them had no parents/education/figures in their life to teach them the things needed to become successful in society. [ENDQ] [NEWLINE] To this I will add what if the skills taught to the children not only disillusioned them about the world they were growing into but also were out dated from the moment those lessons about surviving were taught, given how rapidly the world changes. [NEWLINE] [NEWLINE] Food for thought. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0959, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think for many people the thing that makes Scientology more absurd is the fact that you have to pay absurd amounts of money for access to information. Most (<mask> not all) other mainstream religions provide access to texts for free. The idea of requiring people to pay absurd amounts of money for<mask> in their mind would be life changing(potentially saving) information would be considered immoral by most peoples standards. [NEWLINE] [NEWLINE] Basically the whole thing seems like a giant scam to most people and thats<mask> makes it more absurd in the eyes of the public.</s>
Label encoding: <s>I think for many people the thing that makes Scientology more absurd is the fact that you have to pay absurd amounts of money for access to information. Most ( if not all) other mainstream religions provide access to texts for free. The idea of requiring people to pay absurd amounts of money for what in their mind would be life changing(potentially saving) information would be considered immoral by most peoples standards. [NEWLINE] [NEWLINE] Basically the whole thing seems like a giant scam to most people and thats what makes it more absurd in the eyes of the public.</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>to repeat myself: [NEWLINE] [NEWLINE] [STARTQ] <mask> that's an argument for another day. [ENDQ] [NEWLINE] [NEWLINE] you seem to want to have lots of political arguments and i'm fine with that<mask> that's not the CMV topic. The topic is the right is just bigoted (and Martin-Wilson stuff is an example of this). you have<mask> to defend the assertion that the idea to focus on the wilson stuff is racist (and i know Conor Freisdorf who you cited specifically to support this claim would quickly reject that claim<mask> garbage)</s>
Label encoding: <s>to repeat myself: [NEWLINE] [NEWLINE] [STARTQ] but that's an argument for another day. [ENDQ] [NEWLINE] [NEWLINE] you seem to want to have lots of political arguments and i'm fine with that but that's not the CMV topic. The topic is the right is just bigoted (and Martin-Wilson stuff is an example of this). you have yet to defend the assertion that the idea to focus on the wilson stuff is racist (and i know Conor Freisdorf who you cited specifically to support this claim would quickly reject that claim as garbage)</s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] <mask> I still don't think the movies were good adaptations of the novel, you make a good point about the budget constraints.<mask> there may be a way that they could have made it work (using practical effects instead of relying on CGI like Game of Thrones does for dragons and magic), I don't really have any special knowledge about the financial aspects of special effects to say you're wrong. [NEWLINE] [NEWLINE] <mask> you have changed my view on the effectiveness of a TV series in conveying the magic of the Harry Potter world.</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] While I still don't think the movies were good adaptations of the novel, you make a good point about the budget constraints. While there may be a way that they could have made it work (using practical effects instead of relying on CGI like Game of Thrones does for dragons and magic), I don't really have any special knowledge about the financial aspects of special effects to say you're wrong. [NEWLINE] [NEWLINE] So you have changed my view on the effectiveness of a TV series in conveying the magic of the Harry Potter world.</s>
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> this wouldn't make political speech really expensive, it makes hidden political speech more expensive [ENDQ] [NEWLINE] There are already limits on<mask> much you can give directly to candidates (before and after Citizens United).  You can only give 50k every 4 years.  It doesn't matter<mask> you are anonymous or not, you aren't allowed to give more than that. [NEWLINE] [NEWLINE] Spending money to support a candidate without giving directly to that candidate is<mask> changed under Citizens United.  It doesn't matter whether it's hidden or not.</s>
Label encoding: <s> [STARTQ] But this wouldn't make political speech really expensive, it makes hidden political speech more expensive [ENDQ] [NEWLINE] There are already limits on how much you can give directly to candidates (before and after Citizens United).  You can only give 50k every 4 years.  It doesn't matter if you are anonymous or not, you aren't allowed to give more than that. [NEWLINE] [NEWLINE] Spending money to support a candidate without giving directly to that candidate is what changed under Citizens United.  It doesn't matter whether it's hidden or not.</s>
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0987, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0934, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Those are all good points on the shortcomings of our mental health systems.  Treatments still need to be developed (I don't know anything about treating pedophilia specifically, I'm not a therapist). [NEWLINE] [NEWLINE] It is really terrible that our society still doesn't understand and sympathize with mental illness.  I do think we are getting better,<mask> even things like clinical depression are met with responses like "just get better".  We need to make sure that it's clear that there is no judgement to be found in seeking help.</s>
Label encoding: <s>Those are all good points on the shortcomings of our mental health systems.  Treatments still need to be developed (I don't know anything about treating pedophilia specifically, I'm not a therapist). [NEWLINE] [NEWLINE] It is really terrible that our society still doesn't understand and sympathize with mental illness.  I do think we are getting better, but even things like clinical depression are met with responses like "just get better".  We need to make sure that it's clear that there is no judgement to be found in seeking help.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to disagree about the smell/dirtiness.  It is pretty awful. <mask><mask> some people, mostly city folk, just become accustomed to the smell of garbage from living among it,<mask> it is everywhere.  I'm a connecticutter,<mask> I am close enough to visit often enough.  First thing me and my wife do<mask> we get home is shower to get that awful stench off us.  We have to wash everything, even our jackets after a single day<mask> it permeates everything.</s>
Label encoding: <s>I have to disagree about the smell/dirtiness.  It is pretty awful.  I think some people, mostly city folk, just become accustomed to the smell of garbage from living among it, but it is everywhere.  I'm a connecticutter, so I am close enough to visit often enough.  First thing me and my wife do when we get home is shower to get that awful stench off us.  We have to wash everything, even our jackets after a single day because it permeates everything.</s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0988, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One of the reasons I like it more than other types of welfare is that it brings back some more personal liberty and responsibility back into the mix. [NEWLINE] [NEWLINE] <mask> people get to make their own decisions about<mask> they spend that money, they will (hopefully) learn to prioritize<mask> is important. There won't be excuses for not being able to pay for insurance or food or shelter. They are now responsible for their own well-being. Hopefully, they can spend that money more efficiently than the government programs with tons of overhead.</s>
Label encoding: <s>One of the reasons I like it more than other types of welfare is that it brings back some more personal liberty and responsibility back into the mix. [NEWLINE] [NEWLINE] If people get to make their own decisions about how they spend that money, they will (hopefully) learn to prioritize what is important. There won't be excuses for not being able to pay for insurance or food or shelter. They are now responsible for their own well-being. Hopefully, they can spend that money more efficiently than the government programs with tons of overhead.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Personally, I prefer a translation which does<mask> little of this<mask> possible.<mask> I watch a show set in Japan, about Japanese people, I don't always want the cultural differences "translated away", or converted into something the translator considers roughly equivalent in Western culture.<mask> I don't understand some reference, maybe that's<mask> it relates to something which is actually interesting and unique about the other culture - I'd prefer to realise that this new cultural thing exists than pretend that it doesn't. [ENDQ] [NEWLINE] exactly!</s>
Label encoding: <s> [STARTQ] Personally, I prefer a translation which does as little of this as possible. If I watch a show set in Japan, about Japanese people, I don't always want the cultural differences "translated away", or converted into something the translator considers roughly equivalent in Western culture. If I don't understand some reference, maybe that's because it relates to something which is actually interesting and unique about the other culture - I'd prefer to realise that this new cultural thing exists than pretend that it doesn't. [ENDQ] [NEWLINE] exactly!</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, the police officers are using archaic statistical modelling to justify inane policies that do not deal with the realities of modern day domestic abuse. They *believe* that domestic violence against men isn't real<mask> it doesn't match up to their pre-existing beliefs and notions. These are the same people who believe the legalization of drugs would increase crime and drug use, and brush away evidence to the contrary through the shroud of "experience,"<mask><mask> their experience means nothing<mask> it's based around a policy farce.</s>
Label encoding: <s>Yes, the police officers are using archaic statistical modelling to justify inane policies that do not deal with the realities of modern day domestic abuse. They *believe* that domestic violence against men isn't real because it doesn't match up to their pre-existing beliefs and notions. These are the same people who believe the legalization of drugs would increase crime and drug use, and brush away evidence to the contrary through the shroud of "experience," even though their experience means nothing if it's based around a policy farce.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Any system is going to be gameable. [NEWLINE] [NEWLINE] <mask> you make the fines a percent of their income then people without jobs will cross the streets wherever they want. [NEWLINE] [NEWLINE] <mask><mask> fines are just supposed to be an annoyance that reminds people to act in a reasonable manner. [NEWLINE] [NEWLINE] <mask> we punish people more severely we send them to jail for standard lengths of time (at least that's<mask> is supposed to happen). We don't send rich or poor to jail for differing amounts of time (at least we shouldn't).</s>
Label encoding: <s>Any system is going to be gameable. [NEWLINE] [NEWLINE] If you make the fines a percent of their income then people without jobs will cross the streets wherever they want. [NEWLINE] [NEWLINE] I think fines are just supposed to be an annoyance that reminds people to act in a reasonable manner. [NEWLINE] [NEWLINE] When we punish people more severely we send them to jail for standard lengths of time (at least that's what is supposed to happen). We don't send rich or poor to jail for differing amounts of time (at least we shouldn't).</s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who contributes more to society, a heterosexual couple, both whom work at fast food their whole lives and who have a kid - or a homosexual couple, both of whom are doctors and donate millions to improve public education and who are childless? [NEWLINE] [NEWLINE] I have no idea<mask> you would think that just the act of literally having a child is the only way to "contribute" to the future.  You can improve the lives of people today and in the future without having children, this is plainly obvious.  </s>
Label encoding: <s>Who contributes more to society, a heterosexual couple, both whom work at fast food their whole lives and who have a kid - or a homosexual couple, both of whom are doctors and donate millions to improve public education and who are childless? [NEWLINE] [NEWLINE] I have no idea why you would think that just the act of literally having a child is the only way to "contribute" to the future.  You can improve the lives of people today and in the future without having children, this is plainly obvious.  </s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> the problem is that the app developers concentrate on consuming APIs,<mask> don't expose any new ones. [ENDQ] [NEWLINE] The API ecosystem is exploding at a crazy rate right now, and hackathons<mask> people mash up a half-dozen different APIs to create something totally new are a regular thing. [NEWLINE] [NEWLINE] Seriously,<mask> this is interesting to you, follow ProgrammableWeb's feed or something like it for a<mask>.  It's not unusual to see 20-30 different new API stories in a day.</s><pad>
Label encoding: <s> [STARTQ] I think the problem is that the app developers concentrate on consuming APIs, but don't expose any new ones. [ENDQ] [NEWLINE] The API ecosystem is exploding at a crazy rate right now, and hackathons where people mash up a half-dozen different APIs to create something totally new are a regular thing. [NEWLINE] [NEWLINE] Seriously, if this is interesting to you, follow ProgrammableWeb's feed or something like it for a while.  It's not unusual to see 20-30 different new API stories in a day.</s><pad>
Loss: tensor(0.1008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1390, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think it's ok people lie and do things that aren't news. I read your whole post. It's not obvious to me that this is bad for people, I don't think you've proved that point. You<mask> haven't addressed my concerns about who gets to decide and at<mask> point they can take away the news title and<mask> they enforce that. Do you not have concerns that this could creep into the grounds of censorship? Who is deciding, the government? The FCC? Who?<mask>?</s>
Label encoding: <s>I don't think it's ok people lie and do things that aren't news. I read your whole post. It's not obvious to me that this is bad for people, I don't think you've proved that point. You also haven't addressed my concerns about who gets to decide and at what point they can take away the news title and how they enforce that. Do you not have concerns that this could creep into the grounds of censorship? Who is deciding, the government? The FCC? Who? How?</s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not saying it's the same,<mask> a man saying "I'm not paying" isn't necessarily easy. He still has to live with the fact that there's a child out there, who he created, who's growing up without a father and without proper financial assistance. [NEWLINE] [NEWLINE] I know, it's not at all the same<mask> the emotional weight of having an abortion,<mask> it's not *easy.* Not for everyone. (<mask> then, for some women having an abortion is emotionally easy)</s><pad><pad><pad>
Label encoding: <s>I'm not saying it's the same, but a man saying "I'm not paying" isn't necessarily easy. He still has to live with the fact that there's a child out there, who he created, who's growing up without a father and without proper financial assistance. [NEWLINE] [NEWLINE] I know, it's not at all the same as the emotional weight of having an abortion, but it's not *easy.* Not for everyone. ( But then, for some women having an abortion is emotionally easy)</s><pad><pad><pad>
Loss: tensor(0.2155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0817, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0990, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's the opposite of<mask> we are talking about<mask>. [NEWLINE] [NEWLINE] Churches shouldn't be involved in marriage AT ALL.  Marriage is a state issue. [NEWLINE] [NEWLINE] Churches can make up some other term "Holy Matrimony" is a fine example.  Or"Boo-boo kitty kiss kiss".  That would be fine too. [NEWLINE] [NEWLINE] <mask> your church wants to a "No gay boo-boo-kitty-kiss-kiss" policy, you are welcome to it.</s>
Label encoding: <s>That's the opposite of what we are talking about though. [NEWLINE] [NEWLINE] Churches shouldn't be involved in marriage AT ALL.  Marriage is a state issue. [NEWLINE] [NEWLINE] Churches can make up some other term "Holy Matrimony" is a fine example.  Or"Boo-boo kitty kiss kiss".  That would be fine too. [NEWLINE] [NEWLINE] If your church wants to a "No gay boo-boo-kitty-kiss-kiss" policy, you are welcome to it.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0886, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In addition to the arguments others have presented, I'll show a paper staying that female victims of rape are 4x<mask> likely to retort seriously considering suicide compared to the average female. Similar numbers are purebred for rates of ptsd and severe depression and substance abuse.<mask><mask> these factors combined with the other arguments presented in the thread showing that using the term survivor doesn't really take anything away from others and that the weird itself has a less restrictive meaning than your OP suggests. [NEWLINE] [NEWLINE] [URL].shtml</s>
Label encoding: <s>In addition to the arguments others have presented, I'll show a paper staying that female victims of rape are 4x as likely to retort seriously considering suicide compared to the average female. Similar numbers are purebred for rates of ptsd and severe depression and substance abuse. I think these factors combined with the other arguments presented in the thread showing that using the term survivor doesn't really take anything away from others and that the weird itself has a less restrictive meaning than your OP suggests. [NEWLINE] [NEWLINE] [URL].shtml</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not true: [URL] / [NEWLINE] [NEWLINE] [STARTQ] “The convenience store<mask> the robbery took place was boarded up,<mask> open for business on Friday. A store manager, who declined to give his name, said he fears for his life and pleaded with reporters not to suggest that he called police. [ENDQ] [NEWLINE] [STARTQ] ““It’s very dangerous,” he said. “They kill us<mask> they think we are responsible. People don’t understand that.”” [ENDQ] </s>
Label encoding: <s>Not true: [URL] / [NEWLINE] [NEWLINE] [STARTQ] “The convenience store where the robbery took place was boarded up, but open for business on Friday. A store manager, who declined to give his name, said he fears for his life and pleaded with reporters not to suggest that he called police. [ENDQ] [NEWLINE] [STARTQ] ““It’s very dangerous,” he said. “They kill us if they think we are responsible. People don’t understand that.”” [ENDQ] </s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0824, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0788, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> there were tablet computers before the iPad. [ENDQ] [NEWLINE] Yep, they'd been around for decades and Apple first tried making them in the late 80s. [NEWLINE] [NEWLINE] Difference is the iPad got the formula right and created the huge market that we see today. Given<mask> much Microsoft and others tried (and failed) for years to do the same, that's very impressive. Apple basically showed that usability was more important than a gamut of features that the average person wouldn't use in the first place.</s>
Label encoding: <s> [STARTQ] But there were tablet computers before the iPad. [ENDQ] [NEWLINE] Yep, they'd been around for decades and Apple first tried making them in the late 80s. [NEWLINE] [NEWLINE] Difference is the iPad got the formula right and created the huge market that we see today. Given how much Microsoft and others tried (and failed) for years to do the same, that's very impressive. Apple basically showed that usability was more important than a gamut of features that the average person wouldn't use in the first place.</s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] that reason isn't self-censorship [ENDQ] [NEWLINE] Absolutely, and I had no intention of even implying that. I do think,<mask> [NEWLINE] [NEWLINE] [STARTQ] On the flip side, saying the slur instead of "n-word" isn't adding anything to the conversation. [ENDQ] [NEWLINE] <mask><mask> it really comes down to this point here. I guess to many, it would only serve to make the conversation more uncomfortable. Maybe by not using the slur, more people feel better about joining in the discussion. Thanks!</s><pad>
Label encoding: <s> [STARTQ] that reason isn't self-censorship [ENDQ] [NEWLINE] Absolutely, and I had no intention of even implying that. I do think, however [NEWLINE] [NEWLINE] [STARTQ] On the flip side, saying the slur instead of "n-word" isn't adding anything to the conversation. [ENDQ] [NEWLINE] I think it really comes down to this point here. I guess to many, it would only serve to make the conversation more uncomfortable. Maybe by not using the slur, more people feel better about joining in the discussion. Thanks!</s><pad>
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What<mask> you have the boot (trunk) in the front of your car? [NEWLINE] [NEWLINE] Then reversing into the spot allows you to access items in storage no matter<mask> close to the back of the space you park. [NEWLINE] [NEWLINE] <mask> an actual answer that applies to most people: [NEWLINE] [NEWLINE] <mask> you are trying to reverse out, you aren't able to see very far away from you<mask> your view is blocked by the cars on either side of you, making it much riskier to just pull out.</s>
Label encoding: <s>What if you have the boot (trunk) in the front of your car? [NEWLINE] [NEWLINE] Then reversing into the spot allows you to access items in storage no matter how close to the back of the space you park. [NEWLINE] [NEWLINE] As an actual answer that applies to most people: [NEWLINE] [NEWLINE] If you are trying to reverse out, you aren't able to see very far away from you as your view is blocked by the cars on either side of you, making it much riskier to just pull out.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0928, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's no science to support this position, and plenty to refute it. [NEWLINE] [NEWLINE] <mask> you are<mask> germaphobic<mask> to be terrified of leaving your toothbrush exposed to germs in the air, then bad news<mask> they're *everywhere*, not just in the bathroom. [NEWLINE] [NEWLINE] Leave your brush upside down in a glass of mouthwash<mask> you care that much (<mask> unless you're the only person that ever uses your bathroom then no amount of your own behaviour is going to protect you).</s>
Label encoding: <s>There's no science to support this position, and plenty to refute it. [NEWLINE] [NEWLINE] If you are so germaphobic as to be terrified of leaving your toothbrush exposed to germs in the air, then bad news because they're *everywhere*, not just in the bathroom. [NEWLINE] [NEWLINE] Leave your brush upside down in a glass of mouthwash if you care that much ( because unless you're the only person that ever uses your bathroom then no amount of your own behaviour is going to protect you).</s>
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It was not my intent to silence you.<mask> I have noticed is that people claim to have no problem with asexuals, *<mask><mask><mask> they stay in the closet*. It's like saying transpeople are ok<mask><mask><mask> they present their birth gender. Or I don't mind gays<mask><mask><mask> they act straight/don't flaunt their homosexuality. [NEWLINE] [NEWLINE] <mask> *can't* people be open about their orientations?<mask> do we have to hide it in order to be accepted?</s>
Label encoding: <s>It was not my intent to silence you. What I have noticed is that people claim to have no problem with asexuals, * as long as they stay in the closet*. It's like saying transpeople are ok as long as they present their birth gender. Or I don't mind gays as long as they act straight/don't flaunt their homosexuality. [NEWLINE] [NEWLINE] Why *can't* people be open about their orientations? Why do we have to hide it in order to be accepted?</s>
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course,<mask> I wouldn't be upset at a disinterested third party who told me that the person whom I thought was one of my best friends was actually telling people, "I'm just pretending to be his friend<mask> he's rich and spends money on me." [NEWLINE] [NEWLINE] Sometimes it's not just harmless behind-their-backs talk.  And in NO cases would I be upset at the person who told. [NEWLINE] [NEWLINE] <mask> are you trying to say again?  It doesn't make sense.</s>
Label encoding: <s>Of course, but I wouldn't be upset at a disinterested third party who told me that the person whom I thought was one of my best friends was actually telling people, "I'm just pretending to be his friend because he's rich and spends money on me." [NEWLINE] [NEWLINE] Sometimes it's not just harmless behind-their-backs talk.  And in NO cases would I be upset at the person who told. [NEWLINE] [NEWLINE] What are you trying to say again?  It doesn't make sense.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> is based on little or no knowledge on topics such<mask> human genetics, neuroplasticity, and human psychology.<mask> please CMV! [ENDQ] [NEWLINE] That is<mask> I ended off my original text post with the above disclaimer!<mask><mask> getting me to "realize that your view is based on literally no evidence" is the first step to changing my view, not the biggest factor. It all starts from one's hypothesis, then seeking to be proven or shown otherwise, isn't it?</s>
Label encoding: <s> [STARTQ] My opinion is based on little or no knowledge on topics such as human genetics, neuroplasticity, and human psychology. So please CMV! [ENDQ] [NEWLINE] That is why I ended off my original text post with the above disclaimer! I think getting me to "realize that your view is based on literally no evidence" is the first step to changing my view, not the biggest factor. It all starts from one's hypothesis, then seeking to be proven or shown otherwise, isn't it?</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> it was culturally acceptable to then yes.<mask> it's not. It is culturally acceptable to wear yoga pants in public, to wear spaghetti straps etc. It is culturally acceptable to wear ripped clothing. [NEWLINE] [NEWLINE] If they want a semi-professional enviornment you make a FAIR dress code. require everyone to wear polos, slacks, or dress at knee length. Prohibiting jeggings, and yoga pants,<mask> guys can wear normal everyday attire is not fair. </s>
Label encoding: <s>If it was culturally acceptable to then yes. But it's not. It is culturally acceptable to wear yoga pants in public, to wear spaghetti straps etc. It is culturally acceptable to wear ripped clothing. [NEWLINE] [NEWLINE] If they want a semi-professional enviornment you make a FAIR dress code. require everyone to wear polos, slacks, or dress at knee length. Prohibiting jeggings, and yoga pants, while guys can wear normal everyday attire is not fair. </s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think his point was that there's more to a phone user experience than hardware specs. I'll take a smaller batter and slower processor<mask> the os is well built for them to the point you don't notice any lagging and the battery still lasts all day. [NEWLINE] [NEWLINE] I'm not saying the iPhone is better or worse than any one phone. I'm just saying claiming any device is superior to another based solely on specs is wrong. It ignores other factors that may impact the user experience. </s>
Label encoding: <s>I think his point was that there's more to a phone user experience than hardware specs. I'll take a smaller batter and slower processor if the os is well built for them to the point you don't notice any lagging and the battery still lasts all day. [NEWLINE] [NEWLINE] I'm not saying the iPhone is better or worse than any one phone. I'm just saying claiming any device is superior to another based solely on specs is wrong. It ignores other factors that may impact the user experience. </s>
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ignorance that results in this level is harm is no excuse. There is no excuse to vote with ignorance<mask> you're voting to take away people's rights. And to be that ignorant at least in the US in 2014 is doing<mask> out of *stubborn* ignorance<mask> this information is<mask> readily available and<mask> Pat doesn't have the decency to research his vote before voting to deny his friend rights, then he's a scumbag and not a good friend or person at all.</s><pad>
Label encoding: <s>Ignorance that results in this level is harm is no excuse. There is no excuse to vote with ignorance when you're voting to take away people's rights. And to be that ignorant at least in the US in 2014 is doing so out of *stubborn* ignorance since this information is so readily available and if Pat doesn't have the decency to research his vote before voting to deny his friend rights, then he's a scumbag and not a good friend or person at all.</s><pad>
Loss: tensor(0.0843, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Basically, you're saying that those countries should not learn from the mistakes of others. <mask> they are not  currently abusing or endangering their working class, then it seems perfectly reasonable to create the laws and continue to not abuse them. <mask> wait until people are abused or suffering before protecting them?  Is it not ok to think ahead and prevent abuse? [NEWLINE] Regarding the improved standards<mask> of industrialized nations outsourcing, they can continue to advance<mask> having safe working conditions  and freedom from abuse.</s>
Label encoding: <s>Basically, you're saying that those countries should not learn from the mistakes of others.  If they are not  currently abusing or endangering their working class, then it seems perfectly reasonable to create the laws and continue to not abuse them.  Why wait until people are abused or suffering before protecting them?  Is it not ok to think ahead and prevent abuse? [NEWLINE] Regarding the improved standards because of industrialized nations outsourcing, they can continue to advance while having safe working conditions  and freedom from abuse.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People can still be challenged and not suffer needless pain.  You're just creating a false dichotomy between either having<mask> we have now or living in a universe with no suffering/missteps/challenges whatsoever. There's a huge amount of ground in between.  Free will can account for missteps and some suffering that we bring upon ourselves. <mask> God allowed only suffering that resulted from freewill I wouldn't have a problem. <mask> clearly this isn't the case.  </s>
Label encoding: <s>People can still be challenged and not suffer needless pain.  You're just creating a false dichotomy between either having what we have now or living in a universe with no suffering/missteps/challenges whatsoever. There's a huge amount of ground in between.  Free will can account for missteps and some suffering that we bring upon ourselves.  If God allowed only suffering that resulted from freewill I wouldn't have a problem.  But clearly this isn't the case.  </s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Rape is never okay, normal, nor expected. I never once said that all men are to blame for the actions of a cold-hearted few. I do not believe dressing a certain way means "expressing themselves". It relates back to the man who appeared to be rich walking through the slums. Yes, he should "technically" have the right to do that,<mask> it is not a perfect world and protecting yourself and taking precautions against people who might harm you just makes sense.</s><pad>
Label encoding: <s>Rape is never okay, normal, nor expected. I never once said that all men are to blame for the actions of a cold-hearted few. I do not believe dressing a certain way means "expressing themselves". It relates back to the man who appeared to be rich walking through the slums. Yes, he should "technically" have the right to do that, but it is not a perfect world and protecting yourself and taking precautions against people who might harm you just makes sense.</s><pad>
Loss: tensor(0.0925, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Well yes, a lot of people seem to insist on a etymological fallacy<mask> it comes to bisexual [ENDQ] [NEWLINE] Yes, it is a fallacy,<mask> one of the primary goals of language is to communicate well,<mask><mask> people are going to use a word A to mean X due to the etymological fallacy and almost everyone is doing<mask>, it isn't unreasonable to have other word B to mean X', especially<mask> having the distinction between X and X' is useful. </s>
Label encoding: <s> [STARTQ] Well yes, a lot of people seem to insist on a etymological fallacy when it comes to bisexual [ENDQ] [NEWLINE] Yes, it is a fallacy, but one of the primary goals of language is to communicate well, so if people are going to use a word A to mean X due to the etymological fallacy and almost everyone is doing so, it isn't unreasonable to have other word B to mean X', especially when having the distinction between X and X' is useful. </s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am curious whether the SAT's predictive power is biased<mask> of its influence on peoples futures. [NEWLINE] [NEWLINE] I have not seen any research regarding the SAT's predictive power<mask> I am<mask><mask> higher score correlate to better schools which subsequently correlate to higher paying jobs?<mask> that is the case would it not follow that low scoring students would be unable to qualify for better school and<mask> less opportunity for high paying jobs? [NEWLINE] [NEWLINE] I would be interested in your thoughts on this. [NEWLINE] [NEWLINE] Edit: spelling</s>
Label encoding: <s>I am curious whether the SAT's predictive power is biased because of its influence on peoples futures. [NEWLINE] [NEWLINE] I have not seen any research regarding the SAT's predictive power however I am assuming that higher score correlate to better schools which subsequently correlate to higher paying jobs? If that is the case would it not follow that low scoring students would be unable to qualify for better school and thus less opportunity for high paying jobs? [NEWLINE] [NEWLINE] I would be interested in your thoughts on this. [NEWLINE] [NEWLINE] Edit: spelling</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0998, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0999, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Writing down my private key is vulnerable to theft<mask> its not encrypted, and vulnerable to me forgetting the key<mask> it is encrypted. [ENDQ] [NEWLINE] A gold bar is<mask> vulnerable to theft... This isn't a valid argument against bitcoins or against this relatively safe way of storing them. [NEWLINE] [NEWLINE] With a 'paper wallet', you have the benefit,<mask>, that you could have multiple backups, and that you can try to move the funds before the thief does<mask><mask> you notice the theft. </s>
Label encoding: <s> [STARTQ] Writing down my private key is vulnerable to theft if its not encrypted, and vulnerable to me forgetting the key if it is encrypted. [ENDQ] [NEWLINE] A gold bar is as vulnerable to theft... This isn't a valid argument against bitcoins or against this relatively safe way of storing them. [NEWLINE] [NEWLINE] With a 'paper wallet', you have the benefit, however, that you could have multiple backups, and that you can try to move the funds before the thief does so when you notice the theft. </s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your argument seems to rely on the idea that capitalism is good<mask> it caused<mask> came after. By this logic, surely the Reformation, which is often said to have caused capitalism to emerge, is actually the greatest thing to ever occur? And the Reformation was only possible<mask> of the Council of Nicea, which was only possible<mask> of the teachings of Jesus, which were only possible<mask> of the construction of the Second Temple,  and<mask> on and<mask> forth back in time forever.</s>
Label encoding: <s>Your argument seems to rely on the idea that capitalism is good because it caused what came after. By this logic, surely the Reformation, which is often said to have caused capitalism to emerge, is actually the greatest thing to ever occur? And the Reformation was only possible because of the Council of Nicea, which was only possible because of the teachings of Jesus, which were only possible because of the construction of the Second Temple,  and so on and so forth back in time forever.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe the attitudes you described are hypocrisy embodied, I just don't believe they're a majority view, particularly among the marriage equality movement. [NEWLINE] [NEWLINE] Edit: For background, I identify<mask> a gay man, I still sleep with transgender persons and occasionally cis-women. I catch flak from some other gay men and women,<mask> I understand<mask> you're coming from,<mask> I try to make sure I don't let individuals color my judgement or drive me away from the community. </s>
Label encoding: <s>I believe the attitudes you described are hypocrisy embodied, I just don't believe they're a majority view, particularly among the marriage equality movement. [NEWLINE] [NEWLINE] Edit: For background, I identify as a gay man, I still sleep with transgender persons and occasionally cis-women. I catch flak from some other gay men and women, so I understand where you're coming from, but I try to make sure I don't let individuals color my judgement or drive me away from the community. </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My first thought after reading your comment is [this]( [URL] +max_w-1000/thatssonotgay_retarded_11.png) [NEWLINE] [NEWLINE] I personally feel that there are situations<mask> none of those proclaimed synonyms fit, and the only word that directly drives home the point of the message is'retarded'. [NEWLINE] [NEWLINE] Sometimes the nuclear option is the only option for pointing out just<mask> asinine a situation is,<mask> more often than not it's abused.</s>
Label encoding: <s>My first thought after reading your comment is [this]( [URL] +max_w-1000/thatssonotgay_retarded_11.png) [NEWLINE] [NEWLINE] I personally feel that there are situations where none of those proclaimed synonyms fit, and the only word that directly drives home the point of the message is'retarded'. [NEWLINE] [NEWLINE] Sometimes the nuclear option is the only option for pointing out just how asinine a situation is, but more often than not it's abused.</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>true.  he's got kindof a shitty attitude. <mask><mask> is not changed,<mask> I don't think we disagree<mask> much.  This kind of material can be used for good or bad.  I hope rather than censor it,<mask>, people will engage in more thoughful conversations about it.  And I hope that the women who fear this stuff will give themselves a little more credit and responsibility vis-a-vis their ability to detect genuineness/attraction</s>
Label encoding: <s>true.  he's got kindof a shitty attitude.  My opinion is not changed, but I don't think we disagree so much.  This kind of material can be used for good or bad.  I hope rather than censor it, though, people will engage in more thoughful conversations about it.  And I hope that the women who fear this stuff will give themselves a little more credit and responsibility vis-a-vis their ability to detect genuineness/attraction</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That first link is pretty interesting,<mask> until you get percentages for other races then it's pretty damn useless in refuting OP's point. Context matters. [ENDQ] [NEWLINE] Are you one to say "American Whites are violent",<mask> crime among American Whites is higher than Western Europeans and East Asians? [NEWLINE] [NEWLINE] [STARTQ] <mask>, you are not factoring in children and the elderly, whom we can all safely assume are not committing violent crimes. [ENDQ] [NEWLINE] OP's claim was about Black *people*</s>
Label encoding: <s> [STARTQ] That first link is pretty interesting, but until you get percentages for other races then it's pretty damn useless in refuting OP's point. Context matters. [ENDQ] [NEWLINE] Are you one to say "American Whites are violent", because crime among American Whites is higher than Western Europeans and East Asians? [NEWLINE] [NEWLINE] [STARTQ] Also, you are not factoring in children and the elderly, whom we can all safely assume are not committing violent crimes. [ENDQ] [NEWLINE] OP's claim was about Black *people*</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0996, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0902, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "Its too horrible to use,<mask> we would never use them."  This was true ever<mask> the development of nuclear bombs,<mask> we've have used atomic bombs and pretty close with nuclear. [ENDQ] [NEWLINE] This was said after the bombs were dropped on Hiroshima and Nagasaki, and nuclear weapons have not been used<mask>. I see this<mask> the reason for the Cold War, both sides were unwilling to attack one another directly due to the assured mutual destruction that would result. </s>
Label encoding: <s> [STARTQ] "Its too horrible to use, so we would never use them."  This was true ever since the development of nuclear bombs, yet we've have used atomic bombs and pretty close with nuclear. [ENDQ] [NEWLINE] This was said after the bombs were dropped on Hiroshima and Nagasaki, and nuclear weapons have not been used since. I see this as the reason for the Cold War, both sides were unwilling to attack one another directly due to the assured mutual destruction that would result. </s>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0988, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The US and USSR kept fighting [proxy]( [URL] ) [wars.]( [URL] ) [NEWLINE] [NEWLINE] Being much more involved in events around the world was often very bad (see, 1953 US Iran coup, 1980 Soviet invasion of Afghanistan, 1962 Cuban Missile Crisis). [NEWLINE] [NEWLINE] Plus the USSR was a horrifyingly repressive state which ruled its people with an iron fist, frequently [murdering them en masse.]( [URL] )  Putin's got nothing on Stalin for horribleness.</s>
Label encoding: <s>The US and USSR kept fighting [proxy]( [URL] ) [wars.]( [URL] ) [NEWLINE] [NEWLINE] Being much more involved in events around the world was often very bad (see, 1953 US Iran coup, 1980 Soviet invasion of Afghanistan, 1962 Cuban Missile Crisis). [NEWLINE] [NEWLINE] Plus the USSR was a horrifyingly repressive state which ruled its people with an iron fist, frequently [murdering them en masse.]( [URL] )  Putin's got nothing on Stalin for horribleness.</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, drunk people don't do<mask> much stealing, more violence. <mask> you get in a bar fight with a drunk person, you are most likely an idiot for engaging them. [NEWLINE] [NEWLINE] Theft is more a thing that happens<mask> you get really drunk and lose situational awareness in a way others can exploit. [NEWLINE] [NEWLINE] <mask> then again,<mask> you're like my friend who took $2000 cash with him<mask> going to clubs in Vegas, yes, you're an idiot.</s>
Label encoding: <s>Well, drunk people don't do as much stealing, more violence.  If you get in a bar fight with a drunk person, you are most likely an idiot for engaging them. [NEWLINE] [NEWLINE] Theft is more a thing that happens when you get really drunk and lose situational awareness in a way others can exploit. [NEWLINE] [NEWLINE] But then again, if you're like my friend who took $2000 cash with him when going to clubs in Vegas, yes, you're an idiot.</s>
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] it's unscientific to not accept the existence of alien life. [ENDQ] [NEWLINE] *Accepting the existence* of alien life is premature, and pretty unscientific to say. We can say that there exists a *strong probability* for alien life,<mask> you can't admit the existence of something not<mask> proven - even<mask> lots of circumstantial evidence points to it existing. Otherwise finding the Higgs or tetraquark would not be such a big deal.</s>
Label encoding: <s> [STARTQ] it's unscientific to not accept the existence of alien life. [ENDQ] [NEWLINE] *Accepting the existence* of alien life is premature, and pretty unscientific to say. We can say that there exists a *strong probability* for alien life, but you can't admit the existence of something not yet proven - even if lots of circumstantial evidence points to it existing. Otherwise finding the Higgs or tetraquark would not be such a big deal.</s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's a good point,<mask><mask><mask> people would quickly vote<mask> a policy that cats are unacceptable candidates. [NEWLINE] [NEWLINE] People have the opportunity to write cats in on the ballot<mask> it is,<mask> it doesn't happen<mask> people realize these things affect people's real lives, unlike logo contests and things. [NEWLINE] [NEWLINE] "works just fine" &lt;--debatable [NEWLINE] [NEWLINE] Edit: removed sentence implying that electing a cat might be OK.^(vote 4cat)</s><pad>
Label encoding: <s>That's a good point, but I think people would quickly vote as a policy that cats are unacceptable candidates. [NEWLINE] [NEWLINE] People have the opportunity to write cats in on the ballot as it is, but it doesn't happen because people realize these things affect people's real lives, unlike logo contests and things. [NEWLINE] [NEWLINE] "works just fine" &lt;--debatable [NEWLINE] [NEWLINE] Edit: removed sentence implying that electing a cat might be OK.^(vote 4cat)</s><pad>
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] heteronormativity [ENDQ] [NEWLINE] Well the vast majority of people are heterosexual. [NEWLINE] [NEWLINE] It is not discrimination against gays to say "you are a man, you go to the mens locker room to change". That is basic anatomy. [NEWLINE] [NEWLINE] Not to mention that most people simply would not feel comfortable, women<mask> much<mask> men. Can you imagine having to get naked in front of a group of the opposite gender? My word, that would be embarrassing. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] heteronormativity [ENDQ] [NEWLINE] Well the vast majority of people are heterosexual. [NEWLINE] [NEWLINE] It is not discrimination against gays to say "you are a man, you go to the mens locker room to change". That is basic anatomy. [NEWLINE] [NEWLINE] Not to mention that most people simply would not feel comfortable, women as much as men. Can you imagine having to get naked in front of a group of the opposite gender? My word, that would be embarrassing. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0793, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>With that in mind,  I would be curious on your thoughts of the loose sequel to Ender's Game, Speaker for the Dead. It's really insane to compare the book's themes about acceptance to Card's current views on homosexuality.<mask> it's an entirely different tone to Ender's Game, which I didn't care for terribly much either,<mask> it might appeal to you. Obviously don't give the man more money,<mask> pick it up at a library or something.</s>
Label encoding: <s>With that in mind,  I would be curious on your thoughts of the loose sequel to Ender's Game, Speaker for the Dead. It's really insane to compare the book's themes about acceptance to Card's current views on homosexuality. Also it's an entirely different tone to Ender's Game, which I didn't care for terribly much either, so it might appeal to you. Obviously don't give the man more money, but pick it up at a library or something.</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ]. Most companies get stacks of resumes of equally qualified candidates. [ENDQ] [NEWLINE] Depends on the job. [NEWLINE] [NEWLINE] Menial work that doesn't really require expertise?  Sure, I'll let a typo influenced my decision,<mask><mask> you said, there's 10 other candidates who are just<mask> capable. [NEWLINE] [NEWLINE] Technical jobs?  It's not unlikely one or two candidates will be heads above the rest, I'm not going to discount them<mask> of one typo. </s>
Label encoding: <s> [STARTQ]. Most companies get stacks of resumes of equally qualified candidates. [ENDQ] [NEWLINE] Depends on the job. [NEWLINE] [NEWLINE] Menial work that doesn't really require expertise?  Sure, I'll let a typo influenced my decision, because as you said, there's 10 other candidates who are just as capable. [NEWLINE] [NEWLINE] Technical jobs?  It's not unlikely one or two candidates will be heads above the rest, I'm not going to discount them because of one typo. </s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What does being respectful have anything to do with toilets? [NEWLINE] [NEWLINE] Isn't it just better<mask> everyone in the house only has to do one action<mask> they use the toilet? They can either put it down or lift it up. It's fair to everyone in the house.<mask> makes the people using the toilets<mask> special that they can't afford to do that one action and requires the male to always perform the two actions of lifting it up and putting it  down? </s>
Label encoding: <s>What does being respectful have anything to do with toilets? [NEWLINE] [NEWLINE] Isn't it just better if everyone in the house only has to do one action when they use the toilet? They can either put it down or lift it up. It's fair to everyone in the house. What makes the people using the toilets so special that they can't afford to do that one action and requires the male to always perform the two actions of lifting it up and putting it  down? </s>
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0786, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>!delta I'm awarding this delta<mask> the magnitude of the issue that I perceived is a lot smaller, quantitatively, than I previously thought.<mask>, I'm still uncomfortable with the concept and am realizing that that probably won't be changed.<mask> my own personal feelings, I am generally in favor of a society in which people are free to make their own decisions, including ones that I'd be uncomfortable making. These numbers strengthen my resolve in this particular issue. </s>
Label encoding: <s>!delta I'm awarding this delta because the magnitude of the issue that I perceived is a lot smaller, quantitatively, than I previously thought. However, I'm still uncomfortable with the concept and am realizing that that probably won't be changed. Despite my own personal feelings, I am generally in favor of a society in which people are free to make their own decisions, including ones that I'd be uncomfortable making. These numbers strengthen my resolve in this particular issue. </s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0839, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0861, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who cares<mask> the act of having sex, in isolation, is not immoral?  We're talking about the act of having sex in a specific content. [NEWLINE] [NEWLINE] It's wrong to hurt people for your own benefit,<mask><mask> whether someone else involved<mask> broke a promise. [NEWLINE] [NEWLINE] Just<mask> nobody expects teenagers to rat out their friends doesn't somehow make it okay to hurt people just<mask><mask><mask> you can point your finger at someone else. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Who cares if the act of having sex, in isolation, is not immoral?  We're talking about the act of having sex in a specific content. [NEWLINE] [NEWLINE] It's wrong to hurt people for your own benefit, regardless of whether someone else involved also broke a promise. [NEWLINE] [NEWLINE] Just because nobody expects teenagers to rat out their friends doesn't somehow make it okay to hurt people just as long as you can point your finger at someone else. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0814, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] I'd never heard this theory in a religious context before. Instead, I've always heard it<mask> a philosophical experiment along he lines of "You can't prove the world wasn't created five seconds ago." [NEWLINE] [NEWLINE] <mask><mask> this actually does reconcile a literal Genesis with observed reality,<mask> it's<mask> gotten me reconsidering the "Simulation Argument" and different possible interpretations of it. [NEWLINE] [NEWLINE] Thanks much for this.</s>
Label encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] I'd never heard this theory in a religious context before. Instead, I've always heard it as a philosophical experiment along he lines of "You can't prove the world wasn't created five seconds ago." [NEWLINE] [NEWLINE] I think this actually does reconcile a literal Genesis with observed reality, but it's also gotten me reconsidering the "Simulation Argument" and different possible interpretations of it. [NEWLINE] [NEWLINE] Thanks much for this.</s>
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think country music is definitely a factor (even<mask> it's not the largest one) in the cultural encouragement of wearing ignorance<mask> a badge of pride for teens in the south, and for telling people in their twenties that high school was the best time of your life, and you can never go back<mask> get to drinking and reminiscing. It's not until you get to the songs targeting 30s-50s groups that songs become about family and softer topics.</s>
Label encoding: <s>I think country music is definitely a factor (even if it's not the largest one) in the cultural encouragement of wearing ignorance as a badge of pride for teens in the south, and for telling people in their twenties that high school was the best time of your life, and you can never go back so get to drinking and reminiscing. It's not until you get to the songs targeting 30s-50s groups that songs become about family and softer topics.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0856, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree to an extent, and you do make a good point with centralizing the money.<mask>, my personal belief is that the government would do a better job than the people. And with taxes, we at least have a guaranteed money going *somewhere*. [NEWLINE] [NEWLINE] <mask> given &lt;5% taxes, I doubt you'd see much more of that money go to charities with all the other problems most Americans face, especially medical debt. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I agree to an extent, and you do make a good point with centralizing the money. However, my personal belief is that the government would do a better job than the people. And with taxes, we at least have a guaranteed money going *somewhere*. [NEWLINE] [NEWLINE] If given &lt;5% taxes, I doubt you'd see much more of that money go to charities with all the other problems most Americans face, especially medical debt. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Look mate, abortions often leave the mothers psychologically destroyed. Adoptions may occasionally be even worse. The deal is, sometimes women just get pregnant<mask> of a single error, or even<mask> of bad luck (say, a faulty or otherwise torn condom). And at that point, a woman can either deal with it or potentially enter a period of depression wich may potentially lead to suicide, expecially considering the lack of a husband or boyfriend to confort her.</s>
Label encoding: <s>Look mate, abortions often leave the mothers psychologically destroyed. Adoptions may occasionally be even worse. The deal is, sometimes women just get pregnant because of a single error, or even because of bad luck (say, a faulty or otherwise torn condom). And at that point, a woman can either deal with it or potentially enter a period of depression wich may potentially lead to suicide, expecially considering the lack of a husband or boyfriend to confort her.</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What do you propose instead?  Testing my not be the "best" gauge of whether a student has learned<mask> they are supposed to in a class,<mask> we want *some* way to make sure the students enrolled are actually learning.  Unless you have a better idea than testing, I don't know<mask> you think we should do?  Just abolish all tests and everyone passes every class with no way of separating the top students from the bottom ones?</s><pad>
Label encoding: <s>What do you propose instead?  Testing my not be the "best" gauge of whether a student has learned what they are supposed to in a class, but we want *some* way to make sure the students enrolled are actually learning.  Unless you have a better idea than testing, I don't know what you think we should do?  Just abolish all tests and everyone passes every class with no way of separating the top students from the bottom ones?</s><pad>
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>morality refers to a system of ethics meant to inform decisions.<mask> there are no decisions then morality doesn't exist. You can't teach a wind up toy morality,<mask> it can't think, it can't change it's destiny.<mask> determinism is true then not only do we have no choices, we have no independent thoughts at all. All our thoughts are simply reactions to stimuli that we have absolutely no control over, we are simply wind up dolls</s>
Label encoding: <s>morality refers to a system of ethics meant to inform decisions. If there are no decisions then morality doesn't exist. You can't teach a wind up toy morality, because it can't think, it can't change it's destiny. If determinism is true then not only do we have no choices, we have no independent thoughts at all. All our thoughts are simply reactions to stimuli that we have absolutely no control over, we are simply wind up dolls</s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0987, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0977, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Scores: 7, 3 and 0. No big deal at all. [NEWLINE] [NEWLINE] Like I said, it is nice<mask> it is.<mask> the sub get grows too much AND things get out of control, it might be a nice option to have a stronger moderation.<mask> you making a tempest in a teapot. Again, you see these posts<mask> the sub is small. After it becomes more popular, only high voted stuff will be visible.</s><pad>
Label encoding: <s>Scores: 7, 3 and 0. No big deal at all. [NEWLINE] [NEWLINE] Like I said, it is nice how it is. IF the sub get grows too much AND things get out of control, it might be a nice option to have a stronger moderation. But you making a tempest in a teapot. Again, you see these posts because the sub is small. After it becomes more popular, only high voted stuff will be visible.</s><pad>
Loss: tensor(0.1044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I definitely find the idea of "exotic" lifeforms very interesting (non-DNA, non-carbon, macro-organisms etc). My hunch is that something like this exists out there somewhere- it feels a bit short-sighted to limit our search for life to planets resembling Earth. [NEWLINE] [NEWLINE] <mask> you haven't read it, I can highly recommend Olaf Stapledon's "Star Maker", which is filled with ideas about organisms like this.</s>
Label encoding: <s>I definitely find the idea of "exotic" lifeforms very interesting (non-DNA, non-carbon, macro-organisms etc). My hunch is that something like this exists out there somewhere- it feels a bit short-sighted to limit our search for life to planets resembling Earth. [NEWLINE] [NEWLINE] If you haven't read it, I can highly recommend Olaf Stapledon's "Star Maker", which is filled with ideas about organisms like this.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's neat that you imagine eugenics in that specific example. That's not actually an example of<mask> eugenics even is. [NEWLINE] [NEWLINE] That example is selective breeding. You're trying to emphasize certain characteristics, not eliminate anything particular. [NEWLINE] [NEWLINE] Eugenics is<mask> you're suggesting in your OP: targeted elimination of a gene from a population by either killing individuals with the gene of by preventing thr gene from being passed on in any way. </s>
Label encoding: <s>It's neat that you imagine eugenics in that specific example. That's not actually an example of what eugenics even is. [NEWLINE] [NEWLINE] That example is selective breeding. You're trying to emphasize certain characteristics, not eliminate anything particular. [NEWLINE] [NEWLINE] Eugenics is what you're suggesting in your OP: targeted elimination of a gene from a population by either killing individuals with the gene of by preventing thr gene from being passed on in any way. </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Bombing is easy [ENDQ] [NEWLINE] Creating the right bomb, planting, and detonating them at critical location is definitely not easy. Standalone packages won't last an hour on places like NY subs without being reported, and there are dud bomb cases like the 2010 Times Square car bombing or the 2007 Germany train duds. [NEWLINE] [NEWLINE] We have foiled dozens of terrorists plots,<mask> obviously "the fear mongering" is effective. [NEWLINE] [NEWLINE] [URL] </s>
Label encoding: <s> [STARTQ] Bombing is easy [ENDQ] [NEWLINE] Creating the right bomb, planting, and detonating them at critical location is definitely not easy. Standalone packages won't last an hour on places like NY subs without being reported, and there are dud bomb cases like the 2010 Times Square car bombing or the 2007 Germany train duds. [NEWLINE] [NEWLINE] We have foiled dozens of terrorists plots, so obviously "the fear mongering" is effective. [NEWLINE] [NEWLINE] [URL] </s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do we need a law for that,<mask>? It's a matter of incentives. We want people to wear helmets,<mask> we make a law that says you need to wear a helmet or you'll get a fine. Except there's *already* an incentive to wear a helmet called "not dying." Is a fine really a better incentive than that? And<mask> people are willing to ignore the second incentive,<mask> would they follow the first?</s><pad>
Label encoding: <s>Why do we need a law for that, though? It's a matter of incentives. We want people to wear helmets, so we make a law that says you need to wear a helmet or you'll get a fine. Except there's *already* an incentive to wear a helmet called "not dying." Is a fine really a better incentive than that? And if people are willing to ignore the second incentive, why would they follow the first?</s><pad>
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a pretty good reply. My only response is that longer does not equal better. To me, the extended scenes are lower quality than the rest of the films - that's<mask> they were cut. They dilute the necessary scenes. [NEWLINE] [NEWLINE] I wonder<mask> I'd feel differently<mask> I'd seen the extended cuts first. Except that means YOU NEVER SAW RETURN OF THE KING IN THEATERS!!!!! Dude. You have to. </s>
Label encoding: <s>This is a pretty good reply. My only response is that longer does not equal better. To me, the extended scenes are lower quality than the rest of the films - that's why they were cut. They dilute the necessary scenes. [NEWLINE] [NEWLINE] I wonder if I'd feel differently if I'd seen the extended cuts first. Except that means YOU NEVER SAW RETURN OF THE KING IN THEATERS!!!!! Dude. You have to. </s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0988, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean in the same sense that *every* country is unique, sure.<mask> then<mask>'s the point in even saying we're different from other countries? [NEWLINE] [NEWLINE] Typically someone only brings up the fact that we're different in a context of *and better than everyone else*. At least that's<mask> I'm seeing it. Otherwise I recognize that Earth is occupied by almost 200 or<mask> nation states. Some similar to ours, some very different.</s>
Label encoding: <s>I mean in the same sense that *every* country is unique, sure. But then what's the point in even saying we're different from other countries? [NEWLINE] [NEWLINE] Typically someone only brings up the fact that we're different in a context of *and better than everyone else*. At least that's how I'm seeing it. Otherwise I recognize that Earth is occupied by almost 200 or so nation states. Some similar to ours, some very different.</s>
Loss: tensor(0.1056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0964, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0956, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0937, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If by "cull those infected" you mean kill anyone diagnosed with ebola in any country?  Whelp, now everyone who catches ebola is going to hide it from everyone<mask> those they most deeply trust.  Now ebola will be even more difficult to track and will spread to more people<mask> family members and friends take care of the dying. [NEWLINE] [NEWLINE] Congratulations, you just made a bad situations 10,000 times worse. </s>
Label encoding: <s>If by "cull those infected" you mean kill anyone diagnosed with ebola in any country?  Whelp, now everyone who catches ebola is going to hide it from everyone but those they most deeply trust.  Now ebola will be even more difficult to track and will spread to more people as family members and friends take care of the dying. [NEWLINE] [NEWLINE] Congratulations, you just made a bad situations 10,000 times worse. </s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0958, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL].php/Main/ExiledToTheCouch [NEWLINE] [NEWLINE] There's about 15 Television examples of the Trope. [NEWLINE] [NEWLINE] Like I said, I don't believe there have been peer reviewed studies on this,<mask> it is a popular media trope. [NEWLINE] [NEWLINE] <mask> for household labour, do you really need me to link the top 10 google search results? Is it that contentious to claim that men tend to do outdoor work? [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s> [URL].php/Main/ExiledToTheCouch [NEWLINE] [NEWLINE] There's about 15 Television examples of the Trope. [NEWLINE] [NEWLINE] Like I said, I don't believe there have been peer reviewed studies on this, but it is a popular media trope. [NEWLINE] [NEWLINE] As for household labour, do you really need me to link the top 10 google search results? Is it that contentious to claim that men tend to do outdoor work? [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.1094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>i feel like this is something you should talk about with your daughter<mask> the time comes. point out<mask> you said. ask her about<mask> she thinks of<mask> girls and women are represented on the show. talk about ethnocentrism and<mask> it relates to the real world. i feel like you can enjoy media<mask> still be critical of it, and the show may open up a lot of talking points you may not have had otherwise. </s>
Label encoding: <s>i feel like this is something you should talk about with your daughter when the time comes. point out what you said. ask her about what she thinks of how girls and women are represented on the show. talk about ethnocentrism and how it relates to the real world. i feel like you can enjoy media yet still be critical of it, and the show may open up a lot of talking points you may not have had otherwise. </s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0894, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0826, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0958, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd say that the parameters of your CMV are flawed<mask>, unless I'm misunderstanding the rule of two, Vader was never a Sith lord. At the height of his power he was a mere apprentice. I'd even make the case that he'd never fallen completely to the dark side, and there are hints of his redemption<mask> early<mask> The Empire Strikes Back.<mask> it's precisely these tragic features that make him such a compelling villain.</s>
Label encoding: <s>I'd say that the parameters of your CMV are flawed because, unless I'm misunderstanding the rule of two, Vader was never a Sith lord. At the height of his power he was a mere apprentice. I'd even make the case that he'd never fallen completely to the dark side, and there are hints of his redemption as early as The Empire Strikes Back. But it's precisely these tragic features that make him such a compelling villain.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0829, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I didn't downvote you, I just saw this now. [NEWLINE] [NEWLINE] <mask> yeah, I actually do think that eating them at our current consumption rate is wrong.<mask>, not for the same reasoning<mask> you imply.<mask><mask> the context of eating animals and having sex with them is different<mask> a) many animals themselves eat other animals and b) one is usually tied to survival.<mask><mask> this deserves a different CMV,<mask>.</s>
Label encoding: <s>I didn't downvote you, I just saw this now. [NEWLINE] [NEWLINE] But yeah, I actually do think that eating them at our current consumption rate is wrong. However, not for the same reasoning as you imply. I think the context of eating animals and having sex with them is different because a) many animals themselves eat other animals and b) one is usually tied to survival. I think this deserves a different CMV, though.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0999, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is it<mask> they have a longer stride, taller build, and greater proportion of fast-twitch muscles [ENDQ] [NEWLINE] <mask> do we not see the Black Jamaican men<mask> the best swimming sprinters then. the 50m 100m freestyle, or fly.<mask> they had a greater proportion of fast twitch and could have longer moment arms in which to apply force. They should dominate these events<mask> well,<mask> they don't. </s>
Label encoding: <s> [STARTQ] Is it because they have a longer stride, taller build, and greater proportion of fast-twitch muscles [ENDQ] [NEWLINE] Why do we not see the Black Jamaican men as the best swimming sprinters then. the 50m 100m freestyle, or fly. If they had a greater proportion of fast twitch and could have longer moment arms in which to apply force. They should dominate these events as well, but they don't. </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Not allowing gay people to marry doesnt actually hurt anyone. [ENDQ] [NEWLINE] This is pretty obviously false,<mask> marriage is crucially tied up with taxation, hospital visitation rights, inheritance in the absence of a will, child custody, and other legal constructs. All of these can do more harm (fiscal or emotional) to an individual in a non-marriage relationship than they would<mask> the same individual were married to the same person. </s>
Label encoding: <s> [STARTQ] Not allowing gay people to marry doesnt actually hurt anyone. [ENDQ] [NEWLINE] This is pretty obviously false, because marriage is crucially tied up with taxation, hospital visitation rights, inheritance in the absence of a will, child custody, and other legal constructs. All of these can do more harm (fiscal or emotional) to an individual in a non-marriage relationship than they would if the same individual were married to the same person. </s>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0892, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0846, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In the US,<mask><mask> Wikipedia, [dogs kill 20 to 30 people per year]( [URL] ). At the same time, [suicide kills over 30000 per year]( [URL] ). [NEWLINE] [NEWLINE] Would it be reasonable to guess that<mask> we had no pet dogs that the suicide rate would increase by more than 0.1%?<mask><mask>, then we can estimate that just through suicide prevention dogs save more people than they kill. </s>
Label encoding: <s>In the US, according to Wikipedia, [dogs kill 20 to 30 people per year]( [URL] ). At the same time, [suicide kills over 30000 per year]( [URL] ). [NEWLINE] [NEWLINE] Would it be reasonable to guess that if we had no pet dogs that the suicide rate would increase by more than 0.1%? If so, then we can estimate that just through suicide prevention dogs save more people than they kill. </s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've likely flown about a hundred flights, and I still somewhat disagree with you. Getting in line before your section is called is still relatively pointless. Just sit somewhere close to the gate itself, and<mask> your section is called promptly get up and in line. I have never had an issue getting my full sized carry on to fit on any flight (I usually fly with no checked bags,<mask> my carry on is not small). </s>
Label encoding: <s>I've likely flown about a hundred flights, and I still somewhat disagree with you. Getting in line before your section is called is still relatively pointless. Just sit somewhere close to the gate itself, and when your section is called promptly get up and in line. I have never had an issue getting my full sized carry on to fit on any flight (I usually fly with no checked bags, so my carry on is not small). </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>this is a process<mask>, and<mask><mask> its getting better. racism is still engrained in society,<mask><mask> could it not be? there are still tons of people alive who were raised during the tail end of segregation.<mask>, that's not going to be the case here shortly, and<mask> things like the media and blogs could stop focussing on white vs black mentalities we might actually start seeing everyone<mask> people. </s>
Label encoding: <s>this is a process though, and i think its getting better. racism is still engrained in society, but how could it not be? there are still tons of people alive who were raised during the tail end of segregation. however, that's not going to be the case here shortly, and if things like the media and blogs could stop focussing on white vs black mentalities we might actually start seeing everyone as people. </s>
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Good is relative and subjective.  Good to one culture or society may be atrocious to others.  Average is actually just that<mask>, it's middle of the road<mask><mask> society. <mask> to change your mind there would have to be a gauge, some sort of medium we can all agree equates to good, or average, or bad.  Otherwise we all are speaking different languages with no common point of reference.</s>
Label encoding: <s>Good is relative and subjective.  Good to one culture or society may be atrocious to others.  Average is actually just that though, it's middle of the road regardless of society.  So to change your mind there would have to be a gauge, some sort of medium we can all agree equates to good, or average, or bad.  Otherwise we all are speaking different languages with no common point of reference.</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] She was dressed specifically to give the impression that she wants attention [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] <mask> about, instead of judging her by her cloths and creating your own selfish reasons<mask> to<mask> she chose to dress that way, you judge it by her actions. Make eye contact,<mask> she smiles or aknowledges you then say something,<mask> don't make assumptions about<mask> much attention she wants due to<mask> she is wearing.</s>
Label encoding: <s> [STARTQ] She was dressed specifically to give the impression that she wants attention [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] How about, instead of judging her by her cloths and creating your own selfish reasons as to why she chose to dress that way, you judge it by her actions. Make eye contact, if she smiles or aknowledges you then say something, but don't make assumptions about how much attention she wants due to what she is wearing.</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I completely agree with you, and that's<mask> I tried to make it clear my issue is not with people's beliefs or their right to say it. It is with defenders of free speech over someone who is willing to have a discussion. In other words, "I disapprove of<mask> you say,<mask> I will try to advance humanity's condition with you by communicating with you, or I won't say anything at all". </s><pad>
Label encoding: <s>I completely agree with you, and that's why I tried to make it clear my issue is not with people's beliefs or their right to say it. It is with defenders of free speech over someone who is willing to have a discussion. In other words, "I disapprove of what you say, but I will try to advance humanity's condition with you by communicating with you, or I won't say anything at all". </s><pad>
Loss: tensor(0.1103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0979, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Excellent way to illustrate my point.  The hypothetical theater now has to choose between losing the business of me and my wife (2 tickets) or you (1 ticket). [NEWLINE] [NEWLINE] Economically, which one do you think they would choose?  Especially<mask> they had rules like I've outlined<mask><mask> I act the bad parent and don't bother to try to calm my baby down, they can kick us out with no refund?</s>
Label encoding: <s>Excellent way to illustrate my point.  The hypothetical theater now has to choose between losing the business of me and my wife (2 tickets) or you (1 ticket). [NEWLINE] [NEWLINE] Economically, which one do you think they would choose?  Especially if they had rules like I've outlined where if I act the bad parent and don't bother to try to calm my baby down, they can kick us out with no refund?</s>
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0973, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Teach them to say no,<mask> they don't stop, say 'hey, stop it, I do not want to have sex', then 'STOP, YOU ARE RAPING ME'.  Guys might get confused with "I don't know<mask> I want to", or, "not right now" or something like that, guys won't get confused<mask> she says "<mask> you continue you're raping me".</s>
Label encoding: <s>Teach them to say no, if they don't stop, say 'hey, stop it, I do not want to have sex', then 'STOP, YOU ARE RAPING ME'.  Guys might get confused with "I don't know if I want to", or, "not right now" or something like that, guys won't get confused when she says " if you continue you're raping me".</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>* In short, that study concludes that a health conscious meat eater will live just<mask><mask><mask> a heath conscious vegetarian. People will claim that vegetarians live longer,<mask> this is simply not the case. [NEWLINE] [NEWLINE] * [Up to 80 % of the water we take from rivers and groundwater goes into irrigation]( [URL].aspx) [NEWLINE] [NEWLINE] * Yes,<mask> that's<mask> you don't flush grease down the drain.</s>
Label encoding: <s>* In short, that study concludes that a health conscious meat eater will live just as long as a heath conscious vegetarian. People will claim that vegetarians live longer, but this is simply not the case. [NEWLINE] [NEWLINE] * [Up to 80 % of the water we take from rivers and groundwater goes into irrigation]( [URL].aspx) [NEWLINE] [NEWLINE] * Yes, but that's why you don't flush grease down the drain.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0934, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me just contradict you here on one little comment. [NEWLINE] [NEWLINE] [STARTQ] <mask> I don't think trans-sexuality should be encouraged. [ENDQ] [NEWLINE] I don't think any trans person I know would say they've ever felt "encouraged".  This myth that society is somehow "letting" transsexuality happen is absurd.  From the point of view of just about every trans person, it's very strongly discouraged.</s>
Label encoding: <s>Let me just contradict you here on one little comment. [NEWLINE] [NEWLINE] [STARTQ] But I don't think trans-sexuality should be encouraged. [ENDQ] [NEWLINE] I don't think any trans person I know would say they've ever felt "encouraged".  This myth that society is somehow "letting" transsexuality happen is absurd.  From the point of view of just about every trans person, it's very strongly discouraged.</s>
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] At its core, the goal of a "science" is to increase or refine knowledge through research [ENDQ] [NEWLINE] This seems kind of flimsy to me. Does this mean researching for a literature paper is a science?<mask> about history? There's definitely science in forensics,<mask> is looking at sources constitute a science? I've never thought about those sorts of things<mask> sciences,<mask> they seem to fit your definition. </s>
Label encoding: <s> [STARTQ] At its core, the goal of a "science" is to increase or refine knowledge through research [ENDQ] [NEWLINE] This seems kind of flimsy to me. Does this mean researching for a literature paper is a science? What about history? There's definitely science in forensics, but is looking at sources constitute a science? I've never thought about those sorts of things as sciences, but they seem to fit your definition. </s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's a big part of most people's lives, to ignore understanding that comes from engaging with the social fabric (and not just theory) would,<mask><mask><mask>, limit his potential..<mask> would he write about instead? Would it be real? Would it be deep and believable? Would people relate to it... or want to? [NEWLINE] [NEWLINE] People relate to people - socializing isn't'skillfully fake' to everyone</s>
Label encoding: <s>It's a big part of most people's lives, to ignore understanding that comes from engaging with the social fabric (and not just theory) would, in my opinion, limit his potential.. what would he write about instead? Would it be real? Would it be deep and believable? Would people relate to it... or want to? [NEWLINE] [NEWLINE] People relate to people - socializing isn't'skillfully fake' to everyone</s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> it is possible doesn't make it valuable.<mask> is the point in keep coming up with "meanings" of a two line poem? Nobody who is reading the poem is taking hours trying to find the hidden meanings behind it. More importantly,<mask> it conveys a specific meaning then<mask> does it take ten pages to explain this meaning? And<mask> does it take a ten page analysis to understand the human condition?</s>
Label encoding: <s>Just because it is possible doesn't make it valuable. What is the point in keep coming up with "meanings" of a two line poem? Nobody who is reading the poem is taking hours trying to find the hidden meanings behind it. More importantly, if it conveys a specific meaning then why does it take ten pages to explain this meaning? And why does it take a ten page analysis to understand the human condition?</s>
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0913, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Isn't the term for having been through a matrimony a marriage? That might not come out quite right<mask> hope you can understand it. [NEWLINE] [NEWLINE] I got a good chuckle out of your divorced idea. [NEWLINE] [NEWLINE] <mask><mask> both sides want it and<mask><mask> the easiest way to fix it, is change the legal side. That is quite easy, just change a word on a document and BAM! Done. </s>
Label encoding: <s>Isn't the term for having been through a matrimony a marriage? That might not come out quite right but hope you can understand it. [NEWLINE] [NEWLINE] I got a good chuckle out of your divorced idea. [NEWLINE] [NEWLINE] I agree both sides want it and I think the easiest way to fix it, is change the legal side. That is quite easy, just change a word on a document and BAM! Done. </s>
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0899, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know many bankers,<mask> most the doctors/bankers/professors I've had are split pretty much down the middle male/female. I say "pretty much"<mask> they were mostly women.<mask>, almost all of those jobs require post-secondary education. More women are in college/university than male now,<mask> there is no sufficient evidence to prove that women do these jobs less anymore.</s>
Label encoding: <s>I don't know many bankers, but most the doctors/bankers/professors I've had are split pretty much down the middle male/female. I say "pretty much" because they were mostly women. Also, almost all of those jobs require post-secondary education. More women are in college/university than male now, so there is no sufficient evidence to prove that women do these jobs less anymore.</s>
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is actually a memory and brain structure issue. You make sense on new info and store it, in it's relation to other information you have in your brain. Basically you have to have a mental scaffolding in place to cement new info.<mask> you don't learn a lot of facts you don't have a scaffolding. People without good scaffolding tend to get confused ie Nazis WW1 or 2.  </s>
Label encoding: <s>This is actually a memory and brain structure issue. You make sense on new info and store it, in it's relation to other information you have in your brain. Basically you have to have a mental scaffolding in place to cement new info. If you don't learn a lot of facts you don't have a scaffolding. People without good scaffolding tend to get confused ie Nazis WW1 or 2.  </s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This delta is currently disallowed<mask> your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for<mask> /u/damienrapp98 changed your view.<mask> you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>This delta is currently disallowed as your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for how /u/damienrapp98 changed your view. If you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Came here to say this. My wife uses your strategy,<mask> I am more of your sister's style. After solving a puzzle, my wife really, really knows the painting (we do puzzles of paintings),<mask> my share perception and spacial cognitive skills are better than hers. [NEWLINE] Before I met my wife, it hadn't even crossed my mind to look at the picture in order to solve the puzzle! </s>
Label encoding: <s>Came here to say this. My wife uses your strategy, while I am more of your sister's style. After solving a puzzle, my wife really, really knows the painting (we do puzzles of paintings), while my share perception and spacial cognitive skills are better than hers. [NEWLINE] Before I met my wife, it hadn't even crossed my mind to look at the picture in order to solve the puzzle! </s>
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0926, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0917, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Uh. [NEWLINE] [NEWLINE] /r/Feminism is run by a moderator who has come under fire for allowing MRAs to post there and make the discussion about men. It's actually a huge controversy among Reddit feminists. [NEWLINE] [NEWLINE] The fact that you're honestly claiming that discussing men's issues gets you banned there means that you've never actually visited /r/Feminism and are straight up **lying**.</s>
Label encoding: <s>Uh. [NEWLINE] [NEWLINE] /r/Feminism is run by a moderator who has come under fire for allowing MRAs to post there and make the discussion about men. It's actually a huge controversy among Reddit feminists. [NEWLINE] [NEWLINE] The fact that you're honestly claiming that discussing men's issues gets you banned there means that you've never actually visited /r/Feminism and are straight up **lying**.</s>
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0964, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1308, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Might I ask<mask> agnosticism plays into this? I had always thought that agnosticism is believing that gods are beyond our knowledge, and<mask> such do not claim that they know there is not a god.<mask> I thought atheists held the belief that there is no god at all. I'm sorry<mask> I'm misunderstanding,<mask> I am unfortunately rather ignorant on atheism and agnosticism.</s>
Label encoding: <s>Might I ask where agnosticism plays into this? I had always thought that agnosticism is believing that gods are beyond our knowledge, and as such do not claim that they know there is not a god. But I thought atheists held the belief that there is no god at all. I'm sorry if I'm misunderstanding, but I am unfortunately rather ignorant on atheism and agnosticism.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This delta is currently disallowed<mask> your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for<mask> /u/Scribbles_ changed your view.<mask> you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>This delta is currently disallowed as your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for how /u/Scribbles_ changed your view. If you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0938, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>FUCK! Hahaha. Ok. I knew there was a catch. I have NEVER heard a guy say he was totally cool with sharing that was straight. I am sure they exist and am not<mask> crazy to know that there are probably plenty<mask> you have to admit that was funny. [NEWLINE] Sir that is sometimes into other sirs your point still is helping me see the other side for sure :).</s>
Label encoding: <s>FUCK! Hahaha. Ok. I knew there was a catch. I have NEVER heard a guy say he was totally cool with sharing that was straight. I am sure they exist and am not so crazy to know that there are probably plenty but you have to admit that was funny. [NEWLINE] Sir that is sometimes into other sirs your point still is helping me see the other side for sure :).</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0728, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'd<mask><mask> denying yourself of everything in that manner is immoral to yourself. [ENDQ] [NEWLINE] Unless that is<mask> makes you happy. I recommend you read the book Siddhartha by Herman Hesse. It is about a man who gives up everything in order to find the meaning of life, he suffers greatly,<mask> he doesn't regret any of it. <mask> is that immoral to one's self?</s>
Label encoding: <s> [STARTQ] I'd argue that denying yourself of everything in that manner is immoral to yourself. [ENDQ] [NEWLINE] Unless that is what makes you happy. I recommend you read the book Siddhartha by Herman Hesse. It is about a man who gives up everything in order to find the meaning of life, he suffers greatly, but he doesn't regret any of it.  How is that immoral to one's self?</s>
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it's morally consistent to rank animals along a gradient of intelligence, and mark a cutoff above which one refuses to participate in the harvesting of that animal. One could, for example, eat chicken<mask> refuse to eat dolphin, elephant, or primate. [NEWLINE] [NEWLINE] <mask>,<mask> one decides that dogs are a forbidden food on account of intelligence, one ought to reconsider pigs and octopi<mask> well!</s>
Label encoding: <s>I think it's morally consistent to rank animals along a gradient of intelligence, and mark a cutoff above which one refuses to participate in the harvesting of that animal. One could, for example, eat chicken but refuse to eat dolphin, elephant, or primate. [NEWLINE] [NEWLINE] However, if one decides that dogs are a forbidden food on account of intelligence, one ought to reconsider pigs and octopi as well!</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0973, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0980, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> many people who go out of their way to adopt dogs that have been mistreated [ENDQ] [NEWLINE] Well... it's a lot easier to look after a dog than a kid, too. You don't need daycare for a dog. You don't need school supplies or a university fund. It's just not really comparable. Generally more people have the means to adopt a dog than can adopt a kid.</s><pad>
Label encoding: <s> [STARTQ] so many people who go out of their way to adopt dogs that have been mistreated [ENDQ] [NEWLINE] Well... it's a lot easier to look after a dog than a kid, too. You don't need daycare for a dog. You don't need school supplies or a university fund. It's just not really comparable. Generally more people have the means to adopt a dog than can adopt a kid.</s><pad>
Loss: tensor(0.0914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> with the rights/ responsibilities thing, I'm mostly thinking in terms of individuals dealing with eachother, rather than people's responsibilites to the state. [ENDQ] [NEWLINE] That's pretty much the whole of criminal and tort law. We have a very long list of things people aren't allowed to do to each other and a much shorter list of things the state isn't allowed to do to people.</s>
Label encoding: <s> [STARTQ] But with the rights/ responsibilities thing, I'm mostly thinking in terms of individuals dealing with eachother, rather than people's responsibilites to the state. [ENDQ] [NEWLINE] That's pretty much the whole of criminal and tort law. We have a very long list of things people aren't allowed to do to each other and a much shorter list of things the state isn't allowed to do to people.</s>
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1831, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> should he apologize for? Wanting to have sex with a beautiful young girl? [ENDQ] [NEWLINE] [STARTQ] His manner, to be fair, was boorish and very direct, [ENDQ] [NEWLINE] That? I'd agree that most of this outrage is manufactured<mask> people love to feel morally superior<mask> you don't think his behavior was in appropriate at all? I mean being boorish isn't a compliment.</s><pad>
Label encoding: <s> [STARTQ] What should he apologize for? Wanting to have sex with a beautiful young girl? [ENDQ] [NEWLINE] [STARTQ] His manner, to be fair, was boorish and very direct, [ENDQ] [NEWLINE] That? I'd agree that most of this outrage is manufactured because people love to feel morally superior but you don't think his behavior was in appropriate at all? I mean being boorish isn't a compliment.</s><pad>
Loss: tensor(0.0825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Just<mask> our opinions might change later, does that really give others the right to stop us? [ENDQ] [NEWLINE] <mask><mask> this is a really important aspect of the suicide discussion. Many things aren't considered "rational" decisions<mask> we make them and they can and do negatively affect others.<mask>, we don't have laws against making non-rational decisions, nor should we. Suicide is one of those.</s>
Label encoding: <s> [STARTQ] Just because our opinions might change later, does that really give others the right to stop us? [ENDQ] [NEWLINE] I think this is a really important aspect of the suicide discussion. Many things aren't considered "rational" decisions when we make them and they can and do negatively affect others. However, we don't have laws against making non-rational decisions, nor should we. Suicide is one of those.</s>
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0967, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I assume an association between a day and the date<mask> that's, like,<mask> we have been operating for a very long time. [NEWLINE] [NEWLINE] It<mask> helps for things like scheduling, I know that it's April 25 all day, and it'll be April 25 all day. It didn't suddenly become April 26 in the middle of my day, which is good<mask> I have bills due then.</s>
Label encoding: <s>I assume an association between a day and the date because that's, like, how we have been operating for a very long time. [NEWLINE] [NEWLINE] It also helps for things like scheduling, I know that it's April 25 all day, and it'll be April 25 all day. It didn't suddenly become April 26 in the middle of my day, which is good because I have bills due then.</s>
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with that there is no real way to find 100% proof in these scenarios. [NEWLINE] [NEWLINE] <mask>, my personal trail of thought on this topic is that peoples actions are driven by selfishness, and selfishness is a conclusion that is always possible to come to<mask> discussing<mask> motivates someone to do something. [NEWLINE] [NEWLINE] I<mask> believe that being selfish isn't necassarily bad or wrong.</s>
Label encoding: <s>I agree with that there is no real way to find 100% proof in these scenarios. [NEWLINE] [NEWLINE] However, my personal trail of thought on this topic is that peoples actions are driven by selfishness, and selfishness is a conclusion that is always possible to come to when discussing what motivates someone to do something. [NEWLINE] [NEWLINE] I also believe that being selfish isn't necassarily bad or wrong.</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0924, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0990, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Not everyone is wired the same way. [ENDQ] [NEWLINE] I absolutely agree that not everyone is wired the same way.  Some folks have chemical, hormonal or emotional imbalances. <mask><mask> most folks are trained to shy away from this possibility. <mask> our lifestyle grows,<mask><mask> more and more monogamous people are finding that the lifestyle fulfills their needs more than monogamy does.</s><pad>
Label encoding: <s> [STARTQ] Not everyone is wired the same way. [ENDQ] [NEWLINE] I absolutely agree that not everyone is wired the same way.  Some folks have chemical, hormonal or emotional imbalances.  I think most folks are trained to shy away from this possibility.  As our lifestyle grows, I think more and more monogamous people are finding that the lifestyle fulfills their needs more than monogamy does.</s><pad>
Loss: tensor(0.0962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Could a womans inherent fear of being the smaller sex mean they are more likely to avoid putting themselves in situations<mask> violent crime is possible? I'm a guy and I have no problem walking just about anywhere, simply<mask> of<mask> unlikely it is that I'll actually be assaulted. I have female friends who are very careful about walking alone. Wouldn't I be more likely to be a victim?</s>
Label encoding: <s>Could a womans inherent fear of being the smaller sex mean they are more likely to avoid putting themselves in situations where violent crime is possible? I'm a guy and I have no problem walking just about anywhere, simply because of how unlikely it is that I'll actually be assaulted. I have female friends who are very careful about walking alone. Wouldn't I be more likely to be a victim?</s>
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In that case,<mask> do you explain the situation in Japan,<mask> guns are almost completely prohibited and gun violence is practically nonexistent? [NEWLINE] [NEWLINE] Mexican law enforcement is notorious for corruption and ineptitude. Might that have something to do with it? Perhaps we can<mask> point to the fact that Mexico neighbors a country with profoundly lax gun control and an absurdly high number of firearms in general circulation.</s>
Label encoding: <s>In that case, how do you explain the situation in Japan, where guns are almost completely prohibited and gun violence is practically nonexistent? [NEWLINE] [NEWLINE] Mexican law enforcement is notorious for corruption and ineptitude. Might that have something to do with it? Perhaps we can also point to the fact that Mexico neighbors a country with profoundly lax gun control and an absurdly high number of firearms in general circulation.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0952, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>how is capitalism equal to a religion? religion is irrational, not based on reality or open to testing.<mask><mask> capitalism could work,<mask> some other comments on here have made me look at my views a little more critically and i now agree with them. [NEWLINE] [NEWLINE] <mask> you actually wanted to contribute something then your post should include meaningful dialog and points against pure capitalism. you make none of these.</s>
Label encoding: <s>how is capitalism equal to a religion? religion is irrational, not based on reality or open to testing. I think capitalism could work, however some other comments on here have made me look at my views a little more critically and i now agree with them. [NEWLINE] [NEWLINE] if you actually wanted to contribute something then your post should include meaningful dialog and points against pure capitalism. you make none of these.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆. I do see some weaknesses of my proposed idea now. It seems to be especially incompatible with the notion of free speech in the US. Maybe the idea could be useful in a weaker form,<mask> I cannot think of one for now. [NEWLINE] [NEWLINE] Just a question aside:<mask> is it that swearing gets censored on television in the US? Isn’t swearing free speech too?</s>
Label encoding: <s>∆. I do see some weaknesses of my proposed idea now. It seems to be especially incompatible with the notion of free speech in the US. Maybe the idea could be useful in a weaker form, but I cannot think of one for now. [NEWLINE] [NEWLINE] Just a question aside: Why is it that swearing gets censored on television in the US? Isn’t swearing free speech too?</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1001, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And of course the only stupid politicians are supported by the south.  Sheila Jackson Lee, Nancy Pelosi, Diane Feinstein, Joe (foot in mouth) Biden are all from and supported by the south.<mask> you mean is the only stupid politicians are conservative and supported by the south. [NEWLINE] [NEWLINE] <mask><mask> OP needs to have a meet up with [this guy]( [URL] ) (highlighted).</s>
Label encoding: <s>And of course the only stupid politicians are supported by the south.  Sheila Jackson Lee, Nancy Pelosi, Diane Feinstein, Joe (foot in mouth) Biden are all from and supported by the south. What you mean is the only stupid politicians are conservative and supported by the south. [NEWLINE] [NEWLINE] I think OP needs to have a meet up with [this guy]( [URL] ) (highlighted).</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0979, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually I've got some friends who really appreciate trigger warnings<mask> it gives them a "heads up"<mask> they can take a couple of minutes to mentally prepare themselves before reading or watching the content. Or they can choose to avoid it or put it off until they're in a better state of mind or whatever. Just helps you make a more informed decision and not get taken by surprise, basically.</s>
Label encoding: <s>Actually I've got some friends who really appreciate trigger warnings because it gives them a "heads up" so they can take a couple of minutes to mentally prepare themselves before reading or watching the content. Or they can choose to avoid it or put it off until they're in a better state of mind or whatever. Just helps you make a more informed decision and not get taken by surprise, basically.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] People can only work with the information they have. [ENDQ] [NEWLINE] And I'd argue you'd better default to respect than expect people to prove themselves. Especially in any context in which respect and deference would be expected or required,<mask> is the case in a workplace for instance. [NEWLINE] [NEWLINE] Later on, there's plenty of reasons you might cease to respect someone, I don't disagree. </s>
Label encoding: <s> [STARTQ] People can only work with the information they have. [ENDQ] [NEWLINE] And I'd argue you'd better default to respect than expect people to prove themselves. Especially in any context in which respect and deference would be expected or required, as is the case in a workplace for instance. [NEWLINE] [NEWLINE] Later on, there's plenty of reasons you might cease to respect someone, I don't disagree. </s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0885, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this depends on<mask> political views you are talking about.  I am friends with people with a range of beliefs,<mask> there are a few beliefs that are very important to me that I would have trouble overlooking.  For example,<mask> a gay person, I would find it difficult to be close with someone who was actively campaigning to take away my right to get married.  </s>
Label encoding: <s>I think this depends on what political views you are talking about.  I am friends with people with a range of beliefs, but there are a few beliefs that are very important to me that I would have trouble overlooking.  For example, as a gay person, I would find it difficult to be close with someone who was actively campaigning to take away my right to get married.  </s>
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not exactly. [NEWLINE] Bitcoin has plummeted in value a few times already. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Users can have faith in the security and rarity of bitcoins,<mask> those don't equate to value. [NEWLINE] [NEWLINE] For those who bought Bitcoins at the high level, it has already failed them. An 80% dive in any other currency over that sort of time period would practically cripple an economy.</s>
Label encoding: <s>Not exactly. [NEWLINE] Bitcoin has plummeted in value a few times already. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Users can have faith in the security and rarity of bitcoins, but those don't equate to value. [NEWLINE] [NEWLINE] For those who bought Bitcoins at the high level, it has already failed them. An 80% dive in any other currency over that sort of time period would practically cripple an economy.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean,<mask> a base 10 culture,<mask><mask> it makes a huge amount of sense to make 0-100 be "about the coldest it ever gets to about the hottest it ever gets (in many places).   Yes, I see some value in freezing being 0,<mask> I don't think -20 to 40 works<mask> well for describing the normal range of temperature variation.</s>
Label encoding: <s>I mean, as a base 10 culture, I think it makes a huge amount of sense to make 0-100 be "about the coldest it ever gets to about the hottest it ever gets (in many places).   Yes, I see some value in freezing being 0, but I don't think -20 to 40 works as well for describing the normal range of temperature variation.</s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0986, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But in the end, you've still just left us with a profoundly unequal scenario.<mask> is that any better? Honestly,<mask><mask> you're blowing this concern out of proportion. It should not be<mask> difficult in a functional system to put a child up for adoption. Especially<mask> it's not like they don't have 7-9 months to plan for the child/adoption anyways.</s>
Label encoding: <s>But in the end, you've still just left us with a profoundly unequal scenario. How is that any better? Honestly, I think you're blowing this concern out of proportion. It should not be so difficult in a functional system to put a child up for adoption. Especially since it's not like they don't have 7-9 months to plan for the child/adoption anyways.</s>
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0942, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the simplest thing we can do to make the lives of animals better is to *stop breeding them into captivity*. Once we agree to cease breeding creatures for the sole reason that we intend to kill them<mask> they hit adolescence, maybe I'll take welfarist arguments more seriously.<mask> it stands, all you're doing by buying meat is paying people to breed more animals to kill.</s>
Label encoding: <s>I think the simplest thing we can do to make the lives of animals better is to *stop breeding them into captivity*. Once we agree to cease breeding creatures for the sole reason that we intend to kill them when they hit adolescence, maybe I'll take welfarist arguments more seriously. As it stands, all you're doing by buying meat is paying people to breed more animals to kill.</s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am a male and I personally would like my future wife to take my name, not<mask> its the social norm or that I want to impose misogyny on her. I want it<mask> of the logistical convenience, and the fact for me it is another thing that my wife and I can share in, Im sharing my life with her, and I'd like to share my name<mask> well</s>
Label encoding: <s>I am a male and I personally would like my future wife to take my name, not because its the social norm or that I want to impose misogyny on her. I want it because of the logistical convenience, and the fact for me it is another thing that my wife and I can share in, Im sharing my life with her, and I'd like to share my name as well</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0881, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See that's<mask> I made this subreddit, to try and not ostracize myself!<mask> thank you for that advice I'll take it into consideration and maybe tag a long next time and just observe,<mask> I doubt it'll be a common occurrence. [NEWLINE] Hmm thats interesting. I guess I can still stick strongly to my morals and hang out with them! Thank you! ∆</s>
Label encoding: <s>See that's why I made this subreddit, to try and not ostracize myself! But thank you for that advice I'll take it into consideration and maybe tag a long next time and just observe, but I doubt it'll be a common occurrence. [NEWLINE] Hmm thats interesting. I guess I can still stick strongly to my morals and hang out with them! Thank you! ∆</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What would be the difficulties of hiring mercenaries? People join the military all the time and big businesses burn cash to get military contracts. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] <mask>, history has shown taking over the world tends to not work out in the long run,<mask> that's a different argument. I certainly wouldn't say mercenaries are ideal for taking over<mask> a government really wanted to do that.</s>
Label encoding: <s>What would be the difficulties of hiring mercenaries? People join the military all the time and big businesses burn cash to get military contracts. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Also, history has shown taking over the world tends to not work out in the long run, but that's a different argument. I certainly wouldn't say mercenaries are ideal for taking over if a government really wanted to do that.</s>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0819, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2914, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry LinguaManiac, your post has been removed: [NEWLINE] [NEWLINE] &gt; Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5)</s>
Label encoding: <s>Sorry LinguaManiac, your post has been removed: [NEWLINE] [NEWLINE] &gt; Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5)</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course it does.<mask> the US can literally do anything it wants to, preemptively or otherwise,<mask> would it ever trade that away?  It won't. It will leave the UN before it ever loses the ability to unilaterally assess binding resolutions. This is the same for the other empowered member states and you'll be left with a weak organization of no consequence. </s>
Label encoding: <s>Of course it does. But the US can literally do anything it wants to, preemptively or otherwise, why would it ever trade that away?  It won't. It will leave the UN before it ever loses the ability to unilaterally assess binding resolutions. This is the same for the other empowered member states and you'll be left with a weak organization of no consequence. </s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How is it evidence of that? It's designed, in part, to work with the medical realities of terminal illness, which roughly falls into the six-months-or-less range.<mask> does it bolster your position? You seem to be taking it<mask> an obvious or given fact without elaboration or explanation, and I really can't sufficiently address your concerns without either.</s>
Label encoding: <s>How is it evidence of that? It's designed, in part, to work with the medical realities of terminal illness, which roughly falls into the six-months-or-less range. How does it bolster your position? You seem to be taking it as an obvious or given fact without elaboration or explanation, and I really can't sufficiently address your concerns without either.</s>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am 30 and for the past 15 years all I have dated are 18 and 19 year olds. I am shallow I very much like a hot body and pretty face. I have<mask> dated a 27 year old<mask> she was smokin and could pass for a 20 year old. I too could pass for a 23 or 24 year old is maybe that is part of it.</s>
Label encoding: <s>I am 30 and for the past 15 years all I have dated are 18 and 19 year olds. I am shallow I very much like a hot body and pretty face. I have though dated a 27 year old but she was smokin and could pass for a 20 year old. I too could pass for a 23 or 24 year old is maybe that is part of it.</s>
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh, goodness, now I've fallen down the rabbit hole of videos showing<mask> incredibly ignorant the "<mask> money isn't speech!" objection to *Citizens United* really is. [NEWLINE] [NEWLINE] ETA: Here's another good one which demonstrates that ["more money buys more votes" underlying the  (legitimate,<mask> unfounded) concern about the Money in Politics]( [URL] )</s>
Label encoding: <s>Oh, goodness, now I've fallen down the rabbit hole of videos showing how incredibly ignorant the " But money isn't speech!" objection to *Citizens United* really is. [NEWLINE] [NEWLINE] ETA: Here's another good one which demonstrates that ["more money buys more votes" underlying the  (legitimate, if unfounded) concern about the Money in Politics]( [URL] )</s>
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the last thing a victim's parents want is the police throwing them in prison, and the last thing a victimized child needs is to have his parent(s) taken away<mask> they protected him. [NEWLINE] [NEWLINE] Anyone who is willing to sexually assault a child deserves exactly<mask>'s coming to them. That kind of predatory person is not fit to live in civilized society.</s>
Label encoding: <s>I think the last thing a victim's parents want is the police throwing them in prison, and the last thing a victimized child needs is to have his parent(s) taken away because they protected him. [NEWLINE] [NEWLINE] Anyone who is willing to sexually assault a child deserves exactly what's coming to them. That kind of predatory person is not fit to live in civilized society.</s>
Loss: tensor(0.0966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0810, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> that's just far far far too much power to put into the hands of the free market. [ENDQ] [NEWLINE] I don't want *currently* illegal drugs in the free market. I'll refer to them<mask> hard drugs. I don't want hard drugs in the free market, like you said I did. I want them sold and regulated by the government.</s>
Label encoding: <s> [STARTQ] I think that's just far far far too much power to put into the hands of the free market. [ENDQ] [NEWLINE] I don't want *currently* illegal drugs in the free market. I'll refer to them as hard drugs. I don't want hard drugs in the free market, like you said I did. I want them sold and regulated by the government.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Copying their product could be stealing their livelihood. [ENDQ] [NEWLINE] "Could be" or is? [NEWLINE] [NEWLINE] [STARTQ] <mask> the person who discovers a method to cure cancer should be able to profit from that, just like a musician who makes a song. They've essentially made a product [ENDQ] [NEWLINE] Making a product doesn't entitle someone to make a profit, does it? </s>
Label encoding: <s> [STARTQ] Copying their product could be stealing their livelihood. [ENDQ] [NEWLINE] "Could be" or is? [NEWLINE] [NEWLINE] [STARTQ] but the person who discovers a method to cure cancer should be able to profit from that, just like a musician who makes a song. They've essentially made a product [ENDQ] [NEWLINE] Making a product doesn't entitle someone to make a profit, does it? </s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0869, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0976, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0929, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/rbrychckn. [NEWLINE] [NEWLINE] [^rbrychckn's ^delta ^history](/r/ChangeMyView/wiki/user/rbrychckn) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/rbrychckn. [NEWLINE] [NEWLINE] [^rbrychckn's ^delta ^history](/r/ChangeMyView/wiki/user/rbrychckn) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Kolomgorov complexity is the "perfect" way to measure entropy,<mask> I believe it's not computable.  There are ways to do a reasonable job,<mask>. [NEWLINE] [NEWLINE] One way is to compress a dictionary followed by the password; the number of additional bits required to store the password is a good approximation of the entropy contained in the password.</s><pad>
Label encoding: <s>Kolomgorov complexity is the "perfect" way to measure entropy, but I believe it's not computable.  There are ways to do a reasonable job, however. [NEWLINE] [NEWLINE] One way is to compress a dictionary followed by the password; the number of additional bits required to store the password is a good approximation of the entropy contained in the password.</s><pad>
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0773, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0888, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you drive very little,<mask> you feel its too dangerous on the road, [NEWLINE] you deprive yourself of experience behind the wheel. The more a person [NEWLINE] drives and becomes accustomed to different road conditions, the more [NEWLINE] easier it becomes. A seasoned driver can anticipate danger, do whats [NEWLINE] needed to avoid it and remain calm and cool. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>If you drive very little, because you feel its too dangerous on the road, [NEWLINE] you deprive yourself of experience behind the wheel. The more a person [NEWLINE] drives and becomes accustomed to different road conditions, the more [NEWLINE] easier it becomes. A seasoned driver can anticipate danger, do whats [NEWLINE] needed to avoid it and remain calm and cool. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0988, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0900, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As somebody whose first language isn't latin based,<mask><mask> all latin based languages and germanic languages are inferior and flawed.<mask> you know.. same<mask> computer language, the quality is mostly defined by the user base and available libraries. much more<mask> than the inherit design of the language itself. In that regard english is arguably the best. </s>
Label encoding: <s>As somebody whose first language isn't latin based, I think all latin based languages and germanic languages are inferior and flawed. But you know.. same as computer language, the quality is mostly defined by the user base and available libraries. much more so than the inherit design of the language itself. In that regard english is arguably the best. </s>
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I personally think that people would see that their vote might actually get a policy implemented, rather than possibly getting someone elected, who might vote the way they want, and whose vote on that issue might push a certain issue forward, which then may or may get shot down somewhere else.  And seeing that their vote might count, they would do more research.</s>
Label encoding: <s>I personally think that people would see that their vote might actually get a policy implemented, rather than possibly getting someone elected, who might vote the way they want, and whose vote on that issue might push a certain issue forward, which then may or may get shot down somewhere else.  And seeing that their vote might count, they would do more research.</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0982, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you're in America,<mask><mask>.  In my college town at least getting blackout drunk is encouraged and set<mask> the goal for many.  My friends often hope to forget their Friday nights.  Stories about still being drunk in class or being<mask> hungover you don't get out of bed are shared with pride.  It's a problem.</s>
Label encoding: <s>If you're in America, I disagree.  In my college town at least getting blackout drunk is encouraged and set as the goal for many.  My friends often hope to forget their Friday nights.  Stories about still being drunk in class or being so hungover you don't get out of bed are shared with pride.  It's a problem.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0985, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, yes, I'm sure. Let's all get terrified about the scary Russians invading an island 10,000 miles from their capital with their nonexistent Navy against the wishes of the most powerful military and only nuclear power in the world after they just lost 20 million men. It was the Russians, you see! They made us murder all those people! </s>
Label encoding: <s>Yes, yes, I'm sure. Let's all get terrified about the scary Russians invading an island 10,000 miles from their capital with their nonexistent Navy against the wishes of the most powerful military and only nuclear power in the world after they just lost 20 million men. It was the Russians, you see! They made us murder all those people! </s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's just anti-cyberist.  There's no logical difference between the borg goals and your 'human' goals.  After all the human brain would not really be capable of competing with A.I.<mask> maybe that's the Federation's murky little secret:  Communism is a debt society<mask><mask> is the Federation taking from people?</s>
Label encoding: <s>That's just anti-cyberist.  There's no logical difference between the borg goals and your 'human' goals.  After all the human brain would not really be capable of competing with A.I. so maybe that's the Federation's murky little secret:  Communism is a debt society so what is the Federation taking from people?</s>
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you think they'd have invaded Kuwait<mask> our diplomats message was "<mask> you invade, you will be invaded and your regime replaced" instead of one of (at best) ambiguity and (at worst) duplicity? [NEWLINE] [NEWLINE] Our conventional war, in terms of lives lost, seems to have more or less kept pace with<mask> Saddam was doing anyway.</s>
Label encoding: <s>Do you think they'd have invaded Kuwait if our diplomats message was " If you invade, you will be invaded and your regime replaced" instead of one of (at best) ambiguity and (at worst) duplicity? [NEWLINE] [NEWLINE] Our conventional war, in terms of lives lost, seems to have more or less kept pace with what Saddam was doing anyway.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Edit: did I break the rules? He didn't change my view<mask> I'm looking at this<mask> unethical not illegal. I tried changing his view using a legal argument and failed. Not the same<mask> a delta. [ENDQ] [NEWLINE] Not really,<mask> a case could be made that<mask> I changed your view about the law, you should delta that.</s>
Label encoding: <s> [STARTQ] Edit: did I break the rules? He didn't change my view since I'm looking at this as unethical not illegal. I tried changing his view using a legal argument and failed. Not the same as a delta. [ENDQ] [NEWLINE] Not really, though a case could be made that if I changed your view about the law, you should delta that.</s>
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I just don't see<mask> you can feel the two are the same. [ENDQ] [NEWLINE] <mask> you've already made the choice that you don't value your relationship or respect your partner enough to be faithful. [NEWLINE] [NEWLINE] Killing someone is not even close to the same thing, and has been discussed elsewhere in this thread<mask> you wish to read it.</s>
Label encoding: <s> [STARTQ] I just don't see how you can feel the two are the same. [ENDQ] [NEWLINE] Because you've already made the choice that you don't value your relationship or respect your partner enough to be faithful. [NEWLINE] [NEWLINE] Killing someone is not even close to the same thing, and has been discussed elsewhere in this thread if you wish to read it.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you're gay and choose not to act on your feelings you have chosen a loveless life. Is this<mask> you will be happy with<mask> you look back at your life? You have chosen to remove yourself from one of the things that seems to be part of the essence of being human. [NEWLINE] [NEWLINE] Is that really<mask> you want? </s>
Label encoding: <s>If you're gay and choose not to act on your feelings you have chosen a loveless life. Is this what you will be happy with when you look back at your life? You have chosen to remove yourself from one of the things that seems to be part of the essence of being human. [NEWLINE] [NEWLINE] Is that really what you want? </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0934, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You understand that de facto and de jure have roughly the same effect, right? [NEWLINE] [NEWLINE] <mask> a slave that saw his father be whipped to death will have the same objective beef<mask> a child who saw his father burn to death in a lynching. [NEWLINE] [NEWLINE] It's tough to put a statute of limitations on the sins of the father. </s>
Label encoding: <s>You understand that de facto and de jure have roughly the same effect, right? [NEWLINE] [NEWLINE] So a slave that saw his father be whipped to death will have the same objective beef as a child who saw his father burn to death in a lynching. [NEWLINE] [NEWLINE] It's tough to put a statute of limitations on the sins of the father. </s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> someone supports gay rights doesn't mean they are not homophobic. You might be thinking of homophobia<mask> synonymous with bigotry,<mask> I mean homophobic in a literal sense. One could literally have a phobic aversion to intimacy among gay men, even<mask> knowing on a rational level know that it is morally acceptable for them to engage in that behavior. </s>
Label encoding: <s>Just because someone supports gay rights doesn't mean they are not homophobic. You might be thinking of homophobia as synonymous with bigotry, but I mean homophobic in a literal sense. One could literally have a phobic aversion to intimacy among gay men, even while knowing on a rational level know that it is morally acceptable for them to engage in that behavior. </s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think<mask>. We live in the information age. 2000 years ago, information spread was largely by word of mouth, and written history was much more limited (and disputable) than today. Bar some serious event that destroys a large amount of recent media, the day we live in will be much more clearly recorded than 2000 years ago.</s>
Label encoding: <s>I don't think so. We live in the information age. 2000 years ago, information spread was largely by word of mouth, and written history was much more limited (and disputable) than today. Bar some serious event that destroys a large amount of recent media, the day we live in will be much more clearly recorded than 2000 years ago.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0969, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Read my full post.  I am arguing for a de-coupling of stats from math classes.  The actual number-crunching in elementary stats doesn't require more than a middle-school level of math.  (The underlying mechanics are of course much more complex,<mask> inappropriate for a first course in statistics for secondary students.)</s>
Label encoding: <s>Read my full post.  I am arguing for a de-coupling of stats from math classes.  The actual number-crunching in elementary stats doesn't require more than a middle-school level of math.  (The underlying mechanics are of course much more complex, but inappropriate for a first course in statistics for secondary students.)</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0994, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd call it paraphrasing,<mask> sure, I should have been more precise in my quote.<mask><mask> my point still stands<mask>. Do you think "I should be able to marry whoever I want" without any additional nuance is a compelling argument for anything? Do you think it accurately represents the arguments put forth by gay marriage advocates?</s>
Label encoding: <s>I'd call it paraphrasing, but sure, I should have been more precise in my quote. I think my point still stands though. Do you think "I should be able to marry whoever I want" without any additional nuance is a compelling argument for anything? Do you think it accurately represents the arguments put forth by gay marriage advocates?</s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why does everyone pretend motorcycles are "invisible"? They are only "invisible" to people who are paying zero attention to anything to begin with. And<mask> they aren't paying attention to anything, it wouldn't matter<mask> they were walking, biking, driving a car, driving a bus, on a motorcycle, or driving a giraffe.</s>
Label encoding: <s>Why does everyone pretend motorcycles are "invisible"? They are only "invisible" to people who are paying zero attention to anything to begin with. And if they aren't paying attention to anything, it wouldn't matter if they were walking, biking, driving a car, driving a bus, on a motorcycle, or driving a giraffe.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1968, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be clear, I'm not talking about things like spills (adults do that too). I mean that more often than not, parents would just give their kids food and leave them to their own devices. This would cause an unreasonable mess (sauce on the seats, pasta/crackers on the floor and crushed into the carpet).</s>
Label encoding: <s>To be clear, I'm not talking about things like spills (adults do that too). I mean that more often than not, parents would just give their kids food and leave them to their own devices. This would cause an unreasonable mess (sauce on the seats, pasta/crackers on the floor and crushed into the carpet).</s>
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>By the logic of the excuse, these people are not moral agents.  I don't use or buy the excuse.  By the logic of the excuse, the people using the excuse are subhuman beasts. [NEWLINE] [NEWLINE] <mask> we reject this excuse, then we hold them responsible for their actions.  We kill them for their crimes against humanity.</s>
Label encoding: <s>By the logic of the excuse, these people are not moral agents.  I don't use or buy the excuse.  By the logic of the excuse, the people using the excuse are subhuman beasts. [NEWLINE] [NEWLINE] If we reject this excuse, then we hold them responsible for their actions.  We kill them for their crimes against humanity.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What<mask> those "rules and guidelines" infringe on the personal liberties of others? Or<mask><mask> they, by nature, hold back the progress of humanity (usually in amoral ways).<mask> these "rules and guidelines" are taken<mask> fact brought forth by a power beyond that of humanity...<mask> will change it for the better? </s>
Label encoding: <s>What if those "rules and guidelines" infringe on the personal liberties of others? Or what if they, by nature, hold back the progress of humanity (usually in amoral ways). If these "rules and guidelines" are taken as fact brought forth by a power beyond that of humanity... what will change it for the better? </s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0971, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No way, being sheltered for your entire life does not justify overreacting<mask> you are exposed to adversity. [NEWLINE] [NEWLINE] It is warmongering to excuse the wars in Iraq *and* Afghanistan<mask> results of 9/11.  9/11 was a criminal action and should have been dealt with through the criminal justice system.  </s>
Label encoding: <s>No way, being sheltered for your entire life does not justify overreacting when you are exposed to adversity. [NEWLINE] [NEWLINE] It is warmongering to excuse the wars in Iraq *and* Afghanistan as results of 9/11.  9/11 was a criminal action and should have been dealt with through the criminal justice system.  </s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Lets say *all* government funds do go to education, now we have a hugely educated populous, whats stopping them from developing private organizations to fill the void left by the govt? [NEWLINE] [NEWLINE] It would certainly spark some interesting competition in some areas, and arguably capitalism does a better job of keeping prices low than the govt does.</s>
Label encoding: <s>Lets say *all* government funds do go to education, now we have a hugely educated populous, whats stopping them from developing private organizations to fill the void left by the govt? [NEWLINE] [NEWLINE] It would certainly spark some interesting competition in some areas, and arguably capitalism does a better job of keeping prices low than the govt does.</s>
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> profit is defined<mask> schools football and basketball can earn enough to pay for other sports/facilities. I am including donations and profit was a bad word<mask> I just meant break even or self sustaining. I worded poorly<mask><mask> I meant to say is more than a few programs are not costing the school money.</s>
Label encoding: <s>I'm not sure how profit is defined but schools football and basketball can earn enough to pay for other sports/facilities. I am including donations and profit was a bad word because I just meant break even or self sustaining. I worded poorly but what I meant to say is more than a few programs are not costing the school money.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Killing can serve many purposes. [NEWLINE] [NEWLINE] I can kill you for your car.  I can kill you for your insurance policy.  I can kill you to make a couch out of your skin. [NEWLINE] [NEWLINE] It doesn't make the killing more moral. [NEWLINE] [NEWLINE] <mask><mask>, Jefrey Dahmer wouldn't be a criminal.</s>
Label encoding: <s>Killing can serve many purposes. [NEWLINE] [NEWLINE] I can kill you for your car.  I can kill you for your insurance policy.  I can kill you to make a couch out of your skin. [NEWLINE] [NEWLINE] It doesn't make the killing more moral. [NEWLINE] [NEWLINE] If so, Jefrey Dahmer wouldn't be a criminal.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>P.C. sucks, it either makes us all conforming little pussies or make us blind/oblivious to the real suffering out there in the world. The truth needs to be told, one way or another, there is nothing to gain from a sanitized society, it's just denying/lying about who we are.</s>
Label encoding: <s>P.C. sucks, it either makes us all conforming little pussies or make us blind/oblivious to the real suffering out there in the world. The truth needs to be told, one way or another, there is nothing to gain from a sanitized society, it's just denying/lying about who we are.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0997, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well you guys were talking about women not being taken seriously and you said, "<mask><mask><mask> in Europe it hasn't, and will never change". I was just pointing out that obviously it has changed<mask><mask> it hadn't then women wouldn't be voted<mask> leaders,<mask><mask> you say men subconsciously don't take women seriously.</s>
Label encoding: <s>Well you guys were talking about women not being taken seriously and you said, " In my opinion in Europe it hasn't, and will never change". I was just pointing out that obviously it has changed because if it hadn't then women wouldn't be voted as leaders, since as you say men subconsciously don't take women seriously.</s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/huadpe. [NEWLINE] [NEWLINE] [^huadpe's ^delta ^history](/r/ChangeMyView/wiki/user/huadpe) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/huadpe. [NEWLINE] [NEWLINE] [^huadpe's ^delta ^history](/r/ChangeMyView/wiki/user/huadpe) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] will have to cover a large portion of the costs [ENDQ] [NEWLINE] Unaware of this,<mask><mask> true, I like it. [NEWLINE] <mask>, this is the result of criminal activity, not socially unacceptable behavior. [NEWLINE] These are public services that should be applied universally to the public. Penalties should not be incurred outside of criminal behavior.</s>
Label encoding: <s> [STARTQ] will have to cover a large portion of the costs [ENDQ] [NEWLINE] Unaware of this, but if true, I like it. [NEWLINE] However, this is the result of criminal activity, not socially unacceptable behavior. [NEWLINE] These are public services that should be applied universally to the public. Penalties should not be incurred outside of criminal behavior.</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have just shown you<mask> I can be objectively certain of things, through direct revelation from an objective mind who knows everything. Do you accept my answer? [NEWLINE] [NEWLINE] Today you've reached a dead end. Please show<mask> your questions today lead somewhere.<mask> they didn't, it's time for you to post your conclusion. </s>
Label encoding: <s>I have just shown you how I can be objectively certain of things, through direct revelation from an objective mind who knows everything. Do you accept my answer? [NEWLINE] [NEWLINE] Today you've reached a dead end. Please show how your questions today lead somewhere. If they didn't, it's time for you to post your conclusion. </s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The womb *is the uterus*. It's still the place that babies come from, that hasn't changed. [NEWLINE] [NEWLINE] And, the word "belly" in that sense refers to the innards of a person, not any specific place. [NEWLINE] [NEWLINE] The point is that you're totally wrong<mask> refuse to admit it.</s>
Label encoding: <s>The womb *is the uterus*. It's still the place that babies come from, that hasn't changed. [NEWLINE] [NEWLINE] And, the word "belly" in that sense refers to the innards of a person, not any specific place. [NEWLINE] [NEWLINE] The point is that you're totally wrong but refuse to admit it.</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Δ. This gave me the mental image of a mosaic<mask> a metaphor for human existence.<mask> you take out individual tiles one at a time the picture won't change much. At the same time<mask>, there'd be no picture at all<mask> the tiles weren't placed there one by one in the first place. </s>
Label encoding: <s>Δ. This gave me the mental image of a mosaic as a metaphor for human existence. If you take out individual tiles one at a time the picture won't change much. At the same time though, there'd be no picture at all if the tiles weren't placed there one by one in the first place. </s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've actually made a pretty decent argument. I would give you the delta<mask><mask><mask> to your point about risk assessment you were able to demonstrate that women feel the same degree of intensity in desire<mask> men. Another poster has made the argument effectively,<mask> it's anecdotal,<mask> I can't totally rely on it. </s>
Label encoding: <s>You've actually made a pretty decent argument. I would give you the delta if in addition to your point about risk assessment you were able to demonstrate that women feel the same degree of intensity in desire as men. Another poster has made the argument effectively, but it's anecdotal, so I can't totally rely on it. </s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0954, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My annoyance was my reason for posting, yes. It extends to my in-laws and many of my husband's friends. It'a hard to understand...<mask> I am<mask> with you on the reality TV. Maybe that should be another CMV... except I don't think I want my view changed on that. </s>
Label encoding: <s>My annoyance was my reason for posting, yes. It extends to my in-laws and many of my husband's friends. It'a hard to understand... but I am also with you on the reality TV. Maybe that should be another CMV... except I don't think I want my view changed on that. </s>
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I guess I'm ascribing the full laissez-faire position [ENDQ] [NEWLINE] Even in the full anarcho-capitalist system, "slavery contracts" are expressly illegal.  Your ownership of private property begins with yourself and you cannot forfeit it: it's a philosophical bedrock of the political ideology.  </s>
Label encoding: <s> [STARTQ] I guess I'm ascribing the full laissez-faire position [ENDQ] [NEWLINE] Even in the full anarcho-capitalist system, "slavery contracts" are expressly illegal.  Your ownership of private property begins with yourself and you cannot forfeit it: it's a philosophical bedrock of the political ideology.  </s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's perfectly reasonable, it's definitely not for everyone. I'm sure there are some hunters out there that  have some bloodlust, every group has their wackos,<mask> most of the hunters I've met seem to hold a similar opinion. It's only anecdotal,<mask> that's my experience at least. </s>
Label encoding: <s>That's perfectly reasonable, it's definitely not for everyone. I'm sure there are some hunters out there that  have some bloodlust, every group has their wackos, but most of the hunters I've met seem to hold a similar opinion. It's only anecdotal, but that's my experience at least. </s>
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why would a white guy feel the need to use the word outside of rare song reciting times anyway? [NEWLINE] [NEWLINE] It seems like it's more of a 'well they can use it,<mask> can I' thing for the sake of argument. [NEWLINE] [NEWLINE] Call your friend bro/dude.. it's not hard.</s><pad>
Label encoding: <s>Why would a white guy feel the need to use the word outside of rare song reciting times anyway? [NEWLINE] [NEWLINE] It seems like it's more of a 'well they can use it, so can I' thing for the sake of argument. [NEWLINE] [NEWLINE] Call your friend bro/dude.. it's not hard.</s><pad>
Loss: tensor(0.1265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is already done in Finland and there is nothing that shows cops target the rich excessively. A system to have the county or city disincntivize uneven targeting could easily be put in place by monitoring the average fines paid vs the median income of the area and punishing for a clear statistical bias in any direction. </s>
Label encoding: <s>This is already done in Finland and there is nothing that shows cops target the rich excessively. A system to have the county or city disincntivize uneven targeting could easily be put in place by monitoring the average fines paid vs the median income of the area and punishing for a clear statistical bias in any direction. </s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Suppressing their point of view gives the appearance that yours can't win an argument. It makes undecided people wonder whether they might be right. [NEWLINE] [NEWLINE] Reddit is legally entitled to impose censorship<mask> it would hurt their image with their userbase and possibly their own staff, most of whom seem to strongly favor freedom of speech.</s>
Label encoding: <s>Suppressing their point of view gives the appearance that yours can't win an argument. It makes undecided people wonder whether they might be right. [NEWLINE] [NEWLINE] Reddit is legally entitled to impose censorship but it would hurt their image with their userbase and possibly their own staff, most of whom seem to strongly favor freedom of speech.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, he has. That's<mask> makes this argument stupid. People in the streets who use leviticus 18:22 to try and ban gay marriage are stupid.<mask> you want to believe in that<mask> believe in the other shit it says,<mask> most people don't<mask> it's for the tribe of Levi's</s>
Label encoding: <s>Yes, he has. That's what makes this argument stupid. People in the streets who use leviticus 18:22 to try and ban gay marriage are stupid. If you want to believe in that also believe in the other shit it says, but most people don't because it's for the tribe of Levi's</s>
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0958, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People's opinions follow their reactions.  Left wingers (at least<mask> I live) are much more likely to endorse physician assisted suicide<mask> a right, for example, and are actually disgusted by people who would interfere with<mask> they perceive to be a natural right.  (probably overlap with ancaps and libertarians)</s>
Label encoding: <s>People's opinions follow their reactions.  Left wingers (at least where I live) are much more likely to endorse physician assisted suicide as a right, for example, and are actually disgusted by people who would interfere with what they perceive to be a natural right.  (probably overlap with ancaps and libertarians)</s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In the event of a tyrannical government. It's now to the point<mask> whether you're facing off<mask> a civilian versus the military or a civilian versus the police force you're likely to encounter armored targets that can only be damaged by an explosive device: e.g. [APC's]( [URL] /).</s>
Label encoding: <s>In the event of a tyrannical government. It's now to the point where whether you're facing off as a civilian versus the military or a civilian versus the police force you're likely to encounter armored targets that can only be damaged by an explosive device: e.g. [APC's]( [URL] /).</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/championofobscurity. ^[[History](/r/changemyview/wiki/user/championofobscurity)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/championofobscurity. ^[[History](/r/changemyview/wiki/user/championofobscurity)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've come to realize that the word "insane" is a little strong. I should've been more specific. At the very least<mask><mask> this behavior is delusional. Who they prey for is irrelevant. The fact that they're asking the supernatural for something is enough for me to call them delusional. </s>
Label encoding: <s>I've come to realize that the word "insane" is a little strong. I should've been more specific. At the very least I think this behavior is delusional. Who they prey for is irrelevant. The fact that they're asking the supernatural for something is enough for me to call them delusional. </s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/PineappleSlices. ^[[History](/r/changemyview/wiki/user/PineappleSlices)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/PineappleSlices. ^[[History](/r/changemyview/wiki/user/PineappleSlices)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have read your other posts. You seem to be going with the assumption that everyone who pirates wouldn't have paid<mask> they didn't pirate?<mask><mask>,<mask> someone truly would have paid nothing, then there is no immorality in pirating something.<mask> there are many who would have paid. </s>
Label encoding: <s>I have read your other posts. You seem to be going with the assumption that everyone who pirates wouldn't have paid if they didn't pirate? I agree, if someone truly would have paid nothing, then there is no immorality in pirating something. But there are many who would have paid. </s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While<mask><mask> more and more systems tend to run on a democracy, or at the least a democratic republic, could this not be a symptom of some other change? such<mask> the US and UN pushing for democratic governments? [NEWLINE] [NEWLINE] <mask><mask> doesn't a system like the one you suggest encourage nepotism?</s>
Label encoding: <s>While I agree more and more systems tend to run on a democracy, or at the least a democratic republic, could this not be a symptom of some other change? such as the US and UN pushing for democratic governments? [NEWLINE] [NEWLINE] additionally doesn't a system like the one you suggest encourage nepotism?</s>
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with that definition is that,<mask> you say, it places The Lord of the Rings<mask> middle fantasy,<mask> most people would agree it is most definitely high fantasy. [NEWLINE] [NEWLINE] Personally I thought the existence of elves and dwarves was one major component of high fantasy, along with magic of some sort.</s>
Label encoding: <s>The problem with that definition is that, as you say, it places The Lord of the Rings as middle fantasy, while most people would agree it is most definitely high fantasy. [NEWLINE] [NEWLINE] Personally I thought the existence of elves and dwarves was one major component of high fantasy, along with magic of some sort.</s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh, see, I didn't know that. That's pretty cool.<mask> the downside to that is that I'd have to organize all the music I actually like into a playlist<mask> I want to play it on "shuffle" or the Spotify equivalent, which I do quite frequently with iTunes. </s>
Label encoding: <s>Oh, see, I didn't know that. That's pretty cool. But the downside to that is that I'd have to organize all the music I actually like into a playlist if I want to play it on "shuffle" or the Spotify equivalent, which I do quite frequently with iTunes. </s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/RYouNotEntertained. ^[[History](/r/changemyview/wiki/user/RYouNotEntertained)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/RYouNotEntertained. ^[[History](/r/changemyview/wiki/user/RYouNotEntertained)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Tests aren't that hard.  It's reading comprehension, memory recall, and occasional critical thinking. <mask> you can't handle that then you aren't ready for any job tests are supposed to prepare you for. <mask> you think tests are stressful and intimidating, try doing a job interview.</s>
Label encoding: <s>Tests aren't that hard.  It's reading comprehension, memory recall, and occasional critical thinking.  If you can't handle that then you aren't ready for any job tests are supposed to prepare you for.  If you think tests are stressful and intimidating, try doing a job interview.</s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0823, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Rocalyn3d. ^[[History](/r/changemyview/wiki/user/Rocalyn3d)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Rocalyn3d. ^[[History](/r/changemyview/wiki/user/Rocalyn3d)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Gotcha.<mask> I mentioned to someone else I approached this from an American perspective. [NEWLINE] [NEWLINE] I absolutely don't agree with laws against<mask> called 'hate speech'. I abhor racists,<mask> I under no circumstance want to make racism illegal<mask> that would be a major affront to free speech. </s>
Label encoding: <s>Gotcha. As I mentioned to someone else I approached this from an American perspective. [NEWLINE] [NEWLINE] I absolutely don't agree with laws against so called 'hate speech'. I abhor racists, but I under no circumstance want to make racism illegal as that would be a major affront to free speech. </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This award is currently disallowed<mask> your comment doesn't include enough text ([comment rule 4]( [URL] #wiki_rule_4)). Please add an explanation for<mask> /u/cr0kus changed your view. Responding to this comment will cause me to recheck your delta comment.</s>
Label encoding: <s>This award is currently disallowed as your comment doesn't include enough text ([comment rule 4]( [URL] #wiki_rule_4)). Please add an explanation for how /u/cr0kus changed your view. Responding to this comment will cause me to recheck your delta comment.</s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> lumping trans people in with lgb is kinda insulting to trans people. [ENDQ] [NEWLINE] <mask><mask> with you,<mask> just a suggestion: I would steer away from telling groups of people<mask> they should or shouldn't be insulted about.  That in itself can be seen<mask> insulting.</s>
Label encoding: <s> [STARTQ] In fact lumping trans people in with lgb is kinda insulting to trans people. [ENDQ] [NEWLINE] I agree with you, but just a suggestion: I would steer away from telling groups of people what they should or shouldn't be insulted about.  That in itself can be seen as insulting.</s>
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do you feel "freedom" (it would be helpful<mask> you could define this, are we talking about human rights, economic opportunity...?) is a necessary motivation for war? [ENDQ] [NEWLINE] I'm thinking it's<mask> this is a frequent narrative in the US used to defend overseas conflicts.</s>
Label encoding: <s> [STARTQ] Why do you feel "freedom" (it would be helpful if you could define this, are we talking about human rights, economic opportunity...?) is a necessary motivation for war? [ENDQ] [NEWLINE] I'm thinking it's because this is a frequent narrative in the US used to defend overseas conflicts.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/ninjazzy. ^[[History](/r/changemyview/wiki/user/ninjazzy)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/ninjazzy. ^[[History](/r/changemyview/wiki/user/ninjazzy)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/myhairsreddit. ^[[History](/r/changemyview/wiki/user/myhairsreddit)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/myhairsreddit. ^[[History](/r/changemyview/wiki/user/myhairsreddit)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you mean to tell me that a state that has an area of 97,814 sq mi, with a population density of 5.85 people per square mile has substantially less crime than an area with 10,528 people per square mile? Shocking, I never would have guessed.</s>
Label encoding: <s>Do you mean to tell me that a state that has an area of 97,814 sq mi, with a population density of 5.85 people per square mile has substantially less crime than an area with 10,528 people per square mile? Shocking, I never would have guessed.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've thought of that,<mask><mask><mask><mask> calories in to calories out, chickens are still more efficient than a cow. Of course, not eating meat at all is even more efficient,<mask> scarcity of food isn't really<mask>'s causing world hunger,<mask> I suppose that's okay.</s>
Label encoding: <s>I've thought of that, but as far as calories in to calories out, chickens are still more efficient than a cow. Of course, not eating meat at all is even more efficient, but scarcity of food isn't really what's causing world hunger, so I suppose that's okay.</s>
Loss: tensor(0.1176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/IIIBlackhartIII. ^[[History](/r/changemyview/wiki/user/IIIBlackhartIII)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/IIIBlackhartIII. ^[[History](/r/changemyview/wiki/user/IIIBlackhartIII)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0802, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just curious,<mask> do people assume later historians would be *less* accurate about historical figures? <mask> anything, they should be *more* accurate than contemporaries (up to a certain point),<mask> contemporaries would be more likely to be biased by their own personal stake in the events.</s>
Label encoding: <s>Just curious, why do people assume later historians would be *less* accurate about historical figures?  If anything, they should be *more* accurate than contemporaries (up to a certain point), because contemporaries would be more likely to be biased by their own personal stake in the events.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask>? Sex is a natural thing. [NEWLINE] [NEWLINE] <mask> consent is not. That is entirely a social construct. And society has deemed that children cannot give informed consent. [NEWLINE] [NEWLINE] Unless you have concrete, scientific data that proves otherwise, then there is no way that will ever change.</s>
Label encoding: <s>So what? Sex is a natural thing. [NEWLINE] [NEWLINE] But consent is not. That is entirely a social construct. And society has deemed that children cannot give informed consent. [NEWLINE] [NEWLINE] Unless you have concrete, scientific data that proves otherwise, then there is no way that will ever change.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because governments can request that you call a location by the name they have given it. [NEWLINE] [NEWLINE] Peking = Bejing, Istanbul only became the official name in the 20th century. [NEWLINE] [NEWLINE] And Myanmar, the US officially recognizes it<mask> Burma<mask> it wasn't changed democratically. </s>
Label encoding: <s>Because governments can request that you call a location by the name they have given it. [NEWLINE] [NEWLINE] Peking = Bejing, Istanbul only became the official name in the 20th century. [NEWLINE] [NEWLINE] And Myanmar, the US officially recognizes it as Burma because it wasn't changed democratically. </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆. I hadn't considered that there were non-health related reasons for being against GMO. I wish those people would lead the movement instead of the "Eww *GMO* sounds icky" crowd which<mask><mask> is probably the majority of anti-GMOers.</s>
Label encoding: <s>∆. I hadn't considered that there were non-health related reasons for being against GMO. I wish those people would lead the movement instead of the "Eww *GMO* sounds icky" crowd which I think is probably the majority of anti-GMOers.</s>
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If they could choose whether or not they got injured<mask> they wanted the injury(?metaphor implosion) and the driver couldn't then yes this would be similar. [NEWLINE] [NEWLINE] Sorry for replying to such an old comment I just really don't agree with the analogy. </s>
Label encoding: <s>If they could choose whether or not they got injured if they wanted the injury(?metaphor implosion) and the driver couldn't then yes this would be similar. [NEWLINE] [NEWLINE] Sorry for replying to such an old comment I just really don't agree with the analogy. </s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/garnteller. ^[[History](/r/changemyview/wiki/user/garnteller)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/garnteller. ^[[History](/r/changemyview/wiki/user/garnteller)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ummm... "violence *and* intimidation", not "violence or intimidation". [NEWLINE] [NEWLINE] <mask> you can show me feminists that actually use violence to intimidate women into thinking they're oppressed... I'll call them terrorists. [NEWLINE] [NEWLINE] Until then, let's use our conjunctions correctly.</s>
Label encoding: <s>Ummm... "violence *and* intimidation", not "violence or intimidation". [NEWLINE] [NEWLINE] If you can show me feminists that actually use violence to intimidate women into thinking they're oppressed... I'll call them terrorists. [NEWLINE] [NEWLINE] Until then, let's use our conjunctions correctly.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Minossama. ^[[History](/r/changemyview/wiki/user/Minossama)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Minossama. ^[[History](/r/changemyview/wiki/user/Minossama)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Nepene. ^[[History](/r/changemyview/wiki/user/Nepene)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Nepene. ^[[History](/r/changemyview/wiki/user/Nepene)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"I believe that a woman will never be able to go topless without causing a stir unless women's boobs" [NEWLINE] [NEWLINE] Well you can believe this all you want<mask> that doesn't change the FACT that this is already the case in Polynesia. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [URL] </s>
Label encoding: <s>"I believe that a woman will never be able to go topless without causing a stir unless women's boobs" [NEWLINE] [NEWLINE] Well you can believe this all you want but that doesn't change the FACT that this is already the case in Polynesia. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [URL] </s>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Aren't 3D-printed guns extremely fragile, to the point that they fall apart after just a few shots? Plus, 3D printing is ridiculously expensive. People wouldn't be willing to pay hundreds or thousands of dollars for a gun that will fall apart<mask> you fire it.</s><pad>
Label encoding: <s>Aren't 3D-printed guns extremely fragile, to the point that they fall apart after just a few shots? Plus, 3D printing is ridiculously expensive. People wouldn't be willing to pay hundreds or thousands of dollars for a gun that will fall apart when you fire it.</s><pad>
Loss: tensor(0.1341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Darkstrategy. ^[[History](/r/changemyview/wiki/user/Darkstrategy)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Darkstrategy. ^[[History](/r/changemyview/wiki/user/Darkstrategy)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In a perfect world, every child would be brought up and cared for with adequate resources. [ENDQ] [NEWLINE] <mask> do you assume a single mother couldn't do this? Would you be okay with no rights and no support, only<mask> the woman is financially able to support by herself?</s>
Label encoding: <s> [STARTQ] In a perfect world, every child would be brought up and cared for with adequate resources. [ENDQ] [NEWLINE] Why do you assume a single mother couldn't do this? Would you be okay with no rights and no support, only if the woman is financially able to support by herself?</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Any whistleblower endures accusations and lawsuits of lies or fabrication.  It's part of the deal.  This would be the greatest blown whistle ever.  Has any conspiracy fan ever claimed this has happened?  No one out there with insider details, even<mask> a possible fraud?</s>
Label encoding: <s>Any whistleblower endures accusations and lawsuits of lies or fabrication.  It's part of the deal.  This would be the greatest blown whistle ever.  Has any conspiracy fan ever claimed this has happened?  No one out there with insider details, even if a possible fraud?</s>
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're charged only<mask> the police believe a crime has occurred.<mask> it's clearly self-defense they don't take you in, officer or not.<mask> you're getting shot at you have a right to defend yourself with lethal force. This should not depend on your occupation.</s>
Label encoding: <s>You're charged only if the police believe a crime has occurred. If it's clearly self-defense they don't take you in, officer or not. If you're getting shot at you have a right to defend yourself with lethal force. This should not depend on your occupation.</s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/redmako101. ^[[History](/r/changemyview/wiki/user/redmako101)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/redmako101. ^[[History](/r/changemyview/wiki/user/redmako101)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think you know<mask> tenure is.  Tenure is basically a teacher's right to due process before getting fired.  It's basically making sure they just don't get fired out of the blue with no reason given, or at least without a valid reason.</s>
Label encoding: <s>I don't think you know what tenure is.  Tenure is basically a teacher's right to due process before getting fired.  It's basically making sure they just don't get fired out of the blue with no reason given, or at least without a valid reason.</s>
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Did the mother consent to you punching her in the stomach? Hypothetically, of course. [NEWLINE] [NEWLINE] <mask> no, then it is reasonable that the mother intended to carry the baby to term, and is murder. [NEWLINE] [NEWLINE] Edit: somehow managed to post<mask> typing. </s><pad>
Label encoding: <s>Did the mother consent to you punching her in the stomach? Hypothetically, of course. [NEWLINE] [NEWLINE] If no, then it is reasonable that the mother intended to carry the baby to term, and is murder. [NEWLINE] [NEWLINE] Edit: somehow managed to post while typing. </s><pad>
Loss: tensor(0.1478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the place<mask> you go off the rails is<mask> you place a higher value upon material possessions than you do existence.  Animals do not have money.  Would they be better off dead?  Trees do not have money.  Would they better off dead too?</s><pad>
Label encoding: <s>I think the place where you go off the rails is when you place a higher value upon material possessions than you do existence.  Animals do not have money.  Would they be better off dead?  Trees do not have money.  Would they better off dead too?</s><pad>
Loss: tensor(0.1153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0973, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The number of people who need replacement organs due to circumstances outside of their control is definitely not zero and giving them no choice in<mask> then happens to their body is not ethical. Anything making organ donation compulsory is unethical, even<mask> it is just for a select few. </s>
Label encoding: <s>The number of people who need replacement organs due to circumstances outside of their control is definitely not zero and giving them no choice in what then happens to their body is not ethical. Anything making organ donation compulsory is unethical, even if it is just for a select few. </s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm fine,<mask> thank you for your concern. I'm sorry it comes out that way. This is honestly just something I've kind of struggled to understand for a<mask> (<mask> people think that killing is worse) and I'm honestly trying to change my view.</s>
Label encoding: <s>I'm fine, but thank you for your concern. I'm sorry it comes out that way. This is honestly just something I've kind of struggled to understand for a while ( why people think that killing is worse) and I'm honestly trying to change my view.</s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That was meant<mask> a retort to the above poster who argues that society has done nothing more than your offspring to deserve wealth. [ENDQ] [NEWLINE] <mask> society doesn't deserve your wealth. The individuals responsible for ensuring you could accumulate wealth already got paid for<mask> they did.</s>
Label encoding: <s> [STARTQ] That was meant as a retort to the above poster who argues that society has done nothing more than your offspring to deserve wealth. [ENDQ] [NEWLINE] But society doesn't deserve your wealth. The individuals responsible for ensuring you could accumulate wealth already got paid for what they did.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask> we agree that something being natural or unnatural does not make it ethically wrong or right. [ENDQ] [NEWLINE] There are philosophers who believe that something being natural or unnatural does make it ethically wrong or right.  Jean-Jacques Rousseau for instance.</s>
Label encoding: <s> [STARTQ] But I think we agree that something being natural or unnatural does not make it ethically wrong or right. [ENDQ] [NEWLINE] There are philosophers who believe that something being natural or unnatural does make it ethically wrong or right.  Jean-Jacques Rousseau for instance.</s>
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If that's the only response someone gets, rather than support and empathy, sure.<mask><mask> the person isn't understanding<mask> it happened to them, it's not victim-blaming to explain it<mask> they can avoid repeating the action that exposed them to risk.</s>
Label encoding: <s>If that's the only response someone gets, rather than support and empathy, sure. But if the person isn't understanding why it happened to them, it's not victim-blaming to explain it so they can avoid repeating the action that exposed them to risk.</s>
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Depends on<mask> you mean by sexism or racism. For example,<mask> someone who is 80 years old uses the word "colored" to describe a black person, it doesn't mean they are racist<mask> this term once **was** the politically correct term. </s>
Label encoding: <s>Depends on what you mean by sexism or racism. For example, if someone who is 80 years old uses the word "colored" to describe a black person, it doesn't mean they are racist as this term once **was** the politically correct term. </s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Honestly, I really liked the 9th doctor<mask> well, it was more that i didn't like his companion, and he only ever got one. And by comparison I really liked Martha and Donna,<mask> the 10th was generally a better experience for me. </s>
Label encoding: <s>Honestly, I really liked the 9th doctor as well, it was more that i didn't like his companion, and he only ever got one. And by comparison I really liked Martha and Donna, so the 10th was generally a better experience for me. </s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It may have been honest and humble<mask> I don't think it comes across that way. The point isn't that he wasn't honest nor humble<mask> that it doesn't come off that way<mask> presented<mask> I bought 100 hamburgers and here are the extras.</s>
Label encoding: <s>It may have been honest and humble but I don't think it comes across that way. The point isn't that he wasn't honest nor humble but that it doesn't come off that way when presented as I bought 100 hamburgers and here are the extras.</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They believe Gay marriage would be an expansion of government, and that abortion kills a human. They are pro death penalty,<mask> the convict has committed some heinous crime,<mask> an unborn baby has done nothing wrong. Military spending is nessecary for our global power</s>
Label encoding: <s>They believe Gay marriage would be an expansion of government, and that abortion kills a human. They are pro death penalty, because the convict has committed some heinous crime, while an unborn baby has done nothing wrong. Military spending is nessecary for our global power</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [STARTQ] You've got to know that there are some really weird people in the world, and that we should do our best to make it<mask> comfortable<mask> possible for everybody [ENDQ] [NEWLINE] I admire the thought process that lead to this sentence. Thoroughly decent. </s>
Label encoding: <s> [NEWLINE] [STARTQ] You've got to know that there are some really weird people in the world, and that we should do our best to make it as comfortable as possible for everybody [ENDQ] [NEWLINE] I admire the thought process that lead to this sentence. Thoroughly decent. </s>
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not sure<mask> it is the same in the UK,<mask> in Canada you may refuse your ballot. [NEWLINE] [NEWLINE] This is different than spoiling it,<mask> refused ballots are counted and reported separately. It's the formal way of saying "none of the above". </s>
Label encoding: <s>Not sure if it is the same in the UK, but in Canada you may refuse your ballot. [NEWLINE] [NEWLINE] This is different than spoiling it, as refused ballots are counted and reported separately. It's the formal way of saying "none of the above". </s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>you would be surprised. I use to play on competitive teams and we would always have one wealthy kid on the team who wasn't that good<mask> was on the team<mask> his parents paid for our gear, trips, etc. This happens more than you think. </s>
Label encoding: <s>you would be surprised. I use to play on competitive teams and we would always have one wealthy kid on the team who wasn't that good but was on the team because his parents paid for our gear, trips, etc. This happens more than you think. </s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not<mask> the GP said,<mask>.  He said specifically that they are the highest 5 top bad guys in Guantanamo, and hasn't substantiated that. [NEWLINE] [NEWLINE] Being 'important' and being 'the most important' are two very different things.</s>
Label encoding: <s>That's not what the GP said, though.  He said specifically that they are the highest 5 top bad guys in Guantanamo, and hasn't substantiated that. [NEWLINE] [NEWLINE] Being 'important' and being 'the most important' are two very different things.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All I'm saying is that regional gun control laws have relatively zero effect on<mask> easy it is to obtain a gun.<mask> you have the means to buy a gun in your city, you probably have the means to buy a gun in the next city over.</s><pad>
Label encoding: <s>All I'm saying is that regional gun control laws have relatively zero effect on how easy it is to obtain a gun. If you have the means to buy a gun in your city, you probably have the means to buy a gun in the next city over.</s><pad>
Loss: tensor(0.1274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What about people with fixed delusions?  One of the hallmarks of a delusion is that it cannot be changed simply by presenting contradictory evidence; psychiatrists have largely given up on convincing patients that delusions are false and work instead on controlling the impact of those delusions. [NEWLINE] </s>
Label encoding: <s>What about people with fixed delusions?  One of the hallmarks of a delusion is that it cannot be changed simply by presenting contradictory evidence; psychiatrists have largely given up on convincing patients that delusions are false and work instead on controlling the impact of those delusions. [NEWLINE] </s>
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1795, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who said taxes should be fair? Taxes should give the most money to the government with the least amount of burden on the people. Everyone wins except for the people who feel they should be able to exploit the fruits of society without equivalently contributing back to it.</s>
Label encoding: <s>Who said taxes should be fair? Taxes should give the most money to the government with the least amount of burden on the people. Everyone wins except for the people who feel they should be able to exploit the fruits of society without equivalently contributing back to it.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But aren't the odds of doing something stupid<mask> getting drunk far higher than the odds of negative consequences for most other things people do? [NEWLINE] [NEWLINE] <mask> I would posit that doing anything that increases the likelihood of doing something immoral is, in itself, immoral.</s><pad>
Label encoding: <s>But aren't the odds of doing something stupid while getting drunk far higher than the odds of negative consequences for most other things people do? [NEWLINE] [NEWLINE] Also I would posit that doing anything that increases the likelihood of doing something immoral is, in itself, immoral.</s><pad>
Loss: tensor(0.1420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, checked every one of those, and in every case it applies only to motorcycles and mopeds, and in some cases, bicycles. Might want to double-check anything else your neighbor has told you, before you get a ticket for it.</s>
Label encoding: <s>Yeah, checked every one of those, and in every case it applies only to motorcycles and mopeds, and in some cases, bicycles. Might want to double-check anything else your neighbor has told you, before you get a ticket for it.</s>
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is about corporate freedom, not individual freedom.<mask> you don't want to compensate employees<mask> required by law then don't start a business. [ENDQ] [NEWLINE] <mask>,<mask> is the ruling limited to corportations held by 5 or fewer people? [NEWLINE] </s>
Label encoding: <s> [STARTQ] This is about corporate freedom, not individual freedom. If you don't want to compensate employees as required by law then don't start a business. [ENDQ] [NEWLINE] So, why is the ruling limited to corportations held by 5 or fewer people? [NEWLINE] </s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not claiming that the picture should be tagged<mask> NSFW, I'm saying that the suggestion of hiding all NSFW pictures isn't a solution that accomplishes the goal of hiding sexy pictures of women in subreddits that aren't devoted to sexy pictures.</s>
Label encoding: <s>I'm not claiming that the picture should be tagged as NSFW, I'm saying that the suggestion of hiding all NSFW pictures isn't a solution that accomplishes the goal of hiding sexy pictures of women in subreddits that aren't devoted to sexy pictures.</s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Isn't that sacrificing a move slot on all of your pokemon on dumb luck? It's a high-risk, high-reward situation.<mask> is that imbalanced. It'd be like going for it on 4th down every time in football.</s>
Label encoding: <s>Isn't that sacrificing a move slot on all of your pokemon on dumb luck? It's a high-risk, high-reward situation. How is that imbalanced. It'd be like going for it on 4th down every time in football.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I've made a simple speculation about her motives in filing the petition<mask> I didn't couple that with any vitriol at least not that I could tell. [ENDQ] [NEWLINE] No, you coupled it with contempt for the impact on anyone and everyone else involved.</s>
Label encoding: <s> [STARTQ] I've made a simple speculation about her motives in filing the petition but I didn't couple that with any vitriol at least not that I could tell. [ENDQ] [NEWLINE] No, you coupled it with contempt for the impact on anyone and everyone else involved.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agreed. They showed plenty Sansa's reaction and disposition toward the situation. We recognized that she was being robbed of something. We didn't need to directly see her get violated,<mask> we might<mask> well see the horror through someone else's reaction.</s>
Label encoding: <s>Agreed. They showed plenty Sansa's reaction and disposition toward the situation. We recognized that she was being robbed of something. We didn't need to directly see her get violated, so we might as well see the horror through someone else's reaction.</s>
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What? I wasn't talking about that. You said "the youth are much more accepting and tolerant".<mask> maybe<mask> you're not defending "mother britain" with that much zeal anymore.<mask> its not such a big thing anymore. </s>
Label encoding: <s>What? I wasn't talking about that. You said "the youth are much more accepting and tolerant". So maybe because you're not defending "mother britain" with that much zeal anymore. Because its not such a big thing anymore. </s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Only citizens vote in our elections. We had slaves until recently (relatively speaking) and women couldn't vote until very recently (<mask> relatively speaking). [NEWLINE] [NEWLINE] <mask>'s your point? None of that negates the validity of the central idea. </s>
Label encoding: <s>Only citizens vote in our elections. We had slaves until recently (relatively speaking) and women couldn't vote until very recently ( also relatively speaking). [NEWLINE] [NEWLINE] What's your point? None of that negates the validity of the central idea. </s>
Loss: tensor(0.1707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well the father can't provide breast milk. [NEWLINE] [NEWLINE] I believe that the primary purpose of maternity leave is for recuperation and for ensuring the baby's well being.<mask><mask><mask> it is better suited for the mother to be on maternity leave.</s>
Label encoding: <s>Well the father can't provide breast milk. [NEWLINE] [NEWLINE] I believe that the primary purpose of maternity leave is for recuperation and for ensuring the baby's well being. As a result it is better suited for the mother to be on maternity leave.</s>
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The transition from atheist to religion can be quite the same.  I know many will<mask><mask> nobody goes that direction, or that atheism should be (or is) the default,<mask> forcing atheism on someone is almost exactly the same<mask> forcing religion.</s>
Label encoding: <s>The transition from atheist to religion can be quite the same.  I know many will argue that nobody goes that direction, or that atheism should be (or is) the default, but forcing atheism on someone is almost exactly the same as forcing religion.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it is physiologically harder for some people. Much harder. They deserve sympathy and recognition for this. They deserve help in getting it under control and keeping it there, or at least trying.<mask> they should not be told it's ok.</s>
Label encoding: <s>I think it is physiologically harder for some people. Much harder. They deserve sympathy and recognition for this. They deserve help in getting it under control and keeping it there, or at least trying. But they should not be told it's ok.</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't ever do things for the sake of tradition. [NEWLINE] [NEWLINE] All I can really say is that my wife was incredibly happy<mask> I gave her it. [NEWLINE] [NEWLINE] It makes me happy to make her happy. [NEWLINE] [NEWLINE] It was worth it.</s>
Label encoding: <s>I don't ever do things for the sake of tradition. [NEWLINE] [NEWLINE] All I can really say is that my wife was incredibly happy when I gave her it. [NEWLINE] [NEWLINE] It makes me happy to make her happy. [NEWLINE] [NEWLINE] It was worth it.</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just a small correction: Americans learned about the concentration camps in 1942 and it still took the US 2 years to intervene itself and only<mask> it was attacked.<mask> arguably, the news of atrocities doesn't do much good. See Rwanda<mask> another example.</s>
Label encoding: <s>Just a small correction: Americans learned about the concentration camps in 1942 and it still took the US 2 years to intervene itself and only when it was attacked. So arguably, the news of atrocities doesn't do much good. See Rwanda as another example.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Interesting. <mask> do you say to the amazing results of the Success Academy (charter school) in NYC vs the unionized NYC public schools? [NEWLINE] [NEWLINE] <mask>, I see another study showing the exact opposite of your claim: [NEWLINE] [URL] </s>
Label encoding: <s>Interesting.  What do you say to the amazing results of the Success Academy (charter school) in NYC vs the unionized NYC public schools? [NEWLINE] [NEWLINE] Also, I see another study showing the exact opposite of your claim: [NEWLINE] [URL] </s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because,<mask> I made that comment the post was at 0 karma.<mask> could that be<mask> 16 people had voted? 0 + 16  = 0? Something seems off there. I don't know<mask> it is done, cellphone app maybe?</s><pad>
Label encoding: <s>Because, when I made that comment the post was at 0 karma. How could that be when 16 people had voted? 0 + 16  = 0? Something seems off there. I don't know how it is done, cellphone app maybe?</s><pad>
Loss: tensor(0.1362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>lol do they? I guess gun manufactures do nothing for the economy.<mask> deaths by automobiles are acceptable losses? You threw out a statistic and i threw out a statistic. Fact automobiles harm people much more the firearms. Again<mask> are firearms impractical?</s>
Label encoding: <s>lol do they? I guess gun manufactures do nothing for the economy. So deaths by automobiles are acceptable losses? You threw out a statistic and i threw out a statistic. Fact automobiles harm people much more the firearms. Again how are firearms impractical?</s>
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it was illegal.<mask> they would have refused those customers, surely it would have been met with at least the same amount of support. [ENDQ] [NEWLINE] In most states<mask> they operate, [it would have been perfectly legal.]( [URL] )</s>
Label encoding: <s> [STARTQ] Because it was illegal. If they would have refused those customers, surely it would have been met with at least the same amount of support. [ENDQ] [NEWLINE] In most states where they operate, [it would have been perfectly legal.]( [URL] )</s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1342, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Is losing customers worth the fraction of the time it takes to ring in 17 vs 15 items in terms of monetary loss? [NEWLINE] [NEWLINE] [NEWLINE] Absolutely not. This is a terrible idea an angry cashier thinks up, not a viable business tactic.</s>
Label encoding: <s>Is losing customers worth the fraction of the time it takes to ring in 17 vs 15 items in terms of monetary loss? [NEWLINE] [NEWLINE] [NEWLINE] Absolutely not. This is a terrible idea an angry cashier thinks up, not a viable business tactic.</s>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is the right answer.I know there's not a *right* answer,<mask> this is<mask> close<mask> you can get. You don't play a computer. You can make music on it,<mask> you can't play it.</s>
Label encoding: <s>This is the right answer.I know there's not a *right* answer, but this is as close as you can get. You don't play a computer. You can make music on it, but you can't play it.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When the victim of a crime is society then the victim is everyone. Society is made of people.<mask> the increased risk from drunk drivers is detrimental to society,<mask> society involves you and me, we are both victims. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>When the victim of a crime is society then the victim is everyone. Society is made of people. So the increased risk from drunk drivers is detrimental to society, since society involves you and me, we are both victims. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I really like /u/km89's answer and suggestion.  You should definitely consider the boundaries between your rights to parent your child, and the rights of other parents to bring them up with a Santa legend<mask> they choose too. [NEWLINE] </s>
Label encoding: <s>I really like /u/km89's answer and suggestion.  You should definitely consider the boundaries between your rights to parent your child, and the rights of other parents to bring them up with a Santa legend if they choose too. [NEWLINE] </s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks!  It's the one I've enjoyed most,<mask> I can never remember<mask> to spell it!  (only had it a few times - I'm not rich and most restaurants/bars don't carry it in my experience)</s>
Label encoding: <s>Thanks!  It's the one I've enjoyed most, but I can never remember how to spell it!  (only had it a few times - I'm not rich and most restaurants/bars don't carry it in my experience)</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We don't live in ideality, we live in reality. [NEWLINE] [NEWLINE] It's great to have ideals,<mask> important to adapt them<mask> they actually work. [NEWLINE] [NEWLINE] Many ideals have caused great harm<mask> considerations were not made for reality.</s>
Label encoding: <s>We don't live in ideality, we live in reality. [NEWLINE] [NEWLINE] It's great to have ideals, but important to adapt them so they actually work. [NEWLINE] [NEWLINE] Many ideals have caused great harm when considerations were not made for reality.</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Without support from the People, do you really think those individuals could have done<mask> they are remembered for? An entire nation wasn't shaped by a roomful of people, it's shaped by its people, who were willing to follow.</s>
Label encoding: <s>Without support from the People, do you really think those individuals could have done what they are remembered for? An entire nation wasn't shaped by a roomful of people, it's shaped by its people, who were willing to follow.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And<mask> a community's mods became corrupt and/or biased, you'd let people make a new system that is unbiased. People will flock to the unbiased system and everyone will be able to get non-sensationalized news!</s>
Label encoding: <s>And if a community's mods became corrupt and/or biased, you'd let people make a new system that is unbiased. People will flock to the unbiased system and everyone will be able to get non-sensationalized news!</s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can ignore all<mask> the intoxication helmet and the judge dividing his attention from the case. The rest is window dressing. People can present whatever appearance they like<mask><mask><mask> they do their jobs well. I don't judge by appearances.</s>
Label encoding: <s>I can ignore all but the intoxication helmet and the judge dividing his attention from the case. The rest is window dressing. People can present whatever appearance they like as long as they do their jobs well. I don't judge by appearances.</s>
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> Sherpas can and do prosper without the need to climb Everest,<mask> do any Sherpas do it? Presumably,<mask> a person doesn't have to do something, then the reason they do something is that they want to.</s>
Label encoding: <s>So if Sherpas can and do prosper without the need to climb Everest, why do any Sherpas do it? Presumably, if a person doesn't have to do something, then the reason they do something is that they want to.</s>
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In Australia, even<mask> your electorate is a safe seat, whoever you preference first gets more money at the next election to help with advertising etc.<mask> there is the upper house voting on the same day which is counted a little different.</s>
Label encoding: <s>In Australia, even if your electorate is a safe seat, whoever you preference first gets more money at the next election to help with advertising etc. Also there is the upper house voting on the same day which is counted a little different.</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>True,<mask> is that really the equivalent of<mask> happens with 'not all men'? Of course it is sometimes,<mask> I'm struggling to think of times I've seen NAM used to criticise someone's conception of gender. </s>
Label encoding: <s>True, but is that really the equivalent of what happens with 'not all men'? Of course it is sometimes, but I'm struggling to think of times I've seen NAM used to criticise someone's conception of gender. </s>
Loss: tensor(0.0944, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ for you<mask> you summed up the reasons nicely and I had never thought about that silent all the time concept. [NEWLINE] [NEWLINE] You said your phone estimates 600 hours of battery life.<mask> much do you usually end up getting?</s>
Label encoding: <s>∆ for you because you summed up the reasons nicely and I had never thought about that silent all the time concept. [NEWLINE] [NEWLINE] You said your phone estimates 600 hours of battery life. How much do you usually end up getting?</s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Then don't take them seriously, just ignore them. [NEWLINE] Just saying "You're wrong" achieves nothing at all. [NEWLINE] You won't change their view, you won't change anybody else's view...<mask>'s the point?</s>
Label encoding: <s>Then don't take them seriously, just ignore them. [NEWLINE] Just saying "You're wrong" achieves nothing at all. [NEWLINE] You won't change their view, you won't change anybody else's view... what's the point?</s>
Loss: tensor(0.1358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Citing your observations isn't actual evidence, which makes your argument on the subject not really any more valid than an unfounded opinion.<mask> asking for evidence is "wearing blinders", I suggest you should learn<mask> to debate.</s>
Label encoding: <s>Citing your observations isn't actual evidence, which makes your argument on the subject not really any more valid than an unfounded opinion. If asking for evidence is "wearing blinders", I suggest you should learn how to debate.</s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's fair enough. And<mask> the person with the allergy wants to eat at home,<mask><mask> you mentioned that you would go out and buy new food for them? [NEWLINE] [NEWLINE] <mask><mask>,<mask><mask> that's perfectly reasonable</s>
Label encoding: <s>That's fair enough. And if the person with the allergy wants to eat at home, I think you mentioned that you would go out and buy new food for them? [NEWLINE] [NEWLINE] If so, I think that's perfectly reasonable</s>
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1855, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I've heard that<mask>. That it's a bad idea to go out dressing a certain way. [ENDQ] [NEWLINE] then doesn't that mean it is consistent with "it's a bad idea to not wear your seatbelt"?</s>
Label encoding: <s> [STARTQ] I've heard that though. That it's a bad idea to go out dressing a certain way. [ENDQ] [NEWLINE] then doesn't that mean it is consistent with "it's a bad idea to not wear your seatbelt"?</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I sympathize with whatever you have experienced.<mask>, intersectionality and sensitivity to trans issues is a huge part of mainstream third wave feminism. Whoever would give you any trouble for being who you are is not worthy of consideration.</s>
Label encoding: <s>I sympathize with whatever you have experienced. However, intersectionality and sensitivity to trans issues is a huge part of mainstream third wave feminism. Whoever would give you any trouble for being who you are is not worthy of consideration.</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that everyone has a right to the truth.  It is an unfortunate circumstance that we're in,<mask> sometimes doing nothing is better than doing something wrong.  Thanks for the delta and for keeping an open mind!</s>
Label encoding: <s>I agree that everyone has a right to the truth.  It is an unfortunate circumstance that we're in, but sometimes doing nothing is better than doing something wrong.  Thanks for the delta and for keeping an open mind!</s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was just attempting to point out that there is no scientific consensus that conflicts with the belief/feeling that God exists.<mask> you wouldn't need to trust your own senses *over* scientific consensus. None exists. </s>
Label encoding: <s>I was just attempting to point out that there is no scientific consensus that conflicts with the belief/feeling that God exists. So you wouldn't need to trust your own senses *over* scientific consensus. None exists. </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> there is already a proposed plan to divide California into 6 regions [ENDQ] [NEWLINE]...advocated mostly by rich companies located on the west coast who don't want to see their tax money go to the poor inland regions.</s><pad>
Label encoding: <s> [STARTQ] though there is already a proposed plan to divide California into 6 regions [ENDQ] [NEWLINE]...advocated mostly by rich companies located on the west coast who don't want to see their tax money go to the poor inland regions.</s><pad>
Loss: tensor(0.1431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1456, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Men will always have choice with their gender roles. Women cant choose not to be the primary victims of rape, they cant choose to have more political and social power (aka choose to have more women controlling politics and industry).</s>
Label encoding: <s>Men will always have choice with their gender roles. Women cant choose not to be the primary victims of rape, they cant choose to have more political and social power (aka choose to have more women controlling politics and industry).</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yep, that's actually exactly<mask> I meant.<mask> that's still a huge jump to make. "We can't compute<mask> someone will do" -&gt; "that person can decide<mask> they will do."</s>
Label encoding: <s>Yep, that's actually exactly what I meant. But that's still a huge jump to make. "We can't compute what someone will do" -&gt; "that person can decide what they will do."</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I never disagreed with you (except about demand, I meant the demand for female players would be artificially raised by artificially high salaries creating artificially high incentives, I'm aware this isn't a "correct" definition of demand"</s>
Label encoding: <s>I never disagreed with you (except about demand, I meant the demand for female players would be artificially raised by artificially high salaries creating artificially high incentives, I'm aware this isn't a "correct" definition of demand"</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that being vegan<mask> opposed to vegetarian isn't much more of a sacrifice, relatively speaking,<mask> constantly giving your assets to the poor and sacrificing all of your time to better the world is much more unreasonable.</s>
Label encoding: <s>I think that being vegan as opposed to vegetarian isn't much more of a sacrifice, relatively speaking, while constantly giving your assets to the poor and sacrificing all of your time to better the world is much more unreasonable.</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1420, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everything said here is correct.  I can't fathom a situation<mask> we would have been told to kill someone without knowing<mask> ; or even a situation<mask> we would've been told specifically to kill someone. [NEWLINE] </s>
Label encoding: <s>Everything said here is correct.  I can't fathom a situation where we would have been told to kill someone without knowing why ; or even a situation where we would've been told specifically to kill someone. [NEWLINE] </s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Conscious Rap, Mos Def and Common.<mask> about the lyrical poetry of Rakim vs. Kanye? These are the artists who raise the bar. Defending Kanye is like saying Transformers is your favorite movie.</s>
Label encoding: <s>Conscious Rap, Mos Def and Common. How about the lyrical poetry of Rakim vs. Kanye? These are the artists who raise the bar. Defending Kanye is like saying Transformers is your favorite movie.</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>well wasn't the garden of eden without sin before the apple eating? [NEWLINE] [NEWLINE] <mask>,<mask> god didn't have the ability to create a world without sin<mask> with free will then he's not omnipotent,</s>
Label encoding: <s>well wasn't the garden of eden without sin before the apple eating? [NEWLINE] [NEWLINE] also, if god didn't have the ability to create a world without sin but with free will then he's not omnipotent,</s>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1393, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't get<mask> folks would badmouth Ringo's drumming, it fits really well with their whole sound.  Crazy, flashy drumming would do lots of their songs a disservice.  </s>
Label encoding: <s>I don't get why folks would badmouth Ringo's drumming, it fits really well with their whole sound.  Crazy, flashy drumming would do lots of their songs a disservice.  </s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2898, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Never say never,<mask> I don't think a clickbait advertising emporium based on exploiting the free labor of the writers is going to be<mask> I'd expect to see the Spartacus League.</s>
Label encoding: <s>Never say never, but I don't think a clickbait advertising emporium based on exploiting the free labor of the writers is going to be where I'd expect to see the Spartacus League.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like the genuine simplicity of your argument. Most people see change<mask> a bad sign in general<mask><mask>. [NEWLINE] [NEWLINE] <mask> I don't, again, "in general". I like change in general.</s>
Label encoding: <s>I like the genuine simplicity of your argument. Most people see change as a bad sign in general in fact. [NEWLINE] [NEWLINE] But I don't, again, "in general". I like change in general.</s>
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree,<mask> I would word it<mask> 'was used<mask> an impetus'. Disaster Capitalism by Naomi Klein is an excellent book detailing<mask> the US gov often pushes policy forward in the wake of trauma.</s><pad>
Label encoding: <s>I agree, but I would word it as 'was used as an impetus'. Disaster Capitalism by Naomi Klein is an excellent book detailing how the US gov often pushes policy forward in the wake of trauma.</s><pad>
Loss: tensor(0.1363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you don't want to have a discussion with this person, or you feel like you're having the same discussion else<mask>, then don't reply to this comment. That's politeness 101.</s>
Label encoding: <s>If you don't want to have a discussion with this person, or you feel like you're having the same discussion else where, then don't reply to this comment. That's politeness 101.</s>
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1740, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If a role is broken repeatedly and in many ways, it is clearly not<mask> much a rule<mask> an explanation, one that no longer adequately explains our knowledge and is<mask> not serving its necessary function.</s>
Label encoding: <s>If a role is broken repeatedly and in many ways, it is clearly not so much a rule as an explanation, one that no longer adequately explains our knowledge and is therefore not serving its necessary function.</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Devil's advocate: Could she have felt that all men are to blame from the perspective of policing their own gender? From this perspective she may have not seen it<mask> a generalization? </s>
Label encoding: <s>Devil's advocate: Could she have felt that all men are to blame from the perspective of policing their own gender? From this perspective she may have not seen it as a generalization? </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's there on purpose<mask> they don't seek to close it - there are lawsuits in existence<mask> rapists ARE seeking custody. I just was unable to find an instance<mask> they'd won. </s>
Label encoding: <s>It's there on purpose if they don't seek to close it - there are lawsuits in existence where rapists ARE seeking custody. I just was unable to find an instance where they'd won. </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right, and I'm absolutely not saying that no one should identify<mask> African-American. All I was saying is that using it<mask> a synonym for black won't always work out. </s>
Label encoding: <s>Right, and I'm absolutely not saying that no one should identify as African-American. All I was saying is that using it as a synonym for black won't always work out. </s>
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If he took the Ring at all it would've corrupted him.<mask> even<mask> Elrond had attacked Isildur, the men and family in his army likely would have killed him.</s>
Label encoding: <s>If he took the Ring at all it would've corrupted him. But even if Elrond had attacked Isildur, the men and family in his army likely would have killed him.</s>
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/cyanoacrylate. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/cyanoacrylate. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Boogie went "it's my problem, and I'm the only one that can change this." [NEWLINE] [NEWLINE] The mod replied: "quit making excuses." [NEWLINE] [NEWLINE] Well...okay...</s><pad>
Label encoding: <s>Boogie went "it's my problem, and I'm the only one that can change this." [NEWLINE] [NEWLINE] The mod replied: "quit making excuses." [NEWLINE] [NEWLINE] Well...okay...</s><pad>
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1673, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Perhaps I should have written that every utopian ideology has failed in the real world<mask> far, except for democracy?  And that democracy succeeded by admitting certain impurities that are extremely helpful.</s>
Label encoding: <s>Perhaps I should have written that every utopian ideology has failed in the real world so far, except for democracy?  And that democracy succeeded by admitting certain impurities that are extremely helpful.</s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OK, that's a fair point. I wasn't thinking of time-manner-place restrictions, I was thinking of content restrictions, and you're right to point it out.</s><pad>
Label encoding: <s>OK, that's a fair point. I wasn't thinking of time-manner-place restrictions, I was thinking of content restrictions, and you're right to point it out.</s><pad>
Loss: tensor(0.1813, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is true,<mask> you can cause more good with a (very) large inefficient system than with a handful of people working efficiently.  I personally support both whole-heartedly.</s>
Label encoding: <s>This is true, but you can cause more good with a (very) large inefficient system than with a handful of people working efficiently.  I personally support both whole-heartedly.</s>
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can't tell is we're still joking... we *were* joking, right?... ya know, cause cats aren't commonly known for their insight into the meaning of life</s>
Label encoding: <s>I can't tell is we're still joking... we *were* joking, right?... ya know, cause cats aren't commonly known for their insight into the meaning of life</s>
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think it makes sense to try to apply<mask> I'm talking about with sexuality to arguments about race. They are different things with different sets of historical discrimination and issues.</s>
Label encoding: <s>I don't think it makes sense to try to apply what I'm talking about with sexuality to arguments about race. They are different things with different sets of historical discrimination and issues.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just for a frame of reference, could you give two examples? [NEWLINE] Something that you feel takes a similar amount of skill to photography, and something that takes large amount of skill.</s>
Label encoding: <s>Just for a frame of reference, could you give two examples? [NEWLINE] Something that you feel takes a similar amount of skill to photography, and something that takes large amount of skill.</s>
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's easy to assume you don't need to shoot<mask> the assumption is that the suspect doesn't have a gun. [NEWLINE] [NEWLINE] That's clearly not the case in the US.</s>
Label encoding: <s>It's easy to assume you don't need to shoot when the assumption is that the suspect doesn't have a gun. [NEWLINE] [NEWLINE] That's clearly not the case in the US.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you read through a religious book that people follow and think these parts would negatively affect other people<mask> someone followed the teaching inside to the letter<mask><mask> it is problematic. </s>
Label encoding: <s>If you read through a religious book that people follow and think these parts would negatively affect other people if someone followed the teaching inside to the letter I think it is problematic. </s>
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe, maybe not; money now is worth more than money in the future,<mask> some things that you would spend your money on are time sensitive (eg, car repairs)</s>
Label encoding: <s>Maybe, maybe not; money now is worth more than money in the future, because some things that you would spend your money on are time sensitive (eg, car repairs)</s>
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most politicians and voters are much more moderate and pragmatic. We'd all like to see our policies implemented,<mask><mask> would like to adhere to our democratic principles<mask> well. </s>
Label encoding: <s>Most politicians and voters are much more moderate and pragmatic. We'd all like to see our policies implemented, but also would like to adhere to our democratic principles as well. </s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1706, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In Infinite Crisis it took a handful of speedsters to get Superboy Prime into the speed force and Wally West wasn't capable of staying to the end of that fight.</s>
Label encoding: <s>In Infinite Crisis it took a handful of speedsters to get Superboy Prime into the speed force and Wally West wasn't capable of staying to the end of that fight.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1867, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We absolutely can can police other people's choices [ENDQ] [NEWLINE] You misquoted me. I said we can't police the *reasoning* behind people's choices.</s>
Label encoding: <s> [STARTQ] We absolutely can can police other people's choices [ENDQ] [NEWLINE] You misquoted me. I said we can't police the *reasoning* behind people's choices.</s>
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Perhaps simple preference isn't appropriate for a CMV topic [ENDQ] [NEWLINE] Its not. This has been mentioned here dozens of times and is in the sidebar i believe. </s>
Label encoding: <s> [STARTQ] Perhaps simple preference isn't appropriate for a CMV topic [ENDQ] [NEWLINE] Its not. This has been mentioned here dozens of times and is in the sidebar i believe. </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you<mask> just a quick point.<mask><mask><mask> I know there was a black version of Nick Fury in the comics prior to the MCU version. </s>
Label encoding: <s>I agree with you but just a quick point. As far as I know there was a black version of Nick Fury in the comics prior to the MCU version. </s>
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1791, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1900, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hmm I was picking stupid activities with an inherent social value in order to form a more accurate comparison to OP. And secondhand smoke can kill you realize. </s>
Label encoding: <s>Hmm I was picking stupid activities with an inherent social value in order to form a more accurate comparison to OP. And secondhand smoke can kill you realize. </s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&gt;can be interpreted<mask> a sports picture **by anyone underexposed/asexual enough** to think the picture was actually about speed skating.</s>
Label encoding: <s>&gt;can be interpreted as a sports picture **by anyone underexposed/asexual enough** to think the picture was actually about speed skating.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1748, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1962, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A girlfriend, yes. Significant other, there's really only one that kind of qualifies. <mask>, I am currently in love with someone that's taken.</s>
Label encoding: <s>A girlfriend, yes. Significant other, there's really only one that kind of qualifies.  Though, I am currently in love with someone that's taken.</s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am Mormon (branch from mainstream Christianity in a way). Do you want me to continue with our doctrine? I would be happy to<mask> you wish :)</s>
Label encoding: <s>I am Mormon (branch from mainstream Christianity in a way). Do you want me to continue with our doctrine? I would be happy to if you wish :)</s>
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1919, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You have a gross misunderstanding of quantum mechanics. Nothing about QM suggests that consciousness is some special thing. You have been deluded by proponents of quantum woo.</s>
Label encoding: <s>You have a gross misunderstanding of quantum mechanics. Nothing about QM suggests that consciousness is some special thing. You have been deluded by proponents of quantum woo.</s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yep I see<mask> you're saying, I just wanted to point out the amount of stochasticity that goes into getting to the top of a chart</s>
Label encoding: <s>Yep I see what you're saying, I just wanted to point out the amount of stochasticity that goes into getting to the top of a chart</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Shia LaBeouf: I was raped [NEWLINE] Posted! [NEWLINE] [NEWLINE] A link has been posted to your Facebook feed." [NEWLINE] [NEWLINE] Link again please?</s>
Label encoding: <s>"Shia LaBeouf: I was raped [NEWLINE] Posted! [NEWLINE] [NEWLINE] A link has been posted to your Facebook feed." [NEWLINE] [NEWLINE] Link again please?</s>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1953, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1759, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1990, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1883, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1857, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1868, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1929, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1875, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1891, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not to mention google is worth almost 6 times<mask> much<mask> facebook. <mask> you assume either money or information is power then google wins vs facebook.</s>
Label encoding: <s>Not to mention google is worth almost 6 times as much as facebook.  If you assume either money or information is power then google wins vs facebook.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1852, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1882, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1955, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1835, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1820, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1844, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1838, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1988, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1948, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0848, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is impossible to defend against. An invasion would never be an overt takeover. It would be subtle, slow, and imperceptible.</s>
Label encoding: <s>It is impossible to defend against. An invasion would never be an overt takeover. It would be subtle, slow, and imperceptible.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1896, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1974, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1828, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1943, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1972, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1991, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1987, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The MTV VMA's spoof of that with Will Ferrell was fantastic. [NEWLINE] [NEWLINE] "VIS A VIS! CONCORDANTLY!"</s>
Label encoding: <s>The MTV VMA's spoof of that with Will Ferrell was fantastic. [NEWLINE] [NEWLINE] "VIS A VIS! CONCORDANTLY!"</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1939, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1954, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1929, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1912, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The WNBA doesn't.  Almost every D1 basketball player can dunk.  Almost no WNBA players can even dunk the ball</s><pad>
Label encoding: <s>The WNBA doesn't.  Almost every D1 basketball player can dunk.  Almost no WNBA players can even dunk the ball</s><pad>
Loss: tensor(0.2029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes,<mask> odd an opinion would be on /r/changemyview. Usually isn't very many opinions here. /s</s>
Label encoding: <s>Yes, how odd an opinion would be on /r/changemyview. Usually isn't very many opinions here. /s</s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1992, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1970, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1961, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why does this need to be ensured? Shouldn't an educated electorate be able to decide the most capable candidate for themselves?</s><pad>
Label encoding: <s>Why does this need to be ensured? Shouldn't an educated electorate be able to decide the most capable candidate for themselves?</s><pad>
Loss: tensor(0.2102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Considering that Latin is a dead language, it's usually much more important to read it and write it than to speak it.</s>
Label encoding: <s>Considering that Latin is a dead language, it's usually much more important to read it and write it than to speak it.</s>
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think I'm just about the nerdiest person in my class,<mask> I'm<mask> the social nexus of said class.</s>
Label encoding: <s>I think I'm just about the nerdiest person in my class, but I'm also the social nexus of said class.</s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2000, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In States without Romeo-Juliet laws (California) or<mask> they cross state borders they will and have been prosecuted. </s>
Label encoding: <s>In States without Romeo-Juliet laws (California) or if they cross state borders they will and have been prosecuted. </s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1978, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Depends on the subculture,<mask> yes, some can.<mask> I don't see<mask> that contradicts my point. </s>
Label encoding: <s>Depends on the subculture, but yes, some can. But I don't see how that contradicts my point. </s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People who I see stumble over a mobile OS has just<mask> much trouble on android<mask> iphones in my experience.</s>
Label encoding: <s>People who I see stumble over a mobile OS has just as much trouble on android as iphones in my experience.</s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It doesn't have to serve a purpose for the view to exist. It's not something I set out to believe.</s>
Label encoding: <s>It doesn't have to serve a purpose for the view to exist. It's not something I set out to believe.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's fair.<mask> we're in agreement- either no adblock, or be forced to pay something, yes?</s>
Label encoding: <s>That's fair. So we're in agreement- either no adblock, or be forced to pay something, yes?</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3985, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You can join a club to watch endless Broadway plays and productions for dirt cheap [ENDQ] [NEWLINE] Could you elaborate, please?</s>
Label encoding: <s> [STARTQ] You can join a club to watch endless Broadway plays and productions for dirt cheap [ENDQ] [NEWLINE] Could you elaborate, please?</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For all of those foods, latex gloves are superior to chopsticks. Unless you have a latex allergy, I guess.</s><pad>
Label encoding: <s>For all of those foods, latex gloves are superior to chopsticks. Unless you have a latex allergy, I guess.</s><pad>
Loss: tensor(0.2164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I live in Singapore.<mask> the worst global warming can do is make groceries expensive, I am ok with that.</s>
Label encoding: <s>I live in Singapore. If the worst global warming can do is make groceries expensive, I am ok with that.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right,<mask> the system was built with the idea that not everyone would collect and not collect for long. </s>
Label encoding: <s>Right, but the system was built with the idea that not everyone would collect and not collect for long. </s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1960, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I am currently leaning toward a vocational school.<mask> is the benefit  in going towards a university?</s>
Label encoding: <s>Well I am currently leaning toward a vocational school. What is the benefit  in going towards a university?</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Or<mask> not leave the lid down?  Then everyone has to do something and it looks much better.</s><pad>
Label encoding: <s>Or why not leave the lid down?  Then everyone has to do something and it looks much better.</s><pad>
Loss: tensor(0.2460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2292, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See my other reply. I'm pretty sure there are things I listed that have never been explained. </s>
Label encoding: <s>See my other reply. I'm pretty sure there are things I listed that have never been explained. </s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Streaming music uses data, which is extremely expensive. Local storage is essential until data prices go down</s>
Label encoding: <s>Streaming music uses data, which is extremely expensive. Local storage is essential until data prices go down</s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You ignore their content and vote just based upon the beginning of their comment? That's not cool</s>
Label encoding: <s>You ignore their content and vote just based upon the beginning of their comment? That's not cool</s>
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Where are you making up open relationships from? [NEWLINE] [NEWLINE] I said men having mistresses.</s>
Label encoding: <s>Where are you making up open relationships from? [NEWLINE] [NEWLINE] I said men having mistresses.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your perception of morality would differ radically from mine. Herein lies the dilemma of politics.</s><pad>
Label encoding: <s>Your perception of morality would differ radically from mine. Herein lies the dilemma of politics.</s><pad>
Loss: tensor(0.2770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why is potential for abuse a reason to deny two consenting adults the right to marry?</s>
Label encoding: <s>Why is potential for abuse a reason to deny two consenting adults the right to marry?</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Cookout [ENDQ] [NEWLINE] Hey everybody, we got a southerner right here!</s>
Label encoding: <s> [STARTQ] Cookout [ENDQ] [NEWLINE] Hey everybody, we got a southerner right here!</s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2792, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2818, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2825, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2836, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2950, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2903, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2983, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2785, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2866, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2910, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2916, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0735, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2966, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2940, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2975, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2947, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2949, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2932, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0995, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What do personal goals derived from selfish intent have to do with morality?</s>
Label encoding: <s>What do personal goals derived from selfish intent have to do with morality?</s>
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know two in their 90s who are ready to go.</s>
Label encoding: <s>I know two in their 90s who are ready to go.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed - 1 delta awarded to /u/jennerality</s>
Label encoding: <s>Confirmed - 1 delta awarded to /u/jennerality</s>
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3834, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3779, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6859, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, you can,<mask> you won't win.</s>
Label encoding: <s>Well, you can, but you won't win.</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3933, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And it's compulsory for some reason. </s>
Label encoding: <s>And it's compulsory for some reason. </s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wait....Political ads are from public funds?</s>
Label encoding: <s>Wait....Political ads are from public funds?</s>
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except not. Look at his subsequent comment.</s>
Label encoding: <s>Except not. Look at his subsequent comment.</s>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2860, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4724, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4746, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.4776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(1.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.5783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am the liquor Randy</s><pad>
Label encoding: <s>I am the liquor Randy</s><pad>
Loss: tensor(0.6591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.6447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.7332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.7080e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.0272e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.4695e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.8483e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.4717e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.0021e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
--------------EPOCH 2-------------
Test Accuracy: tensor(0.7568, device='cuda:0')
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Saving model at iteration: 0
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You have a lot of various issues in your post OP,<mask> I'd like to hit the major ones you have mentioned.<mask> a white 30 year old man, I can assure you things aren't easy for us either. It's not easy for anyone, and each race, sex, sexual orientation, etc, has it's own challenges and difficulties in life. [NEWLINE] [NEWLINE] [STARTQ] Nobody is afraid of women.<mask> I go out late at night, there has to be a trusted man with me. There are rapists at night, or<mask> I get into an accident, who's going to change my tire?<mask> I want to go travel extensively in a foreign country, especially the considered "third world" country, it's too dangerous. I am a woman, and I am American, and everyone knows<mask> to to take advantage of that. [ENDQ] [NEWLINE] <mask> a man in my thirties,<mask> I were to interact with a woman I didn't know either at a bar, or the supermarket, or at work there's immediately the assumption that I am simply trying to "get with them". A guy cannot simply be nice to a woman without there being some preconceived notion that it's an attempt at wooing the particular individual. Likewise,<mask> I attempt to help someone out of the goodness of my heart say<mask> they have a broken down on the side of the road or they are visibly lost in a city at night, there's the assumption that I am simply going to try and harm them. People are scared of men. It doesn't matter for race or age,<mask> you're a guy and you're interacting with a woman who you don't know, there's a danger sense women put forth which is entirely unfair, it makes sense,<mask> it's still unfair. [NEWLINE] [NEWLINE] [STARTQ] <mask> I sleep around, I am a slut.<mask> a man sleeps around, he is a stud. No one gets very angry<mask> a man abandons the baby he fathered.<mask>,<mask> I a woman knows that she won't be able to handle this baby, or<mask> it was born out of rape, there are plenty of states<mask> I wouldn't be able to abort it.<mask> I even have sex with one person before they commit their entire being to me (aka: marriage), it's a risk.<mask><mask> a mistake happens and I get pregnant? [ENDQ] [NEWLINE] There is one issue in regards to sex and men, and that issue is that men have to be the sex fiends that most see us<mask> being. We
Label encoding: <s>You have a lot of various issues in your post OP, but I'd like to hit the major ones you have mentioned. As a white 30 year old man, I can assure you things aren't easy for us either. It's not easy for anyone, and each race, sex, sexual orientation, etc, has it's own challenges and difficulties in life. [NEWLINE] [NEWLINE] [STARTQ] Nobody is afraid of women. If I go out late at night, there has to be a trusted man with me. There are rapists at night, or if I get into an accident, who's going to change my tire? If I want to go travel extensively in a foreign country, especially the considered "third world" country, it's too dangerous. I am a woman, and I am American, and everyone knows how to to take advantage of that. [ENDQ] [NEWLINE] As a man in my thirties, if I were to interact with a woman I didn't know either at a bar, or the supermarket, or at work there's immediately the assumption that I am simply trying to "get with them". A guy cannot simply be nice to a woman without there being some preconceived notion that it's an attempt at wooing the particular individual. Likewise, if I attempt to help someone out of the goodness of my heart say if they have a broken down on the side of the road or they are visibly lost in a city at night, there's the assumption that I am simply going to try and harm them. People are scared of men. It doesn't matter for race or age, if you're a guy and you're interacting with a woman who you don't know, there's a danger sense women put forth which is entirely unfair, it makes sense, but it's still unfair. [NEWLINE] [NEWLINE] [STARTQ] If I sleep around, I am a slut. If a man sleeps around, he is a stud. No one gets very angry if a man abandons the baby he fathered. Yet, if I a woman knows that she won't be able to handle this baby, or if it was born out of rape, there are plenty of states where I wouldn't be able to abort it. If I even have sex with one person before they commit their entire being to me (aka: marriage), it's a risk. What if a mistake happens and I get pregnant? [ENDQ] [NEWLINE] There is one issue in regards to sex and men, and that issue is that men have to be the sex fiends that most see us as being. We
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Evolution in<mask> science.<mask><mask>, it is not **A** science,<mask> it is the coalescence of a mind boggling multitude of disciplines.<mask>, Yes, we cannot observe something that happened millions of years ago-<mask> that in no way limits our confidence of an event actually happening. [NEWLINE] [NEWLINE] Let's take the analogy of a robbery. You've been away from you home for weeks, having a lovely vacation in Bermuda.<mask> you come home, your window is broken, your house is trashed and several of your belongings- especially the valuable ones are missing. [NEWLINE] Obviously, someone came into your house.<mask> not being there to see it, you can be fairly certain that is the case. Your belongings could have just up and disappeared,<mask> there is still the more simple explanation that someone took it. [NEWLINE] [NEWLINE] <mask>'s interesting about this robbery is that multiple other robberies occurred in several other nearby houses, all within the time frame of<mask> you left. Even more interesting is that these robberies all occurred in a similar fashion- broken windows, valuables stolen,<mask> well<mask> numerous other peculiarities that tie these robberies together. Again, there exists the explanation that all of these valuables just up and left the houses, causing similar damages. Or, there's the ever more likely explanation that there was a robbery in each of these houses- with evidence building hat it was an repetitive process. Sadly, none of your neighbors saw who robbed their houses either. [NEWLINE] [NEWLINE] You find little comfort in the explanation that a robbery occurred in numerous houses, especially<mask> no one directly observed the culprit-<mask> there ever was one.<mask><mask>, you doubt it was the actions of a culprit at all<mask> the similar crimes occurring in the same time frame-<mask> you look for more evidence. You decide to examine the windows,<mask> they broke or<mask> broke them, you look for finger prints and foot prints. To your surprise, you're able to identify that all of the windows left glass on the inside of the house, indicating that<mask> someone had been in your house, they got in by breaking into your window. You find foot prints that cannot be explained by the shoes typically worn in your house, and you find finger prints that belong to no one in your family. Your neighbors all report the same things. It's become incredibly easy to assume at this point there was<mask> a robber in your house,<mask> you still doubt it. [NEWLINE] [NEWLINE] You have a meeting with your neighbors. All of them complain about<mask> their
Label encoding: <s>Evolution in indeed science. In fact, it is not **A** science, but it is the coalescence of a mind boggling multitude of disciplines. Also, Yes, we cannot observe something that happened millions of years ago- but that in no way limits our confidence of an event actually happening. [NEWLINE] [NEWLINE] Let's take the analogy of a robbery. You've been away from you home for weeks, having a lovely vacation in Bermuda. When you come home, your window is broken, your house is trashed and several of your belongings- especially the valuable ones are missing. [NEWLINE] Obviously, someone came into your house. Despite not being there to see it, you can be fairly certain that is the case. Your belongings could have just up and disappeared, but there is still the more simple explanation that someone took it. [NEWLINE] [NEWLINE] What's interesting about this robbery is that multiple other robberies occurred in several other nearby houses, all within the time frame of when you left. Even more interesting is that these robberies all occurred in a similar fashion- broken windows, valuables stolen, as well as numerous other peculiarities that tie these robberies together. Again, there exists the explanation that all of these valuables just up and left the houses, causing similar damages. Or, there's the ever more likely explanation that there was a robbery in each of these houses- with evidence building hat it was an repetitive process. Sadly, none of your neighbors saw who robbed their houses either. [NEWLINE] [NEWLINE] You find little comfort in the explanation that a robbery occurred in numerous houses, especially since no one directly observed the culprit- if there ever was one. In fact, you doubt it was the actions of a culprit at all despite the similar crimes occurring in the same time frame- so you look for more evidence. You decide to examine the windows, how they broke or what broke them, you look for finger prints and foot prints. To your surprise, you're able to identify that all of the windows left glass on the inside of the house, indicating that if someone had been in your house, they got in by breaking into your window. You find foot prints that cannot be explained by the shoes typically worn in your house, and you find finger prints that belong to no one in your family. Your neighbors all report the same things. It's become incredibly easy to assume at this point there was indeed a robber in your house, yet you still doubt it. [NEWLINE] [NEWLINE] You have a meeting with your neighbors. All of them complain about how their
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0981, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd like to respond to a couple of your points specifically which I,<mask> a Christian, have<mask> wrestled with.<mask><mask> you're asking good questions, and you shouldn't feel ashamed about asking them. [NEWLINE] [NEWLINE] (<mask> an aside regarding questions...)<mask> is the purpose of a question? To find an answer. There is such a thing<mask> truth - either there is a God, or there isn't; either there is a spiritual world of which we are a part, or there isn't, and<mask> forth. Whether or not any of us find, understand, or agree with the truth doesn't change<mask> that truth actually *is*. I could be wrong about<mask> I believe, or my agnostic friend could be wrong,<mask> neither of us, through our belief, has any power to change the reality of<mask> we're seeking to understand.<mask>, I don't think it's reasonable to be satisfied with a position that says that "<mask><mask><mask> you have some kind of belief that's great for you".<mask><mask> it's our duty to try to get<mask> close to the real truth<mask> we can,<mask><mask> I don't believe we'll know everything about that truth until we are face to face with God.<mask> I digress... [NEWLINE] [NEWLINE] [STARTQ] <mask> do I know I have the right God? Maybe I only believe in the American Jesus...<mask> another part of the world believes in Vishnu.<mask><mask> they're right? It seems like it's just fixed on wherever you are.... [ENDQ] [NEWLINE] This is exactly the kind of question to ask. The implications of Jesus<mask> the real truth (and he isn't just American, by the way) are vastly different than the implications<mask> Vishnu is the real truth, and they are not compatible with each other. This leaves us to reason that either one or both of them is wrong at the core, even<mask> they share some external moralities and ideas (which incidentally may themselves be in line with the external moralities and ideas of the real truth). I believe Jesus to be the truth primarily<mask> of the historical evidence surrounding his resurrection. To my understanding, Christianity is the only falsifiable religion. All religions make claims about spiritual realities which we cannot currently see,<mask> Christianity downright hinges on one event: the death and resurrection of Christ. [NEWLINE] [NEWLINE] There are several facts which both Christian and non-Christian scholarly sources attest to be true: [NEWLINE] [NEWLINE] 1.  Jesus was crucified and did actually die [NEWLINE] 2.  The tomb<mask> he was buried was found empty
Label encoding: <s>I'd like to respond to a couple of your points specifically which I, as a Christian, have also wrestled with. I think you're asking good questions, and you shouldn't feel ashamed about asking them. [NEWLINE] [NEWLINE] ( As an aside regarding questions...) What is the purpose of a question? To find an answer. There is such a thing as truth - either there is a God, or there isn't; either there is a spiritual world of which we are a part, or there isn't, and so forth. Whether or not any of us find, understand, or agree with the truth doesn't change what that truth actually *is*. I could be wrong about what I believe, or my agnostic friend could be wrong, but neither of us, through our belief, has any power to change the reality of what we're seeking to understand. Therefore, I don't think it's reasonable to be satisfied with a position that says that " as long as you have some kind of belief that's great for you". I think it's our duty to try to get as close to the real truth as we can, even though I don't believe we'll know everything about that truth until we are face to face with God. But I digress... [NEWLINE] [NEWLINE] [STARTQ] How do I know I have the right God? Maybe I only believe in the American Jesus... While another part of the world believes in Vishnu. What if they're right? It seems like it's just fixed on wherever you are.... [ENDQ] [NEWLINE] This is exactly the kind of question to ask. The implications of Jesus as the real truth (and he isn't just American, by the way) are vastly different than the implications if Vishnu is the real truth, and they are not compatible with each other. This leaves us to reason that either one or both of them is wrong at the core, even if they share some external moralities and ideas (which incidentally may themselves be in line with the external moralities and ideas of the real truth). I believe Jesus to be the truth primarily because of the historical evidence surrounding his resurrection. To my understanding, Christianity is the only falsifiable religion. All religions make claims about spiritual realities which we cannot currently see, but Christianity downright hinges on one event: the death and resurrection of Christ. [NEWLINE] [NEWLINE] There are several facts which both Christian and non-Christian scholarly sources attest to be true: [NEWLINE] [NEWLINE] 1.  Jesus was crucified and did actually die [NEWLINE] 2.  The tomb where he was buried was found empty
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You can only get a useful answer to a search<mask> you can figure out a search string that the search engine will interpret in a way that gives the answer you want. This means that the more you know about a subject, the more information you'll be able to get from the search engine. [ENDQ] [NEWLINE] I'd argue you can find more about<mask> to search for by searching generally and narrowing down until you find specifically<mask> search string will get your answer. Example below. [NEWLINE] [NEWLINE] [STARTQ] For example, you might say: "<mask> bother learning that the name of the game Go is baduk in Korean and wéiqí in Chinese? You can google that."<mask><mask> you ever needed to google anything about Go, you would quickly find that search engines can't do much with strings that include "go",<mask> it's<mask> common. At that point you meed to figure out another way to phrase your question, and the more alternative ways you can state your question, the more likely one<mask> them will work. [ENDQ] [NEWLINE] My example for this is I really like old Japanese cars. I'd like to someday buy one and import into the US and the other day I was curious to see<mask> Japanese used car listings I could find. I knew I wanted to see used Japanese cars for sale, which was my first search. I found a lot of American websites that specialize in importing<mask> that wasn't quite<mask> I was looking for.<mask> I then used translate to find out<mask> that search string would be in Japanese. I copy and pasted that into the search bar, and I found Japanese websites with car listings for cars I had never seen before (which was my goal). I didn't know<mask> "car" was in Japanese, and I still don't,<mask> I didn't need to in order to fulfill my goal. [NEWLINE] [NEWLINE] [STARTQ] In a different thread<mask> a similar issue was being discussed, a middle school teacher described his experience teaching math to students who had been brought up on a calculator-based curriculum. He wanted them to understand or to use distribution, like a(b+c)=ab+ac - and on the blackboard he would have an equation like 5(x+6)=3x. At the next step you were supposed to distribute,<mask> 5x+30=3x, then 30=-2x, then x=-15:<mask> his students were mystified by the entire thing, and he eventually figured out it was<mask> they didn't realize that 5x6=30.<mask> you had
Label encoding: <s> [STARTQ] You can only get a useful answer to a search if you can figure out a search string that the search engine will interpret in a way that gives the answer you want. This means that the more you know about a subject, the more information you'll be able to get from the search engine. [ENDQ] [NEWLINE] I'd argue you can find more about what to search for by searching generally and narrowing down until you find specifically what search string will get your answer. Example below. [NEWLINE] [NEWLINE] [STARTQ] For example, you might say: " Why bother learning that the name of the game Go is baduk in Korean and wéiqí in Chinese? You can google that." But if you ever needed to google anything about Go, you would quickly find that search engines can't do much with strings that include "go", because it's so common. At that point you meed to figure out another way to phrase your question, and the more alternative ways you can state your question, the more likely one if them will work. [ENDQ] [NEWLINE] My example for this is I really like old Japanese cars. I'd like to someday buy one and import into the US and the other day I was curious to see what Japanese used car listings I could find. I knew I wanted to see used Japanese cars for sale, which was my first search. I found a lot of American websites that specialize in importing but that wasn't quite what I was looking for. So I then used translate to find out what that search string would be in Japanese. I copy and pasted that into the search bar, and I found Japanese websites with car listings for cars I had never seen before (which was my goal). I didn't know what "car" was in Japanese, and I still don't, but I didn't need to in order to fulfill my goal. [NEWLINE] [NEWLINE] [STARTQ] In a different thread where a similar issue was being discussed, a middle school teacher described his experience teaching math to students who had been brought up on a calculator-based curriculum. He wanted them to understand or to use distribution, like a(b+c)=ab+ac - and on the blackboard he would have an equation like 5(x+6)=3x. At the next step you were supposed to distribute, so 5x+30=3x, then 30=-2x, then x=-15: but his students were mystified by the entire thing, and he eventually figured out it was because they didn't realize that 5x6=30. If you had
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.3684, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> this gets down to the central disagreement between us: [NEWLINE] [NEWLINE] [STARTQ] I think the problem is<mask> technology progresses, it's going to take a higher level of learning to be a productive member of society.<mask> you spend your first years learning just<mask> to add and subtract and spending time learning specific topics you'll use past school, at some point we're going to outpace our learning tempo and students coming out of school will have no grasp of the technology. [ENDQ] [NEWLINE] In 1850 there were people whose profession was to be *calculators* or *scriveners*. A calculator is a human who did<mask> mechanical calculators do now: you give him a list of sums and he calculates an answer. You could even set up multiple human calculators in a line and get and algorithm.<mask> there were people whose professional skills were completely dependent on arithmetic and completely replaced by the calculator.<mask> algebra was an advanced subject, used in special professions, and calculus (which was nearly two centuries old) was esoteric, whereas statistical analysis was newborn. None of those fields were very economically significant. Likewise with scriveners: they were human copy-pasters, who would copy over documents that needed to be identical. They were making the most of learning to write the alphabet, and were completely replaced by mimeographs.<mask> at the same time, the portion of the economy that required any literacy at all was quite small. [NEWLINE] [NEWLINE] <mask> has happened<mask> then? The jobs that required humans to act<mask> pure mindless arithmetic machines disappeared<mask> mindlessness always get mechanized eventually,<mask> they were replaced by many more jobs that required mindless algebra machines<mask> well<mask> others that required occasional algebra, and an ever-increasing number of professions started to use calculus. Was arithmetic less important in 1950 than in the era of the "calculators"? No, of course not: more students were studying it more thoroughly earlier in the curriculum,<mask> by 1950 huge numbers of wirking-class jobs were starting to require casual arithmetic skills,<mask> for professionals arithmetic was more important than ever. [NEWLINE] [NEWLINE] Now the algebra-machine jobs may be disappearing,<mask> the careers that require casual algebra won't; and<mask> instead of becoming *less* important algebra is becoming *more* important<mask> of the foundational role it plays in other skill-sets, like calculus, programming, and statistics, which used to be the mysterious secrets of a mathematical elite and are now common all over, and becoming more<mask> over time. [NEWLINE] [NEWLINE] Will calculus someday give us calculus machines
Label encoding: <s>I think this gets down to the central disagreement between us: [NEWLINE] [NEWLINE] [STARTQ] I think the problem is as technology progresses, it's going to take a higher level of learning to be a productive member of society. If you spend your first years learning just how to add and subtract and spending time learning specific topics you'll use past school, at some point we're going to outpace our learning tempo and students coming out of school will have no grasp of the technology. [ENDQ] [NEWLINE] In 1850 there were people whose profession was to be *calculators* or *scriveners*. A calculator is a human who did what mechanical calculators do now: you give him a list of sums and he calculates an answer. You could even set up multiple human calculators in a line and get and algorithm. So there were people whose professional skills were completely dependent on arithmetic and completely replaced by the calculator. Meanwhile algebra was an advanced subject, used in special professions, and calculus (which was nearly two centuries old) was esoteric, whereas statistical analysis was newborn. None of those fields were very economically significant. Likewise with scriveners: they were human copy-pasters, who would copy over documents that needed to be identical. They were making the most of learning to write the alphabet, and were completely replaced by mimeographs. But at the same time, the portion of the economy that required any literacy at all was quite small. [NEWLINE] [NEWLINE] What has happened since then? The jobs that required humans to act as pure mindless arithmetic machines disappeared because mindlessness always get mechanized eventually, but they were replaced by many more jobs that required mindless algebra machines as well as others that required occasional algebra, and an ever-increasing number of professions started to use calculus. Was arithmetic less important in 1950 than in the era of the "calculators"? No, of course not: more students were studying it more thoroughly earlier in the curriculum, because by 1950 huge numbers of wirking-class jobs were starting to require casual arithmetic skills, while for professionals arithmetic was more important than ever. [NEWLINE] [NEWLINE] Now the algebra-machine jobs may be disappearing, but the careers that require casual algebra won't; and meanwhile instead of becoming *less* important algebra is becoming *more* important because of the foundational role it plays in other skill-sets, like calculus, programming, and statistics, which used to be the mysterious secrets of a mathematical elite and are now common all over, and becoming more so over time. [NEWLINE] [NEWLINE] Will calculus someday give us calculus machines
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First, no feelings hurt, at least not mine. [NEWLINE] [NEWLINE] Second,<mask><mask> is that "*a farrowing crate, unable to turn around or stand up for her entire life, screaming out of frustration and boredom.*" is torture.<mask> you disagree, then before you read on, tell me<mask>,<mask> we can have the same premise. Sorry<mask> that was a confusion, I thought you would just assume everyone thinks that is torture. I was going from the following definition: [NEWLINE] (en.wikipedia.org/wiki/TortureTorture is the practice or act of deliberately inflicting severe physical pain) [NEWLINE] [NEWLINE] <mask>, got that out of the way. It is confusing sometimes to have more than one discussion at a time,<mask><mask> it is tempting to try and get on the same page right now, I will continue my argument assuming we both agree<mask> far. [NEWLINE] [NEWLINE] Moving on,<mask><mask> that this is not helpful [NEWLINE] [NEWLINE] **I believe that it is quite evident, without any special training or education, to see that this is not a matter** [NEWLINE] [NEWLINE] <mask><mask><mask><mask> it is quite evident, obviously! :) I actually, honestly think that it is quite evident that plants do feel things. I even support this point with evidence.<mask><mask> it is not okay to just dismiss my point, my evidence with "*I believe that it is quite evident, without any special training or education*"<mask> you can not suspend your belief that your point is quite evident, there can be no discussion. [NEWLINE] [NEWLINE] **Now<mask> for the first thing I will hand wave<mask><mask><mask> by definition suffering requires something to be self-aware and the plants like any type of nervous system at all which is required for it.** [NEWLINE] [NEWLINE] Okay,<mask> suffering requires self awareness. That is something I can work with. A nervous system and self awareness are requirements for suffering. Okay, well<mask> that is true would you be opposed to meat that does not come from self aware animals that are lacking nervous systems? I even have an example: [URL] Does In Vitro meat suffer? [NEWLINE] [NEWLINE] **<mask> for the senses of plants -- are you trying to tell me that a Venus Fly Trap's nyctinastic movement is the same<mask> a cow recoiling from a hot brand?** [NEWLINE] [NEWLINE] Yes and no. You are giving my argument the straw man treatment. That said, yes. I will not defend [nyctinastic = hot brand] per say,<mask> those 2 reactions are not similar.
Label encoding: <s>First, no feelings hurt, at least not mine. [NEWLINE] [NEWLINE] Second, my opinion is that "*a farrowing crate, unable to turn around or stand up for her entire life, screaming out of frustration and boredom.*" is torture. If you disagree, then before you read on, tell me why, so we can have the same premise. Sorry if that was a confusion, I thought you would just assume everyone thinks that is torture. I was going from the following definition: [NEWLINE] (en.wikipedia.org/wiki/TortureTorture is the practice or act of deliberately inflicting severe physical pain) [NEWLINE] [NEWLINE] So, got that out of the way. It is confusing sometimes to have more than one discussion at a time, so while it is tempting to try and get on the same page right now, I will continue my argument assuming we both agree so far. [NEWLINE] [NEWLINE] Moving on, I think that this is not helpful [NEWLINE] [NEWLINE] **I believe that it is quite evident, without any special training or education, to see that this is not a matter** [NEWLINE] [NEWLINE] I do not think it is quite evident, obviously! :) I actually, honestly think that it is quite evident that plants do feel things. I even support this point with evidence. I think it is not okay to just dismiss my point, my evidence with "*I believe that it is quite evident, without any special training or education*" If you can not suspend your belief that your point is quite evident, there can be no discussion. [NEWLINE] [NEWLINE] **Now as for the first thing I will hand wave because I think by definition suffering requires something to be self-aware and the plants like any type of nervous system at all which is required for it.** [NEWLINE] [NEWLINE] Okay, so suffering requires self awareness. That is something I can work with. A nervous system and self awareness are requirements for suffering. Okay, well if that is true would you be opposed to meat that does not come from self aware animals that are lacking nervous systems? I even have an example: [URL] Does In Vitro meat suffer? [NEWLINE] [NEWLINE] ** As for the senses of plants -- are you trying to tell me that a Venus Fly Trap's nyctinastic movement is the same as a cow recoiling from a hot brand?** [NEWLINE] [NEWLINE] Yes and no. You are giving my argument the straw man treatment. That said, yes. I will not defend [nyctinastic = hot brand] per say, as those 2 reactions are not similar.
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The risk associated with all of the conspiracy theories is the one thing preventing me from ever going full-blown conspiracy theorist. I'm operating under the "<mask><mask> " assumption. I'm really trying hard not to fall into the trap that a lot of conspiracy theorists find themselves in, and that is the trap of "yeah, okay,<mask>..."<mask> I mean by that is I don't ever want to come off<mask> just discounting all pieces of evidence refuting my position<mask> simultaneously just throwing up another theory. [NEWLINE] [NEWLINE] That being said, I feel your response is based solely on the idea of "it can't be true<mask> no one has come forward<mask>." In terms of the stock market manipulation,<mask><mask>. It would have been incredibly risky and an obvious way of tracking down people involved.<mask>, it still happened. I am absolutely not discounting the possibility of pure coincidence,<mask> you have to admit it is a BIG coincidence. Perhaps the fact that it is just<mask> obvious is enough to pull it off. Hide in plain sight and all that jazz. (Yeah I know that sounds like I'm grasping for straws, I'm just spitballing here.) I'd just like to know<mask> else could potentially cause this. [NEWLINE] [NEWLINE] Before I address the rest of your bullet points I want to say this: I don't think any of us have any idea exactly<mask> powerful/far reaching the government is.<mask><mask>, it would be RIDICULOUSLY difficult to keep all of this under wraps,<mask> that's not to say it's beyond the US Government's capabilities. Look at the NSA for example. It has been around<mask> 1917, and has been conducting mass surveillance on US citizens<mask> 1981. I'm willing to bet that the Snowden-esque surveillance has been going on for quite a<mask>,<mask> it took over 30 years for this information to be leaked and there are THOUSANDS of NSA employees. 9/11 only happened 14 years ago, and thousands of people would not have been needed to make this work. Again, just spitballing. [NEWLINE] [NEWLINE] [STARTQ] Who did the work? [ENDQ] [NEWLINE] I don't know, and I'm not going to pretend to. Who is digging through people's personal files unconstitutionally or carrying out torture at Guantanamo? (Granted those aren't<mask> significant<mask> a large-scale attack on your own people.) There are hundreds of millions of people in this country, and<mask> I'd like to believe there is a vast majority that would be
Label encoding: <s>The risk associated with all of the conspiracy theories is the one thing preventing me from ever going full-blown conspiracy theorist. I'm operating under the " what if " assumption. I'm really trying hard not to fall into the trap that a lot of conspiracy theorists find themselves in, and that is the trap of "yeah, okay, but..." What I mean by that is I don't ever want to come off as just discounting all pieces of evidence refuting my position while simultaneously just throwing up another theory. [NEWLINE] [NEWLINE] That being said, I feel your response is based solely on the idea of "it can't be true because no one has come forward yet." In terms of the stock market manipulation, I agree. It would have been incredibly risky and an obvious way of tracking down people involved. However, it still happened. I am absolutely not discounting the possibility of pure coincidence, but you have to admit it is a BIG coincidence. Perhaps the fact that it is just so obvious is enough to pull it off. Hide in plain sight and all that jazz. (Yeah I know that sounds like I'm grasping for straws, I'm just spitballing here.) I'd just like to know what else could potentially cause this. [NEWLINE] [NEWLINE] Before I address the rest of your bullet points I want to say this: I don't think any of us have any idea exactly how powerful/far reaching the government is. I agree, it would be RIDICULOUSLY difficult to keep all of this under wraps, but that's not to say it's beyond the US Government's capabilities. Look at the NSA for example. It has been around since 1917, and has been conducting mass surveillance on US citizens since 1981. I'm willing to bet that the Snowden-esque surveillance has been going on for quite a while, but it took over 30 years for this information to be leaked and there are THOUSANDS of NSA employees. 9/11 only happened 14 years ago, and thousands of people would not have been needed to make this work. Again, just spitballing. [NEWLINE] [NEWLINE] [STARTQ] Who did the work? [ENDQ] [NEWLINE] I don't know, and I'm not going to pretend to. Who is digging through people's personal files unconstitutionally or carrying out torture at Guantanamo? (Granted those aren't as significant as a large-scale attack on your own people.) There are hundreds of millions of people in this country, and while I'd like to believe there is a vast majority that would be
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>/u/anatcov is incorrect.  The Federal Reserve does increase the money supply. <mask><mask>, this is the global standard and not an anomaly.  Central banks are now buying [more government debt than net new issuance]( [URL] ). <mask> that means is not only are they financing all new debt, they are buying it off private parties now.  It's all new printing. [NEWLINE] [NEWLINE] Now, to address your original viewpoint, I still disagree with you. [NEWLINE] [NEWLINE] [STARTQ] f the general money supply is increased by, say, 3% each year, with all the "new" money going to the government,<mask> is this any different from the government taxing everyone 3%? Everyone's buying power goes down by 3%,<mask> the government's buying power goes up by whatever everyone else lost.<mask><mask> is this considered "sound fiscal policy",<mask> it's just a regressive tax, which is generally frowned upon in the first place? [ENDQ] [NEWLINE] There are a couple key points I would like to make. [NEWLINE] [NEWLINE] - Governments increasing the money supply doesn't have to result in inflation [NEWLINE] - Inflation can actually be a progressive tax [NEWLINE] - The best economic policy actually results in both of the above [NEWLINE] [NEWLINE] <mask> the USA was founded, Washington and Hamilton organized a central bank.  They had a policy of low interest rates and issued lots of new debt/currency.  This wasn't a problem<mask> the recipients of the debt were vastly different than today.  During this period, the debt issuance was based not on political connections / bailout policy (the Fed financing Uncle Sam and NY megabanks),<mask> evaluated<mask> an investment.  The Federal government borrowed money for infrastructure development, the states did the same,<mask> small businesses and individuals<mask> borrowed at below market rates.  Interestingly (and intentionally), this policy resulted in a precipitous drop in interest rates all the way across the pond in Europe to the chagrin of the usury class. [NEWLINE] [NEWLINE] The policy was not nearly<mask> inflationary<mask> you might expect such profligate expansion of the money supply to be today<mask> the debt went to *producers* in the economy.  You had to be investing in tangibles - seeds, farm machinery, factories, roads, etc., with it.  This increases production (goods) and inflation represents the relative change in production versus money supply (velocity X circulation volume).  You can increase the money supply 100x with no inflation,<mask> velocity of money stays the same and production goes up 100x
Label encoding: <s>/u/anatcov is incorrect.  The Federal Reserve does increase the money supply.  In fact, this is the global standard and not an anomaly.  Central banks are now buying [more government debt than net new issuance]( [URL] ).  What that means is not only are they financing all new debt, they are buying it off private parties now.  It's all new printing. [NEWLINE] [NEWLINE] Now, to address your original viewpoint, I still disagree with you. [NEWLINE] [NEWLINE] [STARTQ] f the general money supply is increased by, say, 3% each year, with all the "new" money going to the government, how is this any different from the government taxing everyone 3%? Everyone's buying power goes down by 3%, while the government's buying power goes up by whatever everyone else lost. So why is this considered "sound fiscal policy", when it's just a regressive tax, which is generally frowned upon in the first place? [ENDQ] [NEWLINE] There are a couple key points I would like to make. [NEWLINE] [NEWLINE] - Governments increasing the money supply doesn't have to result in inflation [NEWLINE] - Inflation can actually be a progressive tax [NEWLINE] - The best economic policy actually results in both of the above [NEWLINE] [NEWLINE] When the USA was founded, Washington and Hamilton organized a central bank.  They had a policy of low interest rates and issued lots of new debt/currency.  This wasn't a problem because the recipients of the debt were vastly different than today.  During this period, the debt issuance was based not on political connections / bailout policy (the Fed financing Uncle Sam and NY megabanks), but evaluated as an investment.  The Federal government borrowed money for infrastructure development, the states did the same, but small businesses and individuals also borrowed at below market rates.  Interestingly (and intentionally), this policy resulted in a precipitous drop in interest rates all the way across the pond in Europe to the chagrin of the usury class. [NEWLINE] [NEWLINE] The policy was not nearly as inflationary as you might expect such profligate expansion of the money supply to be today because the debt went to *producers* in the economy.  You had to be investing in tangibles - seeds, farm machinery, factories, roads, etc., with it.  This increases production (goods) and inflation represents the relative change in production versus money supply (velocity X circulation volume).  You can increase the money supply 100x with no inflation, if velocity of money stays the same and production goes up 100x
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Since /u/Froolow has not answered, I will do my best here. Note, I am far from a professional economist,<mask> I'm about to finish my second year of my Economics Major and took a Health Economics course last semester. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> cant the private system get rid of this?<mask> would an insurance company continue to employ a doctor who keeps pushing irrelevant treatment? [ENDQ] [NEWLINE] This is a problem referred to<mask> information asymmetry. It is present in a lot of markets,<mask> it is a very important factor in the healthcare market.<mask> doctors spend a good 8-10 years studying before becoming a doctor, they are experts and very knowledgeable on a large number of health issues. It it simply impossible for the average consumer to come anywhere close to being<mask> knowledgeable<mask> even the least educated doctor.<mask><mask><mask>, the consumer has to trust that the doctor is being honest about his care. The doctor has an incentive to prescribe more expensive treatments<mask> he will make more money doing<mask>. This is a problem in public and private<mask> has been more pronounced in private systems (i.e., the U.S.). The insurance companies<mask> have less knowledge about healthcare than the doctors do<mask> the problem is still present. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Mandatory insurance gets around this. [ENDQ] [NEWLINE] Unfortunately, it does not. Under a mandatory insurance model, users are still charged a fee based on their risk of becoming ill and their expected costs. Users have an incentive to over-state<mask> healthy they are in order to get lower premiums. This drives down the profit for insurance companies who are then forced to raise all premiums to make up for it.<mask> the system is mandatory, healthy individuals, who are now paying higher premiums than they deserve, are forced to keep paying this high premium<mask> they would rather go without insurance. [NEWLINE] [NEWLINE] [STARTQ] 3) Demand is irregular, and I can't predict my need for healthcare in advance, meaning I can't take steps to reduce my other consumption before I consume healthcare [ENDQ] This kind of demand is precisely<mask> the insurance model is designed for? [NEWLINE] [NEWLINE] [STARTQ] 4) Production is irregular; I pay for healthcare<mask> I want health, and I can't guarentee that my purchase of healthcare will result in health [ENDQ] Not sure of the relevance? [NEWLINE] [NEWLINE] For a private market to be completely efficient (absolutely no dead-weight loss), certain conditions must be met. Consumers need to be able to plan out their future expenses (which they cannot here)  and need to know<mask> they
Label encoding: <s>Since /u/Froolow has not answered, I will do my best here. Note, I am far from a professional economist, but I'm about to finish my second year of my Economics Major and took a Health Economics course last semester. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Why cant the private system get rid of this? Why would an insurance company continue to employ a doctor who keeps pushing irrelevant treatment? [ENDQ] [NEWLINE] This is a problem referred to as information asymmetry. It is present in a lot of markets, but it is a very important factor in the healthcare market. Since doctors spend a good 8-10 years studying before becoming a doctor, they are experts and very knowledgeable on a large number of health issues. It it simply impossible for the average consumer to come anywhere close to being as knowledgeable as even the least educated doctor. Because of this, the consumer has to trust that the doctor is being honest about his care. The doctor has an incentive to prescribe more expensive treatments since he will make more money doing so. This is a problem in public and private but has been more pronounced in private systems (i.e., the U.S.). The insurance companies also have less knowledge about healthcare than the doctors do so the problem is still present. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Mandatory insurance gets around this. [ENDQ] [NEWLINE] Unfortunately, it does not. Under a mandatory insurance model, users are still charged a fee based on their risk of becoming ill and their expected costs. Users have an incentive to over-state how healthy they are in order to get lower premiums. This drives down the profit for insurance companies who are then forced to raise all premiums to make up for it. Since the system is mandatory, healthy individuals, who are now paying higher premiums than they deserve, are forced to keep paying this high premium when they would rather go without insurance. [NEWLINE] [NEWLINE] [STARTQ] 3) Demand is irregular, and I can't predict my need for healthcare in advance, meaning I can't take steps to reduce my other consumption before I consume healthcare [ENDQ] This kind of demand is precisely what the insurance model is designed for? [NEWLINE] [NEWLINE] [STARTQ] 4) Production is irregular; I pay for healthcare but I want health, and I can't guarentee that my purchase of healthcare will result in health [ENDQ] Not sure of the relevance? [NEWLINE] [NEWLINE] For a private market to be completely efficient (absolutely no dead-weight loss), certain conditions must be met. Consumers need to be able to plan out their future expenses (which they cannot here)  and need to know what they
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Before you come to any snap judgment, please first let me define "success". I don't mean objectification, like "I can point at any girl on the street and have sex with her<mask> I own her body LOLOL". I don't even mean "playing". All I mean is being able to at some point in my life find a girl to marry who is compatible with me, on terms that are equally respectful to both myself and her. [NEWLINE] [NEWLINE] Let me give a bit of background. I'm a 29 year old successful professional with a full and interesting life. From my point of view (and from taking dating advice), I'd like to think that I have a lot of qualities that women would find attractive. Do you think women are drawn to success or money? I went to a top-5 school, I make money well into the six figures, and I own over $1M worth of real estate.<mask> about creativity or intelligence?<mask> singing, I play 2.5 musical instruments, I'm an internationally award winning dancer, I can speak 3 languages fluently, and I can get by in 10 more. Social skills or being an interesting person? I'm exceedingly polite, a good cook and conversationalist, and I have interesting hobbies that I like to share.<mask><mask> you'll probably infer from this post (due to the nature of it), I'm actually both confident and modest in real life. And finally, the "be attractive/don't be unattractive" dichotomy? Well, I'm certainly not accidentally unattractive -- I'm in great shape, I dress in flattering, fitted clothing, and I take care of my hygiene.<mask><mask><mask> "be attractive", I'm average, which shouldn't hurt or help me either way. I have my minor flaws in appearance,<mask><mask> does everyone. [NEWLINE] [NEWLINE] I'm not looking for a supermodel genius. I actually have incredibly low standards, and my friends and family are often concerned at the quality of women I pursue (Are you sure? Are you drunk?).<mask>, dating seems like a Sisyphean effort. I'm very social and meet on average 5-10 new women a week. I'm dismissed offhand by the vast majority (which is fine, I'm sure that happens to everyone), and with the few remaining I always get the impression from their side that they're not quite sure, that I should be grateful that they're deigning to spend their precious time with me, and that I better work hard and
Label encoding: <s>Before you come to any snap judgment, please first let me define "success". I don't mean objectification, like "I can point at any girl on the street and have sex with her because I own her body LOLOL". I don't even mean "playing". All I mean is being able to at some point in my life find a girl to marry who is compatible with me, on terms that are equally respectful to both myself and her. [NEWLINE] [NEWLINE] Let me give a bit of background. I'm a 29 year old successful professional with a full and interesting life. From my point of view (and from taking dating advice), I'd like to think that I have a lot of qualities that women would find attractive. Do you think women are drawn to success or money? I went to a top-5 school, I make money well into the six figures, and I own over $1M worth of real estate. What about creativity or intelligence? Besides singing, I play 2.5 musical instruments, I'm an internationally award winning dancer, I can speak 3 languages fluently, and I can get by in 10 more. Social skills or being an interesting person? I'm exceedingly polite, a good cook and conversationalist, and I have interesting hobbies that I like to share. Despite what you'll probably infer from this post (due to the nature of it), I'm actually both confident and modest in real life. And finally, the "be attractive/don't be unattractive" dichotomy? Well, I'm certainly not accidentally unattractive -- I'm in great shape, I dress in flattering, fitted clothing, and I take care of my hygiene. As far as "be attractive", I'm average, which shouldn't hurt or help me either way. I have my minor flaws in appearance, but so does everyone. [NEWLINE] [NEWLINE] I'm not looking for a supermodel genius. I actually have incredibly low standards, and my friends and family are often concerned at the quality of women I pursue (Are you sure? Are you drunk?). Yet, dating seems like a Sisyphean effort. I'm very social and meet on average 5-10 new women a week. I'm dismissed offhand by the vast majority (which is fine, I'm sure that happens to everyone), and with the few remaining I always get the impression from their side that they're not quite sure, that I should be grateful that they're deigning to spend their precious time with me, and that I better work hard and
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> the world should agree on a universal timezone. It would make scheduling easier, business meetings would be smoother, and there'd be much less confusion<mask> travelling and trying to synchronise things between countries. It wouldn't make any difference to sleep cycles or work hours or anything like that. We would<mask> abolish daylight saving time. [NEWLINE] [NEWLINE] <mask> I'm **not** suggesting is that the world all wake up at the same time and go to work at the same time. [NEWLINE] [NEWLINE] Let's say that we agree on GMT<mask> the world's timezone (<mask> it doesn't matter which time we choose). London would wake up at about 7am, work from 9am - 5pm, and go to sleep at about 10pm. The sun would rise at about 7am and would set at about 8pm. New York,<mask>, are 5 hours behind London,<mask> the sun would rise there at 2am on the universal clock. This doesn't affect anything<mask>,<mask> they'd wake up at 2am, go to work from 4am - 12pm, and go to bed at 5pm. They'd still get up<mask> the sun rises and go to bed<mask> the sun sets,<mask> *the only difference would be the number they read off the clock*. They'd still have the same sleeping pattern and working hours, and they'd get used to it in a matter of weeks. Instead of waking up and reading the time<mask> 7am, they'd wake up *at the same time* and read the clock<mask> 2am. All that would change is<mask> time they consider nighttime. Sydney,<mask>, are 9 hours in front of London - they wake up<mask> the sun rises at 4pm, they work from 6pm - 2am, see the sun set at 5am, and go to sleep at 7am. It seems weird that they should go to sleep at 7am,<mask> to them it would be the correct time! They'd still be going to sleep shortly after sunset, like the rest of the world, and would go to sleep right on schedule at the same time they would now. They'd just call the time they go to bed 7am rather than 10pm. *The only difference is the time they read off the clock* [NEWLINE] [NEWLINE] The times used above are obviously an arbitrary template used to demonstrate my argument. [NEWLINE] [NEWLINE] This would synchronise the world,<mask><mask> you want to skype your friend in Sydney from London, you can just say "I'll call you at 4am"
Label encoding: <s>I think the world should agree on a universal timezone. It would make scheduling easier, business meetings would be smoother, and there'd be much less confusion when travelling and trying to synchronise things between countries. It wouldn't make any difference to sleep cycles or work hours or anything like that. We would also abolish daylight saving time. [NEWLINE] [NEWLINE] What I'm **not** suggesting is that the world all wake up at the same time and go to work at the same time. [NEWLINE] [NEWLINE] Let's say that we agree on GMT as the world's timezone ( although it doesn't matter which time we choose). London would wake up at about 7am, work from 9am - 5pm, and go to sleep at about 10pm. The sun would rise at about 7am and would set at about 8pm. New York, however, are 5 hours behind London, so the sun would rise there at 2am on the universal clock. This doesn't affect anything though, because they'd wake up at 2am, go to work from 4am - 12pm, and go to bed at 5pm. They'd still get up when the sun rises and go to bed when the sun sets, but *the only difference would be the number they read off the clock*. They'd still have the same sleeping pattern and working hours, and they'd get used to it in a matter of weeks. Instead of waking up and reading the time as 7am, they'd wake up *at the same time* and read the clock as 2am. All that would change is what time they consider nighttime. Sydney, meanwhile, are 9 hours in front of London - they wake up when the sun rises at 4pm, they work from 6pm - 2am, see the sun set at 5am, and go to sleep at 7am. It seems weird that they should go to sleep at 7am, but to them it would be the correct time! They'd still be going to sleep shortly after sunset, like the rest of the world, and would go to sleep right on schedule at the same time they would now. They'd just call the time they go to bed 7am rather than 10pm. *The only difference is the time they read off the clock* [NEWLINE] [NEWLINE] The times used above are obviously an arbitrary template used to demonstrate my argument. [NEWLINE] [NEWLINE] This would synchronise the world, so if you want to skype your friend in Sydney from London, you can just say "I'll call you at 4am"
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0680, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is a very insightful and sympathetic criticism of Rand, which is<mask> I upvoted it,<mask> I don't think it ultimately works. [ENDQ] [NEWLINE] It's hardly a criticism of Rand except with regard to the wording she used.  And<mask> to that: [NEWLINE] [NEWLINE] [STARTQ] Rand's ethics is egoistic<mask> every moral obligation in it stems from the life of the individual. Rand acknowledged flourishing<mask> a value<mask><mask> to survival,<mask> on examination it turns out that flourishing is simply a variety of activities that one engages in in order to maximize one's long term survival prospects.<mask> Rand's egoism does not reduce to virtue ethics, which starts with flourishing rather than life. [ENDQ] [NEWLINE] I don't think the survival/flourishing distinction is<mask>'s essential here,<mask> rather motivation from one's own well-being<mask> distinct from motivation from others' well-being.  Actually, flourishing may be useful here to flesh out<mask> is involved in living *qua man*, which means living in accordance with the dictates of reason.  And that involves the recognition that one cannot logically justify acting on the basis of one's own well-being without acknowledging that the well-being of *any* sentient being carries normative force, in some fashion or other, for anyone capable of recognizing the normative force and acting<mask>.  Applied to the case of a human being making normative decisions<mask> they impact upon the well-being of a non-intelligent animal, say, the human - that is, in acting<mask><mask> reason - identifies<mask> is relevantly (normatively) similar between himself and the other animal.  Both are capable of feeling pain.  And there is some indefensible inconsistency or practical contradiction between upholding the value of my pursuing things that result in less pain for me,<mask> neglecting the significance of the pain for the animal.  It is to recognize a (relevant, normative) similarity and<mask> not act<mask>. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, for a rational being to take into account the normative significance of other sentient beings' pain or well-being,<mask> worthy of respect (in some manner of acting or other - is to live in such a way<mask> not to be solely, narrowly concerned with only one's own well-being, in which case using the terminology of "egoism" is odd and misleading.  The silver lining in this is that Rand's version of "egoism" doesn't merit the many lazy strawman approaches to characterizing her ethics.  And, this is
Label encoding: <s> [STARTQ] This is a very insightful and sympathetic criticism of Rand, which is why I upvoted it, but I don't think it ultimately works. [ENDQ] [NEWLINE] It's hardly a criticism of Rand except with regard to the wording she used.  And as to that: [NEWLINE] [NEWLINE] [STARTQ] Rand's ethics is egoistic because every moral obligation in it stems from the life of the individual. Rand acknowledged flourishing as a value in addition to survival, but on examination it turns out that flourishing is simply a variety of activities that one engages in in order to maximize one's long term survival prospects. Thus Rand's egoism does not reduce to virtue ethics, which starts with flourishing rather than life. [ENDQ] [NEWLINE] I don't think the survival/flourishing distinction is what's essential here, but rather motivation from one's own well-being as distinct from motivation from others' well-being.  Actually, flourishing may be useful here to flesh out what is involved in living *qua man*, which means living in accordance with the dictates of reason.  And that involves the recognition that one cannot logically justify acting on the basis of one's own well-being without acknowledging that the well-being of *any* sentient being carries normative force, in some fashion or other, for anyone capable of recognizing the normative force and acting accordingly.  Applied to the case of a human being making normative decisions as they impact upon the well-being of a non-intelligent animal, say, the human - that is, in acting according to reason - identifies what is relevantly (normatively) similar between himself and the other animal.  Both are capable of feeling pain.  And there is some indefensible inconsistency or practical contradiction between upholding the value of my pursuing things that result in less pain for me, while neglecting the significance of the pain for the animal.  It is to recognize a (relevant, normative) similarity and yet not act accordingly. [NEWLINE] [NEWLINE] On the other hand, for a rational being to take into account the normative significance of other sentient beings' pain or well-being, as worthy of respect (in some manner of acting or other - is to live in such a way as not to be solely, narrowly concerned with only one's own well-being, in which case using the terminology of "egoism" is odd and misleading.  The silver lining in this is that Rand's version of "egoism" doesn't merit the many lazy strawman approaches to characterizing her ethics.  And, this is
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Dubstep is a true minefield. Here we go. [NEWLINE] [NEWLINE] I'm going to blindly assume this is a typical 'dubstep' example you have listened to; [Dr P - Sweet Shop]( [URL] ) [NEWLINE] [NEWLINE] This isn't the true sound. Saying 'oh dubstep is all wob wob screeeech wib wob wob wob' is<mask> false. Let me bring some history in. I'm from the UK, and our underground music, almost wholly from London, is absolute goldmine in terms of music culture. There is a concept of the ['hardcore continuum']( [URL] ), which essentially theorises that underground music is "a musical tradition/subcultural tribe that’s managed to hold it together for nearly 20 years now, negotiating drastic stylistic shifts and significant changes in technology, drugs, and the social/racial composition of its own population". [NEWLINE] [NEWLINE] [NEWLINE] We start in the early 90s, with [hardcore]( [URL] ) which was rave music, which begat drum n bass, [jungle]( [URL] ), garage and 2-step (amongst others). 2-step is a sub strand of garage, focusing on broken swung rhythms and got increasingly dark in tone. Exhibit A; [El-B - Buck and Bury]( [URL] ).<mask> we're in the early 00s, and this sound is rife across London. Imagine a big dark room with a big soundsystem, with these edgey, nervous tunes playing out to a similarly edgy nervous audience. Could you fathom the feeling? [NEWLINE] [NEWLINE] <mask> this sound grew, dubstep emerged. 2-Step swinging rhythms with an emphasis on half-time drums, keeping space in the track and providing the low sub bass that is the blood of underground music. This tune is the foundation, the culmination of all of these influences that have been brewing over the years and something of a realisation that 'dubstep' is it's own genre. I present; [Skream - Midnight Request Line \(2005\)]( [URL] ) and [Digital Mystikz - Antiwar Dub]( [URL] ) [NEWLINE] [NEWLINE] A large part of Dubstep was Dubplate culture. An old concept that roots back to old Dub scenes in Jamaica. The DJ would have a tune cut to a dubplate, a one off vinyl record, which they would have exclusively, or a strict handful of people would have,<mask> fans could hear tunes from specific DJs only, on the night, in the moment. A concept that the commenters on youtube about<mask> '
Label encoding: <s>Dubstep is a true minefield. Here we go. [NEWLINE] [NEWLINE] I'm going to blindly assume this is a typical 'dubstep' example you have listened to; [Dr P - Sweet Shop]( [URL] ) [NEWLINE] [NEWLINE] This isn't the true sound. Saying 'oh dubstep is all wob wob screeeech wib wob wob wob' is so false. Let me bring some history in. I'm from the UK, and our underground music, almost wholly from London, is absolute goldmine in terms of music culture. There is a concept of the ['hardcore continuum']( [URL] ), which essentially theorises that underground music is "a musical tradition/subcultural tribe that’s managed to hold it together for nearly 20 years now, negotiating drastic stylistic shifts and significant changes in technology, drugs, and the social/racial composition of its own population". [NEWLINE] [NEWLINE] [NEWLINE] We start in the early 90s, with [hardcore]( [URL] ) which was rave music, which begat drum n bass, [jungle]( [URL] ), garage and 2-step (amongst others). 2-step is a sub strand of garage, focusing on broken swung rhythms and got increasingly dark in tone. Exhibit A; [El-B - Buck and Bury]( [URL] ). So we're in the early 00s, and this sound is rife across London. Imagine a big dark room with a big soundsystem, with these edgey, nervous tunes playing out to a similarly edgy nervous audience. Could you fathom the feeling? [NEWLINE] [NEWLINE] As this sound grew, dubstep emerged. 2-Step swinging rhythms with an emphasis on half-time drums, keeping space in the track and providing the low sub bass that is the blood of underground music. This tune is the foundation, the culmination of all of these influences that have been brewing over the years and something of a realisation that 'dubstep' is it's own genre. I present; [Skream - Midnight Request Line \(2005\)]( [URL] ) and [Digital Mystikz - Antiwar Dub]( [URL] ) [NEWLINE] [NEWLINE] A large part of Dubstep was Dubplate culture. An old concept that roots back to old Dub scenes in Jamaica. The DJ would have a tune cut to a dubplate, a one off vinyl record, which they would have exclusively, or a strict handful of people would have, so fans could hear tunes from specific DJs only, on the night, in the moment. A concept that the commenters on youtube about how '
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is not easy (for me at the moment at least) to find studies that look at<mask> representation in media influences<mask> we think about and treat certain groups or<mask> certain groups behave and think about themselves. [NEWLINE] For many people it is obvious that<mask> for example a little black kid sees that the hero in the movies he watches is always a white guy, and people that look like him are only side-characters that might influence (not determine) the way he thinks about himself, the things he considers himself capable of and the things he goes on to pursue. And that media<mask> influences<mask> other people think he is capable of. Same could be said for representation of gender. This goes not only for movies<mask> media in general. [NEWLINE] [NEWLINE] The expectation we have of certain groups (gender, race, orientation etc.) come largely from the society we life in and the things we perceive<mask> normal or given. The alternative would be that we are born or naturally develop gender roles or stereotypes even without any interaction, to me that seems highly unlikely. I don't think that we are able to form our subconscious biases only by real world interaction without even being influenced by 'imaginary' content, like for example video games. [NEWLINE] [NEWLINE] <mask>, you are right, that just<mask> something seems obvious that doesn't mean that we don't need to go searching for scientific proof. [NEWLINE] [NEWLINE] Here is one study that I could find; [NEWLINE] [STARTQ] To give one example of this effect, we know that (for reason not relevant here) women and minorities are underrepresented in media especially<mask> it comes to characters that are professionally successful. There was a study done "that finds evidence of a self-esteem boosting effect of television for white boys,<mask> self-esteem damaging effects for white girls, black girls, and black boys.  One primary reason?  White boys see lots of white boys and men in the shows they watch.  And, not just that,<mask> they regularly see these characters and actors in positive, powerful, and central roles." ([quote]( [URL] /), [source]( [URL].abstract)). [ENDQ] [NEWLINE] A quote from a [2000 paper]( [URL] %253A10.1023%252FA%253A1007046204478.pdf?originUrl=http%3A%2F%2Flink.springer.com%2Farticle%2F10.1023%2FA%3A1007046204478&amp;token2=exp=1436542114~
Label encoding: <s>It is not easy (for me at the moment at least) to find studies that look at how representation in media influences how we think about and treat certain groups or how certain groups behave and think about themselves. [NEWLINE] For many people it is obvious that if for example a little black kid sees that the hero in the movies he watches is always a white guy, and people that look like him are only side-characters that might influence (not determine) the way he thinks about himself, the things he considers himself capable of and the things he goes on to pursue. And that media also influences what other people think he is capable of. Same could be said for representation of gender. This goes not only for movies but media in general. [NEWLINE] [NEWLINE] The expectation we have of certain groups (gender, race, orientation etc.) come largely from the society we life in and the things we perceive as normal or given. The alternative would be that we are born or naturally develop gender roles or stereotypes even without any interaction, to me that seems highly unlikely. I don't think that we are able to form our subconscious biases only by real world interaction without even being influenced by 'imaginary' content, like for example video games. [NEWLINE] [NEWLINE] However, you are right, that just because something seems obvious that doesn't mean that we don't need to go searching for scientific proof. [NEWLINE] [NEWLINE] Here is one study that I could find; [NEWLINE] [STARTQ] To give one example of this effect, we know that (for reason not relevant here) women and minorities are underrepresented in media especially when it comes to characters that are professionally successful. There was a study done "that finds evidence of a self-esteem boosting effect of television for white boys, but self-esteem damaging effects for white girls, black girls, and black boys.  One primary reason?  White boys see lots of white boys and men in the shows they watch.  And, not just that, but they regularly see these characters and actors in positive, powerful, and central roles." ([quote]( [URL] /), [source]( [URL].abstract)). [ENDQ] [NEWLINE] A quote from a [2000 paper]( [URL] %253A10.1023%252FA%253A1007046204478.pdf?originUrl=http%3A%2F%2Flink.springer.com%2Farticle%2F10.1023%2FA%3A1007046204478&amp;token2=exp=1436542114~
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> I recognize that global warming is real and happening.  And I recognize that human activity appears to be the main driver of it. <mask> I don't think the consequences will be<mask> dire<mask> are portrayed for humanity.  In the following, most of my source material is going to come from the IPCC, which is<mask><mask> a fairly good source for data on this, and<mask> anything has an institutional bias in the direction of warning of dire consequences. [NEWLINE] [NEWLINE] There are ~~four~~three avenues that<mask><mask> are of primary concern: [NEWLINE] [NEWLINE] * Sea level rise.  [The IPCC estimates]( [URL].pdf) a sea level rise in the range of 0.5m over the next century. <mask> that's not trivial, it's<mask> not dire.  Most coastal communities can manage that level of rise with levees and sea walls.  Occasionally someone shows a map with a city like New York or Miami under a 3 or 4 meter rise in sea levels,<mask> the IPCC estimates give that a very low probability of happening. [NEWLINE] [NEWLINE] * Drought and other negative impacts on agricultural production.  I do not deny that some areas will see reduced rainfall,<mask> I don't think that can be true for everywhere (<mask>, it would seem like higher aggregate temperatures would result in more atmospheric water vapor and more aggregate rainfall). <mask><mask> some areas may be negatively impacted, others will be positively impacted.  Further, we've been getting progressively more efficient agriculturally<mask> time goes on.  Across the world, the amount of agricultural land per capita has been falling for decades, and in developed nations,<mask> population growth is slow, is falling in absolute terms.  See pg. 502 in [this IPCC report]( [URL].pdf)  This [other report]( [URL].gov/publications/SAR/SAR_Chapter%2013.pdf) on the impact of climate change on crop output says, with<mask>'s described<mask> medium confidence that: [NEWLINE] [STARTQ] Global agricultural production can be maintained relative to base production under climate change<mask> expressed by general circulation models under doubled CO2 equilibrium climate scenarios. [ENDQ] [NEWLINE] * Severe weather events. <mask> I don't deny that global warming can cause more severe weather events such<mask> hurricanes, I question whether this is a very dire consequence. <mask> our weather forecasting improves, and our disaster preparedness improves, the loss of life from weather events falls.  Even a very bad storm like Katrina was not nearly<mask> devastating to human life<mask> a storm [with
Label encoding: <s>So I recognize that global warming is real and happening.  And I recognize that human activity appears to be the main driver of it.  But I don't think the consequences will be as dire as are portrayed for humanity.  In the following, most of my source material is going to come from the IPCC, which is I think a fairly good source for data on this, and if anything has an institutional bias in the direction of warning of dire consequences. [NEWLINE] [NEWLINE] There are ~~four~~three avenues that I think are of primary concern: [NEWLINE] [NEWLINE] * Sea level rise.  [The IPCC estimates]( [URL].pdf) a sea level rise in the range of 0.5m over the next century.  While that's not trivial, it's also not dire.  Most coastal communities can manage that level of rise with levees and sea walls.  Occasionally someone shows a map with a city like New York or Miami under a 3 or 4 meter rise in sea levels, but the IPCC estimates give that a very low probability of happening. [NEWLINE] [NEWLINE] * Drought and other negative impacts on agricultural production.  I do not deny that some areas will see reduced rainfall, but I don't think that can be true for everywhere ( indeed, it would seem like higher aggregate temperatures would result in more atmospheric water vapor and more aggregate rainfall).  So while some areas may be negatively impacted, others will be positively impacted.  Further, we've been getting progressively more efficient agriculturally as time goes on.  Across the world, the amount of agricultural land per capita has been falling for decades, and in developed nations, where population growth is slow, is falling in absolute terms.  See pg. 502 in [this IPCC report]( [URL].pdf)  This [other report]( [URL].gov/publications/SAR/SAR_Chapter%2013.pdf) on the impact of climate change on crop output says, with what's described as medium confidence that: [NEWLINE] [STARTQ] Global agricultural production can be maintained relative to base production under climate change as expressed by general circulation models under doubled CO2 equilibrium climate scenarios. [ENDQ] [NEWLINE] * Severe weather events.  While I don't deny that global warming can cause more severe weather events such as hurricanes, I question whether this is a very dire consequence.  As our weather forecasting improves, and our disaster preparedness improves, the loss of life from weather events falls.  Even a very bad storm like Katrina was not nearly as devastating to human life as a storm [with
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure I can directly challenge your view in this case,<mask> I feel that this kind of policy or attitude is a judgement call for corporate. I understand your frustrations,<mask> I don't think it's a black and white problem.<mask><mask><mask><mask> this kind of policy or attitude is just,<mask> you say, "a holdover from the work schedules of farms, factories and the military." That probably has something to do with it,<mask> there are three primary reasons I see that this kind of attitude persists: conventionality, culture, and politics. [NEWLINE] [NEWLINE] Please keep in mind<mask> reading this that I know little about your job and understand that you may feel none of these things apply to you.<mask> these are my views on<mask> this attitude is<mask> prevalent in jobs today. [NEWLINE] [NEWLINE] Conventionality dictates that there is an efficiency gain to force an entire team on the same schedule to minimize the chance of any holdups<mask> a member of the team can't be reached. This theoretically improves the dependability of every employee both internally and externally to clients. This is,<mask> I said, a judgement call by corporate.<mask> this policy didn't have any consequences, it would be black and white --<mask> risk any hold-ups<mask> someone's not here<mask> I need him to be?<mask> obviously it's ambiguous -- you believe there would be little-to-no hold-ups and<mask> that your efficiency is sacrificed by being forced to work at non-ideal hours.<mask> this is all very hard to quantify at a corporate level,<mask> the policy is set and the attitude persists. [NEWLINE] [NEWLINE] That persistent and unquestioned attitude of a 7am schedule is a bit of a cultural thing in places too. East coast vs. West coast US have very different cultural attitude on this, driven in part by the required sporadic hours that scientists in silicon valley have experienced vs. a need to be reachable on a consistent schedule that the east coast business world has traditionally desired. Other countries have different cultural attitudes. On a pragmatic basis, you could easily<mask><mask> decisions like this shouldn't be driven just by cultural traditions, and a careful analysis should be made instead.<mask> there are reasons that cultural traditions persist, including the personal comfort that people get from having a consistent attitude with different jobs, rather than drastically different expectations from different bosses. And the personal opinions that some have that an early schedule is just more "professional." [NEWLINE] [NEWLINE] Which leads me to the political driving factor. Many do view an early schedule -- or at least
Label encoding: <s>I'm not sure I can directly challenge your view in this case, because I feel that this kind of policy or attitude is a judgement call for corporate. I understand your frustrations, but I don't think it's a black and white problem. I do not think this kind of policy or attitude is just, as you say, "a holdover from the work schedules of farms, factories and the military." That probably has something to do with it, but there are three primary reasons I see that this kind of attitude persists: conventionality, culture, and politics. [NEWLINE] [NEWLINE] Please keep in mind while reading this that I know little about your job and understand that you may feel none of these things apply to you. But these are my views on why this attitude is so prevalent in jobs today. [NEWLINE] [NEWLINE] Conventionality dictates that there is an efficiency gain to force an entire team on the same schedule to minimize the chance of any holdups because a member of the team can't be reached. This theoretically improves the dependability of every employee both internally and externally to clients. This is, as I said, a judgement call by corporate. If this policy didn't have any consequences, it would be black and white -- why risk any hold-ups because someone's not here when I need him to be? But obviously it's ambiguous -- you believe there would be little-to-no hold-ups and also that your efficiency is sacrificed by being forced to work at non-ideal hours. But this is all very hard to quantify at a corporate level, so the policy is set and the attitude persists. [NEWLINE] [NEWLINE] That persistent and unquestioned attitude of a 7am schedule is a bit of a cultural thing in places too. East coast vs. West coast US have very different cultural attitude on this, driven in part by the required sporadic hours that scientists in silicon valley have experienced vs. a need to be reachable on a consistent schedule that the east coast business world has traditionally desired. Other countries have different cultural attitudes. On a pragmatic basis, you could easily argue that decisions like this shouldn't be driven just by cultural traditions, and a careful analysis should be made instead. But there are reasons that cultural traditions persist, including the personal comfort that people get from having a consistent attitude with different jobs, rather than drastically different expectations from different bosses. And the personal opinions that some have that an early schedule is just more "professional." [NEWLINE] [NEWLINE] Which leads me to the political driving factor. Many do view an early schedule -- or at least
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"<mask> you believe something is bad, you should outlaw it." [NEWLINE] [NEWLINE] There is a difference between legal sanction and social sanction. I hate okra.<mask><mask> its bad.<mask><mask> that people with okra on their breath are encroaching on my happiness<mask> they breath around me and have to smell it.<mask> should it be Illegal?<mask><mask> all the people in the north thought the same<mask> me? majority of the population thinks okra is bad, should it be illegal? [NEWLINE] [NEWLINE] No. [NEWLINE] [NEWLINE] The people who, for whatever reason wish to eat okra have every right to do<mask><mask> it makes them happy.<mask> it's not hurting me, and the cost is having men with guns, attack, capture, and take people who eat and grow okra into cages, and remove them from society. [NEWLINE] [NEWLINE] That is<mask> we are talking about with legal sanction. The government only knows one way to handle law, and that is with force. They would have to take money, by force, from people who like and dislike okra alike, to fund operations to find people who grow, transport and consume okra. People won't just stop eating the okra,<mask> they do like it and have no moral issue with it,<mask> they will risk men with guns kicking down their door to consume it. Not only that<mask> there is a going to be a taboo about okra, and status that comes to some people who do enjoy it. The consuming of okra very well might increase. [NEWLINE] [NEWLINE] Now lets look at corn. Corn is huge, everyone eats corn, and I love corn. God i love corn in everything. Now lets say that there are people who think corn is bad, like<mask><mask> okra is bad. Rather than make it illegal, they make a large campaign that points out<mask> much corn farming destroys the environment,<mask> well<mask><mask> completely obnoxious people who eat corn on the cob and smack their lips all over the place are. They then come out with study's that show that corn leads to diabetes, leads to kidney and liver issues, and<mask> are empty calories that lead to being just a fat ass. Now the corn industry will have tons of issues with this. They will fight tooth and nail to keep corn on the cob the main stay of every Americans side dish. Year pass, and people see that corn really is bad to grow, and finally see that all the smacking of lips and slurps are obnoxious. Other side dishes become more popular<mask> replacements for corn, and
Label encoding: <s>" If you believe something is bad, you should outlaw it." [NEWLINE] [NEWLINE] There is a difference between legal sanction and social sanction. I hate okra. I think its bad. I think that people with okra on their breath are encroaching on my happiness when they breath around me and have to smell it. So should it be Illegal? What if all the people in the north thought the same as me? majority of the population thinks okra is bad, should it be illegal? [NEWLINE] [NEWLINE] No. [NEWLINE] [NEWLINE] The people who, for whatever reason wish to eat okra have every right to do so if it makes them happy. Because it's not hurting me, and the cost is having men with guns, attack, capture, and take people who eat and grow okra into cages, and remove them from society. [NEWLINE] [NEWLINE] That is what we are talking about with legal sanction. The government only knows one way to handle law, and that is with force. They would have to take money, by force, from people who like and dislike okra alike, to fund operations to find people who grow, transport and consume okra. People won't just stop eating the okra, because they do like it and have no moral issue with it, so they will risk men with guns kicking down their door to consume it. Not only that but there is a going to be a taboo about okra, and status that comes to some people who do enjoy it. The consuming of okra very well might increase. [NEWLINE] [NEWLINE] Now lets look at corn. Corn is huge, everyone eats corn, and I love corn. God i love corn in everything. Now lets say that there are people who think corn is bad, like i think okra is bad. Rather than make it illegal, they make a large campaign that points out how much corn farming destroys the environment, as well as how completely obnoxious people who eat corn on the cob and smack their lips all over the place are. They then come out with study's that show that corn leads to diabetes, leads to kidney and liver issues, and also are empty calories that lead to being just a fat ass. Now the corn industry will have tons of issues with this. They will fight tooth and nail to keep corn on the cob the main stay of every Americans side dish. Year pass, and people see that corn really is bad to grow, and finally see that all the smacking of lips and slurps are obnoxious. Other side dishes become more popular as replacements for corn, and
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many problems with all this. [NEWLINE] [NEWLINE] (1) Most obviously<mask> god is all powerful then god could simply create people in an ascended state to begin with whatever that means. <mask> any religion that says infants or babies who die get to go to heaven or whatever, clearly already believes this happens a lot. [NEWLINE] [NEWLINE] (2) The principle itself just isn't true. [NEWLINE] [NEWLINE] [STARTQ] By enduring and responding to evil, we grow and become virtuous [ENDQ] [NEWLINE] Generally this just doesn't happen. <mask> it did then wouldn't we all be virtuous by now?  You see a lot of virtue around? [NEWLINE] [NEWLINE] (2a)<mask> it worked then people who have suffered a lot would be the most virtuous.  This is false. [NEWLINE] [NEWLINE] (2b) Can we agree that generally people who lived a long time ago had a lot more nasty shit happen to them?  Does that mean that people in the 1800s were more moral than people today? <mask><mask> they were far more racist, sexist, prejudiced in all sorts of ways and felt OK with slavery? [NEWLINE] [NEWLINE] (2c)<mask><mask><mask> society's morals have increased<mask> life has got easier. [NEWLINE] [NEWLINE] (2d) I will concede that rich people (subject to the least pain) are less moral than the middle class,<mask> at the same time being subjected to poverty often thrusts people towards a more violent, prejudiced and criminal life style.  This is not<mask> the theory would predict. [NEWLINE] [NEWLINE] (2e) Are older people more moral than the young?  They ought to be<mask> this theory is right<mask> they already lived through a lifetime of god's training program.  Older people tend to be much more prejudicial eg on gay rights issues. [NEWLINE] [NEWLINE] (3) Even<mask> it was true that pain creates virtue, the theory doesn't explain the pattern of the distribution of pain [NEWLINE] [NEWLINE] <mask> pain is god's training program then it ought to be more evenly distributed. <mask> is the point of having some go through life without much pain.  That would be counter-productive.  This theory would predict a particular pattern of the distribution of pain that is simply nothing like that which is observed.  For example<mask> someone had a very painful childhood then logically they learned to be virtuous at an early age and don't need repeat lessons.  Contrariwise children born to an easy life would require extra pain in later life to catch up. [NEWLINE] [NEWLINE] We see the opposite.  People with shitty childhoods are mostly those
Label encoding: <s>Many problems with all this. [NEWLINE] [NEWLINE] (1) Most obviously if god is all powerful then god could simply create people in an ascended state to begin with whatever that means.  Indeed any religion that says infants or babies who die get to go to heaven or whatever, clearly already believes this happens a lot. [NEWLINE] [NEWLINE] (2) The principle itself just isn't true. [NEWLINE] [NEWLINE] [STARTQ] By enduring and responding to evil, we grow and become virtuous [ENDQ] [NEWLINE] Generally this just doesn't happen.  If it did then wouldn't we all be virtuous by now?  You see a lot of virtue around? [NEWLINE] [NEWLINE] (2a) If it worked then people who have suffered a lot would be the most virtuous.  This is false. [NEWLINE] [NEWLINE] (2b) Can we agree that generally people who lived a long time ago had a lot more nasty shit happen to them?  Does that mean that people in the 1800s were more moral than people today?  Even though they were far more racist, sexist, prejudiced in all sorts of ways and felt OK with slavery? [NEWLINE] [NEWLINE] (2c) On the contrary society's morals have increased as life has got easier. [NEWLINE] [NEWLINE] (2d) I will concede that rich people (subject to the least pain) are less moral than the middle class, but at the same time being subjected to poverty often thrusts people towards a more violent, prejudiced and criminal life style.  This is not what the theory would predict. [NEWLINE] [NEWLINE] (2e) Are older people more moral than the young?  They ought to be if this theory is right because they already lived through a lifetime of god's training program.  Older people tend to be much more prejudicial eg on gay rights issues. [NEWLINE] [NEWLINE] (3) Even if it was true that pain creates virtue, the theory doesn't explain the pattern of the distribution of pain [NEWLINE] [NEWLINE] If pain is god's training program then it ought to be more evenly distributed.  What is the point of having some go through life without much pain.  That would be counter-productive.  This theory would predict a particular pattern of the distribution of pain that is simply nothing like that which is observed.  For example if someone had a very painful childhood then logically they learned to be virtuous at an early age and don't need repeat lessons.  Contrariwise children born to an easy life would require extra pain in later life to catch up. [NEWLINE] [NEWLINE] We see the opposite.  People with shitty childhoods are mostly those
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>? I'm not arguing that it necessarily follows that NAM is in some way false, just that people have reason to be sceptical. [ENDQ] [NEWLINE] I don't want to deny you your skepticism<mask> at the same time you can't criticize its use. [NEWLINE] [NEWLINE] [STARTQ] hijacking the discussion to be about<mask> men might feel unfairly treated by the phrasing seems petty and disingenuous [ENDQ] [NEWLINE] I suppose it could be, can you give me an example that you think fairly represents that kind of a conversation? [NEWLINE] [NEWLINE] I'm thinking a comment that could come up in the domestic abuse situation might be, "Men are just looking to control women and they do it through physical violence." deserves a NAM comment<mask> it is flat out wrong and assumes all men are this way or that enough of the majority are that we can avoid discussing the minority there. I use the "men are ____" format<mask> that was the example the OP gave.  To talk about another format is arguing something else the OP didn't talk about. [NEWLINE] [NEWLINE] <mask> you used another format like, "Some men are controlling." and someone says NAM then yes you have a point<mask> again that was not the point of the OP. [NEWLINE] [NEWLINE] Focusing on one gender is<mask> created the imbalance we have now, continuing to focus on only one of the genders won't create balance it'll just create a different imbalance. [NEWLINE] [NEWLINE] [STARTQ] <mask> 'illogical' isn't a term that can adequately apply to a gender at all. [ENDQ] [NEWLINE] Neither can 99% of all statements that begin with "Men/ Women are _____" Exactly my point. <mask> people stopped saying, "Men/ Women are..." no one would feel the need to correct it. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> there's something to be said about this point. I feel like 'not all men' holds more weight in an academic context. On twitter,<mask>,<mask> much discussion is between friends and informal, I'm inclined to think that this scenario is closer to the POW/guard one, not<mask> men and women are opposed to each other<mask><mask> of the way the language operates within a given context that isn't necessarily supposed to be interpreted by people outside of a certain group [ENDQ] [NEWLINE] I would replace academic with public.<mask> you want to have a private venting session with your friends no matter your gender etc...and you are just mad and angry I accept we will all make statements that are BS...the purpose there is entirely different<mask> it is semi private...and<mask> someone
Label encoding: <s> [STARTQ] So? I'm not arguing that it necessarily follows that NAM is in some way false, just that people have reason to be sceptical. [ENDQ] [NEWLINE] I don't want to deny you your skepticism but at the same time you can't criticize its use. [NEWLINE] [NEWLINE] [STARTQ] hijacking the discussion to be about how men might feel unfairly treated by the phrasing seems petty and disingenuous [ENDQ] [NEWLINE] I suppose it could be, can you give me an example that you think fairly represents that kind of a conversation? [NEWLINE] [NEWLINE] I'm thinking a comment that could come up in the domestic abuse situation might be, "Men are just looking to control women and they do it through physical violence." deserves a NAM comment because it is flat out wrong and assumes all men are this way or that enough of the majority are that we can avoid discussing the minority there. I use the "men are ____" format because that was the example the OP gave.  To talk about another format is arguing something else the OP didn't talk about. [NEWLINE] [NEWLINE] If you used another format like, "Some men are controlling." and someone says NAM then yes you have a point but again that was not the point of the OP. [NEWLINE] [NEWLINE] Focusing on one gender is what created the imbalance we have now, continuing to focus on only one of the genders won't create balance it'll just create a different imbalance. [NEWLINE] [NEWLINE] [STARTQ] because 'illogical' isn't a term that can adequately apply to a gender at all. [ENDQ] [NEWLINE] Neither can 99% of all statements that begin with "Men/ Women are _____" Exactly my point.  If people stopped saying, "Men/ Women are..." no one would feel the need to correct it. [NEWLINE] [NEWLINE] [STARTQ] I think there's something to be said about this point. I feel like 'not all men' holds more weight in an academic context. On twitter, however, where much discussion is between friends and informal, I'm inclined to think that this scenario is closer to the POW/guard one, not because men and women are opposed to each other but because of the way the language operates within a given context that isn't necessarily supposed to be interpreted by people outside of a certain group [ENDQ] [NEWLINE] I would replace academic with public. If you want to have a private venting session with your friends no matter your gender etc...and you are just mad and angry I accept we will all make statements that are BS...the purpose there is entirely different because it is semi private...and if someone
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Many Nazis were atheists [ENDQ] [NEWLINE] Theoretically, yes.  [Practically, no.]( [URL] #Atheists) [NEWLINE] [NEWLINE] [STARTQ] and Himmler even wanted a neopagan revival. [ENDQ] [NEWLINE] Neopaganism is a religion. [NEWLINE] [NEWLINE] [STARTQ] Hitler had all of the Crucifixes in Catholic schools replaced with a picture of him. He literally set himself up above Christ, making him the opposite of a Christian. [ENDQ] [NEWLINE] Not only is that false, it's backwards.  From [here]( [URL] #Catholic_resistance): [NEWLINE] [NEWLINE] *Catholic anger was further fuelled by actions of the Gauleiter of Upper Bavaria, Adolf* ***Wagner, a militantly anti-Catholic Nazi****, who in June 1941 ordered the removal of crucifixes from all schools in his Gau. This attack on Catholicism provoked the first public demonstrations against government policy<mask> the Nazis had come to power, and the mass signing of petitions, including by Catholic soldiers serving at the front.* ***<mask> Hitler heard of this he ordered Wagner to rescind his decree****,<mask> the damage had been done – German Catholics had learned that the regime could be successfully opposed.* [NEWLINE] [NEWLINE] [STARTQ] Stalin really hated Christianity,<mask> did Lenin, Marxism is inherently anti religious, with millions killed in the name of progress, an atheistic, materialistic progress. [ENDQ] [NEWLINE] *<mask> * much wrong with this.  First off, Soviet leaders didn't hate religion<mask> they were atheists, they hated religion<mask> they felt ["religion was an opiate that needed to be removed in order to construct the ideal communist society"]( [URL] #Religion) and ["religious organizations are always considered by Marxism<mask> the organs of bourgeois reaction, used for the protection of the exploitation and the stupefaction of the working class."]( [URL] #Communist_states)  Their personal beliefs were irrelevant; they ruthlessly oppressed [artists and intellectuals]( [URL] #Intelligentsia) just<mask> much<mask> the religious, and for the same reasons. [NEWLINE] [NEWLINE] <mask>, not only is materialism not solely an atheistic philosophy (religious leaders can be [very]( [URL] ) [materialistic]( [URL] /) [<mask> ]( [URL] %27s_evangelical_university_getting_filthy_rich_off_your_tax_money)) or even one embraced by all atheists (there are plenty of [non-materialist atheists]( [URL] )),<mask> communism explicitly rejects materialism in the progress-bound capitalistic sense and the Soviet
Label encoding: <s> [STARTQ] Many Nazis were atheists [ENDQ] [NEWLINE] Theoretically, yes.  [Practically, no.]( [URL] #Atheists) [NEWLINE] [NEWLINE] [STARTQ] and Himmler even wanted a neopagan revival. [ENDQ] [NEWLINE] Neopaganism is a religion. [NEWLINE] [NEWLINE] [STARTQ] Hitler had all of the Crucifixes in Catholic schools replaced with a picture of him. He literally set himself up above Christ, making him the opposite of a Christian. [ENDQ] [NEWLINE] Not only is that false, it's backwards.  From [here]( [URL] #Catholic_resistance): [NEWLINE] [NEWLINE] *Catholic anger was further fuelled by actions of the Gauleiter of Upper Bavaria, Adolf* ***Wagner, a militantly anti-Catholic Nazi****, who in June 1941 ordered the removal of crucifixes from all schools in his Gau. This attack on Catholicism provoked the first public demonstrations against government policy since the Nazis had come to power, and the mass signing of petitions, including by Catholic soldiers serving at the front.* *** When Hitler heard of this he ordered Wagner to rescind his decree****, but the damage had been done – German Catholics had learned that the regime could be successfully opposed.* [NEWLINE] [NEWLINE] [STARTQ] Stalin really hated Christianity, as did Lenin, Marxism is inherently anti religious, with millions killed in the name of progress, an atheistic, materialistic progress. [ENDQ] [NEWLINE] * So * much wrong with this.  First off, Soviet leaders didn't hate religion because they were atheists, they hated religion because they felt ["religion was an opiate that needed to be removed in order to construct the ideal communist society"]( [URL] #Religion) and ["religious organizations are always considered by Marxism as the organs of bourgeois reaction, used for the protection of the exploitation and the stupefaction of the working class."]( [URL] #Communist_states)  Their personal beliefs were irrelevant; they ruthlessly oppressed [artists and intellectuals]( [URL] #Intelligentsia) just as much as the religious, and for the same reasons. [NEWLINE] [NEWLINE] Secondly, not only is materialism not solely an atheistic philosophy (religious leaders can be [very]( [URL] ) [materialistic]( [URL] /) [ indeed ]( [URL] %27s_evangelical_university_getting_filthy_rich_off_your_tax_money)) or even one embraced by all atheists (there are plenty of [non-materialist atheists]( [URL] )), but communism explicitly rejects materialism in the progress-bound capitalistic sense and the Soviet
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with your first point about the difficulty in making utilitarian value judgments.<mask>, humans in general seem to be pretty terrible at knowing<mask> will make them happy.<mask><mask> that the consumer culture which seems to have spread into every economic class after the industrial revolution may be due in part to the creation of a need<mask> there wasn't before, simply by virtue of learning about a new product. I've<mask> read articles to the effect that some popular magazines are designed in such a way that the readers feel inadequate and bad about themselves, then the magazine displays the advertisements of products that promise to fix these inadequacies in comparison to the people portrayed in the magazine. The readers then feel they need to buy the product in order to be complete and happy. We are very prone to social manipulation. [NEWLINE] [NEWLINE] Your dessert example is one that has been used in an effort to convince me to have sex with men! I'm afraid we just have to play the probabilities<mask> making decisions and try to keep a level head. It's one of life's frustrations. [NEWLINE] [NEWLINE]... [NEWLINE] [NEWLINE] Humanity's desire for ultimate meaning is very interesting. Religion seems to be the only institution that caters to it. There might be a deep relationship there. From your posts, it seems that you are agnostic or atheist,<mask> I'm going to use that assumption and share my personal musings on the evolutionary value of a desire for objective or transcendent meaning. [NEWLINE] [NEWLINE] Early humans had to deal with both a mysterious and unpredictable natural environment and mercurial, violent social orders. Two things that religions accomplish are creating explanations for the workings of the natural world (<mask> well<mask> cultivating a sense that humans can influence it through prayer) and creating an inviolable social order with deities at the top and humans who interpret the will and laws of the deities. Humans who believe in the deities respect the laws of the religious order and fulfill their biological drive for taking action to change unfavorable circumstances through ritual and prayer, which is far less damaging to established social orders than staging a coup on the current alphas in the community. [NEWLINE] [NEWLINE] Societies built on long term, objective, transcendent concepts are stable and have high levels of cooperation,<mask> instead of being united under only a mortal human leader or within a constantly changing collection, the underlying, unifying concept stays the same for many generations. This stability helps societies to grow and prosper. It<mask> encourages people within the societies to kill or exile iconoclasts who try to upset the order,<mask> it would upset the balance of
Label encoding: <s>I agree with your first point about the difficulty in making utilitarian value judgments. Also, humans in general seem to be pretty terrible at knowing what will make them happy. I think that the consumer culture which seems to have spread into every economic class after the industrial revolution may be due in part to the creation of a need where there wasn't before, simply by virtue of learning about a new product. I've also read articles to the effect that some popular magazines are designed in such a way that the readers feel inadequate and bad about themselves, then the magazine displays the advertisements of products that promise to fix these inadequacies in comparison to the people portrayed in the magazine. The readers then feel they need to buy the product in order to be complete and happy. We are very prone to social manipulation. [NEWLINE] [NEWLINE] Your dessert example is one that has been used in an effort to convince me to have sex with men! I'm afraid we just have to play the probabilities when making decisions and try to keep a level head. It's one of life's frustrations. [NEWLINE] [NEWLINE]... [NEWLINE] [NEWLINE] Humanity's desire for ultimate meaning is very interesting. Religion seems to be the only institution that caters to it. There might be a deep relationship there. From your posts, it seems that you are agnostic or atheist, so I'm going to use that assumption and share my personal musings on the evolutionary value of a desire for objective or transcendent meaning. [NEWLINE] [NEWLINE] Early humans had to deal with both a mysterious and unpredictable natural environment and mercurial, violent social orders. Two things that religions accomplish are creating explanations for the workings of the natural world ( as well as cultivating a sense that humans can influence it through prayer) and creating an inviolable social order with deities at the top and humans who interpret the will and laws of the deities. Humans who believe in the deities respect the laws of the religious order and fulfill their biological drive for taking action to change unfavorable circumstances through ritual and prayer, which is far less damaging to established social orders than staging a coup on the current alphas in the community. [NEWLINE] [NEWLINE] Societies built on long term, objective, transcendent concepts are stable and have high levels of cooperation, because instead of being united under only a mortal human leader or within a constantly changing collection, the underlying, unifying concept stays the same for many generations. This stability helps societies to grow and prosper. It also encourages people within the societies to kill or exile iconoclasts who try to upset the order, because it would upset the balance of
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me try to tell you about beauty standards for women. You say they *only* need to look thin. A lot of women would disagree with you. [NEWLINE] [NEWLINE] Let's start with the basics: body hair. Women have body hair just like men,<mask> it's socially unacceptable. Men have to shave their faces or keep their beards trimmed,<mask> women have to shave their arm pits and legs.<mask> a woman has dark arm hair, she generally has to shave that, too. Many women get hair on their upper lip or chin, and those need to be plucked or waxed. Eyebrows<mask> need to be a certain shape. Some women pluck them,<mask> others get them threaded or waxed. Women often get their pubic hair waxed, too. [NEWLINE] [NEWLINE] Now<mask> about the face and skin? Women are often told they have to wear makeup to work to look professional. I know I've been told that. That means you have to spend time and money on chemicals to put on your eyelashes and lips every single day. Many women<mask> put makeup over their entire faces and then color on their cheeks and eyelids. I haven't gotten to wrinkles<mask>. A man with wrinkles is mature; a woman with wrinkles is old. Makeup doesn't cover wrinkles very well,<mask> a lot of women get injections and surgical procedures to hide wrinkles. [NEWLINE] [NEWLINE] <mask> for hair, men are basically expected to keep theirs short. It's OK<mask> their hair goes gray<mask> it shows sophistication. Women are generally told that long hair is more attractive, and they need to wear it down. I know many women who have been told by strangers to dye their hair<mask> they start showing gray. Think of politicians. There are many with white hair.<mask> would you think of Hilary Clinton<mask> she showed her white hair? You'd probably just see her<mask> an old lady.<mask>, women have to keep their hair dyed *every few weeks* after a certain age. They<mask> have to keep it styled which takes a *lot* of time in the morning. Women have to blow dry their hair and add products to add volume and shine and tame flyaways &amp; frizz. [NEWLINE] [NEWLINE] We're<mask> usually expected to add jewels to our bodies.<mask> professional women skip on the necklace and earrings, it might seem that she's not finished getting dressed. Better stick those jewels into the earlobes! [NEWLINE] [NEWLINE] Women<mask> know that having a defined clavicle is attractive. We know that
Label encoding: <s>Let me try to tell you about beauty standards for women. You say they *only* need to look thin. A lot of women would disagree with you. [NEWLINE] [NEWLINE] Let's start with the basics: body hair. Women have body hair just like men, but it's socially unacceptable. Men have to shave their faces or keep their beards trimmed, but women have to shave their arm pits and legs. If a woman has dark arm hair, she generally has to shave that, too. Many women get hair on their upper lip or chin, and those need to be plucked or waxed. Eyebrows also need to be a certain shape. Some women pluck them, while others get them threaded or waxed. Women often get their pubic hair waxed, too. [NEWLINE] [NEWLINE] Now what about the face and skin? Women are often told they have to wear makeup to work to look professional. I know I've been told that. That means you have to spend time and money on chemicals to put on your eyelashes and lips every single day. Many women also put makeup over their entire faces and then color on their cheeks and eyelids. I haven't gotten to wrinkles yet. A man with wrinkles is mature; a woman with wrinkles is old. Makeup doesn't cover wrinkles very well, so a lot of women get injections and surgical procedures to hide wrinkles. [NEWLINE] [NEWLINE] As for hair, men are basically expected to keep theirs short. It's OK if their hair goes gray because it shows sophistication. Women are generally told that long hair is more attractive, and they need to wear it down. I know many women who have been told by strangers to dye their hair when they start showing gray. Think of politicians. There are many with white hair. How would you think of Hilary Clinton if she showed her white hair? You'd probably just see her as an old lady. So, women have to keep their hair dyed *every few weeks* after a certain age. They also have to keep it styled which takes a *lot* of time in the morning. Women have to blow dry their hair and add products to add volume and shine and tame flyaways &amp; frizz. [NEWLINE] [NEWLINE] We're also usually expected to add jewels to our bodies. If professional women skip on the necklace and earrings, it might seem that she's not finished getting dressed. Better stick those jewels into the earlobes! [NEWLINE] [NEWLINE] Women also know that having a defined clavicle is attractive. We know that
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can agree with you on cars,<mask> I cannot on motorcycles. [NEWLINE] [NEWLINE] [NEWLINE] I have been riding for 8 years. I ride 90% of the year, most of January is covered in ice. 5 of those years has been on a nice quiet soft bike, a Vulcan 550 LTD, that everyone is<mask> happy to see go by. People would tell me, "thanks for the quiet pipes" all the<mask> I am seething.<mask> am I seething? [NEWLINE] [NEWLINE] Almost daily I would get run off into another lane or on to the shoulder. People would pull out in front of me all the time. I am not exaggerating about it, I live in the city. Riding in the city is dangerous, you have to have your head on a swivel, and with filtering and splitting being outlawed, it's even more dangerous. All too often people in cars do not look before they signal and come over<mask> they are busy looking for a car. [NEWLINE] [NEWLINE] Then I bought a Harley. [NEWLINE] [NEWLINE] Three times this year I have been run off the road into the emergency lane. Twice now have I been pulled out in front of. I have successfully avoided one left hook<mask> well. It's easy to remember these incidents<mask> they are scary<mask> hell. [NEWLINE] [NEWLINE] Once I got loud pipes I noticed that people started looking<mask> they'd flick their blinker on<mask> they HEARD something. Ordinarily the most common move for people is to flick the blinker and come on over<mask> they don't see a CAR<mask> they are not LOOKING for a MOTORCYCLE. People who would have pulled out in front of me don't<mask> often<mask> they are going to take a right into traffic,<mask> they heard a noise AND they saw me flashing my brights preparing for a SMIDSY. I see careless drivers at a stop, jog the vehicle and stop<mask> they HEARD something then NOTICED me. [NEWLINE] [NEWLINE] I see some people drift<mask> they hear something,<mask> not into my lane, they drift<mask> they HEAR something and now are aware that something is around them. Let them drift a little, we are trained to drive for the cagers. They are now aware and LOOKING FOR A MOTORCYCLE or something, or anything. You see them drift, it is your job to get to a position<mask> they can see you, we are taught this<mask> you take a safety course. Most of the time you postition yourself<mask> they can see you in the
Label encoding: <s>I can agree with you on cars, but I cannot on motorcycles. [NEWLINE] [NEWLINE] [NEWLINE] I have been riding for 8 years. I ride 90% of the year, most of January is covered in ice. 5 of those years has been on a nice quiet soft bike, a Vulcan 550 LTD, that everyone is so happy to see go by. People would tell me, "thanks for the quiet pipes" all the while I am seething. Why am I seething? [NEWLINE] [NEWLINE] Almost daily I would get run off into another lane or on to the shoulder. People would pull out in front of me all the time. I am not exaggerating about it, I live in the city. Riding in the city is dangerous, you have to have your head on a swivel, and with filtering and splitting being outlawed, it's even more dangerous. All too often people in cars do not look before they signal and come over because they are busy looking for a car. [NEWLINE] [NEWLINE] Then I bought a Harley. [NEWLINE] [NEWLINE] Three times this year I have been run off the road into the emergency lane. Twice now have I been pulled out in front of. I have successfully avoided one left hook as well. It's easy to remember these incidents because they are scary as hell. [NEWLINE] [NEWLINE] Once I got loud pipes I noticed that people started looking when they'd flick their blinker on because they HEARD something. Ordinarily the most common move for people is to flick the blinker and come on over if they don't see a CAR because they are not LOOKING for a MOTORCYCLE. People who would have pulled out in front of me don't as often because they are going to take a right into traffic, but they heard a noise AND they saw me flashing my brights preparing for a SMIDSY. I see careless drivers at a stop, jog the vehicle and stop because they HEARD something then NOTICED me. [NEWLINE] [NEWLINE] I see some people drift when they hear something, but not into my lane, they drift because they HEAR something and now are aware that something is around them. Let them drift a little, we are trained to drive for the cagers. They are now aware and LOOKING FOR A MOTORCYCLE or something, or anything. You see them drift, it is your job to get to a position so they can see you, we are taught this when you take a safety course. Most of the time you postition yourself so they can see you in the
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Consent,<mask><mask> we can agree, is a basic requirement for sex to be a positive act. Consent allows for there to be communication and reassurance between whoever is involved that<mask> they're doing, and<mask> they're doing it, is okay with their partner(s). Consent<mask> allows for things that would otherwise be creepy or weird to come into a positive light and expressed in a healthy way. I would consider the act of objectifying to fall under this category. For example, a woman may be creeped out<mask> a dude right in front of her at a party is just staring at her boobs without any concern for her at all. The same act of staring at the boobs is 100% fine<mask> it's with her boyfriend<mask> they're having consensual sexual relations. [NEWLINE] [NEWLINE] <mask> we take this same line of logic of consent that happens with sex and put it with masturbation, that's<mask> I'm going with this.<mask> you're thinking about another person in a sexual manner without them knowing about it, and then acting out that thought process by masturbating, it just seems a bit creepy to me.<mask> I'm not allowed to objectify or do other sexual acts in person without having consent from the other person, I don't see<mask> that wouldn't<mask> translate over to masturbation. It's the same sexual thoughts, feelings, and intentions with the only difference being that the partner is not physically with you.<mask> they're not there with you, you can't ask for proper consent unless you physically ask them some other time<mask> you can masturbate about them. [NEWLINE] [NEWLINE] This line of thought has logic to it,<mask> it's a conclusion that I don't like to face. This line of thought potentially makes masturbation an awful act<mask> done without consent. It<mask> makes it an act that wouldn't be allowed<mask> no one gave any consent to you. You might be able to say that pornography that was paid for might be consensual masturbation<mask> the model is being paid, and is<mask> willing to consent to you masturbating. Even<mask>, this doesn't allow you to explore your sexual desires outside of straight up porn unless someone says it's okay.<mask> you asked me whether I'm for or against masturbation, I would say that I'm very pro-masturbation. The act has health benefits, relieves stress, is pleasurable, and all sorts of other fun goodies. It makes it<mask> I don't want to come to the conclusion that I've come to that it needs strict consent about the person(s)
Label encoding: <s>Consent, I think we can agree, is a basic requirement for sex to be a positive act. Consent allows for there to be communication and reassurance between whoever is involved that what they're doing, and how they're doing it, is okay with their partner(s). Consent also allows for things that would otherwise be creepy or weird to come into a positive light and expressed in a healthy way. I would consider the act of objectifying to fall under this category. For example, a woman may be creeped out if a dude right in front of her at a party is just staring at her boobs without any concern for her at all. The same act of staring at the boobs is 100% fine if it's with her boyfriend while they're having consensual sexual relations. [NEWLINE] [NEWLINE] If we take this same line of logic of consent that happens with sex and put it with masturbation, that's where I'm going with this. If you're thinking about another person in a sexual manner without them knowing about it, and then acting out that thought process by masturbating, it just seems a bit creepy to me. If I'm not allowed to objectify or do other sexual acts in person without having consent from the other person, I don't see how that wouldn't also translate over to masturbation. It's the same sexual thoughts, feelings, and intentions with the only difference being that the partner is not physically with you. Because they're not there with you, you can't ask for proper consent unless you physically ask them some other time if you can masturbate about them. [NEWLINE] [NEWLINE] This line of thought has logic to it, but it's a conclusion that I don't like to face. This line of thought potentially makes masturbation an awful act if done without consent. It also makes it an act that wouldn't be allowed if no one gave any consent to you. You might be able to say that pornography that was paid for might be consensual masturbation because the model is being paid, and is therefore willing to consent to you masturbating. Even so, this doesn't allow you to explore your sexual desires outside of straight up porn unless someone says it's okay. If you asked me whether I'm for or against masturbation, I would say that I'm very pro-masturbation. The act has health benefits, relieves stress, is pleasurable, and all sorts of other fun goodies. It makes it so I don't want to come to the conclusion that I've come to that it needs strict consent about the person(s)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Taking a quick moment to address your addendum to the OP. [NEWLINE] [NEWLINE] I feel you are leaving a crucial component of the argument out.  The reason that it is<mask> difficult for you to find anything wrong with the cannibalism in principal is<mask> you are framing it devoid of context.  Cannibalism in "the right" context already has widespread acceptance, albeit begrudgingly. [NEWLINE] [NEWLINE] <mask> people hear about cannibalism used<mask> a last ditch means to survive most see an occurrence that is sad, disgusting and regrettable. <mask> is it wrong? <mask> the standard in society against cannibalism most can be convinced of the validity of the action considering the extreme circumstances.  Some even see it<mask> being worse<mask> they HADN'T eaten the other person, viewing it<mask> better that somebody survived rather than nobody.  It may even be that it is the last great thing someone accomplished for their comrade in suffering, providing them with the needed nutrition to survive. [NEWLINE] [NEWLINE] Their exist today numerous tribes that still practice ritual cannibalism within their community<mask> a part of death rites and such.  We see it<mask> disgusting of course<mask> only evangelical bigots invade their community and attempt to make them stop.  The practices are in a way just<mask> crazy<mask> our habits of preserving them and burying them in a box<mask> likely stem from the same goal.  Dispose of the body in a way that causes the least disease and attracts the least predators and scavengers. [NEWLINE] ***** [NEWLINE] At this point you may be wondering "Aren't you supposed to be changing my view?  This is pretty strong for my case."  At this point I am challenging not the issues the question brings up<mask> the question itself. [NEWLINE] [NEWLINE] You cannot possibly have a discussion about the morality of a given event<mask> that event is taking place in a vacuum.  Once someone has reached a certain point in logical thinking statements like "(Insert context free event here) is not immoral<mask> done in the right way." become essentially no-brainers.  Without context anything that can be imagined is possible and the arguments ceases to be about the event and instead becomes a tennis match wherein the participants throw various hypothetical contexts at each other attempting to reach favorable conclusions.  Just<mask> their is no human without society, there are no morals without context. [NEWLINE] [NEWLINE] By presupposing the context<mask> you have done in point 4 (consent), the Edit (new society starting from scratch) and your latest addendum (cleanliness, safety and access are now non-
Label encoding: <s>Taking a quick moment to address your addendum to the OP. [NEWLINE] [NEWLINE] I feel you are leaving a crucial component of the argument out.  The reason that it is so difficult for you to find anything wrong with the cannibalism in principal is because you are framing it devoid of context.  Cannibalism in "the right" context already has widespread acceptance, albeit begrudgingly. [NEWLINE] [NEWLINE] When people hear about cannibalism used as a last ditch means to survive most see an occurrence that is sad, disgusting and regrettable.  But is it wrong?  Despite the standard in society against cannibalism most can be convinced of the validity of the action considering the extreme circumstances.  Some even see it as being worse if they HADN'T eaten the other person, viewing it as better that somebody survived rather than nobody.  It may even be that it is the last great thing someone accomplished for their comrade in suffering, providing them with the needed nutrition to survive. [NEWLINE] [NEWLINE] Their exist today numerous tribes that still practice ritual cannibalism within their community as a part of death rites and such.  We see it as disgusting of course but only evangelical bigots invade their community and attempt to make them stop.  The practices are in a way just as crazy as our habits of preserving them and burying them in a box but likely stem from the same goal.  Dispose of the body in a way that causes the least disease and attracts the least predators and scavengers. [NEWLINE] ***** [NEWLINE] At this point you may be wondering "Aren't you supposed to be changing my view?  This is pretty strong for my case."  At this point I am challenging not the issues the question brings up but the question itself. [NEWLINE] [NEWLINE] You cannot possibly have a discussion about the morality of a given event when that event is taking place in a vacuum.  Once someone has reached a certain point in logical thinking statements like "(Insert context free event here) is not immoral if done in the right way." become essentially no-brainers.  Without context anything that can be imagined is possible and the arguments ceases to be about the event and instead becomes a tennis match wherein the participants throw various hypothetical contexts at each other attempting to reach favorable conclusions.  Just as their is no human without society, there are no morals without context. [NEWLINE] [NEWLINE] By presupposing the context as you have done in point 4 (consent), the Edit (new society starting from scratch) and your latest addendum (cleanliness, safety and access are now non-
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As someone who moved from the UK to the US right after graduating, I can understand. [NEWLINE] [NEWLINE] At first, I thought it would be amazing. After I graduated, I spent 2 months visiting a bunch of the big cities in the US, and we fell in love. It wasn't long after that, that I got a job offer to move over here (San Francisco). [NEWLINE] [NEWLINE] Before I actually lived here, the idea of living here was very romantic. [NEWLINE] [NEWLINE] In truth, it's just a country. There are some things I absolutely love about the US, that are less about the politics. The geography is excellent. There aren't really any places left in the UK that aren't completely inhabited, and farmed for centuries. And<mask> it does exist, it's only a couple of miles to civilization. (Ok, let's not count Scotland.<mask>, that shit's way too cold). I love<mask> different parts of the US are.<mask> you look down the street in California, it looks different to<mask> it is in New York, Florida, Washington, etc. In that sense, there's a lot to explore. I've<mask> spent some time in Hawaii, and to be honest, it's beautiful. There isn't quite the equivalent in the UK, Gibraltar doesn't really compare. [NEWLINE] [NEWLINE] On the flip side; I can see a lot of problems. Large parts of San Francisco are a ghetto. The number of homeless people I walk past in the 15 minute walk to work, astonishes me.<mask> does the number of $250k cars I see every day. [NEWLINE] [NEWLINE] One common problem people have, is the medical care. No, it's not free. My employer pays my insurance, my deductable (the excess, to the more civilized people), is $500 a year. I don't consider that too bad really, my taxes are lower than they would be in the UK. Things are generally cheaper over here; eating out is almost on par with cooking at home, clothes are cheaper, electronics are cheaper, cars are a lot cheaper. My salary is higher; I make<mask> is about market-rate for<mask> I do, in my location, with my experience and age. At today's exchange rate, it's a little shy of £100k.<mask> I moved to London -<mask> living costs (rent, at least) is on par with SF, I'd expect to make about half that,<mask> I'm lucky. [NEWLINE] [NEWLINE] Flip side; there are a lot of jobs
Label encoding: <s>As someone who moved from the UK to the US right after graduating, I can understand. [NEWLINE] [NEWLINE] At first, I thought it would be amazing. After I graduated, I spent 2 months visiting a bunch of the big cities in the US, and we fell in love. It wasn't long after that, that I got a job offer to move over here (San Francisco). [NEWLINE] [NEWLINE] Before I actually lived here, the idea of living here was very romantic. [NEWLINE] [NEWLINE] In truth, it's just a country. There are some things I absolutely love about the US, that are less about the politics. The geography is excellent. There aren't really any places left in the UK that aren't completely inhabited, and farmed for centuries. And where it does exist, it's only a couple of miles to civilization. (Ok, let's not count Scotland. But, that shit's way too cold). I love how different parts of the US are. If you look down the street in California, it looks different to how it is in New York, Florida, Washington, etc. In that sense, there's a lot to explore. I've also spent some time in Hawaii, and to be honest, it's beautiful. There isn't quite the equivalent in the UK, Gibraltar doesn't really compare. [NEWLINE] [NEWLINE] On the flip side; I can see a lot of problems. Large parts of San Francisco are a ghetto. The number of homeless people I walk past in the 15 minute walk to work, astonishes me. So does the number of $250k cars I see every day. [NEWLINE] [NEWLINE] One common problem people have, is the medical care. No, it's not free. My employer pays my insurance, my deductable (the excess, to the more civilized people), is $500 a year. I don't consider that too bad really, my taxes are lower than they would be in the UK. Things are generally cheaper over here; eating out is almost on par with cooking at home, clothes are cheaper, electronics are cheaper, cars are a lot cheaper. My salary is higher; I make what is about market-rate for what I do, in my location, with my experience and age. At today's exchange rate, it's a little shy of £100k. If I moved to London - where living costs (rent, at least) is on par with SF, I'd expect to make about half that, if I'm lucky. [NEWLINE] [NEWLINE] Flip side; there are a lot of jobs
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The books experienced a very significant change in pace after the third book.<mask><mask> whether you liked them or hated them, they introduced many new characters in multiple new regions, had to catch up with all of the previous characters, and provided more backstory about the world that was essential to understanding all of the plotting and magic that was going on. The last two books had a lot of jobs to do involving introduction, and the introductory parts of characters are rarely the most interesting (especially in a series<mask> the audience can take multiple seasons to warm up to characters). The fact that the show hasn't slowed to an absolute crawl<mask> adapting this much slower material already elevates it above the books in a certain respect. [NEWLINE] [NEWLINE] <mask> for your points: [NEWLINE] [NEWLINE] Rapey suckage: Dany's rape scene doesn't substantially differentiate her from the books. Frankly, that scene in the books was still rape (<mask> you give a 13-year-old girl to the general of a massive barbarian army and have her try and say no, it's probably not going to work),<mask> the books don't stand up comparatively. The rape even served a purpose, highlighting the starting point of Dany's journey from powerless to powerful. Cersei's rape in the show wasn't even supposed to be a rape (just a thoroughly appalling sex scene),<mask> I hope we can chalk that up to bad editing. The most recent rape scene is still too fresh to fairly judge (for all we know, it works perfectly into the context of the rest of the season). [NEWLINE] [NEWLINE] Staging: There are definitely notable uses of good staging,<mask> I will admit that Game of Thrones can be lackluster in this respect more often that I'd like. To be fair, Game of Thrones does strive for a certain level of realism (<mask> best<mask> a show with dragons and zombies can). The prosaic staging could be seen<mask> adherence to this guideline, with the director not distorting our perceptions of the characters to make the fantasy world seem more mundane and relatable. [NEWLINE] [NEWLINE] Dialogue: Yeah, they can't all be winners. The writers for the show have to deal with a massive handicap<mask> : they need to convey the books' storyline and characters,<mask> their main viewing audience has the attention span of a goldfish. Maybe you didn't suffer from this,<mask> among the show watchers I know,<mask> a character doesn't show up for a few episodes, they will likely forget their name (<mask> they even knew it in the first place). It
Label encoding: <s>The books experienced a very significant change in pace after the third book. Regardless of whether you liked them or hated them, they introduced many new characters in multiple new regions, had to catch up with all of the previous characters, and provided more backstory about the world that was essential to understanding all of the plotting and magic that was going on. The last two books had a lot of jobs to do involving introduction, and the introductory parts of characters are rarely the most interesting (especially in a series where the audience can take multiple seasons to warm up to characters). The fact that the show hasn't slowed to an absolute crawl while adapting this much slower material already elevates it above the books in a certain respect. [NEWLINE] [NEWLINE] As for your points: [NEWLINE] [NEWLINE] Rapey suckage: Dany's rape scene doesn't substantially differentiate her from the books. Frankly, that scene in the books was still rape ( if you give a 13-year-old girl to the general of a massive barbarian army and have her try and say no, it's probably not going to work), so the books don't stand up comparatively. The rape even served a purpose, highlighting the starting point of Dany's journey from powerless to powerful. Cersei's rape in the show wasn't even supposed to be a rape (just a thoroughly appalling sex scene), so I hope we can chalk that up to bad editing. The most recent rape scene is still too fresh to fairly judge (for all we know, it works perfectly into the context of the rest of the season). [NEWLINE] [NEWLINE] Staging: There are definitely notable uses of good staging, but I will admit that Game of Thrones can be lackluster in this respect more often that I'd like. To be fair, Game of Thrones does strive for a certain level of realism ( as best as a show with dragons and zombies can). The prosaic staging could be seen as adherence to this guideline, with the director not distorting our perceptions of the characters to make the fantasy world seem more mundane and relatable. [NEWLINE] [NEWLINE] Dialogue: Yeah, they can't all be winners. The writers for the show have to deal with a massive handicap though : they need to convey the books' storyline and characters, when their main viewing audience has the attention span of a goldfish. Maybe you didn't suffer from this, but among the show watchers I know, if a character doesn't show up for a few episodes, they will likely forget their name ( if they even knew it in the first place). It
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've seen a lot in the comments, relating to<mask> you compared homosexuality and incest. Most of it arguing that people are born gay, and<mask> its natural, and people aren't born incestuous,<mask> its unnatural. [NEWLINE] [NEWLINE] [STARTQ] some studies are suggesting that homosexuality is not a choice or preference<mask> just a part of them<mask> conception. Incest is a choice. [ENDQ] [NEWLINE] [STARTQ] People are born gay. Incest has a lot of factors psychologically that play into it. Being gay is like being born a certain race or in a certain country they do not choose it. Pretty sure incest doesn't work that way. [ENDQ] [NEWLINE] Those are just a few examples I've seen. And I can tell you, that with 100% certainty, that you are not born gay, it is a choice.<mask>?<mask> I went through this choice in my own way. I have to explain this in a way that can be understood... Up until a few years ago, I was 100%, positively heterosexual. I became a furry recently and suddenly realized that 75% of anything furry would have some homosexual innuendo or something. Rather than shunning the furry fandom, berating them, leaving and never returning, or anything like that, I accepted this to be a fact of the fandom. I followed my motto, live and let live. I don't bother them, they don't bother me. (This<mask> ties in to<mask> I'm not against homosexual marriage, I wouldn't want them hating me for marrying the person I love,<mask><mask> should I hate them for marrying the person they love?) [NEWLINE] [NEWLINE] It became every day stuff for me. I'd see more homosexual stuff than heterosexual stuff. I accepted it. Keep in mind, I'm not 100%, absolutely homosexual now,<mask> I am much more open to it. 5 years ago I'd have kicked a homosexual person, now I open to the idea of homosexuality.<mask><mask>, just recently (like this week) I've "fallen in love" with a character from a show I've been watching, and its a man. I'd never have done that 5 years ago. [NEWLINE] [NEWLINE] **TL;DR:** I made the CHOICE to become more open to homosexuality and now I'm slowly working my way towards it, and<mask> I wanted to, I could deny it all I wanted and I'd never move towards it again, once again, MY CHOICE. [NEWLINE] [NEWLINE] Anyways, back to incest. Almost every case stems from love. And I'm pretty sure
Label encoding: <s>I've seen a lot in the comments, relating to how you compared homosexuality and incest. Most of it arguing that people are born gay, and thus its natural, and people aren't born incestuous, so its unnatural. [NEWLINE] [NEWLINE] [STARTQ] some studies are suggesting that homosexuality is not a choice or preference but just a part of them since conception. Incest is a choice. [ENDQ] [NEWLINE] [STARTQ] People are born gay. Incest has a lot of factors psychologically that play into it. Being gay is like being born a certain race or in a certain country they do not choose it. Pretty sure incest doesn't work that way. [ENDQ] [NEWLINE] Those are just a few examples I've seen. And I can tell you, that with 100% certainty, that you are not born gay, it is a choice. Why? Because I went through this choice in my own way. I have to explain this in a way that can be understood... Up until a few years ago, I was 100%, positively heterosexual. I became a furry recently and suddenly realized that 75% of anything furry would have some homosexual innuendo or something. Rather than shunning the furry fandom, berating them, leaving and never returning, or anything like that, I accepted this to be a fact of the fandom. I followed my motto, live and let live. I don't bother them, they don't bother me. (This also ties in to why I'm not against homosexual marriage, I wouldn't want them hating me for marrying the person I love, so why should I hate them for marrying the person they love?) [NEWLINE] [NEWLINE] It became every day stuff for me. I'd see more homosexual stuff than heterosexual stuff. I accepted it. Keep in mind, I'm not 100%, absolutely homosexual now, but I am much more open to it. 5 years ago I'd have kicked a homosexual person, now I open to the idea of homosexuality. In fact, just recently (like this week) I've "fallen in love" with a character from a show I've been watching, and its a man. I'd never have done that 5 years ago. [NEWLINE] [NEWLINE] **TL;DR:** I made the CHOICE to become more open to homosexuality and now I'm slowly working my way towards it, and if I wanted to, I could deny it all I wanted and I'd never move towards it again, once again, MY CHOICE. [NEWLINE] [NEWLINE] Anyways, back to incest. Almost every case stems from love. And I'm pretty sure
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>After a couple philosophy classes and some long winded discussions it seems<mask><mask> many people find it hard to believe that free will doesn't exist in any real scientific way. [NEWLINE] [NEWLINE] Let's start off with premise #1 [NEWLINE] [NEWLINE] Everything that is observable in the universe follows laws of nature. All laws of nature follow the universal law of causation, which states that everything in the universe has a cause and is<mask> an effect of that cause. [NEWLINE] [NEWLINE] <mask> premise #1 is true then everything in the physical world must be determined. [NEWLINE] [NEWLINE] <mask> everything is determined than there is no such thing<mask> free will. [NEWLINE] <mask> every behaviour is determined by laws of causation it is physically impossible for someone to be free,<mask> that freedom would require the nullification of the laws of causation. In other words to believe in free will is<mask> to believe in spontaneous magic, it's to believe that humans possess powers that could only be explained by the supernatural. [NEWLINE] [NEWLINE] For example: In most legal traditions it is customary to judge whether someone is guilty of x crime by accumulating evidence in order to prove that  x person intentionally did this crime. Let's assume that there was enough evidence to prove without doubt that Greg Smith robbed a convenience store. Now that we have proof that it was Greg, we should ask ourselves<mask> Greg really chose to rob the store.<mask> soon<mask> you ask yourself that kind of question the only scientific recourse you have is to understand Greg<mask> a causal agent.<mask> caused Greg to rob the store? Was it the fact that Greg is from a low socio-economic class? Was it the fact that Greg lost his job 2 weeks ago and Greg wont be able to feed his kids<mask> he doesn't find a way to get more money? Is it the fact that Greg was abused<mask> child? Did Greg rob the store due to complex laws of causality that are reducible to both biological and environmental reasons? Or did Greg simply rob the store<mask> he felt like it? [NEWLINE] [NEWLINE] Even more simple decisions like choosing between a chocolate bar and an apple can be reducible to complex causal interplay between biological and environmental laws. Did you choose the chocolate bar<mask> you like it better? Or is it<mask> we have an evolutionary pull towards sugar and fat dense food? Is it<mask> you read an article on naturalnews.com talking about<mask> dark chocolate can cure your pancreas cancer? Or did you chose the apple<mask> you grew up on an orchard farm and apples remind you of your innocent childhood? or did you choose the apple<mask>
Label encoding: <s>After a couple philosophy classes and some long winded discussions it seems as though many people find it hard to believe that free will doesn't exist in any real scientific way. [NEWLINE] [NEWLINE] Let's start off with premise #1 [NEWLINE] [NEWLINE] Everything that is observable in the universe follows laws of nature. All laws of nature follow the universal law of causation, which states that everything in the universe has a cause and is thus an effect of that cause. [NEWLINE] [NEWLINE] If premise #1 is true then everything in the physical world must be determined. [NEWLINE] [NEWLINE] If everything is determined than there is no such thing as free will. [NEWLINE] If every behaviour is determined by laws of causation it is physically impossible for someone to be free, as that freedom would require the nullification of the laws of causation. In other words to believe in free will is also to believe in spontaneous magic, it's to believe that humans possess powers that could only be explained by the supernatural. [NEWLINE] [NEWLINE] For example: In most legal traditions it is customary to judge whether someone is guilty of x crime by accumulating evidence in order to prove that  x person intentionally did this crime. Let's assume that there was enough evidence to prove without doubt that Greg Smith robbed a convenience store. Now that we have proof that it was Greg, we should ask ourselves if Greg really chose to rob the store. As soon as you ask yourself that kind of question the only scientific recourse you have is to understand Greg as a causal agent. What caused Greg to rob the store? Was it the fact that Greg is from a low socio-economic class? Was it the fact that Greg lost his job 2 weeks ago and Greg wont be able to feed his kids if he doesn't find a way to get more money? Is it the fact that Greg was abused as child? Did Greg rob the store due to complex laws of causality that are reducible to both biological and environmental reasons? Or did Greg simply rob the store because he felt like it? [NEWLINE] [NEWLINE] Even more simple decisions like choosing between a chocolate bar and an apple can be reducible to complex causal interplay between biological and environmental laws. Did you choose the chocolate bar because you like it better? Or is it because we have an evolutionary pull towards sugar and fat dense food? Is it because you read an article on naturalnews.com talking about how dark chocolate can cure your pancreas cancer? Or did you chose the apple because you grew up on an orchard farm and apples remind you of your innocent childhood? or did you choose the apple because
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] dishwashers require the dishes to be "pre cleaned". This already defeats the purpose,<mask> did I pay hundreds of dollars for a machine that doesn't have food processing capabilities? [ENDQ] [NEWLINE] Incorrect. My dishwasher doesn't require things to be "pre-cleaned".<mask> even<mask> it were true, you've got a bad assumption. Something that relieves you from X amount of time and work is a bad idea<mask> it doesn't relieve you from all of it? [NEWLINE] [NEWLINE] [STARTQ] dishwashers rarely removed ingrained or dried gunk. This only worsens<mask> you charge the dishwasher over a couple of days, and the older plates have already dried up. You have to take the still dirty plates and clean them manually. [ENDQ] [NEWLINE] Incorrect. I try not to let dishes get bad, and dishwashers won't handle large pieces of food properly,<mask> they're good on everything else, including plates that have some tough to get off stuff. [NEWLINE] [NEWLINE] [STARTQ] dishwashers don't properly dry "deep" items like Tupperware and pots,<mask> you aren't careful taking these out, you'll get everything else wet, once again making the process useless. [ENDQ] [NEWLINE] Anything that corrects water should be places<mask> that the concave area faces down.<mask> you do that properly, then you should have no problem washing tupperware and pots. I do<mask> regularly. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> overall that dishwashers are overrated, undercapable, overpriced appliances that have no place in a modern kitchen. [ENDQ] [NEWLINE] It depends on who's kitchen you're talking about. It sounds to me like a couple of things are going on. You don't understand<mask> to properly load and use a dishwasher, you have a bad performing dishwasher and you have a lot of bad assumptions about<mask> a dishwasher is and<mask> it's supposed to help. [NEWLINE] [NEWLINE] The ordinary process of cleaning dishes is hard and timestaking for a lot of people. There are people with large families who go through enormous amounts of dirty dishes in a day. There are disabled people who can't physically stand at the sink long enough to do dishes. There are busy people who want their chores to take a little less long. A dishwasher isn't magical. It isn't going to scrub your pots or dry everything perfectly. It's not going to load itself or put your dishes away. It isn't going to tuck your kids into bed at night.<mask><mask> you use it properly, a
Label encoding: <s> [STARTQ] dishwashers require the dishes to be "pre cleaned". This already defeats the purpose, why did I pay hundreds of dollars for a machine that doesn't have food processing capabilities? [ENDQ] [NEWLINE] Incorrect. My dishwasher doesn't require things to be "pre-cleaned". But even if it were true, you've got a bad assumption. Something that relieves you from X amount of time and work is a bad idea because it doesn't relieve you from all of it? [NEWLINE] [NEWLINE] [STARTQ] dishwashers rarely removed ingrained or dried gunk. This only worsens when you charge the dishwasher over a couple of days, and the older plates have already dried up. You have to take the still dirty plates and clean them manually. [ENDQ] [NEWLINE] Incorrect. I try not to let dishes get bad, and dishwashers won't handle large pieces of food properly, but they're good on everything else, including plates that have some tough to get off stuff. [NEWLINE] [NEWLINE] [STARTQ] dishwashers don't properly dry "deep" items like Tupperware and pots, if you aren't careful taking these out, you'll get everything else wet, once again making the process useless. [ENDQ] [NEWLINE] Anything that corrects water should be places so that the concave area faces down. If you do that properly, then you should have no problem washing tupperware and pots. I do so regularly. [NEWLINE] [NEWLINE] [STARTQ] I think overall that dishwashers are overrated, undercapable, overpriced appliances that have no place in a modern kitchen. [ENDQ] [NEWLINE] It depends on who's kitchen you're talking about. It sounds to me like a couple of things are going on. You don't understand how to properly load and use a dishwasher, you have a bad performing dishwasher and you have a lot of bad assumptions about what a dishwasher is and how it's supposed to help. [NEWLINE] [NEWLINE] The ordinary process of cleaning dishes is hard and timestaking for a lot of people. There are people with large families who go through enormous amounts of dirty dishes in a day. There are disabled people who can't physically stand at the sink long enough to do dishes. There are busy people who want their chores to take a little less long. A dishwasher isn't magical. It isn't going to scrub your pots or dry everything perfectly. It's not going to load itself or put your dishes away. It isn't going to tuck your kids into bed at night. But if you use it properly, a
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0703, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can't speak for every fat person,<mask> I can tell you my experience. I'm not really sure<mask> you draw the line for extreme cases - I don't need a scooter at the store or a walking stick to get around, I can walk across campus to my classes (<mask> this is Florida and it's hot<mask> hell,<mask> I take the bus often, at least to the classes that are more than a mile away - we have a big campus). I can walk up the flight of stairs to my apartment without getting winded,<mask> I'll get a little tired after 3 or 4 flights.<mask>, you can decide for yourself whether I'm the person you're thinking of. (Incidentally, I do have multiple symptoms of PCOS, a hormone imbalance that can make it hard to lose weight,<mask> I know that's not the main cause of my weight. The main cause is that I don't spend time focusing on losing weight.) [NEWLINE] [NEWLINE] I've always been overweight, ever<mask> I was a little kid. I didn't have siblings to play with, or neighbors my age,<mask> I never learned very good social skills or<mask> to play with others.<mask>,<mask> there was no one for me to play with at home, I read. All the time. In 4th grade, I remember going to the library instead of recess<mask> I didn't know<mask> to make friends or play with other kids. Add a case of asthma into the mix, and I never learned good exercise habits. [NEWLINE] [NEWLINE] In high school, my weight problems were compounded by stress eating and spending most of my time working on school or reading instead of doing physical things.<mask> these habits weren't good for my weight, they got me a National Merit Scholarship that covers all of my tuition, books, and housing, and leaves me a little more money for other living expenses during the fall and spring semesters.<mask><mask> that probably proves that I'm not stupid or lazy, just that I have different priorities than other people. [NEWLINE] [NEWLINE] <mask> I spend<mask> much time working on school/extracurriculars, I don't have time to work a regular job (I'm on my school's debate team and we travel a lot.<mask> we're gone<mask> much, I can't find a job during the school year). That means I don't have a lot of extra cash lying around. Unhealthy food tends to keep me from being hungry longer than healthy food for the same amount of money or less.<mask> of that
Label encoding: <s>I can't speak for every fat person, but I can tell you my experience. I'm not really sure where you draw the line for extreme cases - I don't need a scooter at the store or a walking stick to get around, I can walk across campus to my classes ( although this is Florida and it's hot as hell, so I take the bus often, at least to the classes that are more than a mile away - we have a big campus). I can walk up the flight of stairs to my apartment without getting winded, but I'll get a little tired after 3 or 4 flights. So, you can decide for yourself whether I'm the person you're thinking of. (Incidentally, I do have multiple symptoms of PCOS, a hormone imbalance that can make it hard to lose weight, but I know that's not the main cause of my weight. The main cause is that I don't spend time focusing on losing weight.) [NEWLINE] [NEWLINE] I've always been overweight, ever since I was a little kid. I didn't have siblings to play with, or neighbors my age, so I never learned very good social skills or how to play with others. Also, because there was no one for me to play with at home, I read. All the time. In 4th grade, I remember going to the library instead of recess because I didn't know how to make friends or play with other kids. Add a case of asthma into the mix, and I never learned good exercise habits. [NEWLINE] [NEWLINE] In high school, my weight problems were compounded by stress eating and spending most of my time working on school or reading instead of doing physical things. While these habits weren't good for my weight, they got me a National Merit Scholarship that covers all of my tuition, books, and housing, and leaves me a little more money for other living expenses during the fall and spring semesters. I think that probably proves that I'm not stupid or lazy, just that I have different priorities than other people. [NEWLINE] [NEWLINE] Because I spend so much time working on school/extracurriculars, I don't have time to work a regular job (I'm on my school's debate team and we travel a lot. Because we're gone so much, I can't find a job during the school year). That means I don't have a lot of extra cash lying around. Unhealthy food tends to keep me from being hungry longer than healthy food for the same amount of money or less. Because of that
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Isn't your entire post based on your personal experience?<mask> we take that away from your argument here,<mask> do you really have left? [NEWLINE] [NEWLINE] <mask> seriously. Experience tends to have a tempering effect on information that can often be difficult to quantify. Just<mask> you've arrived at your own point of view on experience (through your experience), events change the way we can view identical information. [NEWLINE] [NEWLINE] <mask> someone researches loss for instance, it may be easy to develop an academic view of the subject,<mask> that knowledge does not confer<mask> the actual experience of loss is like, or<mask> difficult it can be to deal with. In the same way that knowing<mask> to lose weight doesn't often translate into success with losing weight. [NEWLINE] [NEWLINE] The greatest benefit my own experience has conferred on me is a great awareness of<mask> it's like to be wrong. I've been wrong about a lot. It has gradually made me less and less certain over the years and<mask> I'm a lot less concerned than I once was about whether or not I'm right. It<mask> tends to make me more open to being wrong,<mask><mask> someone points out a flaw in my reasoning I'm quicker to give up my old point of view in favor of seeing things differently. [NEWLINE] [NEWLINE] A study conducted by Richard West at James Madison University found that contrary to<mask> one might expect, people who scored very high on SATs or other standardized methods of evaluating cognitive or collegiate ability had a surprising vulnerability to certain kinds of test questions, such<mask> : [NEWLINE] [NEWLINE] A bat and ball cost a dollar and ten cents. The bat costs a dollar more than the ball.<mask> much does the ball cost? [NEWLINE] [NEWLINE] <mask> our initial thought might be that the ball costs ten cents, this is actually incorrect. The correct answer is that the bat costs a dollar and five cents and that the ball costs five cents. [NEWLINE] [NEWLINE] Intelligent people, Richard West found, were more prone to relying on heuristics (or mental short cuts) to solve problems and that this made them MORE likely than the average person to be vulnerable to certain cognitive biases. [NEWLINE] [NEWLINE] In some instances, it is the very clarity with which we see that world that is our biggest barrier to seeing it accurately. This is<mask> experience is most helpful. It's not just having the book knowledge that is important. It's having the relevant life experience that brings that knowledge into its full three dimensions. [NEWLINE] [NEWLINE] <mask> your own arguments have developed, you've become frustrated with the ability of others, who often may not argue
Label encoding: <s>Isn't your entire post based on your personal experience? If we take that away from your argument here, what do you really have left? [NEWLINE] [NEWLINE] But seriously. Experience tends to have a tempering effect on information that can often be difficult to quantify. Just as you've arrived at your own point of view on experience (through your experience), events change the way we can view identical information. [NEWLINE] [NEWLINE] If someone researches loss for instance, it may be easy to develop an academic view of the subject, but that knowledge does not confer what the actual experience of loss is like, or how difficult it can be to deal with. In the same way that knowing how to lose weight doesn't often translate into success with losing weight. [NEWLINE] [NEWLINE] The greatest benefit my own experience has conferred on me is a great awareness of what it's like to be wrong. I've been wrong about a lot. It has gradually made me less and less certain over the years and so I'm a lot less concerned than I once was about whether or not I'm right. It also tends to make me more open to being wrong, so if someone points out a flaw in my reasoning I'm quicker to give up my old point of view in favor of seeing things differently. [NEWLINE] [NEWLINE] A study conducted by Richard West at James Madison University found that contrary to what one might expect, people who scored very high on SATs or other standardized methods of evaluating cognitive or collegiate ability had a surprising vulnerability to certain kinds of test questions, such as : [NEWLINE] [NEWLINE] A bat and ball cost a dollar and ten cents. The bat costs a dollar more than the ball. How much does the ball cost? [NEWLINE] [NEWLINE] While our initial thought might be that the ball costs ten cents, this is actually incorrect. The correct answer is that the bat costs a dollar and five cents and that the ball costs five cents. [NEWLINE] [NEWLINE] Intelligent people, Richard West found, were more prone to relying on heuristics (or mental short cuts) to solve problems and that this made them MORE likely than the average person to be vulnerable to certain cognitive biases. [NEWLINE] [NEWLINE] In some instances, it is the very clarity with which we see that world that is our biggest barrier to seeing it accurately. This is where experience is most helpful. It's not just having the book knowledge that is important. It's having the relevant life experience that brings that knowledge into its full three dimensions. [NEWLINE] [NEWLINE] As your own arguments have developed, you've become frustrated with the ability of others, who often may not argue
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Harming and killing animals for pleasure. Just<mask> shooting someone in self-defense is not commensurate with shooting someone to satisfy a sadistic urge.Killing animals for food<mask> we have no other choice for survival is not morally equivalent to killing animals<mask> we have plentiful alternatives. Violence committed to save a life is never analogous to violence committed for pleasure or profit. [NEWLINE] This distinction is crucial for several reasons, the first of which is that it clarifies a serious category error in the thinking of people who insist that meat-eating is “natural”—and<mask> morally neutral—<mask> other animals eat animals. It’s important to realize that, with a few exceptions,<mask> humans kill other animals for food, we’re not doing<mask> animals do in nature. Humans have no biological or nutritional need to consume meat or any animal products.<mask> animals kill other animals for food, they do<mask> they must, in order to survive; they have no choice in the matter. Many humans,<mask><mask><mask><mask>, do have a choice, and<mask> people with access to non-animal food options choose to consume animals anyway–<mask> they can, or<mask> they like the taste– they are not killing from necessity,<mask> animals (and some humans in crisis or subsistence situations) do. Whether we’re talking about a lion taking down a water buffalo, or a human in some remote or impoverished location forced to hunt in order to feed her family: these are acts of necessity, and do not equate to, nor justify, wholly unnecessary harm to animals. There is no analogy to be found in nature for the massive harm we inflict on animals merely for pleasure.Another reason it’s important to recognize the necessity/pleasure distinction is that harming animals for pleasure goes against core values most of us hold in common—which is<mask>, for example, millions of us are outraged over dog fighting, and<mask> we oppose dog fighting on principle. The notion of deriving pleasure from violence toward animals is repulsive to us;<mask><mask> can we justify harming animals for the taste of their flesh?<mask> can it be wrong to harm for pleasure in one instance, and not the other? The same reasons that compel us to oppose dog fighting compel us to abstain from killing animals we don’t need to eat: namely, that it is wrong to harm animals for pleasure, and it is wrong to kill animals for pleasure.Finally, to harm animals for pleasure is<mask>, ultimately, to harm ourselves. Constantly
Label encoding: <s>Harming and killing animals for pleasure. Just as shooting someone in self-defense is not commensurate with shooting someone to satisfy a sadistic urge.Killing animals for food when we have no other choice for survival is not morally equivalent to killing animals when we have plentiful alternatives. Violence committed to save a life is never analogous to violence committed for pleasure or profit. [NEWLINE] This distinction is crucial for several reasons, the first of which is that it clarifies a serious category error in the thinking of people who insist that meat-eating is “natural”—and therefore morally neutral— because other animals eat animals. It’s important to realize that, with a few exceptions, when humans kill other animals for food, we’re not doing what animals do in nature. Humans have no biological or nutritional need to consume meat or any animal products. When animals kill other animals for food, they do as they must, in order to survive; they have no choice in the matter. Many humans, on the other hand, do have a choice, and when people with access to non-animal food options choose to consume animals anyway– because they can, or because they like the taste– they are not killing from necessity, as animals (and some humans in crisis or subsistence situations) do. Whether we’re talking about a lion taking down a water buffalo, or a human in some remote or impoverished location forced to hunt in order to feed her family: these are acts of necessity, and do not equate to, nor justify, wholly unnecessary harm to animals. There is no analogy to be found in nature for the massive harm we inflict on animals merely for pleasure.Another reason it’s important to recognize the necessity/pleasure distinction is that harming animals for pleasure goes against core values most of us hold in common—which is why, for example, millions of us are outraged over dog fighting, and why we oppose dog fighting on principle. The notion of deriving pleasure from violence toward animals is repulsive to us; so how can we justify harming animals for the taste of their flesh? How can it be wrong to harm for pleasure in one instance, and not the other? The same reasons that compel us to oppose dog fighting compel us to abstain from killing animals we don’t need to eat: namely, that it is wrong to harm animals for pleasure, and it is wrong to kill animals for pleasure.Finally, to harm animals for pleasure is also, ultimately, to harm ourselves. Constantly
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I got a job and the office was decorated with posters of girls in bikinis on cars, [ENDQ] [NEWLINE] That would be<mask> it is inappropriate, not<mask> it is degrading. It's just<mask> inappropriate to show up in a bathing suit, or pajamas. Certainly this guy was dressed inappropriately for an office job; that's not the issue. The issue was that he was attacked for it being *sexist*. There's nothing sexist about liking to look at scantily clad women. That is the baseline of male sexuality and it is nothing to be ashamed of. Nor should women be ashamed of their sexuality or sexual interests. It is inappropriate to display them at work, yes, even for women.<mask> again, that isn't the issue here. [NEWLINE] [NEWLINE] <mask> this wasn't an office job. This was a once-in-a-lifetime event -- a world's first, and the fruition of a decade or more of work. It isn't "Wednesday at the office". People do unique and meaningful things to mark those occasions. Canadian astronaut Steve MacLean wore a Canadian hockey jersey at NASA Mission Control the day he was Capcom for fellow Canadian astronaut Chris Hadfield installing the Canadarm2 to the International Space Station, using the shuttle's Canadarm. A hockey jersey is<mask> not appropriate office attire,<mask> it was meaningful for a meaningful event. [NEWLINE] [NEWLINE] In this case, Dr. Taylor wore a birthday gift shirt given to him by a close (female) friend, based on his quirky and tacky wardrobe. To him it shows scientists and engineers<mask> less stuffy and more fun, something we actually need more of. It was a quirky thing, and with some sentimental value. There is nothing sexist about wearing the shirt. It took no rights from women, made no suggestion that women were of lesser importance. The content of the shirt was simply a tacky and cheesy display, no different<mask> it were Velvet Elvis or dogs playing cards. [NEWLINE] [NEWLINE] That people would call it sexist is the problem.<mask> a woman wears a jersey of a men's sports team, is she saying that all men are good for is playing sports -- that this all they can contribute? Of course not. It is simply wrong to say this is sexist. [NEWLINE] [NEWLINE] [STARTQ] The main way my gender was thought of was T&amp;A. [ENDQ] [NEWLINE] Hold on. The problem wasn't that women were (and are)<mask> of in terms of beauty. The problem was that women weren't<mask> actively thought of, or treated,
Label encoding: <s> [STARTQ] If I got a job and the office was decorated with posters of girls in bikinis on cars, [ENDQ] [NEWLINE] That would be because it is inappropriate, not because it is degrading. It's just as inappropriate to show up in a bathing suit, or pajamas. Certainly this guy was dressed inappropriately for an office job; that's not the issue. The issue was that he was attacked for it being *sexist*. There's nothing sexist about liking to look at scantily clad women. That is the baseline of male sexuality and it is nothing to be ashamed of. Nor should women be ashamed of their sexuality or sexual interests. It is inappropriate to display them at work, yes, even for women. But again, that isn't the issue here. [NEWLINE] [NEWLINE] But this wasn't an office job. This was a once-in-a-lifetime event -- a world's first, and the fruition of a decade or more of work. It isn't "Wednesday at the office". People do unique and meaningful things to mark those occasions. Canadian astronaut Steve MacLean wore a Canadian hockey jersey at NASA Mission Control the day he was Capcom for fellow Canadian astronaut Chris Hadfield installing the Canadarm2 to the International Space Station, using the shuttle's Canadarm. A hockey jersey is also not appropriate office attire, but it was meaningful for a meaningful event. [NEWLINE] [NEWLINE] In this case, Dr. Taylor wore a birthday gift shirt given to him by a close (female) friend, based on his quirky and tacky wardrobe. To him it shows scientists and engineers as less stuffy and more fun, something we actually need more of. It was a quirky thing, and with some sentimental value. There is nothing sexist about wearing the shirt. It took no rights from women, made no suggestion that women were of lesser importance. The content of the shirt was simply a tacky and cheesy display, no different if it were Velvet Elvis or dogs playing cards. [NEWLINE] [NEWLINE] That people would call it sexist is the problem. If a woman wears a jersey of a men's sports team, is she saying that all men are good for is playing sports -- that this all they can contribute? Of course not. It is simply wrong to say this is sexist. [NEWLINE] [NEWLINE] [STARTQ] The main way my gender was thought of was T&amp;A. [ENDQ] [NEWLINE] Hold on. The problem wasn't that women were (and are) though of in terms of beauty. The problem was that women weren't as actively thought of, or treated,
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just a note, your argument would be even more convincing<mask> you mention the fact that pretty much everyone alive now has some ancestors that were probably in the same situation. Think back just about half a century and a 35 and 17 year old getting married in a rural area was more helpful to running a business or farm. This kind of behavior was part of a cultural norm,<mask><mask> culture has changed we now demonize this behavior. [NEWLINE] [NEWLINE] <mask> you were to look at sex<mask> purely biological needs or for the purpose of reproduction (rather than life experiences, ages, demographics, etc) then sex should start for a female once her body has hit maturity. The biological window for the female becomes more limited<mask> we apply cultural boundaries and tell women to wait on having children (Note: Having women wait is still a good idea in my mind<mask> they ~~may not~~ don't have enough life experience to raise a child).<mask>, this does not negate the fact that girls after reaching maturity will start to have urges or seek out sex. Just ask most girls now<mask> they first lost their virginity. Many of them will say between the ages of 14-20. And that is a little terrifying for me to think about.<mask><mask> about<mask> much I have learned about myself in just the last year and<mask> much I have changed/grown,<mask> between those years you grow and change exponentially. I remember being 19 and thinking the girl I had been having sex with for over a year was going to be the one I married. And now I look back and I realize<mask> little I knew. I am grateful that didn't work out. I see /r/relationships and /r/sex posts with people who are 16-20 freaking out about their relationship not working, and the only thing I can think is *"You are<mask> fucking young! You don't have to settle with someone now, you still have your whole life ahead of you!"* [NEWLINE] [NEWLINE] The problem is that (i.e. thinking they are too young) is the mentality we adopt. We often forget the urges we had<mask> we were in that same position.<mask><mask> the older we get the harder it is for us to remember the emotions and cravings we had<mask> teens, and our bodies were ready,<mask> our minds were torn. And<mask><mask><mask> thinking we tend to blame the male ("who should know better"),<mask> a 17 year old girl can be just<mask>,<mask> not more than, promiscuous<mask> someone 5-10 years older;
Label encoding: <s>Just a note, your argument would be even more convincing if you mention the fact that pretty much everyone alive now has some ancestors that were probably in the same situation. Think back just about half a century and a 35 and 17 year old getting married in a rural area was more helpful to running a business or farm. This kind of behavior was part of a cultural norm, but since culture has changed we now demonize this behavior. [NEWLINE] [NEWLINE] If you were to look at sex as purely biological needs or for the purpose of reproduction (rather than life experiences, ages, demographics, etc) then sex should start for a female once her body has hit maturity. The biological window for the female becomes more limited as we apply cultural boundaries and tell women to wait on having children (Note: Having women wait is still a good idea in my mind since they ~~may not~~ don't have enough life experience to raise a child). However, this does not negate the fact that girls after reaching maturity will start to have urges or seek out sex. Just ask most girls now when they first lost their virginity. Many of them will say between the ages of 14-20. And that is a little terrifying for me to think about. I think about how much I have learned about myself in just the last year and how much I have changed/grown, but between those years you grow and change exponentially. I remember being 19 and thinking the girl I had been having sex with for over a year was going to be the one I married. And now I look back and I realize how little I knew. I am grateful that didn't work out. I see /r/relationships and /r/sex posts with people who are 16-20 freaking out about their relationship not working, and the only thing I can think is *"You are so fucking young! You don't have to settle with someone now, you still have your whole life ahead of you!"* [NEWLINE] [NEWLINE] The problem is that (i.e. thinking they are too young) is the mentality we adopt. We often forget the urges we had when we were in that same position. I think the older we get the harder it is for us to remember the emotions and cravings we had as teens, and our bodies were ready, but our minds were torn. And because of this thinking we tend to blame the male ("who should know better"), but a 17 year old girl can be just as, if not more than, promiscuous as someone 5-10 years older;
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, people with more money buy more stuff.<mask>'s your point? [NEWLINE] <mask> economic activity is based entirely on<mask> gets bought. <mask> people do not get more money to buy stuff, the economy stagnates. <mask> 'd the economy crash in 2008?  Credit dried up, and people had to make up the negative savings rate with wages and not buying stuff.  GDP went severely negative, basically until enough people had paid off enough debt that they could start buying stuff again.  The fact that you can't make the connection between people buying stuff and people having enough MONEY to buy stuff is astounding. [NEWLINE] Here's another fun fact.  Your pay has little to nothing to do with the value you add to a product.  Your pay is determined by your NEGOTIATING POSITION. <mask> you can add millions of dollars of value,<mask> have a crappy negotiating position, you're not going to get paid squat (Nicola Tesla). <mask><mask><mask><mask>,<mask> you are marginal in<mask> you do,<mask> you can convince a company to shell out big bucks for you, then you're going to be rolling in dough.  After all, CEO pay has almost ZERO correlation with company performance. [NEWLINE] Commodity prices. <mask> are they an indicator of utilization of the means of production?  Supply and demand. <mask> mines worldwide are able to produce X amount of iron ore a year,<mask> companies only need x-25%, then there will be mines that are closed, and the price of iron ore (and iron, and steel) will be correspondingly low,<mask> well. <mask>,<mask> demand for iron ore is x+ 25%, then the price of iron ore, steel, and iron will be sky high, ALL the mines that can produce iron ore will be running<mask> hard<mask> they can, and commodity prices will be high. [NEWLINE] <mask> they're not. [NEWLINE] The market is underutilized. [NEWLINE] Wages weren't stagnant for anyone. [NEWLINE] You see, you're going to have to show<mask> you're getting your numbers.  I'm getting mine from Pew. [NEWLINE] [URL] / [NEWLINE] Other measures of income inequality are nasty,<mask> well. [NEWLINE] [URL] [NEWLINE] Rich people spend just<mask> much of their income<mask> poor people. [NEWLINE] No.  Period.  They don't.  Period.  One of the fundamentals ways you get rich is by spending LESS than you make. [NEWLINE] The fundamental way you stay poor is you spend EVERYTHING you make.
Label encoding: <s>So, people with more money buy more stuff. What's your point? [NEWLINE] Because economic activity is based entirely on what gets bought.  If people do not get more money to buy stuff, the economy stagnates.  Why 'd the economy crash in 2008?  Credit dried up, and people had to make up the negative savings rate with wages and not buying stuff.  GDP went severely negative, basically until enough people had paid off enough debt that they could start buying stuff again.  The fact that you can't make the connection between people buying stuff and people having enough MONEY to buy stuff is astounding. [NEWLINE] Here's another fun fact.  Your pay has little to nothing to do with the value you add to a product.  Your pay is determined by your NEGOTIATING POSITION.  If you can add millions of dollars of value, but have a crappy negotiating position, you're not going to get paid squat (Nicola Tesla).  On the other hand, if you are marginal in what you do, but you can convince a company to shell out big bucks for you, then you're going to be rolling in dough.  After all, CEO pay has almost ZERO correlation with company performance. [NEWLINE] Commodity prices.  How are they an indicator of utilization of the means of production?  Supply and demand.  If mines worldwide are able to produce X amount of iron ore a year, but companies only need x-25%, then there will be mines that are closed, and the price of iron ore (and iron, and steel) will be correspondingly low, as well.  But, if demand for iron ore is x+ 25%, then the price of iron ore, steel, and iron will be sky high, ALL the mines that can produce iron ore will be running as hard as they can, and commodity prices will be high. [NEWLINE] But they're not. [NEWLINE] The market is underutilized. [NEWLINE] Wages weren't stagnant for anyone. [NEWLINE] You see, you're going to have to show where you're getting your numbers.  I'm getting mine from Pew. [NEWLINE] [URL] / [NEWLINE] Other measures of income inequality are nasty, as well. [NEWLINE] [URL] [NEWLINE] Rich people spend just as much of their income as poor people. [NEWLINE] No.  Period.  They don't.  Period.  One of the fundamentals ways you get rich is by spending LESS than you make. [NEWLINE] The fundamental way you stay poor is you spend EVERYTHING you make.
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Legally, women have more rights then men do [ENDQ] [NEWLINE] This is false, let's break it down. [NEWLINE] [NEWLINE] [STARTQ] They get children in a divorce [ENDQ] [NEWLINE] Men who request custody get at least joint custody more than 80% of the time. The only reason that the "women get the kids" idea persists is<mask> the vast majority of the time the father doesn't request custody<mask> people see the woman get the kids more often.<mask> more than 94% of the time the custody arrangement was a mutual decision by both parents and had nothing to do with the courts. [NEWLINE] [NEWLINE] [STARTQ] they get child support and alimony (<mask> often do you hear of a male getting that?) [ENDQ] [NEWLINE] Well<mask> most of the time it's the mother who is agreed to have the kids, of course they will get child support. Women get alimony more often<mask><mask> often does she make more than her husband? Child support goes to the custodial parent, there's no gender bias. Alimony goes to the spouse who made less money and<mask> women are expected to give up their jobs to take care of the kids, much more often they are the ones who make less.<mask> you want to fix these problems, then we need to fix society. [NEWLINE] [NEWLINE] [STARTQ] any report of domestic abuse might land the male in jail, even<mask> it was the woman hitting him without him hitting back [ENDQ] [NEWLINE] This one depends on<mask> in the country you are.<mask><mask> that there is a problem in assumption here<mask>. [NEWLINE] [NEWLINE] [STARTQ] You haven't even mentioned woman on male sexual assault, which a lot of people even claim is impossible. [ENDQ] [NEWLINE] This one is due to toxic masculinity. The idea that it's shameful for a woman to control a man and emasculating or makes them less of a man causing victims to be less likely to come forward and less likely to be believed. Huzzah patriarchy hurting us guys too! [NEWLINE] [NEWLINE] [STARTQ] Chances are a consensual sex act with a female will land you in jail<mask> she changed her mind the next day, even<mask> she pounced on you the day before. [ENDQ] [NEWLINE] This is just plain false. It's extremely unlikely for this to happen. Chances are a consensual sex act will just be a consensual sex act. [NEWLINE] [NEWLINE] [STARTQ] Anita Sarkeesian, who actively lies about the content of video games to advocate for them changing, and criticizes any game that matches her proposed changes<mask> being sexist anyway [ENDQ] [NEWLINE] False. [NEWLINE] [NEWLINE] [STARTQ] And<mask> anyone points out politely that she is
Label encoding: <s> [STARTQ] Legally, women have more rights then men do [ENDQ] [NEWLINE] This is false, let's break it down. [NEWLINE] [NEWLINE] [STARTQ] They get children in a divorce [ENDQ] [NEWLINE] Men who request custody get at least joint custody more than 80% of the time. The only reason that the "women get the kids" idea persists is because the vast majority of the time the father doesn't request custody so people see the woman get the kids more often. But more than 94% of the time the custody arrangement was a mutual decision by both parents and had nothing to do with the courts. [NEWLINE] [NEWLINE] [STARTQ] they get child support and alimony ( how often do you hear of a male getting that?) [ENDQ] [NEWLINE] Well because most of the time it's the mother who is agreed to have the kids, of course they will get child support. Women get alimony more often because how often does she make more than her husband? Child support goes to the custodial parent, there's no gender bias. Alimony goes to the spouse who made less money and since women are expected to give up their jobs to take care of the kids, much more often they are the ones who make less. If you want to fix these problems, then we need to fix society. [NEWLINE] [NEWLINE] [STARTQ] any report of domestic abuse might land the male in jail, even if it was the woman hitting him without him hitting back [ENDQ] [NEWLINE] This one depends on where in the country you are. I agree that there is a problem in assumption here though. [NEWLINE] [NEWLINE] [STARTQ] You haven't even mentioned woman on male sexual assault, which a lot of people even claim is impossible. [ENDQ] [NEWLINE] This one is due to toxic masculinity. The idea that it's shameful for a woman to control a man and emasculating or makes them less of a man causing victims to be less likely to come forward and less likely to be believed. Huzzah patriarchy hurting us guys too! [NEWLINE] [NEWLINE] [STARTQ] Chances are a consensual sex act with a female will land you in jail because she changed her mind the next day, even if she pounced on you the day before. [ENDQ] [NEWLINE] This is just plain false. It's extremely unlikely for this to happen. Chances are a consensual sex act will just be a consensual sex act. [NEWLINE] [NEWLINE] [STARTQ] Anita Sarkeesian, who actively lies about the content of video games to advocate for them changing, and criticizes any game that matches her proposed changes as being sexist anyway [ENDQ] [NEWLINE] False. [NEWLINE] [NEWLINE] [STARTQ] And if anyone points out politely that she is
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is naturally an incredibly controversial topic and I highly doubt that many people will agree with me,<mask> I feel compelled to discuss this. [NEWLINE] [NEWLINE] This needs to be prefaced by saying that all of this is based on averages, not individuals. I'm sure many women are more capable than many men,<mask> I'm saying is that the general trend across the board is that on average a male worker is more productive. [NEWLINE] [NEWLINE] Men work almost triple the amount of overtime hours than women. [NEWLINE] [NEWLINE] Men work more standard hours than women. [NEWLINE] [NEWLINE] Women take more sick days off work than men. [NEWLINE] [NEWLINE] Women are more likely than men to take several years of maternity leave off of work. [NEWLINE] [NEWLINE] A small percentage of women undergo severe PMS and<mask> work at a diminished capacity for several days of each month. [NEWLINE] [NEWLINE] To use anecdotal evidence, yesterday i came in to work, to find that all 8 garbages were completely filled. This was strange,<mask> the closing crew is supposed to take out the garbages every night. I asked my coworker<mask> no one had done the garbages last night, and I was effectively told that only girls were working the closing shift, and none of them wanted to touch the garbages<mask> they thought it was gross.<mask> I had to take out 8 garbages first thing in the morning,<mask> I was a man.<mask>, the parking lot was overflowing with trash that had spilled out of the over-packed garbage cans,<mask> none of the girls had done the usual nightly responsibility of sweeping,<mask> they thought it was a man's job<mask> well. I'm not talking about one or two co-workers here, there were 7 women on shift, and every single one of them followed this gender stereotyped idea of<mask> closing responsibilities they should have. Its one thing<mask> men and women divide tasks based on relative strengths and weaknesses,<mask><mask> someone neglects to do a task<mask> there's no one of the opposite sex to assign it to, that is ridiculous.<mask> only men are closing, someone still ends up washing the dishes and shining the countertops. [NEWLINE] [NEWLINE] [NEWLINE] Due to the biological, sociological, and psychological factors that women face, they are on average less suited to performing excessive labour, and in an economic system<mask> more work = more money, it is no surprise that men make more than women. [NEWLINE] [NEWLINE] <mask> I find most surprising is that<mask> I bring up this issue, I am met with incredible opposition,<mask><mask> I am trying to
Label encoding: <s>This is naturally an incredibly controversial topic and I highly doubt that many people will agree with me, but I feel compelled to discuss this. [NEWLINE] [NEWLINE] This needs to be prefaced by saying that all of this is based on averages, not individuals. I'm sure many women are more capable than many men, what I'm saying is that the general trend across the board is that on average a male worker is more productive. [NEWLINE] [NEWLINE] Men work almost triple the amount of overtime hours than women. [NEWLINE] [NEWLINE] Men work more standard hours than women. [NEWLINE] [NEWLINE] Women take more sick days off work than men. [NEWLINE] [NEWLINE] Women are more likely than men to take several years of maternity leave off of work. [NEWLINE] [NEWLINE] A small percentage of women undergo severe PMS and consequently work at a diminished capacity for several days of each month. [NEWLINE] [NEWLINE] To use anecdotal evidence, yesterday i came in to work, to find that all 8 garbages were completely filled. This was strange, since the closing crew is supposed to take out the garbages every night. I asked my coworker why no one had done the garbages last night, and I was effectively told that only girls were working the closing shift, and none of them wanted to touch the garbages because they thought it was gross. So I had to take out 8 garbages first thing in the morning, because I was a man. Additionally, the parking lot was overflowing with trash that had spilled out of the over-packed garbage cans, but none of the girls had done the usual nightly responsibility of sweeping, since they thought it was a man's job as well. I'm not talking about one or two co-workers here, there were 7 women on shift, and every single one of them followed this gender stereotyped idea of what closing responsibilities they should have. Its one thing if men and women divide tasks based on relative strengths and weaknesses, but if someone neglects to do a task because there's no one of the opposite sex to assign it to, that is ridiculous. When only men are closing, someone still ends up washing the dishes and shining the countertops. [NEWLINE] [NEWLINE] [NEWLINE] Due to the biological, sociological, and psychological factors that women face, they are on average less suited to performing excessive labour, and in an economic system where more work = more money, it is no surprise that men make more than women. [NEWLINE] [NEWLINE] What I find most surprising is that when I bring up this issue, I am met with incredible opposition, as if I am trying to
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think inequality is an inevitablity in an organized society like ours, at least to a certain extent.<mask> would you spend the time, effort, and money to go to medical school<mask> you could make the same salary working the checkout line at a grocery store? We need to provide adequate incentive for people to want to take up certain careers. Some equality is even a good thing,<mask> it gives people a reason to work hard to advance their careers, innovate in their field, and try to better society<mask> they can better their own lives. [NEWLINE] [NEWLINE] <mask>, a system of gross inequality isn't fair. Those with power and wealth do things to increase their power and wealth, that kind of hegemony is arguably inevitable, and at the very least it's a persistent feature of societies with widespread inequality. [NEWLINE] [NEWLINE] A system isn't fair<mask> the inequality is<mask> widespread that people on the bottom range of incomes aren't able to support themselves and the top range of incomes are earning more than they're worth. Is a CEO's job really<mask> much more important and difficult than a rank-and-file employee that they deserve to make a thousand times more money? Or is that simply those at the top seeking to preserve and enhance their own power on the backs of those who truly work hard? Surely they deserve to earn enough to attract quality candidates to the job,<mask> do they really deserve<mask> much that it puts the employees that keep the company moving under persistent financial pressure? [NEWLINE] [NEWLINE] [STARTQ] it's achieved through a fair system and legal means [ENDQ] [NEWLINE] I wouldn't call it fair, necessarily. Economic mobility isn't<mask> high<mask> we'd like to think, and the majority of people remain in the social class their parents were in. The kids of those at the top are more likely to get jobs at the top, meaning they've likely never experienced work at minimum wage and have no idea<mask> life is like without tons of money. There's a disconnect between the rich and the poor in part due to a lack of understanding, and that disconnect leads to an insensitivity towards the poor from the rich. They think they got<mask> they were purely due to hard work, and think that the poor are there purely<mask> of laziness, and neither of those things are true (at least most of the time, obviously there are exceptions). <mask> you're led to believe that you earned your money in a fair system, you're going to do everything you can within the confines of that system to make more money, not understanding that societal barriers often prevent
Label encoding: <s>I think inequality is an inevitablity in an organized society like ours, at least to a certain extent. Why would you spend the time, effort, and money to go to medical school if you could make the same salary working the checkout line at a grocery store? We need to provide adequate incentive for people to want to take up certain careers. Some equality is even a good thing, as it gives people a reason to work hard to advance their careers, innovate in their field, and try to better society so they can better their own lives. [NEWLINE] [NEWLINE] However, a system of gross inequality isn't fair. Those with power and wealth do things to increase their power and wealth, that kind of hegemony is arguably inevitable, and at the very least it's a persistent feature of societies with widespread inequality. [NEWLINE] [NEWLINE] A system isn't fair if the inequality is so widespread that people on the bottom range of incomes aren't able to support themselves and the top range of incomes are earning more than they're worth. Is a CEO's job really so much more important and difficult than a rank-and-file employee that they deserve to make a thousand times more money? Or is that simply those at the top seeking to preserve and enhance their own power on the backs of those who truly work hard? Surely they deserve to earn enough to attract quality candidates to the job, but do they really deserve so much that it puts the employees that keep the company moving under persistent financial pressure? [NEWLINE] [NEWLINE] [STARTQ] it's achieved through a fair system and legal means [ENDQ] [NEWLINE] I wouldn't call it fair, necessarily. Economic mobility isn't as high as we'd like to think, and the majority of people remain in the social class their parents were in. The kids of those at the top are more likely to get jobs at the top, meaning they've likely never experienced work at minimum wage and have no idea what life is like without tons of money. There's a disconnect between the rich and the poor in part due to a lack of understanding, and that disconnect leads to an insensitivity towards the poor from the rich. They think they got where they were purely due to hard work, and think that the poor are there purely because of laziness, and neither of those things are true (at least most of the time, obviously there are exceptions).  If you're led to believe that you earned your money in a fair system, you're going to do everything you can within the confines of that system to make more money, not understanding that societal barriers often prevent
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with your premise,<mask> don't agree with your conclusion. [NEWLINE] [NEWLINE] On the premise, you're right: young people can't possibly relate to the adult themes and experiences, particularly of modern literature (ancient epic poetry, Norse sagas and the like, are another story). In general, life experience helps you relate more fully and more powerfully to all art. In a very real sense, great works of art &amp; literature are literally wasted on the young. [NEWLINE] [NEWLINE] On the conclusion,<mask><mask> you're missing some of the *other* reasons for having young people read literature which have nothing to do with the themes or characters in the works. [NEWLINE] [NEWLINE] 1. Reading skills. General literacy is probably the single most important thing schoolkids go to school to acquire. They need to be able to recognize and interpret complex sentence structures, rhetorical devices, literary tropes like similes, metaphors, irony, sarcasm, literary allusion, novel uses of words, literary styles from different eras &amp; periods of language or dialects, etc. Kids are supposed to become *sophisticated* language users by the time they get out of school, not simply "read book. get information." automatons. They rely on school to give them the skills to communicate effectively later in life. There's no better way to learn sophisticated language than by puzzling through the most sophisticated examples we know of.<mask> whether or not kids are able to relate to the themes of great literature, they are hopefully picking up the reading skills they're going to need later in life<mask> they *can* appreciate those more exalted artistic ideas. [NEWLINE] [NEWLINE] 2. Critical thinking. Kids don't go to school to study things they already understand. The act of reading, picking up a dictionary, doing some research, scratching your head and trying to figure out exactly<mask> something means, is valuable. You wouldn't take kids who are ready for trigonometry and have them repeat basic algebra again. Similarly you don't want to give kids ready for a linguistic challenge copies of *Everybody Poops*. Trying to figure out language that you don't understand is a valuable<mask> painfully challenging exercise: it develops critical thinking skills. Very few people today can understand every line of Shakespeare,<mask> we've all had to struggle through it and try to wring some meaning out of it, and it's a valuable exercise in itself. Just seeing<mask> much our own language has changed in 400 years is an education in itself. [NEWLINE] [NEWLINE] 3. Writing skills. Reading, writing
Label encoding: <s>I agree with your premise, but don't agree with your conclusion. [NEWLINE] [NEWLINE] On the premise, you're right: young people can't possibly relate to the adult themes and experiences, particularly of modern literature (ancient epic poetry, Norse sagas and the like, are another story). In general, life experience helps you relate more fully and more powerfully to all art. In a very real sense, great works of art &amp; literature are literally wasted on the young. [NEWLINE] [NEWLINE] On the conclusion, I think you're missing some of the *other* reasons for having young people read literature which have nothing to do with the themes or characters in the works. [NEWLINE] [NEWLINE] 1. Reading skills. General literacy is probably the single most important thing schoolkids go to school to acquire. They need to be able to recognize and interpret complex sentence structures, rhetorical devices, literary tropes like similes, metaphors, irony, sarcasm, literary allusion, novel uses of words, literary styles from different eras &amp; periods of language or dialects, etc. Kids are supposed to become *sophisticated* language users by the time they get out of school, not simply "read book. get information." automatons. They rely on school to give them the skills to communicate effectively later in life. There's no better way to learn sophisticated language than by puzzling through the most sophisticated examples we know of. So whether or not kids are able to relate to the themes of great literature, they are hopefully picking up the reading skills they're going to need later in life when they *can* appreciate those more exalted artistic ideas. [NEWLINE] [NEWLINE] 2. Critical thinking. Kids don't go to school to study things they already understand. The act of reading, picking up a dictionary, doing some research, scratching your head and trying to figure out exactly what something means, is valuable. You wouldn't take kids who are ready for trigonometry and have them repeat basic algebra again. Similarly you don't want to give kids ready for a linguistic challenge copies of *Everybody Poops*. Trying to figure out language that you don't understand is a valuable if painfully challenging exercise: it develops critical thinking skills. Very few people today can understand every line of Shakespeare, but we've all had to struggle through it and try to wring some meaning out of it, and it's a valuable exercise in itself. Just seeing how much our own language has changed in 400 years is an education in itself. [NEWLINE] [NEWLINE] 3. Writing skills. Reading, writing
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for the delta and the thoughtful reply.  To reply to you: [NEWLINE] [STARTQ] Democracies do not work on the principle of an "optimal" government,<mask> rather on government of the people,<mask> inefficiency is accepted<mask> long the people get to decide. [ENDQ] [NEWLINE] That's true in a true democracy,<mask> the US government was specifically founded NOT to be an absolute democracy<mask> of fear of mob rule (and for other, less favorable motivations, like classism, sexism, and racism,<mask> still).  We elect representatives who more or less are tasked with becoming experts in our national policies and making important decisions for us.  That has always existed in American government and has,<mask> anything, decreased in the past century.  We can't complain about the idea that government officials sometimes choose policies that are unpopular with the public without complaining about the very foundation of representative democracy.  And it's not fair to ignore instances<mask> unpopular government policies were arguably the right call, such<mask> Reconstruction policies which aimed to limit discrimination against black people in post-Civil-War society.  These policies often weren't popular, especially in the South,<mask> in retrospect we can see that they were more fair than allowing Southern popular votes to decide policy,<mask> blacks were a disenfranchised minority in the South and it was easy for the majority white voters to devise ways to discriminate against them. <mask> I would be wary of endorsing any policy that amounts to "let the people have<mask> they want,<mask><mask><mask> a majority agrees." [NEWLINE] [STARTQ] MKUltra happened. Should we trust an agency that does that, simply<mask> at times it does good? [ENDQ] [NEWLINE] Asking whether we should "trust" an organization like the CIA is not the same<mask> asking whether we should give it the power it has.  It might be fair that,<mask> an individual, you would mistrust agencies like the CIA and the NSA<mask> you know that they have records of violating the rights of individual citizens (rights that, it should be noted, are afforded you and protected by other parts of the government).  This isn't the same<mask> arguing that the CIA and NSA should no longer be allowed to exist, along with any other covert intelligence agencies we might not know about,<mask> that certainly wouldn't be in our nation's best interest.  Being the only powerful nation without intelligence and counterintelligence forces would put us at a huge disadvantage on the world scale, and leave us vulnerable.  It seems to me that we can't simply talk about "limiting government" all
Label encoding: <s>Thanks for the delta and the thoughtful reply.  To reply to you: [NEWLINE] [STARTQ] Democracies do not work on the principle of an "optimal" government, but rather on government of the people, so inefficiency is accepted as long the people get to decide. [ENDQ] [NEWLINE] That's true in a true democracy, but the US government was specifically founded NOT to be an absolute democracy because of fear of mob rule (and for other, less favorable motivations, like classism, sexism, and racism, but still).  We elect representatives who more or less are tasked with becoming experts in our national policies and making important decisions for us.  That has always existed in American government and has, if anything, decreased in the past century.  We can't complain about the idea that government officials sometimes choose policies that are unpopular with the public without complaining about the very foundation of representative democracy.  And it's not fair to ignore instances where unpopular government policies were arguably the right call, such as Reconstruction policies which aimed to limit discrimination against black people in post-Civil-War society.  These policies often weren't popular, especially in the South, but in retrospect we can see that they were more fair than allowing Southern popular votes to decide policy, since blacks were a disenfranchised minority in the South and it was easy for the majority white voters to devise ways to discriminate against them.  So I would be wary of endorsing any policy that amounts to "let the people have what they want, as long as a majority agrees." [NEWLINE] [STARTQ] MKUltra happened. Should we trust an agency that does that, simply because at times it does good? [ENDQ] [NEWLINE] Asking whether we should "trust" an organization like the CIA is not the same as asking whether we should give it the power it has.  It might be fair that, as an individual, you would mistrust agencies like the CIA and the NSA because you know that they have records of violating the rights of individual citizens (rights that, it should be noted, are afforded you and protected by other parts of the government).  This isn't the same as arguing that the CIA and NSA should no longer be allowed to exist, along with any other covert intelligence agencies we might not know about, because that certainly wouldn't be in our nation's best interest.  Being the only powerful nation without intelligence and counterintelligence forces would put us at a huge disadvantage on the world scale, and leave us vulnerable.  It seems to me that we can't simply talk about "limiting government" all
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So I've distilled<mask> you wrote down to three points: [NEWLINE] [NEWLINE] 1) **Respect costs nothing.** [NEWLINE] [NEWLINE] 2) **Experience deserves respect.** [NEWLINE] [NEWLINE] 3) **<mask> you've proved yourself, you shouldn't need to prove yourself again.** (This is more<mask> respect buys you than not needing to give respect,<mask> it's related). [NEWLINE] [NEWLINE] Addressing these points: [NEWLINE] [NEWLINE] 1) **<mask> respect costs you nothing,<mask> don't you respect everyone?** Assuming I'm a more junior, inexperienced person,<mask> doesn't the experienced person respect me equally?<mask> do people spend<mask> much time and money to earn credentials which are supposed to equate to respect in their fields? [NEWLINE] [NEWLINE] The answer to these points is that respect does cost me something. We don't have the luxury, with our limited time on this planet, to evaluate all corners of a decision.<mask> someone whom I respect tells me something, I rank it higher in my decision making process. That creates an opportunity cost corresponding with any other decisions I might have made.<mask> that more experienced person was wrong, I've suffered in some way. People will take advantage of being respected to further themselves. Look at politicians or religious figures that charge large sums of money for speaking privileges etc.<mask> I use my respect influences my limited resources including<mask> I spend my time; it certainly isn't free. [NEWLINE] [NEWLINE] 2)<mask> you accept my first point that respect does cost me something, then we have to assume it's a somewhat fixed resource. **<mask> experience == respect, then I'd have no choice<mask> to spend that respect.** Take two politicians who have been in office the same amount of time. Chances are, you respect one more than the other;<mask>? [NEWLINE] [NEWLINE] <mask> one of them was likely right more or made decisions which you agreed with more. That means that not all experience is the same. There is some experience which is of higher quality than other experience.<mask> let's say we work for the same company across teams and you've been around more than I have, you could say you've had more projects than I have,<mask> experience,<mask> I've been right more. You've kept your business around,<mask> it's limping more often than not. Is your experience<mask> good<mask> mine<mask> I've been right a higher percentage? Should I defer to you? Heck, even<mask> you were right<mask> often<mask> I was - I don't know your motives, the only ones I know are my own. The
Label encoding: <s>So I've distilled what you wrote down to three points: [NEWLINE] [NEWLINE] 1) **Respect costs nothing.** [NEWLINE] [NEWLINE] 2) **Experience deserves respect.** [NEWLINE] [NEWLINE] 3) ** If you've proved yourself, you shouldn't need to prove yourself again.** (This is more what respect buys you than not needing to give respect, but it's related). [NEWLINE] [NEWLINE] Addressing these points: [NEWLINE] [NEWLINE] 1) ** If respect costs you nothing, why don't you respect everyone?** Assuming I'm a more junior, inexperienced person, why doesn't the experienced person respect me equally? Why do people spend so much time and money to earn credentials which are supposed to equate to respect in their fields? [NEWLINE] [NEWLINE] The answer to these points is that respect does cost me something. We don't have the luxury, with our limited time on this planet, to evaluate all corners of a decision. When someone whom I respect tells me something, I rank it higher in my decision making process. That creates an opportunity cost corresponding with any other decisions I might have made. If that more experienced person was wrong, I've suffered in some way. People will take advantage of being respected to further themselves. Look at politicians or religious figures that charge large sums of money for speaking privileges etc. How I use my respect influences my limited resources including how I spend my time; it certainly isn't free. [NEWLINE] [NEWLINE] 2) If you accept my first point that respect does cost me something, then we have to assume it's a somewhat fixed resource. ** If experience == respect, then I'd have no choice how to spend that respect.** Take two politicians who have been in office the same amount of time. Chances are, you respect one more than the other; why? [NEWLINE] [NEWLINE] Because one of them was likely right more or made decisions which you agreed with more. That means that not all experience is the same. There is some experience which is of higher quality than other experience. So let's say we work for the same company across teams and you've been around more than I have, you could say you've had more projects than I have, thus experience, but I've been right more. You've kept your business around, but it's limping more often than not. Is your experience as good as mine if I've been right a higher percentage? Should I defer to you? Heck, even if you were right as often as I was - I don't know your motives, the only ones I know are my own. The
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>like I said, personal experience. In my area there's a great amount of racism that has led to people essentially being hounded out of homes. [NEWLINE] [NEWLINE] <mask><mask> the city gives a more varied and nebulous set of values. You learn to tolerate and even enjoy hearing different languages, different cultures around you. It gives you more of a wholesome value of humanity - no one's that different, we can all coexist together.<mask> yes, it doesn't give you the close neighborly love unless you move to the right neighborhood.<mask> then,<mask><mask> it<mask> instills a set of values that encourage people to seek out friends.<mask> you're not all living in a community together, it teaches actual active friendship<mask> you have to track them down, make plans and really desire that person in your life. Whereas country life gives you that close neighborly feel,<mask> only with the people who are literally your close neighbors. It doesn't teach you<mask> to find people who really are good people or friends, it doesn't teach you<mask> to have these social values without convenient proximity. Again, personal experience,<mask> you move out of the countryside you essentially disappear unless you go back and visit, no one is able to 'keep in touch', maybe<mask> you weren't actually good friends (just people living near each other)<mask> maybe<mask> the social values are only ingrained<mask> far<mask> proximity. [NEWLINE] [NEWLINE] I'd<mask> point out that there is a lot of caring in cities,<mask> it's perhaps harder to see<mask> the city isn't one community,<mask> you never get the whole community rallying around.<mask> you see constantly people feeding the homeless, volunteering, jumping onto train tracks to save people. [NEWLINE] [NEWLINE] Being anonymous isn't an indication of values, nor is keeping kids on leashes. It's simple practicality.<mask> are you not going to be anonymous<mask> there's more than 1 million people around you? Just on a practical level. Even<mask> everyone loved the same number of people<mask> they do in a countryside town, you'll still be anonymous to more than 99% of the population of a city<mask> it's actually impossible to not be unless you're famous. And leashes are just<mask> of crowds and dangerous busy streets - these are practical issues not moral or value issues. [NEWLINE] [NEWLINE] <mask> for the environment issue, I sort of already tackled that. Individuals in a city can't really do much<mask> there are<mask> many individuals - the actions of one is micro.<mask> there are plenty of people who do those micro things
Label encoding: <s>like I said, personal experience. In my area there's a great amount of racism that has led to people essentially being hounded out of homes. [NEWLINE] [NEWLINE] I think the city gives a more varied and nebulous set of values. You learn to tolerate and even enjoy hearing different languages, different cultures around you. It gives you more of a wholesome value of humanity - no one's that different, we can all coexist together. But yes, it doesn't give you the close neighborly love unless you move to the right neighborhood. But then, I think it also instills a set of values that encourage people to seek out friends. Since you're not all living in a community together, it teaches actual active friendship where you have to track them down, make plans and really desire that person in your life. Whereas country life gives you that close neighborly feel, but only with the people who are literally your close neighbors. It doesn't teach you how to find people who really are good people or friends, it doesn't teach you how to have these social values without convenient proximity. Again, personal experience, if you move out of the countryside you essentially disappear unless you go back and visit, no one is able to 'keep in touch', maybe because you weren't actually good friends (just people living near each other) but maybe because the social values are only ingrained so far as proximity. [NEWLINE] [NEWLINE] I'd also point out that there is a lot of caring in cities, but it's perhaps harder to see because the city isn't one community, so you never get the whole community rallying around. But you see constantly people feeding the homeless, volunteering, jumping onto train tracks to save people. [NEWLINE] [NEWLINE] Being anonymous isn't an indication of values, nor is keeping kids on leashes. It's simple practicality. How are you not going to be anonymous when there's more than 1 million people around you? Just on a practical level. Even if everyone loved the same number of people as they do in a countryside town, you'll still be anonymous to more than 99% of the population of a city because it's actually impossible to not be unless you're famous. And leashes are just because of crowds and dangerous busy streets - these are practical issues not moral or value issues. [NEWLINE] [NEWLINE] As for the environment issue, I sort of already tackled that. Individuals in a city can't really do much because there are so many individuals - the actions of one is micro. But there are plenty of people who do those micro things
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was in a similar boat and ended up becoming Catholic (after marrying a Catholic).  One thing that surprised me was<mask> little I knew about the most popular Christian church before I started researching it closely.  Here are my responses to your issues.  Granted I can't speak for the church,<mask> this is<mask> I see things: [NEWLINE] [NEWLINE] [STARTQ] Creationism and certain views that contradict plausible science that has been proven again and again. [ENDQ] [NEWLINE] The Catholic church has no problem with the theory of evolution or any of the other commonly accepted scientific theories<mask><mask><mask> I'm aware.  We do not hold to any strict creationist view of the world,<mask> there is no problem here at all. [NEWLINE] [NEWLINE] [STARTQ] The fact that there is such a harsh distinction between the two possible afterlives: endless punishment or endless paradise, and the fact that someone evil like Ted Bundy, "confessed" and apologized for his sins before execution, would be sent to Heaven and a virtuous, amazing human being like Bill Gates would be sent to Hell<mask> he is an atheist. [ENDQ] [NEWLINE] I'm not an expert on this,<mask> I believe the Catholic church does not actually believe in a strict heaven vs. hell.  There is at least heaven, purgatory, the resurrection, and hell.  Personally,<mask><mask><mask><mask> most people who are non-Christian or unbaptized<mask> they die go to hell,<mask> I'm not sure<mask> the official church statement on that is. [NEWLINE] [NEWLINE] [STARTQ] Anti-gay, anti-contraception and other repressive movements in the name of religion. [ENDQ] [NEWLINE] This is a bit of an issue admittedly in the church. <mask>, in our defense I would say that the Catholic church permits more disagreement within it than most churches.  Officially the church is against contraception, for example,<mask><mask><mask> you can still be OK with contraception and be admitted to the church.  No one asked me my view on that before accepting me in. [NEWLINE] [NEWLINE] [STARTQ] The fact that there are<mask> many religions, and within those, many different denominations and inconsistencies.<mask> can one possible say their belief is the absolute truth, held above the others? [ENDQ] [NEWLINE] Again, I like the Catholic church position here.  The Catholic church accepts a significant amount of disagreement in the knowledge that we have<mask> to find absolute truth on most topics.  Ideally<mask><mask> we should have one church which permits all reasonable opinions within it. [NEWLINE] [NEWLINE] [STARTQ] Various miracles and stories (like Noah's Ark) in the Bible are
Label encoding: <s>I was in a similar boat and ended up becoming Catholic (after marrying a Catholic).  One thing that surprised me was how little I knew about the most popular Christian church before I started researching it closely.  Here are my responses to your issues.  Granted I can't speak for the church, but this is how I see things: [NEWLINE] [NEWLINE] [STARTQ] Creationism and certain views that contradict plausible science that has been proven again and again. [ENDQ] [NEWLINE] The Catholic church has no problem with the theory of evolution or any of the other commonly accepted scientific theories as far as I'm aware.  We do not hold to any strict creationist view of the world, so there is no problem here at all. [NEWLINE] [NEWLINE] [STARTQ] The fact that there is such a harsh distinction between the two possible afterlives: endless punishment or endless paradise, and the fact that someone evil like Ted Bundy, "confessed" and apologized for his sins before execution, would be sent to Heaven and a virtuous, amazing human being like Bill Gates would be sent to Hell because he is an atheist. [ENDQ] [NEWLINE] I'm not an expert on this, but I believe the Catholic church does not actually believe in a strict heaven vs. hell.  There is at least heaven, purgatory, the resurrection, and hell.  Personally, I do not think most people who are non-Christian or unbaptized when they die go to hell, but I'm not sure what the official church statement on that is. [NEWLINE] [NEWLINE] [STARTQ] Anti-gay, anti-contraception and other repressive movements in the name of religion. [ENDQ] [NEWLINE] This is a bit of an issue admittedly in the church.  However, in our defense I would say that the Catholic church permits more disagreement within it than most churches.  Officially the church is against contraception, for example, but I think you can still be OK with contraception and be admitted to the church.  No one asked me my view on that before accepting me in. [NEWLINE] [NEWLINE] [STARTQ] The fact that there are so many religions, and within those, many different denominations and inconsistencies. How can one possible say their belief is the absolute truth, held above the others? [ENDQ] [NEWLINE] Again, I like the Catholic church position here.  The Catholic church accepts a significant amount of disagreement in the knowledge that we have yet to find absolute truth on most topics.  Ideally I think we should have one church which permits all reasonable opinions within it. [NEWLINE] [NEWLINE] [STARTQ] Various miracles and stories (like Noah's Ark) in the Bible are
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL] [NEWLINE] [NEWLINE] [STARTQ] The Debt Collection Improvement Act of 1996 (DCIA) authorizes the Secretary of the Treasury to collect past-due child support by the administrative offset of federal payments. Executive Order 13019-Supporting Families: Collecting Delinquent Child Support Obligations (September 1996), requires the Secretary of the Treasury to promptly develop and implement procedures necessary for the collection of past-due child support debts by administrative offset (the reduction or withholding of a payment). [ENDQ] [NEWLINE] The US government calls child support debt. [NEWLINE] [NEWLINE] [URL].pdf [NEWLINE] [NEWLINE] [STARTQ] <mask><mask>, the majority of people recognise the ti [ENDQ] me to pay child maintenance is<mask> it’s due, [NEWLINE] not years later<mask> the debt has escalated in [NEWLINE] to thousands of pounds. We should not have to [NEWLINE] use draconian remedies like forcing the sale of par [NEWLINE] ents’ homes in order to [NEWLINE] get them to face their [NEWLINE] responsibilities. [NEWLINE] [NEWLINE] The UK government calls child support debt, and notes my sort of scenario, that people can be forced to sell their home to pay child support. [NEWLINE] [NEWLINE] I'm getting rather frustrated- I don't enjoy semantics debates and you are arguing that the word debt can't be used for child support<mask> official government websites use the term debt for child support.<mask> you're from Canada or Australia I could<mask> easily provide a cite that their governments use the term debt for child support. An entirely pointless debate<mask> you are arguing for a non standard definition of debt is very frustrating for me. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> you are treating your view<mask> a completely radical new idea that has never been discussed before, with a completely new use of the word "debt" [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [STARTQ] another law, specifying indicia of fraud which create a prima facie case that a debtor transferred income or property to avoid payment to a child support creditor, which the Secretary finds affords comparable rights to child support creditors; and [ENDQ] [NEWLINE] Here, in the law<mask> they ban reduction of child support, they call it a debt.<mask> I noted before, I am getting rather frustrated at being accused of using the term debt in a non standard way<mask> a few seconds of googling could easily show that this is the official way to use the term. This definitely isn't going to change my view- I'm not going to change my view based off a false fact that I can disprove by googling it. [NEWLINE] [NEWLINE] [STARTQ] <mask> your objection to bankruptcy (<mask> opposed to "jubilee") is that you
Label encoding: <s> [URL] [NEWLINE] [NEWLINE] [STARTQ] The Debt Collection Improvement Act of 1996 (DCIA) authorizes the Secretary of the Treasury to collect past-due child support by the administrative offset of federal payments. Executive Order 13019-Supporting Families: Collecting Delinquent Child Support Obligations (September 1996), requires the Secretary of the Treasury to promptly develop and implement procedures necessary for the collection of past-due child support debts by administrative offset (the reduction or withholding of a payment). [ENDQ] [NEWLINE] The US government calls child support debt. [NEWLINE] [NEWLINE] [URL].pdf [NEWLINE] [NEWLINE] [STARTQ] In addition, the majority of people recognise the ti [ENDQ] me to pay child maintenance is when it’s due, [NEWLINE] not years later when the debt has escalated in [NEWLINE] to thousands of pounds. We should not have to [NEWLINE] use draconian remedies like forcing the sale of par [NEWLINE] ents’ homes in order to [NEWLINE] get them to face their [NEWLINE] responsibilities. [NEWLINE] [NEWLINE] The UK government calls child support debt, and notes my sort of scenario, that people can be forced to sell their home to pay child support. [NEWLINE] [NEWLINE] I'm getting rather frustrated- I don't enjoy semantics debates and you are arguing that the word debt can't be used for child support when official government websites use the term debt for child support. If you're from Canada or Australia I could also easily provide a cite that their governments use the term debt for child support. An entirely pointless debate where you are arguing for a non standard definition of debt is very frustrating for me. [NEWLINE] [NEWLINE] [STARTQ] But if you are treating your view as a completely radical new idea that has never been discussed before, with a completely new use of the word "debt" [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [STARTQ] another law, specifying indicia of fraud which create a prima facie case that a debtor transferred income or property to avoid payment to a child support creditor, which the Secretary finds affords comparable rights to child support creditors; and [ENDQ] [NEWLINE] Here, in the law where they ban reduction of child support, they call it a debt. As I noted before, I am getting rather frustrated at being accused of using the term debt in a non standard way when a few seconds of googling could easily show that this is the official way to use the term. This definitely isn't going to change my view- I'm not going to change my view based off a false fact that I can disprove by googling it. [NEWLINE] [NEWLINE] [STARTQ] If your objection to bankruptcy ( as opposed to "jubilee") is that you
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is a secondary discussion,<mask> we agree aggression is not fundamental. The challenge is defining aggression in a non-arbitrary way which supports ancap (or libertarianism)<mask> nothing more. [ENDQ] [NEWLINE] No, the definition of aggression need not exclude every other philosophy.<mask><mask> a perfectly reasonable definition of aggression is "to deprive someone of<mask> they are entitled to". Of course,<mask> your author pointed out, it isn't clear<mask> people are entitled to<mask> that is a separate argument, one that<mask><mask> libertarians are equipped to win. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Well, I guess that's<mask> just about everyone with different political views disagree on :). [ENDQ] [NEWLINE] No,<mask><mask> that<mask> some people disagree with libertarians<mask> they have different theories of entitlement,<mask><mask> that the vast majority disagree with libertarianism<mask> they are for aggression in certain cases. Suppose its a time of war, and we can bomb an enemy munitions depot,<mask> doing<mask> would kill 5 innocent civilians,<mask><mask> most people recognize that those civilians have a right to life,<mask> the munitions depot were sufficiently important (like, it would allow the enemy to murder 1000 civilians), most people would think we should bomb it. Clearly this is an act of aggression against those civilians,<mask> they are entitled to their lives,<mask> I suspect that most people  would be for the bombing anyway. I'm against this sort of aggression, most people are not. Sure, some non-libertarians are against aggression too, and I'm happy to engage with them about theories of entitlement,<mask><mask><mask> libertarians talk a lot about the NAP,<mask> most people are pro-aggression, whatever their theory of entitlement. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It appears to me that we've started the whack-a-mole. You seem to be using a utilitarian argument. The problem with that is there are cases<mask> free markets don't work very well, and even an imperfect government can improve outcomes. A utilitarian framework will not justify an ancap world. [ENDQ] [NEWLINE] [NEWLINE] No, I fully recognize that utilitarianism doesn't justify ancapism, and that ancapism probably yields less than maximal utility. I understand<mask> you think that my argument seems consequentialist, and I apologize for not fleshing out the framework that I'm working off of. Let me expand here: [NEWLINE] [NEWLINE] <mask><mask> ethics should tell us<mask> to live our lives. This means that ethics can only tell me to do things that I am capable of actually doing. For example,<mask> someone showed me
Label encoding: <s> [STARTQ] This is a secondary discussion, since we agree aggression is not fundamental. The challenge is defining aggression in a non-arbitrary way which supports ancap (or libertarianism) but nothing more. [ENDQ] [NEWLINE] No, the definition of aggression need not exclude every other philosophy. I think a perfectly reasonable definition of aggression is "to deprive someone of what they are entitled to". Of course, as your author pointed out, it isn't clear what people are entitled to but that is a separate argument, one that I think libertarians are equipped to win. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Well, I guess that's what just about everyone with different political views disagree on :). [ENDQ] [NEWLINE] No, I think that while some people disagree with libertarians because they have different theories of entitlement, I think that the vast majority disagree with libertarianism because they are for aggression in certain cases. Suppose its a time of war, and we can bomb an enemy munitions depot, but doing so would kill 5 innocent civilians, even though most people recognize that those civilians have a right to life, if the munitions depot were sufficiently important (like, it would allow the enemy to murder 1000 civilians), most people would think we should bomb it. Clearly this is an act of aggression against those civilians, since they are entitled to their lives, but I suspect that most people  would be for the bombing anyway. I'm against this sort of aggression, most people are not. Sure, some non-libertarians are against aggression too, and I'm happy to engage with them about theories of entitlement, but I think libertarians talk a lot about the NAP, because most people are pro-aggression, whatever their theory of entitlement. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It appears to me that we've started the whack-a-mole. You seem to be using a utilitarian argument. The problem with that is there are cases where free markets don't work very well, and even an imperfect government can improve outcomes. A utilitarian framework will not justify an ancap world. [ENDQ] [NEWLINE] [NEWLINE] No, I fully recognize that utilitarianism doesn't justify ancapism, and that ancapism probably yields less than maximal utility. I understand why you think that my argument seems consequentialist, and I apologize for not fleshing out the framework that I'm working off of. Let me expand here: [NEWLINE] [NEWLINE] I think ethics should tell us how to live our lives. This means that ethics can only tell me to do things that I am capable of actually doing. For example, if someone showed me
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To the OP: [NEWLINE] [NEWLINE] See, posts like this are exactly<mask> most MRA's don't like Feminists. There's little attempt to understand<mask>'s motivating the people in the movement and instead the easy route of just slandering a large percentage of the movement<mask> misogynistic. [NEWLINE] [NEWLINE] The biggest difference I've personally seen behind the movements, and the one that speaks to<mask> MRA's are more vitriolic, is that generally speaking they have been actually slighted by the system in a way that favors women. [NEWLINE] [NEWLINE] Most feminists I know haven't actually suffered any major hardships<mask><mask><mask> of the system. Almost every one that I know developed their views in college, many<mask><mask><mask> of Women's Studies classes. Almost none has meet with any sort of real "oppression" in their lives. [NEWLINE] [NEWLINE] <mask>, almost every single MRA member I've talked to has one thing in common. Their lives have been DIRECTLY effected by some form of sexism. Whether it's losing a job, losing custody of their kids, being falsely accused of a crime, being jailed just for being a man, or losing a loved one. [NEWLINE] [NEWLINE] These men then have to deal with a lack of a support system. There are almost no men's shelters. Often even "co-ed" or family shelters won't take in men. In many places Domestic Abuse hotlines won't even talk to men, or worse just assume the man is the aggressor. Then<mask> many of those men turn to Feminism,<mask> that's the only "acceptable" social justice movement, they find themselves marginalized. Policies like men speaking last, men only allowed to be "allies", the complete and utter lack of campaigning on men's issues, men being told that they're the problem. Even society runs with this, think<mask> many PSA's you've seen about men, telling men not to rape, telling men not to hit women, telling men THEY'RE the problem. [NEWLINE] [NEWLINE] Even then,<mask> they do find themselves a community that accepts them in the MRA community, they STILL find themselves under attack. They can't go to a lecture/discussion without picket lines of feminists calling them names and screaming at them. They can't have a discussion online without being called a hate group or misogyinists or being told<mask> "MRA's aren't REALLY needed, men's issues are feminist's issues!". They can't fundraise (to, you know, open those shelter's that men desperately need
Label encoding: <s>To the OP: [NEWLINE] [NEWLINE] See, posts like this are exactly why most MRA's don't like Feminists. There's little attempt to understand what's motivating the people in the movement and instead the easy route of just slandering a large percentage of the movement as misogynistic. [NEWLINE] [NEWLINE] The biggest difference I've personally seen behind the movements, and the one that speaks to WHY MRA's are more vitriolic, is that generally speaking they have been actually slighted by the system in a way that favors women. [NEWLINE] [NEWLINE] Most feminists I know haven't actually suffered any major hardships as a result of the system. Almost every one that I know developed their views in college, many as a result of Women's Studies classes. Almost none has meet with any sort of real "oppression" in their lives. [NEWLINE] [NEWLINE] Meanwhile, almost every single MRA member I've talked to has one thing in common. Their lives have been DIRECTLY effected by some form of sexism. Whether it's losing a job, losing custody of their kids, being falsely accused of a crime, being jailed just for being a man, or losing a loved one. [NEWLINE] [NEWLINE] These men then have to deal with a lack of a support system. There are almost no men's shelters. Often even "co-ed" or family shelters won't take in men. In many places Domestic Abuse hotlines won't even talk to men, or worse just assume the man is the aggressor. Then when many of those men turn to Feminism, since that's the only "acceptable" social justice movement, they find themselves marginalized. Policies like men speaking last, men only allowed to be "allies", the complete and utter lack of campaigning on men's issues, men being told that they're the problem. Even society runs with this, think how many PSA's you've seen about men, telling men not to rape, telling men not to hit women, telling men THEY'RE the problem. [NEWLINE] [NEWLINE] Even then, if they do find themselves a community that accepts them in the MRA community, they STILL find themselves under attack. They can't go to a lecture/discussion without picket lines of feminists calling them names and screaming at them. They can't have a discussion online without being called a hate group or misogyinists or being told how "MRA's aren't REALLY needed, men's issues are feminist's issues!". They can't fundraise (to, you know, open those shelter's that men desperately need
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are many excellent arguments in this thread already,<mask> I will add my own thoughts.  The regret that we feel about abortion derives from a feeling of wasted potential.  We know perfectly well that the fetus has no knowledge, no consciousness, no experience, no accomplishments (other than growing, something a vegetable can<mask> do).  It does,<mask>, have the potential to be born and to grow up and do all the things that people do.  Every one of us was at one time a fetus, and we know that we too could have been aborted, had our mother felt disinclined to allow the pregnancy to reach completion. <mask>, there is always going to be wasted potential of this type.  A woman might be biologically capable of having 20 babies in the course of her lifetime (some women have done<mask> )<mask><mask> any woman fails to have that many babies (not counting those women who have medical problems that would prevent this number of pregnancies) then some potential human beings have been lost to the world.  In a sense, the failure to fertilize an ovum that has reached the uterus does exactly the same thing that abortion does, by depriving the world of a potential person. <mask> we recognize that a woman has the right to decide whether she does or does not want another baby.  We would not force someone to become pregnant,<mask> many people would force a woman to remain pregnant, rather than terminating an unwanted pregnancy.  The idea, in particular, that abortion should not be allowed even in cases of rape, is a terrible violation of a woman's right to control her own life.  Let us say that a woman is married to a man that she loves, a wonderful man, carefully chosen<mask> the correct mate for her, and she wants to have a child with that man and<mask> has stopped taking birth control pills. <mask> now she is fair game for rapists to impregnate at will, forcing her to have some other man's child instead (and even<mask> the rapist goes to jail or conceivably is executed, the woman<mask><mask> your morality must still bear his child).  This approach means in effect that women do not own their own bodies; other people are free to hijack a woman's reproductive system and use it for their own purposes.  That cannot be moral. [NEWLINE] <mask> others have noted, the consequences for the unborn fetus, of forcing the mother to carry the fetus to term, are not necessarily good.  Woman often seek abortions for very practical reasons. 
Label encoding: <s>There are many excellent arguments in this thread already, but I will add my own thoughts.  The regret that we feel about abortion derives from a feeling of wasted potential.  We know perfectly well that the fetus has no knowledge, no consciousness, no experience, no accomplishments (other than growing, something a vegetable can also do).  It does, however, have the potential to be born and to grow up and do all the things that people do.  Every one of us was at one time a fetus, and we know that we too could have been aborted, had our mother felt disinclined to allow the pregnancy to reach completion.  However, there is always going to be wasted potential of this type.  A woman might be biologically capable of having 20 babies in the course of her lifetime (some women have done so ) so if any woman fails to have that many babies (not counting those women who have medical problems that would prevent this number of pregnancies) then some potential human beings have been lost to the world.  In a sense, the failure to fertilize an ovum that has reached the uterus does exactly the same thing that abortion does, by depriving the world of a potential person.  Yet we recognize that a woman has the right to decide whether she does or does not want another baby.  We would not force someone to become pregnant, yet many people would force a woman to remain pregnant, rather than terminating an unwanted pregnancy.  The idea, in particular, that abortion should not be allowed even in cases of rape, is a terrible violation of a woman's right to control her own life.  Let us say that a woman is married to a man that she loves, a wonderful man, carefully chosen as the correct mate for her, and she wants to have a child with that man and therefore has stopped taking birth control pills.  But now she is fair game for rapists to impregnate at will, forcing her to have some other man's child instead (and even if the rapist goes to jail or conceivably is executed, the woman according to your morality must still bear his child).  This approach means in effect that women do not own their own bodies; other people are free to hijack a woman's reproductive system and use it for their own purposes.  That cannot be moral. [NEWLINE] As others have noted, the consequences for the unborn fetus, of forcing the mother to carry the fetus to term, are not necessarily good.  Woman often seek abortions for very practical reasons. 
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0911, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I heard Mike Huckabee say this during the first Republican Debate, and until I went online the next morning, I couldn't imagine anyone would disagree with him. People were saying this was shameful, and embarrassing to the US military, a ridiculous simplification. To be clear, I am no fan of Huckabee, and I have nothing against our military or militaries in general. That said,<mask><mask> he was dead on. [NEWLINE] [NEWLINE] <mask><mask> most people who don't accept this do<mask><mask> they think it's crass, and brutal. Well, yeah,<mask> it's<mask> absolutely true. That's<mask> the military has all those rifles and artillery and tanks and battleships and bombs and chemical weapons and knives and humvees and machine guns: for the killing of people and the breaking of things. Sure, not every member of the military's job is to be a killer,<mask> those people are there to support the killers. The cooks, mechanics, engineers, and secretaries are all there to let everyone else kill people and break things<mask> safely and efficiently<mask> possible. [NEWLINE] [NEWLINE] Again, I have absolutely no problem with this from a moral perspective. I am certainly not condemning anybody, just stating facts. Most people I've seen disagree with Huckabee are just dancing around this. "Soldiers exist to protect the United States and her interests!" Sure, using violence or the threat of violence. "The army doesn't just kill people, they developed the Internet!" Yeah,<mask> a weapon to coordinate their violence in the most efficient way possible. The internet we have now is just an unintended side effect. [NEWLINE] [NEWLINE] The US military is in a bit of a unique position,<mask> they haven't had a lot of opportunities to do their job recently. Our military is<mask> badass, there are not a lot of people with enough courage or stupidity to take us on.<mask>, a lot of time is spent running practice drills and handing out food to people after earthquakes. That's great,<mask> it's all a displacement activity until they need to do their real job. It's a sideshow, a distraction. The alternative would be these men sitting around, waiting for something to need destroying. We don't keep our military around and spend billions of dollars for disaster relief. We keep them around to kill people and break things. That's their real purpose, simple<mask> that. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.
Label encoding: <s>I heard Mike Huckabee say this during the first Republican Debate, and until I went online the next morning, I couldn't imagine anyone would disagree with him. People were saying this was shameful, and embarrassing to the US military, a ridiculous simplification. To be clear, I am no fan of Huckabee, and I have nothing against our military or militaries in general. That said, I think he was dead on. [NEWLINE] [NEWLINE] I think most people who don't accept this do so because they think it's crass, and brutal. Well, yeah, but it's also absolutely true. That's why the military has all those rifles and artillery and tanks and battleships and bombs and chemical weapons and knives and humvees and machine guns: for the killing of people and the breaking of things. Sure, not every member of the military's job is to be a killer, but those people are there to support the killers. The cooks, mechanics, engineers, and secretaries are all there to let everyone else kill people and break things as safely and efficiently as possible. [NEWLINE] [NEWLINE] Again, I have absolutely no problem with this from a moral perspective. I am certainly not condemning anybody, just stating facts. Most people I've seen disagree with Huckabee are just dancing around this. "Soldiers exist to protect the United States and her interests!" Sure, using violence or the threat of violence. "The army doesn't just kill people, they developed the Internet!" Yeah, as a weapon to coordinate their violence in the most efficient way possible. The internet we have now is just an unintended side effect. [NEWLINE] [NEWLINE] The US military is in a bit of a unique position, since they haven't had a lot of opportunities to do their job recently. Our military is so badass, there are not a lot of people with enough courage or stupidity to take us on. Thus, a lot of time is spent running practice drills and handing out food to people after earthquakes. That's great, but it's all a displacement activity until they need to do their real job. It's a sideshow, a distraction. The alternative would be these men sitting around, waiting for something to need destroying. We don't keep our military around and spend billions of dollars for disaster relief. We keep them around to kill people and break things. That's their real purpose, simple as that. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not interested in exhaustively sourcing and explaining my views in a reddit comment forum.  I will add a few more details,<mask>. [NEWLINE] [NEWLINE] [STARTQ] Ibn Khaldun predicted the rise of non-governmental corporations.<mask> exactly are you contesting here? [ENDQ] [NEWLINE] I have never heard of Ibn Khaldun,<mask> the quote you presented constitutes a weak prediction<mask> it does not explain itself in reductionist terms.  In other words, it doesn't explain *<mask> * Khaldun thought private enterprise would surpass kingdoms in size or scope.  He merely asserted that it would and gave no time-limit or environmental contingency on his prediction.  His prediction came 200 years before the emergence of the technology that would eventually topple those kingdoms (i.e., gunpowder weaponry),<mask> I cannot fathom on<mask> basis he formed his conjecture.  And regardless, even the emergence of gunpowder weaponry did not *guarantee* the usurpation of kings and their elite entourage,<mask> despotic nations can use their existing domination to monopolize coercive technology.  The Soviet Union is<mask> one example of many. [NEWLINE] [NEWLINE] [STARTQ] <mask> is the reductionist pyramid? [ENDQ] [NEWLINE] The reductionist pyramid is simply the observation that scientific knowledge can be broken into smaller and smaller pieces.  Physics gives rise to chemistry.  Chemistry gives rise to biology.  And biology, in theory, gives rise to "sociology"...<mask> the connection has not<mask> been made.  Sociologists, rather than working to connect their understanding to the rest of science, have worked independently on theories that do not rely on biology, and<mask> almost assuredly are doomed to failure. [NEWLINE] [NEWLINE] [STARTQ] Diamond's approach to history is not unique at all. [ENDQ] [NEWLINE] I did not claim that Jared's approach was unique,<mask> I maintain that it is not mainstream<mask><mask> the fact that Adam Smith apparently dabbled in it. [NEWLINE] [NEWLINE] [STARTQ] any Marxist worth her salt could tell you that the environment (aka "material conditions") impact human behavior/development... [ENDQ] [NEWLINE] That is an untestable hypothesis and<mask> outside the realm of science.  It's not enough to abstractly appreciate the role of the environment in human affairs.  Science requires the grunt-work of forming specific, testable hypotheses--and then testing them!  This is especially difficult for historians<mask> the evidence is often hard to come by,<mask> Jared Diamond proves that it's not impossible. [NEWLINE] [NEWLINE] [STARTQ] they are saying the same thing... [ENDQ] [NEWLINE] No they don't.
Label encoding: <s>I'm not interested in exhaustively sourcing and explaining my views in a reddit comment forum.  I will add a few more details, though. [NEWLINE] [NEWLINE] [STARTQ] Ibn Khaldun predicted the rise of non-governmental corporations. What exactly are you contesting here? [ENDQ] [NEWLINE] I have never heard of Ibn Khaldun, but the quote you presented constitutes a weak prediction because it does not explain itself in reductionist terms.  In other words, it doesn't explain * why * Khaldun thought private enterprise would surpass kingdoms in size or scope.  He merely asserted that it would and gave no time-limit or environmental contingency on his prediction.  His prediction came 200 years before the emergence of the technology that would eventually topple those kingdoms (i.e., gunpowder weaponry), so I cannot fathom on what basis he formed his conjecture.  And regardless, even the emergence of gunpowder weaponry did not *guarantee* the usurpation of kings and their elite entourage, since despotic nations can use their existing domination to monopolize coercive technology.  The Soviet Union is but one example of many. [NEWLINE] [NEWLINE] [STARTQ] what is the reductionist pyramid? [ENDQ] [NEWLINE] The reductionist pyramid is simply the observation that scientific knowledge can be broken into smaller and smaller pieces.  Physics gives rise to chemistry.  Chemistry gives rise to biology.  And biology, in theory, gives rise to "sociology"... but the connection has not yet been made.  Sociologists, rather than working to connect their understanding to the rest of science, have worked independently on theories that do not rely on biology, and so almost assuredly are doomed to failure. [NEWLINE] [NEWLINE] [STARTQ] Diamond's approach to history is not unique at all. [ENDQ] [NEWLINE] I did not claim that Jared's approach was unique, but I maintain that it is not mainstream regardless of the fact that Adam Smith apparently dabbled in it. [NEWLINE] [NEWLINE] [STARTQ] any Marxist worth her salt could tell you that the environment (aka "material conditions") impact human behavior/development... [ENDQ] [NEWLINE] That is an untestable hypothesis and thus outside the realm of science.  It's not enough to abstractly appreciate the role of the environment in human affairs.  Science requires the grunt-work of forming specific, testable hypotheses--and then testing them!  This is especially difficult for historians because the evidence is often hard to come by, but Jared Diamond proves that it's not impossible. [NEWLINE] [NEWLINE] [STARTQ] they are saying the same thing... [ENDQ] [NEWLINE] No they don't.
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The brain is a pattern maker and pattern matcher. And the patterns it makes, it finds in the outside world with ease. And the more deeply held and treasured a pattern (belief), the more it must find them out there in the world,<mask> that they can be matched are validated to be true. And the more invested it becomes, the more it sees<mask> it believes, even<mask> "out there" is the very poorest of matches. [NEWLINE] [NEWLINE] To challenge your own beliefs is to threaten your very identity! For it's our patterns/beliefs/values that make us who we are, and to change them for something else is an act of self-destruction without the guarantee of catharsis at the other end!<mask> kudos to you! [NEWLINE] [NEWLINE] My Dad was a mathematician by training and a hobby astrologer/astronomer, and he made thousands of charts. He sort a statistical proof of astrology for about a decade, collecting thousands of birth dates/times/locations for people in different professions. I remember there was a famous sportsmen set, and I remember he had a huge amount of data on volcanic eruptions and earthquakes. He eventually discounted the popular field<mask> completely rubbish, and was left with two statistical correlations (mars in some house for sportsmen and for earthquakes!) That was 20 years ago, and<mask> remains from my perspective, and conversations with him, is his spiritual sense that "existential causation" is "top down", not "bottom-up"... [NEWLINE] [NEWLINE] Ahh, Dad. [NEWLINE] [NEWLINE] Anyhow, you discount confirmation bias,<mask><mask> sure are you with the validity of your self-diagnosis? It's<mask> a small part of known [cognitive biases]( [URL] ) -<mask> many ways for the mind to trick itself to retain it's values/beliefs! [NEWLINE] [NEWLINE] You love science and thinking critically,<mask> no doubt you have a hunch that there is a contradiction you need to resolve. [NEWLINE] [NEWLINE] Science doesn't do<mask> well proving the non-existence of something,<mask> that something doesn't exist or is untestable  -<mask> there is no proof to find! There's nothing for science to point to and say "Oh, there it isn't!"<mask><mask> you make a very specific claim,<mask> your words have set meanings and definitions, then Science or just well applied logic can find proof of absence (or proof of impossibility) much easier. For example, for Science, you have to make a specific claim like
Label encoding: <s>The brain is a pattern maker and pattern matcher. And the patterns it makes, it finds in the outside world with ease. And the more deeply held and treasured a pattern (belief), the more it must find them out there in the world, so that they can be matched are validated to be true. And the more invested it becomes, the more it sees what it believes, even if "out there" is the very poorest of matches. [NEWLINE] [NEWLINE] To challenge your own beliefs is to threaten your very identity! For it's our patterns/beliefs/values that make us who we are, and to change them for something else is an act of self-destruction without the guarantee of catharsis at the other end! So kudos to you! [NEWLINE] [NEWLINE] My Dad was a mathematician by training and a hobby astrologer/astronomer, and he made thousands of charts. He sort a statistical proof of astrology for about a decade, collecting thousands of birth dates/times/locations for people in different professions. I remember there was a famous sportsmen set, and I remember he had a huge amount of data on volcanic eruptions and earthquakes. He eventually discounted the popular field as completely rubbish, and was left with two statistical correlations (mars in some house for sportsmen and for earthquakes!) That was 20 years ago, and what remains from my perspective, and conversations with him, is his spiritual sense that "existential causation" is "top down", not "bottom-up"... [NEWLINE] [NEWLINE] Ahh, Dad. [NEWLINE] [NEWLINE] Anyhow, you discount confirmation bias, but how sure are you with the validity of your self-diagnosis? It's but a small part of known [cognitive biases]( [URL] ) - so many ways for the mind to trick itself to retain it's values/beliefs! [NEWLINE] [NEWLINE] You love science and thinking critically, so no doubt you have a hunch that there is a contradiction you need to resolve. [NEWLINE] [NEWLINE] Science doesn't do so well proving the non-existence of something, if that something doesn't exist or is untestable  - because there is no proof to find! There's nothing for science to point to and say "Oh, there it isn't!" But if you make a very specific claim, where your words have set meanings and definitions, then Science or just well applied logic can find proof of absence (or proof of impossibility) much easier. For example, for Science, you have to make a specific claim like
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*IN CASE IT'S NOT OBVIOUS, PORN LINKS IN THIS POST ARE NSFW [NEWLINE] [NEWLINE] [STARTQ] Have you actually used Bing recently? I realize it's a joke to many people,<mask> that's mostly<mask> it sucked for<mask> long. Well, it doesn't suck anymore. [ENDQ] [NEWLINE] Admittedly, prior to posting this CMV I had not tried Bing<mask> very early on. <mask> tonight I went to the "BingItOn" site and took the "challenge"...turns out I still prefer google 5-0. <mask>,<mask><mask> I tried to deliberately use queries I hadn't googled recently, I still could be subconsciously picking results that feel in some intangible way more familiar to me.  Nonetheless, upon further research it looks like [a majority of internet users taking the blind "BingItOn" test]( [URL] %209.pdf) still preferred Google<mask> of earlier this year. [NEWLINE] [NEWLINE] [STARTQ] For one thing, Bing doesn't censor porn results - unlike Google. [ENDQ] [NEWLINE] [NEWLINE] One of the most surprising claims I've heard ITT is that google censors porn from its image search results.  It seems like<mask> they've actually done is refine their algorithm<mask> that [you only find porn<mask> you're actually searching for porn.]( [URL] ) [NEWLINE] [NEWLINE] This policy makes sense to me.  There's no reason to expect that the majority of users searching "<mask> to put a diaper on a baby" would want their top search results to include [cockshots and photos of girls fucking themselves with dildos.]( [URL] ) <mask> you *are* into diaper porn, google makes it easy to find.  All you need to do is enter a slightly more suggestive query like ["diaper girls."]( [URL] )   ("Diaper porn" results are even more explicit,<mask> honestly I got tired of screencapping disturbing diaper porn.  Search for yourself<mask> you're curious). [NEWLINE] [NEWLINE] For more conventional porn tastes, there is<mask> plenty of porn on Google.   "Sasha grey anal," a query that a porn-seeking user might actually type -- and a non-porn-seeking user would probably never type -- [does not disappoint]( [URL] ).   &lt;-- That screencap is not the type of thing you'd see from a search engine that actually censored porn. [NEWLINE] [NEWLINE] [STARTQ] For another thing, you can still use the "+" operator on Bing - unlike Google. [ENDQ] [NEWLINE] I don't see "+
Label encoding: <s>*IN CASE IT'S NOT OBVIOUS, PORN LINKS IN THIS POST ARE NSFW [NEWLINE] [NEWLINE] [STARTQ] Have you actually used Bing recently? I realize it's a joke to many people, but that's mostly because it sucked for so long. Well, it doesn't suck anymore. [ENDQ] [NEWLINE] Admittedly, prior to posting this CMV I had not tried Bing since very early on.  So tonight I went to the "BingItOn" site and took the "challenge"...turns out I still prefer google 5-0.  However, even though I tried to deliberately use queries I hadn't googled recently, I still could be subconsciously picking results that feel in some intangible way more familiar to me.  Nonetheless, upon further research it looks like [a majority of internet users taking the blind "BingItOn" test]( [URL] %209.pdf) still preferred Google as of earlier this year. [NEWLINE] [NEWLINE] [STARTQ] For one thing, Bing doesn't censor porn results - unlike Google. [ENDQ] [NEWLINE] [NEWLINE] One of the most surprising claims I've heard ITT is that google censors porn from its image search results.  It seems like what they've actually done is refine their algorithm so that [you only find porn if you're actually searching for porn.]( [URL] ) [NEWLINE] [NEWLINE] This policy makes sense to me.  There's no reason to expect that the majority of users searching " how to put a diaper on a baby" would want their top search results to include [cockshots and photos of girls fucking themselves with dildos.]( [URL] )  If you *are* into diaper porn, google makes it easy to find.  All you need to do is enter a slightly more suggestive query like ["diaper girls."]( [URL] )   ("Diaper porn" results are even more explicit, but honestly I got tired of screencapping disturbing diaper porn.  Search for yourself if you're curious). [NEWLINE] [NEWLINE] For more conventional porn tastes, there is also plenty of porn on Google.   "Sasha grey anal," a query that a porn-seeking user might actually type -- and a non-porn-seeking user would probably never type -- [does not disappoint]( [URL] ).   &lt;-- That screencap is not the type of thing you'd see from a search engine that actually censored porn. [NEWLINE] [NEWLINE] [STARTQ] For another thing, you can still use the "+" operator on Bing - unlike Google. [ENDQ] [NEWLINE] I don't see "+
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would like to give an example. Here in Sweden we have free education and medicine.<mask> medical care is free the state takes tax money and gives it to the medical providers. This means the salaries are the same for basically everyone in medical care. Same goes for the people in education. Same actually goes for most educated people in Sweden<mask> of the high taxes. The starting salary for software engineers in the US is almost the same<mask> the end of life salary for a software engineer in Sweden. A physician here earns less than $120k a year<mask> they are 65 and then you take away 40-50% off that for tax and everything you buy is expensive<mask> we have high VAT. [NEWLINE] [NEWLINE] The starting salary for lawyers in the US is a third of<mask> the top lawyer in Sweden earned last year. That is the most famous and successful lawyer. [NEWLINE] [NEWLINE] A teacher has almost<mask> long an education<mask> an engineer<mask> their end of life salary is the same<mask> the engineer's starting salary. Physicians in Sweden work hellish hours, many times with no pay<mask> they need to take care of more and more patients with less and less physicians<mask> many go to other sectors such<mask> consulting or emigrate<mask> of the shitty pay. We have physicians who barely speak the language<mask> there aren't enough physicians educated here. Nurses never have any significant salary developments and specialist nurses have to pay for the education themselves with a very minor raise afterwards. [NEWLINE] [NEWLINE] Now<mask> are the advantages of this system? We have free medical care. Free education.<mask> you need help with paying for your apartment you can get some money from the state. No job? Money. Handicapped and will never be able to work? Money. Sick? Money.<mask> you haven't saved enough money you get a guaranteed pension of atleast $1k roughly. Students get a pay check for attending school. You can get money for having children. Free school lunches. Free dental care until you're 20 and after that heavily subsidized. [NEWLINE] [NEWLINE] All of this comes at the cost that you're basically raking in 2-3x<mask> much money<mask> the bottom earners<mask> you're a top earner. A lot of people don't feel they're paid well enough for their knowledge. Now I'm both grateful and mad for this. My dad is a taxi driver and can basically work 8 to 5, Mondays through Fridays/sometimes Saturdays and make enough for us to live 20 minutes from the capital's center in one of the most sought after areas with some
Label encoding: <s>I would like to give an example. Here in Sweden we have free education and medicine. Since medical care is free the state takes tax money and gives it to the medical providers. This means the salaries are the same for basically everyone in medical care. Same goes for the people in education. Same actually goes for most educated people in Sweden because of the high taxes. The starting salary for software engineers in the US is almost the same as the end of life salary for a software engineer in Sweden. A physician here earns less than $120k a year when they are 65 and then you take away 40-50% off that for tax and everything you buy is expensive as we have high VAT. [NEWLINE] [NEWLINE] The starting salary for lawyers in the US is a third of what the top lawyer in Sweden earned last year. That is the most famous and successful lawyer. [NEWLINE] [NEWLINE] A teacher has almost as long an education as an engineer but their end of life salary is the same as the engineer's starting salary. Physicians in Sweden work hellish hours, many times with no pay because they need to take care of more and more patients with less and less physicians since many go to other sectors such as consulting or emigrate because of the shitty pay. We have physicians who barely speak the language because there aren't enough physicians educated here. Nurses never have any significant salary developments and specialist nurses have to pay for the education themselves with a very minor raise afterwards. [NEWLINE] [NEWLINE] Now what are the advantages of this system? We have free medical care. Free education. If you need help with paying for your apartment you can get some money from the state. No job? Money. Handicapped and will never be able to work? Money. Sick? Money. If you haven't saved enough money you get a guaranteed pension of atleast $1k roughly. Students get a pay check for attending school. You can get money for having children. Free school lunches. Free dental care until you're 20 and after that heavily subsidized. [NEWLINE] [NEWLINE] All of this comes at the cost that you're basically raking in 2-3x as much money as the bottom earners if you're a top earner. A lot of people don't feel they're paid well enough for their knowledge. Now I'm both grateful and mad for this. My dad is a taxi driver and can basically work 8 to 5, Mondays through Fridays/sometimes Saturdays and make enough for us to live 20 minutes from the capital's center in one of the most sought after areas with some
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just my point of view: [NEWLINE] <mask>, in economic terms, healthcare is NOT a public good. This means that it is neither non-excludable, (this means you CAN prevent people from having healthcare, whereas you cant effectively stop someone from say, enjoying a sunset.) nor non-rivalry( that means that providing an additional unit of the good will decrease its quality). [NEWLINE] [NEWLINE] You pointed out that healthcare is provided to everyone equally,<mask><mask> income level. [NEWLINE] This is true,<mask>, I do not see it being of equal standard. This is<mask> some people wait longer in queues than others.<mask>, depending on the queue, the standard of the healthcare provided changes.<mask> a person who has a stomacheache may recieve better healthcare (treated more swiftly) than a person with a tumor.  This system is inefficient. [NEWLINE] [NEWLINE] <mask><mask>, you mentioned that<mask> one does not know, better ask a doctor. This is the mentality of many,<mask> of the free healthcare. This worsens the queues, and<mask><mask><mask>, worsens the standard of healtcare for EVERYONE.<mask><mask>, those who end up with nothing wrong has effectively wasted the country's resources, by wasting the time of the doctors, ect (this is the opportunity cost, that is, instead of treating the idiot who has a nosebleed in the summer, the doctor uses his time to diagnose a flu patient or something.) This is pretty inefficient,<mask> it wastes the earths finite resources. [NEWLINE] [NEWLINE] <mask> on the topic of wastage, in free public healthcare, doctors are more likely to recommend the MOST EFFECTIVE TREATMENT,<mask> its free.<mask> your chdst hurts, BETTER GO FOR X-RAY SCANS, MRI SCANS JUST IN CASE. [NEWLINE] [NEWLINE] <mask> whats wrong with healthcare? Well it seems that in most countries, due to the personal interests of the consumer and the producers, there is an underproduction of healthcare. [NEWLINE] <mask>? Well its<mask><mask> you cure a guy, there are POSITIVE EXTERNALITIES, this means that additional benefit is granted to a third party. (<mask> you're intrested I  this portion please message me, ill elaborate to the best of my abilities!) [NEWLINE] [NEWLINE] For instance, by curing a guy(quickly), you make sure that they do not spread the disease to others!<mask><mask>, they can return to work, contributing to the society instead of resting at home (opportunity cost) [NEWLINE] <mask>
Label encoding: <s>Just my point of view: [NEWLINE] Firstly, in economic terms, healthcare is NOT a public good. This means that it is neither non-excludable, (this means you CAN prevent people from having healthcare, whereas you cant effectively stop someone from say, enjoying a sunset.) nor non-rivalry( that means that providing an additional unit of the good will decrease its quality). [NEWLINE] [NEWLINE] You pointed out that healthcare is provided to everyone equally, regardless of income level. [NEWLINE] This is true, however, I do not see it being of equal standard. This is because some people wait longer in queues than others. Thus, depending on the queue, the standard of the healthcare provided changes. So a person who has a stomacheache may recieve better healthcare (treated more swiftly) than a person with a tumor.  This system is inefficient. [NEWLINE] [NEWLINE] In addition, you mentioned that if one does not know, better ask a doctor. This is the mentality of many, because of the free healthcare. This worsens the queues, and as a result, worsens the standard of healtcare for EVERYONE. In addition, those who end up with nothing wrong has effectively wasted the country's resources, by wasting the time of the doctors, ect (this is the opportunity cost, that is, instead of treating the idiot who has a nosebleed in the summer, the doctor uses his time to diagnose a flu patient or something.) This is pretty inefficient, as it wastes the earths finite resources. [NEWLINE] [NEWLINE] While on the topic of wastage, in free public healthcare, doctors are more likely to recommend the MOST EFFECTIVE TREATMENT, since its free. So your chdst hurts, BETTER GO FOR X-RAY SCANS, MRI SCANS JUST IN CASE. [NEWLINE] [NEWLINE] So whats wrong with healthcare? Well it seems that in most countries, due to the personal interests of the consumer and the producers, there is an underproduction of healthcare. [NEWLINE] Why? Well its because when you cure a guy, there are POSITIVE EXTERNALITIES, this means that additional benefit is granted to a third party. ( If you're intrested I  this portion please message me, ill elaborate to the best of my abilities!) [NEWLINE] [NEWLINE] For instance, by curing a guy(quickly), you make sure that they do not spread the disease to others! In addition, they can return to work, contributing to the society instead of resting at home (opportunity cost) [NEWLINE] TH
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Personally I would<mask><mask> individuals within the regime abused their power to destroy knowledge. [NEWLINE] [NEWLINE] I don't think many religions or gods that people worship would actually support the destruction of knowledge,<mask> rules and laws have never completely stopped corruption and the few abusing the many. [NEWLINE] [NEWLINE] <mask> you've stated elsewhere, religion is often a guise by which the rich, or the few, abuse the poor, or the many. Personally<mask><mask> the detrimental parts of religion would have happened without religion. There would simply be another guise the deed was put under. [NEWLINE] [NEWLINE] Some very messed up things have been done in the name of religion.<mask> I could say the same with science. Look at<mask> some people have done in the name of social Darwinism. Look at some of the experiments the Nazis performed on people. (Of course you can<mask><mask> those experiments and deeds were not real science, they were simply done in the name of science,<mask> I would argue in return that the gross deeds done by religious institutions were done in the name of religion, and not the'real religion') [NEWLINE] [NEWLINE] <mask> an example, the crusades were a pretty terrible things. Huge battles, lots of death, terrible deeds done by the crusaders. And yes, religion was a motivator. Perhaps the pope of the time really thought it was their God-given duty to take back Jerusalem. And many of the crusaders were certainly motivated by religion. They were told that God would absolve them of all their sins, and they would have certain success with God at their backs.<mask> religion was not close to being the only factor. You know<mask> the middle east had? Riches. Gold, tapestries, trade routes, architecture, sciences, mathematics, literature. The middle east was filthy rich with exactly<mask> the west wanted. And<mask> the kings and the pope used religion to motivate the people to go fight for Jerusalem.<mask> I don't doubt that the crusades could have happened without religion<mask> a motivator. Would it have been more difficult? Yeah. Religion is easily twisted, especially<mask> only the few have the ability to read the bible, and<mask> have the only information on the God all the people in Europe were essentially forced to worship, or claim to worship (semantics).<mask> the crusades still could have happened (<mask><mask><mask>, and honestly we'll never know for sure). [NEWLINE] [NEWLINE] The all being said, I could even make an argument that the crusades were beneficial for the world. The suffering of the relatively few in the
Label encoding: <s>Personally I would argue that individuals within the regime abused their power to destroy knowledge. [NEWLINE] [NEWLINE] I don't think many religions or gods that people worship would actually support the destruction of knowledge, but rules and laws have never completely stopped corruption and the few abusing the many. [NEWLINE] [NEWLINE] As you've stated elsewhere, religion is often a guise by which the rich, or the few, abuse the poor, or the many. Personally I think the detrimental parts of religion would have happened without religion. There would simply be another guise the deed was put under. [NEWLINE] [NEWLINE] Some very messed up things have been done in the name of religion. But I could say the same with science. Look at what some people have done in the name of social Darwinism. Look at some of the experiments the Nazis performed on people. (Of course you can argue that those experiments and deeds were not real science, they were simply done in the name of science, but I would argue in return that the gross deeds done by religious institutions were done in the name of religion, and not the'real religion') [NEWLINE] [NEWLINE] As an example, the crusades were a pretty terrible things. Huge battles, lots of death, terrible deeds done by the crusaders. And yes, religion was a motivator. Perhaps the pope of the time really thought it was their God-given duty to take back Jerusalem. And many of the crusaders were certainly motivated by religion. They were told that God would absolve them of all their sins, and they would have certain success with God at their backs. But religion was not close to being the only factor. You know what the middle east had? Riches. Gold, tapestries, trade routes, architecture, sciences, mathematics, literature. The middle east was filthy rich with exactly what the west wanted. And so the kings and the pope used religion to motivate the people to go fight for Jerusalem. But I don't doubt that the crusades could have happened without religion as a motivator. Would it have been more difficult? Yeah. Religion is easily twisted, especially when only the few have the ability to read the bible, and therefore have the only information on the God all the people in Europe were essentially forced to worship, or claim to worship (semantics). But the crusades still could have happened ( in my opinion, and honestly we'll never know for sure). [NEWLINE] [NEWLINE] The all being said, I could even make an argument that the crusades were beneficial for the world. The suffering of the relatively few in the
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>American here,<mask> I have dual views<mask> I did part of my schooling in the UK at the University of Luton (I know). [NEWLINE] [NEWLINE] I still think that the US is the land of opportunity. Goods are cheap in comparison to the other countries you mentioned. Australia and NZ are both very expensive places to live. I knew many Brits who used their holiday to travel to the US and came back with loads of retail such<mask> Levi's (this was 98-99) that were much cheaper than they could have bought in the UK [NEWLINE] [NEWLINE] You mentioned the poor. That's a conundrum<mask> it is in many parts of the world. There are welfare benefits such<mask> food stamps, medicaid (which gives lower income classes free medical care) and WIC (women infants and children) which allows for free medical care, prescriptions and childbirth.<mask> one of the advantages that I don't see the poor communities taking enough advantage of is the US Military. Aside from the Reddit cries of Hegemony! This is an opportunity to earn a living, learn a skill, or multiple ones, get a paid for education, and<mask> you make a career of it, you're out at 40 with a pension for the rest of your life and medical care. There are plenty of places in the military<mask> you're not going to wind up on the front lines of Afghanistan. [NEWLINE] [NEWLINE] Opportunity: from the undermotivated that seem to be the majority posters of Reddit, my experience is they are wrong. I grew up in a middle class household, got an education, was lucky enough to have parents help with college. I was never a straight A student,<mask> my personality and work ethic has allowed me to find a job that has moved me to the upper income brackets. I truly believe that anyone can do anything they want here. Check out the statistics, immigrants that come here legally tend to fare in the top earning echelons of American society. [NEWLINE] [NEWLINE] Medical:<mask> everything you see on Reddit, you can get medical treatment in the US.<mask> you show up at an ER, you will be treated, you will be charged,<mask> a lot of hospitals don't ever collect on those charges. I've already noted WIC and medicaid. It's not necessarily going to end your financial life to get sick. And<mask><mask>, your care is going to be world class. That's<mask> lots of people fly from all over the world to get treated in the US. That's capitalism, it costs more,<mask> it
Label encoding: <s>American here, though I have dual views as I did part of my schooling in the UK at the University of Luton (I know). [NEWLINE] [NEWLINE] I still think that the US is the land of opportunity. Goods are cheap in comparison to the other countries you mentioned. Australia and NZ are both very expensive places to live. I knew many Brits who used their holiday to travel to the US and came back with loads of retail such as Levi's (this was 98-99) that were much cheaper than they could have bought in the UK [NEWLINE] [NEWLINE] You mentioned the poor. That's a conundrum as it is in many parts of the world. There are welfare benefits such as food stamps, medicaid (which gives lower income classes free medical care) and WIC (women infants and children) which allows for free medical care, prescriptions and childbirth. But one of the advantages that I don't see the poor communities taking enough advantage of is the US Military. Aside from the Reddit cries of Hegemony! This is an opportunity to earn a living, learn a skill, or multiple ones, get a paid for education, and if you make a career of it, you're out at 40 with a pension for the rest of your life and medical care. There are plenty of places in the military where you're not going to wind up on the front lines of Afghanistan. [NEWLINE] [NEWLINE] Opportunity: from the undermotivated that seem to be the majority posters of Reddit, my experience is they are wrong. I grew up in a middle class household, got an education, was lucky enough to have parents help with college. I was never a straight A student, but my personality and work ethic has allowed me to find a job that has moved me to the upper income brackets. I truly believe that anyone can do anything they want here. Check out the statistics, immigrants that come here legally tend to fare in the top earning echelons of American society. [NEWLINE] [NEWLINE] Medical: despite everything you see on Reddit, you can get medical treatment in the US. If you show up at an ER, you will be treated, you will be charged, but a lot of hospitals don't ever collect on those charges. I've already noted WIC and medicaid. It's not necessarily going to end your financial life to get sick. And in addition, your care is going to be world class. That's why lots of people fly from all over the world to get treated in the US. That's capitalism, it costs more, but it
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You really have a warped view of<mask> people actually think. [NEWLINE] [NEWLINE] [STARTQ] Nobody is afraid of women. [ENDQ] [NEWLINE] You're just<mask> likely to get mugged<mask> a guy is. 99% of rapes aren't a rapist hiding in a parking lot or in the bushes, they're by acquaintances, by someone you know, usually under the influence of alcohol. [NEWLINE] [NEWLINE] [STARTQ] <mask> I sleep around, I am a slut.<mask> a man sleeps around, he is a stud. [ENDQ] [NEWLINE] Most people don't think highly of either gender that sleeps around a lot. [NEWLINE] [NEWLINE] [STARTQ] No one gets very angry<mask> a man abandons the baby he fathered. [ENDQ] [NEWLINE] Yes, they do. That's complete nonsense<mask> you think that. [NEWLINE] [NEWLINE] [STARTQ] Men have freedom of opportunity. [ENDQ] [NEWLINE] It's actually a lot easier to enter engineering fields<mask> a women. With employers required to have 1:1 ratio of men to women in the workplace<mask> 1:20 of women to men actually major in STEM courses, they're at a much greater advantage to being selected for a job. I constantly listen to people complaining that a girl with a 2.0 gets an internship over them,<mask> they have a 4.0, solely<mask> of her gender. [NEWLINE] [NEWLINE] In class my nerd TA's are at the beckon call of any of the females in the class<mask> they give females more attention than males<mask> they have little to no female contact at all, which further puts males at a disadvantage. [NEWLINE] [NEWLINE] The problem is that women simply, and unfortunately, don't choose to be STEM majors. It's really a shame, too, we're missing out on about 50% of the great minds that could be in majoring in things that benefit mankind greater than the others. [NEWLINE] [NEWLINE] [STARTQ] Some people may say that<mask> I really wanted to, I could still be go into tech or engineering,<mask> there are extra difficulties involved that make one ask -- is it worth it? [ENDQ] [NEWLINE] Yes, and it'll be far easier for you than a white male. [NEWLINE] [NEWLINE] [STARTQ] <mask> I go anywhere looking fashionable and attractive, I will get harassed. [ENDQ] [NEWLINE] This does not happen.<mask> you're being insulted, it's not<mask> of<mask> you're wearing. [NEWLINE] [NEWLINE] [STARTQ] We still don't know<mask> there is a real bias against Asians at top colleges, and<mask> it will change prospects. I know that there was a problem, too many whites at colleges,<mask> that was due to blatant racism against minorities. There were minorities who outperformed
Label encoding: <s>You really have a warped view of what people actually think. [NEWLINE] [NEWLINE] [STARTQ] Nobody is afraid of women. [ENDQ] [NEWLINE] You're just as likely to get mugged as a guy is. 99% of rapes aren't a rapist hiding in a parking lot or in the bushes, they're by acquaintances, by someone you know, usually under the influence of alcohol. [NEWLINE] [NEWLINE] [STARTQ] If I sleep around, I am a slut. If a man sleeps around, he is a stud. [ENDQ] [NEWLINE] Most people don't think highly of either gender that sleeps around a lot. [NEWLINE] [NEWLINE] [STARTQ] No one gets very angry if a man abandons the baby he fathered. [ENDQ] [NEWLINE] Yes, they do. That's complete nonsense if you think that. [NEWLINE] [NEWLINE] [STARTQ] Men have freedom of opportunity. [ENDQ] [NEWLINE] It's actually a lot easier to enter engineering fields as a women. With employers required to have 1:1 ratio of men to women in the workplace while 1:20 of women to men actually major in STEM courses, they're at a much greater advantage to being selected for a job. I constantly listen to people complaining that a girl with a 2.0 gets an internship over them, while they have a 4.0, solely because of her gender. [NEWLINE] [NEWLINE] In class my nerd TA's are at the beckon call of any of the females in the class because they give females more attention than males because they have little to no female contact at all, which further puts males at a disadvantage. [NEWLINE] [NEWLINE] The problem is that women simply, and unfortunately, don't choose to be STEM majors. It's really a shame, too, we're missing out on about 50% of the great minds that could be in majoring in things that benefit mankind greater than the others. [NEWLINE] [NEWLINE] [STARTQ] Some people may say that if I really wanted to, I could still be go into tech or engineering, but there are extra difficulties involved that make one ask -- is it worth it? [ENDQ] [NEWLINE] Yes, and it'll be far easier for you than a white male. [NEWLINE] [NEWLINE] [STARTQ] If I go anywhere looking fashionable and attractive, I will get harassed. [ENDQ] [NEWLINE] This does not happen. If you're being insulted, it's not because of what you're wearing. [NEWLINE] [NEWLINE] [STARTQ] We still don't know if there is a real bias against Asians at top colleges, and if it will change prospects. I know that there was a problem, too many whites at colleges, but that was due to blatant racism against minorities. There were minorities who outperformed
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see redditors from the USA posting and commenting about<mask> they believe their country to be the best in the world. I talk to people in Australia who believe the same thing about this country. I believe these people are completely out of touch with the global community and are<mask><mask> damaging the global community by holding views that their country is better than others. I expect to see this behaviour<mask> analysing the ancient history of humans and nations,<mask> I expect better from modern society. CMV [NEWLINE] [NEWLINE] **EDIT**: It has been rightly pointed out that I have misused the word patriotic. I cannot pinpoint a word that encapsulates<mask> I mean<mask> I am referring to the seemingly extreme love for one's country of origin<mask> they believe that it is superior in all ways to all other countries and will make decisions and have discussions under that assumption no matter<mask>. [NEWLINE] [NEWLINE] **EDIT 2**: The word I'm looking for is Chauvinism ( [URL] ) and the damage I believe that can happen is better described by the word Jingoism ( [URL] ).<mask><mask> somebody commented the term Exceptionalism ( [URL] ) can be used somewhat to describe the attitude<mask> well. [NEWLINE] [NEWLINE] **EDIT 3**: I've been asked to clarify a few terms I have used. [NEWLINE] [NEWLINE] **Immaturity**: In retrospect this wasn't the most appropriate word to use<mask> I still stand by analysis that it can make the nation 'look' immature in the attitude it holds, in the same way that an immature person would not hold rational or mature views about the world<mask> of inexperience or otherwise. [NEWLINE] [NEWLINE] **Global Community**:<mask> many may not actually acknowledge the existence of a global community, I do see more an more evidence of said global community. I would use Reddit<mask> a prime example of this global community (<mask> it has a western majority)<mask> it shows that people from around the world can come and have discussions and often show that we aren't dissimilar in a lot of key ways. I want to<mask> plug my experiences in the open source programming scene<mask> I see collaboration between people of many different nationalities and backgrounds on common projects and goals. There are many other examples especially in regards to scientific groups (eg. CERN) and governmental summits. [NEWLINE] [NEWLINE] **Damaging**: An example of the damage I am referring to is like that which we are seeing at the moment between Australia and Indonesia. We had a very good and strong relationship between the two neighbouring countries until recently<mask> it became apparent that the Australian
Label encoding: <s>I see redditors from the USA posting and commenting about how they believe their country to be the best in the world. I talk to people in Australia who believe the same thing about this country. I believe these people are completely out of touch with the global community and are in fact damaging the global community by holding views that their country is better than others. I expect to see this behaviour when analysing the ancient history of humans and nations, but I expect better from modern society. CMV [NEWLINE] [NEWLINE] **EDIT**: It has been rightly pointed out that I have misused the word patriotic. I cannot pinpoint a word that encapsulates what I mean but I am referring to the seemingly extreme love for one's country of origin where they believe that it is superior in all ways to all other countries and will make decisions and have discussions under that assumption no matter what. [NEWLINE] [NEWLINE] **EDIT 2**: The word I'm looking for is Chauvinism ( [URL] ) and the damage I believe that can happen is better described by the word Jingoism ( [URL] ). Also as somebody commented the term Exceptionalism ( [URL] ) can be used somewhat to describe the attitude as well. [NEWLINE] [NEWLINE] **EDIT 3**: I've been asked to clarify a few terms I have used. [NEWLINE] [NEWLINE] **Immaturity**: In retrospect this wasn't the most appropriate word to use but I still stand by analysis that it can make the nation 'look' immature in the attitude it holds, in the same way that an immature person would not hold rational or mature views about the world because of inexperience or otherwise. [NEWLINE] [NEWLINE] **Global Community**: While many may not actually acknowledge the existence of a global community, I do see more an more evidence of said global community. I would use Reddit as a prime example of this global community ( though it has a western majority) as it shows that people from around the world can come and have discussions and often show that we aren't dissimilar in a lot of key ways. I want to also plug my experiences in the open source programming scene where I see collaboration between people of many different nationalities and backgrounds on common projects and goals. There are many other examples especially in regards to scientific groups (eg. CERN) and governmental summits. [NEWLINE] [NEWLINE] **Damaging**: An example of the damage I am referring to is like that which we are seeing at the moment between Australia and Indonesia. We had a very good and strong relationship between the two neighbouring countries until recently when it became apparent that the Australian
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Should we do it** [NEWLINE] [NEWLINE] This is a HUGE can of worms that would dominate the whole debate before most people even considered whether it was even feasible. [NEWLINE] [NEWLINE] and this alone asks even more questions. [NEWLINE] [NEWLINE] *Is this punishing success?* [NEWLINE] [NEWLINE] *whats to say it won't happen again a few years down the line?* [NEWLINE] [NEWLINE] *<mask> effect will occur in the job market<mask><mask><mask> * [NEWLINE] [NEWLINE] *Is it in the best interest of the general populace to receive a large amount of money the likes of which they are unaccustomed to?* [NEWLINE] [NEWLINE] Those questions would absolutely need to be answered<mask> even they do not address the moral dillemma, is it ok to forcibly remove money from one person and give it to somebody else. In all honestly thats exactly<mask> would be occuring. [NEWLINE] [NEWLINE] Before I even address the other parts of the question, I'm going to say that this idea goes against everything America stands for, not only that,<mask> its likely doomed from the start<mask> the people who are part of the 1% are not stupid, I wouldn't put it past any of them to leave the country rather than give up a HUGE sum of money for no real reason. [NEWLINE] [NEWLINE] **Basic Income** [NEWLINE] [NEWLINE] <mask> does this mean? does it mean we give everyone a monthly government check based on cost of living? does it mean a federal minimum wage that is adjusted<mask> that everyone with a job can live comfortably? [NEWLINE] [NEWLINE] Neither of those two have real moral issues,<mask> they both sound highly idealistic, and likely not feasible. That money from the government would need to come from somewhere. And not every business is going to be able to afford federal minimum wage requirements that are artificially inflated by places like California. [NEWLINE] [NEWLINE] <mask> you decided to let it go state by state you introduce a lot of red tape to the whole process, and the idea of basic income is lost (at least my interpretation of it)<mask> it would be nearly the same thing that we have now. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] **Outlawing renting** [NEWLINE] [NEWLINE] <mask>? there are some people, like myself that do not want to own a home, ever. [NEWLINE] [NEWLINE] [NEWLINE] In a moral frame,<mask> should we prohibit people from having a place to live that is paid in a manner that is suitable for both the original owner and the tenant,<mask><mask> of apartments, would we require people to own the individual rooms? [NEWLINE] [NEWLINE] [NEWLINE] In order for the amount of homeless to not skyrocket you would have to
Label encoding: <s>**Should we do it** [NEWLINE] [NEWLINE] This is a HUGE can of worms that would dominate the whole debate before most people even considered whether it was even feasible. [NEWLINE] [NEWLINE] and this alone asks even more questions. [NEWLINE] [NEWLINE] *Is this punishing success?* [NEWLINE] [NEWLINE] *whats to say it won't happen again a few years down the line?* [NEWLINE] [NEWLINE] * What effect will occur in the job market as a result * [NEWLINE] [NEWLINE] *Is it in the best interest of the general populace to receive a large amount of money the likes of which they are unaccustomed to?* [NEWLINE] [NEWLINE] Those questions would absolutely need to be answered but even they do not address the moral dillemma, is it ok to forcibly remove money from one person and give it to somebody else. In all honestly thats exactly what would be occuring. [NEWLINE] [NEWLINE] Before I even address the other parts of the question, I'm going to say that this idea goes against everything America stands for, not only that, but its likely doomed from the start because the people who are part of the 1% are not stupid, I wouldn't put it past any of them to leave the country rather than give up a HUGE sum of money for no real reason. [NEWLINE] [NEWLINE] **Basic Income** [NEWLINE] [NEWLINE] What does this mean? does it mean we give everyone a monthly government check based on cost of living? does it mean a federal minimum wage that is adjusted so that everyone with a job can live comfortably? [NEWLINE] [NEWLINE] Neither of those two have real moral issues, but they both sound highly idealistic, and likely not feasible. That money from the government would need to come from somewhere. And not every business is going to be able to afford federal minimum wage requirements that are artificially inflated by places like California. [NEWLINE] [NEWLINE] If you decided to let it go state by state you introduce a lot of red tape to the whole process, and the idea of basic income is lost (at least my interpretation of it) because it would be nearly the same thing that we have now. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] **Outlawing renting** [NEWLINE] [NEWLINE] Why? there are some people, like myself that do not want to own a home, ever. [NEWLINE] [NEWLINE] [NEWLINE] In a moral frame, why should we prohibit people from having a place to live that is paid in a manner that is suitable for both the original owner and the tenant, also what of apartments, would we require people to own the individual rooms? [NEWLINE] [NEWLINE] [NEWLINE] In order for the amount of homeless to not skyrocket you would have to
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "You're not wrong Walter. You're just an asshole." [ENDQ] [NEWLINE] Honestly, I would've been a little disappointing<mask> this *wasn't* one of the first comments. Congrats! [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] It seems like you think that the manager bending over backwards to accommodate you means that you're behavior is justified. [ENDQ] [NEWLINE] I believe my behavior was "justified"<mask> they had three opportunities to refuse service: [NEWLINE] [NEWLINE] *<mask> I asked<mask> they were open initially. [NEWLINE] *<mask> I told them there would be 12 of us. [NEWLINE] *<mask> I told them we'd go somewhere else<mask> the kitchen wasn't open. [NEWLINE] [NEWLINE] I didn't tell them "we'll just be quick" or that "we'd hurry", I even specifically told them that<mask> they couldn't accommodate us, we'd leave- they *chose* to seat us instead,<mask> I am neither "wrong", nor am I an "asshole" for expecting them to *actually provide* the service for which I was paying. [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] <mask> the real reason is the toxic "The customer is always right" mantra of capitalism. [ENDQ] [NEWLINE] I'll take the Pepsi challenge with Capitalism against any other system any old day of the week. It's not perfect,<mask> it's far better than any alternative. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] The manager was accommodating you<mask> she was terrified of losing her job<mask> you complained. [ENDQ] [NEWLINE] <mask> that's the type of environment that she chooses to work in, then that's between her and her employer; it doesn't make it my problem<mask>. Again, the restaurant had several chances to say that they couldn't, or were unwilling to, accommodate us- they didn't. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] There was no reason you couldn't meet the individuals who were cooking and serving your food halfway and order something simple and leave in a reasonable amount of time. [ENDQ] [NEWLINE] Did they offer me a discount? Was my food going to be any cheaper?<mask> a business tells me that they can provide a service, and I pay for that service, then I expect to get<mask> I'm paying for. Period. Again, saying that they were closed would've been okay. Saying that they couldn't handle 12 people, would've been okay. Letting us leave<mask> we started to, would've been okay. They chose not to do any of those things. [NEWLINE] [NEWLINE]
Label encoding: <s> [STARTQ] "You're not wrong Walter. You're just an asshole." [ENDQ] [NEWLINE] Honestly, I would've been a little disappointing if this *wasn't* one of the first comments. Congrats! [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] It seems like you think that the manager bending over backwards to accommodate you means that you're behavior is justified. [ENDQ] [NEWLINE] I believe my behavior was "justified" because they had three opportunities to refuse service: [NEWLINE] [NEWLINE] * When I asked if they were open initially. [NEWLINE] * When I told them there would be 12 of us. [NEWLINE] * When I told them we'd go somewhere else if the kitchen wasn't open. [NEWLINE] [NEWLINE] I didn't tell them "we'll just be quick" or that "we'd hurry", I even specifically told them that if they couldn't accommodate us, we'd leave- they *chose* to seat us instead, so I am neither "wrong", nor am I an "asshole" for expecting them to *actually provide* the service for which I was paying. [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] But the real reason is the toxic "The customer is always right" mantra of capitalism. [ENDQ] [NEWLINE] I'll take the Pepsi challenge with Capitalism against any other system any old day of the week. It's not perfect, but it's far better than any alternative. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] The manager was accommodating you because she was terrified of losing her job if you complained. [ENDQ] [NEWLINE] If that's the type of environment that she chooses to work in, then that's between her and her employer; it doesn't make it my problem however. Again, the restaurant had several chances to say that they couldn't, or were unwilling to, accommodate us- they didn't. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] There was no reason you couldn't meet the individuals who were cooking and serving your food halfway and order something simple and leave in a reasonable amount of time. [ENDQ] [NEWLINE] Did they offer me a discount? Was my food going to be any cheaper? When a business tells me that they can provide a service, and I pay for that service, then I expect to get what I'm paying for. Period. Again, saying that they were closed would've been okay. Saying that they couldn't handle 12 people, would've been okay. Letting us leave when we started to, would've been okay. They chose not to do any of those things. [NEWLINE] [NEWLINE]
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've outright said "no" before to women I was interested in. Sometimes, it was in bed with them, completely naked. Somehow, I ended up having sex anyway. [NEWLINE] [NEWLINE] With a few girls I've seen over the course of my life, I've taken them back to either my or their apartment, got into bed, started feeling each other up, and then stopped them<mask> I didn't want to have sex. I've then gotten pressured by them to continue. [NEWLINE] [NEWLINE] I am clearly aroused--I just don't want to have sex for personal reasons. I like to get to know someone pretty well before I sleep with them, just to make sure there's not gonna be problems<mask> we don't work out at some point (that's happened in the past). [NEWLINE] [NEWLINE] <mask> some girls don't like hearing that. One girl started coercing me, saying, "Come on, I do yoga. Don't you want to see<mask> flexible I am?" and she started rubbing on me. I say I really shouldn't do this,<mask> I *am* getting turned on.<mask> eventually I just say fuck it (without explicit consent) and go for it. Another time, a girl just put the condom on me, and I was like, well, let's just get this over with. [NEWLINE] [NEWLINE] <mask>, here's<mask> I'm gonna get controversial with this. I know there is the "Don't blame the victim" mentality,<mask> few things are black and white to me. No undeniably means no,<mask> there are things I can do to not send mixed signals to a partner, which, objectively speaking I did. I've taken steps to stop sending those signals. [NEWLINE] [NEWLINE] <mask> I mean,<mask> I said no, that's by definition rape is it not? [NEWLINE] [NEWLINE] Here's another situation that's a bit sketchy:<mask> about<mask> I'm with my current girlfriend, whom I've had sex with many times. I've told her outright no before<mask> I'm not turned on,<mask> I care for her, and I care for her needs.<mask>, without explicit consent after saying no, I have sex with her just<mask> I care about satisfying her. Is that rape, too,<mask> I didn't want it? [NEWLINE] [NEWLINE] At the very least, I don't consider the situations I've described worth reporting to anyone. I mean, it's my body we're talking about here, aren't we--not the law's, who is sometimes less than trust worthy? I'm not
Label encoding: <s>I've outright said "no" before to women I was interested in. Sometimes, it was in bed with them, completely naked. Somehow, I ended up having sex anyway. [NEWLINE] [NEWLINE] With a few girls I've seen over the course of my life, I've taken them back to either my or their apartment, got into bed, started feeling each other up, and then stopped them because I didn't want to have sex. I've then gotten pressured by them to continue. [NEWLINE] [NEWLINE] I am clearly aroused--I just don't want to have sex for personal reasons. I like to get to know someone pretty well before I sleep with them, just to make sure there's not gonna be problems if we don't work out at some point (that's happened in the past). [NEWLINE] [NEWLINE] But some girls don't like hearing that. One girl started coercing me, saying, "Come on, I do yoga. Don't you want to see how flexible I am?" and she started rubbing on me. I say I really shouldn't do this, but I *am* getting turned on. So eventually I just say fuck it (without explicit consent) and go for it. Another time, a girl just put the condom on me, and I was like, well, let's just get this over with. [NEWLINE] [NEWLINE] So, here's where I'm gonna get controversial with this. I know there is the "Don't blame the victim" mentality, but few things are black and white to me. No undeniably means no, but there are things I can do to not send mixed signals to a partner, which, objectively speaking I did. I've taken steps to stop sending those signals. [NEWLINE] [NEWLINE] But I mean, if I said no, that's by definition rape is it not? [NEWLINE] [NEWLINE] Here's another situation that's a bit sketchy: how about when I'm with my current girlfriend, whom I've had sex with many times. I've told her outright no before when I'm not turned on, but I care for her, and I care for her needs. So, without explicit consent after saying no, I have sex with her just because I care about satisfying her. Is that rape, too, when I didn't want it? [NEWLINE] [NEWLINE] At the very least, I don't consider the situations I've described worth reporting to anyone. I mean, it's my body we're talking about here, aren't we--not the law's, who is sometimes less than trust worthy? I'm not
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Could you expand on this?<mask> I look it up, I might still be confused<mask> to<mask> you actually mean here. [ENDQ] [NEWLINE] [URL] #Corporate_personhood [NEWLINE] [NEWLINE] Basically, the government treats corporations<mask> people,<mask><mask> that's not logical (<mask> the government controls the court system,<mask> legally they're people) [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> a monopoly arises? (easily done by keeping wages high at first, then decreasing after monopoly is attained) New businesses cannot compete with a monopoly, especially not<mask> they offer higher wages to their employees. [ENDQ] [NEWLINE] I'd like to challenge your belief that a company can become a monopoly by cutting prices way below profit for an extended period of time and recover enough afterward to keep all other potential competitors out. Many people believe this,<mask> we shouldn't base our beliefs on<mask> others think. [NEWLINE] [NEWLINE] Let's say there are many competing companies and you are an executive in the largest one. You believe that you can form a monopoly,<mask> you gather the other executives for a meeting and address them<mask> follows: I know<mask> we can become a monopoly. We will cut our prices by 50% (or whatever) for 6 months to a year until smaller companies can't make ends meet. We will take a loss of $X million<mask> we will be operating at a loss,<mask> we will emerge<mask> the only company available, at which time we can raise prices higher than they were originally and make a killing. [NEWLINE] [NEWLINE] <mask> would they say? [NEWLINE] [NEWLINE] "Sounds like a bold plan, do you have any case studies of this working?" [NEWLINE] [NEWLINE] "Sounds very risky,<mask> it takes longer than you say we could go bankrupt too" [NEWLINE] [NEWLINE] And<mask> on [NEWLINE] [NEWLINE] <mask> let's say the company's management goes along with it and takes the risk. Give me your version of<mask> would happen next. [NEWLINE] [NEWLINE] [STARTQ] That threat of violence is present<mask> you profit of society and infrastructure. It is the only way to ensure the rules in exchange for those benefits are met. [ENDQ] [NEWLINE] Taxes don't only pay for infrastructure. And<mask> does profiting from society mean? [NEWLINE] [NEWLINE] [STARTQ] It is exactly the same for a company. You accept the rules that company provides in exchange for the benefits of using its means of production. [ENDQ] [NEWLINE] Except you get to choose the company and leave it whenever you want. [NEWLINE] [NEWLINE] [STARTQ] Breaking those rules will be met with violence, too. [ENDQ] [NEWLINE] No, misbehavior in a company is met with firing someone, not threatening them with violence.
Label encoding: <s> [STARTQ] Could you expand on this? If I look it up, I might still be confused as to what you actually mean here. [ENDQ] [NEWLINE] [URL] #Corporate_personhood [NEWLINE] [NEWLINE] Basically, the government treats corporations as people, even though that's not logical ( but the government controls the court system, so legally they're people) [NEWLINE] [NEWLINE] [STARTQ] What if a monopoly arises? (easily done by keeping wages high at first, then decreasing after monopoly is attained) New businesses cannot compete with a monopoly, especially not if they offer higher wages to their employees. [ENDQ] [NEWLINE] I'd like to challenge your belief that a company can become a monopoly by cutting prices way below profit for an extended period of time and recover enough afterward to keep all other potential competitors out. Many people believe this, but we shouldn't base our beliefs on what others think. [NEWLINE] [NEWLINE] Let's say there are many competing companies and you are an executive in the largest one. You believe that you can form a monopoly, so you gather the other executives for a meeting and address them as follows: I know how we can become a monopoly. We will cut our prices by 50% (or whatever) for 6 months to a year until smaller companies can't make ends meet. We will take a loss of $X million because we will be operating at a loss, but we will emerge as the only company available, at which time we can raise prices higher than they were originally and make a killing. [NEWLINE] [NEWLINE] What would they say? [NEWLINE] [NEWLINE] "Sounds like a bold plan, do you have any case studies of this working?" [NEWLINE] [NEWLINE] "Sounds very risky, if it takes longer than you say we could go bankrupt too" [NEWLINE] [NEWLINE] And so on [NEWLINE] [NEWLINE] But let's say the company's management goes along with it and takes the risk. Give me your version of what would happen next. [NEWLINE] [NEWLINE] [STARTQ] That threat of violence is present because you profit of society and infrastructure. It is the only way to ensure the rules in exchange for those benefits are met. [ENDQ] [NEWLINE] Taxes don't only pay for infrastructure. And what does profiting from society mean? [NEWLINE] [NEWLINE] [STARTQ] It is exactly the same for a company. You accept the rules that company provides in exchange for the benefits of using its means of production. [ENDQ] [NEWLINE] Except you get to choose the company and leave it whenever you want. [NEWLINE] [NEWLINE] [STARTQ] Breaking those rules will be met with violence, too. [ENDQ] [NEWLINE] No, misbehavior in a company is met with firing someone, not threatening them with violence.
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are couple of problems with this: [NEWLINE] [NEWLINE] * No Deviant Candidates: The vast majority of people deviate from social and moral norms in various ways. This system would tend select for incredibly bland people (<mask> it were democratic) - any deviation from social norms (hates sports), or moral norms (has open relationship with wife) would be  disadvantages over more "normal" candidates. This would result in incredibly non-representative candidates - legitimately "normal" people are a tiny minority in our society, and selecting for people who didn't understand deviations from the norm might be a disaster. [NEWLINE] [NEWLINE] * Compromising Privacy of Family and Friends: The privacy of the candidate's family and friends would be compromised - communications have (at least) two sides. Even<mask> it is reasonable to ask the candidate to renounce her privacy, is it fair to allow the candidate make that decision for everyone she has ever communicated with? Imagine your brother runs for office - under this system, every e-mail that you have ever sent him becomes public without your consent. I don't think it is reasonable to subject the family and friends of political figures to that - particularly<mask> they would get no choice in the matter. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on Family: This is related to the norms point above,<mask> candidates with a difficult family life (e.g. a racist uncle, abusive parent, or sister with mental illness) would probably be at a huge disadvantage in elections - they would be put in a position<mask> they would likely be judged based the errant communication of others. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on History: Like above,<mask> from the perspective of a candidate who went through a difficult time in her life or who was a huge jerk<mask> she was young. This hypothetical candidate would be at an extreme disadvantage to someone who grew up in an easy situation and never had anything bad happen to them. [NEWLINE] [NEWLINE] * Renouncing Future Privacy: Positions of power are not for life - imagine someone wanted to serve in parliament/congress for a term. All deviations (e.g. furry, brony, etc.) and family history (e.g. racist uncle, huge fights between family members, etc.) would be exposed forever. Even<mask> they actually managed to get elected, this would be in the public record every time they applied for a job or went out on a date - for the rest of their life. [NEWLINE] [NEWLINE] * Electing Deceitful/Cautious People:
Label encoding: <s>There are couple of problems with this: [NEWLINE] [NEWLINE] * No Deviant Candidates: The vast majority of people deviate from social and moral norms in various ways. This system would tend select for incredibly bland people ( if it were democratic) - any deviation from social norms (hates sports), or moral norms (has open relationship with wife) would be  disadvantages over more "normal" candidates. This would result in incredibly non-representative candidates - legitimately "normal" people are a tiny minority in our society, and selecting for people who didn't understand deviations from the norm might be a disaster. [NEWLINE] [NEWLINE] * Compromising Privacy of Family and Friends: The privacy of the candidate's family and friends would be compromised - communications have (at least) two sides. Even if it is reasonable to ask the candidate to renounce her privacy, is it fair to allow the candidate make that decision for everyone she has ever communicated with? Imagine your brother runs for office - under this system, every e-mail that you have ever sent him becomes public without your consent. I don't think it is reasonable to subject the family and friends of political figures to that - particularly when they would get no choice in the matter. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on Family: This is related to the norms point above, but candidates with a difficult family life (e.g. a racist uncle, abusive parent, or sister with mental illness) would probably be at a huge disadvantage in elections - they would be put in a position where they would likely be judged based the errant communication of others. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on History: Like above, but from the perspective of a candidate who went through a difficult time in her life or who was a huge jerk when she was young. This hypothetical candidate would be at an extreme disadvantage to someone who grew up in an easy situation and never had anything bad happen to them. [NEWLINE] [NEWLINE] * Renouncing Future Privacy: Positions of power are not for life - imagine someone wanted to serve in parliament/congress for a term. All deviations (e.g. furry, brony, etc.) and family history (e.g. racist uncle, huge fights between family members, etc.) would be exposed forever. Even if they actually managed to get elected, this would be in the public record every time they applied for a job or went out on a date - for the rest of their life. [NEWLINE] [NEWLINE] * Electing Deceitful/Cautious People:
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The plea for equal rights, is bullshit<mask> we already have equal rights, I can't marry a man. And gays can marry the opposite sex.<mask> our rights are quite equal. It's just I want to marry someone I can. [ENDQ] [NEWLINE] It doesn't matter to you<mask> you can't marry a man,<mask> you're sexuality doesn't preclude you to marrying a man.<mask> straight marriage was outlawed and you could only marry someone of your own sex, would you agree to a gay person saying "you've got equality,<mask> I can't marry someone of the opposite sex either"? Of course not. [NEWLINE] [NEWLINE] [STARTQ] Which brings me to the reason<mask> marriage exists: it's the societies tool to support its own reproduction. [ENDQ] [NEWLINE] Ignoring the fact that the purpose of marriage has changed hugely in the past centuries from a forced political union to the current set-up, you seem to be under the impression that gay couples are somewhat infertile. Just like infertile heterosexual couples, there are many options available to same-sex couples which allow them to conceive.<mask> you bar them from marriage<mask> normal intercourse between those two individuals won't result in pregnancy, then by that logic you must outlaw marriage to infertile couples. Two fully fertile same-sex spouses are far more likely to produce children, with help, than a heterosexual couple<mask> one or both are infertile. [NEWLINE] [NEWLINE] [STARTQ] Now,<mask> same sex couples can't have children in any natural way, and most of them don't want to [ENDQ] [NEWLINE] <mask> are you possibly basing that on? You don't think gay people love children<mask> much<mask> straight people do? You don't think they want to nurture a child<mask> much<mask> a straight person does? My wife's cousin (a lesbian) has two children of her own whom she loves and adores with every bit of affection<mask> a straight woman. [NEWLINE] [NEWLINE] [STARTQ] here comes in the fact that we don't know<mask> problems that might cause to the child,<mask> I'll leave it [ENDQ] [NEWLINE] We absolutely do know the answer to this: Children raised in same-sex households are just<mask> healthy and well adjusted<mask> children in an equivalent heterosexual household. [Here is a study that backs that up]( [URL] %20of%20Same-Sex%20Couples%20Position%20Statement%20-%20October%202006%20(1).pdf). [NEWLINE] [NEWLINE] [And another]( [URL].pdf) [NEWLINE] [NEWLINE] [And another]( [URL].pdf) [NEWLINE] [NEWLINE] [
Label encoding: <s> [STARTQ] The plea for equal rights, is bullshit because we already have equal rights, I can't marry a man. And gays can marry the opposite sex. So our rights are quite equal. It's just I want to marry someone I can. [ENDQ] [NEWLINE] It doesn't matter to you if you can't marry a man, because you're sexuality doesn't preclude you to marrying a man. If straight marriage was outlawed and you could only marry someone of your own sex, would you agree to a gay person saying "you've got equality, because I can't marry someone of the opposite sex either"? Of course not. [NEWLINE] [NEWLINE] [STARTQ] Which brings me to the reason why marriage exists: it's the societies tool to support its own reproduction. [ENDQ] [NEWLINE] Ignoring the fact that the purpose of marriage has changed hugely in the past centuries from a forced political union to the current set-up, you seem to be under the impression that gay couples are somewhat infertile. Just like infertile heterosexual couples, there are many options available to same-sex couples which allow them to conceive. If you bar them from marriage because normal intercourse between those two individuals won't result in pregnancy, then by that logic you must outlaw marriage to infertile couples. Two fully fertile same-sex spouses are far more likely to produce children, with help, than a heterosexual couple where one or both are infertile. [NEWLINE] [NEWLINE] [STARTQ] Now, as same sex couples can't have children in any natural way, and most of them don't want to [ENDQ] [NEWLINE] What are you possibly basing that on? You don't think gay people love children as much as straight people do? You don't think they want to nurture a child as much as a straight person does? My wife's cousin (a lesbian) has two children of her own whom she loves and adores with every bit of affection as a straight woman. [NEWLINE] [NEWLINE] [STARTQ] here comes in the fact that we don't know what problems that might cause to the child, but I'll leave it [ENDQ] [NEWLINE] We absolutely do know the answer to this: Children raised in same-sex households are just as healthy and well adjusted as children in an equivalent heterosexual household. [Here is a study that backs that up]( [URL] %20of%20Same-Sex%20Couples%20Position%20Statement%20-%20October%202006%20(1).pdf). [NEWLINE] [NEWLINE] [And another]( [URL].pdf) [NEWLINE] [NEWLINE] [And another]( [URL].pdf) [NEWLINE] [NEWLINE] [
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not sure<mask> you are thinking that a shock collar is used,<mask> it may be far more humane than you think.<mask> I was a kid, we had many dogs chase cats off of our property. We had two dogs killed by cars. We don't think tethering a dog is humane (<mask><mask>, it's illegal now<mask> we live). We can't walk the dogs all day long, either. [NEWLINE] [NEWLINE] We never think of our dogs<mask> sacks of meat. They are beloved members of our family, and we want to keep them safe without keeping them on a chain or in a crate. [NEWLINE] [NEWLINE] <mask>,<mask> a supplement to walking dogs with a leash and occasionally visits to a dog park, we have used an "invisible fence" for our dogs for 23 years now. This is simply a wire loop buried a few inches all around our property. This wire loop creates an antenna that activates a collar<mask> it comes within a few feet. [NEWLINE] [NEWLINE] To train the dog, we first mark the fence line with flags. We train the dog to avoid the fence line using a traditional leash with no shock collar. We teach her that the fence line is a no-go area,<mask> that the rest of the property is her domain. [NEWLINE] [NEWLINE] Then we use the shock collar in combination with the leash.<mask> the dog approaches the fence, a warning tone sounds and we tug the dog back away from the fence line, reminding her that this is a no-go zone. She associates the flags, the tone, the tug, and our commands<mask> a reminder that she can't go there. [NEWLINE] [NEWLINE] The final step in training is to remove the traditional leash and use only the shock collar.<mask> the unleashed dog continues toward the fence, the familiar tone sounds and the dog is reminded to move back. [NEWLINE] [NEWLINE] <mask>,<mask>, some temptation is too great and the dog continues toward the fence line<mask> the warning tone, a brief shock is delivered. The dog then scampers back to us for comfort and reassurance. [NEWLINE] [NEWLINE] Once we're sure the dog understands the relationship of the property line, the tone, the shock, and the no-go status, we removed the flags gradually: first every third one, then a few more, then a few more until they are all gone. The lesson,<mask>, remains. [NEWLINE] [NEWLINE] Our current dog has received exactly ONE shock<mask> far in her entire life, and this was during the final stage of training. In exchange for this ONE shock, she has been completely
Label encoding: <s>Not sure how you are thinking that a shock collar is used, but it may be far more humane than you think. When I was a kid, we had many dogs chase cats off of our property. We had two dogs killed by cars. We don't think tethering a dog is humane ( in fact, it's illegal now where we live). We can't walk the dogs all day long, either. [NEWLINE] [NEWLINE] We never think of our dogs as sacks of meat. They are beloved members of our family, and we want to keep them safe without keeping them on a chain or in a crate. [NEWLINE] [NEWLINE] Therefore, as a supplement to walking dogs with a leash and occasionally visits to a dog park, we have used an "invisible fence" for our dogs for 23 years now. This is simply a wire loop buried a few inches all around our property. This wire loop creates an antenna that activates a collar when it comes within a few feet. [NEWLINE] [NEWLINE] To train the dog, we first mark the fence line with flags. We train the dog to avoid the fence line using a traditional leash with no shock collar. We teach her that the fence line is a no-go area, but that the rest of the property is her domain. [NEWLINE] [NEWLINE] Then we use the shock collar in combination with the leash. If the dog approaches the fence, a warning tone sounds and we tug the dog back away from the fence line, reminding her that this is a no-go zone. She associates the flags, the tone, the tug, and our commands as a reminder that she can't go there. [NEWLINE] [NEWLINE] The final step in training is to remove the traditional leash and use only the shock collar. If the unleashed dog continues toward the fence, the familiar tone sounds and the dog is reminded to move back. [NEWLINE] [NEWLINE] If, however, some temptation is too great and the dog continues toward the fence line despite the warning tone, a brief shock is delivered. The dog then scampers back to us for comfort and reassurance. [NEWLINE] [NEWLINE] Once we're sure the dog understands the relationship of the property line, the tone, the shock, and the no-go status, we removed the flags gradually: first every third one, then a few more, then a few more until they are all gone. The lesson, however, remains. [NEWLINE] [NEWLINE] Our current dog has received exactly ONE shock so far in her entire life, and this was during the final stage of training. In exchange for this ONE shock, she has been completely
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not only are you not sexist, you've convinced me that the policy is not *necessarily* sexist either. The policy extends to both men and women, in some form. [NEWLINE] [NEWLINE] <mask> for your other counterpoints, my idea of appropriate attire may not be palatable to you or most people, and it adds another layer of disagreement to this discussion.<mask> here it is... I actually think an ideal society *would* permit individuals to go out naked for the sake of comfort, in most environments. [NEWLINE] [NEWLINE] I don't expect you to be convinced<mask>,<mask> in the spirit of a CMV, I'll try to convince you now. First, please examine these pictures. [NEWLINE] [NEWLINE] [Exhibit A]( [URL].jpg) [NEWLINE] [NEWLINE] [Exhibit B]( [URL].jpg) [NEWLINE] [NEWLINE] [Exhibit C]( [URL].jpg) [NEWLINE] [NEWLINE] [Exhibit D]( [URL].jpg) [NEWLINE] [NEWLINE] Please do not read the rest of my response until the condensation that has gathered on your screen momentarily evaporates into steam (just kidding). In all seriousness, these pictures of Victoria Justice are not just provocative, they are essential to the issue of appropriate attire. [NEWLINE] [NEWLINE] <mask> developing a policy, in order to achieve fairness you have to imagine that everyone is "handicapped." In a classroom<mask> every female student is "handicapped" with Victoria's level of attractiveness,<mask> would you propose the dress code should be? Should she cover more skin? She is already almost completely covered up<mask> it is. Should she wear looser fitting clothing? [NEWLINE] [NEWLINE] [Exhibit E]( [URL].jpg) [NEWLINE] [NEWLINE] Unfortunately, this student would still be a distraction, even<mask> the dress were to cover all of her torso. The dress code demands something even looser. [NEWLINE] [NEWLINE] [Exhibit F]( [URL].jpg) [NEWLINE] [NEWLINE] Make no mistake, I would never accuse you of wanting women to wear burkas. You just want to help, and you have rational reasons for believing<mask> you do.<mask> the idea that women are only alluring<mask> they show skin is a paradox.<mask> that were true, nudists would be aroused most of the time (<mask> they most certainly are not). A woman is alluring<mask> she is alluring. Period. Rather than fixing a real problem, dress codes only perpetuate the Pavlovian connection between women dressing comfortably and a man's "right" to neglect his responsibility to operate with sexual restraint. [NEWLINE] [NEWLINE] We should decide whether we want to fix one of two problems. Problem
Label encoding: <s>Not only are you not sexist, you've convinced me that the policy is not *necessarily* sexist either. The policy extends to both men and women, in some form. [NEWLINE] [NEWLINE] As for your other counterpoints, my idea of appropriate attire may not be palatable to you or most people, and it adds another layer of disagreement to this discussion. But here it is... I actually think an ideal society *would* permit individuals to go out naked for the sake of comfort, in most environments. [NEWLINE] [NEWLINE] I don't expect you to be convinced though, so in the spirit of a CMV, I'll try to convince you now. First, please examine these pictures. [NEWLINE] [NEWLINE] [Exhibit A]( [URL].jpg) [NEWLINE] [NEWLINE] [Exhibit B]( [URL].jpg) [NEWLINE] [NEWLINE] [Exhibit C]( [URL].jpg) [NEWLINE] [NEWLINE] [Exhibit D]( [URL].jpg) [NEWLINE] [NEWLINE] Please do not read the rest of my response until the condensation that has gathered on your screen momentarily evaporates into steam (just kidding). In all seriousness, these pictures of Victoria Justice are not just provocative, they are essential to the issue of appropriate attire. [NEWLINE] [NEWLINE] When developing a policy, in order to achieve fairness you have to imagine that everyone is "handicapped." In a classroom where every female student is "handicapped" with Victoria's level of attractiveness, what would you propose the dress code should be? Should she cover more skin? She is already almost completely covered up as it is. Should she wear looser fitting clothing? [NEWLINE] [NEWLINE] [Exhibit E]( [URL].jpg) [NEWLINE] [NEWLINE] Unfortunately, this student would still be a distraction, even if the dress were to cover all of her torso. The dress code demands something even looser. [NEWLINE] [NEWLINE] [Exhibit F]( [URL].jpg) [NEWLINE] [NEWLINE] Make no mistake, I would never accuse you of wanting women to wear burkas. You just want to help, and you have rational reasons for believing what you do. But the idea that women are only alluring when they show skin is a paradox. If that were true, nudists would be aroused most of the time ( but they most certainly are not). A woman is alluring because she is alluring. Period. Rather than fixing a real problem, dress codes only perpetuate the Pavlovian connection between women dressing comfortably and a man's "right" to neglect his responsibility to operate with sexual restraint. [NEWLINE] [NEWLINE] We should decide whether we want to fix one of two problems. Problem
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ok, your wording had thrown me off. I don't think "intent" exists in any crime name except<mask> an addon, like assault with intent to kill. From here on I'll just talk conspiracy and attempted. [ENDQ] [NEWLINE] You may be outside the US.  In the US,<mask> you plot to kill someone and take reasonable steps on that journey to do<mask>, you've committed "intent to commit murder." [NEWLINE] [NEWLINE] <mask> you do it with other people you've committed "conspiracy to commit murder." [NEWLINE] [NEWLINE] <mask> you attempt it (<mask> fail) you've committed "attempted murder." [NEWLINE] [NEWLINE] And<mask> you do it, you've committed "murder." [NEWLINE] [NEWLINE] [STARTQ] I still have objections to your internal consistency on this point. Lets say I hire a killer, have a committed a crime now, or only<mask> the killer does the job? Now<mask><mask> I try to hire an undercover cop<mask> a killer? All he can do is tell me not to do it and inform the would be victim? [ENDQ] [NEWLINE] You've talked about killing someone and taken steps towards that goal,<mask> there is no victim<mask>. [NEWLINE] [NEWLINE] <mask> is this any different with current laws?  It's only a charge that would be added on later and<mask> there was an undercover cop, then you still have a way to stop the crime from happening. [NEWLINE] [NEWLINE] [STARTQ] <mask> does one test for this? Who designs the test? Who administers the test? [ENDQ] [NEWLINE] Same way any test is done, psychologists, courthouses. [NEWLINE] [NEWLINE] [STARTQ] <mask> long can they deny you this right? [ENDQ] [NEWLINE] Never.  You begin adulthood<mask> soon<mask> you are self sufficient to decide this is something you want to undertake. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] You state some 30 year olds should be protected.<mask> other age based rights can we deny? The ability to deny someone the right to consent to sex is at least a major obstacle to their right to reproduction. [ENDQ] [NEWLINE] Who said we're denying them the right?  We're just removing legal consent from the equation.  That 30 year old,<mask> they are incapable of taking and passing this test is no more able to consent than a 5 year old. [NEWLINE] [NEWLINE] <mask> we use age<mask> a metric of<mask> is okay. <mask> the actual ABILITY to consent is<mask> is important, NOT age. [NEWLINE] [NEWLINE] [STARTQ] <mask> about the right to vote? [ENDQ] [NEWLINE] Not sure I even really believe in voting.  I like the idea of a meritocracy\technocracy a lot more. 
Label encoding: <s> [STARTQ] Ok, your wording had thrown me off. I don't think "intent" exists in any crime name except as an addon, like assault with intent to kill. From here on I'll just talk conspiracy and attempted. [ENDQ] [NEWLINE] You may be outside the US.  In the US, if you plot to kill someone and take reasonable steps on that journey to do so, you've committed "intent to commit murder." [NEWLINE] [NEWLINE] If you do it with other people you've committed "conspiracy to commit murder." [NEWLINE] [NEWLINE] If you attempt it ( but fail) you've committed "attempted murder." [NEWLINE] [NEWLINE] And if you do it, you've committed "murder." [NEWLINE] [NEWLINE] [STARTQ] I still have objections to your internal consistency on this point. Lets say I hire a killer, have a committed a crime now, or only when the killer does the job? Now what if I try to hire an undercover cop as a killer? All he can do is tell me not to do it and inform the would be victim? [ENDQ] [NEWLINE] You've talked about killing someone and taken steps towards that goal, but there is no victim yet. [NEWLINE] [NEWLINE] How is this any different with current laws?  It's only a charge that would be added on later and if there was an undercover cop, then you still have a way to stop the crime from happening. [NEWLINE] [NEWLINE] [STARTQ] How does one test for this? Who designs the test? Who administers the test? [ENDQ] [NEWLINE] Same way any test is done, psychologists, courthouses. [NEWLINE] [NEWLINE] [STARTQ] How long can they deny you this right? [ENDQ] [NEWLINE] Never.  You begin adulthood as soon as you are self sufficient to decide this is something you want to undertake. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] You state some 30 year olds should be protected. What other age based rights can we deny? The ability to deny someone the right to consent to sex is at least a major obstacle to their right to reproduction. [ENDQ] [NEWLINE] Who said we're denying them the right?  We're just removing legal consent from the equation.  That 30 year old, if they are incapable of taking and passing this test is no more able to consent than a 5 year old. [NEWLINE] [NEWLINE] Yet we use age as a metric of what is okay.  When the actual ABILITY to consent is what is important, NOT age. [NEWLINE] [NEWLINE] [STARTQ] What about the right to vote? [ENDQ] [NEWLINE] Not sure I even really believe in voting.  I like the idea of a meritocracy\technocracy a lot more. 
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't disagree with this: [NEWLINE] [NEWLINE] [STARTQ] <mask> it were up to nature and there were no doctors, medicine, or anything like that, people like me would have likely died before we ever got the chance to bear children. [ENDQ] [NEWLINE] or this either: [NEWLINE] [NEWLINE] [STARTQ] <mask> people voluntarily chose to not reproduce like I am doing, there would be far less people with genetic disorders. [ENDQ] [NEWLINE] And, of course, I don't want to go around convincing anyone to have children<mask> I say it is okay. [NEWLINE] [NEWLINE] <mask> I don't agree that it is necessarily morally wrong, nor would I discourage it, and this is<mask> : [NEWLINE] [NEWLINE] We,<mask> human beings have more to offer than can be singled down to any one trait, or possibly even set of traits. [NEWLINE] [NEWLINE] Someone who is horribly unhealthy, and in great pain and misery<mask> of it, isn't necessarily someone that has nothing to contribute to society. Stephen Hawking and ALS, for example. *And any other "great" person that had any other genetic disorder for<mask> many examples you like* [NEWLINE] [NEWLINE] It isn't possible to know, (at least now, and quite likely ever), everything a person can possibly contribute based on their genes. [NEWLINE] [NEWLINE] In my own modest bad health experience, I have taken away that the pain and suffering I've experienced have<mask><mask> enhanced some of my positive characteristics, and help me contribute more to others.<mask><mask> similar things could be said for pain and suffering of many kinds. [NEWLINE] [NEWLINE] Now,<mask> I happen to pass on my condition, (which, admittedly, I would think differently of,<mask> it were known to be fully genetic and not a combination of environmental and genetic influences) I won't be happy about it. I'm sure I'll feel horrible<mask> I see them in pain the way I am. [NEWLINE] [NEWLINE] <mask> I believe suffering is a part of the human condition. I believe it is meant to be. I believe we suffer for reasons, and learn from suffering. And I firmly believe people suffer in one way or another.<mask> it isn't health, it's something else. [NEWLINE] [NEWLINE] And I believe that my husband and I have many positive traits, and that we offer quite a bit to society around us. And I believe that our child would be a good contribution for us to make, and have good things to offer, even<mask> they<mask> have health issues. [NEWLINE] [NEWLINE] Now, I could be wrong, we might have crap kids that do nothing at all for society, and keep more bad health genes
Label encoding: <s>I don't disagree with this: [NEWLINE] [NEWLINE] [STARTQ] If it were up to nature and there were no doctors, medicine, or anything like that, people like me would have likely died before we ever got the chance to bear children. [ENDQ] [NEWLINE] or this either: [NEWLINE] [NEWLINE] [STARTQ] If people voluntarily chose to not reproduce like I am doing, there would be far less people with genetic disorders. [ENDQ] [NEWLINE] And, of course, I don't want to go around convincing anyone to have children because I say it is okay. [NEWLINE] [NEWLINE] But I don't agree that it is necessarily morally wrong, nor would I discourage it, and this is why : [NEWLINE] [NEWLINE] We, as human beings have more to offer than can be singled down to any one trait, or possibly even set of traits. [NEWLINE] [NEWLINE] Someone who is horribly unhealthy, and in great pain and misery because of it, isn't necessarily someone that has nothing to contribute to society. Stephen Hawking and ALS, for example. *And any other "great" person that had any other genetic disorder for however many examples you like* [NEWLINE] [NEWLINE] It isn't possible to know, (at least now, and quite likely ever), everything a person can possibly contribute based on their genes. [NEWLINE] [NEWLINE] In my own modest bad health experience, I have taken away that the pain and suffering I've experienced have in fact enhanced some of my positive characteristics, and help me contribute more to others. I think similar things could be said for pain and suffering of many kinds. [NEWLINE] [NEWLINE] Now, if I happen to pass on my condition, (which, admittedly, I would think differently of, if it were known to be fully genetic and not a combination of environmental and genetic influences) I won't be happy about it. I'm sure I'll feel horrible when I see them in pain the way I am. [NEWLINE] [NEWLINE] But I believe suffering is a part of the human condition. I believe it is meant to be. I believe we suffer for reasons, and learn from suffering. And I firmly believe people suffer in one way or another. When it isn't health, it's something else. [NEWLINE] [NEWLINE] And I believe that my husband and I have many positive traits, and that we offer quite a bit to society around us. And I believe that our child would be a good contribution for us to make, and have good things to offer, even if they also have health issues. [NEWLINE] [NEWLINE] Now, I could be wrong, we might have crap kids that do nothing at all for society, and keep more bad health genes
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I guess I try to make new guy friends<mask> I hope they'll have the same interests<mask> I do and will be there to communicate and commiserate with. Among other reasons, I try to make<mask> many<mask> possible<mask> it becomes easy to find somebody to help or be helped by. It<mask> builds up my social standing<mask> I'm going to be perfectly honest. [NEWLINE] [NEWLINE] I don't know<mask> I'd be<mask> I were gay-- it's a good question,<mask> I don't know<mask> my perception of all my friends would be changed. [NEWLINE] [NEWLINE] At my all-boys school, I have friends who are gay with whom I interact all the time. I have one in particular who intrigues me<mask><mask> ostensibly having gay friends outside our school who send him suggestive pictures (and from<mask> I hear, they put just about every boy at my school to shame),<mask> drunk, he admits that he wants to be with some of the boys who go to my school<mask><mask> they are heterosexual. I can't help<mask> wonder<mask> just being friends with guys feels like it isn't enough. [NEWLINE] [NEWLINE] Thinking of my guy friends<mask> girls...<mask><mask> I would find some of their personalities cute<mask> not exactly my type. There are girls I know<mask> am not friends with nor really think of them romantically<mask> their personalities are off-putting. I suppose it doesn't help to be in the environment I'm in that I'm only exposed to a small facet of personality from the girls I interact with from the all-girls school.<mask> my best friend were a girl,<mask><mask> I'd rather not date that person primarily for fear of losing that person once we separated. Other friends would make very odd girls... Overall, I'm not sure<mask> I would date them-- it's a good hypothetical question<mask> a difficult one to contemplate. [NEWLINE] [NEWLINE] More importantly, I wonder<mask> we'd still be friends. The social system in my environment is very segregated between boys and girls.<mask> we were to simply do something together, it'd be considered some sort of date and not a friendly outing, which is probably weird to most people. The fringe friends would probably not be my friends. I imagine it'd be hard to relate with them<mask> they were girls<mask><mask><mask> I'd stay in touch with my closer friends. This is a tough thought-experiment,<mask> I'm sorry<mask> I'm repetitive or giving conflicting answers. [NEWLINE] [NEWLINE] <mask> I were to turn those girls at the all-girls school into guys, I
Label encoding: <s>I guess I try to make new guy friends because I hope they'll have the same interests as I do and will be there to communicate and commiserate with. Among other reasons, I try to make as many as possible because it becomes easy to find somebody to help or be helped by. It also builds up my social standing if I'm going to be perfectly honest. [NEWLINE] [NEWLINE] I don't know how I'd be if I were gay-- it's a good question, but I don't know how my perception of all my friends would be changed. [NEWLINE] [NEWLINE] At my all-boys school, I have friends who are gay with whom I interact all the time. I have one in particular who intrigues me because despite ostensibly having gay friends outside our school who send him suggestive pictures (and from what I hear, they put just about every boy at my school to shame), when drunk, he admits that he wants to be with some of the boys who go to my school even though they are heterosexual. I can't help but wonder if just being friends with guys feels like it isn't enough. [NEWLINE] [NEWLINE] Thinking of my guy friends as girls... I think I would find some of their personalities cute but not exactly my type. There are girls I know but am not friends with nor really think of them romantically because their personalities are off-putting. I suppose it doesn't help to be in the environment I'm in that I'm only exposed to a small facet of personality from the girls I interact with from the all-girls school. If my best friend were a girl, I think I'd rather not date that person primarily for fear of losing that person once we separated. Other friends would make very odd girls... Overall, I'm not sure if I would date them-- it's a good hypothetical question but a difficult one to contemplate. [NEWLINE] [NEWLINE] More importantly, I wonder if we'd still be friends. The social system in my environment is very segregated between boys and girls. If we were to simply do something together, it'd be considered some sort of date and not a friendly outing, which is probably weird to most people. The fringe friends would probably not be my friends. I imagine it'd be hard to relate with them if they were girls while I think I'd stay in touch with my closer friends. This is a tough thought-experiment, so I'm sorry if I'm repetitive or giving conflicting answers. [NEWLINE] [NEWLINE] If I were to turn those girls at the all-girls school into guys, I
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You are ignoring the fact that cars take you to places and energy production creates quality of life,<mask> smoking is just a useless indulgence.<mask><mask> that, I do actually think that petrol-burning cars and power stations are toxic. [ENDQ] [NEWLINE] <mask><mask> /u/garnteller really hit the nail on the head. It seems that you have created a system of moving the goal-posts and are actually quite solidified in your position. I<mask> think /u/peacekitty<mask> hit it<mask> saying "individual smokers are the ones being considerate or inconsiderate." [NEWLINE] [NEWLINE] Your denial of the benefits of smoking<mask> highlights the extremity of your position. Don't get me wrong, smoking obviously has *overall* negative effects,<mask> your claim that the state brought about by smoking is the normal state of a non-smoker is factually inaccurate. That would be like telling someone that drinking a cup of coffee has no effect, it merely brings a person to the normal state of a non-coffee drinker. Cigarettes contain stimulants (<mask> coffee does) and they have a stimulating effect. This is a benefit. In your view (and mine), the stimulating effect of a cigarette is outweighed by the negative health effects. That does not mean the benefit is not real, and your denial of it shows a rigidity of mind not amenable to having your view changed. [NEWLINE] [NEWLINE] <mask><mask> that there are many times that smoking is inconsiderate, in confined spaces being the main one and the other main situation would be being in a place<mask> someone else needs to remain (like a bus stop).<mask> there are other cases<mask> it is fine, like in one's own home or out for a walk in a forest. The middle ground is<mask> walking down the street. Is it "inconsiderate"? Well, that depends on<mask> you consider "due regard for the rights or feelings of others". **Do you consider it *due* that this absolute stranger walking down this open public area concede to your desire rather than fulfil their own?** Perhaps they have fully considered you and other non-smokers and have come to the conclusion that their smoking on the street is no more harmful than the car-fumes you allow yourself to be around. Perhaps they consider the air quality and judge that, unless you too close to them (6ft<mask><mask> that article), you will be breathing mostly non-smoke and<mask> will be fine. **Who's responsibility is it to keep you
Label encoding: <s> [STARTQ] You are ignoring the fact that cars take you to places and energy production creates quality of life, while smoking is just a useless indulgence. But besides that, I do actually think that petrol-burning cars and power stations are toxic. [ENDQ] [NEWLINE] I think /u/garnteller really hit the nail on the head. It seems that you have created a system of moving the goal-posts and are actually quite solidified in your position. I also think /u/peacekitty also hit it when saying "individual smokers are the ones being considerate or inconsiderate." [NEWLINE] [NEWLINE] Your denial of the benefits of smoking also highlights the extremity of your position. Don't get me wrong, smoking obviously has *overall* negative effects, but your claim that the state brought about by smoking is the normal state of a non-smoker is factually inaccurate. That would be like telling someone that drinking a cup of coffee has no effect, it merely brings a person to the normal state of a non-coffee drinker. Cigarettes contain stimulants ( as coffee does) and they have a stimulating effect. This is a benefit. In your view (and mine), the stimulating effect of a cigarette is outweighed by the negative health effects. That does not mean the benefit is not real, and your denial of it shows a rigidity of mind not amenable to having your view changed. [NEWLINE] [NEWLINE] I agree that there are many times that smoking is inconsiderate, in confined spaces being the main one and the other main situation would be being in a place where someone else needs to remain (like a bus stop). But there are other cases where it is fine, like in one's own home or out for a walk in a forest. The middle ground is when walking down the street. Is it "inconsiderate"? Well, that depends on what you consider "due regard for the rights or feelings of others". **Do you consider it *due* that this absolute stranger walking down this open public area concede to your desire rather than fulfil their own?** Perhaps they have fully considered you and other non-smokers and have come to the conclusion that their smoking on the street is no more harmful than the car-fumes you allow yourself to be around. Perhaps they consider the air quality and judge that, unless you too close to them (6ft according to that article), you will be breathing mostly non-smoke and thus will be fine. **Who's responsibility is it to keep you
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [STARTQ] You insisted on having access to the entire menu or you would leave: Slightly douchy<mask> I gave you a pass<mask> it didn't seem like you made a scene about it. [ENDQ] [NEWLINE] It's hard to relate tone<mask> typing,<mask> I was in no way "douchey" about it. I simply said we weren't interested in a half menu, and that<mask> that's all they had, we'd go somewhere else. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] Then, after all of this,<mask> do you choose to repay the people who bent over backwards to ensure you had a good meal? You and your party proceeded to stay at the restaurant for THREE FUCKING HOURS AND THIRTY MINUTES after their scheduled closing time?!?! [ENDQ] [NEWLINE] Two and half hours: 11pm - 1:30am. Not that you'd find 2 and half any more palatable than three and a half,<mask> I want to keep things accurate. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] <mask> you seriously need a Reddit CMV in order to see exactly<mask> this shouldn't be acceptable, than I don't feel there is any way to change your view other than to sentence you to 1 year of customer service work<mask> you can see the flip side of<mask> you did to people who work long hours for shit pay and deal with obnoxious customers who think they are God's gift to retail/dining. [ENDQ] [NEWLINE] I'm from Orlando. My first job, back in high school, was selling popcorn at Disney World, and I worked in the Hospitality/Service Industries all the way through college,<mask> I've got plenty of experience in restaurants. [NEWLINE] [NEWLINE] And yes, I've been on the other side of this situation. Every server has rolled the dice and lost<mask> it comes to late tables: it starts dying down and you figure you're about to get cut,<mask> you start cleaning up, doing side-work, maybe even start rolling silver, and then BAM you get triple sat and your section fills up. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] My only hope is that the bill was one check for 12 people and they were tipped appropriately (more than the standard 10-20%). [ENDQ] [NEWLINE] No, actually, the restaurant is a "Pay first" type place,<mask> there was only one check, and I paid it. I tipped 20% on the check and gave the manager some cash to send back to the kitchen staff (<mask>
Label encoding: <s> [NEWLINE] [STARTQ] You insisted on having access to the entire menu or you would leave: Slightly douchy but I gave you a pass because it didn't seem like you made a scene about it. [ENDQ] [NEWLINE] It's hard to relate tone when typing, but I was in no way "douchey" about it. I simply said we weren't interested in a half menu, and that if that's all they had, we'd go somewhere else. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] Then, after all of this, how do you choose to repay the people who bent over backwards to ensure you had a good meal? You and your party proceeded to stay at the restaurant for THREE FUCKING HOURS AND THIRTY MINUTES after their scheduled closing time?!?! [ENDQ] [NEWLINE] Two and half hours: 11pm - 1:30am. Not that you'd find 2 and half any more palatable than three and a half, but I want to keep things accurate. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] If you seriously need a Reddit CMV in order to see exactly why this shouldn't be acceptable, than I don't feel there is any way to change your view other than to sentence you to 1 year of customer service work so you can see the flip side of what you did to people who work long hours for shit pay and deal with obnoxious customers who think they are God's gift to retail/dining. [ENDQ] [NEWLINE] I'm from Orlando. My first job, back in high school, was selling popcorn at Disney World, and I worked in the Hospitality/Service Industries all the way through college, so I've got plenty of experience in restaurants. [NEWLINE] [NEWLINE] And yes, I've been on the other side of this situation. Every server has rolled the dice and lost when it comes to late tables: it starts dying down and you figure you're about to get cut, so you start cleaning up, doing side-work, maybe even start rolling silver, and then BAM you get triple sat and your section fills up. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] My only hope is that the bill was one check for 12 people and they were tipped appropriately (more than the standard 10-20%). [ENDQ] [NEWLINE] No, actually, the restaurant is a "Pay first" type place, so there was only one check, and I paid it. I tipped 20% on the check and gave the manager some cash to send back to the kitchen staff ( as
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] NTSC says it was destroyed by fire that melted beams that caused it to collapse.<mask> you could point to any other steel building pre or post 9/11 that collapsed into itself in a single event due to fire, show me. The whole premise doesn't stack up and neither does the structural physics of the event. [ENDQ] [NEWLINE] Show me another skyscraper of similar size and construction with a failed sprinkler system, filled with combustibles, extensive structural damage (very large gash on South face) that had a fire and didn't collapse. [NEWLINE] [NEWLINE] <mask> for the "free fall collapse": [NEWLINE] [NEWLINE] [STARTQ] This analysis showed that the 40 percent longer descent time—compared to the 3.9 second free fall time—was due primarily to Stage 1, which corresponded to the buckling of the exterior columns in the lower stories of the north face. During Stage 2, the north face descended essentially in free fall, indicating negligible support from the structure below. This is consistent with the structural analysis model, which showed the exterior columns buckling and losing their capacity to support the loads from the structure above. In Stage 3, the acceleration decreased<mask> the upper portion of the north face encountered increased resistance from the collapsed structure and the debris pile below. [ENDQ] [NEWLINE] [Source]( [URL].cfm) [NEWLINE] [NEWLINE] [STARTQ] Again this needs a lot more investigation. [ENDQ] [NEWLINE] No it doesn't.  We have a very plausible explanation and most objections are just unusual things created by a lack of understanding of the Physics of this kind of collapse.  The rest is just playing off of coincidence and ominous statements. [NEWLINE] [NEWLINE] Our time, money, and effort are better spent on other things. [NEWLINE] [NEWLINE] [STARTQ] The girders and debris from the site were immediately removed from the scene of the crime and disposed of.<mask> was this done? [ENDQ] [NEWLINE] They wanted to clean up<mask> soon<mask> possible.  This "crime scene" happened to be in the heart of one of the largest and most important cities in the world.  Still, it took considerable time to clean out all the debris.  The steel was a valuable metal,<mask> it was melted and repurposed. [NEWLINE] [NEWLINE] [STARTQ] <mask> isn't any of the material available for analysts, apart from the few samples of dust collected by bystanders? [ENDQ] [NEWLINE] <mask> there was no evidence to suspect anything.  They didn't test for volcanic activity or meteor strikes either,<mask><mask> do you know it wasn't that? [NEWLINE] [NEWLINE] [STARTQ] <mask> can you explain the areas of extreme thermal activity present in debris
Label encoding: <s> [STARTQ] NTSC says it was destroyed by fire that melted beams that caused it to collapse. If you could point to any other steel building pre or post 9/11 that collapsed into itself in a single event due to fire, show me. The whole premise doesn't stack up and neither does the structural physics of the event. [ENDQ] [NEWLINE] Show me another skyscraper of similar size and construction with a failed sprinkler system, filled with combustibles, extensive structural damage (very large gash on South face) that had a fire and didn't collapse. [NEWLINE] [NEWLINE] As for the "free fall collapse": [NEWLINE] [NEWLINE] [STARTQ] This analysis showed that the 40 percent longer descent time—compared to the 3.9 second free fall time—was due primarily to Stage 1, which corresponded to the buckling of the exterior columns in the lower stories of the north face. During Stage 2, the north face descended essentially in free fall, indicating negligible support from the structure below. This is consistent with the structural analysis model, which showed the exterior columns buckling and losing their capacity to support the loads from the structure above. In Stage 3, the acceleration decreased as the upper portion of the north face encountered increased resistance from the collapsed structure and the debris pile below. [ENDQ] [NEWLINE] [Source]( [URL].cfm) [NEWLINE] [NEWLINE] [STARTQ] Again this needs a lot more investigation. [ENDQ] [NEWLINE] No it doesn't.  We have a very plausible explanation and most objections are just unusual things created by a lack of understanding of the Physics of this kind of collapse.  The rest is just playing off of coincidence and ominous statements. [NEWLINE] [NEWLINE] Our time, money, and effort are better spent on other things. [NEWLINE] [NEWLINE] [STARTQ] The girders and debris from the site were immediately removed from the scene of the crime and disposed of. Why was this done? [ENDQ] [NEWLINE] They wanted to clean up as soon as possible.  This "crime scene" happened to be in the heart of one of the largest and most important cities in the world.  Still, it took considerable time to clean out all the debris.  The steel was a valuable metal, so it was melted and repurposed. [NEWLINE] [NEWLINE] [STARTQ] Why isn't any of the material available for analysts, apart from the few samples of dust collected by bystanders? [ENDQ] [NEWLINE] Because there was no evidence to suspect anything.  They didn't test for volcanic activity or meteor strikes either, so how do you know it wasn't that? [NEWLINE] [NEWLINE] [STARTQ] How can you explain the areas of extreme thermal activity present in debris
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well it's a good thing I don't have a "get up and just do it" attitude. Bro, you're preaching to the choir. I didn't respond to your other post<mask><mask><mask> with some of your finer points,<mask> overall,<mask><mask> your head is in the right place. I<mask> didn't know that bit about food deserts. I mean, the nearest grocery store to me is about a mile,<mask> in the city, people walk that like it's nothing. [NEWLINE] [NEWLINE] I just hate it<mask> people walk around spewing "well they just don't know better<mask> they're poor and uneducated!!" [NEWLINE] [NEWLINE] Poor people aren't children or idiots. They usually do not want help, and use the same excuses that we hand them by accident through pity.<mask><mask> the best thing you can do to encourage someone to change their behavior is give them responsibility. [NEWLINE] [NEWLINE] I mean,<mask> anyone ever asks me<mask> is healthy, I will certainly help them out.<mask>, I am not going to tell the guy spending his food stamps on kool aid and potato chips that he is making poor decisions. He likely doesn't give a fuck, and it is not my place to tell him<mask> to live his life. [NEWLINE] [NEWLINE] <mask>,<mask> he then starts complaining about<mask> unhealthy he is<mask> he is poor, I will not let him get away with being<mask> irresponsible for his decisions. "No sir, it isn't<mask> you're poor, it's<mask> you'd rather drink cool aid than water like I do, and it's<mask> you'd rather eat potato chips than salad." [NEWLINE] [NEWLINE] And it isn't like they like or dislike being unhealthy, they just don't care.<mask> the fuck are they trying to live longer for? They just want to live<mask> life is hard, and<mask> they work harder to make a hard life longer instead of relaxing, well that's just stupid, and "<mask> do you think I am an idiot?" [NEWLINE] [NEWLINE] I have<mask> struggled (and still do) with depression/bipolar disorder.<mask> I know<mask> it feels. I was anorexic for a time<mask> the depression was particularly bad. I didn't feel anything, food tasted like cardboard, and I certainly didn't care that<mask> I don't eat something I could lose an organ and "shut the fuck up you have to put this down your throat or your grounded". [NEWLINE] [NEWLINE] Just saying that making excuses for people means they don't even have to excuse themselves. They can go right on doing their
Label encoding: <s>Well it's a good thing I don't have a "get up and just do it" attitude. Bro, you're preaching to the choir. I didn't respond to your other post because I disagree with some of your finer points, but overall, I think your head is in the right place. I also didn't know that bit about food deserts. I mean, the nearest grocery store to me is about a mile, but in the city, people walk that like it's nothing. [NEWLINE] [NEWLINE] I just hate it when people walk around spewing "well they just don't know better because they're poor and uneducated!!" [NEWLINE] [NEWLINE] Poor people aren't children or idiots. They usually do not want help, and use the same excuses that we hand them by accident through pity. I think the best thing you can do to encourage someone to change their behavior is give them responsibility. [NEWLINE] [NEWLINE] I mean, if anyone ever asks me what is healthy, I will certainly help them out. However, I am not going to tell the guy spending his food stamps on kool aid and potato chips that he is making poor decisions. He likely doesn't give a fuck, and it is not my place to tell him how to live his life. [NEWLINE] [NEWLINE] However, if he then starts complaining about how unhealthy he is because he is poor, I will not let him get away with being so irresponsible for his decisions. "No sir, it isn't because you're poor, it's because you'd rather drink cool aid than water like I do, and it's because you'd rather eat potato chips than salad." [NEWLINE] [NEWLINE] And it isn't like they like or dislike being unhealthy, they just don't care. What the fuck are they trying to live longer for? They just want to live because life is hard, and if they work harder to make a hard life longer instead of relaxing, well that's just stupid, and " what do you think I am an idiot?" [NEWLINE] [NEWLINE] I have also struggled (and still do) with depression/bipolar disorder. So I know how it feels. I was anorexic for a time when the depression was particularly bad. I didn't feel anything, food tasted like cardboard, and I certainly didn't care that if I don't eat something I could lose an organ and "shut the fuck up you have to put this down your throat or your grounded". [NEWLINE] [NEWLINE] Just saying that making excuses for people means they don't even have to excuse themselves. They can go right on doing their
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My school county in high school actually did make this a requirement (Broward County, FL). I<mask> graduated from an IB school,<mask> I can relate to your experience very well.<mask>, I am completely against community service. Here's<mask>, in a nutshell: [NEWLINE] [NEWLINE] * The community was never REALLY helped in any significant way by any community service either I or any of my friends did (aside from CAS project fundraisers). [NEWLINE] [NEWLINE] * Nothing I could do for service was interesting AT ALL. for it to count<mask> "service" it must be by yourself or in the service of a non-profit organization.<mask>, you could not do any meaningful, or professional work.<mask> basically, all service activities consisted of selling things at a concession stand, handing out flyers, washing cars, or cleaning up horse shit (I *shit* you not). None of these activities taught me any important lessons or gave me any valuable skills, or really any skills I would use in the future. [NEWLINE] [NEWLINE] * Responsibility is already taught thoroughly through homework and other already-mandated school activities and requirements. [NEWLINE] [NEWLINE] * The "extracurricular activities" they involve really are not looked upon favorably by colleges anyways. Honestly,<mask> I picked up horse shit for 30 hours during high school,<mask> would any university or institute look upon me more favorably? In my case, I'm fairly certain an institute of technology could give two shits about<mask> I washed cars every weekend my senior year. [NEWLINE] [NEWLINE] * It may not require much time in theory,<mask> the only time you can preform these activities severely inhibits your ability to take part in more meaningful and professionally beneficial work. In my district you could not preform service hours during school hours. This means it's limited to after school and weekends. Being an IB student, I'm sure you can relate: there was *no fucking way* I was going to do all that work at school, do my homework, and still have time for service after school.<mask> that leaves the weekends. Well, I gave up about 3 years of working professionally (in a data center,<mask> I'm currently employed, having graduated this year) which could have much better improved my college applications, provided a source of income to save for college, and taught me responsibility much better than any community service requirement. [NEWLINE] [NEWLINE] <mask>, I've always felt that<mask> the opportunity to use some of my real skills for service (something IT-related) it seriously teaches the youth of our country to not value their abilities
Label encoding: <s>My school county in high school actually did make this a requirement (Broward County, FL). I also graduated from an IB school, so I can relate to your experience very well. However, I am completely against community service. Here's why, in a nutshell: [NEWLINE] [NEWLINE] * The community was never REALLY helped in any significant way by any community service either I or any of my friends did (aside from CAS project fundraisers). [NEWLINE] [NEWLINE] * Nothing I could do for service was interesting AT ALL. for it to count as "service" it must be by yourself or in the service of a non-profit organization. Thus, you could not do any meaningful, or professional work. So basically, all service activities consisted of selling things at a concession stand, handing out flyers, washing cars, or cleaning up horse shit (I *shit* you not). None of these activities taught me any important lessons or gave me any valuable skills, or really any skills I would use in the future. [NEWLINE] [NEWLINE] * Responsibility is already taught thoroughly through homework and other already-mandated school activities and requirements. [NEWLINE] [NEWLINE] * The "extracurricular activities" they involve really are not looked upon favorably by colleges anyways. Honestly, if I picked up horse shit for 30 hours during high school, why would any university or institute look upon me more favorably? In my case, I'm fairly certain an institute of technology could give two shits about how I washed cars every weekend my senior year. [NEWLINE] [NEWLINE] * It may not require much time in theory, but the only time you can preform these activities severely inhibits your ability to take part in more meaningful and professionally beneficial work. In my district you could not preform service hours during school hours. This means it's limited to after school and weekends. Being an IB student, I'm sure you can relate: there was *no fucking way* I was going to do all that work at school, do my homework, and still have time for service after school. So that leaves the weekends. Well, I gave up about 3 years of working professionally (in a data center, where I'm currently employed, having graduated this year) which could have much better improved my college applications, provided a source of income to save for college, and taught me responsibility much better than any community service requirement. [NEWLINE] [NEWLINE] Also, I've always felt that if the opportunity to use some of my real skills for service (something IT-related) it seriously teaches the youth of our country to not value their abilities
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to speak anecdotally here. I had similar feelings about antidepressants (specifically SSRIs) for a very long time. Then I started to struggle with OCD. I didn't know<mask> was wrong with me at first,<mask> my doctor and I worked out a dosage of Zoloft that worked for me.<mask> you don't know much about OCD, I was basically suffering from extreme bouts of anxious thoughts that would in turn leave me depressed and almost unable to move. I felt bad/guilty about *everything*. [NEWLINE] [NEWLINE] I went on and off Zoloft a few times, and I tried different dosages<mask> well. After a<mask>, it did feel to me that the pills made everything a bit "numb" -<mask><mask> I felt *better* for a time, I<mask> didn't feel completely like myself. To clarify, they did not put me into "happy land" and make everything seem like sunshine and rainbows. They just made it easier for me *not to care* about the things that used to make me anxious or upset.<mask>, the side effects were a little too much for me (this is an *individual* problem - not everyone experiences negative side effects). I made an effort to step down safely off the medication (with doctor supervision), and work on behavioral therapy techniques to combat my anxiety in the future. [NEWLINE] [NEWLINE] Which leads me to my main point to you. It's easy<mask> you've not taken antidepressants to look at them<mask> a "quick fix" or a band-aid for your problems.<mask> that's not<mask> they work. They are more of a tool to pull you out of a negative feedback loop of anxiety and depression<mask> you just cannot do it yourself. You<mask> have to keep in mind that they tend not to work "by themselves". You should<mask> see a therapist to help change your thought processes. There are many techniques you can learn that just involve changing your thinking pattern. It's really just training yourself to handle depressive/anxious thoughts in a constructive way. [NEWLINE] [NEWLINE] For some people, anti-depressants are only a temporary solution, and they can eventually stop taking them. For other people, it's a lot easier to slip back into old patterns<mask> they are not taking medication. And that *okay*; people are different, they have different bodies, and they require different tools to manage their illness.<mask> taking an antidepressant is no less valid than using any other medication to manage any other illness. [NEWLINE] [NEWLINE] It sounds
Label encoding: <s>I'm going to speak anecdotally here. I had similar feelings about antidepressants (specifically SSRIs) for a very long time. Then I started to struggle with OCD. I didn't know what was wrong with me at first, but my doctor and I worked out a dosage of Zoloft that worked for me. If you don't know much about OCD, I was basically suffering from extreme bouts of anxious thoughts that would in turn leave me depressed and almost unable to move. I felt bad/guilty about *everything*. [NEWLINE] [NEWLINE] I went on and off Zoloft a few times, and I tried different dosages as well. After a while, it did feel to me that the pills made everything a bit "numb" - so while I felt *better* for a time, I also didn't feel completely like myself. To clarify, they did not put me into "happy land" and make everything seem like sunshine and rainbows. They just made it easier for me *not to care* about the things that used to make me anxious or upset. However, the side effects were a little too much for me (this is an *individual* problem - not everyone experiences negative side effects). I made an effort to step down safely off the medication (with doctor supervision), and work on behavioral therapy techniques to combat my anxiety in the future. [NEWLINE] [NEWLINE] Which leads me to my main point to you. It's easy when you've not taken antidepressants to look at them as a "quick fix" or a band-aid for your problems. But that's not how they work. They are more of a tool to pull you out of a negative feedback loop of anxiety and depression when you just cannot do it yourself. You also have to keep in mind that they tend not to work "by themselves". You should also see a therapist to help change your thought processes. There are many techniques you can learn that just involve changing your thinking pattern. It's really just training yourself to handle depressive/anxious thoughts in a constructive way. [NEWLINE] [NEWLINE] For some people, anti-depressants are only a temporary solution, and they can eventually stop taking them. For other people, it's a lot easier to slip back into old patterns when they are not taking medication. And that *okay*; people are different, they have different bodies, and they require different tools to manage their illness. But taking an antidepressant is no less valid than using any other medication to manage any other illness. [NEWLINE] [NEWLINE] It sounds
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe "Thou shalt not kill" is a fundamental rule in life,<mask><mask> I'm not that religious i just believe killing another person is the most heinous act a person can commit,<mask> soldiers go out to other countries and are paid to take lives<mask> they return<mask> heroes. Now i know there are other jobs to do in the army, not just killing<mask> the key part for most of the soldiers is to kill. [ENDQ] [NEWLINE] The concept of "Thou shalt not kill" is<mask>... misinterpreted.  It's technically, "Thou shalt not murder",<mask> murder is an unlawful killing. [NEWLINE] [NEWLINE] War... is lawful by nature.  The reasons for war are always just in the two or more parties involved.  For the US, sending troops overseas is done to kill people who intend to do harm to us in our homes... for those they're killing, they likewise respond that they are fighting to repel the US from their lands or that we are being punished for another transgression (stationing troops near their holy lands or supporting Israel). [NEWLINE] [NEWLINE] Soldiers, by definition, have no choice to kill.  For them, its typically a binary solution: Kill or Be Killed. <mask> we could win wars with pool noodles alone, we would,<mask> we can't.  To expand on the lack of choice: soldiers are<mask> tools of the State.  They cannot choose<mask> they are deployed and<mask>.  That is the nature of a soldier. [NEWLINE] [NEWLINE] [STARTQ] <mask> are they any different to murderers who kill over jealousy or greed?<mask> are they shown<mask> much respect for killing in 'the name of their country'? [ENDQ] [NEWLINE] Simply put:<mask> they are not acting out of those petty needs.  Soldiers rarely want to be in some shithole far from home<mask> their wife is probably sucking their neighbors dick.  They typically signed up<mask> they were either: idealistic, in poor financial situations, or couldn't do anything else.  For them, a few years in uniform with housing and food is better than the alternatives of poverty they faced at home. [NEWLINE] [NEWLINE] Soldiers<mask> kill in order to avoid being killed or to protect their countrymen. <mask> they find out terrorists have a chemical weapon, they will try and stop them before that weapon can be delivered.  The terrorists will fight back, and people will die.  There is no other choice. [NEWLINE] [NEWLINE] [STARTQ] It is<mask><mask> that taking lives, for whatever reason, is wrong. Soldiers don't deserve respect
Label encoding: <s> [STARTQ] I believe "Thou shalt not kill" is a fundamental rule in life, even though I'm not that religious i just believe killing another person is the most heinous act a person can commit, yet soldiers go out to other countries and are paid to take lives but they return as heroes. Now i know there are other jobs to do in the army, not just killing but the key part for most of the soldiers is to kill. [ENDQ] [NEWLINE] The concept of "Thou shalt not kill" is also... misinterpreted.  It's technically, "Thou shalt not murder", where murder is an unlawful killing. [NEWLINE] [NEWLINE] War... is lawful by nature.  The reasons for war are always just in the two or more parties involved.  For the US, sending troops overseas is done to kill people who intend to do harm to us in our homes... for those they're killing, they likewise respond that they are fighting to repel the US from their lands or that we are being punished for another transgression (stationing troops near their holy lands or supporting Israel). [NEWLINE] [NEWLINE] Soldiers, by definition, have no choice to kill.  For them, its typically a binary solution: Kill or Be Killed.  If we could win wars with pool noodles alone, we would, but we can't.  To expand on the lack of choice: soldiers are also tools of the State.  They cannot choose where they are deployed and why.  That is the nature of a soldier. [NEWLINE] [NEWLINE] [STARTQ] How are they any different to murderers who kill over jealousy or greed? Why are they shown so much respect for killing in 'the name of their country'? [ENDQ] [NEWLINE] Simply put: because they are not acting out of those petty needs.  Soldiers rarely want to be in some shithole far from home where their wife is probably sucking their neighbors dick.  They typically signed up because they were either: idealistic, in poor financial situations, or couldn't do anything else.  For them, a few years in uniform with housing and food is better than the alternatives of poverty they faced at home. [NEWLINE] [NEWLINE] Soldiers also kill in order to avoid being killed or to protect their countrymen.  If they find out terrorists have a chemical weapon, they will try and stop them before that weapon can be delivered.  The terrorists will fight back, and people will die.  There is no other choice. [NEWLINE] [NEWLINE] [STARTQ] It is my opinion that taking lives, for whatever reason, is wrong. Soldiers don't deserve respect
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Dude.  You've essentially issued a non-denial denial: a number of wild accusations/assertions without debating the actual merits of the analogy at hand. [NEWLINE] [NEWLINE] Mistake #1: You imply I'm a woman.  I'm a guy. [NEWLINE] Mistake #2: You've somehow assumed I am casting "a minority group in a negative stereotype."<mask> an ethnic minority in the US myself, I'd be *delighted* to hear<mask> my analogy does that, exactly. Seems rather that you've made a number of stereotypical assumptions yourself.  Are you willing to come clean on this? [NEWLINE] [NEWLINE] Mistake #3:  "Men are the demographic that actually has the highest rate of victimisation..." First, you seem unwilling to back that assertion up with a citation (and the burden is on the asserter, not everyone else). <mask> for the sake of argument let's say you're right.  It's still disingenuous<mask> a more nuanced understanding of crime and women reveals that women have the highest rate of *sexual* assault, which is the more relevant crime at hand here<mask> we're talking about unwanted sexual attention. <mask> you've cherry-picked your facts, such<mask> they may be. [NEWLINE] [NEWLINE] Mistake #4:  You won't debate the actual merits of the argument.  I would infer that it is<mask> you're actually having trouble picking holes in the mugging analogy.  To show you<mask> it's done, here's an example. [NEWLINE] [NEWLINE] Your "women are just like celebrities" analogy,<mask> apt in a general sense, fails on closer inspection in the following ways: [NEWLINE] [NEWLINE] First, being an actor/celebrity is a chosen avocation.  Women do not get to choose being a woman. [NEWLINE] [NEWLINE] Second,<mask> you would be absolutely correct in saying that the law is against celebrities<mask> it comes to getting unwanted attention in public, the analogy fails<mask> considering that you are essentially defending paparazzi.  In other words,<mask> unwanted attention may be legal, it still makes you an asshole. [NEWLINE] [NEWLINE] Thirdly, your analogies and examples make narrow assumptions around women's motivations.  You assume that<mask> women look nice, it is solely to "get attention"<mask> ignoring other factors such<mask> *wanting to look nice for themselves.*  Again, consider the "nice watch" analogy.  Everyone's got some item they like to wear *<mask> they like it* not necessarily<mask> they're demanding attention from others
Label encoding: <s>Dude.  You've essentially issued a non-denial denial: a number of wild accusations/assertions without debating the actual merits of the analogy at hand. [NEWLINE] [NEWLINE] Mistake #1: You imply I'm a woman.  I'm a guy. [NEWLINE] Mistake #2: You've somehow assumed I am casting "a minority group in a negative stereotype." As an ethnic minority in the US myself, I'd be *delighted* to hear how my analogy does that, exactly. Seems rather that you've made a number of stereotypical assumptions yourself.  Are you willing to come clean on this? [NEWLINE] [NEWLINE] Mistake #3:  "Men are the demographic that actually has the highest rate of victimisation..." First, you seem unwilling to back that assertion up with a citation (and the burden is on the asserter, not everyone else).  But for the sake of argument let's say you're right.  It's still disingenuous as a more nuanced understanding of crime and women reveals that women have the highest rate of *sexual* assault, which is the more relevant crime at hand here since we're talking about unwanted sexual attention.  So you've cherry-picked your facts, such as they may be. [NEWLINE] [NEWLINE] Mistake #4:  You won't debate the actual merits of the argument.  I would infer that it is because you're actually having trouble picking holes in the mugging analogy.  To show you how it's done, here's an example. [NEWLINE] [NEWLINE] Your "women are just like celebrities" analogy, while apt in a general sense, fails on closer inspection in the following ways: [NEWLINE] [NEWLINE] First, being an actor/celebrity is a chosen avocation.  Women do not get to choose being a woman. [NEWLINE] [NEWLINE] Second, while you would be absolutely correct in saying that the law is against celebrities when it comes to getting unwanted attention in public, the analogy fails when considering that you are essentially defending paparazzi.  In other words, while unwanted attention may be legal, it still makes you an asshole. [NEWLINE] [NEWLINE] Thirdly, your analogies and examples make narrow assumptions around women's motivations.  You assume that when women look nice, it is solely to "get attention" while ignoring other factors such as *wanting to look nice for themselves.*  Again, consider the "nice watch" analogy.  Everyone's got some item they like to wear * because they like it* not necessarily because they're demanding attention from others
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] &gt; Men who have been forcefully penetrated are still considered to have been raped [ENDQ] [NEWLINE] [STARTQ] <mask> women who rape men<mask> often forcefully penetrate them... [ENDQ] [NEWLINE] I'm just establishing facts here, these are the "CDC definitions" in question. Anything further you're reading into this was unintended. [NEWLINE] [NEWLINE] [STARTQ] &gt; most of these men who have been assaulted in this way will not identify with or label their experience<mask> "rape", and<mask> the reporting statistics won't accurately reflect the incidence of sexual violence against men. [ENDQ] [NEWLINE] [STARTQ] <mask> now "forced intercourse without consent" isn't actually rape. I mean shit,<mask> we want to reduce female rape victimization rates, we just need to indoctrinate them into thinking they can't be raped and then VOILA... they aren't actually rape victims anymore<mask> it happens<mask> they don't think<mask>. [ENDQ] [NEWLINE] [STARTQ] I like your logic here. /sarcasm [ENDQ] [NEWLINE] You're misrepresenting my position. [NEWLINE] [NEWLINE] Look, again,<mask> you have a semantic problem with<mask> it is called, I can respect that position. I can agree that in an ideal world, we should view sexual assaults against men and women<mask> equally heinous, and probably collectively use the same word to describe them.<mask> I am specifically saying that the word choice here is to make sure these crimes against men *are* documented, and not to sweep them under the rug - the contention is that the cases would go unreported<mask> rape,<mask> make a new category<mask> that these assaults are reported somehow. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] &gt;<mask>, for the first time, men who were surveyed about sexual violence were specifically asked<mask> they had ever been made to penetrate against their will separately from being asked about rape, and in doing<mask> the survey uncovered a whole new category of men who were the victims of sexual violence. [ENDQ] [NEWLINE] [STARTQ] OR, you know, they could just treat male victims of rape<mask> actual rape victims instead of grouping them in the same category<mask> "non-contact unwanted sexual experiences".<mask> instead, people like Koss **want to be able to "conclude"** that women are victims of rape 10x more than men. [ENDQ] [STARTQ] She's **willfully** misrepresented studies/statistics every chance she gets to inflate female victimization rates to be significantly higher than male's. This has result in a massive difference in policy towards both of them (note<mask> it's VAWA and not VAPA) and funding for victim services. **I don't really
Label encoding: <s> [STARTQ] &gt; Men who have been forcefully penetrated are still considered to have been raped [ENDQ] [NEWLINE] [STARTQ] Because women who rape men so often forcefully penetrate them... [ENDQ] [NEWLINE] I'm just establishing facts here, these are the "CDC definitions" in question. Anything further you're reading into this was unintended. [NEWLINE] [NEWLINE] [STARTQ] &gt; most of these men who have been assaulted in this way will not identify with or label their experience as "rape", and so the reporting statistics won't accurately reflect the incidence of sexual violence against men. [ENDQ] [NEWLINE] [STARTQ] So now "forced intercourse without consent" isn't actually rape. I mean shit, if we want to reduce female rape victimization rates, we just need to indoctrinate them into thinking they can't be raped and then VOILA... they aren't actually rape victims anymore when it happens because they don't think so. [ENDQ] [NEWLINE] [STARTQ] I like your logic here. /sarcasm [ENDQ] [NEWLINE] You're misrepresenting my position. [NEWLINE] [NEWLINE] Look, again, if you have a semantic problem with what it is called, I can respect that position. I can agree that in an ideal world, we should view sexual assaults against men and women as equally heinous, and probably collectively use the same word to describe them. But I am specifically saying that the word choice here is to make sure these crimes against men *are* documented, and not to sweep them under the rug - the contention is that the cases would go unreported as rape, so make a new category so that these assaults are reported somehow. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] &gt; where, for the first time, men who were surveyed about sexual violence were specifically asked if they had ever been made to penetrate against their will separately from being asked about rape, and in doing so the survey uncovered a whole new category of men who were the victims of sexual violence. [ENDQ] [NEWLINE] [STARTQ] OR, you know, they could just treat male victims of rape as actual rape victims instead of grouping them in the same category as "non-contact unwanted sexual experiences". But instead, people like Koss **want to be able to "conclude"** that women are victims of rape 10x more than men. [ENDQ] [STARTQ] She's **willfully** misrepresented studies/statistics every chance she gets to inflate female victimization rates to be significantly higher than male's. This has result in a massive difference in policy towards both of them (note how it's VAWA and not VAPA) and funding for victim services. **I don't really
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People have already laid out these points,<mask> I feel like understanding an unfamiliar perspective can sometimes be a simple matter of wording,<mask> lemme give it a shot. [NEWLINE] [NEWLINE] Race is real. Race is real like money, and national borders, and adoption, and the meaning of these squiggles I'm writing with--it's real<mask> enough people collectively agreed that it was real and gave it importance. It isn't biologically or intrinsically real (<mask><mask> we tried to find its 'location'<mask> to speak) and it *shouldn't* matter,<mask> thanks to the actions of the past (and present) it continues to be real. Racism makes no sense<mask> these differences are based in our treatment of each other, our preference for one person or another, our limiting of people by their colour or nationality...not on something inherent. Racism treats people<mask><mask><mask> we *think* they are<mask> a category, not who they are<mask> a person or a biological being. [NEWLINE] [NEWLINE] These erroneous thoughts have consequences.<mask><mask> we can agree that it would be foolish to say "race has no biological basis,<mask> slavery had nothing to do with dark skin." Equally, these past actions continued to have effects with each passing generation. For example, we know that (<mask><mask> race) people whose parents didn't go to college are themselves less likely to go to college. We know that people whose parents are poor will not inherit much money, and<mask> have less assistance. We know that people whose parents were incarcerated are more likely to be incarcerated.<mask><mask> your slave great grandparents couldn't own property and it was illegal to teach them to read...you're starting off a little behind everyone else. [NEWLINE] [NEWLINE] <mask> those are ripples of the past, and hopefully they'll catch up to the present, right? Once the differences of the past are smoothed out, perhaps some racism will diminish.<mask>, people still discriminate based on race, even most of us think racism is bullshit. My grandpa wouldn't talk to me<mask> I brought a black girlfriend home. There are literal neo-nazis out there in the world. My girlfriend is 3 times more likely to be raped than my sister,<mask> my girlfriend is Native American.<mask> I'm saying is that there are not only the effects of the past that people have to catch up to,<mask> the stereotypes of the present that make it<mask> that people are still hurt or disliked<mask> of their colour. Are you doing it? I doubt it. Especially not
Label encoding: <s>People have already laid out these points, but I feel like understanding an unfamiliar perspective can sometimes be a simple matter of wording, so lemme give it a shot. [NEWLINE] [NEWLINE] Race is real. Race is real like money, and national borders, and adoption, and the meaning of these squiggles I'm writing with--it's real because enough people collectively agreed that it was real and gave it importance. It isn't biologically or intrinsically real ( even though we tried to find its 'location' so to speak) and it *shouldn't* matter, but thanks to the actions of the past (and present) it continues to be real. Racism makes no sense because these differences are based in our treatment of each other, our preference for one person or another, our limiting of people by their colour or nationality...not on something inherent. Racism treats people according to how we *think* they are as a category, not who they are as a person or a biological being. [NEWLINE] [NEWLINE] These erroneous thoughts have consequences. I think we can agree that it would be foolish to say "race has no biological basis, therefore slavery had nothing to do with dark skin." Equally, these past actions continued to have effects with each passing generation. For example, we know that ( regardless of race) people whose parents didn't go to college are themselves less likely to go to college. We know that people whose parents are poor will not inherit much money, and therefore have less assistance. We know that people whose parents were incarcerated are more likely to be incarcerated. So when your slave great grandparents couldn't own property and it was illegal to teach them to read...you're starting off a little behind everyone else. [NEWLINE] [NEWLINE] But those are ripples of the past, and hopefully they'll catch up to the present, right? Once the differences of the past are smoothed out, perhaps some racism will diminish. However, people still discriminate based on race, even most of us think racism is bullshit. My grandpa wouldn't talk to me if I brought a black girlfriend home. There are literal neo-nazis out there in the world. My girlfriend is 3 times more likely to be raped than my sister, because my girlfriend is Native American. What I'm saying is that there are not only the effects of the past that people have to catch up to, but the stereotypes of the present that make it so that people are still hurt or disliked because of their colour. Are you doing it? I doubt it. Especially not
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That is not true. He complained to people in higher positions in the NSA,<mask> he never actually attempted to become a legally defined whistleblower. [ENDQ] [NEWLINE] <mask> you are threatened by an organization that makes disinformation and manufacturing discredit, this is like saying "he didn't do<mask> it took to get the corrupt mayor to investigate the chief of police". You're forgetting that character assassination and counterintelligence is one of the NSA's most skillfully wielded tools. You don't snitch on the mob to the don just<mask> the law says you have to. [NEWLINE] [NEWLINE] [STARTQ] Should the NSA be invading the privacy of US citizens? Not at all. [ENDQ] [NEWLINE] Agreed. [NEWLINE] [NEWLINE] [STARTQ] Should Snowden release US strategies for spying on foreign powers? [ENDQ] [NEWLINE] <mask> they are being used against American Citizens, yes. Yes they should.<mask> the NSA doesn't want people to disclose their methods, then they need to not use them against the American people. The moment the NSA spies on us, their tactics should be subject to scrutiny. Seeing<mask> there is no check or balance on<mask> they chose to do, exposure became a moral imperative greater then the damage that disclosure would do to our foreign intelligence efforts. [NEWLINE] [NEWLINE] [STARTQ] We already know we can't trust our enemies [ENDQ] [NEWLINE] Our enemies are not the American People. I wouldn't support Snowden<mask> he disclosed legal programs that spied solely on foreign nationals (friendly or otherwise). The NSA created the situation, Snowden performed a highly patriotic duty. [NEWLINE] [NEWLINE] [STARTQ]. You believe that the domestic spying actions he exposed makes up for that, I suppose. [ENDQ] [NEWLINE] I very much do think the domestic spying actions he exposed made up for any damage he may have done to our spying program. It would be different were the abuses not<mask> incredibly overreaching,<mask> they were. We can't choose<mask> actions we will allow to be done to us 'for our protection'<mask> we don't know<mask> they are. [NEWLINE] [NEWLINE] [STARTQ] In that case, does a woman showing her chest to a police officer get her out of the speeding ticket? It shouldn't. No matter<mask>, that woman, just like Snowden, broke the law.<mask> things they do before/after does not matter. The law was broken. Someone who breaks the law, like espionage and treason, is not patriot. They are despicable criminals to our society and government in general. [ENDQ] [NEWLINE] This is not a black-and-white issue and your failure to understand scale and seriousness is harming your argument. I assume you think that watergate was a bad thing
Label encoding: <s> [STARTQ] That is not true. He complained to people in higher positions in the NSA, but he never actually attempted to become a legally defined whistleblower. [ENDQ] [NEWLINE] When you are threatened by an organization that makes disinformation and manufacturing discredit, this is like saying "he didn't do what it took to get the corrupt mayor to investigate the chief of police". You're forgetting that character assassination and counterintelligence is one of the NSA's most skillfully wielded tools. You don't snitch on the mob to the don just because the law says you have to. [NEWLINE] [NEWLINE] [STARTQ] Should the NSA be invading the privacy of US citizens? Not at all. [ENDQ] [NEWLINE] Agreed. [NEWLINE] [NEWLINE] [STARTQ] Should Snowden release US strategies for spying on foreign powers? [ENDQ] [NEWLINE] When they are being used against American Citizens, yes. Yes they should. If the NSA doesn't want people to disclose their methods, then they need to not use them against the American people. The moment the NSA spies on us, their tactics should be subject to scrutiny. Seeing as there is no check or balance on what they chose to do, exposure became a moral imperative greater then the damage that disclosure would do to our foreign intelligence efforts. [NEWLINE] [NEWLINE] [STARTQ] We already know we can't trust our enemies [ENDQ] [NEWLINE] Our enemies are not the American People. I wouldn't support Snowden if he disclosed legal programs that spied solely on foreign nationals (friendly or otherwise). The NSA created the situation, Snowden performed a highly patriotic duty. [NEWLINE] [NEWLINE] [STARTQ]. You believe that the domestic spying actions he exposed makes up for that, I suppose. [ENDQ] [NEWLINE] I very much do think the domestic spying actions he exposed made up for any damage he may have done to our spying program. It would be different were the abuses not so incredibly overreaching, but they were. We can't choose what actions we will allow to be done to us 'for our protection' if we don't know what they are. [NEWLINE] [NEWLINE] [STARTQ] In that case, does a woman showing her chest to a police officer get her out of the speeding ticket? It shouldn't. No matter what, that woman, just like Snowden, broke the law. What things they do before/after does not matter. The law was broken. Someone who breaks the law, like espionage and treason, is not patriot. They are despicable criminals to our society and government in general. [ENDQ] [NEWLINE] This is not a black-and-white issue and your failure to understand scale and seriousness is harming your argument. I assume you think that watergate was a bad thing
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually, he *can* do both, and judges do both just about every single day. [NEWLINE] [NEWLINE] They are bound by the caselaw that came before them,<mask> they have leeway<mask> it comes to certain discretionary items: for instance, bond conditions (determined based on flight risk and threat to community, etc.) and sentencing (determined based on seriousness of crime, history of defendant, sentencing guidelines, and argument by counsel). I'm not sure of your knowledge of the justice system<mask> I'll expound just a little bit here, forgive me<mask> it's unnecessary. [NEWLINE] [NEWLINE] In regards to sentencing; in the United States, judges decide criminal sentences based on the criminal sentencing guidelines, which amounts to a series of checks that help them analyze a defendant's role in the commission of a certain crime<mask> well<mask> the defendant's criminal history. Essentially, the prosecutor<mask> well<mask> the defense attorney will "score" the Defendant and compare the numbers they come up with; there is room for argument here, which is<mask> prep for sentencing can be a laborious procedure for an attorney. The "score" that each attorney comes up with will correspond to a table of values, for each of which there is a *high* end and a *low* end. The judge can, based on his/her take on the case and the arguments of counsel, decide<mask> within that bracket to sentence the Defendant. [NEWLINE] [NEWLINE] I wholeheartedly disagree that for a law to be "just" that it must carry with it a universal standard (assuming you mean the sentence/punishment here). The rigidity you suggest is untenable for two reasons:<mask> each and every single court case is unique and has factors both mitigating and aggravating; and, I would posit, this flexibility is one of the core tenets of the advocative justice system (i.e. you choose someone to represent you in a court of law<mask> you trust they will fight like hell for you, knowing that on the other side there is someone doing the same thing to fight like hell against you). The nature of the system necessitates a certain amount of "judgment calls." [NEWLINE] [NEWLINE] <mask><mask><mask> stare decisis goes, a good way to look at it,<mask><mask><mask>, is to view caselaw<mask> the material that fills in the "negative space" around which the judge determines a sentence: they cannot go outside the boundaries of the caselaw the precedes them.<mask><mask><mask> they are operating within those boundaries,<mask>, and within
Label encoding: <s>Actually, he *can* do both, and judges do both just about every single day. [NEWLINE] [NEWLINE] They are bound by the caselaw that came before them, but they have leeway when it comes to certain discretionary items: for instance, bond conditions (determined based on flight risk and threat to community, etc.) and sentencing (determined based on seriousness of crime, history of defendant, sentencing guidelines, and argument by counsel). I'm not sure of your knowledge of the justice system so I'll expound just a little bit here, forgive me if it's unnecessary. [NEWLINE] [NEWLINE] In regards to sentencing; in the United States, judges decide criminal sentences based on the criminal sentencing guidelines, which amounts to a series of checks that help them analyze a defendant's role in the commission of a certain crime as well as the defendant's criminal history. Essentially, the prosecutor as well as the defense attorney will "score" the Defendant and compare the numbers they come up with; there is room for argument here, which is why prep for sentencing can be a laborious procedure for an attorney. The "score" that each attorney comes up with will correspond to a table of values, for each of which there is a *high* end and a *low* end. The judge can, based on his/her take on the case and the arguments of counsel, decide where within that bracket to sentence the Defendant. [NEWLINE] [NEWLINE] I wholeheartedly disagree that for a law to be "just" that it must carry with it a universal standard (assuming you mean the sentence/punishment here). The rigidity you suggest is untenable for two reasons: because each and every single court case is unique and has factors both mitigating and aggravating; and, I would posit, this flexibility is one of the core tenets of the advocative justice system (i.e. you choose someone to represent you in a court of law because you trust they will fight like hell for you, knowing that on the other side there is someone doing the same thing to fight like hell against you). The nature of the system necessitates a certain amount of "judgment calls." [NEWLINE] [NEWLINE] As far as stare decisis goes, a good way to look at it, in my opinion, is to view caselaw as the material that fills in the "negative space" around which the judge determines a sentence: they cannot go outside the boundaries of the caselaw the precedes them. So long as they are operating within those boundaries, though, and within
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Polar bears mostly eat seals.  Without polar bears we will see an increase in the population of the seals they eat.  This will result in a decrease in the population of animals they eat, which does partially contain some animals that are common foods for humans.  At this point, the chain of species affected becomes too broad to properly assess the impact,<mask> it is likely that further species will be directly affected and some of which might be species that humans interact with more directly.  We rarely know<mask> species are [keystone species]( [URL] ) before there is an issue with them,<mask> apex predators like the polar bear are more likely to have that status due to their influence on an entire food web. [NEWLINE] [NEWLINE] <mask> will be the end result to you and people like you?  A possible rise in the price of some sea food, with some species disappearing from the menu completely.  That might not seem like much,<mask> depending on<mask> widespread the changes int he ecosystem are, the rise in price could be very significant and many species could be off the menu.  Even<mask> you don't eat sea food, it forms the primary diet for enough people that them having to eat other things could raise the price of other foods.  There is<mask> the potential lose of profits in the fishing industry that can have wide ranging economic repercussions. [NEWLINE] [NEWLINE] All of this ignores the fact that the main reason that polar bears disappearing gets<mask> much attention<mask> it does is that they are a very visual species that is highly affected by global warming.  The effects of global warming are much wider reaching than just polar bears and will impact you in many other ways (rise in sea levels, pollinator die offs resulting in poor crop harvests, etc.). [NEWLINE] [NEWLINE] Edit:  I forgot to mention the fact that sometimes we reach a point<mask> we find that a species hold some key feature for research in another field.  Sometimes it is medical research base on their DNA or hormones, sometimes it is engineering base on a substance they secrete.  We don't know<mask> animals we will need for research until we need it. <mask> an animal goes extinct, then we have no way to research them,<mask><mask> the animal sticks around, then they will be there<mask> we ever find a reason to research them in detail.  You use the Dodo<mask> an example of an animal whose extinction doesn't affect you,<mask> for all we know, the Dodo held the key to the cures for cancer, AIDS, and the common cold,
Label encoding: <s>Polar bears mostly eat seals.  Without polar bears we will see an increase in the population of the seals they eat.  This will result in a decrease in the population of animals they eat, which does partially contain some animals that are common foods for humans.  At this point, the chain of species affected becomes too broad to properly assess the impact, but it is likely that further species will be directly affected and some of which might be species that humans interact with more directly.  We rarely know what species are [keystone species]( [URL] ) before there is an issue with them, but apex predators like the polar bear are more likely to have that status due to their influence on an entire food web. [NEWLINE] [NEWLINE] What will be the end result to you and people like you?  A possible rise in the price of some sea food, with some species disappearing from the menu completely.  That might not seem like much, but depending on how widespread the changes int he ecosystem are, the rise in price could be very significant and many species could be off the menu.  Even if you don't eat sea food, it forms the primary diet for enough people that them having to eat other things could raise the price of other foods.  There is also the potential lose of profits in the fishing industry that can have wide ranging economic repercussions. [NEWLINE] [NEWLINE] All of this ignores the fact that the main reason that polar bears disappearing gets as much attention as it does is that they are a very visual species that is highly affected by global warming.  The effects of global warming are much wider reaching than just polar bears and will impact you in many other ways (rise in sea levels, pollinator die offs resulting in poor crop harvests, etc.). [NEWLINE] [NEWLINE] Edit:  I forgot to mention the fact that sometimes we reach a point where we find that a species hold some key feature for research in another field.  Sometimes it is medical research base on their DNA or hormones, sometimes it is engineering base on a substance they secrete.  We don't know what animals we will need for research until we need it.  If an animal goes extinct, then we have no way to research them, but if the animal sticks around, then they will be there if we ever find a reason to research them in detail.  You use the Dodo as an example of an animal whose extinction doesn't affect you, but for all we know, the Dodo held the key to the cures for cancer, AIDS, and the common cold,
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sure... these are all reasonable questions/points. And we use other mechanisms<mask> well. And our multi-round interview and tests still yield some "bad hires". We do a couple of things about this (beyond generous outplacement,<mask> we assumed the blame for any "bad hire")... and one is to QA our hiring process. We always ask ourselves --<mask> could we have asked to have detected *that* earlier on. We iterated our process probably 4 times/year. It actually got "pretty good" --<mask> there is never going to be 100%<mask> humans are involved. [NEWLINE] [NEWLINE] [STARTQ] Is it really possible to give more than a basic test in an interview setting, [ENDQ] [NEWLINE] It was a realistic,<mask> stretch problem set.<mask>, maybe 12-16 hours of work for an above average person with a 4 hour time-window. Nobody (well, one) finished,<mask> our goal was to see<mask> they went about it all. It helped us filter people who put buzz-words on their resumes vs. being able to hit it hard pretty well. I'm guessing that this may be particular to production coding environments.<mask> I interviewed in semiconductor, optics (physics), o&amp;g refining situations myself -- a few of them had very good/rigorous tests. [NEWLINE] [NEWLINE] Would it detect whether I cheated in college -- probably not.<mask> it would detect whether I was full of shit or not. Sometimes that is all that matters in a production work-place. The cheating/trust stuff usually does emerge in less than the 3 month probation period<mask>. That's not skill -- it's character. [NEWLINE] [NEWLINE] Everything verbal was "cultural fit" -- which comes down to: 1) Do you like them?; 2) Can you learn to trust them to have your back (in your opinion)?; 3) Do (you) think they would fit? We had about 5-6 of those cross-functional verbal ones -- and any "no" or "maybe" was a "no hire". It wasn't democratic from that point of view. [NEWLINE] [NEWLINE] [STARTQ] will you immediately be able to spot the difference between someone who cheated their way through school and someone who didn't? [ENDQ] [NEWLINE] Absolutely not with 100% accuracy. You are right. We had a culture<mask> our employees would "poach" some of their former co-workers. We didn't pay them (much; for long -- it didn't work)<mask> of perverse incentives that get set up. We gave
Label encoding: <s>Sure... these are all reasonable questions/points. And we use other mechanisms as well. And our multi-round interview and tests still yield some "bad hires". We do a couple of things about this (beyond generous outplacement, because we assumed the blame for any "bad hire")... and one is to QA our hiring process. We always ask ourselves -- what could we have asked to have detected *that* earlier on. We iterated our process probably 4 times/year. It actually got "pretty good" -- but there is never going to be 100% where humans are involved. [NEWLINE] [NEWLINE] [STARTQ] Is it really possible to give more than a basic test in an interview setting, [ENDQ] [NEWLINE] It was a realistic, but stretch problem set. So, maybe 12-16 hours of work for an above average person with a 4 hour time-window. Nobody (well, one) finished, but our goal was to see how they went about it all. It helped us filter people who put buzz-words on their resumes vs. being able to hit it hard pretty well. I'm guessing that this may be particular to production coding environments. When I interviewed in semiconductor, optics (physics), o&amp;g refining situations myself -- a few of them had very good/rigorous tests. [NEWLINE] [NEWLINE] Would it detect whether I cheated in college -- probably not. But it would detect whether I was full of shit or not. Sometimes that is all that matters in a production work-place. The cheating/trust stuff usually does emerge in less than the 3 month probation period though. That's not skill -- it's character. [NEWLINE] [NEWLINE] Everything verbal was "cultural fit" -- which comes down to: 1) Do you like them?; 2) Can you learn to trust them to have your back (in your opinion)?; 3) Do (you) think they would fit? We had about 5-6 of those cross-functional verbal ones -- and any "no" or "maybe" was a "no hire". It wasn't democratic from that point of view. [NEWLINE] [NEWLINE] [STARTQ] will you immediately be able to spot the difference between someone who cheated their way through school and someone who didn't? [ENDQ] [NEWLINE] Absolutely not with 100% accuracy. You are right. We had a culture where our employees would "poach" some of their former co-workers. We didn't pay them (much; for long -- it didn't work) because of perverse incentives that get set up. We gave
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me give you a hypothetical situation. [NEWLINE] [NEWLINE] You and the guys are out at a local bar.  It's not too busy and you have no intention of hooking up with anyone - tonight is about getting drunk with the boys.  You're drinking, watching the game, talking, drinking, talking more, drinking more.  You've had too much; you're out of your wits.  You realize you're really drunk,<mask> you've just had another shot and a beer,<mask> this is going to get worse before it gets better.  You stumble outside without thinking about paying the tab (you have paid actually, just five minutes ago,<mask> you're blacking out and can't recall).  Not only are you having trouble focusing your vision, you're having trouble putting one foot in front of the other and gravity is all wrong.  Someone puts their arm around you and helps you walk a few steps.  They're talking and you pick up the word "Home".  Yes, that's the word!  Home.  Home is<mask> you want to go. <mask> after a few more stumbles you fall comfortably into a taxi.  (you have no idea<mask> it got there, you don't even ponder this,  it just feels<mask> good to be sitting down) You pass out for the entire drive.  Your new friend helps you lean far enough forward that you can stand up and get out of the cab.  You stumble across the unfamiliar terrain with a little guidance.  You black out again. <mask> next you wake, it feels like someone's jumping on your bed.  You feel pressure on your legs and your stomach, gotta piss, bladder hurts.  You open your eyes and between the colors and the shadows you’re not sure which wall you’re looking at or even<mask> it’s the ceiling.  It takes a moment to realize this isn’t your room. <mask> your vision goes from bleary to slightly less bleary you notice this is the cause of the rhythmic feeling. (imagine her naked, this is totally SFW - don't worry) [URL].jpg [NEWLINE] [NEWLINE] Now tell me, would you have chosen to sleep with that woman<mask> you were sober?  Even<mask> you agreed to sleep with her in one of those memory gaps, were you in any shape to give consent?  Could you reasonably say that you wouldn't feel violated?  Have you contracted an STD from this person? <mask>
Label encoding: <s>Let me give you a hypothetical situation. [NEWLINE] [NEWLINE] You and the guys are out at a local bar.  It's not too busy and you have no intention of hooking up with anyone - tonight is about getting drunk with the boys.  You're drinking, watching the game, talking, drinking, talking more, drinking more.  You've had too much; you're out of your wits.  You realize you're really drunk, but you've just had another shot and a beer, so this is going to get worse before it gets better.  You stumble outside without thinking about paying the tab (you have paid actually, just five minutes ago, but you're blacking out and can't recall).  Not only are you having trouble focusing your vision, you're having trouble putting one foot in front of the other and gravity is all wrong.  Someone puts their arm around you and helps you walk a few steps.  They're talking and you pick up the word "Home".  Yes, that's the word!  Home.  Home is where you want to go.  So after a few more stumbles you fall comfortably into a taxi.  (you have no idea how it got there, you don't even ponder this,  it just feels so good to be sitting down) You pass out for the entire drive.  Your new friend helps you lean far enough forward that you can stand up and get out of the cab.  You stumble across the unfamiliar terrain with a little guidance.  You black out again.  When next you wake, it feels like someone's jumping on your bed.  You feel pressure on your legs and your stomach, gotta piss, bladder hurts.  You open your eyes and between the colors and the shadows you’re not sure which wall you’re looking at or even if it’s the ceiling.  It takes a moment to realize this isn’t your room.  As your vision goes from bleary to slightly less bleary you notice this is the cause of the rhythmic feeling. (imagine her naked, this is totally SFW - don't worry) [URL].jpg [NEWLINE] [NEWLINE] Now tell me, would you have chosen to sleep with that woman if you were sober?  Even if you agreed to sleep with her in one of those memory gaps, were you in any shape to give consent?  Could you reasonably say that you wouldn't feel violated?  Have you contracted an STD from this person?  What
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I been using Google Chrome<mask><mask><mask> i remember (probably<mask> 2010) and it's been my default browser on all the computers/laptops I've used. Chrome was fast, reliable and most of all it had a great materials design. [NEWLINE] Having said all that, recently Chromes has failed to live up to its name. Many browsers out there have the same<mask> not better look and feel which<mask> hold up in the speed department. To put it simply, other browsers has closed,<mask> not overtaken the Chrome and everything it stood for. [NEWLINE] Saying that you use Chrome means nothing anymore, and might<mask> well use IE (or the incarnation Spartan) [NEWLINE] Edit:<mask> the memory usage by Chrome is crazy. I can't comment on the memory usage by Firefox haven't used it that much [NEWLINE] [NEWLINE] Please change my view [NEWLINE] [NEWLINE] Edit 1: I'm off to bed. I'll be back in the morning (Australian morning) [NEWLINE] Edit 2: I do see now that Chrome is still innovating,<mask> not<mask> much on the aesthetics,<mask> rather helping developers and making better websites.<mask>, after reading most a lot of comments (thank you for that), I see now that Chrome is more than a browser<mask> rather a platform. It offers a wide range of features (Hangouts, Sync, etc) that no other browser can match at the moment. [NEWLINE] Thank you for your replies [NEWLINE] Edit 3: My intentions were never to show that Chrome or Firefox *insert any browser name* are bad and you shouldn't use them. My intentions were to understand<mask> or<mask> not is Chrome overrated by society (especially on the internet). [NEWLINE] Edit 4: RIP my inbox. Thank you, for the replies [NEWLINE] <mask>, I am aware that the word overrated is misspelled [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to*
Label encoding: <s>I been using Google Chrome as far as i remember (probably since 2010) and it's been my default browser on all the computers/laptops I've used. Chrome was fast, reliable and most of all it had a great materials design. [NEWLINE] Having said all that, recently Chromes has failed to live up to its name. Many browsers out there have the same if not better look and feel which also hold up in the speed department. To put it simply, other browsers has closed, if not overtaken the Chrome and everything it stood for. [NEWLINE] Saying that you use Chrome means nothing anymore, and might as well use IE (or the incarnation Spartan) [NEWLINE] Edit: Also the memory usage by Chrome is crazy. I can't comment on the memory usage by Firefox haven't used it that much [NEWLINE] [NEWLINE] Please change my view [NEWLINE] [NEWLINE] Edit 1: I'm off to bed. I'll be back in the morning (Australian morning) [NEWLINE] Edit 2: I do see now that Chrome is still innovating, but not so much on the aesthetics, but rather helping developers and making better websites. Also, after reading most a lot of comments (thank you for that), I see now that Chrome is more than a browser but rather a platform. It offers a wide range of features (Hangouts, Sync, etc) that no other browser can match at the moment. [NEWLINE] Thank you for your replies [NEWLINE] Edit 3: My intentions were never to show that Chrome or Firefox *insert any browser name* are bad and you shouldn't use them. My intentions were to understand why or why not is Chrome overrated by society (especially on the internet). [NEWLINE] Edit 4: RIP my inbox. Thank you, for the replies [NEWLINE] Also, I am aware that the word overrated is misspelled [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to*
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the notion that we should "retire" the term "white privilege"<mask> we don't want to hurt white peoples' feeling is... you guessed it... a great example of white privilege. [ENDQ] [NEWLINE] Using a rhetorical tactic in a discussion about whether we should use that tactic is a bit meta, isn't it? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the reality is that in general, we're<mask> used to living in a world<mask> we don't have to think about our race, that<mask> it is brought up, and there's anything unpleasant about it, there's an urge to say "hey hey... come on. let's not be<mask> negative. can't we steer this conversation in a more constructive direction?" [ENDQ] [NEWLINE] Isn't the ideal situation that everyone is treated in such a way that they don't have to think about their race?<mask> is it a privilege to be treated properly? [NEWLINE] [NEWLINE] I'm not saying "hey hey... come on. let's not be<mask> negative. can't we steer this conversation in a more constructive direction?"<mask> I'm uncomfortable acknowledging that I have privileges other people don't, I'm arguing against casting these things<mask> privileges at all. [NEWLINE] [NEWLINE] Rather than looking at it<mask> "I have privileges<mask> I'm white", look at it<mask> "this other person is being denied basic rights and treatment<mask> he's black." It isn't a privilege to feel<mask><mask> your race doesn't affect your life, that should be<mask> *everyone* experiences. [NEWLINE] [NEWLINE] <mask> we focus instead on<mask> other people don't have it<mask> good<mask> you, rather than<mask> you have it better than other people, we kill several birds with one stone: People who have learned that 'privilege' is often used<mask> a shaming tactic won't argue against that language, people who do use 'privilege'<mask> a shaming tactic won't sully reputations of people having honest conversations about disadvantage, and it explicitly puts focus on giving all people *more privileges* rather than pointing out that cis white males have more than their fair share. [NEWLINE] [NEWLINE] [STARTQ] I'm asking you to basically "grow up" and have a mature conversation about a real thing that happens in our world without us having to worry about your delicate feelings. [ENDQ] [NEWLINE] <mask> you are trying to change public opinion, you need to worry about<mask> people perceive different rhetoric. I don't care<mask> they are *wrong* to feel that way, I care about actually improving the treatment of the underprivileged.<mask> a slight change of language dodges
Label encoding: <s> [STARTQ] the notion that we should "retire" the term "white privilege" because we don't want to hurt white peoples' feeling is... you guessed it... a great example of white privilege. [ENDQ] [NEWLINE] Using a rhetorical tactic in a discussion about whether we should use that tactic is a bit meta, isn't it? [NEWLINE] [NEWLINE] [STARTQ] I think the reality is that in general, we're so used to living in a world where we don't have to think about our race, that when it is brought up, and there's anything unpleasant about it, there's an urge to say "hey hey... come on. let's not be so negative. can't we steer this conversation in a more constructive direction?" [ENDQ] [NEWLINE] Isn't the ideal situation that everyone is treated in such a way that they don't have to think about their race? Why is it a privilege to be treated properly? [NEWLINE] [NEWLINE] I'm not saying "hey hey... come on. let's not be so negative. can't we steer this conversation in a more constructive direction?" because I'm uncomfortable acknowledging that I have privileges other people don't, I'm arguing against casting these things as privileges at all. [NEWLINE] [NEWLINE] Rather than looking at it as "I have privileges because I'm white", look at it as "this other person is being denied basic rights and treatment because he's black." It isn't a privilege to feel as though your race doesn't affect your life, that should be what *everyone* experiences. [NEWLINE] [NEWLINE] If we focus instead on how other people don't have it as good as you, rather than how you have it better than other people, we kill several birds with one stone: People who have learned that 'privilege' is often used as a shaming tactic won't argue against that language, people who do use 'privilege' as a shaming tactic won't sully reputations of people having honest conversations about disadvantage, and it explicitly puts focus on giving all people *more privileges* rather than pointing out that cis white males have more than their fair share. [NEWLINE] [NEWLINE] [STARTQ] I'm asking you to basically "grow up" and have a mature conversation about a real thing that happens in our world without us having to worry about your delicate feelings. [ENDQ] [NEWLINE] If you are trying to change public opinion, you need to worry about how people perceive different rhetoric. I don't care if they are *wrong* to feel that way, I care about actually improving the treatment of the underprivileged. If a slight change of language dodges
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Off-topic can be fun! I recognize that my view will be radically unpopular on such a libertarian site,<mask> I'll give it a shot. To begin, I reject the notion that equality is the goal of just treatment. It is,<mask>, often a side effect. That is, being treated equally with another is not valuable unless there are other relevant extant reasons for doing<mask>. For example, I have friend with perfect vision. He has a Driver's Licence and I have a Driver's Licence with a restriction (I have to wear corrective eyewear). This is a form of discrimination,<mask> it is good<mask> driving without corrective lenses would be dangerous for me.<mask>,<mask> I had perfect eyesight, and I was still given the restriction, it would be unjust<mask> there would be no legitimate grounds for that different treatment. Boiled down,<mask><mask> that<mask> A and B are similar in all relevant ways and are treated justly, then their treatment will be equal.<mask><mask> they are treated justly, and A and B are dissimilar in relevant ways, then they would be treated differently. This reasoning is similar to,<mask> not the same<mask>, that of [Aristotle]( [URL] ), especially looking at section 6 of the Nicomachean Ethics. [NEWLINE] [NEWLINE] <mask> you accept this sort of justice, then gender roles become judicially tricky. On the one hand, you want to avoid keeping people from living their own best life<mask> of over-generalized rules.<mask><mask><mask><mask>, gender roles exist<mask> they often reflect mathematically significant differences between groups of people. Regarding reproduction, the difference between male and female role is significant enough that I believe it justifies legislative difference. The fighting in wars case is much tougher<mask> the difference is much less stark. [NEWLINE] [NEWLINE] I've<mask> been thinking about<mask> the "losing" in society works. The reality is that not everyone can fulfill their heart's desire. Some people enjoy murder, torture, pedophilia, stealing, and<mask> on.<mask><mask> people don't choose their fundamental orientations (cf. the whole homosexual marriage debate), the reality is that those desires that,<mask> fulfilled, cause<mask> we believe to be harm, should at least not be encouraged.<mask> we are encouraging beneficial forms of social interaction and cultivating helpful desires, then those who have the unhelpful traits will<mask> "lose". [NEWLINE] [NEWLINE] In conclusion, I like the intentions behind egalitarianism and the desire for fairness.<mask><mask><mask> it is weak<mask> people aren
Label encoding: <s>Off-topic can be fun! I recognize that my view will be radically unpopular on such a libertarian site, but I'll give it a shot. To begin, I reject the notion that equality is the goal of just treatment. It is, however, often a side effect. That is, being treated equally with another is not valuable unless there are other relevant extant reasons for doing so. For example, I have friend with perfect vision. He has a Driver's Licence and I have a Driver's Licence with a restriction (I have to wear corrective eyewear). This is a form of discrimination, but it is good because driving without corrective lenses would be dangerous for me. However, if I had perfect eyesight, and I was still given the restriction, it would be unjust because there would be no legitimate grounds for that different treatment. Boiled down, I think that if A and B are similar in all relevant ways and are treated justly, then their treatment will be equal. But if they are treated justly, and A and B are dissimilar in relevant ways, then they would be treated differently. This reasoning is similar to, but not the same as, that of [Aristotle]( [URL] ), especially looking at section 6 of the Nicomachean Ethics. [NEWLINE] [NEWLINE] If you accept this sort of justice, then gender roles become judicially tricky. On the one hand, you want to avoid keeping people from living their own best life because of over-generalized rules. On the other hand, gender roles exist because they often reflect mathematically significant differences between groups of people. Regarding reproduction, the difference between male and female role is significant enough that I believe it justifies legislative difference. The fighting in wars case is much tougher since the difference is much less stark. [NEWLINE] [NEWLINE] I've also been thinking about how the "losing" in society works. The reality is that not everyone can fulfill their heart's desire. Some people enjoy murder, torture, pedophilia, stealing, and so on. Even though people don't choose their fundamental orientations (cf. the whole homosexual marriage debate), the reality is that those desires that, when fulfilled, cause what we believe to be harm, should at least not be encouraged. If we are encouraging beneficial forms of social interaction and cultivating helpful desires, then those who have the unhelpful traits will indeed "lose". [NEWLINE] [NEWLINE] In conclusion, I like the intentions behind egalitarianism and the desire for fairness. But I think it is weak because people aren
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, your comment raises a good point, in that now that information is freely available online (legally or not), this question is a lot more academic than it once was. [NEWLINE] [NEWLINE] That said,<mask> worries me more is less the idea of an intended rapist walking in and reading the book:<mask> someone has already decided to commit a crime like that, will the book change their view? [NEWLINE] [NEWLINE] Honestly? I believe that yes, yes it could.<mask> that's a question for another day,<mask> I don't think it will be possible to be proved on a factual basis.<mask>, let's continue using the child abuser metaphor. [NEWLINE] [NEWLINE] Large numbers of pedophiles never act on their desires. A book that explains<mask> to get away with it (and, to extend the metaphor, explains<mask> it's not such a bad thing after all) is absolutely going to raise the likelihood of children being abused,<mask> it will raise the likelihood of a pedophile<mask> non-abuser becoming an abuser. [NEWLINE] [NEWLINE] I'm equally worried about the reaction of mentally ill people to certain stimuli. Oftentimes, people with abnormal thought processes make leaps that the rest of us do not.<mask> they read a book about<mask> and<mask> you should abuse kids, I can see that leading to an increased possibility for abuse. [NEWLINE] [NEWLINE] Ultimately, of course, it's entirely up to you. Do<mask><mask> it will raise the likelihood of children being abused?<mask> I see<mask> people could argue against that, I absolutely think it will, and<mask><mask> anyone who tries to<mask><mask> there's no chance it will is being very disingenuous. [NEWLINE] [NEWLINE] <mask>, whether or not restricting that particular freedom of speech is worth lower chances of violence/pedophilia/mass murder/whatever is a completely personal choice.<mask><mask> in extreme cases such<mask> these,<mask> the clear danger seems to outweigh any minor benefits, it's worth making an exception. [NEWLINE] [NEWLINE] That said, I've had the slippery slope fallacy shoved in my face over and over<mask><mask><mask>, and I am fully aware that many other people may not consider it "worth it." It all depends on<mask> you view/calculate the utility of other people, and the utility of freedom. [NEWLINE] [NEWLINE] TL;DR:<mask><mask> this information being publically available will raise the likelihood of abuse,<mask> it will certainly raise the likelihood of easily persuadable people (mentally ill or otherwise) to commit abuse.<mask>, whether or not the restriction of freedom that would be required
Label encoding: <s>Firstly, your comment raises a good point, in that now that information is freely available online (legally or not), this question is a lot more academic than it once was. [NEWLINE] [NEWLINE] That said, what worries me more is less the idea of an intended rapist walking in and reading the book: if someone has already decided to commit a crime like that, will the book change their view? [NEWLINE] [NEWLINE] Honestly? I believe that yes, yes it could. But that's a question for another day, because I don't think it will be possible to be proved on a factual basis. However, let's continue using the child abuser metaphor. [NEWLINE] [NEWLINE] Large numbers of pedophiles never act on their desires. A book that explains how to get away with it (and, to extend the metaphor, explains why it's not such a bad thing after all) is absolutely going to raise the likelihood of children being abused, as it will raise the likelihood of a pedophile but non-abuser becoming an abuser. [NEWLINE] [NEWLINE] I'm equally worried about the reaction of mentally ill people to certain stimuli. Oftentimes, people with abnormal thought processes make leaps that the rest of us do not. If they read a book about how and why you should abuse kids, I can see that leading to an increased possibility for abuse. [NEWLINE] [NEWLINE] Ultimately, of course, it's entirely up to you. Do i think it will raise the likelihood of children being abused? While I see why people could argue against that, I absolutely think it will, and I think anyone who tries to argue that there's no chance it will is being very disingenuous. [NEWLINE] [NEWLINE] However, whether or not restricting that particular freedom of speech is worth lower chances of violence/pedophilia/mass murder/whatever is a completely personal choice. I think in extreme cases such as these, where the clear danger seems to outweigh any minor benefits, it's worth making an exception. [NEWLINE] [NEWLINE] That said, I've had the slippery slope fallacy shoved in my face over and over because of this, and I am fully aware that many other people may not consider it "worth it." It all depends on how you view/calculate the utility of other people, and the utility of freedom. [NEWLINE] [NEWLINE] TL;DR: I think this information being publically available will raise the likelihood of abuse, but it will certainly raise the likelihood of easily persuadable people (mentally ill or otherwise) to commit abuse. However, whether or not the restriction of freedom that would be required
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My argument is simple.<mask> I tip, the server benefits by the extra cash.<mask> I don't tip, they will at least get minimum wage<mask> required by law. Life might be harder with minimum wage,<mask> people can certainly survive, especially<mask> they make good financial decisions like you're supposed to. [NEWLINE] [NEWLINE] Serving food to customers is part of the server's job description and I'm not required to tip other minimum wage workers following their job descriptions for providing their time and service. An example would be non-commissioned sales associates in stores. I used to work at Best Buy for $8/hour (not exactly minimum wage<mask> pretty close to it) and I personally know<mask> much time we can spend per customer. Most interactions usually last about 5min or less<mask> it's not that uncommon to be with a customer for up to a hour (or more in rare cases). Even then, sales aren't guaranteed unlike a restaurant server. [NEWLINE] [NEWLINE] Over time, price of meals might go up<mask> I<mask> should I care about that? Ultimately there's not a significant difference<mask> a meal is cheap and I tip 20% vs a more expensive meal<mask> I don't tip? You might<mask><mask><mask> employers might fire servers who don't get enough tips<mask> a sign of bad service<mask> I don't think that should be the case.<mask> a server is providing excellent service, I'm more likely to spend more like with dessert or drinks.<mask> I'm getting bad service, I just want to leave<mask> soon<mask> possible resulting in less sale. Even<mask> this system of evaluating servers doesn't work for whatever reason, employers could always evaluate based on customer surveys or by sending out undercover people to eat at their restaurant. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***.
Label encoding: <s>My argument is simple. If I tip, the server benefits by the extra cash. If I don't tip, they will at least get minimum wage as required by law. Life might be harder with minimum wage, but people can certainly survive, especially if they make good financial decisions like you're supposed to. [NEWLINE] [NEWLINE] Serving food to customers is part of the server's job description and I'm not required to tip other minimum wage workers following their job descriptions for providing their time and service. An example would be non-commissioned sales associates in stores. I used to work at Best Buy for $8/hour (not exactly minimum wage but pretty close to it) and I personally know how much time we can spend per customer. Most interactions usually last about 5min or less but it's not that uncommon to be with a customer for up to a hour (or more in rare cases). Even then, sales aren't guaranteed unlike a restaurant server. [NEWLINE] [NEWLINE] Over time, price of meals might go up but I why should I care about that? Ultimately there's not a significant difference if a meal is cheap and I tip 20% vs a more expensive meal but I don't tip? You might also argue that employers might fire servers who don't get enough tips as a sign of bad service but I don't think that should be the case. If a server is providing excellent service, I'm more likely to spend more like with dessert or drinks. If I'm getting bad service, I just want to leave as soon as possible resulting in less sale. Even if this system of evaluating servers doesn't work for whatever reason, employers could always evaluate based on customer surveys or by sending out undercover people to eat at their restaurant. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***.
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0678, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You can't have it both ways [ENDQ] [NEWLINE] I do. I eat plants and animals. [NEWLINE] [NEWLINE] [STARTQ] <mask> by choosing to eat an animal, you choose to monetarily support the way it was raised and slaughtered. [ENDQ] [NEWLINE] I do. [NEWLINE] [NEWLINE] [STARTQ] Just<mask> you didn't kill it yourself doesn't make it<mask>. [ENDQ] [NEWLINE] I do not see it<mask> immoral to kill and eat meat<mask><mask> would I use this argument you made up on my behalf. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] That's like buying a shirt made by children and saying that you have no responsibility for child labor. "All I did was buy it!" [ENDQ] [NEWLINE] I support child labor in developing countries.<mask> should they be kept in poverty<mask> of my cultural concerns about child labor. There is nothing immoral about a child working -<mask> you close the factory the same child will have to work just hard for just<mask> long on a farm. The stability and pay of factory work might allow this working child's children to go to school instead of toiling away on the farm. [NEWLINE] [NEWLINE] [STARTQ] The difference is, by eating an animal, I'm choosing the suffering of ten times<mask> many plants, plus the suffering of an animal. [ENDQ] [NEWLINE] and [NEWLINE] [NEWLINE] [STARTQ] I'll admit it: I have no absolutely no idea<mask> you mean by this. Would you mind explaining this? [ENDQ] [NEWLINE] [NEWLINE] No you are not. The animal would have eaten the plants<mask><mask> your decision. By eating the animal you are not participating in the death or the potential suffering of the plants. Animals are more efficient. [NEWLINE] [NEWLINE] [STARTQ] My standard of suffering is more or less this: is there any physiological evidence of physical pain or emotional suffering? For animals, the answer is overwhelmingly yes. They react to the things that cause pain to humans in pretty much the same way: vocalizing, rapid breathing, spikes in adrenaline and blood pressure, dilated pupils, attempting to avoid the source of the pain. Plants do none of this. [ENDQ] [NEWLINE] [NEWLINE] This is exactly my point. Your choice is entirely emotional and simply chooses the least anthropomorphical option. It is not a moral decision. [NEWLINE] [NEWLINE] [STARTQ] I might be misunderstanding entirely,<mask> your argument seems to be "<mask> you're not entirely sure, you'd better err on the side of deliberately causing more destruction." [ENDQ] [NEWLINE] My point is that your choice is valid<mask> it is not a moral one. The moral mandate for humans is that you will perish without taking other life. The choice is kill or perish<mask> the choice to kill cannot be immoral
Label encoding: <s> [STARTQ] You can't have it both ways [ENDQ] [NEWLINE] I do. I eat plants and animals. [NEWLINE] [NEWLINE] [STARTQ] But by choosing to eat an animal, you choose to monetarily support the way it was raised and slaughtered. [ENDQ] [NEWLINE] I do. [NEWLINE] [NEWLINE] [STARTQ] Just because you didn't kill it yourself doesn't make it so. [ENDQ] [NEWLINE] I do not see it as immoral to kill and eat meat so why would I use this argument you made up on my behalf. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] That's like buying a shirt made by children and saying that you have no responsibility for child labor. "All I did was buy it!" [ENDQ] [NEWLINE] I support child labor in developing countries. Why should they be kept in poverty because of my cultural concerns about child labor. There is nothing immoral about a child working - if you close the factory the same child will have to work just hard for just as long on a farm. The stability and pay of factory work might allow this working child's children to go to school instead of toiling away on the farm. [NEWLINE] [NEWLINE] [STARTQ] The difference is, by eating an animal, I'm choosing the suffering of ten times as many plants, plus the suffering of an animal. [ENDQ] [NEWLINE] and [NEWLINE] [NEWLINE] [STARTQ] I'll admit it: I have no absolutely no idea what you mean by this. Would you mind explaining this? [ENDQ] [NEWLINE] [NEWLINE] No you are not. The animal would have eaten the plants regardless of your decision. By eating the animal you are not participating in the death or the potential suffering of the plants. Animals are more efficient. [NEWLINE] [NEWLINE] [STARTQ] My standard of suffering is more or less this: is there any physiological evidence of physical pain or emotional suffering? For animals, the answer is overwhelmingly yes. They react to the things that cause pain to humans in pretty much the same way: vocalizing, rapid breathing, spikes in adrenaline and blood pressure, dilated pupils, attempting to avoid the source of the pain. Plants do none of this. [ENDQ] [NEWLINE] [NEWLINE] This is exactly my point. Your choice is entirely emotional and simply chooses the least anthropomorphical option. It is not a moral decision. [NEWLINE] [NEWLINE] [STARTQ] I might be misunderstanding entirely, but your argument seems to be " if you're not entirely sure, you'd better err on the side of deliberately causing more destruction." [ENDQ] [NEWLINE] My point is that your choice is valid but it is not a moral one. The moral mandate for humans is that you will perish without taking other life. The choice is kill or perish therefore the choice to kill cannot be immoral
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just for clarification, do you mean it's destined to fail forever? Or simply that it won't catch on in the near future? [NEWLINE] [NEWLINE] Most of your technical qualms are things that increased power, engineering and miniaturization will take on in the near future.<mask> hardware designers think it's a priority, then we aren't too far away from wireless VR goggles that aren't much bigger than regular sunglasses,<mask> bulky isn't an issue. [NEWLINE] [NEWLINE] <mask> for control, we're already going down the path with kinect that you can use your body<mask> a control. For FPS, hold the gun and pull the trigger.<mask>'s better than that?<mask> for expensive and confining. The cost will go down<mask> it becomes popular. The same thing that happens with every broad entertainment technology. And gaming is already a confined activity. People don't mind sitting on their couches or at desks to play games. It's wildly popular. [NEWLINE] [NEWLINE] That,<mask><mask><mask> I can see, takes care of #1 and #5. Maybe not tomorrow or next year,<mask> within a decade<mask> people put research into it, almost certainly. [NEWLINE] [NEWLINE] <mask> for #2, you could say the same thing about any number of wildly popular technologies. Some grandparents love their smartphones and facebook, some have no idea<mask> they do. I don't see VR<mask> meaningfully different from a lot of technologies that have thrived in that regard. [NEWLINE] [NEWLINE] <mask> for #3 and #4,<mask><mask>, concerts and chat won't be huge markets for VR. I don't think this kills the technology, it's focus is more likely to be games and other kinds of entertainment. [NEWLINE] [NEWLINE] For #6 I thought the same thing about 3d movies,<mask> they seem to be here to stay. The novelty may wear off,<mask> the pleasure of the experience isn't fully reliant on novelty. Adding quality 3d vision to games gives players a better appreciation for the physical space their avatar is occupying, alowing depth and distance to become more of a factor. It<mask> makes environments more immersive.  Snow swirls around you etc etc. The 3ds was a novelty<mask> it used 3d mostly<mask> a novelty. A more powerful system can use it to real effect. [NEWLINE] [NEWLINE] I may share your view that technology that's rolling out right now isn't going to make a huge market impact,<mask> I feel fairly certain that we'll see another push in the fairly near future<mask> technological advancement has ironed the kinks that make VR
Label encoding: <s>Just for clarification, do you mean it's destined to fail forever? Or simply that it won't catch on in the near future? [NEWLINE] [NEWLINE] Most of your technical qualms are things that increased power, engineering and miniaturization will take on in the near future. If hardware designers think it's a priority, then we aren't too far away from wireless VR goggles that aren't much bigger than regular sunglasses, so bulky isn't an issue. [NEWLINE] [NEWLINE] As for control, we're already going down the path with kinect that you can use your body as a control. For FPS, hold the gun and pull the trigger. What's better than that? As for expensive and confining. The cost will go down when it becomes popular. The same thing that happens with every broad entertainment technology. And gaming is already a confined activity. People don't mind sitting on their couches or at desks to play games. It's wildly popular. [NEWLINE] [NEWLINE] That, as far as I can see, takes care of #1 and #5. Maybe not tomorrow or next year, but within a decade if people put research into it, almost certainly. [NEWLINE] [NEWLINE] As for #2, you could say the same thing about any number of wildly popular technologies. Some grandparents love their smartphones and facebook, some have no idea what they do. I don't see VR as meaningfully different from a lot of technologies that have thrived in that regard. [NEWLINE] [NEWLINE] As for #3 and #4, I agree, concerts and chat won't be huge markets for VR. I don't think this kills the technology, it's focus is more likely to be games and other kinds of entertainment. [NEWLINE] [NEWLINE] For #6 I thought the same thing about 3d movies, but they seem to be here to stay. The novelty may wear off, but the pleasure of the experience isn't fully reliant on novelty. Adding quality 3d vision to games gives players a better appreciation for the physical space their avatar is occupying, alowing depth and distance to become more of a factor. It also makes environments more immersive.  Snow swirls around you etc etc. The 3ds was a novelty because it used 3d mostly as a novelty. A more powerful system can use it to real effect. [NEWLINE] [NEWLINE] I may share your view that technology that's rolling out right now isn't going to make a huge market impact, but I feel fairly certain that we'll see another push in the fairly near future when technological advancement has ironed the kinks that make VR
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>On point 3: [NEWLINE] [NEWLINE] The structure of the hunger games was in part based on the television reality show "Survivor" [NEWLINE] [NEWLINE] I *have* spent quite a lot of time studying game theory, and from the Russian revolution to the fall of the USSR - from the reign of terror in France to the reconquista in Spain - it is completely, absolutely normal for alliance to spring up among people who have every intention of betraying one another later, and all of whom are aware that all the others plan to betray them. [NEWLINE] [NEWLINE] None of that matters.  You ally with the most powerful group,<mask> ifyou ally with the weaker group you lose right away.  Later,<mask> you're winning, that's<mask> you have to worry about the guys on your own team. [NEWLINE] [NEWLINE] In the TV show survivor this scnario plays out in nearly every season.  The two separate tribes, who have a degree of shared history, get tossed in together (in this case, inner and outer districts, who share two separate, common cultures). [NEWLINE] [NEWLINE] The tribe with the advantage (the larger survivor tribe, or the inner districts), takes out a few members of the other tribe first,<mask> it being easy pickings.  Once they have an unassailable advantage, they turn on themselves and take out on insider.  Now the weaker team is closer to catching up,<mask> they take one of them out. [NEWLINE] [NEWLINE] <mask>, with two teams (dominant) and (weak) the kill order looks something like this: [NEWLINE] [NEWLINE] Cast: 1, 2, 3, 4, 5, 6 (all dominant), 7, 8, 9, 10, 11 (submissive) [NEWLINE] [NEWLINE] You can expect them to lose/die in this order: [NEWLINE] [NEWLINE] * 7S [NEWLINE] * 8S [NEWLINE] * 1D [NEWLINE] * 9S [NEWLINE] * 2D [NEWLINE] * 3D [NEWLINE] * 10S [NEWLINE] * 4D [NEWLINE] [NEWLINE] final showdown between 5D and 6D and 11S! [NEWLINE] [NEWLINE] This way the dominant group remains dominant.  Anyone foolish enough to exterminate the submissive group entirely will lose access to their only viable ally in the end game,<mask> the strong men turn on one another.  A submissive ally is important<mask>,<mask> you get to the final 2, you can be assured of beating them.  They'll agree<mask>, without your help they're dead either way,<mask> a single hail-mary at the end is better odds than certain death against a team of superior enemies.
Label encoding: <s>On point 3: [NEWLINE] [NEWLINE] The structure of the hunger games was in part based on the television reality show "Survivor" [NEWLINE] [NEWLINE] I *have* spent quite a lot of time studying game theory, and from the Russian revolution to the fall of the USSR - from the reign of terror in France to the reconquista in Spain - it is completely, absolutely normal for alliance to spring up among people who have every intention of betraying one another later, and all of whom are aware that all the others plan to betray them. [NEWLINE] [NEWLINE] None of that matters.  You ally with the most powerful group, because ifyou ally with the weaker group you lose right away.  Later, when you're winning, that's when you have to worry about the guys on your own team. [NEWLINE] [NEWLINE] In the TV show survivor this scnario plays out in nearly every season.  The two separate tribes, who have a degree of shared history, get tossed in together (in this case, inner and outer districts, who share two separate, common cultures). [NEWLINE] [NEWLINE] The tribe with the advantage (the larger survivor tribe, or the inner districts), takes out a few members of the other tribe first, despite it being easy pickings.  Once they have an unassailable advantage, they turn on themselves and take out on insider.  Now the weaker team is closer to catching up, so they take one of them out. [NEWLINE] [NEWLINE] So, with two teams (dominant) and (weak) the kill order looks something like this: [NEWLINE] [NEWLINE] Cast: 1, 2, 3, 4, 5, 6 (all dominant), 7, 8, 9, 10, 11 (submissive) [NEWLINE] [NEWLINE] You can expect them to lose/die in this order: [NEWLINE] [NEWLINE] * 7S [NEWLINE] * 8S [NEWLINE] * 1D [NEWLINE] * 9S [NEWLINE] * 2D [NEWLINE] * 3D [NEWLINE] * 10S [NEWLINE] * 4D [NEWLINE] [NEWLINE] final showdown between 5D and 6D and 11S! [NEWLINE] [NEWLINE] This way the dominant group remains dominant.  Anyone foolish enough to exterminate the submissive group entirely will lose access to their only viable ally in the end game, when the strong men turn on one another.  A submissive ally is important because, when you get to the final 2, you can be assured of beating them.  They'll agree because, without your help they're dead either way, so a single hail-mary at the end is better odds than certain death against a team of superior enemies.
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know I wasn't very specific about the scenario in my original post,<mask> this is my first time posting in this subreddit. [NEWLINE] [NEWLINE] I believe that it becomes a two way street once the child is able to competently understand<mask> to have arguments. The age most certainly would not be identical in every case,<mask> there are very astute 8-year-olds and very stupid 18-year-olds. [NEWLINE] [NEWLINE] <mask> the discussion cannot happen<mask>, it is a lack of logical development in the child. Then it would make sense for the '<mask> I told you<mask>'to be in place,<mask> they can't operate<mask> rational agents. [NEWLINE] [NEWLINE] I do agree with you in expecting for parents to keep their kids in line.<mask>,<mask><mask> some areas of behavior are more black and white than others. For example,<mask> their child punches other kids for asking to share their toy, that is a problem.<mask><mask> a child was constantly bullying another child, then the one on the receiving end decided to finally hit back,<mask><mask> some parents could misconstrue that<mask>'misbehaving.' [NEWLINE] [NEWLINE] <mask><mask> your police officer example is very helpful to my own reflection.<mask><mask><mask> that the relationship between police officers and civilians isn't identical to parents and children. Of course the civilians need to listen to the police officer in that instance,<mask> there are officers who take the law into their own hands, and<mask><mask> that the civilian should not be penalized for speaking up<mask> that is the case. [NEWLINE] [NEWLINE] <mask><mask> with you to a degree on the employee example. That relationship seems to be<mask> things should work in the workplace.<mask>, I don't think it can quite be applied to parenting. The employee is being paid for this deference to the employer.<mask> at any point they get sick of the employer's way of running things, they can leave the company. Of course one could say that the 'payment' children receive is food, shelter, etc.<mask> parents are required by law to provide these things,<mask><mask> the nature of disagreement.<mask> the parent provides such things to the best of their ability and has a sound moral character, then<mask><mask> they deserve the respect of the child.<mask> the parent is doing the bare minimum of interaction, providing the bare minimum of basic [NEWLINE] human necessities, and operates the household without the possibility of ever making a mistake, then there is a problem, and I wouldn't blame the child for not respecting them once they are older to think for themselves.</s>
Label encoding: <s>I know I wasn't very specific about the scenario in my original post, but this is my first time posting in this subreddit. [NEWLINE] [NEWLINE] I believe that it becomes a two way street once the child is able to competently understand how to have arguments. The age most certainly would not be identical in every case, as there are very astute 8-year-olds and very stupid 18-year-olds. [NEWLINE] [NEWLINE] If the discussion cannot happen yet, it is a lack of logical development in the child. Then it would make sense for the'because I told you so'to be in place, since they can't operate as rational agents. [NEWLINE] [NEWLINE] I do agree with you in expecting for parents to keep their kids in line. However, I think some areas of behavior are more black and white than others. For example, if their child punches other kids for asking to share their toy, that is a problem. But if a child was constantly bullying another child, then the one on the receiving end decided to finally hit back, I think some parents could misconstrue that as'misbehaving.' [NEWLINE] [NEWLINE] I think your police officer example is very helpful to my own reflection. But I think that the relationship between police officers and civilians isn't identical to parents and children. Of course the civilians need to listen to the police officer in that instance, but there are officers who take the law into their own hands, and I think that the civilian should not be penalized for speaking up if that is the case. [NEWLINE] [NEWLINE] I agree with you to a degree on the employee example. That relationship seems to be how things should work in the workplace. However, I don't think it can quite be applied to parenting. The employee is being paid for this deference to the employer. If at any point they get sick of the employer's way of running things, they can leave the company. Of course one could say that the 'payment' children receive is food, shelter, etc. but parents are required by law to provide these things, regardless of the nature of disagreement. If the parent provides such things to the best of their ability and has a sound moral character, then I think they deserve the respect of the child. If the parent is doing the bare minimum of interaction, providing the bare minimum of basic [NEWLINE] human necessities, and operates the household without the possibility of ever making a mistake, then there is a problem, and I wouldn't blame the child for not respecting them once they are older to think for themselves.</s>
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] &gt; they're an ideal situation that we're trying to have.<mask> they can't be 'natural', 'inalienable' or 'god-given'. [ENDQ] [NEWLINE] [STARTQ] Philosophy jargon failure. [ENDQ] [NEWLINE] You're<mask> kind. [NEWLINE] [NEWLINE] [STARTQ] That something is a natural right does not mean that it is inviolable. It means that it is inherent in our core being. "Nature" here is a term of art (that varies in meaning somewhat with the  particular schools of philosophy of the speaker" and basically references some idealized abstract notion of a perfect situation. [ENDQ] [NEWLINE] <mask><mask> entirely. First, *nature* does not at all carry the meaning of an idealized perfect situation (especially in this context). It rather means the the state of affairs unaffected by humans. People can and often do strive for a truly ideal situation that can't at all be called *natural*. [NEWLINE] [NEWLINE] [STARTQ] For example, saying that it is human nature to be rational beings in no way precludes the possibility that some individual human beings, or even all individual human beings are at times and places exercising reason improperly. Rather, it is to say that an ideal human, who has maximized their potential and not failed to live up to the ideal of humanity is a rational being. And that this is something we have a reason to aspire towards<mask> it is in some way our natural (again using the word philosophically) inclination. [ENDQ] [NEWLINE] This would follow is the concept of 'perfect rationality<mask> human nature' was<mask><mask> an axiom in reality,<mask> instead it's ridiculously optimistic and remains unsupported in the face of plenty of evidence to the contrary. A human nature to be rational is the basis of several thought experiments related to pure practical reason (exe: the categorical imperative) that means nothing<mask> removed from it's vacuum. It's like the physics joke about assuming cows are spheres; it is useful for conceptual deduction, <mask> it has no basis in reality, and should not be applied<mask> such. [NEWLINE] [NEWLINE] It is important to realize that rights (similar to spherical cows) are useful in determining<mask> *should* be done (<mask> in practice it is an incredibly flawed notion perpetuated by irrational humans in the legislative system and by extension everyone),<mask> they do not actually exist in that they have no enforcement other than the proxy enforcement provided by law. [NEWLINE] [NEWLINE] <mask><mask> : Natural rights are the product of thought experiments using hypothetical axioms. Legal rights are the product of natural rights plus bias. </s><pad>
Label encoding: <s> [STARTQ] &gt; they're an ideal situation that we're trying to have. So they can't be 'natural', 'inalienable' or 'god-given'. [ENDQ] [NEWLINE] [STARTQ] Philosophy jargon failure. [ENDQ] [NEWLINE] You're so kind. [NEWLINE] [NEWLINE] [STARTQ] That something is a natural right does not mean that it is inviolable. It means that it is inherent in our core being. "Nature" here is a term of art (that varies in meaning somewhat with the  particular schools of philosophy of the speaker" and basically references some idealized abstract notion of a perfect situation. [ENDQ] [NEWLINE] I disagree entirely. First, *nature* does not at all carry the meaning of an idealized perfect situation (especially in this context). It rather means the the state of affairs unaffected by humans. People can and often do strive for a truly ideal situation that can't at all be called *natural*. [NEWLINE] [NEWLINE] [STARTQ] For example, saying that it is human nature to be rational beings in no way precludes the possibility that some individual human beings, or even all individual human beings are at times and places exercising reason improperly. Rather, it is to say that an ideal human, who has maximized their potential and not failed to live up to the ideal of humanity is a rational being. And that this is something we have a reason to aspire towards because it is in some way our natural (again using the word philosophically) inclination. [ENDQ] [NEWLINE] This would follow is the concept of 'perfect rationality as human nature' was in fact an axiom in reality, but instead it's ridiculously optimistic and remains unsupported in the face of plenty of evidence to the contrary. A human nature to be rational is the basis of several thought experiments related to pure practical reason (exe: the categorical imperative) that means nothing when removed from it's vacuum. It's like the physics joke about assuming cows are spheres; it is useful for conceptual deduction,  but it has no basis in reality, and should not be applied as such. [NEWLINE] [NEWLINE] It is important to realize that rights (similar to spherical cows) are useful in determining what *should* be done ( although in practice it is an incredibly flawed notion perpetuated by irrational humans in the legislative system and by extension everyone), but they do not actually exist in that they have no enforcement other than the proxy enforcement provided by law. [NEWLINE] [NEWLINE] TLDR : Natural rights are the product of thought experiments using hypothetical axioms. Legal rights are the product of natural rights plus bias. </s><pad>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me start off by stating<mask> I believe the reddit basic income theory is. From<mask> I have read from /r/basicIncome, the idea is that in a future utopian society computers will take all of our jobs and there will be a huge number of people out of work. Basic income would be a government-issued wage that would act<mask> a safety net for these newly underemployed/unemployed workers. [NEWLINE] [NEWLINE] I am not making this post to argue against future automation problems getting rid of jobs. Yes, jobs will require more skills and basic menial tasks may soon be replaced.<mask> that is frictional unemployment, not structural. Its incorrect to say that<mask> a McDonalds cashier got laid off<mask> of a touchscreen order screen that said McDonald's worker will never work again. [NEWLINE] [NEWLINE] Many people on reddit would like to implement Basic Income today. They state that it would rid our problems with a complicated tax code, food stamps and other welfare programs. The huge issue is<mask> you have a dead wage safety net you create a massive systematic disincentive to people to work.<mask> people make a certain wage<mask> working zero hours a week,<mask> go get a job at the grocery store? The marginal value of a full time job becomes diluted. [NEWLINE] [NEWLINE] Next off, wages are considered elastic. Wages are often sticky,<mask> they will still react to the market, especially new hires.<mask> everyone is getting basic income, employers would simply cut costs and pay people less in wages<mask><mask> not? Employees would be indifferent<mask> they aren't netting less money,<mask> basic income would make up the difference. Competitive wages would suffer. [NEWLINE] [NEWLINE] CMV [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Let me start off by stating what I believe the reddit basic income theory is. From what I have read from /r/basicIncome, the idea is that in a future utopian society computers will take all of our jobs and there will be a huge number of people out of work. Basic income would be a government-issued wage that would act as a safety net for these newly underemployed/unemployed workers. [NEWLINE] [NEWLINE] I am not making this post to argue against future automation problems getting rid of jobs. Yes, jobs will require more skills and basic menial tasks may soon be replaced. But that is frictional unemployment, not structural. Its incorrect to say that because a McDonalds cashier got laid off because of a touchscreen order screen that said McDonald's worker will never work again. [NEWLINE] [NEWLINE] Many people on reddit would like to implement Basic Income today. They state that it would rid our problems with a complicated tax code, food stamps and other welfare programs. The huge issue is when you have a dead wage safety net you create a massive systematic disincentive to people to work. If people make a certain wage while working zero hours a week, why go get a job at the grocery store? The marginal value of a full time job becomes diluted. [NEWLINE] [NEWLINE] Next off, wages are considered elastic. Wages are often sticky, but they will still react to the market, especially new hires. If everyone is getting basic income, employers would simply cut costs and pay people less in wages because why not? Employees would be indifferent because they aren't netting less money, since basic income would make up the difference. Competitive wages would suffer. [NEWLINE] [NEWLINE] CMV [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't understand your argument concerning the utility function. In the thread you linked, Amarkov specifically made four points, and you only answered one (the disparity between the intent and the actual effects of one's action). It remains that: [NEWLINE] [NEWLINE] * you didn't tell<mask> would be an utility or happiness function.<mask> do you measure and quantify it?<mask> you can't put a number on it, then the whole concept of "maximizing utility" is meaningless. I'm not even sure anybody has a coherent utility function; the human irrationality being<mask> it is, I could very well prefer A to B, B to C and C to A (replace A, B and C with foods, houses, cars, romantic relations...). NB: you must quantify utility, and not only rank preferences, or later you run into the [voting paradox]( [URL] ) and [Arrow's impossibility theorem]( [URL] ). [NEWLINE] [NEWLINE] * let us assume that anybody has an utility function. The idea that your actions must "maximize the total happiness" implies that you must know<mask> other people desire<mask> well<mask> them. That may work only on a very basic level, for needs (people want food, shelter, no pain...),<mask> I don't see it going much further. [NEWLINE] [NEWLINE] * let us assume that everyone has an utility function which is public. Then you run into the problem of comparing different people's utilities (see h1ppophagist's post), which was unduly glossed over by Amarkov. [NEWLINE] [NEWLINE] * OK, let us assume that the three issues above are solved. Then you run into the problem that there is no good order in dimension larger than 2. That is,<mask> you have an action which improves everyone's happiness, then of course it is morally good.<mask> I don't think that any other moral system would have given you a different answer. And<mask> about actions which improve one person's happiness at the expense of another? You believe that the good thing to maximize is the "total utility".<mask>?<mask> not the minimum utility (" society is only<mask> good<mask> its most destitute people"), or the maximum utility, or the geometric mean of the utility, or any other kind of weighted average?<mask> would the total be *special*? [NEWLINE] [NEWLINE] <mask> these question can't be answered, then utilitarianism has to be rejected not<mask> it is, in some naive sense, good or bad,<mask><mask> it just does not make sense.</s>
Label encoding: <s>I don't understand your argument concerning the utility function. In the thread you linked, Amarkov specifically made four points, and you only answered one (the disparity between the intent and the actual effects of one's action). It remains that: [NEWLINE] [NEWLINE] * you didn't tell what would be an utility or happiness function. How do you measure and quantify it? If you can't put a number on it, then the whole concept of "maximizing utility" is meaningless. I'm not even sure anybody has a coherent utility function; the human irrationality being as it is, I could very well prefer A to B, B to C and C to A (replace A, B and C with foods, houses, cars, romantic relations...). NB: you must quantify utility, and not only rank preferences, or later you run into the [voting paradox]( [URL] ) and [Arrow's impossibility theorem]( [URL] ). [NEWLINE] [NEWLINE] * let us assume that anybody has an utility function. The idea that your actions must "maximize the total happiness" implies that you must know what other people desire as well as them. That may work only on a very basic level, for needs (people want food, shelter, no pain...), but I don't see it going much further. [NEWLINE] [NEWLINE] * let us assume that everyone has an utility function which is public. Then you run into the problem of comparing different people's utilities (see h1ppophagist's post), which was unduly glossed over by Amarkov. [NEWLINE] [NEWLINE] * OK, let us assume that the three issues above are solved. Then you run into the problem that there is no good order in dimension larger than 2. That is, if you have an action which improves everyone's happiness, then of course it is morally good. But I don't think that any other moral system would have given you a different answer. And what about actions which improve one person's happiness at the expense of another? You believe that the good thing to maximize is the "total utility". Why? Why not the minimum utility (" society is only as good as its most destitute people"), or the maximum utility, or the geometric mean of the utility, or any other kind of weighted average? Why would the total be *special*? [NEWLINE] [NEWLINE] If these question can't be answered, then utilitarianism has to be rejected not because it is, in some naive sense, good or bad, but because it just does not make sense.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are massive differences between teaching philosophy, and teaching critical thinking.  All the benefits you discuss stem from critical thinking. [NEWLINE] [NEWLINE] It is easy to imagine a Philosophy class that doesn't help students improve their critical thinking skills.  Most class discussions I've ever seen involve perhaps 4-6 serious participants in a class of 30 people, with the majority listening,<mask> not contributing. <mask> an essay is written later on the topic, those 30 people will primarily write rehashed versions of<mask> was discussed in the debate,<mask> they are primarily remembering, rewording, and reviewing<mask> was discussed, without necessarily critically examining it, and independently coming up with new insights. [NEWLINE] [NEWLINE] There has been substantial amounts of research over the last 50 years on improving critical thinking, and most of the methods that have been tried have been surprisingly ineffective.  The most promising method I've seen is the one adopted in a 1st year critical thinking course at the university of Melbourne, known<mask> [LAMP \(Lots of Argument Mapping\).]( [URL] /) [NEWLINE] [NEWLINE] This [paper]( [URL] ;pid=sites&amp;srcid=ZGVmYXVsdGRvbWFpbnx0aW12YW5nZWxkZXJ8Z3g6NmY2MjY3MjE4YjI2MzNmOQ) analyzed the effectiveness of philosophy courses on teaching critical thinking and came to the conclusion: "The meta-analysis results indicate that students do improve<mask> studying philosophy, and apparently more<mask> than general university students,<mask> we cannot be very confidant that this difference is not just the result of random variation.  More importantly, studying philosophy is less effective than studying critical thinking,<mask><mask> whether one is being taught in a philosophy department or some other department.  finally, studying philosophy is much less effective than studying critical thinking using techniques known to be particularly effective such<mask> LAMP." [NEWLINE] [NEWLINE] <mask> you skip forward to page 25 of that study, there's a chart showing the effect sizes of various critical thinking interventions.  Notable is that the [LAMP method]( [URL] /), focused purely on critical thinking skills, and almost completely devoid of anything resembling traditional philosophy, is twice<mask> effective<mask> the non-argument mapping philosophy courses, and more effective then the argument mapping + philosophy courses. [NEWLINE] [NEWLINE] <mask> you want to teach critical thinking<mask> a core subject, then ditch philosophy and instead integrate argument mapping into regular subjects.</s>
Label encoding: <s>There are massive differences between teaching philosophy, and teaching critical thinking.  All the benefits you discuss stem from critical thinking. [NEWLINE] [NEWLINE] It is easy to imagine a Philosophy class that doesn't help students improve their critical thinking skills.  Most class discussions I've ever seen involve perhaps 4-6 serious participants in a class of 30 people, with the majority listening, but not contributing.  When an essay is written later on the topic, those 30 people will primarily write rehashed versions of what was discussed in the debate, but they are primarily remembering, rewording, and reviewing what was discussed, without necessarily critically examining it, and independently coming up with new insights. [NEWLINE] [NEWLINE] There has been substantial amounts of research over the last 50 years on improving critical thinking, and most of the methods that have been tried have been surprisingly ineffective.  The most promising method I've seen is the one adopted in a 1st year critical thinking course at the university of Melbourne, known as [LAMP \(Lots of Argument Mapping\).]( [URL] /) [NEWLINE] [NEWLINE] This [paper]( [URL] ;pid=sites&amp;srcid=ZGVmYXVsdGRvbWFpbnx0aW12YW5nZWxkZXJ8Z3g6NmY2MjY3MjE4YjI2MzNmOQ) analyzed the effectiveness of philosophy courses on teaching critical thinking and came to the conclusion: "The meta-analysis results indicate that students do improve while studying philosophy, and apparently more so than general university students, though we cannot be very confidant that this difference is not just the result of random variation.  More importantly, studying philosophy is less effective than studying critical thinking, regardless of whether one is being taught in a philosophy department or some other department.  finally, studying philosophy is much less effective than studying critical thinking using techniques known to be particularly effective such as LAMP." [NEWLINE] [NEWLINE] If you skip forward to page 25 of that study, there's a chart showing the effect sizes of various critical thinking interventions.  Notable is that the [LAMP method]( [URL] /), focused purely on critical thinking skills, and almost completely devoid of anything resembling traditional philosophy, is twice as effective as the non-argument mapping philosophy courses, and more effective then the argument mapping + philosophy courses. [NEWLINE] [NEWLINE] If you want to teach critical thinking as a core subject, then ditch philosophy and instead integrate argument mapping into regular subjects.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't see<mask> this is relevant. [NEWLINE] [NEWLINE] <mask> are my attempts at trying to defy reality to prove the non-existence of reality of any use to anyone? [NEWLINE] [NEWLINE] Religion can make me happy<mask> well, shit, heroin can make me "happy". All kinds of things in this world can make me happy from<mask>'s considered "good" to "bad" by whomever. That's entirely beside the point and I'm not even sure<mask> you bring it up. [NEWLINE] [NEWLINE] <mask> for the sake of an argument, let me put these things in my perspective. Let's say there is a bear in front of me. Here are the thoughts in my head about that: [NEWLINE] [NEWLINE] * I can't be sure<mask> the bear is real. I can't be sure<mask> the bear can actually hurt me, or<mask> death that may follow is real. Can't ever be sure. [NEWLINE] * From whatever possibly false knowledge in my head that I can gather, I only have this frame of reference, this "world" that I know of and I understand some of its current rules (or at least, I understand them<mask> best<mask> I currently can) [NEWLINE] * From these rules and some of my past experiences, I gather that rather than running at the bear, trying to see<mask> the bear is real, and possibly risking death (<mask> I can't be sure<mask> anything follows, or<mask> I can even die), is not the solution I'm looking for. I enjoy my time in this world (or think that I do), and do not want to risk whatever is after this,<mask> anything whatsoever. [NEWLINE] *<mask>, I come to the conclusion that I do not want to experience, at this moment, the idea of "death". Whether it's possible (that I can die), whether anything follows (better or worse, afterlife or "respawn"), and whether the bear is real is irrelevant at this moment. [NEWLINE] * I turn to the thing that seems to follow the rules of this "reality" that could help me, let's say, a gun. From<mask> I know of the rules, it's likely that a warning shot will deter the bear.<mask> the bear is a figment of my imagination doesn't matter at this point. I will take the route I deem safest<mask><mask> the odds and rules of this world. [NEWLINE] [NEWLINE] That's it. I've already mentioned that "I accept the frame of reference around me." It's all I can do. </s>
Label encoding: <s>I don't see how this is relevant. [NEWLINE] [NEWLINE] How are my attempts at trying to defy reality to prove the non-existence of reality of any use to anyone? [NEWLINE] [NEWLINE] Religion can make me happy as well, shit, heroin can make me "happy". All kinds of things in this world can make me happy from what's considered "good" to "bad" by whomever. That's entirely beside the point and I'm not even sure why you bring it up. [NEWLINE] [NEWLINE] But for the sake of an argument, let me put these things in my perspective. Let's say there is a bear in front of me. Here are the thoughts in my head about that: [NEWLINE] [NEWLINE] * I can't be sure if the bear is real. I can't be sure if the bear can actually hurt me, or if death that may follow is real. Can't ever be sure. [NEWLINE] * From whatever possibly false knowledge in my head that I can gather, I only have this frame of reference, this "world" that I know of and I understand some of its current rules (or at least, I understand them as best as I currently can) [NEWLINE] * From these rules and some of my past experiences, I gather that rather than running at the bear, trying to see if the bear is real, and possibly risking death ( where I can't be sure if anything follows, or if I can even die), is not the solution I'm looking for. I enjoy my time in this world (or think that I do), and do not want to risk whatever is after this, if anything whatsoever. [NEWLINE] * Thus, I come to the conclusion that I do not want to experience, at this moment, the idea of "death". Whether it's possible (that I can die), whether anything follows (better or worse, afterlife or "respawn"), and whether the bear is real is irrelevant at this moment. [NEWLINE] * I turn to the thing that seems to follow the rules of this "reality" that could help me, let's say, a gun. From what I know of the rules, it's likely that a warning shot will deter the bear. If the bear is a figment of my imagination doesn't matter at this point. I will take the route I deem safest according to the odds and rules of this world. [NEWLINE] [NEWLINE] That's it. I've already mentioned that "I accept the frame of reference around me." It's all I can do. </s>
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Or sometimes these people roll in with a broken spine and end up being quadriplegics. [ENDQ] [NEWLINE] Yep,<mask> averages here. Chronic costs from being obese are large. Their shortened life expectancy doesn't cover the increased healthcare cost at living. [NEWLINE] [NEWLINE] <mask>, the majority of people are overweight (even obese) in the US. The majority of people do not ski, do not rock climb and do not sky dive. The flat costs on the systems aren't even on the same scale. [NEWLINE] [NEWLINE] [STARTQ] <mask> these people are in the health care system, the system should accommodate them. The system has to accommodate people of all shapes, sizes, and abilities even<mask> those things cost more. [ENDQ] [NEWLINE] Of course it does. That's the issue. We have to accommodate for them,<mask> it's one thing to have to have **all** your beds accommodate obese people rather than have only a few for the special cases. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> it's a noble goal to try to lower obesity in the aggregate to improve people's quality of life and lower health care costs,<mask> shaming individual fat people<mask> their health care costs more is<mask> unfair<mask> shaming a quadriplegic for his increased burden on the health care system<mask> he chose to go skiing that day. [ENDQ] [NEWLINE] The difference is that the former can "learn from their mistakes" and lose weight. They can do it **right now**. That's the issue here, this problem is directly solvable at all stages in the journey. On a skiing holiday everything is done to minimize risk. They skis are set for maximum safety. There are rules to follow. Nobody goes "oh look, he's a quadriplegic, well it's an acceptable risk" they go "Holy shit,<mask> can we do to stop that from happening to us!<mask> can we minimize the risks". [NEWLINE] [NEWLINE] The issue here is fat enabling. Making fat "acceptable" is the issue that **has to stop**, and in the end this means targeting people who aren't doing anything to improve. I'm thin,<mask> I'm regularly told by family that I should exercise more to be more fit (rather than sit in front of my PC all day).<mask><mask> with them and have started daily workout routines<mask><mask><mask>.<mask> was important<mask> was them not ignoring the issue and just saying they were happy with the situation. [NEWLINE] [NEWLINE] The real big reason comes back to cost associated with fat acceptance.</s>
Label encoding: <s> [STARTQ] Or sometimes these people roll in with a broken spine and end up being quadriplegics. [ENDQ] [NEWLINE] Yep, but averages here. Chronic costs from being obese are large. Their shortened life expectancy doesn't cover the increased healthcare cost at living. [NEWLINE] [NEWLINE] Also, the majority of people are overweight (even obese) in the US. The majority of people do not ski, do not rock climb and do not sky dive. The flat costs on the systems aren't even on the same scale. [NEWLINE] [NEWLINE] [STARTQ] If these people are in the health care system, the system should accommodate them. The system has to accommodate people of all shapes, sizes, and abilities even if those things cost more. [ENDQ] [NEWLINE] Of course it does. That's the issue. We have to accommodate for them, but it's one thing to have to have **all** your beds accommodate obese people rather than have only a few for the special cases. [NEWLINE] [NEWLINE] [STARTQ] I think it's a noble goal to try to lower obesity in the aggregate to improve people's quality of life and lower health care costs, but shaming individual fat people because their health care costs more is as unfair as shaming a quadriplegic for his increased burden on the health care system because he chose to go skiing that day. [ENDQ] [NEWLINE] The difference is that the former can "learn from their mistakes" and lose weight. They can do it **right now**. That's the issue here, this problem is directly solvable at all stages in the journey. On a skiing holiday everything is done to minimize risk. They skis are set for maximum safety. There are rules to follow. Nobody goes "oh look, he's a quadriplegic, well it's an acceptable risk" they go "Holy shit, what can we do to stop that from happening to us! How can we minimize the risks". [NEWLINE] [NEWLINE] The issue here is fat enabling. Making fat "acceptable" is the issue that **has to stop**, and in the end this means targeting people who aren't doing anything to improve. I'm thin, but I'm regularly told by family that I should exercise more to be more fit (rather than sit in front of my PC all day). I agree with them and have started daily workout routines as a result. What was important though was them not ignoring the issue and just saying they were happy with the situation. [NEWLINE] [NEWLINE] The real big reason comes back to cost associated with fat acceptance.</s>
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your reasoning is valid<mask><mask><mask> with the response you choose. [NEWLINE] [NEWLINE] Let's talk about the Facebook link you showed. To be blunt, the woman would not be considered physically beautiful out of context.<mask> your point is that it is inaccurate to say she is objectively beautiful, then you're correct. Try to consider,<mask>, that the people commenting on her beauty may be doing<mask> in an effort to be compassionate.<mask> the situation might seem simple, there's actually quite a few factors involved that influence the response. First, does she look arrogant or attention seeking? She's not the one posting the image, nor did she necessarily know<mask> many people would see it. She's showing a vulnerable moment and people empathize with her. Saying "you're beautiful" is another way of saying "we care about you," even<mask> it doesn't literally translate to that in the English language. And before you begin preparing counter-arguments, yes there are certainly situations<mask> telling a person they're beautiful is more harmful than helpful. And yes, a person might misinterpret people's intentions. Communication isn't perfect,<mask> in this situation<mask><mask> the comments were appropriate. Try to see peoples' intentions and consider that in your viewpoint. [NEWLINE] [NEWLINE] You<mask> mention<mask> allowing people to see their own flaws allows them to improve themselves. I completely agree.<mask> look back at the Facebook photo. The woman already has makeup and decent clothing.<mask> she were wearing rags I guarantee the comments would be more critical.<mask> people see in this picture is a woman who is dissatisfied with her image,<mask> trying the best with<mask> she has. Criticism won't get her much farther. Having positive feedback can make a tremendous difference in a person's day, and that's<mask> most of the comments are aiming for. Call it lying or dishonesty or false hope; at the end of the day everyone benefits from a kind gesture. [NEWLINE] [NEWLINE] <mask> many people don't realize it, the purpose of this subreddit is to change a person's view, not necessarily reverse it. Like I said before,<mask><mask> with your points on these kinds of comments having the potential to sound condescending or blind a person from their problems.<mask>'s important is that you don't generalize this logic to every situation<mask> a person shows their worries about being attractive. Your post makes it seem like nobody should EVER call an unattractive individual beautiful; there are a range of appropriate and inappropriate situations.</s>
Label encoding: <s>Your reasoning is valid but I disagree with the response you choose. [NEWLINE] [NEWLINE] Let's talk about the Facebook link you showed. To be blunt, the woman would not be considered physically beautiful out of context. If your point is that it is inaccurate to say she is objectively beautiful, then you're correct. Try to consider, however, that the people commenting on her beauty may be doing so in an effort to be compassionate. Although the situation might seem simple, there's actually quite a few factors involved that influence the response. First, does she look arrogant or attention seeking? She's not the one posting the image, nor did she necessarily know how many people would see it. She's showing a vulnerable moment and people empathize with her. Saying "you're beautiful" is another way of saying "we care about you," even if it doesn't literally translate to that in the English language. And before you begin preparing counter-arguments, yes there are certainly situations where telling a person they're beautiful is more harmful than helpful. And yes, a person might misinterpret people's intentions. Communication isn't perfect, but in this situation I think the comments were appropriate. Try to see peoples' intentions and consider that in your viewpoint. [NEWLINE] [NEWLINE] You also mention how allowing people to see their own flaws allows them to improve themselves. I completely agree. But look back at the Facebook photo. The woman already has makeup and decent clothing. If she were wearing rags I guarantee the comments would be more critical. What people see in this picture is a woman who is dissatisfied with her image, despite trying the best with what she has. Criticism won't get her much farther. Having positive feedback can make a tremendous difference in a person's day, and that's what most of the comments are aiming for. Call it lying or dishonesty or false hope; at the end of the day everyone benefits from a kind gesture. [NEWLINE] [NEWLINE] Although many people don't realize it, the purpose of this subreddit is to change a person's view, not necessarily reverse it. Like I said before, I agree with your points on these kinds of comments having the potential to sound condescending or blind a person from their problems. What's important is that you don't generalize this logic to every situation where a person shows their worries about being attractive. Your post makes it seem like nobody should EVER call an unattractive individual beautiful; there are a range of appropriate and inappropriate situations.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me quote *The Hogfather* by Terry Pratchett (a secular humanist).  This is a conversation between Death and his granddaughter Susan concerning the importance of the Hogfather (that world's version of Santa Claus). [NEWLINE] [NEWLINE] [STARTQ] “All right," said Susan. "I'm not stupid. You're saying humans need... fantasies to make life bearable." [ENDQ] [NEWLINE] [STARTQ] NO. HUMANS NEED FANTASY TO BE HUMAN. TO BE THE PLACE<mask> THE FALLING ANGEL MEETS THE RISING APE. [ENDQ] [NEWLINE] [STARTQ] "Tooth fairies? Hogfathers? Little—" [ENDQ] [NEWLINE] [STARTQ] YES.<mask> PRACTICE. YOU HAVE TO START OUT LEARNING TO BELIEVE THE LITTLE LIES. [ENDQ] [NEWLINE] [STARTQ] "<mask> we can believe the big ones?" [ENDQ] [NEWLINE] [STARTQ] YES. JUSTICE. MERCY. DUTY. THAT SORT OF THING. [ENDQ] [NEWLINE] [STARTQ] "They're not the same at all!" [ENDQ] [NEWLINE] [STARTQ] YOU THINK<mask>? THEN TAKE THE UNIVERSE AND GRIND IT DOWN TO THE FINEST POWDER AND SIEVE IT THROUGH THE FINEST SIEVE AND THEN SHOW ME ONE ATOM OF JUSTICE, ONE MOLECULE OF MERCY. AND<mask><mask> —Death waved a hand. AND<mask><mask> YOU ACT<mask><mask> THERE IS SOME IDEAL ORDER IN THE WORLD,<mask><mask> THERE IS SOME RIGHTNESS IN THE UNIVERSE BY WHICH IT MAY BE JUDGED. [ENDQ] [NEWLINE] [STARTQ] "Yes,<mask> people have got to believe that, or<mask>'s the point—" [ENDQ] [NEWLINE] [STARTQ] "MY POINT EXACTLY.” [ENDQ] [NEWLINE] Basically, stories like Santa cultivate belief, train little human beings<mask> to believe.  And humans need to believe in *something* in life.  Don't know<mask> this makes any sense,<mask> I'll squeeze Terry Pratchett in anywhere I can.  Sue me. [NEWLINE] [NEWLINE] EDIT:  Santa is a prototypical story.  Stories are some of the most beautiful and *human* parts of existence.  Santa lets children wholeheartedly participate in a fantastical story, making their lives more magical.  Which they'll be doing on a smaller scale the rest of their lives (<mask> you don't have a genuine emotional reaction to a book or a TV show you like,<mask> are you watching?).</s>
Label encoding: <s>Let me quote *The Hogfather* by Terry Pratchett (a secular humanist).  This is a conversation between Death and his granddaughter Susan concerning the importance of the Hogfather (that world's version of Santa Claus). [NEWLINE] [NEWLINE] [STARTQ] “All right," said Susan. "I'm not stupid. You're saying humans need... fantasies to make life bearable." [ENDQ] [NEWLINE] [STARTQ] NO. HUMANS NEED FANTASY TO BE HUMAN. TO BE THE PLACE WHERE THE FALLING ANGEL MEETS THE RISING APE. [ENDQ] [NEWLINE] [STARTQ] "Tooth fairies? Hogfathers? Little—" [ENDQ] [NEWLINE] [STARTQ] YES. AS PRACTICE. YOU HAVE TO START OUT LEARNING TO BELIEVE THE LITTLE LIES. [ENDQ] [NEWLINE] [STARTQ] " So we can believe the big ones?" [ENDQ] [NEWLINE] [STARTQ] YES. JUSTICE. MERCY. DUTY. THAT SORT OF THING. [ENDQ] [NEWLINE] [STARTQ] "They're not the same at all!" [ENDQ] [NEWLINE] [STARTQ] YOU THINK SO? THEN TAKE THE UNIVERSE AND GRIND IT DOWN TO THE FINEST POWDER AND SIEVE IT THROUGH THE FINEST SIEVE AND THEN SHOW ME ONE ATOM OF JUSTICE, ONE MOLECULE OF MERCY. AND YET —Death waved a hand. AND YET YOU ACT AS IF THERE IS SOME IDEAL ORDER IN THE WORLD, AS IF THERE IS SOME RIGHTNESS IN THE UNIVERSE BY WHICH IT MAY BE JUDGED. [ENDQ] [NEWLINE] [STARTQ] "Yes, but people have got to believe that, or what's the point—" [ENDQ] [NEWLINE] [STARTQ] "MY POINT EXACTLY.” [ENDQ] [NEWLINE] Basically, stories like Santa cultivate belief, train little human beings how to believe.  And humans need to believe in *something* in life.  Don't know if this makes any sense, but I'll squeeze Terry Pratchett in anywhere I can.  Sue me. [NEWLINE] [NEWLINE] EDIT:  Santa is a prototypical story.  Stories are some of the most beautiful and *human* parts of existence.  Santa lets children wholeheartedly participate in a fantastical story, making their lives more magical.  Which they'll be doing on a smaller scale the rest of their lives ( if you don't have a genuine emotional reaction to a book or a TV show you like, why are you watching?).</s>
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I apologize about the article. You've clearly given it far more attention than it deserved, and quite frankly I feel silly having linked to it now. Man, I really feel bad that you wrote all that out. [ENDQ] [NEWLINE] No big deal really. It's occasionally fun to tear apart anti vaccine articles like this. [NEWLINE] [NEWLINE] [STARTQ] you make really good points for the most part (<mask> have many others in this thread),<mask> I'm still not sure I feel comfortable with them. [ENDQ] [NEWLINE] Thanks and that's perfectly understandable. [NEWLINE] [NEWLINE] [STARTQ] I don't think it's fair to throw individual concerns in the dumpster. [ENDQ] [NEWLINE] And under no means should we.<mask> there is a risk involved with a procedure (<mask> there is with vaccination) it requires a fair amount of justification to force someone to partake, they should have decision making abilities. [NEWLINE] [NEWLINE] [STARTQ] The macroscopic decisions and policies should still be justifiable on a microscopic or individual basis [ENDQ] [NEWLINE] That they should.<mask> we often look at the *bad* parts of events without seeing the good. You got to experience the bad, the one in a very large number negative side effects<mask> let me ask you this,<mask> that one month of vertigo resulted in you *not* passing on HPV to someone you are intimate with with the end consequence being they *don't* get cervical cancer that they otherwise would have, would it have been worth it? I would make that sacrifice (having seen my father go through vertigo, I get<mask> you went through. It really, really sucks) 10 time out of 10.<mask> would you think of someone looking at your case thinking "I don't want to go through that" and being the direct cause of a partners cervical cancer? I'm not trying to use rhetorical devices here, it's an honest question. [NEWLINE] [NEWLINE] I guess<mask> I'm trying to get across is that it can be looked at<mask> a microscopic level of *harm* that is caused,<mask> it can<mask> be flipped and you can look at the microscopic benefit. Good that comes from vaccines at an individual level. A mother who gets to be there for her children, a child who doesn't die from small pox, etc. A vaccine that harms 0.1%, probably from minor or non life threatening effects, may be the direct cause of benefit, sometimes life saving, for many millions  of individuals.</s>
Label encoding: <s> [STARTQ] I apologize about the article. You've clearly given it far more attention than it deserved, and quite frankly I feel silly having linked to it now. Man, I really feel bad that you wrote all that out. [ENDQ] [NEWLINE] No big deal really. It's occasionally fun to tear apart anti vaccine articles like this. [NEWLINE] [NEWLINE] [STARTQ] you make really good points for the most part ( as have many others in this thread), but I'm still not sure I feel comfortable with them. [ENDQ] [NEWLINE] Thanks and that's perfectly understandable. [NEWLINE] [NEWLINE] [STARTQ] I don't think it's fair to throw individual concerns in the dumpster. [ENDQ] [NEWLINE] And under no means should we. If there is a risk involved with a procedure ( as there is with vaccination) it requires a fair amount of justification to force someone to partake, they should have decision making abilities. [NEWLINE] [NEWLINE] [STARTQ] The macroscopic decisions and policies should still be justifiable on a microscopic or individual basis [ENDQ] [NEWLINE] That they should. But we often look at the *bad* parts of events without seeing the good. You got to experience the bad, the one in a very large number negative side effects but let me ask you this, if that one month of vertigo resulted in you *not* passing on HPV to someone you are intimate with with the end consequence being they *don't* get cervical cancer that they otherwise would have, would it have been worth it? I would make that sacrifice (having seen my father go through vertigo, I get what you went through. It really, really sucks) 10 time out of 10. What would you think of someone looking at your case thinking "I don't want to go through that" and being the direct cause of a partners cervical cancer? I'm not trying to use rhetorical devices here, it's an honest question. [NEWLINE] [NEWLINE] I guess what I'm trying to get across is that it can be looked at as a microscopic level of *harm* that is caused, but it can also be flipped and you can look at the microscopic benefit. Good that comes from vaccines at an individual level. A mother who gets to be there for her children, a child who doesn't die from small pox, etc. A vaccine that harms 0.1%, probably from minor or non life threatening effects, may be the direct cause of benefit, sometimes life saving, for many millions  of individuals.</s>
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*“<mask><mask><mask> I know most people regard their holy texts<mask> the "word of their god".<mask> over time, cherry picking and deviating from the word of that go seems the same to me<mask> deviating from the religion itself.”* [NEWLINE] [NEWLINE] In this topic and your posts you have been making the logical error of taking a few things some people think about religion and applying them to religion<mask> a whole.<mask> you are arguing about the widespread belief in religion being a sign of weakness that is different than arguing about<mask> a few people may see things being contradictory. That is the point I was getting at mentioning the Westbro Baptist Church, that you aren't arguing about the details and not the idea of religion. [NEWLINE] [NEWLINE] I strongly disagree with nihilism and the idea that without a supreme authority there is no morality or system of being. There has never been a universal purpose for life (not counting evolution) or system of morality. Even<mask> looking at the same religious texts people will end up with different views about them. Morality has always been created and maintained by society. Religion was the dominant force for enforcing morality before, now we have the media to do that. The idea that you will go to heaven<mask> you follow the rules is very similar to The American Dream. [NEWLINE] [NEWLINE] There is no true morality or purpose except for the ones we make for ourselves. In the grand scheme of things our lives won't mean anything<mask> we still go and try to achieve the things we want to achieve. People make thier own morality and meaning out of life.<mask> you are a nihilist with this logic<mask> is stopping you from committing all the thefts, rapes, and murders your heart desires? [NEWLINE] [NEWLINE] *A life without meaning,<mask> nothing is "good" or "bad" can be a dangerous ideology for some people, and<mask> a dangerous ideology for the entire world.* [NEWLINE] [NEWLINE] This is a vague statement making an extremely illogical sweeping generalization. Capitalism can be a dangerous ideology for some people, and<mask> is a dangerous ideology for the entire world. Communism can be a dangerous ideology for some people, and<mask> is a dangerous ideology for the entire world. Feminism can be a dangerous ideology for some people, and<mask> is a dangerous ideology for the entire world. Alcohol can be a dangerous substance for some people, and<mask> a dangerous substance for the entire world. </s>
Label encoding: <s>*“ As far as I know most people regard their holy texts as the "word of their god". so over time, cherry picking and deviating from the word of that go seems the same to me as deviating from the religion itself.”* [NEWLINE] [NEWLINE] In this topic and your posts you have been making the logical error of taking a few things some people think about religion and applying them to religion as a whole. If you are arguing about the widespread belief in religion being a sign of weakness that is different than arguing about how a few people may see things being contradictory. That is the point I was getting at mentioning the Westbro Baptist Church, that you aren't arguing about the details and not the idea of religion. [NEWLINE] [NEWLINE] I strongly disagree with nihilism and the idea that without a supreme authority there is no morality or system of being. There has never been a universal purpose for life (not counting evolution) or system of morality. Even when looking at the same religious texts people will end up with different views about them. Morality has always been created and maintained by society. Religion was the dominant force for enforcing morality before, now we have the media to do that. The idea that you will go to heaven if you follow the rules is very similar to The American Dream. [NEWLINE] [NEWLINE] There is no true morality or purpose except for the ones we make for ourselves. In the grand scheme of things our lives won't mean anything but we still go and try to achieve the things we want to achieve. People make thier own morality and meaning out of life. If you are a nihilist with this logic what is stopping you from committing all the thefts, rapes, and murders your heart desires? [NEWLINE] [NEWLINE] *A life without meaning, where nothing is "good" or "bad" can be a dangerous ideology for some people, and therefore a dangerous ideology for the entire world.* [NEWLINE] [NEWLINE] This is a vague statement making an extremely illogical sweeping generalization. Capitalism can be a dangerous ideology for some people, and therefore is a dangerous ideology for the entire world. Communism can be a dangerous ideology for some people, and therefore is a dangerous ideology for the entire world. Feminism can be a dangerous ideology for some people, and therefore is a dangerous ideology for the entire world. Alcohol can be a dangerous substance for some people, and therefore a dangerous substance for the entire world. </s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right by<mask> metric? [NEWLINE] [NEWLINE] The Na'vi clearly value harmony with nature and traditionalism over colonialism and exploitation. From their perspective, the humans are clearly in the wrong. [NEWLINE] [NEWLINE] The humans are only 'right'<mask> we consider conquest and colonialism<mask> the measures of a civilizations success, which arguably are shitty measures<mask><mask> they are at odds with most of nature (<mask> other animals mine?<mask> other animals detonate nukes?), not to mention threaten the life of humanity itself. [NEWLINE] [NEWLINE] Think of the Matrix,<mask> Smith compares humanity to a virus, spreading like wildfire and killing everything it touches, including the very biosphere that makes humanity even possible. Would you say this is "right"? Is it a great success<mask> a virus wipes out it's host, killing itself in the process? A fire that burns fast and bright is great to look at,<mask> it won't keep you warm throughout the winter. [NEWLINE] [NEWLINE] <mask> for Quaritch's actions themselves, well, guy was working with<mask> he had. It's worth mentioning that he was a mercenary, not a soldier, and that the army that took down Hometree was a re-fitted mining op, not a professional army. Given<mask> he had to work with, I'd say he did okay. The issue isn't with him,<mask> rather his superiors<mask> well<mask> the larger human society that make his actions necessary. [NEWLINE] [NEWLINE] And<mask>, on that quote specifically,<mask><mask> it's worth noting that this is Quaritch justifying his actions after they've already began. He didn't come to Pandora to further human civilization; he came<mask> it was his job. His superiors didn't start mining on Pandora<mask> they want to further civilization; they did it<mask> they wanted to make money (it was a corporation<mask> I recall). Humanities' forays into space weren't done<mask> voyages of discovery or to bring civilization to the galaxy; it was done<mask> earth was running out of space/resources (<mask> I recall correctly the hero mentions at one point that Earth is a shit hole). [NEWLINE] [NEWLINE] Quaritch killed a lot of aliens, burned down some forest, and they decided to justify it after the fact<mask> bringing the light of civilization into the cosmos. He didn't decide to bring civilization to space and figured the best way to do it was to blow shit up</s>
Label encoding: <s>Right by what metric? [NEWLINE] [NEWLINE] The Na'vi clearly value harmony with nature and traditionalism over colonialism and exploitation. From their perspective, the humans are clearly in the wrong. [NEWLINE] [NEWLINE] The humans are only 'right' if we consider conquest and colonialism as the measures of a civilizations success, which arguably are shitty measures given that they are at odds with most of nature ( what other animals mine? What other animals detonate nukes?), not to mention threaten the life of humanity itself. [NEWLINE] [NEWLINE] Think of the Matrix, when Smith compares humanity to a virus, spreading like wildfire and killing everything it touches, including the very biosphere that makes humanity even possible. Would you say this is "right"? Is it a great success when a virus wipes out it's host, killing itself in the process? A fire that burns fast and bright is great to look at, but it won't keep you warm throughout the winter. [NEWLINE] [NEWLINE] As for Quaritch's actions themselves, well, guy was working with what he had. It's worth mentioning that he was a mercenary, not a soldier, and that the army that took down Hometree was a re-fitted mining op, not a professional army. Given what he had to work with, I'd say he did okay. The issue isn't with him, but rather his superiors as well as the larger human society that make his actions necessary. [NEWLINE] [NEWLINE] And also, on that quote specifically, I think it's worth noting that this is Quaritch justifying his actions after they've already began. He didn't come to Pandora to further human civilization; he came because it was his job. His superiors didn't start mining on Pandora because they want to further civilization; they did it because they wanted to make money (it was a corporation if I recall). Humanities' forays into space weren't done as voyages of discovery or to bring civilization to the galaxy; it was done because earth was running out of space/resources ( if I recall correctly the hero mentions at one point that Earth is a shit hole). [NEWLINE] [NEWLINE] Quaritch killed a lot of aliens, burned down some forest, and they decided to justify it after the fact as bringing the light of civilization into the cosmos. He didn't decide to bring civilization to space and figured the best way to do it was to blow shit up</s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I chose a dog,<mask> they tend to be the animal most likely to sexually assault their owner, and many other things...whether or not it's wanted... [NEWLINE] [NEWLINE] We can<mask> talk dolphins,<mask> you'd prefer? [NEWLINE] [NEWLINE] Then there are all the breeding programs that traditionally involve getting an orgasm from an animal... [NEWLINE] [NEWLINE] Look, I don't enjoy this conversation any more than you do. I already feel ill, and this is only going to get worse, from here on out.<mask> I know a girl, who, once<mask> she was masturbating, was surprised by her cat's tongue. And she let him finish. [NEWLINE] [NEWLINE] By your logic, freezing up made her a rapist. Or was she a rapist<mask> she confessed she enjoyed it? [NEWLINE] [NEWLINE] Or was it<mask> she let him do it again? She loved that cat, and not<mask> a sex toy. They were the best of friends - I never had reason to doubt he was happy. [NEWLINE] [NEWLINE] Every living creature should be that fortunate. [NEWLINE] [NEWLINE] Please don't misunderstand. I'm not arguing that sex with animals is something that should be pursued. Reddit,<mask> it forgets to mark pictures NSFW, has shown me the horrible things some women will do to mice, eels and octopi,<mask> a camera is around to record it all, and she thinks there's an audience for it. Their last moments were spent suffocating inside her, terrified. Even a predator isn't usually that pointlessly cruel. [NEWLINE] [NEWLINE] And I don't need to explain<mask> men can do to an animal with their genitals. There's no question much of this is rape, and torture, and needs to be stopped at all costs. [NEWLINE] [NEWLINE] <mask><mask> a rape victim, I am opposed to cheapening the meaning of the word rape, by blindly applying it to other situations, and dumbing down the conversation. [NEWLINE] [NEWLINE] Or, let me put it another way... [NEWLINE] [NEWLINE] Do you think 18 is old enough for someone to consent to sex,<mask><mask> their brain hasn't finished developing, and won't until their twenties?<mask><mask>,<mask>?<mask> about 17? 16?<mask> makes someone able to consent,<mask> a random number? [NEWLINE] [NEWLINE] <mask> it's not old enough,<mask> aren't you trying to prevent all those rapes? </s>
Label encoding: <s>I chose a dog, because they tend to be the animal most likely to sexually assault their owner, and many other things...whether or not it's wanted... [NEWLINE] [NEWLINE] We can also talk dolphins, if you'd prefer? [NEWLINE] [NEWLINE] Then there are all the breeding programs that traditionally involve getting an orgasm from an animal... [NEWLINE] [NEWLINE] Look, I don't enjoy this conversation any more than you do. I already feel ill, and this is only going to get worse, from here on out. But I know a girl, who, once when she was masturbating, was surprised by her cat's tongue. And she let him finish. [NEWLINE] [NEWLINE] By your logic, freezing up made her a rapist. Or was she a rapist because she confessed she enjoyed it? [NEWLINE] [NEWLINE] Or was it when she let him do it again? She loved that cat, and not as a sex toy. They were the best of friends - I never had reason to doubt he was happy. [NEWLINE] [NEWLINE] Every living creature should be that fortunate. [NEWLINE] [NEWLINE] Please don't misunderstand. I'm not arguing that sex with animals is something that should be pursued. Reddit, because it forgets to mark pictures NSFW, has shown me the horrible things some women will do to mice, eels and octopi, if a camera is around to record it all, and she thinks there's an audience for it. Their last moments were spent suffocating inside her, terrified. Even a predator isn't usually that pointlessly cruel. [NEWLINE] [NEWLINE] And I don't need to explain what men can do to an animal with their genitals. There's no question much of this is rape, and torture, and needs to be stopped at all costs. [NEWLINE] [NEWLINE] But as a rape victim, I am opposed to cheapening the meaning of the word rape, by blindly applying it to other situations, and dumbing down the conversation. [NEWLINE] [NEWLINE] Or, let me put it another way... [NEWLINE] [NEWLINE] Do you think 18 is old enough for someone to consent to sex, even though their brain hasn't finished developing, and won't until their twenties? If so, why? What about 17? 16? What makes someone able to consent, besides a random number? [NEWLINE] [NEWLINE] If it's not old enough, why aren't you trying to prevent all those rapes? </s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0930, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From an institutional perspective (meaning your employers perspective) standards like dress codes help create a baseline of acceptable presentation and behavior. Once that baseline is established, typically some occasional variation is allowable<mask><mask><mask> no one abuses the privilege. Dress codes<mask> serve the purpose of creating a specific and universal set of rules, which alleviates the need of having to individually address each and every employees personal interpretation of<mask> is or isn't acceptable.<mask> a manager, I don't want waste time arguing with an employee or my boss over whether their definition of "naughty bits" is the same<mask> mine,<mask> I set a dress code of slacks and a button up shirt and avoid the problem altogether. These are a few of the many benefits that an employer may find in dress codes. They may not be benefits *you* care about,<mask> they are benefits none the less. [NEWLINE] [NEWLINE] From am individual perspective, dressing up a bit **does** change the way people perceive and treat you. The company I currently work for has a set dress code: Black, logo monogrammed polo and slacks. I never fucking wear that polo. From years of experience I've learned that<mask> you're wearing a logo'd polo shirt people treat you like the hired help,<mask><mask> you wear a button up shirt and a tie (even an oversized, untucked button up) they treat you like someone who can get things done for them. The difference is stunning. I've gotten more raises than my peers, better gigs, more compliments and client recommendations, etc, etc. You might not believe that this should be the case,<mask> it is. Dressing to impress is a thing, it works, and you should do it whenever you get the chance. This is a clear and definite benefit. Dressing up a bit<mask> changes<mask> you feel about yourself, it can increase your confidence, [NEWLINE] [NEWLINE] You aren't going to find any clear evidence that dress codes have a universal benefit,<mask> they don't, nothing does. Some companies may find that a relaxed dress code increases moral, for others it may lead to lower productivity.<mask> with everything their are benefits and detriments a plenty that will manifest themselves in accordance with the other variables that influence any group of people.<mask> there **are** benefits.</s><pad>
Label encoding: <s>From an institutional perspective (meaning your employers perspective) standards like dress codes help create a baseline of acceptable presentation and behavior. Once that baseline is established, typically some occasional variation is allowable as long as no one abuses the privilege. Dress codes also serve the purpose of creating a specific and universal set of rules, which alleviates the need of having to individually address each and every employees personal interpretation of what is or isn't acceptable. As a manager, I don't want waste time arguing with an employee or my boss over whether their definition of "naughty bits" is the same as mine, so I set a dress code of slacks and a button up shirt and avoid the problem altogether. These are a few of the many benefits that an employer may find in dress codes. They may not be benefits *you* care about, but they are benefits none the less. [NEWLINE] [NEWLINE] From am individual perspective, dressing up a bit **does** change the way people perceive and treat you. The company I currently work for has a set dress code: Black, logo monogrammed polo and slacks. I never fucking wear that polo. From years of experience I've learned that if you're wearing a logo'd polo shirt people treat you like the hired help, but if you wear a button up shirt and a tie (even an oversized, untucked button up) they treat you like someone who can get things done for them. The difference is stunning. I've gotten more raises than my peers, better gigs, more compliments and client recommendations, etc, etc. You might not believe that this should be the case, but it is. Dressing to impress is a thing, it works, and you should do it whenever you get the chance. This is a clear and definite benefit. Dressing up a bit also changes how you feel about yourself, it can increase your confidence, [NEWLINE] [NEWLINE] You aren't going to find any clear evidence that dress codes have a universal benefit, because they don't, nothing does. Some companies may find that a relaxed dress code increases moral, for others it may lead to lower productivity. As with everything their are benefits and detriments a plenty that will manifest themselves in accordance with the other variables that influence any group of people. But there **are** benefits.</s><pad>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't see any good at all from keeping them around [ENDQ] [NEWLINE] Determining the value of one human being over another is an inherently sticky problem, especially<mask> you're bringing in the idea of getting rid of those who aren't "valuable enough". There are always those who will see particular groups of people<mask> "less than" those around them.<mask> who should be granted that wonderful task of determining such a thing? You?<mask> makes you more qualified than others to determine the worth of another person's life, especially<mask> you're not even the one responsible for taking care of it? [NEWLINE] [NEWLINE] <mask><mask> those in charge of making such decisions started doing<mask> on the basis of race?<mask> about eliminating people who practice particular religions? Sometimes the reason not to do something (like eliminating inconvenient groups of people) isn't<mask> it adds a direct benefit to society to keep them,<mask><mask> to engage in such an act would make that society less worth living in (<mask> of who that society would require us to be in order to be a part of it). [NEWLINE] [NEWLINE] Your dislike of those who are mentally handicapped is subjective,<mask> are the other examples mentioned here.<mask> I have someone in my family who is mentally handicapped,<mask> feel<mask><mask> his or her presence adds to the overall quality of our life experiences, who are you to tell us that this family member isn't contributing to our overall well-being? Who are you to decide that this family member of mine should be eliminated simply<mask> it makes you uncomfortable? [NEWLINE] [NEWLINE] Sometimes the act of living in society is itself a social commitment to one another, that in times<mask> we are not at our best requires the people<mask> a whole to maintain an obligation to one another that says we will help to take up the slack in times<mask> we the individual may not always be able to carry the load. It's this social contract to one another that allows us to have a certain level of existential anxiety relieved,<mask> even<mask> we are struggling ourselves, we know that our friends, our family, and sometimes even the society<mask> a whole will help us through.<mask> it's fine to condemn others and say "<mask> I've never needed such services or support from others", this doesn't mean that you or those you love never will.</s>
Label encoding: <s> [STARTQ] I don't see any good at all from keeping them around [ENDQ] [NEWLINE] Determining the value of one human being over another is an inherently sticky problem, especially if you're bringing in the idea of getting rid of those who aren't "valuable enough". There are always those who will see particular groups of people as "less than" those around them. So who should be granted that wonderful task of determining such a thing? You? What makes you more qualified than others to determine the worth of another person's life, especially if you're not even the one responsible for taking care of it? [NEWLINE] [NEWLINE] What if those in charge of making such decisions started doing so on the basis of race? What about eliminating people who practice particular religions? Sometimes the reason not to do something (like eliminating inconvenient groups of people) isn't because it adds a direct benefit to society to keep them, but because to engage in such an act would make that society less worth living in ( because of who that society would require us to be in order to be a part of it). [NEWLINE] [NEWLINE] Your dislike of those who are mentally handicapped is subjective, as are the other examples mentioned here. If I have someone in my family who is mentally handicapped, but feel as though his or her presence adds to the overall quality of our life experiences, who are you to tell us that this family member isn't contributing to our overall well-being? Who are you to decide that this family member of mine should be eliminated simply because it makes you uncomfortable? [NEWLINE] [NEWLINE] Sometimes the act of living in society is itself a social commitment to one another, that in times when we are not at our best requires the people as a whole to maintain an obligation to one another that says we will help to take up the slack in times when we the individual may not always be able to carry the load. It's this social contract to one another that allows us to have a certain level of existential anxiety relieved, because even if we are struggling ourselves, we know that our friends, our family, and sometimes even the society as a whole will help us through. While it's fine to condemn others and say " but I've never needed such services or support from others", this doesn't mean that you or those you love never will.</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Marriage is costly and arbitrary.  The marital ceremony plays no role in a couple's happiness. <mask><mask><mask><mask><mask>, marriage can actually compromise a couple's happiness. [NEWLINE] [NEWLINE] Marriage is confining.  Evolutionarily speaking, most people are serial monogamists.  The fact that at least forty percent of American marriages are marked by adultery makes this apparent. [NEWLINE] [NEWLINE] The majority of marriages are unhappy.  In the West, over half of  marriages end in separation or divorce. One cannot pretend that remaining marriages survive<mask> their participants are<mask> happy.  Often, marriages persist due to religious, financial, and emotional threats. [NEWLINE] [NEWLINE] Separations and divorces can be crippling. <mask> one partner is the "breadwinner" and the other is the homemaker, the homemaker is often left financially dependent.  Children can suffer immensely from relocation and emotional strife that develops within the family. [NEWLINE] [NEWLINE] I understand the financial perks of marriage and possible benefits of marriage<mask> there are children.  I am not against anybody's *right* to marry. <mask>,<mask> the son of two parents who had a nasty split, I am reminded of the pitfalls of marriage everyday. [NEWLINE] [NEWLINE] I don't want to feel this way,<mask>.  I hope the evidence is there to change my view.  I would love a long-term partner,<mask> think many girls would see a guy against marriage<mask> one who cannot commit. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than just downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Marriage is costly and arbitrary.  The marital ceremony plays no role in a couple's happiness.  As a matter of fact, marriage can actually compromise a couple's happiness. [NEWLINE] [NEWLINE] Marriage is confining.  Evolutionarily speaking, most people are serial monogamists.  The fact that at least forty percent of American marriages are marked by adultery makes this apparent. [NEWLINE] [NEWLINE] The majority of marriages are unhappy.  In the West, over half of  marriages end in separation or divorce. One cannot pretend that remaining marriages survive because their participants are so happy.  Often, marriages persist due to religious, financial, and emotional threats. [NEWLINE] [NEWLINE] Separations and divorces can be crippling.  When one partner is the "breadwinner" and the other is the homemaker, the homemaker is often left financially dependent.  Children can suffer immensely from relocation and emotional strife that develops within the family. [NEWLINE] [NEWLINE] I understand the financial perks of marriage and possible benefits of marriage when there are children.  I am not against anybody's *right* to marry.  However, as the son of two parents who had a nasty split, I am reminded of the pitfalls of marriage everyday. [NEWLINE] [NEWLINE] I don't want to feel this way, though.  I hope the evidence is there to change my view.  I would love a long-term partner, but think many girls would see a guy against marriage as one who cannot commit. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than just downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The idea that there is some natural balance that humans are disturbing. The world has never been at a balance. It has always oscillated violently between extremes. It has happened before the humans and will continue after we are gone. [ENDQ] [NEWLINE] Making this argument (<mask> you are) regarding CO2 concentration in the atmosphere is just plain wrong.  We know the past concentrations extremely well. [NEWLINE] [NEWLINE] CO2 is a gas, it exists in the atmosphere, and it has a definable mass<mask> such.  We have a decent "scale" with which to measure this mass.  We have directly put CO2 gas into the atmosphere, and this is consistent with measurements that it has increased, and will be double the pre-industrial level at some point in the future. [NEWLINE] [NEWLINE] It absolutely blows my mind that your current position lies in contradiction to this.  We produced a gas through a chemical reaction, then there was more gas in the container<mask> we put it.  Do you understand that?  I hesitate to even call it an "argument",<mask> it's an appeal to permanency of matter itself. <mask> I'm playing with a infant, and I put a toy behind my back, the toy doesn't disappear.  Over time, the child learns that objects don't come into and out of existence easily.  Nothing about this changes<mask> you talk about gases.  Gases are<mask> things. [NEWLINE] [NEWLINE] I understand that you have some trouble with the fact that CO2 gas goes into and out of the atmosphere seasonally, and at a greater rate that human emissions. <mask> the natural uptake of CO2 isn't up-taking all that we emit - evidenced by the simple fact that *there is now more up there*. <mask> did it come from?  It came from geologic sources.  We put a pipe into the ground, extracted hydrocarbon compounds, refined it, and burned it.  The Carbon was completely stable in the ground.  It had not moved for a long time.  We added it to the atmosphere. <mask> about this do you disagree with? [NEWLINE] [NEWLINE] I am constantly flabbergasted by the breakdown of basic object permanency on the part of AWG skeptics.  I swear, you learned this in preschool.</s><pad>
Label encoding: <s> [STARTQ] The idea that there is some natural balance that humans are disturbing. The world has never been at a balance. It has always oscillated violently between extremes. It has happened before the humans and will continue after we are gone. [ENDQ] [NEWLINE] Making this argument ( as you are) regarding CO2 concentration in the atmosphere is just plain wrong.  We know the past concentrations extremely well. [NEWLINE] [NEWLINE] CO2 is a gas, it exists in the atmosphere, and it has a definable mass as such.  We have a decent "scale" with which to measure this mass.  We have directly put CO2 gas into the atmosphere, and this is consistent with measurements that it has increased, and will be double the pre-industrial level at some point in the future. [NEWLINE] [NEWLINE] It absolutely blows my mind that your current position lies in contradiction to this.  We produced a gas through a chemical reaction, then there was more gas in the container where we put it.  Do you understand that?  I hesitate to even call it an "argument", because it's an appeal to permanency of matter itself.  If I'm playing with a infant, and I put a toy behind my back, the toy doesn't disappear.  Over time, the child learns that objects don't come into and out of existence easily.  Nothing about this changes when you talk about gases.  Gases are also things. [NEWLINE] [NEWLINE] I understand that you have some trouble with the fact that CO2 gas goes into and out of the atmosphere seasonally, and at a greater rate that human emissions.  But the natural uptake of CO2 isn't up-taking all that we emit - evidenced by the simple fact that *there is now more up there*.  Where did it come from?  It came from geologic sources.  We put a pipe into the ground, extracted hydrocarbon compounds, refined it, and burned it.  The Carbon was completely stable in the ground.  It had not moved for a long time.  We added it to the atmosphere.  What about this do you disagree with? [NEWLINE] [NEWLINE] I am constantly flabbergasted by the breakdown of basic object permanency on the part of AWG skeptics.  I swear, you learned this in preschool.</s><pad>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is called [The Transparent Society]( [URL] ) (short version in Wired [here]( [URL].12/fftransparent_pr.html)). I took a law and technology course and we discussed this for about 3 classes. [NEWLINE] [NEWLINE] The problem we have in society right now is *information asymmetry*. Some people (usually those who have political power) know more about you than you know about them. Or otherwise they have the capability to know more about you than you can ever know about them. This gives them power over you -- they know<mask> you are,<mask> you're doing, who you're doing it with,<mask> you're thinking etc.. and they can use that against you should they chose to. [NEWLINE] [NEWLINE] In short, this creates a massive *imbalance* of power between people who have information and people who dont, and is one of many factors that helps totalitarian states to stay in power. The Transparent Society solves this problem by making information asymmetry go away. [NEWLINE] [NEWLINE] <mask> the problem just shifts. The problem with The Transparent Society becomes *technology asymmetry*. Sure, everyone *can* know everything about everyone,<mask> only those with the knowledge and resources to put that information to proper use will actually do anything with it. The vast majority of<mask> is recorded is useless. They can *find* the needles in the haystack, and you can't. [NEWLINE] [NEWLINE] Those who can use technology to filter out the important stuff now have the power.<mask> then you say, well<mask><mask> everyone could do that too? [NEWLINE] [NEWLINE] Then that's asking for everyone to have equal opportunity to education and resources, and the problem shifts again. Maybe<mask> the Internet becomes entirely autonomous, we have 3D printers and machine learning will endlessly beef up Wikipedia that'll happen. Lots more has to be discovered and implemented before that'll ever happen. [NEWLINE] [NEWLINE] The Transparent Society solves some problems, and creates new ones. Unclear<mask> its a good thing or bad thing overall,<mask> it almost certainly would be horrible in the transition phase<mask> of the total loss of privacy, and the legacy of those who hold power now (i.e. good luck getting those in power to give it up). [NEWLINE] [NEWLINE] edit: fixes of crap sentences</s>
Label encoding: <s>This is called [The Transparent Society]( [URL] ) (short version in Wired [here]( [URL].12/fftransparent_pr.html)). I took a law and technology course and we discussed this for about 3 classes. [NEWLINE] [NEWLINE] The problem we have in society right now is *information asymmetry*. Some people (usually those who have political power) know more about you than you know about them. Or otherwise they have the capability to know more about you than you can ever know about them. This gives them power over you -- they know where you are, what you're doing, who you're doing it with, what you're thinking etc.. and they can use that against you should they chose to. [NEWLINE] [NEWLINE] In short, this creates a massive *imbalance* of power between people who have information and people who dont, and is one of many factors that helps totalitarian states to stay in power. The Transparent Society solves this problem by making information asymmetry go away. [NEWLINE] [NEWLINE] But the problem just shifts. The problem with The Transparent Society becomes *technology asymmetry*. Sure, everyone *can* know everything about everyone, but only those with the knowledge and resources to put that information to proper use will actually do anything with it. The vast majority of what is recorded is useless. They can *find* the needles in the haystack, and you can't. [NEWLINE] [NEWLINE] Those who can use technology to filter out the important stuff now have the power. So then you say, well what if everyone could do that too? [NEWLINE] [NEWLINE] Then that's asking for everyone to have equal opportunity to education and resources, and the problem shifts again. Maybe if the Internet becomes entirely autonomous, we have 3D printers and machine learning will endlessly beef up Wikipedia that'll happen. Lots more has to be discovered and implemented before that'll ever happen. [NEWLINE] [NEWLINE] The Transparent Society solves some problems, and creates new ones. Unclear if its a good thing or bad thing overall, but it almost certainly would be horrible in the transition phase because of the total loss of privacy, and the legacy of those who hold power now (i.e. good luck getting those in power to give it up). [NEWLINE] [NEWLINE] edit: fixes of crap sentences</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think there are several layers of misconception going on here (and one of the layers is that thought on these matters is shifting, and I'm not certain that anyone has the "right" idea about mental illness at the moment). [NEWLINE] [NEWLINE] Layer 1: a depressed person is still responsible for their actions.<mask> they do something to hurt someone else, it is still their fault.<mask>'s important to recognize is that the depressed person is not depressed by choice: they cannot flip a switch and stop being depressed. They need help digging themselves out, and that help should usually take a professional, medical form. [NEWLINE] [NEWLINE] Perhaps analogy might help:<mask> you're driving your car around with a fuel leak, and do nothing to get it fixed, you're going to be responsible for any accidents you might cause. The principle isn't too different<mask> you're piloting a brain with some medical issues. It's your responsibility to fix the issue,<mask> you probably have to go to a professional to get it fixed. [NEWLINE] [NEWLINE] Layer 2: "lazy" isn't a medical diagnosis,<mask> depression and anxiety are, and<mask><mask> that most people who you might call lazy are suffering from some level of depression or anxiety. Lazy is a symptom; anxiety might be the cause. [NEWLINE] [NEWLINE] Layer 3: The boundary between "clinically diagnosable", and within the bounds of "normal", is a little bit hard to pin down. Everybody exhibits depressive or anxious behaviors sometimes. Diagnosis usually happens<mask> those behaviors begin to seriously interfere in that person's life.<mask><mask><mask> that there are some profound implications of the disease model of mental disorders that we haven't fully worked through<mask> a society --<mask><mask> that we assume that people can just power through some things that they really can't (this relates to obesity, school schedules and teen sleep deprivation, the influence of advertising and shady sales tactics, etc.) [NEWLINE] [NEWLINE] Layer 4: none of this absolves people of personal responsibility for their actions.<mask> it can inform the ways in which we seek to better ourselves, and the ways in which we see each others' "character", and "character flaws".<mask><mask> that we should do less judging of each other, and more offering help and support.</s>
Label encoding: <s>I think there are several layers of misconception going on here (and one of the layers is that thought on these matters is shifting, and I'm not certain that anyone has the "right" idea about mental illness at the moment). [NEWLINE] [NEWLINE] Layer 1: a depressed person is still responsible for their actions. If they do something to hurt someone else, it is still their fault. What's important to recognize is that the depressed person is not depressed by choice: they cannot flip a switch and stop being depressed. They need help digging themselves out, and that help should usually take a professional, medical form. [NEWLINE] [NEWLINE] Perhaps analogy might help: if you're driving your car around with a fuel leak, and do nothing to get it fixed, you're going to be responsible for any accidents you might cause. The principle isn't too different when you're piloting a brain with some medical issues. It's your responsibility to fix the issue, though you probably have to go to a professional to get it fixed. [NEWLINE] [NEWLINE] Layer 2: "lazy" isn't a medical diagnosis, but depression and anxiety are, and I think that most people who you might call lazy are suffering from some level of depression or anxiety. Lazy is a symptom; anxiety might be the cause. [NEWLINE] [NEWLINE] Layer 3: The boundary between "clinically diagnosable", and within the bounds of "normal", is a little bit hard to pin down. Everybody exhibits depressive or anxious behaviors sometimes. Diagnosis usually happens when those behaviors begin to seriously interfere in that person's life. But I think that there are some profound implications of the disease model of mental disorders that we haven't fully worked through as a society -- I think that we assume that people can just power through some things that they really can't (this relates to obesity, school schedules and teen sleep deprivation, the influence of advertising and shady sales tactics, etc.) [NEWLINE] [NEWLINE] Layer 4: none of this absolves people of personal responsibility for their actions. But it can inform the ways in which we seek to better ourselves, and the ways in which we see each others' "character", and "character flaws". I think that we should do less judging of each other, and more offering help and support.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> you should look at it from a standpoint of<mask> precious life is.<mask> recently we have concluded that life isn't that rare in the cosmos it is rare in the aspect that the odds are insurmountable against a certain individual,<mask><mask> the individual being a human/animal, from being born. Just think about it.<mask> many things must have happened perfectly in order for that individual to be born. Theoretically speaking you could trace any lineage back to the first cell ever created. In that aspect<mask> there was even a single change or break in the lineage the individual wound not exists. Here I'm going to attach a quote<mask> the speaker compares life to a miracle. It's from the movie watchmen and I feel it explains it way better than I can. [NEWLINE] -----"Miracles. Events with astronomical odds of occurring, like oxygen turning into gold. I've longed to witness such an event, and<mask> I neglect that in human coupling, millions upon millions of cells compete to create life, for generation after generation until, finally, your mother loves a man, Edward Blake, the Comedian, a man she has every reason to hate, and out of that contradiction, against unfathomable odds, it's you - only you - that emerged. To distill<mask> specific a form, from all that chaos. It's like turning air into gold. A miracle." -Jon Osterman [NEWLINE] [NEWLINE] <mask> the odds are<mask> astoundingly against any individual from being born it just goes to show<mask> precious life is. The same theory can be applied to groups of animals such<mask> a species<mask> their mere existence by evolution defies logic. Its<mask><mask> they were chosen in the grand sense of things to exists. [NEWLINE] [NEWLINE] I think we<mask> humans have an obligation simply<mask> we can understand<mask> precious life is.<mask> a species goes extinct, its gone forever and that lineage no longer exists. Its like buying the winning ticket to infinite wealth and then ripping it up without checking<mask> you won.<mask> humans are the cause of extinction and we have it in their power to prevent it from happening then we should simply<mask> only we can. The animals can't protect themselves from us.  </s><pad>
Label encoding: <s>I think you should look at it from a standpoint of how precious life is. Although recently we have concluded that life isn't that rare in the cosmos it is rare in the aspect that the odds are insurmountable against a certain individual, regardless of the individual being a human/animal, from being born. Just think about it. How many things must have happened perfectly in order for that individual to be born. Theoretically speaking you could trace any lineage back to the first cell ever created. In that aspect if there was even a single change or break in the lineage the individual wound not exists. Here I'm going to attach a quote where the speaker compares life to a miracle. It's from the movie watchmen and I feel it explains it way better than I can. [NEWLINE] -----"Miracles. Events with astronomical odds of occurring, like oxygen turning into gold. I've longed to witness such an event, and yet I neglect that in human coupling, millions upon millions of cells compete to create life, for generation after generation until, finally, your mother loves a man, Edward Blake, the Comedian, a man she has every reason to hate, and out of that contradiction, against unfathomable odds, it's you - only you - that emerged. To distill so specific a form, from all that chaos. It's like turning air into gold. A miracle." -Jon Osterman [NEWLINE] [NEWLINE] Since the odds are so astoundingly against any individual from being born it just goes to show how precious life is. The same theory can be applied to groups of animals such as a species since their mere existence by evolution defies logic. Its as if they were chosen in the grand sense of things to exists. [NEWLINE] [NEWLINE] I think we as humans have an obligation simply because we can understand how precious life is. If a species goes extinct, its gone forever and that lineage no longer exists. Its like buying the winning ticket to infinite wealth and then ripping it up without checking if you won. If humans are the cause of extinction and we have it in their power to prevent it from happening then we should simply because only we can. The animals can't protect themselves from us.  </s><pad>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're oversimplifying the issue. A simple comparison of "it work for the Czechs" or "it worked for the Estonians" doesn't work<mask> they have drastically different economies and industries which complicates things. [NEWLINE] [NEWLINE] Side note: I don't know much about the issues the Czech Republic and Estonia had<mask> I googled "Czech Republic financial crisis" and the very first result is from Forbes entitled ["The Stunning Failure of the Czech Republic's Austerity Experiment"]( [URL] /). Needless to say I laughed. [NEWLINE] [NEWLINE] <mask> back to your original point, there is wide agreement that, from a cultural perspective, Greek economic policy needs to change. Everyone agrees that they need to stop living beyond their means and actually pay taxes,<mask> living beyond their means is a completely different statement for countries than it is for people. In the context of Greece, having a balanced budget doesn't mean their problems will be solved.<mask><mask>, it's really quite the opposite. A very common Econ 101 question is "<mask> such and such economy runs a policy of a purely balanced budget,<mask> will this affect long term-growth AEE?" In almost every scenario the answer is unequivocally negative. [NEWLINE] [NEWLINE] In the case of Greece, austerity would most directly harm the correction they desperately need to make, namely tax evasion.<mask> the government cuts pensions it means less money available for everyday citizens.<mask><mask><mask>, the people become less inclined to pay taxes<mask> tax evasion is already somewhat engrained in the culture. This in turn means less money for the government which means they have to cut more funding which means less money for the people, etc. etc. etc. It's a cycle of negativity. [NEWLINE] [NEWLINE] Furthermore, I'm left to question your understanding of the issue at hand a bit<mask> you seem to think the Greek people want Germany and France to give them money.<mask><mask>, it is quite the opposite. One of the things the vote showed is that they no longer want to rely on funding from the EU powerhouses and would rather go it alone.<mask> this means for the preexisting Greek debt to those countries I don't know, and I don't think most economists know either.</s>
Label encoding: <s>I think you're oversimplifying the issue. A simple comparison of "it work for the Czechs" or "it worked for the Estonians" doesn't work because they have drastically different economies and industries which complicates things. [NEWLINE] [NEWLINE] Side note: I don't know much about the issues the Czech Republic and Estonia had so I googled "Czech Republic financial crisis" and the very first result is from Forbes entitled ["The Stunning Failure of the Czech Republic's Austerity Experiment"]( [URL] /). Needless to say I laughed. [NEWLINE] [NEWLINE] But back to your original point, there is wide agreement that, from a cultural perspective, Greek economic policy needs to change. Everyone agrees that they need to stop living beyond their means and actually pay taxes, but living beyond their means is a completely different statement for countries than it is for people. In the context of Greece, having a balanced budget doesn't mean their problems will be solved. In fact, it's really quite the opposite. A very common Econ 101 question is " if such and such economy runs a policy of a purely balanced budget, how will this affect long term-growth AEE?" In almost every scenario the answer is unequivocally negative. [NEWLINE] [NEWLINE] In the case of Greece, austerity would most directly harm the correction they desperately need to make, namely tax evasion. If the government cuts pensions it means less money available for everyday citizens. As a result, the people become less inclined to pay taxes as tax evasion is already somewhat engrained in the culture. This in turn means less money for the government which means they have to cut more funding which means less money for the people, etc. etc. etc. It's a cycle of negativity. [NEWLINE] [NEWLINE] Furthermore, I'm left to question your understanding of the issue at hand a bit as you seem to think the Greek people want Germany and France to give them money. In fact, it is quite the opposite. One of the things the vote showed is that they no longer want to rely on funding from the EU powerhouses and would rather go it alone. What this means for the preexisting Greek debt to those countries I don't know, and I don't think most economists know either.</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> trade and relations are vital to most countries survival. [ENDQ] [NEWLINE] This is an opinion.  Many countries are protectionist of certain industries to incubate them, and gain competitive advantage in them.  Most countries are protectionist of their food industries,<mask> that's kind of common sense.  Not all globalism is good.  The last decade or two have shown some of its major drawbacks. [NEWLINE] [NEWLINE] [STARTQ] <mask> then the solution would be to have more referendums, not to completely overhaul the entire system. [ENDQ] [NEWLINE] I was agreeing that it might work... It might<mask> not.  It would depend on its implementation. <mask><mask> you're trying to grow tree in an empty pool filled with dirt, eventually it will just hit the edges of the pool and kill itself with its own roots. [NEWLINE] [NEWLINE] [STARTQ] And this brings up the danger of people having no clue about these issues. [ENDQ] [NEWLINE] You could require people to take a quiz before voting. <mask><mask> we have<mask> little connection to a country that nobody knows anything about it, maybe we shouldn't be getting involved and acting like they're under our sphere of control. <mask> the average person knows nothing, do you really think our politicians know enough for any good to come from it? [NEWLINE] [NEWLINE] [STARTQ] That's a vastly oversimplified reason. [ENDQ] [NEWLINE] OK, well all I'm saying is<mask> Russia was bombing my state and everyone in my state knew someone that those bombs had killed, people would start to get pretty pissed at Russia. [NEWLINE] [NEWLINE] [STARTQ] It turns out that even<mask> you don't want it, many others might. Before the start of the Iraq war, between 47-60% of people supported an invasion. [ENDQ] [NEWLINE] Sorry,<mask><mask><mask> there should be a higher threshold than we have now for causing people to murder each other.  Didn't they teach you in school that violence doesn't solve anything?  It turns out, that's actually pretty true.  It pretty much always results in more escalation.  Unless we really are trying to build an empire like every other country seems to think (which is<mask> they're trying to stop us), then we should CHILL THE FUCK OUT WITH THE BOMBS!!!</s>
Label encoding: <s> [STARTQ] since trade and relations are vital to most countries survival. [ENDQ] [NEWLINE] This is an opinion.  Many countries are protectionist of certain industries to incubate them, and gain competitive advantage in them.  Most countries are protectionist of their food industries, because that's kind of common sense.  Not all globalism is good.  The last decade or two have shown some of its major drawbacks. [NEWLINE] [NEWLINE] [STARTQ] So then the solution would be to have more referendums, not to completely overhaul the entire system. [ENDQ] [NEWLINE] I was agreeing that it might work... It might also not.  It would depend on its implementation.  But if you're trying to grow tree in an empty pool filled with dirt, eventually it will just hit the edges of the pool and kill itself with its own roots. [NEWLINE] [NEWLINE] [STARTQ] And this brings up the danger of people having no clue about these issues. [ENDQ] [NEWLINE] You could require people to take a quiz before voting.  But if we have so little connection to a country that nobody knows anything about it, maybe we shouldn't be getting involved and acting like they're under our sphere of control.  If the average person knows nothing, do you really think our politicians know enough for any good to come from it? [NEWLINE] [NEWLINE] [STARTQ] That's a vastly oversimplified reason. [ENDQ] [NEWLINE] OK, well all I'm saying is if Russia was bombing my state and everyone in my state knew someone that those bombs had killed, people would start to get pretty pissed at Russia. [NEWLINE] [NEWLINE] [STARTQ] It turns out that even if you don't want it, many others might. Before the start of the Iraq war, between 47-60% of people supported an invasion. [ENDQ] [NEWLINE] Sorry, but I think there should be a higher threshold than we have now for causing people to murder each other.  Didn't they teach you in school that violence doesn't solve anything?  It turns out, that's actually pretty true.  It pretty much always results in more escalation.  Unless we really are trying to build an empire like every other country seems to think (which is why they're trying to stop us), then we should CHILL THE FUCK OUT WITH THE BOMBS!!!</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"This is kind of comparable to the whole "like a girl" thing.<mask> a guy says you "run like a girl", they don't mean to insult all girls.<mask><mask> a girl, growing up in a world<mask> you are basically the example of "not good enough", it can take a toll on you and make you feel<mask><mask> you actually aren't good enough." [NEWLINE] [NEWLINE] <mask> that sort of thing does<mask> take a toll on girls and woman, that is unfortunate. I sympathize with them greatly.<mask> people need to be a bit more analytical with<mask> language is used. The phrase "run like a girl," and other phrases like that, are,<mask> you mentioned, not meant to be taken literally. Those feelings of sadness come<mask><mask><mask> of taking the expression literally,<mask> we need to explain that they are not supposed to be taken literally. [NEWLINE] [NEWLINE] "The thing is, nowadays "gay" is used to describe anything that is bad.<mask> your lamp stops working "that lamp is<mask> gay".<mask> you computer crashes, "that's<mask> gay". You are using it to replace words like "stupid" and "useless". I consider m self a conservative and I'm not a big fan of political correctness,<mask> come on. That is pretty insulting." [NEWLINE] [NEWLINE] I understand your view,<mask> using the word Gay to describe something<mask> bad is simply using a separate definition of Gay. For the reasons you stated above,<mask>, it is pretty obvious using the word gay<mask> an insult is not very productive. I would encourage refraining from using the word that way,<mask> that doesn't mean it is immoral. [NEWLINE] [NEWLINE] "<mask> just<mask> they don't mean any harm doesn't mean it isn't hurtful to others." [NEWLINE] [NEWLINE] Agreed. It can be hurtful,<mask><mask><mask> it would be better<mask> we didn't use it<mask> an insult.<mask> I<mask> wouldn't classify it<mask> morally wrong to continue using the word<mask> an insult,<mask><mask> people misunderstand<mask> you're trying to say, that shouldn't be on you. I actually thought your comment was very well thought it, I just disagree with portions of it.</s><pad><pad>
Label encoding: <s>"This is kind of comparable to the whole "like a girl" thing. When a guy says you "run like a girl", they don't mean to insult all girls. But as a girl, growing up in a world where you are basically the example of "not good enough", it can take a toll on you and make you feel as though you actually aren't good enough." [NEWLINE] [NEWLINE] If that sort of thing does indeed take a toll on girls and woman, that is unfortunate. I sympathize with them greatly. But people need to be a bit more analytical with how language is used. The phrase "run like a girl," and other phrases like that, are, as you mentioned, not meant to be taken literally. Those feelings of sadness come as a result of taking the expression literally, so we need to explain that they are not supposed to be taken literally. [NEWLINE] [NEWLINE] "The thing is, nowadays "gay" is used to describe anything that is bad. If your lamp stops working "that lamp is so gay". If you computer crashes, "that's so gay". You are using it to replace words like "stupid" and "useless". I consider m self a conservative and I'm not a big fan of political correctness, but come on. That is pretty insulting." [NEWLINE] [NEWLINE] I understand your view, but using the word Gay to describe something as bad is simply using a separate definition of Gay. For the reasons you stated above, though, it is pretty obvious using the word gay as an insult is not very productive. I would encourage refraining from using the word that way, but that doesn't mean it is immoral. [NEWLINE] [NEWLINE] " But just because they don't mean any harm doesn't mean it isn't hurtful to others." [NEWLINE] [NEWLINE] Agreed. It can be hurtful, so I think it would be better if we didn't use it as an insult. But I also wouldn't classify it as morally wrong to continue using the word as an insult, because if people misunderstand what you're trying to say, that shouldn't be on you. I actually thought your comment was very well thought it, I just disagree with portions of it.</s><pad><pad>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Both "Jesus" and "aliens" are emotional beliefs, and things people believe are real. [ENDQ] [NEWLINE] <mask> I don't get discriminated, unfortunately I have to begin by saying that I don't believe in Jesus and all that taboos that would disqualify me from having a voice among the completely rational and open-minded. [NEWLINE] [NEWLINE] A couple of things: First, all beliefs and most thoughts are "emotional". Very little,<mask> anything, is based on empiricism and most things cannot be (most things aren't testable, repeatable, falsifiable, peer-reviewed, etc.) and things thought to be empirical, like belief in "logic and reason", is known<mask> [naive realism.]( [URL] %C3%AFve_realism_%28psychology%29) There are huge bodies of work dedicated to this in neuroscience, behaviorism, and epistemology--<mask> you can do away with the idea that some things are "true" on a dichotomy and realize that it's<mask><mask> a continuum that you are *perceiving* and not calculating or cataloging. [NEWLINE] [NEWLINE] <mask> for your view, the reason religions are protected is twofold: (1) they're large and powerful groups capable of defending themselves, (2) they have rivalries. [NEWLINE] [NEWLINE] <mask> a government wishing to govern wants to prevent these groups from turning inward for protection, gaining support, and becoming like gangs within society which fight. Radical Islam is an example of this,<mask> terrorist activities are meant to polarize cultures and governments against Muslims,<mask> that Muslims will turn to Radical Islam for support. Governments *don't* demonize Muslims<mask><mask><mask> and have various methods for diffusing the situation (labeling "terrorists", focusing on the victims, responding with excessive police presences to give a sense of order, etc.). [NEWLINE] [NEWLINE] <mask><mask> car guys became a huge and powerful group that were at ends with farmers for some bizarre reason, the government would produce a way to diffuse conflict before it starts with anti-discrimination laws,<mask> they don't turn inward. [NEWLINE] [NEWLINE] That's<mask> the turban can and must get a pass.</s>
Label encoding: <s> [STARTQ] Both "Jesus" and "aliens" are emotional beliefs, and things people believe are real. [ENDQ] [NEWLINE] So I don't get discriminated, unfortunately I have to begin by saying that I don't believe in Jesus and all that taboos that would disqualify me from having a voice among the completely rational and open-minded. [NEWLINE] [NEWLINE] A couple of things: First, all beliefs and most thoughts are "emotional". Very little, if anything, is based on empiricism and most things cannot be (most things aren't testable, repeatable, falsifiable, peer-reviewed, etc.) and things thought to be empirical, like belief in "logic and reason", is known as [naive realism.]( [URL] %C3%AFve_realism_%28psychology%29) There are huge bodies of work dedicated to this in neuroscience, behaviorism, and epistemology-- so you can do away with the idea that some things are "true" on a dichotomy and realize that it's in fact a continuum that you are *perceiving* and not calculating or cataloging. [NEWLINE] [NEWLINE] As for your view, the reason religions are protected is twofold: (1) they're large and powerful groups capable of defending themselves, (2) they have rivalries. [NEWLINE] [NEWLINE] Therefore a government wishing to govern wants to prevent these groups from turning inward for protection, gaining support, and becoming like gangs within society which fight. Radical Islam is an example of this, as terrorist activities are meant to polarize cultures and governments against Muslims, so that Muslims will turn to Radical Islam for support. Governments *don't* demonize Muslims for this reason and have various methods for diffusing the situation (labeling "terrorists", focusing on the victims, responding with excessive police presences to give a sense of order, etc.). [NEWLINE] [NEWLINE] So if car guys became a huge and powerful group that were at ends with farmers for some bizarre reason, the government would produce a way to diffuse conflict before it starts with anti-discrimination laws, so they don't turn inward. [NEWLINE] [NEWLINE] That's why the turban can and must get a pass.</s>
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it cannot be manually driven I don't really see<mask> the point<mask> software failures, malfunctions or whatever would mean you can't go anywhere. [ENDQ] [NEWLINE] Just like things on regular cars can malfunction<mask> well to the point that the car can't be driven. [NEWLINE] [NEWLINE] [STARTQ] It seems like google is pumping a fair amount of money into this<mask> there are more important issues. Say<mask> they had focused on inventing a zero carbon emission car or whatever. [ENDQ] [NEWLINE] Google is a software company, meaning that the expertise of the company itself isn't to create a machine with lower emissions<mask> rather to create software which has various uses. Other car companies are working on creating zero-emission cars which one day may be able to use the software that Google creates<mask> that we can have zero-emitting self-driving cars. Saying that Google should be working on a zero-emission car instead of self-driving software is like saying that an electrician should fix your pipes<mask> a plumber rewires your house. [NEWLINE] [NEWLINE] [STARTQ] It just doesn't seem like something we need. [ENDQ] [NEWLINE] <mask> should all companies that make things that we don't need stop their production of these things? For example, Sharp makes TVs. We really don't need TVs,<mask> should Sharp stop making them? Should Bose stop making headphones<mask> we really don't need them? [NEWLINE] [NEWLINE] Aside from these points, there are various positives to self-driving cars. Once the software is perfected to the point that it is ready for consumers to purchase and its use is widespread, we will most likely see a decrease in car accidents<mask><mask><mask> of driver error. You won't have to worry about someone falling asleep at the wheel, someone getting road rage, or drunk drivers<mask> cars will be driving themselves. Aside from this,<mask> every car is connected through a computer software, each car can take different routes to the same destinations to lower traffic. For example, instead of having everyone cram onto a highway during rush hour going in or out of a major city, self-driving cars can all determine<mask> the best possible route is for them based on the routes of other self-driving cars.</s>
Label encoding: <s> [STARTQ] if it cannot be manually driven I don't really see why the point as software failures, malfunctions or whatever would mean you can't go anywhere. [ENDQ] [NEWLINE] Just like things on regular cars can malfunction as well to the point that the car can't be driven. [NEWLINE] [NEWLINE] [STARTQ] It seems like google is pumping a fair amount of money into this when there are more important issues. Say if they had focused on inventing a zero carbon emission car or whatever. [ENDQ] [NEWLINE] Google is a software company, meaning that the expertise of the company itself isn't to create a machine with lower emissions but rather to create software which has various uses. Other car companies are working on creating zero-emission cars which one day may be able to use the software that Google creates so that we can have zero-emitting self-driving cars. Saying that Google should be working on a zero-emission car instead of self-driving software is like saying that an electrician should fix your pipes while a plumber rewires your house. [NEWLINE] [NEWLINE] [STARTQ] It just doesn't seem like something we need. [ENDQ] [NEWLINE] So should all companies that make things that we don't need stop their production of these things? For example, Sharp makes TVs. We really don't need TVs, so should Sharp stop making them? Should Bose stop making headphones since we really don't need them? [NEWLINE] [NEWLINE] Aside from these points, there are various positives to self-driving cars. Once the software is perfected to the point that it is ready for consumers to purchase and its use is widespread, we will most likely see a decrease in car accidents as a result of driver error. You won't have to worry about someone falling asleep at the wheel, someone getting road rage, or drunk drivers since cars will be driving themselves. Aside from this, if every car is connected through a computer software, each car can take different routes to the same destinations to lower traffic. For example, instead of having everyone cram onto a highway during rush hour going in or out of a major city, self-driving cars can all determine what the best possible route is for them based on the routes of other self-driving cars.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the tradition of "going to the movies" will soon fall out of favor [ENDQ] [NEWLINE] <mask>  you want that to change (and be a "world changer"), then you need to focus on the movie-going experience rather than the artistic side of movie making.  "Going to the moves" is more of a hassle than it is worth.  Honestly, the last movie that I was excited about seeing in a movie theatre on a big screen was Independence Day. [NEWLINE] [NEWLINE] Now I can watch movies in the privacy of my own home, with a better picture and sound quality than in a theatre, on my own schedule, without having annoying people all around me.  And I can do it for considerably less money. [NEWLINE] [NEWLINE] I don't know<mask> *anyone* enjoys going to the movies anymore.  It is stupid expensive ($50 for a couple after buying snacks and sodas).  You feel like you need a shower after leaving<mask> who knows<mask> kind of nasty unshowered person had that seat before you.  The at-home experience is *<mask> much* more enjoyable and cheaper.  The only thing theatres have going for them is you get the new releases there first.  I can wait 3 months for the bluray to come out; that's not a problem. [NEWLINE] [NEWLINE] <mask> you can still be a film maker.  You just need to change your expectations of<mask> your film will be seen and gear your product toward that audience. <mask> young, single people may not get together and watch a movie at home (like they go out<mask> a group to the theatre), virtually every other demographic does:  Families have "family move night" and pop popcorn and watch a move.  Couples snuggle up on the couch and watch a movie.  Old people turn on a move and go to sleep. [NEWLINE] [NEWLINE] People want to be entertained (pause for a moment and think about<mask> much of the US economy is based upon entertainment). <mask> you produce a product that entertains people, there is going to be a market for it; even<mask> the path to that market isn't the same path that gets to it today.</s>
Label encoding: <s> [STARTQ] the tradition of "going to the movies" will soon fall out of favor [ENDQ] [NEWLINE] If  you want that to change (and be a "world changer"), then you need to focus on the movie-going experience rather than the artistic side of movie making.  "Going to the moves" is more of a hassle than it is worth.  Honestly, the last movie that I was excited about seeing in a movie theatre on a big screen was Independence Day. [NEWLINE] [NEWLINE] Now I can watch movies in the privacy of my own home, with a better picture and sound quality than in a theatre, on my own schedule, without having annoying people all around me.  And I can do it for considerably less money. [NEWLINE] [NEWLINE] I don't know why *anyone* enjoys going to the movies anymore.  It is stupid expensive ($50 for a couple after buying snacks and sodas).  You feel like you need a shower after leaving because who knows what kind of nasty unshowered person had that seat before you.  The at-home experience is * so much* more enjoyable and cheaper.  The only thing theatres have going for them is you get the new releases there first.  I can wait 3 months for the bluray to come out; that's not a problem. [NEWLINE] [NEWLINE] So you can still be a film maker.  You just need to change your expectations of where your film will be seen and gear your product toward that audience.  While young, single people may not get together and watch a movie at home (like they go out as a group to the theatre), virtually every other demographic does:  Families have "family move night" and pop popcorn and watch a move.  Couples snuggle up on the couch and watch a movie.  Old people turn on a move and go to sleep. [NEWLINE] [NEWLINE] People want to be entertained (pause for a moment and think about how much of the US economy is based upon entertainment).  If you produce a product that entertains people, there is going to be a market for it; even if the path to that market isn't the same path that gets to it today.</s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Reddit has a lot of *really goddamn shitty* anti-women views that<mask> not universal are fairly common. One of the experiences I had going to university was that a lot of my worldviews were challenged very strongly and vocally, and more importantly I learned that some shit I believed in was<mask> not okay that I shouldn't even articulate it unless I'm *really* prepared to defend it.<mask> forced to defend casual statements that people called me out on, I was forced to re-evaluate my views and through this grew and matured<mask> a person. [NEWLINE] [NEWLINE] I can't help<mask> notice that people who I interact with who are under 20 or<mask> and frequent reddit enough to mention it in public for some reason seem to not only hold some pretty strongly anti-woman views ("feminism is evil" without having a tiny grasp of<mask> feminism is, the "friendzone" is a real thing and not an awful manipulative concept, etc.) which they can't even begin to articulate a defence of<mask> called on.<mask><mask> that tying opinions like this consequence-free to an anonymous internet points community does nothing more than validate these opinions without encouraging critical thinking about them at all, and that to a lot of younger members on this site they're lead to think that these extreme views are the norm and are acceptable. [NEWLINE] [NEWLINE] I do think SRS goes overboard sometimes, I do think that there's an extent to which they miss that sometimes jokes do make fun of people and that's an uphill fight you'll never win.<mask> I<mask> have a huge problem with the idea that they should be totally hands off. Maybe<mask> people want their internet points they should have to contend fairly with gangs of people who, oh, say, think joking about raping babies should be frowned upon in a public place. It's not a freedom of speech issue, it seems like the rule is set up for freedom from consequence of speech<mask> the people who would normally most strongly object are already frequenting the same part of the site and<mask> most likely to be perceived<mask> "vote brigading" and<mask> their voice is basically cut off. [NEWLINE] [NEWLINE] edit: these downvotes are definitely changing my view.</s>
Label encoding: <s>Reddit has a lot of *really goddamn shitty* anti-women views that while not universal are fairly common. One of the experiences I had going to university was that a lot of my worldviews were challenged very strongly and vocally, and more importantly I learned that some shit I believed in was so not okay that I shouldn't even articulate it unless I'm *really* prepared to defend it. When forced to defend casual statements that people called me out on, I was forced to re-evaluate my views and through this grew and matured as a person. [NEWLINE] [NEWLINE] I can't help but notice that people who I interact with who are under 20 or so and frequent reddit enough to mention it in public for some reason seem to not only hold some pretty strongly anti-woman views ("feminism is evil" without having a tiny grasp of what feminism is, the "friendzone" is a real thing and not an awful manipulative concept, etc.) which they can't even begin to articulate a defence of when called on. I think that tying opinions like this consequence-free to an anonymous internet points community does nothing more than validate these opinions without encouraging critical thinking about them at all, and that to a lot of younger members on this site they're lead to think that these extreme views are the norm and are acceptable. [NEWLINE] [NEWLINE] I do think SRS goes overboard sometimes, I do think that there's an extent to which they miss that sometimes jokes do make fun of people and that's an uphill fight you'll never win. But I also have a huge problem with the idea that they should be totally hands off. Maybe if people want their internet points they should have to contend fairly with gangs of people who, oh, say, think joking about raping babies should be frowned upon in a public place. It's not a freedom of speech issue, it seems like the rule is set up for freedom from consequence of speech where the people who would normally most strongly object are already frequenting the same part of the site and thus most likely to be perceived as "vote brigading" and thus their voice is basically cut off. [NEWLINE] [NEWLINE] edit: these downvotes are definitely changing my view.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you think people who are ugly should be shamed for being ugly? [NEWLINE] [NEWLINE] More to the point:<mask> *shouldn't* being fat be accepted?  Sure there are *some* health risks being associated with being morbidly obese,<mask> slightly overweight?  And<mask> would health risks even matter? <mask> do you care about someone's health, and<mask> would that give you the right to *shame* them? [NEWLINE] [NEWLINE] [STARTQ] For the majority of overweight and obese people, it isn't medical factors controlling it, it's weak self control. [ENDQ] [NEWLINE] I don't know<mask> I buy this.  I'd imagine that for *the majority* of overweight people, it's having a unfortunately low metabolism in a culture<mask> the cheapest and most readily available food are incredibly unhealthy. <mask> this, you can't possibly know, for any given fat person, whether they are<mask><mask> of<mask> they were born or<mask> of<mask> they eat. [NEWLINE] [NEWLINE] [STARTQ] And I wouldn't have gained the self control<mask> it weren't for a little teasing/bullying<mask> a kid. [ENDQ] [NEWLINE] For most people, shaming them has the opposite effect, and can even drive them to suicide.  There are far better motivational forces than shame. [NEWLINE] [NEWLINE] [STARTQ] <mask> shouldn't we be able to shame fat people? I'm not saying necessarily to their face, that's just dickish to do to anyone really,<mask><mask> should being fat be accepted? [ENDQ] [NEWLINE] Probably<mask> someone being fat affects you *zero*.  And really, it's asymmetrical judgement,<mask> nowhere near the amount of shame levied towards fat people are levied towards skinny or underweight people.  Western cultural standards would rather have someone be unhealthily underweight than unhealthily overweight, or overweight at all. [NEWLINE] [NEWLINE] Finally, your penultimate paragraph is irrelevant.  Beauty isn't an objective measurement, and you can't decide that<mask> *you* find someone's body unappealing that they *are* unappealing. <mask> you see a "grossly misshapen" body, someone else might see a beautiful one.  And<mask>,<mask> do you even care?</s>
Label encoding: <s>Do you think people who are ugly should be shamed for being ugly? [NEWLINE] [NEWLINE] More to the point: why *shouldn't* being fat be accepted?  Sure there are *some* health risks being associated with being morbidly obese, but slightly overweight?  And why would health risks even matter?  Why do you care about someone's health, and why would that give you the right to *shame* them? [NEWLINE] [NEWLINE] [STARTQ] For the majority of overweight and obese people, it isn't medical factors controlling it, it's weak self control. [ENDQ] [NEWLINE] I don't know if I buy this.  I'd imagine that for *the majority* of overweight people, it's having a unfortunately low metabolism in a culture where the cheapest and most readily available food are incredibly unhealthy.  Besides this, you can't possibly know, for any given fat person, whether they are so because of how they were born or because of how they eat. [NEWLINE] [NEWLINE] [STARTQ] And I wouldn't have gained the self control if it weren't for a little teasing/bullying as a kid. [ENDQ] [NEWLINE] For most people, shaming them has the opposite effect, and can even drive them to suicide.  There are far better motivational forces than shame. [NEWLINE] [NEWLINE] [STARTQ] Why shouldn't we be able to shame fat people? I'm not saying necessarily to their face, that's just dickish to do to anyone really, but why should being fat be accepted? [ENDQ] [NEWLINE] Probably because someone being fat affects you *zero*.  And really, it's asymmetrical judgement, since nowhere near the amount of shame levied towards fat people are levied towards skinny or underweight people.  Western cultural standards would rather have someone be unhealthily underweight than unhealthily overweight, or overweight at all. [NEWLINE] [NEWLINE] Finally, your penultimate paragraph is irrelevant.  Beauty isn't an objective measurement, and you can't decide that because *you* find someone's body unappealing that they *are* unappealing.  Where you see a "grossly misshapen" body, someone else might see a beautiful one.  And besides, why do you even care?</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok I'll answer point two first. The reason piracy doesn't hurt something like game of thrones is<mask> it increases exposure. Something like that lives on<mask> many people are watching it and piracy gets the word out by having more people talk about it and getting others who are willing to pay or unable to pirate to watch it.<mask> it may be surprising a very low percentage of the population even knows<mask> to pirate things. It seems like a lot more<mask><mask> you're on this site you're probably more tech oriented than a vast majority of the population. [NEWLINE] [NEWLINE] <mask> for your first point, piracy has been around for quite a<mask>,<mask> the biggest blow that has been landed to dvd sales was by online streaming content. Netflix, Hulu, Amazon Prime Instant. These are the things that have reduced both piracy and dvd sales, its got the data that shows people are willing to pay for media<mask> its a decent price for<mask> they're receiving. The huge success of them has shown that its a very good option. The reason<mask> execs haven't just gone to that type of model<mask> is it reduces their profit margin. It might end up being more profitable for them to go to that type of model,<mask> they see less money per person, not looking at<mask> many more people they could have by doing<mask>. Instead they stay with<mask> they know and keep the profits they can expect rather than potentially risking money for a larger reward based off of quantity over the price of the sale. [NEWLINE] [NEWLINE] By trying to hold prices hostage they alienate a lot of people who either don't have the disposable income to buy movies like that or just think that the price is too high for the content. Netflix has really helped out the indie community by actually making it somewhat easy<mask> an indie producer to get your movie seen by people. Before it was near impossible unless you have a really explosive hit to get much outside a festival such<mask> sundance. [NEWLINE] [NEWLINE] The market is shifting. Piracy rates show that people are unhappy and unwilling to stick with prior business models and that companies are going to have to change to the times<mask> they really want to stay in business.</s>
Label encoding: <s>Ok I'll answer point two first. The reason piracy doesn't hurt something like game of thrones is because it increases exposure. Something like that lives on how many people are watching it and piracy gets the word out by having more people talk about it and getting others who are willing to pay or unable to pirate to watch it. While it may be surprising a very low percentage of the population even knows how to pirate things. It seems like a lot more because if you're on this site you're probably more tech oriented than a vast majority of the population. [NEWLINE] [NEWLINE] As for your first point, piracy has been around for quite a while, but the biggest blow that has been landed to dvd sales was by online streaming content. Netflix, Hulu, Amazon Prime Instant. These are the things that have reduced both piracy and dvd sales, its got the data that shows people are willing to pay for media if its a decent price for what they're receiving. The huge success of them has shown that its a very good option. The reason why execs haven't just gone to that type of model though is it reduces their profit margin. It might end up being more profitable for them to go to that type of model, but they see less money per person, not looking at how many more people they could have by doing so. Instead they stay with what they know and keep the profits they can expect rather than potentially risking money for a larger reward based off of quantity over the price of the sale. [NEWLINE] [NEWLINE] By trying to hold prices hostage they alienate a lot of people who either don't have the disposable income to buy movies like that or just think that the price is too high for the content. Netflix has really helped out the indie community by actually making it somewhat easy as an indie producer to get your movie seen by people. Before it was near impossible unless you have a really explosive hit to get much outside a festival such as sundance. [NEWLINE] [NEWLINE] The market is shifting. Piracy rates show that people are unhappy and unwilling to stick with prior business models and that companies are going to have to change to the times if they really want to stay in business.</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Allow me to preface my comments by saying I hope you don't take this (or anything else I've written) offensively.  I'm just giving you my thoughts on this topic. [NEWLINE] [NEWLINE] Once again you're over-complicating things by bringing up concepts like neuroscience.  Notwithstanding your own psychology there is a simple fact about each of us (save for hermaphrodites) that we are physically (or biologically which would be interchangeable here) born<mask> either a man or a woman.  No amount of wordsmithing can avoid this plain fact. [NEWLINE] [NEWLINE] That being said some of us (apparently yourself included) have a psychological disorder whereby their *gender identity* does not comport with their *physical form*.  Once again you can have your views about this<mask> realize that I am making no moral judgment against those afflicted<mask> I say psychological disorder.  Taking your own view concerning the nature of this "feeling" it is plain and simply a deviation from the norm and a mental disorder.  Some might argue environmental factors may lead to someone developing this "feeling"<mask> others may argue they are entirely biological.  No one can be certain one way or the other,<mask> is often the case<mask> dealing with the workings of our minds. [NEWLINE] [NEWLINE] I was using the term "internal self" in reference to<mask> someone thought about themselves e.g. they are a man inhabiting a woman's body. <mask> I was using the term "external desires" to refer to<mask> someone desired irrespective of their own body e.g. the desire to have intercourse with a particular gender.  To me the difference is that the former involves rejection of ones physical self (through surgery or other means)<mask> the latter involves acceptance of ones true desires (through sex preference).  I would hope people could find happiness without surgery and hormones and doctors<mask> maybe I'm wrong, and I accept that I have no frame of reference here. [NEWLINE] [NEWLINE] In light of that, I still support your decision to "transition" - everyone should be afforded the opportunity to do with their body that which they see fit - even suicide.</s>
Label encoding: <s>Allow me to preface my comments by saying I hope you don't take this (or anything else I've written) offensively.  I'm just giving you my thoughts on this topic. [NEWLINE] [NEWLINE] Once again you're over-complicating things by bringing up concepts like neuroscience.  Notwithstanding your own psychology there is a simple fact about each of us (save for hermaphrodites) that we are physically (or biologically which would be interchangeable here) born as either a man or a woman.  No amount of wordsmithing can avoid this plain fact. [NEWLINE] [NEWLINE] That being said some of us (apparently yourself included) have a psychological disorder whereby their *gender identity* does not comport with their *physical form*.  Once again you can have your views about this but realize that I am making no moral judgment against those afflicted when I say psychological disorder.  Taking your own view concerning the nature of this "feeling" it is plain and simply a deviation from the norm and a mental disorder.  Some might argue environmental factors may lead to someone developing this "feeling" while others may argue they are entirely biological.  No one can be certain one way or the other, as is often the case when dealing with the workings of our minds. [NEWLINE] [NEWLINE] I was using the term "internal self" in reference to how someone thought about themselves e.g. they are a man inhabiting a woman's body.  While I was using the term "external desires" to refer to what someone desired irrespective of their own body e.g. the desire to have intercourse with a particular gender.  To me the difference is that the former involves rejection of ones physical self (through surgery or other means) while the latter involves acceptance of ones true desires (through sex preference).  I would hope people could find happiness without surgery and hormones and doctors but maybe I'm wrong, and I accept that I have no frame of reference here. [NEWLINE] [NEWLINE] In light of that, I still support your decision to "transition" - everyone should be afforded the opportunity to do with their body that which they see fit - even suicide.</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point is not that everyone has to experience America. The point is that it is costly to travel<mask> far<mask><mask> many people never get to do it. [NEWLINE] [NEWLINE] And the cost is no<mask> near<mask> small<mask> you are making it out. First you have to pick vacation time or unpaid time. Time constraints mean you may only get two weeks and 2-4 days will just be a series of plane rides. Then there is the cost of food and shelter and local transport not to mention entry to any sites you may wish to visit. It adds up fast. Plus the exchange rate is not favorable to Americans. [NEWLINE] [NEWLINE] I've been to Europe and plan to return next year. I'm not putting it down. I'm enamored with travel. I budgeted 2k originally and quickly realized it was unrealistic and upped it to 3k.<mask> all was said and done<mask> spent closer to 4k. And that's ok<mask> I can afford it<mask> not everyone can. [NEWLINE] [NEWLINE] Then consider the people who have started a family. Lugging children around Europe isn't necessarily worth it and way more costly. That means 18 years of little to no travel and the. You have to pay for college. Life costs a lot of money. Many people just cannot justify the cost of a short and glossed over Europe trip<mask> seeing the states or Caribbean or someplace closer is more manageable. [NEWLINE] [NEWLINE] Oh and the language barrier means many Americans will have a hard time even planning such a trip let alone doing it. Especially of you're not in your twenties it becomes harder and harder to have a first trip overseas. You can talk about<mask> cheap it *can* be<mask> that requires a lot of sacrifices and it's time consuming to nickel and dime your way across the continent. [NEWLINE] [NEWLINE] Many Americans live paycheck to paycheck. Even your fantasy 2k trip is a pipe dream for millions of Americans. again I'm not trying to discourage people from going<mask> encourage it.<mask> I'm realistic and many<mask> not most Americans simply cannot afford to go or cannot justify the cost compared to cheaper and better local options. </s>
Label encoding: <s>The point is not that everyone has to experience America. The point is that it is costly to travel so far hence why many people never get to do it. [NEWLINE] [NEWLINE] And the cost is no where near as small as you are making it out. First you have to pick vacation time or unpaid time. Time constraints mean you may only get two weeks and 2-4 days will just be a series of plane rides. Then there is the cost of food and shelter and local transport not to mention entry to any sites you may wish to visit. It adds up fast. Plus the exchange rate is not favorable to Americans. [NEWLINE] [NEWLINE] I've been to Europe and plan to return next year. I'm not putting it down. I'm enamored with travel. I budgeted 2k originally and quickly realized it was unrealistic and upped it to 3k. When all was said and done if spent closer to 4k. And that's ok because I can afford it but not everyone can. [NEWLINE] [NEWLINE] Then consider the people who have started a family. Lugging children around Europe isn't necessarily worth it and way more costly. That means 18 years of little to no travel and the. You have to pay for college. Life costs a lot of money. Many people just cannot justify the cost of a short and glossed over Europe trip when seeing the states or Caribbean or someplace closer is more manageable. [NEWLINE] [NEWLINE] Oh and the language barrier means many Americans will have a hard time even planning such a trip let alone doing it. Especially of you're not in your twenties it becomes harder and harder to have a first trip overseas. You can talk about how cheap it *can* be but that requires a lot of sacrifices and it's time consuming to nickel and dime your way across the continent. [NEWLINE] [NEWLINE] Many Americans live paycheck to paycheck. Even your fantasy 2k trip is a pipe dream for millions of Americans. again I'm not trying to discourage people from going if encourage it. But I'm realistic and many if not most Americans simply cannot afford to go or cannot justify the cost compared to cheaper and better local options. </s>
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see human society<mask> a macro level cohesive entity moving along some kind of flux or wave of time and change. Each part of the organism has different functions, many of the humans are valuable for maintaining cohesive structures and getting the stuff done. There are other proportionally smaller groups of humans that contribute other functions, one such group is the group of people with high IQ's, whatever that's supposed to mean. You could<mask> call the visionaries, geniuses, artists, sages, wise men, holy men,<mask> have you. [NEWLINE] [NEWLINE] I must state that all functions of the collective human organism are equally valuable<mask> without one, the whole would simply not be the same collective human organism. [NEWLINE] [NEWLINE] <mask> I had to guess at<mask> the purpose of the high IQ individuals were, they'd sort of be agents of the liminal spaces. My metaphor for this is the cell wall of single celled organisms. The cell wall includes lots of specialized structures that allow certain molecules to pass through to the rest of the cell body and disallow others. This cell wall is necessary for the cell to continue it's life cycle and regulate the metabolic function<mask> the cell. Certainly a very useful thing. [NEWLINE] [NEWLINE] <mask>,<mask> you compare the mass of the cell wall to mass of the rest of the cell, you'd find that there is a whole lot less of the ell wall than any other cell parts. This is<mask> the way it should be, living organisms appear to be very self contained sort of things and it simply is not possible to have most of the living organism in contact with the world. [NEWLINE] [NEWLINE] <mask><mask> with their flexibility of thought and ability to adapt to new situations on the fly, people with high IQ's are almost certainly on that liminal border between collective humanity and the rest of the universe and this is our biological adaptation to the fact that the universe is constantly changing. The people with high IQ's can adapt to respond to this change<mask> maintaining homeostasis for the rest. [NEWLINE] [NEWLINE] I suppose<mask> you don't think that's important then you probably haven't been out to the limits<mask>. </s>
Label encoding: <s>I see human society as a macro level cohesive entity moving along some kind of flux or wave of time and change. Each part of the organism has different functions, many of the humans are valuable for maintaining cohesive structures and getting the stuff done. There are other proportionally smaller groups of humans that contribute other functions, one such group is the group of people with high IQ's, whatever that's supposed to mean. You could also call the visionaries, geniuses, artists, sages, wise men, holy men, what have you. [NEWLINE] [NEWLINE] I must state that all functions of the collective human organism are equally valuable because without one, the whole would simply not be the same collective human organism. [NEWLINE] [NEWLINE] If I had to guess at what the purpose of the high IQ individuals were, they'd sort of be agents of the liminal spaces. My metaphor for this is the cell wall of single celled organisms. The cell wall includes lots of specialized structures that allow certain molecules to pass through to the rest of the cell body and disallow others. This cell wall is necessary for the cell to continue it's life cycle and regulate the metabolic function so the cell. Certainly a very useful thing. [NEWLINE] [NEWLINE] However, if you compare the mass of the cell wall to mass of the rest of the cell, you'd find that there is a whole lot less of the ell wall than any other cell parts. This is as the way it should be, living organisms appear to be very self contained sort of things and it simply is not possible to have most of the living organism in contact with the world. [NEWLINE] [NEWLINE] I think with their flexibility of thought and ability to adapt to new situations on the fly, people with high IQ's are almost certainly on that liminal border between collective humanity and the rest of the universe and this is our biological adaptation to the fact that the universe is constantly changing. The people with high IQ's can adapt to respond to this change while maintaining homeostasis for the rest. [NEWLINE] [NEWLINE] I suppose if you don't think that's important then you probably haven't been out to the limits yet. </s>
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] which shouldn't be valid<mask> youre killing another body...which should have rights. [ENDQ] [NEWLINE] Lets say that the fetus should have rights.<mask> should the fetus get to extend his/her will over their mother's?<mask> do the mother's rights suddenly not matter? [NEWLINE] [NEWLINE] <mask> now the question is,<mask> should the fetus have rights to begin with? It's neither aware nor independent- meaning it needs the mother to survive (up until week 24 or<mask>,<mask> it has a 50% survival rate outside the womb) [NEWLINE] [NEWLINE] [STARTQ] Then there is the justifications, rape, disease, etc.<mask><mask> in some of those cases. [ENDQ] [NEWLINE] <mask> should these be justified? Doesn't the fetus have rights that extend over the mother's?<mask> would these situations make it any different? [NEWLINE] [NEWLINE] [STARTQ] The lack of responsibility should not result in a life being snuffed out [ENDQ] [NEWLINE] <mask> you want the government to tell people that they can't have sex unless they are looking to reproduce? or have some sort of birth control (which can fail from time to time)? [NEWLINE] [NEWLINE] ------------------------------------------------------------------ [NEWLINE] [NEWLINE] Ultimately, your argument boils down to "fetuses are alive and have rights" [NEWLINE] [NEWLINE] Unless I am wrong,<mask> do you believe that? [NEWLINE] [NEWLINE] From my perspective, fetuses can be considered a type of parasite<mask> they are unwanted- they grow inside or on a host, steal nutrients, and can cause significant pain and suffering along with other health problems. [NEWLINE] [NEWLINE] Furthermore,<mask> should the rights of the mother suddenly not matter? She's the one that has to lug around the fetus for 9 months and<mask> she doesn't want to,<mask> force her to? [NEWLINE] [NEWLINE] And even furthermore,<mask> do you consider a fetus "alive" in the traditional sense? A fetus isn't even viable until around the end of the second trimester (24 weeks ish), and that's at only 50%. For nearly 100% viability, you'd have to be closing on 28-30 weeks.<mask><mask> we can agree that the vast majority of abortions occur significantly before the 24 week mark.</s>
Label encoding: <s> [STARTQ] which shouldn't be valid because youre killing another body...which should have rights. [ENDQ] [NEWLINE] Lets say that the fetus should have rights. Why should the fetus get to extend his/her will over their mother's? Why do the mother's rights suddenly not matter? [NEWLINE] [NEWLINE] But now the question is, why should the fetus have rights to begin with? It's neither aware nor independent- meaning it needs the mother to survive (up until week 24 or so, where it has a 50% survival rate outside the womb) [NEWLINE] [NEWLINE] [STARTQ] Then there is the justifications, rape, disease, etc. I agree in some of those cases. [ENDQ] [NEWLINE] why should these be justified? Doesn't the fetus have rights that extend over the mother's? Why would these situations make it any different? [NEWLINE] [NEWLINE] [STARTQ] The lack of responsibility should not result in a life being snuffed out [ENDQ] [NEWLINE] so you want the government to tell people that they can't have sex unless they are looking to reproduce? or have some sort of birth control (which can fail from time to time)? [NEWLINE] [NEWLINE] ------------------------------------------------------------------ [NEWLINE] [NEWLINE] Ultimately, your argument boils down to "fetuses are alive and have rights" [NEWLINE] [NEWLINE] Unless I am wrong, why do you believe that? [NEWLINE] [NEWLINE] From my perspective, fetuses can be considered a type of parasite if they are unwanted- they grow inside or on a host, steal nutrients, and can cause significant pain and suffering along with other health problems. [NEWLINE] [NEWLINE] Furthermore, why should the rights of the mother suddenly not matter? She's the one that has to lug around the fetus for 9 months and if she doesn't want to, why force her to? [NEWLINE] [NEWLINE] And even furthermore, why do you consider a fetus "alive" in the traditional sense? A fetus isn't even viable until around the end of the second trimester (24 weeks ish), and that's at only 50%. For nearly 100% viability, you'd have to be closing on 28-30 weeks. I think we can agree that the vast majority of abortions occur significantly before the 24 week mark.</s>
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Theft: copying something can't possibly be theft<mask> theft requires taking something away and copying doesn't do that. [ENDQ] [NEWLINE] Thinking about IP violations<mask> theft (defined<mask> depriving someone of their rightfully owned property) is, in my mind, a wrong way to go about it. Ownership of something doesn't just mean you have possession of it,<mask> that you have the right to *control its use.* We talk about violations of IP<mask> "theft"<mask> in reality, it is closer to a violation of the right to control the property. [NEWLINE] [NEWLINE] <mask><mask> a person has a patent on a widget, he has the right to license and control the use of that widget.<mask> he doesn't think it should be used in, say, weapons manufacturing- he has the right to deny the use of his IP in that field. [NEWLINE] [NEWLINE] Now say that a weapons company ignores this and uses the widget anyway. He has been wronged in two senses: first by loss of profits from the use of his widget (wherein the thing stolen is the *profit*, not the widget itself). Second, in the *control of the use* of the widget. The weapons manufacturer didn't deprive the inventor of *possession,*<mask> he did deprive the inventor of *control.* [NEWLINE] [NEWLINE] <mask><mask> it may not be theft of the property (on your definition of theft), it *is still a violation of property rights.* [NEWLINE] [NEWLINE] Edit- To the economic utility point: I'm not sure I get your point. The argument in favor of IP is that IP will not be developed without protection for use and profit. This seems clearly correct: say a corporation could steal any patent from any inventor and deprive the inventor of profit or control without fear of reprimand.<mask> incentive would the inventor have to devote time to inventing?<mask> incentive would any person have to do research,<mask> they would not be able to profit from the research. [NEWLINE] [NEWLINE] [STARTQ] <mask> we do know for sure,<mask> a prohibition on commercial copying is implemented, is that individual rights will be violated. [ENDQ] [NEWLINE] <mask>?</s>
Label encoding: <s> [STARTQ] Theft: copying something can't possibly be theft because theft requires taking something away and copying doesn't do that. [ENDQ] [NEWLINE] Thinking about IP violations as theft (defined as depriving someone of their rightfully owned property) is, in my mind, a wrong way to go about it. Ownership of something doesn't just mean you have possession of it, but that you have the right to *control its use.* We talk about violations of IP as "theft" when in reality, it is closer to a violation of the right to control the property. [NEWLINE] [NEWLINE] So if a person has a patent on a widget, he has the right to license and control the use of that widget. If he doesn't think it should be used in, say, weapons manufacturing- he has the right to deny the use of his IP in that field. [NEWLINE] [NEWLINE] Now say that a weapons company ignores this and uses the widget anyway. He has been wronged in two senses: first by loss of profits from the use of his widget (wherein the thing stolen is the *profit*, not the widget itself). Second, in the *control of the use* of the widget. The weapons manufacturer didn't deprive the inventor of *possession,* but he did deprive the inventor of *control.* [NEWLINE] [NEWLINE] So while it may not be theft of the property (on your definition of theft), it *is still a violation of property rights.* [NEWLINE] [NEWLINE] Edit- To the economic utility point: I'm not sure I get your point. The argument in favor of IP is that IP will not be developed without protection for use and profit. This seems clearly correct: say a corporation could steal any patent from any inventor and deprive the inventor of profit or control without fear of reprimand. What incentive would the inventor have to devote time to inventing? What incentive would any person have to do research, when they would not be able to profit from the research. [NEWLINE] [NEWLINE] [STARTQ] What we do know for sure, if a prohibition on commercial copying is implemented, is that individual rights will be violated. [ENDQ] [NEWLINE] How?</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Note that the FT does not state that the second law of thermodynamics is wrong or invalid. The second law of thermodynamics is a statement about macroscopic systems. The FT is more general. It can be applied to both microscopic and macroscopic systems. **<mask> applied to macroscopic systems, the FT is equivalent to the Second Law of Thermodynamics.** [ENDQ] [NEWLINE] In other words, FT predicts small aberrations in a system that otherwise only increases in entropy. It does not predict large scale regenerations. The chance of your machine being anywhere near the incredibly rare regeneration of a star would be miniscule. The universe<mask> a whole would never restore itself to the point<mask> one could be said to have survived and even<mask> you did get lucky enough to be near a regenerated star once (and I'm talking about the star being part of your observable universe before it burns out), it wouldn't keep happening. [NEWLINE] [NEWLINE] The following is a requirement of the Poincare Recurrence Theorem: [NEWLINE] [NEWLINE] [STARTQ] A finite upper bound can be set on the total potentially accessible phase space volume. For a mechanical system, this bound can be provided by requiring that the system is contained in a bounded physical region of space (<mask> that it cannot, for example, eject particles that never return) — combined with the conservation of energy, this locks the system into a finite region in phase space. [ENDQ] [NEWLINE] No such upper bound can be placed on the expanding universe that plays such a pivotal role in creating the heat death of the universe. [NEWLINE] [NEWLINE] There are theories that predict that a survivable universe will eventually and spontaneously reoccur.<mask>, these theories hinge on an infinite time scale and the chance of a survivable universe reoccurring shortly (with respect to the lifetime of the universe) after the death of the old universe is negligible. Furthermore, your incredibly engineered machine that is designed to hold on to<mask> much energy<mask> possible probably wouldn't survive the intense levels of radiation released in a second big bang. Eliminating excess energy just wouldn't be<mask> it was designed for.</s>
Label encoding: <s> [STARTQ] Note that the FT does not state that the second law of thermodynamics is wrong or invalid. The second law of thermodynamics is a statement about macroscopic systems. The FT is more general. It can be applied to both microscopic and macroscopic systems. ** When applied to macroscopic systems, the FT is equivalent to the Second Law of Thermodynamics.** [ENDQ] [NEWLINE] In other words, FT predicts small aberrations in a system that otherwise only increases in entropy. It does not predict large scale regenerations. The chance of your machine being anywhere near the incredibly rare regeneration of a star would be miniscule. The universe as a whole would never restore itself to the point where one could be said to have survived and even if you did get lucky enough to be near a regenerated star once (and I'm talking about the star being part of your observable universe before it burns out), it wouldn't keep happening. [NEWLINE] [NEWLINE] The following is a requirement of the Poincare Recurrence Theorem: [NEWLINE] [NEWLINE] [STARTQ] A finite upper bound can be set on the total potentially accessible phase space volume. For a mechanical system, this bound can be provided by requiring that the system is contained in a bounded physical region of space ( so that it cannot, for example, eject particles that never return) — combined with the conservation of energy, this locks the system into a finite region in phase space. [ENDQ] [NEWLINE] No such upper bound can be placed on the expanding universe that plays such a pivotal role in creating the heat death of the universe. [NEWLINE] [NEWLINE] There are theories that predict that a survivable universe will eventually and spontaneously reoccur. However, these theories hinge on an infinite time scale and the chance of a survivable universe reoccurring shortly (with respect to the lifetime of the universe) after the death of the old universe is negligible. Furthermore, your incredibly engineered machine that is designed to hold on to as much energy as possible probably wouldn't survive the intense levels of radiation released in a second big bang. Eliminating excess energy just wouldn't be what it was designed for.</s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I battle with my friends we don't use these rules and are still able to play a balanced enough game. [ENDQ] [NEWLINE] <mask> you're just playing with a team you like, sure.<mask><mask> you play with a team you've specifically bred with maximum winning potential in mind, it would be different. [NEWLINE] [NEWLINE] There are a huge number of strategies in Pokemon, and there are ways to counter each strategy. There's various risks and tradeoffs for each approach you make. Generally speaking, this is a good thing. You want game play to be competitive and varied and for people to have to think about the way they're going to play both in the game and in the metagame. [NEWLINE] [NEWLINE] <mask> you allow for broken strategies to be used the game becomes less varied. Imagine,<mask> an extreme example,<mask> there was a pokemon that had stats that were extroadinarily high. The defense of shuckle, the attack power of Machamp, the HP of Blissy, etc. And it had an ability that prevented the opponent from switching, and it had a type combo that made it weak to<mask> few things<mask> possible. In this extreme case,<mask> you do not include this pokemon on your team, you *will* lose. You've made the competitive scene less interesting<mask> there's simply no way to counter the fact that you need one of these incredibly strong pokemon on your team. [NEWLINE] [NEWLINE] <mask><mask><mask> you would agree that you'd be justified in banning it - it makes the game suck. You effectively only have 5 pokemon you have any freedom to craft your team around, and the 6th will always be this Ubermon. [NEWLINE] [NEWLINE] This is<mask> Smogon is trying to prevent. There are some Pokemon and moves that are<mask> broken that either everyone has to account for that pokemon being on the opponents team or they lose. That decreases the number of viable strategies and the number of viable pokemon, and makes the game less interesting. Somewhat counter-intuitively, by removing some pokemon from the competitive scene you increase<mask> many pokemon are viable in the cometitive scene. </s>
Label encoding: <s> [STARTQ] When I battle with my friends we don't use these rules and are still able to play a balanced enough game. [ENDQ] [NEWLINE] If you're just playing with a team you like, sure. But if you play with a team you've specifically bred with maximum winning potential in mind, it would be different. [NEWLINE] [NEWLINE] There are a huge number of strategies in Pokemon, and there are ways to counter each strategy. There's various risks and tradeoffs for each approach you make. Generally speaking, this is a good thing. You want game play to be competitive and varied and for people to have to think about the way they're going to play both in the game and in the metagame. [NEWLINE] [NEWLINE] When you allow for broken strategies to be used the game becomes less varied. Imagine, as an extreme example, if there was a pokemon that had stats that were extroadinarily high. The defense of shuckle, the attack power of Machamp, the HP of Blissy, etc. And it had an ability that prevented the opponent from switching, and it had a type combo that made it weak to as few things as possible. In this extreme case, if you do not include this pokemon on your team, you *will* lose. You've made the competitive scene less interesting because there's simply no way to counter the fact that you need one of these incredibly strong pokemon on your team. [NEWLINE] [NEWLINE] So I think you would agree that you'd be justified in banning it - it makes the game suck. You effectively only have 5 pokemon you have any freedom to craft your team around, and the 6th will always be this Ubermon. [NEWLINE] [NEWLINE] This is what Smogon is trying to prevent. There are some Pokemon and moves that are so broken that either everyone has to account for that pokemon being on the opponents team or they lose. That decreases the number of viable strategies and the number of viable pokemon, and makes the game less interesting. Somewhat counter-intuitively, by removing some pokemon from the competitive scene you increase how many pokemon are viable in the cometitive scene. </s>
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Pretty much by definition unlawful arrests aren't a policeman's job.  Official misconduct is a crime, falsifying evidence is a crime, perjuring oneself is a crime.  These are routine behaviors of a significant percentage of police.  Not every cop does it,<mask> enough do that it is a threat to civil order.  The glue that holds us together<mask> a society is brittle, we have made it a couple of lifetimes<mask> it was completely torn to shit.  We've got a good thing going in the U.S.,<mask> allow such an insignificant part of our society to run rampant and possibly ruin it for us all? <mask> the commission of those crimes is to be excused,<mask> does it stop?  We have to draw a line at some point, we can't just allow police to arrest anyone and then throw money at them, or much worse, and more commonly, railroad them to justify the arrest, without criminal repercussions. [NEWLINE] [NEWLINE] I don't expect the police to be perfect, and I don't support OP's argument,<mask> I will not stand for corruption and say it's a necessary evil. <mask> police and prosecutors enforced existing laws on each other like they do to the public, police would quit in droves and we could keep the honest ones.  <mask> until that point, we're dealing with a toxic cancer that has been ignored for far too long. [NEWLINE] [NEWLINE] This is not some pie in the sky dream, this could be done. Right now we have a magnifying glass on the issues of police brutality, and excessive force.  It's not that the police have gotten worse, it's that technology has given us unprecedented access to undisputed facts for the first time. [NEWLINE] [NEWLINE] <mask> we have the knowledge, and it's given us the will to change.  We can't stick the genie back in the bottle, the world knows now. <mask> we let this fester it won't be Ferguson burning,<mask> L.A. and New York, and Chicago. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Pretty much by definition unlawful arrests aren't a policeman's job.  Official misconduct is a crime, falsifying evidence is a crime, perjuring oneself is a crime.  These are routine behaviors of a significant percentage of police.  Not every cop does it, but enough do that it is a threat to civil order.  The glue that holds us together as a society is brittle, we have made it a couple of lifetimes since it was completely torn to shit.  We've got a good thing going in the U.S., why allow such an insignificant part of our society to run rampant and possibly ruin it for us all?  If the commission of those crimes is to be excused, where does it stop?  We have to draw a line at some point, we can't just allow police to arrest anyone and then throw money at them, or much worse, and more commonly, railroad them to justify the arrest, without criminal repercussions. [NEWLINE] [NEWLINE] I don't expect the police to be perfect, and I don't support OP's argument, but I will not stand for corruption and say it's a necessary evil.  If police and prosecutors enforced existing laws on each other like they do to the public, police would quit in droves and we could keep the honest ones.   But until that point, we're dealing with a toxic cancer that has been ignored for far too long. [NEWLINE] [NEWLINE] This is not some pie in the sky dream, this could be done. Right now we have a magnifying glass on the issues of police brutality, and excessive force.  It's not that the police have gotten worse, it's that technology has given us unprecedented access to undisputed facts for the first time. [NEWLINE] [NEWLINE] So we have the knowledge, and it's given us the will to change.  We can't stick the genie back in the bottle, the world knows now.  If we let this fester it won't be Ferguson burning, but L.A. and New York, and Chicago. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Shouldn't job/internship/school placement be based on merit? [ENDQ] [NEWLINE] Here's the thing. The only people that get successful 'networking' are the people that are able to actually do good work in the first place. Let me give you an example: [NEWLINE] [NEWLINE] I've been working for about 10 years now post-college, and I've had lots and lots of friends ask me about openings,<mask> to get their resume'seen', etc.<mask> unless those people actually going to be a good fit for the job, I'm not going to risk my job hiring someone incompetent,<mask> their incompetence reflects poorly on my ability to find/hire good people. I have seen a few colleagues who were fired, or who missed promotion, due to the fact that they let their friends into the organization that weren't the best people for the job. [NEWLINE] [NEWLINE] Business schools likely harp on the 'networking' thing<mask> it's a buzzword, not<mask> its the only thing that matters. And,<mask> networking is literally just 'being nice to people and not burning bridges,' it's not like they can make a whole class out of it.<mask> professor A will mention it, without realizing that *other* professors (B, C, and D) are<mask> mentioning it, and<mask> it gets repeated more often than it should. Compare that to a Marketing class, which isn't likely to discuss Computer Science,<mask> the college presumably has a class about that already. [NEWLINE] [NEWLINE] I'm not sure about the morality of this. It could be argued that you should only allow people you know into an organization,<mask> some people are crazy, lie on their resume, etc.<mask>, it could be immoral to pick random people,<mask> they could turn out to be serial killers. [NEWLINE] [NEWLINE] <mask><mask> networking is *slightly* important, it's really just about being nice, cultivating relationships, and ensuring that your organization hires good people. And, actually being good and competent at your job is absolutely more important than any possible 'networking' you can do.</s>
Label encoding: <s> [STARTQ] Shouldn't job/internship/school placement be based on merit? [ENDQ] [NEWLINE] Here's the thing. The only people that get successful 'networking' are the people that are able to actually do good work in the first place. Let me give you an example: [NEWLINE] [NEWLINE] I've been working for about 10 years now post-college, and I've had lots and lots of friends ask me about openings, how to get their resume'seen', etc. But unless those people actually going to be a good fit for the job, I'm not going to risk my job hiring someone incompetent, because their incompetence reflects poorly on my ability to find/hire good people. I have seen a few colleagues who were fired, or who missed promotion, due to the fact that they let their friends into the organization that weren't the best people for the job. [NEWLINE] [NEWLINE] Business schools likely harp on the 'networking' thing because it's a buzzword, not because its the only thing that matters. And, since networking is literally just 'being nice to people and not burning bridges,' it's not like they can make a whole class out of it. So professor A will mention it, without realizing that *other* professors (B, C, and D) are also mentioning it, and thus it gets repeated more often than it should. Compare that to a Marketing class, which isn't likely to discuss Computer Science, since the college presumably has a class about that already. [NEWLINE] [NEWLINE] I'm not sure about the morality of this. It could be argued that you should only allow people you know into an organization, since some people are crazy, lie on their resume, etc. Therefore, it could be immoral to pick random people, since they could turn out to be serial killers. [NEWLINE] [NEWLINE] So while networking is *slightly* important, it's really just about being nice, cultivating relationships, and ensuring that your organization hires good people. And, actually being good and competent at your job is absolutely more important than any possible 'networking' you can do.</s>
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As I posted a few minutes ago, I once lived in a quiet neighborhood and rode a Harley.<mask> 2013 I moved from that neighborhood to my mother's house<mask> she is very ill and I take care of her. Her house is on a very busy avenue and there is traffic going by at all hours of the day and night. Big dump trucks, 18 wheelers, metro buses, motorcycles, you name it. The speed limit is 30mph<mask> you would never know it. People speed up and down like a freeway and there is no passing<mask> I've seen people pass each other many many times. Not only that, there is a high school across the street and an elementary school down the street. School is out for the summer of course<mask><mask> it is back in session it is horrible. Traffic backs up and I can smell the exhaust coming in under the front door I guess<mask> I have the windows closed and the air on. People get impatient<mask> they sit out front and blow their horn. At the end of the school day the kids who drive use this street like it's the Daytona 500. Every fucking day I have to listen to them rev their engines and burn rubber down the street. I have called the police department many times and told them<mask> they put an officer out here they would make a shit ton of money on speeding tickets. It is rare<mask> to see a cop here. It is amazing to me that no pedestrians have ever been hit by speeding cars<mask> I guess it's just a matter of time. [NEWLINE] [NEWLINE] Oh and I almost forgot. Across the way is a small airport that accommodates small jets. Every day I hear them starting the engines and it gets louder and louder and louder until they take off. Every year there is an airshow.<mask> wonderful to have to listen to that shit all day. There is<mask> train tracks not far from here and of course I can hear the train whistle. I have gotten used to hearing that<mask><mask> it's not an issue. The other noise is<mask> bad I have lost my fucking mind. </s>
Label encoding: <s>As I posted a few minutes ago, I once lived in a quiet neighborhood and rode a Harley. Since 2013 I moved from that neighborhood to my mother's house because she is very ill and I take care of her. Her house is on a very busy avenue and there is traffic going by at all hours of the day and night. Big dump trucks, 18 wheelers, metro buses, motorcycles, you name it. The speed limit is 30mph but you would never know it. People speed up and down like a freeway and there is no passing but I've seen people pass each other many many times. Not only that, there is a high school across the street and an elementary school down the street. School is out for the summer of course but when it is back in session it is horrible. Traffic backs up and I can smell the exhaust coming in under the front door I guess because I have the windows closed and the air on. People get impatient so they sit out front and blow their horn. At the end of the school day the kids who drive use this street like it's the Daytona 500. Every fucking day I have to listen to them rev their engines and burn rubber down the street. I have called the police department many times and told them if they put an officer out here they would make a shit ton of money on speeding tickets. It is rare however to see a cop here. It is amazing to me that no pedestrians have ever been hit by speeding cars but I guess it's just a matter of time. [NEWLINE] [NEWLINE] Oh and I almost forgot. Across the way is a small airport that accommodates small jets. Every day I hear them starting the engines and it gets louder and louder and louder until they take off. Every year there is an airshow. So wonderful to have to listen to that shit all day. There is also train tracks not far from here and of course I can hear the train whistle. I have gotten used to hearing that though so it's not an issue. The other noise is so bad I have lost my fucking mind. </s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've had an argument with a friend of mine, who is religious, about the whole gay marriage issue (I live in Washington,<mask> it is legal now, and we've<mask> to fall into a pit of Oblivion).  I brought up the separation of church and state and she said she didn't care and that she'd vote based on<mask> she believed,<mask> that's<mask> she thought was the right thing to do. [NEWLINE] [NEWLINE] <mask> I asked her<mask> she's choosing this particular thing to make a stance on,<mask> she ignores<mask> many other parts of the bible that include things that are not socially or lawfully acceptable, her only answer was that she recognizes she's a sinner,<mask> is doing the best she can. [NEWLINE] [NEWLINE] That conversation aside, which was an obvious cop-out on her part, I'm wondering<mask> religious people (in general, in this country, Christians) can honestly feel like they're doing the right thing<mask> they just pick and choose convenient things from the bible to try to enforce<mask> completely ignoring the things that are now seen<mask> ridiculous.  Does the bible list different things by priority? <mask> are religious folks allowed to ignore certain things,<mask> do everything they can to enfore others? [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I've had an argument with a friend of mine, who is religious, about the whole gay marriage issue (I live in Washington, where it is legal now, and we've yet to fall into a pit of Oblivion).  I brought up the separation of church and state and she said she didn't care and that she'd vote based on what she believed, as that's what she thought was the right thing to do. [NEWLINE] [NEWLINE] When I asked her why she's choosing this particular thing to make a stance on, when she ignores so many other parts of the bible that include things that are not socially or lawfully acceptable, her only answer was that she recognizes she's a sinner, but is doing the best she can. [NEWLINE] [NEWLINE] That conversation aside, which was an obvious cop-out on her part, I'm wondering how religious people (in general, in this country, Christians) can honestly feel like they're doing the right thing when they just pick and choose convenient things from the bible to try to enforce while completely ignoring the things that are now seen as ridiculous.  Does the bible list different things by priority?  Why are religious folks allowed to ignore certain things, but do everything they can to enfore others? [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a few thought experiments, I dunno<mask> they've been mentioned<mask>. [NEWLINE] [NEWLINE] Consider that you designed and built a trolley car,<mask> you know the exact specifications of it. You're sitting on a bridge next to a lever that diverts the car onto an alternative track. You're watching six workers, 5 on one track and 1 on the alternate, work on the track. All of a sudden you notice a rogue trolley car zooming down the track and will kill the five workers, and you don't have enough time to do anything<mask> pull the lever.<mask> do you do? Clearly,<mask><mask> utilitarianism, you pull the lever and the one person dies. [NEWLINE] [NEWLINE] Now consider the same setup, except in this case there's no lever. You're on the bridge with another person who is the exact mass needed to stop the trolley car in time to save the five workers. All you have to do to stop the rogue trolley car is push the man over.<mask> do you do? [NEWLINE] [NEWLINE] Most people have a very big qualm with pushing the man over,<mask> it seems that maximizing happiness does not guide our moral compass.<mask> you have no qualms about it, you are a true utilitarian, and I'll concede this one to you. [NEWLINE] [NEWLINE] In which case, consider this thought experiment. You are in a situation and you have two choices. You can lie, or tell the truth. Either way produces the same outcome, 100 utils.<mask> a utilitarian, you must concede that lying and telling the truth have no moral value on their own, and in this case lying is equally<mask> moral<mask> telling the truth,<mask><mask><mask> you do one of the two. [NEWLINE] [NEWLINE] And a third. Consider that you promised the neighbor twenty dollars to mow your lawn. After he is done and knocks on your door, it occurs to you that giving the same twenty dollars to starving children in Africa will overall produce more utils than this kid that lives across from you. Do you break your promise to this kid?</s>
Label encoding: <s>I have a few thought experiments, I dunno if they've been mentioned yet. [NEWLINE] [NEWLINE] Consider that you designed and built a trolley car, so you know the exact specifications of it. You're sitting on a bridge next to a lever that diverts the car onto an alternative track. You're watching six workers, 5 on one track and 1 on the alternate, work on the track. All of a sudden you notice a rogue trolley car zooming down the track and will kill the five workers, and you don't have enough time to do anything but pull the lever. What do you do? Clearly, according to utilitarianism, you pull the lever and the one person dies. [NEWLINE] [NEWLINE] Now consider the same setup, except in this case there's no lever. You're on the bridge with another person who is the exact mass needed to stop the trolley car in time to save the five workers. All you have to do to stop the rogue trolley car is push the man over. What do you do? [NEWLINE] [NEWLINE] Most people have a very big qualm with pushing the man over, so it seems that maximizing happiness does not guide our moral compass. If you have no qualms about it, you are a true utilitarian, and I'll concede this one to you. [NEWLINE] [NEWLINE] In which case, consider this thought experiment. You are in a situation and you have two choices. You can lie, or tell the truth. Either way produces the same outcome, 100 utils. As a utilitarian, you must concede that lying and telling the truth have no moral value on their own, and in this case lying is equally as moral as telling the truth, so long as you do one of the two. [NEWLINE] [NEWLINE] And a third. Consider that you promised the neighbor twenty dollars to mow your lawn. After he is done and knocks on your door, it occurs to you that giving the same twenty dollars to starving children in Africa will overall produce more utils than this kid that lives across from you. Do you break your promise to this kid?</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You're just trying to convince me that censorship is better, and I'm just saying that this site is now heavily censored. [ENDQ] [NEWLINE] Parts of it are heavily censored,<mask> it's about<mask> useful<mask> saying that the restaurants in a posh part of town are heavily censored<mask> you gotta have a tie and jacket. [NEWLINE] [NEWLINE] (Now I want to watch _<mask> Good<mask> It Gets_) [NEWLINE] [NEWLINE] Starting a new sub is free. Try it. The link is right there on reddit's front page.<mask> you look around, you'll find a few subs that are utterly unmodded and the spam filter is turned off. [NEWLINE] [NEWLINE] They're all toilets. [NEWLINE] [NEWLINE] <mask> reddit isn't meant to be a graffiti wall in general, it just provides a mechanism to make a million walls<mask> you can spray<mask> you want. Or take ownership of your own and censor<mask> you want. That's just<mask> reddit went about the way of solving the--arguably insolvable--problem of having places without X, and places<mask> X and every other letter of the alphabet get their chance. [NEWLINE] [NEWLINE] Many of those who get their posts removed aren't satisfied with the suggestion to take the post elsewhere for two reasons: [NEWLINE] [NEWLINE] 1. "Elsewhere" doesn't have enough subscribers. [NEWLINE] [NEWLINE] 2. It's imperative that the person they're arguing with _must_ see the reply, and<mask> must everyone else. [NEWLINE] [NEWLINE] Freedom of speech doesn't include the freedom to be heard. It just means you won't be arrested and jailed for whatever you say.<mask> reddit's mechanism for parceling out virtual land isn't perfect, at the very least the admins very deliberately don't give a toss<mask> goes on in a sub. [NEWLINE] [NEWLINE] That has been to their detriment several times, such<mask> with /r/niggers and /r/jailbait. It took *fucking CNN* to get those subs yanked.<mask> now there's /r/greatapes. Go to that sub. Now tell me about reddit censorship.</s>
Label encoding: <s> [STARTQ] You're just trying to convince me that censorship is better, and I'm just saying that this site is now heavily censored. [ENDQ] [NEWLINE] Parts of it are heavily censored, but it's about as useful as saying that the restaurants in a posh part of town are heavily censored because you gotta have a tie and jacket. [NEWLINE] [NEWLINE] (Now I want to watch _ As Good As It Gets_) [NEWLINE] [NEWLINE] Starting a new sub is free. Try it. The link is right there on reddit's front page. If you look around, you'll find a few subs that are utterly unmodded and the spam filter is turned off. [NEWLINE] [NEWLINE] They're all toilets. [NEWLINE] [NEWLINE] But reddit isn't meant to be a graffiti wall in general, it just provides a mechanism to make a million walls where you can spray what you want. Or take ownership of your own and censor what you want. That's just how reddit went about the way of solving the--arguably insolvable--problem of having places without X, and places where X and every other letter of the alphabet get their chance. [NEWLINE] [NEWLINE] Many of those who get their posts removed aren't satisfied with the suggestion to take the post elsewhere for two reasons: [NEWLINE] [NEWLINE] 1. "Elsewhere" doesn't have enough subscribers. [NEWLINE] [NEWLINE] 2. It's imperative that the person they're arguing with _must_ see the reply, and so must everyone else. [NEWLINE] [NEWLINE] Freedom of speech doesn't include the freedom to be heard. It just means you won't be arrested and jailed for whatever you say. While reddit's mechanism for parceling out virtual land isn't perfect, at the very least the admins very deliberately don't give a toss what goes on in a sub. [NEWLINE] [NEWLINE] That has been to their detriment several times, such as with /r/niggers and /r/jailbait. It took *fucking CNN* to get those subs yanked. But now there's /r/greatapes. Go to that sub. Now tell me about reddit censorship.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A human designed it. A human designed the machine that builds it. A human oversees the machine, repairs the machine, and make's sure the machine has energy to function. A human tests the things that the machine builds, moves it to a market, and convinces other people to buy it. [ENDQ] [NEWLINE] <mask>, the more advanced the machines become, the fewer people need to be involved. For instance automobile manufacturing was once largely assembly lines of hundreds of workers, now it is largely automated, with some human input,<mask> comparatively little.<mask> technology continues to advance the greater degree of mechanisation we get and the fewer humans are required,<mask> it stands we still need humans to do tasks such<mask> maintenance on said machines,<mask> it not unlikely that an automated system could eventually be created that allows machines to maintain themselves/each other. [NEWLINE] [NEWLINE] Are we all to become designers and salesmen? [NEWLINE] [NEWLINE] [STARTQ] <mask> we see is a changing demand for human labor. Instead of a demand for a person who works on an assembly line, we need people who can repair the machine that works the line. We're just transitioning to a service-based economy. [ENDQ] [NEWLINE] This is true, currently,<mask><mask> I said above, more and more work will become mechanised,<mask> the technology is advanced and affordable enough, which it will be eventually, it will be adopted. You don't need a mechanic<mask> you have two mechanic-bots that can repair other machines and each other, you don't need builders<mask> you have pre-fab buildings that can be assembled by fully automated machines. You may still need people at some points,<mask> far fewer. [NEWLINE] [NEWLINE] I'm not talking about the current state of affairs and you seem to dismissing technological advance and adoption, one of the main ideas of mechanisation is that it requires fewer (expensive and unreliable) humans, instead of 200 factory workers you have a machine and 10 factory workers, that's the point, you don't get 200 engineers and managers instead. We can't all be lawyers, managers and salesmen.</s>
Label encoding: <s> [STARTQ] A human designed it. A human designed the machine that builds it. A human oversees the machine, repairs the machine, and make's sure the machine has energy to function. A human tests the things that the machine builds, moves it to a market, and convinces other people to buy it. [ENDQ] [NEWLINE] However, the more advanced the machines become, the fewer people need to be involved. For instance automobile manufacturing was once largely assembly lines of hundreds of workers, now it is largely automated, with some human input, but comparatively little. As technology continues to advance the greater degree of mechanisation we get and the fewer humans are required, as it stands we still need humans to do tasks such as maintenance on said machines, but it not unlikely that an automated system could eventually be created that allows machines to maintain themselves/each other. [NEWLINE] [NEWLINE] Are we all to become designers and salesmen? [NEWLINE] [NEWLINE] [STARTQ] What we see is a changing demand for human labor. Instead of a demand for a person who works on an assembly line, we need people who can repair the machine that works the line. We're just transitioning to a service-based economy. [ENDQ] [NEWLINE] This is true, currently, but as I said above, more and more work will become mechanised, when the technology is advanced and affordable enough, which it will be eventually, it will be adopted. You don't need a mechanic when you have two mechanic-bots that can repair other machines and each other, you don't need builders when you have pre-fab buildings that can be assembled by fully automated machines. You may still need people at some points, but far fewer. [NEWLINE] [NEWLINE] I'm not talking about the current state of affairs and you seem to dismissing technological advance and adoption, one of the main ideas of mechanisation is that it requires fewer (expensive and unreliable) humans, instead of 200 factory workers you have a machine and 10 factory workers, that's the point, you don't get 200 engineers and managers instead. We can't all be lawyers, managers and salesmen.</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To caveat:<mask> we lived in a free market, I'd agree.<mask> we don't. [NEWLINE] [NEWLINE] Some people have already made the argument that the 1% rely on government infrastructure to make their money, and I'm going to make a related argument. I don't think it matters<mask> you use public roads to make your fortune, or hire people educated by public education,<mask> once you begin colluding with the government to make public goods serve you disproportionately, that's<mask> you can ask serious questions about whether the 1% deserve their money.<mask>, the question is: Do members of the 1% receive protection or assistance from the government that makes their professions more lucrative? Yes, they do. [NEWLINE] [NEWLINE] [Here's]( [URL] /) an infographic breakdown of the professions held by the 1%. Let's look at one of the larger blocks: Lawyers. [NEWLINE] [NEWLINE] Lawyers make money by participating in trials, and providing legal services to wealthy corporations and individuals who want to be compliant or appear compliant with the law. These laws are written by congresspeople -- many of whom are former lawyers themselves, and will return to legal practice<mask> they fail to be re-elected. The expanding federal and state code that all businesses must adhere to are a direct result of the tacit partnership between the legal profession and the federal government. The cost of this partnership to U.S. businesses can be read about in this report called [10,000 Commandments]( [URL] ). The cost for American businesses to comply with these regulations is $1.8 trillion. Not all of this money goes to lawyers,<mask> a good chunk surely does. [NEWLINE] [NEWLINE] This public/private cronyism is pervasive throughout the U.S. economy. It's hard to find a fortune that hasn't been built through collusion with federal and state governments. [NEWLINE] [NEWLINE] TL;DR: Private individuals and businesses collude with the government to rig the game in their favor against their competitors and against consumers.<mask> they do that, it's completely ridiculous for them to claim their fortunes are theirs alone. [NEWLINE] </s>
Label encoding: <s>To caveat: If we lived in a free market, I'd agree. But we don't. [NEWLINE] [NEWLINE] Some people have already made the argument that the 1% rely on government infrastructure to make their money, and I'm going to make a related argument. I don't think it matters if you use public roads to make your fortune, or hire people educated by public education, but once you begin colluding with the government to make public goods serve you disproportionately, that's when you can ask serious questions about whether the 1% deserve their money. So, the question is: Do members of the 1% receive protection or assistance from the government that makes their professions more lucrative? Yes, they do. [NEWLINE] [NEWLINE] [Here's]( [URL] /) an infographic breakdown of the professions held by the 1%. Let's look at one of the larger blocks: Lawyers. [NEWLINE] [NEWLINE] Lawyers make money by participating in trials, and providing legal services to wealthy corporations and individuals who want to be compliant or appear compliant with the law. These laws are written by congresspeople -- many of whom are former lawyers themselves, and will return to legal practice when they fail to be re-elected. The expanding federal and state code that all businesses must adhere to are a direct result of the tacit partnership between the legal profession and the federal government. The cost of this partnership to U.S. businesses can be read about in this report called [10,000 Commandments]( [URL] ). The cost for American businesses to comply with these regulations is $1.8 trillion. Not all of this money goes to lawyers, but a good chunk surely does. [NEWLINE] [NEWLINE] This public/private cronyism is pervasive throughout the U.S. economy. It's hard to find a fortune that hasn't been built through collusion with federal and state governments. [NEWLINE] [NEWLINE] TL;DR: Private individuals and businesses collude with the government to rig the game in their favor against their competitors and against consumers. When they do that, it's completely ridiculous for them to claim their fortunes are theirs alone. [NEWLINE] </s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's a fine choice<mask> it's<mask> you actually want. Denying who you are and refusing to act on your sexuality<mask> can be extremely unhealthy psychologically. I can understand a desire to not come out, "Hi, I'm a member of one of the most abused, hated, ostracised, and distrusted groups in the world; let's have a parade about it", never made much sense to me either.<mask><mask><mask> wrong it is, homeophobia is sadly commonplace. [NEWLINE] [NEWLINE] That said<mask> you're in pretty much any Western European country, Australia, New Zealand, Canada, or the more enlightened parts of USA you can be easily welcomed by your peers without issue<mask><mask> your sexuality. You've made it clear you have no religious issues preventing you from acknowledging your orientation,<mask> just do it. You don't have to actively go looking for a sexual partner, and you don't have to be a camp lisping puff, or dress like the S&amp;M rainbow apocalypse is coming. Many gay people even find those stereotypes insulting (<mask>, they started<mask> subtle body language code back<mask> being gay was illegal, and the traits just got exaggerated;<mask> knowing their history it's less irritating). [NEWLINE] [NEWLINE] Rather,<mask> you should do is simply accept yourself for who you are, and should a guy you like turn up, don't stop yourself from seeking a relationship<mask> it happens naturally. It might not,<mask> then again, maybe it will. It's not "giving in", it's just being yourself. You're not doing anything wrong,<mask><mask> deny yourself happiness? This is the best way to live, and it changes nothing about your current lifestyle except your personal outlook. [NEWLINE] [NEWLINE] Do<mask> makes you happy, and<mask> being single makes you happy that is a perfectly valid choice. You don't have to proudly screw men<mask> you're gay and proud (groan),<mask> you don't have to deny yourself a happy relationship just to avoid the theatrics and politics. [NEWLINE] [NEWLINE] I hope this helped.</s>
Label encoding: <s>That's a fine choice if it's what you actually want. Denying who you are and refusing to act on your sexuality however can be extremely unhealthy psychologically. I can understand a desire to not come out, "Hi, I'm a member of one of the most abused, hated, ostracised, and distrusted groups in the world; let's have a parade about it", never made much sense to me either. Regardless of how wrong it is, homeophobia is sadly commonplace. [NEWLINE] [NEWLINE] That said if you're in pretty much any Western European country, Australia, New Zealand, Canada, or the more enlightened parts of USA you can be easily welcomed by your peers without issue regardless of your sexuality. You've made it clear you have no religious issues preventing you from acknowledging your orientation, so just do it. You don't have to actively go looking for a sexual partner, and you don't have to be a camp lisping puff, or dress like the S&amp;M rainbow apocalypse is coming. Many gay people even find those stereotypes insulting ( although, they started as subtle body language code back when being gay was illegal, and the traits just got exaggerated; so knowing their history it's less irritating). [NEWLINE] [NEWLINE] Rather, what you should do is simply accept yourself for who you are, and should a guy you like turn up, don't stop yourself from seeking a relationship if it happens naturally. It might not, but then again, maybe it will. It's not "giving in", it's just being yourself. You're not doing anything wrong, so why deny yourself happiness? This is the best way to live, and it changes nothing about your current lifestyle except your personal outlook. [NEWLINE] [NEWLINE] Do what makes you happy, and if being single makes you happy that is a perfectly valid choice. You don't have to proudly screw men because you're gay and proud (groan), but you don't have to deny yourself a happy relationship just to avoid the theatrics and politics. [NEWLINE] [NEWLINE] I hope this helped.</s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That is literally<mask> "The Emperor's New Clothes" is about, yes,<mask> the moral is much broader in its scope. I am obviously not arguing that the painting isn't literally present. I am suggesting that people who consider themselves informed about art are afraid to admit that these paintings are simply not that good for fear of seeming ignorant in front of others, much like the people in TENC were afraid to admit that they couldn't see the emperor's clothes. [NEWLINE] [NEWLINE] Your defense of these paintings includes: [NEWLINE] [NEWLINE] 1) It is strongly influenced by Nietzche [NEWLINE] [NEWLINE] 2) It confronts the viewer with something uncomfortable [NEWLINE] [NEWLINE] 3) Viewed in the right setting it overwhelms the senses due to its size [NEWLINE] [NEWLINE] 4) It is original [NEWLINE] [NEWLINE] 5) It takes technical skill to create [NEWLINE] [NEWLINE] Let's see<mask> I can illustrate my point with an example: [NEWLINE] [NEWLINE] Bob is a musician and amateur philosopher. He spends many years studying Nihilism and pondering<mask> to reflect the view that life is without objective or meaning in his music. [NEWLINE] [NEWLINE] He finally has an epiphany and creates his materpiece: 20 minutes of static and cymbal crashes. He uses all his skill<mask> a musician to make high quality recordings of the sounds and masters the track expertly. He notes that his music is meant to be listened to on very loud and in complete darkness. Bob's song now satisfies all the above criteria. [NEWLINE] [NEWLINE] 1) It is strongly influenced by and attempts to convey a philosophy, namely Nihilism [NEWLINE] [NEWLINE] 2) It confronts the viewer with something uncomfortable (loud, meaningless noise) [NEWLINE] [NEWLINE] 3) Listened to in the right setting (loud and dark) it overwhelms the senses due to its volume [NEWLINE] [NEWLINE] 4) It is certainly original. [NEWLINE] [NEWLINE] 5) It takes technical skill to create. Bob did expert-level mastering and recording. [NEWLINE] [NEWLINE] <mask>, it is still just 20 minutes of static and cymbal crashes and would almost unanimously be considered to not be art.</s>
Label encoding: <s>That is literally what "The Emperor's New Clothes" is about, yes, but the moral is much broader in its scope. I am obviously not arguing that the painting isn't literally present. I am suggesting that people who consider themselves informed about art are afraid to admit that these paintings are simply not that good for fear of seeming ignorant in front of others, much like the people in TENC were afraid to admit that they couldn't see the emperor's clothes. [NEWLINE] [NEWLINE] Your defense of these paintings includes: [NEWLINE] [NEWLINE] 1) It is strongly influenced by Nietzche [NEWLINE] [NEWLINE] 2) It confronts the viewer with something uncomfortable [NEWLINE] [NEWLINE] 3) Viewed in the right setting it overwhelms the senses due to its size [NEWLINE] [NEWLINE] 4) It is original [NEWLINE] [NEWLINE] 5) It takes technical skill to create [NEWLINE] [NEWLINE] Let's see if I can illustrate my point with an example: [NEWLINE] [NEWLINE] Bob is a musician and amateur philosopher. He spends many years studying Nihilism and pondering how to reflect the view that life is without objective or meaning in his music. [NEWLINE] [NEWLINE] He finally has an epiphany and creates his materpiece: 20 minutes of static and cymbal crashes. He uses all his skill as a musician to make high quality recordings of the sounds and masters the track expertly. He notes that his music is meant to be listened to on very loud and in complete darkness. Bob's song now satisfies all the above criteria. [NEWLINE] [NEWLINE] 1) It is strongly influenced by and attempts to convey a philosophy, namely Nihilism [NEWLINE] [NEWLINE] 2) It confronts the viewer with something uncomfortable (loud, meaningless noise) [NEWLINE] [NEWLINE] 3) Listened to in the right setting (loud and dark) it overwhelms the senses due to its volume [NEWLINE] [NEWLINE] 4) It is certainly original. [NEWLINE] [NEWLINE] 5) It takes technical skill to create. Bob did expert-level mastering and recording. [NEWLINE] [NEWLINE] Yet, it is still just 20 minutes of static and cymbal crashes and would almost unanimously be considered to not be art.</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The guide can contribute to acquiantance rape and manipulation of the woman. Just<mask> she is sexually interested does not mean anything sexual is allowed.<mask><mask> the girl doesnt like giving blowjobs? Or she doesnt like handjobs? Or she's not ready to take it that far? Or she's just not in the mood?<mask><mask> she was sexually assaulted, raped, and/or sexually abused in the past? That's not something a girl goes around telling everyone. Presuming that she's up for anything is dangerous and rapey behavior. You should give her a chance to say yes instead of no. It's better to communicate vocally asking than to psychologicay damage your partner by misreading the situation. Sex should be enjoyed by both people.<mask> one person ends up psychologically damaged, it reflects poorly on your skills. It's not worth the risk. [NEWLINE] [NEWLINE] There are ways to seductively give consent. Sex talk does that. Ask her<mask> would she like you to do. Or tell her<mask> you'd like her to do without grabbing her and starting it for her.<mask> a guy decides for us, it can be extremely terrifying and girls are taught to react submissively. [NEWLINE] [NEWLINE] It's hard to see it unless you experience it. I had a guy who manipulated me by using the same.techniques the book says. This close friend ended up sexually assaulting me. He treated it like an accident, an oopsies. He's not the one who is paranoid about the male gender, dislike being touch,  is emotionally unstable now, disgusted with his body, lost sexual interest, fears trusting others, and is jumpy. I am. This guide teaches men to not be careful and tells men<mask> women want-<mask> really woman is too vague of a term to address<mask> sexual actions a partnet might want. It is rapey. It's the fact that we dont acknowledge<mask> rape and instead claim it's her changing her mind that make people think it's okay.</s>
Label encoding: <s>The guide can contribute to acquiantance rape and manipulation of the woman. Just because she is sexually interested does not mean anything sexual is allowed. What if the girl doesnt like giving blowjobs? Or she doesnt like handjobs? Or she's not ready to take it that far? Or she's just not in the mood? What if she was sexually assaulted, raped, and/or sexually abused in the past? That's not something a girl goes around telling everyone. Presuming that she's up for anything is dangerous and rapey behavior. You should give her a chance to say yes instead of no. It's better to communicate vocally asking than to psychologicay damage your partner by misreading the situation. Sex should be enjoyed by both people. If one person ends up psychologically damaged, it reflects poorly on your skills. It's not worth the risk. [NEWLINE] [NEWLINE] There are ways to seductively give consent. Sex talk does that. Ask her what would she like you to do. Or tell her what you'd like her to do without grabbing her and starting it for her. When a guy decides for us, it can be extremely terrifying and girls are taught to react submissively. [NEWLINE] [NEWLINE] It's hard to see it unless you experience it. I had a guy who manipulated me by using the same.techniques the book says. This close friend ended up sexually assaulting me. He treated it like an accident, an oopsies. He's not the one who is paranoid about the male gender, dislike being touch,  is emotionally unstable now, disgusted with his body, lost sexual interest, fears trusting others, and is jumpy. I am. This guide teaches men to not be careful and tells men what women want- when really woman is too vague of a term to address what sexual actions a partnet might want. It is rapey. It's the fact that we dont acknowledge as rape and instead claim it's her changing her mind that make people think it's okay.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Little errors are going to exist on resumes. It's common for individuals to be blind to<mask> they have written, even after they have taken the time to review their work. This is<mask> professionals in writing, even<mask> they write short blurbs, require editors to go over their work. Perfect resumes may be the result of multiple sets of eyes reviewing it rather than meticulously reviewing the document. [NEWLINE] [NEWLINE] A perfect resume isn't a reflection of one's focus on detail or even conscientiousness. There are people who just find it difficult to write about themselves. Other aspects of their output, or their ability to review the output of others, are separate from their ability to put together a one page document about their job history and fit with your company. [NEWLINE] [NEWLINE] This is<mask> interviews exist. This is<mask> some professions may have assessments or some form of testing to determine<mask> the individual meets the minimum requirements for the position. (We have an ATS in place to weed out candidates who don't meet the minimum qualifications for our positions.) A resume's main purpose is to provide you with a candidate's work history. Additional information should come from the candidate. I'll be honest,<mask>. It is a little bit funny<mask> someone writes, "Detal orientated," or something similar.<mask> you talk to candidates who have the work history you want. Or perhaps the potential you want. [NEWLINE] [NEWLINE] This is just<mask> I've picked up from being involved in the hiring/recruiting portion of HR at a large corporation in the US. I am not directly involved in hiring candidates,<mask> I've learned some things about<mask> we look for in hiring individuals at all levels. And this is an interesting time period.<mask> unemployment is going down, people are still having trouble finding jobs. They want to present the best face possible for every position,<mask> they<mask> want to and need to apply for<mask> many positions<mask> possible. And some people are anxious about the whole process. [NEWLINE] [NEWLINE] To err is human. Please remember that.</s>
Label encoding: <s>Little errors are going to exist on resumes. It's common for individuals to be blind to what they have written, even after they have taken the time to review their work. This is why professionals in writing, even when they write short blurbs, require editors to go over their work. Perfect resumes may be the result of multiple sets of eyes reviewing it rather than meticulously reviewing the document. [NEWLINE] [NEWLINE] A perfect resume isn't a reflection of one's focus on detail or even conscientiousness. There are people who just find it difficult to write about themselves. Other aspects of their output, or their ability to review the output of others, are separate from their ability to put together a one page document about their job history and fit with your company. [NEWLINE] [NEWLINE] This is why interviews exist. This is why some professions may have assessments or some form of testing to determine if the individual meets the minimum requirements for the position. (We have an ATS in place to weed out candidates who don't meet the minimum qualifications for our positions.) A resume's main purpose is to provide you with a candidate's work history. Additional information should come from the candidate. I'll be honest, though. It is a little bit funny if someone writes, "Detal orientated," or something similar. But you talk to candidates who have the work history you want. Or perhaps the potential you want. [NEWLINE] [NEWLINE] This is just what I've picked up from being involved in the hiring/recruiting portion of HR at a large corporation in the US. I am not directly involved in hiring candidates, but I've learned some things about what we look for in hiring individuals at all levels. And this is an interesting time period. While unemployment is going down, people are still having trouble finding jobs. They want to present the best face possible for every position, but they also want to and need to apply for as many positions as possible. And some people are anxious about the whole process. [NEWLINE] [NEWLINE] To err is human. Please remember that.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your approach doesn't hold much hope for this,<mask> all it does is postpone your death until the actual heat-death of the universe. I would challenge that part of your view. [NEWLINE] [NEWLINE] <mask>, it does seem possible that at least some scenarios could allow survival. [NEWLINE] [NEWLINE] One is the creation of another universe that we could somehow transfer ourselves into. We don't currently understand<mask> universes come to be,<mask> this seems implausible today,<mask> we have billions of years to figure it out. [NEWLINE] [NEWLINE] Another is that quantum mechanics might possibly make it impossible for the universe to *actually* achieve perfect heat-death. It might be the case that there is always a fluctuating quantum field that can be used<mask> a minute energy source. A sufficiently clever species might find a way to spread their consciousness over a sufficiently large area that this minute amount of energy could sustain them. [NEWLINE] [NEWLINE] Another is that we *really* don't understand anything about the missing majority of the universe we're currently calling Dark Matter and Dark Energy. We don't know<mask> laws they follow, we don't really know anything about them, or even for certain know that they exist (they are basically just an argument from ignorance at present). Something about one or both of these might allow survival somehow. [NEWLINE] [NEWLINE] All of these,<mask> (and, I believe, any other way that humans could survive heat death), rely on physics that we don't understand well enough today to say that we know it's possible. [NEWLINE] [NEWLINE] By the physics we know today, it's impossible. You can't have a perfect perpetual motion machine, and<mask> anything you devise will eventually run down. [NEWLINE] [NEWLINE] Positing that something will stop the heat death of the universe in time to survive is not really "surviving the heat death of the universe", it's saying that you think there won't ever *be* a "heat death of the universe". That's a very different statement. Might be true,<mask> our current theories say otherwise. </s>
Label encoding: <s>I think your approach doesn't hold much hope for this, because all it does is postpone your death until the actual heat-death of the universe. I would challenge that part of your view. [NEWLINE] [NEWLINE] However, it does seem possible that at least some scenarios could allow survival. [NEWLINE] [NEWLINE] One is the creation of another universe that we could somehow transfer ourselves into. We don't currently understand how universes come to be, so this seems implausible today, but we have billions of years to figure it out. [NEWLINE] [NEWLINE] Another is that quantum mechanics might possibly make it impossible for the universe to *actually* achieve perfect heat-death. It might be the case that there is always a fluctuating quantum field that can be used as a minute energy source. A sufficiently clever species might find a way to spread their consciousness over a sufficiently large area that this minute amount of energy could sustain them. [NEWLINE] [NEWLINE] Another is that we *really* don't understand anything about the missing majority of the universe we're currently calling Dark Matter and Dark Energy. We don't know what laws they follow, we don't really know anything about them, or even for certain know that they exist (they are basically just an argument from ignorance at present). Something about one or both of these might allow survival somehow. [NEWLINE] [NEWLINE] All of these, though (and, I believe, any other way that humans could survive heat death), rely on physics that we don't understand well enough today to say that we know it's possible. [NEWLINE] [NEWLINE] By the physics we know today, it's impossible. You can't have a perfect perpetual motion machine, and so anything you devise will eventually run down. [NEWLINE] [NEWLINE] Positing that something will stop the heat death of the universe in time to survive is not really "surviving the heat death of the universe", it's saying that you think there won't ever *be* a "heat death of the universe". That's a very different statement. Might be true, but our current theories say otherwise. </s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I try to follow a general rule in life: don't act like you know<mask> other peoples' lives are like unless you've lived them. [NEWLINE] [NEWLINE] You're making a pretty bold claim about racism solely<mask> you-<mask> a white male living in a rural area-haven't experienced it. Of course you aren't going to experience it. You're white. [NEWLINE] [NEWLINE] About the Ebonics/rap aesthetic, consider this:  most white people have a pretty good idea<mask> they came from. My great-grandmother emigrated here from the Netherlands. I<mask> have relatives who came from Scotland. I have a cultural identity. [NEWLINE] [NEWLINE] Now, ask a black person<mask> in Africa their ancestors came from. The vast majority of them can't answer<mask> the only way to find out is with a DNA test, and that barely gives you an answer more specific than "Africa."  Every black person in America has a gaping hole in their history the rest of us don't have. And that has consequences. [NEWLINE] [NEWLINE] We like to think of America<mask> a melting pot of cultures. We have input from Italians, Germans, Dutch, Irish, Japanese, Chinese, etc. And we celebrate these contributions. Black people want to and deserve to have input<mask> well,<mask> instead of using ancient traditions like the rest of us, they have to invent new ones.<mask> they don't know<mask> they're Ethiopian or Kenyan or Somali or Nigerian. All they know is that they're "black."<mask> they invent things like rap, and baggy pants, and Kwanzaa, and<mask> happens? Derision, sneering, condemnation.<mask> white people complain about black cultural memes, it sends the message "<mask> can't you just be WHITE?" [NEWLINE] [NEWLINE] Tl;dr.<mask> of slavery, blacks have to invent cultural traditions to have input on the more universal American culture. Expecting them to just "act white"<mask> everybody else gets to have their own cultural input is insensitive and unfair. </s>
Label encoding: <s>I try to follow a general rule in life: don't act like you know what other peoples' lives are like unless you've lived them. [NEWLINE] [NEWLINE] You're making a pretty bold claim about racism solely because you- as a white male living in a rural area-haven't experienced it. Of course you aren't going to experience it. You're white. [NEWLINE] [NEWLINE] About the Ebonics/rap aesthetic, consider this:  most white people have a pretty good idea where they came from. My great-grandmother emigrated here from the Netherlands. I also have relatives who came from Scotland. I have a cultural identity. [NEWLINE] [NEWLINE] Now, ask a black person where in Africa their ancestors came from. The vast majority of them can't answer because the only way to find out is with a DNA test, and that barely gives you an answer more specific than "Africa."  Every black person in America has a gaping hole in their history the rest of us don't have. And that has consequences. [NEWLINE] [NEWLINE] We like to think of America as a melting pot of cultures. We have input from Italians, Germans, Dutch, Irish, Japanese, Chinese, etc. And we celebrate these contributions. Black people want to and deserve to have input as well, but instead of using ancient traditions like the rest of us, they have to invent new ones. Because they don't know If they're Ethiopian or Kenyan or Somali or Nigerian. All they know is that they're "black." So they invent things like rap, and baggy pants, and Kwanzaa, and what happens? Derision, sneering, condemnation. When white people complain about black cultural memes, it sends the message " why can't you just be WHITE?" [NEWLINE] [NEWLINE] Tl;dr. Because of slavery, blacks have to invent cultural traditions to have input on the more universal American culture. Expecting them to just "act white" when everybody else gets to have their own cultural input is insensitive and unfair. </s>
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Unless there's evidence to say otherwise<mask><mask> its fair to assume that most cases the right person went to prison. [ENDQ] [NEWLINE] I'd agree that in *most cases* they did,<mask> I'm not sure "most" is enough.  I mean, you're talking "over 50%" at that point. And<mask> you were to claim it's a 99% success ratio I'd start questioning it. [NEWLINE] [NEWLINE] [STARTQ] <mask> the person is innocent then they know this to be the case and should continue to argue back until they disprove the case the police put forward. [ENDQ] [NEWLINE] That's a great theory,<mask> it's not<mask> happens in reality. There are *plenty* of cases<mask> someone admits to something they didn't do<mask> they think it will result in a better outcome for them; there are even cases<mask> someone's mind is changed and they honestly think they did something which they didn't. [NEWLINE] [NEWLINE] [STARTQ] <mask> the person is guilty the police (<mask> they're doing this effectively) will be able to sketch out the details of<mask> the suspect actually did. The suspect knowing the details of<mask> happened and hearing the police describe it to them with the correct detail will believe that he's been found out<mask><mask> else would they know it? [ENDQ] [NEWLINE] Through [cold reading]( [URL] )? [NEWLINE] [NEWLINE] I mean there's plenty of ways you can read data from someone without them being aware you're doing it, and there's plenty of ways you can feed data to someone without them being aware it's happening. [NEWLINE] [NEWLINE] [STARTQ] I would agree that the protections aren't enforced<mask> that's a different problem than the one OP has asked. [ENDQ] [NEWLINE] I'd argue it's part of the same problem. The only way police interrogations are justified is<mask> there's protection against overly aggressive interrogations. Empirically, we can see these protections are not sufficient. Unless you have an idea on<mask> to *massively* strengthen these protections, I'd<mask><mask> police interrogation is not justified.</s>
Label encoding: <s> [STARTQ] Unless there's evidence to say otherwise I think its fair to assume that most cases the right person went to prison. [ENDQ] [NEWLINE] I'd agree that in *most cases* they did, but I'm not sure "most" is enough.  I mean, you're talking "over 50%" at that point. And if you were to claim it's a 99% success ratio I'd start questioning it. [NEWLINE] [NEWLINE] [STARTQ] If the person is innocent then they know this to be the case and should continue to argue back until they disprove the case the police put forward. [ENDQ] [NEWLINE] That's a great theory, but it's not what happens in reality. There are *plenty* of cases where someone admits to something they didn't do because they think it will result in a better outcome for them; there are even cases where someone's mind is changed and they honestly think they did something which they didn't. [NEWLINE] [NEWLINE] [STARTQ] If the person is guilty the police ( if they're doing this effectively) will be able to sketch out the details of what the suspect actually did. The suspect knowing the details of what happened and hearing the police describe it to them with the correct detail will believe that he's been found out because how else would they know it? [ENDQ] [NEWLINE] Through [cold reading]( [URL] )? [NEWLINE] [NEWLINE] I mean there's plenty of ways you can read data from someone without them being aware you're doing it, and there's plenty of ways you can feed data to someone without them being aware it's happening. [NEWLINE] [NEWLINE] [STARTQ] I would agree that the protections aren't enforced but that's a different problem than the one OP has asked. [ENDQ] [NEWLINE] I'd argue it's part of the same problem. The only way police interrogations are justified is if there's protection against overly aggressive interrogations. Empirically, we can see these protections are not sufficient. Unless you have an idea on how to *massively* strengthen these protections, I'd argue that police interrogation is not justified.</s>
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for your comments. I do fit the description of first world non-farmer. I'm a critical geographer by training, and I'm here to get more information,<mask> I have a fundamental understanding that in any structural shift of economics or agriculture, there are winners and losers,<mask> the saying goes.<mask> there is a continued shift towards designer varieties of seed that remove local autonomy in sourcing seed stocks, does it make sense that many farmers will lose some part of their agricultural/economic network and relationships? Some who do not want to shift will be forced to, or manipulated in unfair ways to get them to comply? [NEWLINE] [NEWLINE] I guess I need to understand<mask> it's a misconception that GMO business relationships with farmers are of any substantial difference compared to designer/hybrid varieties sold by corporations, otherwise GMO business practices are not really the issue. I am under the impression that GMO seed are unique in that they are usually engineered to work with a specific herbicide/pesticide that is proprietary, and must<mask> be bought from the same corporation that makes the seed. [NEWLINE] [NEWLINE] From the conclusions of the study in India,<mask> showing positive results from the sample of farmers who use Bt Cotton, [NEWLINE] [STARTQ] Appropriate policy and regulatory frameworks are required to ensure that the needs of poor farmers and consumers are taken into account and that undesirable social consequences are avoided. [ENDQ] [NEWLINE] I wish they would have addressed<mask> they mean here,<mask> this may be partly<mask> I'm interested in. The study mentioned that "93% of the country’s total cotton area" is already growing Bt Cotton. I wonder<mask> these other 7% have not switched over? [NEWLINE] [NEWLINE] Paternalistic? I don't know.<mask> many more articles about lack of resources in India do I need to see to assume small farmers are having a hard time? [NEWLINE] [NEWLINE] I need to collect more information about experiences with other GMO staple crops like rice and wheat, which<mask><mask> they grow a lot of in India. </s>
Label encoding: <s>Thanks for your comments. I do fit the description of first world non-farmer. I'm a critical geographer by training, and I'm here to get more information, but I have a fundamental understanding that in any structural shift of economics or agriculture, there are winners and losers, as the saying goes. If there is a continued shift towards designer varieties of seed that remove local autonomy in sourcing seed stocks, does it make sense that many farmers will lose some part of their agricultural/economic network and relationships? Some who do not want to shift will be forced to, or manipulated in unfair ways to get them to comply? [NEWLINE] [NEWLINE] I guess I need to understand if it's a misconception that GMO business relationships with farmers are of any substantial difference compared to designer/hybrid varieties sold by corporations, otherwise GMO business practices are not really the issue. I am under the impression that GMO seed are unique in that they are usually engineered to work with a specific herbicide/pesticide that is proprietary, and must also be bought from the same corporation that makes the seed. [NEWLINE] [NEWLINE] From the conclusions of the study in India, while showing positive results from the sample of farmers who use Bt Cotton, [NEWLINE] [STARTQ] Appropriate policy and regulatory frameworks are required to ensure that the needs of poor farmers and consumers are taken into account and that undesirable social consequences are avoided. [ENDQ] [NEWLINE] I wish they would have addressed what they mean here, because this may be partly what I'm interested in. The study mentioned that "93% of the country’s total cotton area" is already growing Bt Cotton. I wonder why these other 7% have not switched over? [NEWLINE] [NEWLINE] Paternalistic? I don't know. How many more articles about lack of resources in India do I need to see to assume small farmers are having a hard time? [NEWLINE] [NEWLINE] I need to collect more information about experiences with other GMO staple crops like rice and wheat, which I think they grow a lot of in India. </s>
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0694, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your question is framed in the same manner that most politicians do, i.e. relative to the status quo. It always sounds like a compelling idea to say "shouldn't we tax the 'rich' slightly more?" Shouldn't we punish violent criminals slightly more? Shouldn't we help the poor slightly more?<mask> on and<mask> forth. The real question that needs to be asked is one that isn't burdened and anchored to prior assumptions.<mask> is the appropriate level of taxation? And<mask> progressive should the tax structure be? [NEWLINE] [NEWLINE] <mask> it has been said in other responses, the answer (somewhat) depends on<mask> you mean by "slightly". The current tax structure in the US is very progressive. Check out page 7 of this report. Effective marginal tax rates for the poor are upward of 90%! [NEWLINE] [URL].pdf [NEWLINE] [NEWLINE] <mask><mask> the real issue with your question is this.<mask> you want people to do less of something, tax it. It works with almost unfailing efficiency. With economic issues, you must always consider the marginal example.<mask>, at very high income levels, marginal income doesn't necessarily impact quality of life in the same way<mask> it does at lower income levels. Essentially there diminishing marginal returns on income.<mask> I make $1mm a year, an extra $5,000 doesn't really move the needle for me. Certainly not in the way it would for someone making $20,000 a year.<mask> this high earner doesn't derive<mask> much utility from that extra income, he has less incentive to work for it. And now<mask> you tax it at an increasing rate, you've really started to impact that motivation. [NEWLINE] [NEWLINE] I can tell you from my own personal experience that today's tax rates at the top bracket definitely negatively affect my willingness and motivation to work harder. I make plenty of money, I am comfortable.<mask> would I bust my ass even more to give 40% of it away?</s>
Label encoding: <s>Your question is framed in the same manner that most politicians do, i.e. relative to the status quo. It always sounds like a compelling idea to say "shouldn't we tax the 'rich' slightly more?" Shouldn't we punish violent criminals slightly more? Shouldn't we help the poor slightly more? So on and so forth. The real question that needs to be asked is one that isn't burdened and anchored to prior assumptions. What is the appropriate level of taxation? And how progressive should the tax structure be? [NEWLINE] [NEWLINE] As it has been said in other responses, the answer (somewhat) depends on what you mean by "slightly". The current tax structure in the US is very progressive. Check out page 7 of this report. Effective marginal tax rates for the poor are upward of 90%! [NEWLINE] [URL].pdf [NEWLINE] [NEWLINE] I think the real issue with your question is this. If you want people to do less of something, tax it. It works with almost unfailing efficiency. With economic issues, you must always consider the marginal example. Also, at very high income levels, marginal income doesn't necessarily impact quality of life in the same way as it does at lower income levels. Essentially there diminishing marginal returns on income. If I make $1mm a year, an extra $5,000 doesn't really move the needle for me. Certainly not in the way it would for someone making $20,000 a year. Since this high earner doesn't derive as much utility from that extra income, he has less incentive to work for it. And now if you tax it at an increasing rate, you've really started to impact that motivation. [NEWLINE] [NEWLINE] I can tell you from my own personal experience that today's tax rates at the top bracket definitely negatively affect my willingness and motivation to work harder. I make plenty of money, I am comfortable. Why would I bust my ass even more to give 40% of it away?</s>
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm only referring to federal politics, by the way [ENDQ] [NEWLINE] It's kind of hard to separate out state and federal politics like this.  Winning state elections gives you control over redistricting which helps you to win in the House, and means you have plenty of experienced politicians in the state to choose from<mask> the time comes to nominate candidates for federal office.  Doing well in federal elections gets you the media exposure, activists and donors you need to win at the state level. [NEWLINE] [NEWLINE] [STARTQ] I'm pro-choice in all cases<mask> it's pretty much a fact that Roe v. Wade will never get overturned. [ENDQ] [NEWLINE] The Supreme Court came very close to overturning Roe v Wade in 1992, and Thomas and Scalia have consistently made it clear that they want to. <mask> we were to elect a Republican president or two, it wouldn't be that surprising<mask> they managed to appoint enough conservative justices for it to be overturned. <mask> there have been, and will probably continue to be, court cases and federal legislation regarding particular aspects of regulating abortion that fall short of completely banning it. [NEWLINE] [NEWLINE] [STARTQ] I'm pro-gay marriage<mask> its national spread is inevitable. [ENDQ] [NEWLINE] The spread will be very slow unless the Supreme Court finds a right to same-sex marriage.  With the Court's current make-up, it seems reasonably likely that there would be a 5-4 decision in favor of that.  Again, a Republican president could appoint another Scalia to replace Ginsburg, Breyer or Kennedy, destroying that possibility. [NEWLINE] [NEWLINE] There are<mask> other LGBT issues that are harder to advance through the states or the courts - a pretty obvious one being ENDA, which adds sexual orientation and gender identity to the protected classes covered by civil rights legislation (preventing businesses discriminating against gay employees, etc.).  It passed the Senate with unanimous support from Democrats and a few votes from moderate Republicans,<mask> seems unlikely to get anywhere in the House<mask> it is under Republican control.</s>
Label encoding: <s> [STARTQ] I'm only referring to federal politics, by the way [ENDQ] [NEWLINE] It's kind of hard to separate out state and federal politics like this.  Winning state elections gives you control over redistricting which helps you to win in the House, and means you have plenty of experienced politicians in the state to choose from when the time comes to nominate candidates for federal office.  Doing well in federal elections gets you the media exposure, activists and donors you need to win at the state level. [NEWLINE] [NEWLINE] [STARTQ] I'm pro-choice in all cases but it's pretty much a fact that Roe v. Wade will never get overturned. [ENDQ] [NEWLINE] The Supreme Court came very close to overturning Roe v Wade in 1992, and Thomas and Scalia have consistently made it clear that they want to.  If we were to elect a Republican president or two, it wouldn't be that surprising if they managed to appoint enough conservative justices for it to be overturned.  Also there have been, and will probably continue to be, court cases and federal legislation regarding particular aspects of regulating abortion that fall short of completely banning it. [NEWLINE] [NEWLINE] [STARTQ] I'm pro-gay marriage but its national spread is inevitable. [ENDQ] [NEWLINE] The spread will be very slow unless the Supreme Court finds a right to same-sex marriage.  With the Court's current make-up, it seems reasonably likely that there would be a 5-4 decision in favor of that.  Again, a Republican president could appoint another Scalia to replace Ginsburg, Breyer or Kennedy, destroying that possibility. [NEWLINE] [NEWLINE] There are also other LGBT issues that are harder to advance through the states or the courts - a pretty obvious one being ENDA, which adds sexual orientation and gender identity to the protected classes covered by civil rights legislation (preventing businesses discriminating against gay employees, etc.).  It passed the Senate with unanimous support from Democrats and a few votes from moderate Republicans, but seems unlikely to get anywhere in the House while it is under Republican control.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As to whether Christianity consciously stacked the deck in their favor or the system naturally evolved over time, I'm not entirely sure. [NEWLINE] [NEWLINE] <mask> I had to guess, I would say it's a combination of both. Christianity is an evolved religion.<mask> anything evolves, it adapts to the new climates the world takes on. [NEWLINE] [NEWLINE] Christianity is one of the dominant religions of our current culture, and<mask> you study the history of christianity you can see<mask> different points have been emphasized through propaganda at different points in time. [NEWLINE] [NEWLINE] For example, around the time of the crusades the church began commissioning the best painters in the world at the time to depict detailed scenes of hell. You can see<mask> they would need to do this in order to convince people to go to war. [NEWLINE] [NEWLINE] This is basic positive/negative reinforcement.<mask> you tell someone: [NEWLINE] [NEWLINE] -<mask> you perform X, then I will give you Y. [NEWLINE] [NEWLINE] Then their motivation will only run<mask> high<mask> all your using is positive reinforcement.<mask>,<mask> you add in negative reinforcement along with it: [NEWLINE] [NEWLINE] -<mask> you do X, I will give you Y.<mask> you don't do X, (insert bad thing) will happen to you. [NEWLINE] [NEWLINE] Telling someone they need to die for their religion requires you to play with both forms of motivation, positive and negative reinforcement. [NEWLINE] [NEWLINE] In my personal opinion,<mask><mask> that over time there were some popes and high officials in the church that were aware of the power Christianity has to persuade people simply by<mask> the framing of the Bible has evolved over time, and used that for their own personal gain.<mask><mask><mask> the majority of people who are using Christianity's framing power to convert people are doing it subconsciously. [NEWLINE] [NEWLINE] <mask>, I will admit I'm no expert on the history of Christianity and I would need to study up on it more before I feel my comment on the motion can be taken seriously. </s><pad>
Label encoding: <s>As to whether Christianity consciously stacked the deck in their favor or the system naturally evolved over time, I'm not entirely sure. [NEWLINE] [NEWLINE] If I had to guess, I would say it's a combination of both. Christianity is an evolved religion. As anything evolves, it adapts to the new climates the world takes on. [NEWLINE] [NEWLINE] Christianity is one of the dominant religions of our current culture, and if you study the history of christianity you can see how different points have been emphasized through propaganda at different points in time. [NEWLINE] [NEWLINE] For example, around the time of the crusades the church began commissioning the best painters in the world at the time to depict detailed scenes of hell. You can see why they would need to do this in order to convince people to go to war. [NEWLINE] [NEWLINE] This is basic positive/negative reinforcement. If you tell someone: [NEWLINE] [NEWLINE] - If you perform X, then I will give you Y. [NEWLINE] [NEWLINE] Then their motivation will only run so high because all your using is positive reinforcement. However, if you add in negative reinforcement along with it: [NEWLINE] [NEWLINE] - If you do X, I will give you Y. If you don't do X, (insert bad thing) will happen to you. [NEWLINE] [NEWLINE] Telling someone they need to die for their religion requires you to play with both forms of motivation, positive and negative reinforcement. [NEWLINE] [NEWLINE] In my personal opinion, I think that over time there were some popes and high officials in the church that were aware of the power Christianity has to persuade people simply by how the framing of the Bible has evolved over time, and used that for their own personal gain. But I think the majority of people who are using Christianity's framing power to convert people are doing it subconsciously. [NEWLINE] [NEWLINE] However, I will admit I'm no expert on the history of Christianity and I would need to study up on it more before I feel my comment on the motion can be taken seriously. </s><pad>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The "bullying" people are talking about, and the "coddling" people are talking about, are almost exclusively things we would call "felony"<mask> those involved are [STARTQ] 18. [ENDQ] [NEWLINE] <mask> someone locks you in a small closet and denies you escape, it's called felony Kidnapping.  In school, it's called a good laugh. [NEWLINE] [NEWLINE] <mask> someone jumps you with a baseball bat, it's called armed assault.  In school, it's called a scuffle. [NEWLINE] [NEWLINE] <mask> your peers ridicule and demean you, day in and day out, that's called harassment.  Corporate HR is<mask> zero-tolerance on that it's not funny.  One of my coworkers confronts me and calls me useless, their cube will be empty tomorrow.  In school, it's called "kids learning to socialize". [NEWLINE] [NEWLINE] <mask> you haven't seen *this* kind of bullying, and the way teachers have historically laughed and supported it, you don't know the real reason we're fighting it, and you obviously can't understand<mask> people are willing to end their own lives over it. [NEWLINE] [NEWLINE] Imagine<mask> people you had to see every day were openly allowed to commit violent felonies against you, and the cops would laugh it off.  You'd probably end up taking the law into your own hands before someone killed you. [NEWLINE] [NEWLINE] The school bully in my highschool escalated to jumping people with baseball bats at their homes, and attempted to stab his father... until then, people laughed even<mask> he walked down halls at 16 years of age punching random people at full strength<mask><mask><mask> HE COULD. <mask>? <mask> it's "part of growing up and we can't coddle our children". [NEWLINE] [NEWLINE] <mask> is it that our police have come to the point<mask> almost every adult physical conflict is considered a criminal act,<mask> kids can be hospitalized and those who opposed bullying are still being ridiculed for it?</s>
Label encoding: <s>The "bullying" people are talking about, and the "coddling" people are talking about, are almost exclusively things we would call "felony" when those involved are [STARTQ] 18. [ENDQ] [NEWLINE] If someone locks you in a small closet and denies you escape, it's called felony Kidnapping.  In school, it's called a good laugh. [NEWLINE] [NEWLINE] If someone jumps you with a baseball bat, it's called armed assault.  In school, it's called a scuffle. [NEWLINE] [NEWLINE] When your peers ridicule and demean you, day in and day out, that's called harassment.  Corporate HR is so zero-tolerance on that it's not funny.  One of my coworkers confronts me and calls me useless, their cube will be empty tomorrow.  In school, it's called "kids learning to socialize". [NEWLINE] [NEWLINE] If you haven't seen *this* kind of bullying, and the way teachers have historically laughed and supported it, you don't know the real reason we're fighting it, and you obviously can't understand why people are willing to end their own lives over it. [NEWLINE] [NEWLINE] Imagine if people you had to see every day were openly allowed to commit violent felonies against you, and the cops would laugh it off.  You'd probably end up taking the law into your own hands before someone killed you. [NEWLINE] [NEWLINE] The school bully in my highschool escalated to jumping people with baseball bats at their homes, and attempted to stab his father... until then, people laughed even as he walked down halls at 16 years of age punching random people at full strength BECAUSE HE COULD.  Why?  Because it's "part of growing up and we can't coddle our children". [NEWLINE] [NEWLINE] Why is it that our police have come to the point where almost every adult physical conflict is considered a criminal act, but kids can be hospitalized and those who opposed bullying are still being ridiculed for it?</s>
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Psychologically and physiologically our first reaction to aggressive actions is defense,<mask> resisting arrest statutes are legally punishing citizens for their natural instincts. [ENDQ] [NEWLINE] This seems to be your primary reason for<mask> resisting arrest should be reasonable. I'm not saying its wrong,<mask><mask><mask><mask><mask> this is the best reasoning for<mask> arrests are<mask> bad. Humans have other "natural instincts"<mask><mask><mask> following one's mind, logically, is a better measure of a situation than naturally following your innate desire to rebel against arrest. [NEWLINE] [NEWLINE] Now,<mask><mask> that jail time/arrest time is bad and could be significantly bad in some cases (bear in mind, I've never been in jail and this is only things I've read). Time wasted in jail<mask> your family, friends, or coworkers may need you, potential beatings/r***, and just mental trauma are<mask> awful and ideally no one would have to go through that. In that sense, I can understand many people not wanting to be arrested. (<mask> I missed any crucial details, please let me know) [NEWLINE] [NEWLINE] <mask>, society isn't perfect, and there's no way (currently at least) to immediately test a suspect.<mask><mask><mask> it comes down to a choice between the police's suspicions of an individual versus the individual's needs.<mask> I believe that<mask> individuals are given the right to resist arrest and police aren't able to stop them, this can lead to a lot of scenarios of truly bad people getting away with crimes. In exchange, you'll get some good people put away for hopefully short periods of time, and<mask><mask> that's a worthwhile tradeoff.<mask> something horrible happens to the innocent person in jail/he's treated poorly, they would hopefully be able to recoup their losses in court. [NEWLINE] [NEWLINE] I would<mask> address<mask> many think a lot of police seem to get off easily/not<mask> easily,<mask><mask><mask> that's a whole 'nother conversation.</s>
Label encoding: <s> [STARTQ] Psychologically and physiologically our first reaction to aggressive actions is defense, so resisting arrest statutes are legally punishing citizens for their natural instincts. [ENDQ] [NEWLINE] This seems to be your primary reason for why resisting arrest should be reasonable. I'm not saying its wrong, but I do not think this is the best reasoning for why arrests are so bad. Humans have other "natural instincts" but I think following one's mind, logically, is a better measure of a situation than naturally following your innate desire to rebel against arrest. [NEWLINE] [NEWLINE] Now, I agree that jail time/arrest time is bad and could be significantly bad in some cases (bear in mind, I've never been in jail and this is only things I've read). Time wasted in jail when your family, friends, or coworkers may need you, potential beatings/r***, and just mental trauma are indeed awful and ideally no one would have to go through that. In that sense, I can understand many people not wanting to be arrested. ( if I missed any crucial details, please let me know) [NEWLINE] [NEWLINE] But, society isn't perfect, and there's no way (currently at least) to immediately test a suspect. So I think it comes down to a choice between the police's suspicions of an individual versus the individual's needs. But I believe that if individuals are given the right to resist arrest and police aren't able to stop them, this can lead to a lot of scenarios of truly bad people getting away with crimes. In exchange, you'll get some good people put away for hopefully short periods of time, and I think that's a worthwhile tradeoff. If something horrible happens to the innocent person in jail/he's treated poorly, they would hopefully be able to recoup their losses in court. [NEWLINE] [NEWLINE] I would also address how many think a lot of police seem to get off easily/not so easily, but I think that's a whole 'nother conversation.</s>
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*Meant* by whom. By your same argument, we're meant to eat cake and pizza<mask> there are good evolutionary reasons<mask> we crave sugars and fats. [NEWLINE] [NEWLINE] The availability of sex today is nothing like it was back then. You can't judge an *ought* ("meant to") from an *is*, or this case, a *was*. [NEWLINE] [NEWLINE] <mask>, to counter your point, *love* evolved<mask> a way to keep us in pairs rather than serial partners. [NEWLINE] [NEWLINE] Edit: Generally speaking, males and females have different best interests<mask> it comes to reproduction,<mask> there isn't even a human "meant to". Males tend to maximize reproductive success,<mask> a baseline, by impregnating<mask> many females<mask> possible. This is<mask> the cost of reproduction is<mask> low for men, and handful of calories and a ~~30~~ 2 minutes of their time. For females it's a huge investment of calories, time, risk of predators, difficult in gathering calories, then raising the offspring. [NEWLINE] [NEWLINE] <mask> females benefit by only having sex with males who are willing to stay around to help acquire calories and protect the woman and child from predators, and help raise the child. (We're talking over evolutionary times and wild conditions, not modern society.)<mask> men end up maximizing reproductive success by actually being that type of guy that women chose to have sex with, which includes signs of commitment. [NEWLINE] [NEWLINE] <mask> note that men have no value in committing to an unfaithful woman and end up raising some other guy's child without knowing it.<mask> guys will tend to chose women who<mask> show signs of commitment. [NEWLINE] [NEWLINE] <mask><mask><mask> we have at least serial monogamy<mask><mask><mask> -- love that keeps people together and monogamous for "long enough". There are other strategies that work. Orangutans mate for life. Bonobos have sex often with everybody and tend to raise children communally, for instance. [NEWLINE] </s>
Label encoding: <s>*Meant* by whom. By your same argument, we're meant to eat cake and pizza because there are good evolutionary reasons why we crave sugars and fats. [NEWLINE] [NEWLINE] The availability of sex today is nothing like it was back then. You can't judge an *ought* ("meant to") from an *is*, or this case, a *was*. [NEWLINE] [NEWLINE] However, to counter your point, *love* evolved as a way to keep us in pairs rather than serial partners. [NEWLINE] [NEWLINE] Edit: Generally speaking, males and females have different best interests when it comes to reproduction, so there isn't even a human "meant to". Males tend to maximize reproductive success, as a baseline, by impregnating as many females as possible. This is because the cost of reproduction is so low for men, and handful of calories and a ~~30~~ 2 minutes of their time. For females it's a huge investment of calories, time, risk of predators, difficult in gathering calories, then raising the offspring. [NEWLINE] [NEWLINE] So females benefit by only having sex with males who are willing to stay around to help acquire calories and protect the woman and child from predators, and help raise the child. (We're talking over evolutionary times and wild conditions, not modern society.) Hence men end up maximizing reproductive success by actually being that type of guy that women chose to have sex with, which includes signs of commitment. [NEWLINE] [NEWLINE] Also note that men have no value in committing to an unfaithful woman and end up raising some other guy's child without knowing it. So guys will tend to chose women who also show signs of commitment. [NEWLINE] [NEWLINE] As a result we have at least serial monogamy as a result -- love that keeps people together and monogamous for "long enough". There are other strategies that work. Orangutans mate for life. Bonobos have sex often with everybody and tend to raise children communally, for instance. [NEWLINE] </s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>,<mask> of the consequences for family, friends, society<mask> a whole? Certainly kidnapping the world's top AIDS researcher wouldn't be justified,<mask><mask> he would maximize his own happiness. [ENDQ] [NEWLINE] <mask>? The researcher would experience the greatest possible happiness possible. In his virtual world he'd cure AIDS and every other disease and fulfill every dream he has. Maybe he'd regain the use of his leg, get married to the love of his life. Denying this happiness to him seems unjustified to me. [NEWLINE] [NEWLINE] [STARTQ] <mask>, assume that the hook up is only temporary.<mask> plans have you interfered with? Did the person miss something important? Alternatively,<mask> you take someone OFF the machine, life will undoubtedly seem unbearable after a time of pure happiness. [ENDQ] [NEWLINE] I concede that a temporary experience machine would make you want to kill yourself afterwards. Assume in this case either that the machine lasts<mask><mask><mask> the person connected. [NEWLINE] [NEWLINE] [STARTQ] Another point to consider is that not everyone prioritizes pure hedons. I'm not convinced<mask> well that a life with more hedons is more valuable. [ENDQ] [NEWLINE] The machine is not a 24/7 sex drugs and rocknroll party. It gives you the most valuable mental state you have.<mask> you get happiness from a life of monk-like ascetic living, you will get that in the machine. [NEWLINE] [NEWLINE] [STARTQ] &gt;<mask>,<mask> such an experience machine is invented, it would always be justified to plug<mask> many people into the machine<mask> possible, no matter<mask> pain is involved in the process. It would be immoral to deny the greatest possible happiness to someone. [ENDQ] [NEWLINE] [STARTQ] You can maximize individual happiness<mask> decreasing total happiness. [ENDQ] [NEWLINE] I'm not sure I understand this part. Plugging everyone into the machine would seem like the most moral act imaginable to me. Everyone would be experiencing the greatest possible happiness. The end of suffering.<mask> can you argue against that? [NEWLINE] </s>
Label encoding: <s> [STARTQ] However, what of the consequences for family, friends, society as a whole? Certainly kidnapping the world's top AIDS researcher wouldn't be justified, even though he would maximize his own happiness. [ENDQ] [NEWLINE] Why? The researcher would experience the greatest possible happiness possible. In his virtual world he'd cure AIDS and every other disease and fulfill every dream he has. Maybe he'd regain the use of his leg, get married to the love of his life. Denying this happiness to him seems unjustified to me. [NEWLINE] [NEWLINE] [STARTQ] Also, assume that the hook up is only temporary. What plans have you interfered with? Did the person miss something important? Alternatively, if you take someone OFF the machine, life will undoubtedly seem unbearable after a time of pure happiness. [ENDQ] [NEWLINE] I concede that a temporary experience machine would make you want to kill yourself afterwards. Assume in this case either that the machine lasts as long as the person connected. [NEWLINE] [NEWLINE] [STARTQ] Another point to consider is that not everyone prioritizes pure hedons. I'm not convinced as well that a life with more hedons is more valuable. [ENDQ] [NEWLINE] The machine is not a 24/7 sex drugs and rocknroll party. It gives you the most valuable mental state you have. If you get happiness from a life of monk-like ascetic living, you will get that in the machine. [NEWLINE] [NEWLINE] [STARTQ] &gt; Thus, when such an experience machine is invented, it would always be justified to plug as many people into the machine as possible, no matter what pain is involved in the process. It would be immoral to deny the greatest possible happiness to someone. [ENDQ] [NEWLINE] [STARTQ] You can maximize individual happiness while decreasing total happiness. [ENDQ] [NEWLINE] I'm not sure I understand this part. Plugging everyone into the machine would seem like the most moral act imaginable to me. Everyone would be experiencing the greatest possible happiness. The end of suffering. How can you argue against that? [NEWLINE] </s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> it would be immoral to bring a child into the world without being capable of providing for it.<mask><mask> it is<mask> immoral to force a man to pay for a child he didn't want.<mask> a woman can't raise the child without the father's support the moral thing to do would be to abort the child. [ENDQ] [NEWLINE] Sure, sure.<mask> on a societal scale, we can't ignore results - it isn't enough to focus on<mask> is moral. [NEWLINE] [NEWLINE] The likely practical outcome of this policy would be more children growing up with less financial support.<mask> of a variety of reasons: [NEWLINE] [NEWLINE] 1. The mother is wishful and thinks the father will "come around". [NEWLINE] 2. The mother sees abortion<mask> immoral. [NEWLINE] 3. The mother has a hard time deciding, and accepts the "default" - not to medically intervene, i.e., carry to term. [NEWLINE] 4. The mother has medical concerns that make abortion risky for some reason. [NEWLINE] [NEWLINE] and on the father's side [NEWLINE] [NEWLINE] 1. Having a foolproof method of getting out of child support means men don't need to be<mask> careful about using condoms ("<mask> she gets pregnant, I'll sign a piece of paper"). That means more unsafe sex, leading to both more pregnancies and more STDs and<mask> forth. [NEWLINE] [NEWLINE] We don't live in a perfect world. It would be nice<mask> men could have all the sex they want, and<mask> the woman happens to get pregnant<mask> taking precautions, they can avoid that affecting the rest of their life. That's great for the men.<mask> for society, it means more children growing up with less financial support. Which we know causes serious problems. [NEWLINE] [NEWLINE] Now sure, you have some guesses about this all working out for the best.<mask> even in the **most** optimistic case, you are talking about an experiment that no human society has ever done, with potentially grevious results.</s>
Label encoding: <s> [STARTQ] I think it would be immoral to bring a child into the world without being capable of providing for it. I think it is also immoral to force a man to pay for a child he didn't want. If a woman can't raise the child without the father's support the moral thing to do would be to abort the child. [ENDQ] [NEWLINE] Sure, sure. But on a societal scale, we can't ignore results - it isn't enough to focus on what is moral. [NEWLINE] [NEWLINE] The likely practical outcome of this policy would be more children growing up with less financial support. Because of a variety of reasons: [NEWLINE] [NEWLINE] 1. The mother is wishful and thinks the father will "come around". [NEWLINE] 2. The mother sees abortion as immoral. [NEWLINE] 3. The mother has a hard time deciding, and accepts the "default" - not to medically intervene, i.e., carry to term. [NEWLINE] 4. The mother has medical concerns that make abortion risky for some reason. [NEWLINE] [NEWLINE] and on the father's side [NEWLINE] [NEWLINE] 1. Having a foolproof method of getting out of child support means men don't need to be as careful about using condoms (" if she gets pregnant, I'll sign a piece of paper"). That means more unsafe sex, leading to both more pregnancies and more STDs and so forth. [NEWLINE] [NEWLINE] We don't live in a perfect world. It would be nice if men could have all the sex they want, and if the woman happens to get pregnant despite taking precautions, they can avoid that affecting the rest of their life. That's great for the men. But for society, it means more children growing up with less financial support. Which we know causes serious problems. [NEWLINE] [NEWLINE] Now sure, you have some guesses about this all working out for the best. But even in the **most** optimistic case, you are talking about an experiment that no human society has ever done, with potentially grevious results.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Drum kits are pretty common for 12 gauges. You can use regular vertically stacked magazines too, once you install the kit. [ENDQ] [NEWLINE] I don't doubt this,<mask> many 12 gauge shotguns are pump action which do not take vertical magazines. My Remington 870 has a 6+1 magazine tube. [NEWLINE] [NEWLINE] [STARTQ] I understand the issue,<mask> at the range, me and another guy were shooting at target with our shotguns. His KelTek shotgun ran out,<mask> mine was still going.<mask>? I kept shooting by quickly grabbing a round out of the box and tossing it into the ejector port, chambering a round.<mask> it's in reach, consider it loaded. [ENDQ] [NEWLINE] You're not always going to be standing next to a box of shotgun shells.<mask> you get woken up in the middle of the night<mask> you heard someone break into your house, do you think you'll be in the best shape to reload something like a pump action shotgun? Sure, it's not terribly difficult,<mask> it's nowhere<mask> simple<mask> loading an AR15. [NEWLINE] [NEWLINE] Plus, depending on which US State you live in, you might only need one magazine (in terms of round capacity). [NEWLINE] [NEWLINE] AR15's are very straightforward to reload, and can easily become muscle memory with a bit of practice. [NEWLINE] [NEWLINE] You could<mask> stick a mag in the waistband of your underwear (<mask> that's<mask> you roll) or hold a mag in your non-dominant hand<mask> shooting. It's not ideal,<mask><mask> is a home defense situation ever ideal? [NEWLINE] [NEWLINE] The idea of using a gun in self defense revolves around situations<mask> your life (or the lives of your loved ones) is in danger. These situations are never ideal, which is<mask> you try to minimize any risks.<mask><mask> an AR15 is easier to operate, and<mask> more reliable than other options such<mask> a 12 gauge shotgun.</s>
Label encoding: <s> [STARTQ] Drum kits are pretty common for 12 gauges. You can use regular vertically stacked magazines too, once you install the kit. [ENDQ] [NEWLINE] I don't doubt this, but many 12 gauge shotguns are pump action which do not take vertical magazines. My Remington 870 has a 6+1 magazine tube. [NEWLINE] [NEWLINE] [STARTQ] I understand the issue, but at the range, me and another guy were shooting at target with our shotguns. His KelTek shotgun ran out, but mine was still going. Why? I kept shooting by quickly grabbing a round out of the box and tossing it into the ejector port, chambering a round. If it's in reach, consider it loaded. [ENDQ] [NEWLINE] You're not always going to be standing next to a box of shotgun shells. If you get woken up in the middle of the night because you heard someone break into your house, do you think you'll be in the best shape to reload something like a pump action shotgun? Sure, it's not terribly difficult, but it's nowhere as simple as loading an AR15. [NEWLINE] [NEWLINE] Plus, depending on which US State you live in, you might only need one magazine (in terms of round capacity). [NEWLINE] [NEWLINE] AR15's are very straightforward to reload, and can easily become muscle memory with a bit of practice. [NEWLINE] [NEWLINE] You could also stick a mag in the waistband of your underwear ( if that's how you roll) or hold a mag in your non-dominant hand while shooting. It's not ideal, but when is a home defense situation ever ideal? [NEWLINE] [NEWLINE] The idea of using a gun in self defense revolves around situations where your life (or the lives of your loved ones) is in danger. These situations are never ideal, which is why you try to minimize any risks. IMO an AR15 is easier to operate, and thus more reliable than other options such as a 12 gauge shotgun.</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] I feel that I'm not making it simplistic<mask> you're making it complicated<mask> it really is about things that people willingly take in their hand, willingly put in their mouth and willingly chew and swallow. [NEWLINE] [NEWLINE] <mask> I already was presenting a very effective solution. People that feel that  they are fat should eat less. I understand that you feel that this is making it too easy. I<mask><mask><mask> think you're making it to easy for you by pointing out all the factors in society that promote obesity, acting<mask><mask> it was not you who was in charge of your life. [NEWLINE] [NEWLINE] It's not that we really dissent, we just view the problem from very different perspectives. You argue from the larger picture of cultural influences on the individual, I'm viewing it from the other end of the spectrum, the physical deeds of the obese individual. You say I'm to simplistic and you might be right,<mask> right<mask> I am<mask> I say that you generalizing don't help the problem either. [NEWLINE] [NEWLINE] <mask><mask> your arguments didn't change my view this far, I recognize that the behavior that I criticize is influenced by external factors. That brings us down to the question of free will, which is the only valid attack vector on my argument I see this far. [NEWLINE] [NEWLINE] EDIT:<mask><mask><mask> about it,  bringing me to this realization is supposed to be rewarded a delta<mask><mask><mask> I understand.<mask> my arguments holds up under the preface that free will is a fact, it's obviously totally naught<mask> free will is an illusion and our behavior is determined by conditioning, socialization and genetics.<mask> this would be the case, you'd be totally right, and by now I have no certain means to come to a conclusion about the question of the free will. Still I<mask> everyone else live under the impression that I act free,<mask> I have to argue from that viewpoint.</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] I feel that I'm not making it simplistic but you're making it complicated because it really is about things that people willingly take in their hand, willingly put in their mouth and willingly chew and swallow. [NEWLINE] [NEWLINE] So I already was presenting a very effective solution. People that feel that  they are fat should eat less. I understand that you feel that this is making it too easy. I on the contrary think you're making it to easy for you by pointing out all the factors in society that promote obesity, acting as if it was not you who was in charge of your life. [NEWLINE] [NEWLINE] It's not that we really dissent, we just view the problem from very different perspectives. You argue from the larger picture of cultural influences on the individual, I'm viewing it from the other end of the spectrum, the physical deeds of the obese individual. You say I'm to simplistic and you might be right, as right as I am when I say that you generalizing don't help the problem either. [NEWLINE] [NEWLINE] So while your arguments didn't change my view this far, I recognize that the behavior that I criticize is influenced by external factors. That brings us down to the question of free will, which is the only valid attack vector on my argument I see this far. [NEWLINE] [NEWLINE] EDIT: While I think about it,  bringing me to this realization is supposed to be rewarded a delta as far as I understand. While my arguments holds up under the preface that free will is a fact, it's obviously totally naught if free will is an illusion and our behavior is determined by conditioning, socialization and genetics. If this would be the case, you'd be totally right, and by now I have no certain means to come to a conclusion about the question of the free will. Still I as everyone else live under the impression that I act free, so I have to argue from that viewpoint.</s>
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The existence of other states is an existential threat to the openness of sea lanes,<mask> /u/OneBildoNation and /u/SignoreReddit discuss in general terms above. [NEWLINE] [NEWLINE] There are several geographic chokepoints that could be easily closed off to basically individually obliterate the world economy, such<mask> the Straits of Malacca or Hormuz. Without fear of an American reaction, other states such<mask> China for Malacca or Iran or Saudi Arabia for Hormuz would be tempted to seize them for themselves to impose a tax on the shipping of goods through them or more likely to use the threat of reducing transit through them<mask> leverage over other nations that rely on them. Even<mask> states don't actually have such an aggressive intent, the states that rely on shipping through the chokepoints would still feel compelled to seize control of them to prevent the possibility of that situation arising, which in turn would raise the same fears in other states. Either way creates a likelihood of war, possibly extraordinarily damaging wars given the importance of the chokepoints. [NEWLINE] [NEWLINE] <mask><mask> your personal opinion of this analysis, these are precisely the assumptions that the world's governments are operating under. The key element of China's military development is strengthening its already formidable anti-access/area-denial capabilities to prevent U.S. weapons platforms from deploying to the area and more generally limit freedom of movement in the region. The U.S. military's effort to respond to this development is the core of Obama's "pivot to Asia." [NEWLINE] [NEWLINE] On a historical note, military struggles over those chokepoints were common, with the Portuguese and then British achieving particular success at controlling them. The U.S. obviously doesn't directly control them,<mask> its overwhelming military power and historically benign exercise of it (e.g., not seizing direct control and imposing a tax on the movement of goods through it) prevents such struggles from arising.</s><pad>
Label encoding: <s>The existence of other states is an existential threat to the openness of sea lanes, as /u/OneBildoNation and /u/SignoreReddit discuss in general terms above. [NEWLINE] [NEWLINE] There are several geographic chokepoints that could be easily closed off to basically individually obliterate the world economy, such as the Straits of Malacca or Hormuz. Without fear of an American reaction, other states such as China for Malacca or Iran or Saudi Arabia for Hormuz would be tempted to seize them for themselves to impose a tax on the shipping of goods through them or more likely to use the threat of reducing transit through them as leverage over other nations that rely on them. Even if states don't actually have such an aggressive intent, the states that rely on shipping through the chokepoints would still feel compelled to seize control of them to prevent the possibility of that situation arising, which in turn would raise the same fears in other states. Either way creates a likelihood of war, possibly extraordinarily damaging wars given the importance of the chokepoints. [NEWLINE] [NEWLINE] Regardless of your personal opinion of this analysis, these are precisely the assumptions that the world's governments are operating under. The key element of China's military development is strengthening its already formidable anti-access/area-denial capabilities to prevent U.S. weapons platforms from deploying to the area and more generally limit freedom of movement in the region. The U.S. military's effort to respond to this development is the core of Obama's "pivot to Asia." [NEWLINE] [NEWLINE] On a historical note, military struggles over those chokepoints were common, with the Portuguese and then British achieving particular success at controlling them. The U.S. obviously doesn't directly control them, but its overwhelming military power and historically benign exercise of it (e.g., not seizing direct control and imposing a tax on the movement of goods through it) prevents such struggles from arising.</s><pad>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Training is a lot more than expenses. It's<mask> a time and trust thing too. I may love a new puppy<mask> much<mask> I loved the dog I just lost.<mask> it's going to take time to trust them. For some dogs it can take upwards of 2 years. And by trust I mean trust around the younger and older members of my family, trust that they will come back<mask> they get loose, trust that they won't be defensive of other pets, etc. Not to mention the trust the other way. My dog needs to learn to trust that I'll be back<mask> I leave, that I will feed them. All of this comes with a good relationship and a good relationship involves at least basic training and is worth much more than just the cost of training. [NEWLINE] [NEWLINE] The general rule with horses is you save up at least<mask> the horse is worth. A well trained horse can be worth $10,000-$15,000.<mask> anything you pay<mask> it would cost to get a new horse.<mask><mask> this applies to dogs and cats, the issue is people don't usually sell dogs and cats the way people sell horses,<mask> there's no understood value.<mask> you can try to work out a value.<mask> you were going to lose your dog in its prime,<mask> would it cost you to get the exact dog again or at least something close enough (a dog that fits in your family and is<mask> trusted<mask> this dog)? The cost of the dog, the cost of the required vet visits, the cost of any new equipment you'd need, the cost of training, and the value of your time put into 1-2 years of training, and<mask> much it's worth to not deal with the sadness of losing your pet. Even a $200 Mutt from the pound can end up being worth $1,500-$2,000<mask> you factor those things in.</s>
Label encoding: <s>Training is a lot more than expenses. It's also a time and trust thing too. I may love a new puppy as much as I loved the dog I just lost. But it's going to take time to trust them. For some dogs it can take upwards of 2 years. And by trust I mean trust around the younger and older members of my family, trust that they will come back if they get loose, trust that they won't be defensive of other pets, etc. Not to mention the trust the other way. My dog needs to learn to trust that I'll be back when I leave, that I will feed them. All of this comes with a good relationship and a good relationship involves at least basic training and is worth much more than just the cost of training. [NEWLINE] [NEWLINE] The general rule with horses is you save up at least what the horse is worth. A well trained horse can be worth $10,000-$15,000. If anything you pay what it would cost to get a new horse. I think this applies to dogs and cats, the issue is people don't usually sell dogs and cats the way people sell horses, so there's no understood value. But you can try to work out a value. If you were going to lose your dog in its prime, what would it cost you to get the exact dog again or at least something close enough (a dog that fits in your family and is as trusted as this dog)? The cost of the dog, the cost of the required vet visits, the cost of any new equipment you'd need, the cost of training, and the value of your time put into 1-2 years of training, and how much it's worth to not deal with the sadness of losing your pet. Even a $200 Mutt from the pound can end up being worth $1,500-$2,000 when you factor those things in.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I realized this several comments down the chain, and gave /u/DHCKris some examples. [NEWLINE] [NEWLINE] I'll give them to you<mask> well: [NEWLINE] [NEWLINE] GINO – (a) Godzilla In Name Only [NEWLINE] GEMA – (a) Gas and Electricity Markets Authority – Georgia Emergency Management Agency – (p) Gesellschaft für musikalische Aufführungs- und mechanische Vervielfältigungsrechte (German performance rights organisation) – (a) Global Engine Manufacturing Alliance [NEWLINE] GIMP – (a) GNU Image Manipulation Program [NEWLINE] GLAC – (a) General Ledger Accounting Code [NEWLINE] GLUT – (a/i) OpenGL Utility Toolkit [NEWLINE] GOES – (a) Geostationary Operational Environmental Satellite [NEWLINE] CASA - (a) Civil Aviation Safety Authority [NEWLINE] CASE (a) Cellular ammunition storage equipment [NEWLINE] CBASSE – (a) Commission on Behavioral And Social Sciences and Education ("sea bass") [NEWLINE] CERT – (a) Computer Emergency Response Team [NEWLINE] CEPT – (a/i) Conférence européenne des administrations des postes et des télécommunications (French, "European Conference of Postal and Telecommunications Administrations") [NEWLINE] CFES – (a) Continuous Flow Electrophoresis System [NEWLINE] CHAPS - (a) Clearing House Automated Payment System [NEWLINE] CHAMPUS – (a) (U.S.) Civilian Health and Medical Program of the Uniformed Services (now known<mask> TRICARE) [NEWLINE] CINEOS – (a) Campo Imperatore Near-Earth Object Survey [NEWLINE] CIVETS – (a) Colombia, Indonesia, Vietnam, Egypt, Turkey, South Africa (economics) [NEWLINE] CME – (a) Coronal Mass Ejection (usually Sun) [NEWLINE] CNI – (a) Clinical Nursing Intern</s>
Label encoding: <s>I realized this several comments down the chain, and gave /u/DHCKris some examples. [NEWLINE] [NEWLINE] I'll give them to you as well: [NEWLINE] [NEWLINE] GINO – (a) Godzilla In Name Only [NEWLINE] GEMA – (a) Gas and Electricity Markets Authority – Georgia Emergency Management Agency – (p) Gesellschaft für musikalische Aufführungs- und mechanische Vervielfältigungsrechte (German performance rights organisation) – (a) Global Engine Manufacturing Alliance [NEWLINE] GIMP – (a) GNU Image Manipulation Program [NEWLINE] GLAC – (a) General Ledger Accounting Code [NEWLINE] GLUT – (a/i) OpenGL Utility Toolkit [NEWLINE] GOES – (a) Geostationary Operational Environmental Satellite [NEWLINE] CASA - (a) Civil Aviation Safety Authority [NEWLINE] CASE (a) Cellular ammunition storage equipment [NEWLINE] CBASSE – (a) Commission on Behavioral And Social Sciences and Education ("sea bass") [NEWLINE] CERT – (a) Computer Emergency Response Team [NEWLINE] CEPT – (a/i) Conférence européenne des administrations des postes et des télécommunications (French, "European Conference of Postal and Telecommunications Administrations") [NEWLINE] CFES – (a) Continuous Flow Electrophoresis System [NEWLINE] CHAPS - (a) Clearing House Automated Payment System [NEWLINE] CHAMPUS – (a) (U.S.) Civilian Health and Medical Program of the Uniformed Services (now known as TRICARE) [NEWLINE] CINEOS – (a) Campo Imperatore Near-Earth Object Survey [NEWLINE] CIVETS – (a) Colombia, Indonesia, Vietnam, Egypt, Turkey, South Africa (economics) [NEWLINE] CME – (a) Coronal Mass Ejection (usually Sun) [NEWLINE] CNI – (a) Clinical Nursing Intern</s>
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you don't have to hold back the overachievers<mask> you want to help those who stumble. [ENDQ] [NEWLINE] I don't know, it's a matter of attitude<mask> to whether you look at it<mask> "holding back",<mask> the help (money) has to come from somewhere. [NEWLINE] [NEWLINE] [STARTQ] I'd like to explore this idea that the wealthy "degrade the legislature" some more, it seems promising. [ENDQ] [NEWLINE] It's a truism that money buys political influence.  For example, the Koch brothers are continually implicated in spending on advocacy for legislation of questionable utility that they themselves favor.  It's understood that they are only able to do<mask> much of this<mask> they are billionaires. [NEWLINE] [NEWLINE] Or look at the bankruptcy reform in 2005, passed at the behest of bankers, that made discharging student loan debt in bankruptcy illegal.  Bankers... rich guys.  Students... mostly poor.  And look at the situation with student loan debt today. <mask> a country, we owe about<mask> much on student loans<mask> we do on mortgages for houses. [NEWLINE] [NEWLINE] [URL].phtml [NEWLINE] [NEWLINE] That's only one example, it had a very real negative effect on the "99%", for the benefit of the "1%". [NEWLINE] [NEWLINE] <mask> there is more than one avenue for this sort of influence to be acquired, suffice it to say that they all take money to make use of. [NEWLINE] [NEWLINE] Throughout history the main concern about people being "too rich" is that they would<mask> become too influential<mask> a matter of course, and wreck political progress.  It's not always a matter of "he's got more than me, no fair" - it's "he's got enough money to change the rules in his favor, no fair".  These quotations are about banks and not wealth per se,<mask> pertain somewhat to that line of reasoning: [URL] / [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] But you don't have to hold back the overachievers if you want to help those who stumble. [ENDQ] [NEWLINE] I don't know, it's a matter of attitude as to whether you look at it as "holding back", but the help (money) has to come from somewhere. [NEWLINE] [NEWLINE] [STARTQ] I'd like to explore this idea that the wealthy "degrade the legislature" some more, it seems promising. [ENDQ] [NEWLINE] It's a truism that money buys political influence.  For example, the Koch brothers are continually implicated in spending on advocacy for legislation of questionable utility that they themselves favor.  It's understood that they are only able to do so much of this because they are billionaires. [NEWLINE] [NEWLINE] Or look at the bankruptcy reform in 2005, passed at the behest of bankers, that made discharging student loan debt in bankruptcy illegal.  Bankers... rich guys.  Students... mostly poor.  And look at the situation with student loan debt today.  As a country, we owe about as much on student loans as we do on mortgages for houses. [NEWLINE] [NEWLINE] [URL].phtml [NEWLINE] [NEWLINE] That's only one example, it had a very real negative effect on the "99%", for the benefit of the "1%". [NEWLINE] [NEWLINE] While there is more than one avenue for this sort of influence to be acquired, suffice it to say that they all take money to make use of. [NEWLINE] [NEWLINE] Throughout history the main concern about people being "too rich" is that they would also become too influential as a matter of course, and wreck political progress.  It's not always a matter of "he's got more than me, no fair" - it's "he's got enough money to change the rules in his favor, no fair".  These quotations are about banks and not wealth per se, but pertain somewhat to that line of reasoning: [URL] / [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Human sexuality is incredibly complex and multifaceted. You like feet? Cool. There are people who like vore, fat, hands, farting, sneezing, giants, being tied up, pain, biting, urination... This is a list that is probably hundreds of thousands of categories long. There is absolutely nothing wrong with you being into it. It harms no one and it adds a new flavor to your sexuality. It is not something that you can change about yourself,<mask> for your own personal contentment you need to embrace it. [NEWLINE] [NEWLINE] <mask> from the sound of it, you need reassurance. I have a crazier fetish than feet (<mask> I do like those too). And I have connected with people who are willing to indulge me, they exist. And in situations that have been mixed in attitudes toward fetishism, I have pursued some acceptable compromises. Having a good sex life (especially<mask> you have defined preferences) is about communication. You shouldn't hide something like that about yourself for 3 years. You need to find a level of comfort<mask> you can talk about it earlier. I'm not saying that it is first date material,<mask> a few months into  a relationship you need to be able to talk about sex.<mask> you skirt this kind of communication, it's going to hurt your relationship. It's a bad sign.<mask> you have to repress something that is part of you to make a relationship work, that's a huge problem. [NEWLINE] [NEWLINE] I'm really sorry that things have not been going well for you.<mask> I feel that<mask> adults people should be able to tolerate the idea of fetishes and especially your partner's fetishes and sexual desires (even<mask> you don't immediately like them yourself). Try communicating earlier in relationships and be comfortable with who you are.<mask> you're looking for more resources try /r/sex or fetlife.</s>
Label encoding: <s>Human sexuality is incredibly complex and multifaceted. You like feet? Cool. There are people who like vore, fat, hands, farting, sneezing, giants, being tied up, pain, biting, urination... This is a list that is probably hundreds of thousands of categories long. There is absolutely nothing wrong with you being into it. It harms no one and it adds a new flavor to your sexuality. It is not something that you can change about yourself, so for your own personal contentment you need to embrace it. [NEWLINE] [NEWLINE] But from the sound of it, you need reassurance. I have a crazier fetish than feet ( though I do like those too). And I have connected with people who are willing to indulge me, they exist. And in situations that have been mixed in attitudes toward fetishism, I have pursued some acceptable compromises. Having a good sex life (especially if you have defined preferences) is about communication. You shouldn't hide something like that about yourself for 3 years. You need to find a level of comfort where you can talk about it earlier. I'm not saying that it is first date material, but a few months into  a relationship you need to be able to talk about sex. If you skirt this kind of communication, it's going to hurt your relationship. It's a bad sign. If you have to repress something that is part of you to make a relationship work, that's a huge problem. [NEWLINE] [NEWLINE] I'm really sorry that things have not been going well for you. But I feel that as adults people should be able to tolerate the idea of fetishes and especially your partner's fetishes and sexual desires (even if you don't immediately like them yourself). Try communicating earlier in relationships and be comfortable with who you are. If you're looking for more resources try /r/sex or fetlife.</s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can definitely see this being a consideration in real life. Obviously, the monitor is placed in the center of the player's real vision<mask><mask><mask>. [NEWLINE] [NEWLINE] After just now taking a rather overwhelming dip into the depths of the anatomy of the fovea, foveal/retinal resolution and it's relation to pixel density at given viewing distance, and the viewing distance's effect on viewing angle, I finally found the part I should have read first: [NEWLINE] [STARTQ] The fovea sees only the central two degrees of the visual field, (approximately twice the width of your thumbnail at arm's length). [^[cite]]( [URL] #cite_note-12) [ENDQ] [NEWLINE] <mask> 2&amp;deg; is a larger percentage of a smaller viewing angle, it's still only the area immediately next to the crosshair. [Relative acuity of the eye drops to 50% approximately 3&amp;deg; from the fovea]( [URL].svg/724px-AcuityHumanEye.svg.png) (after potentially faulty math, 8&amp;deg; from center of FOV), which is is 4 times the fovea alone. Completely guessing,<mask><mask> all I changed was my resolution (*particularly* not changing distance-to-display (or, more favorably, not changing resolution and simply moving the monitor further away)), that angle would be more like 45% of the effective display rather than the (another guess) 20% it's at now. [NEWLINE] [NEWLINE] I went into this with the intention of dissuading you,<mask> I *think* this is delta worthy. I really don't know<mask>'s supposed to qualify. :p You didn't change my view significantly,<mask> this is a valid point I hadn't considered. [NEWLINE] [NEWLINE] &amp;#8710; for you..<mask><mask>?</s>
Label encoding: <s>I can definitely see this being a consideration in real life. Obviously, the monitor is placed in the center of the player's real vision for this reason. [NEWLINE] [NEWLINE] After just now taking a rather overwhelming dip into the depths of the anatomy of the fovea, foveal/retinal resolution and it's relation to pixel density at given viewing distance, and the viewing distance's effect on viewing angle, I finally found the part I should have read first: [NEWLINE] [STARTQ] The fovea sees only the central two degrees of the visual field, (approximately twice the width of your thumbnail at arm's length). [^[cite]]( [URL] #cite_note-12) [ENDQ] [NEWLINE] While 2&amp;deg; is a larger percentage of a smaller viewing angle, it's still only the area immediately next to the crosshair. [Relative acuity of the eye drops to 50% approximately 3&amp;deg; from the fovea]( [URL].svg/724px-AcuityHumanEye.svg.png) (after potentially faulty math, 8&amp;deg; from center of FOV), which is is 4 times the fovea alone. Completely guessing, but if all I changed was my resolution (*particularly* not changing distance-to-display (or, more favorably, not changing resolution and simply moving the monitor further away)), that angle would be more like 45% of the effective display rather than the (another guess) 20% it's at now. [NEWLINE] [NEWLINE] I went into this with the intention of dissuading you, but I *think* this is delta worthy. I really don't know what's supposed to qualify. :p You didn't change my view significantly, but this is a valid point I hadn't considered. [NEWLINE] [NEWLINE] &amp;#8710; for you.. I think?</s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well<mask> the website is being both touted<mask> a platform for "free" speech, expression, content and is used by people around the world to spread information of thousands of viewpoints on millions of topics<mask> should it be up to the website to determine<mask> is allowed or<mask> isn't<mask> it doesn't break the law. [NEWLINE] [NEWLINE] Just<mask> there are small communities that perpetuate ideas that I don't subscribe or the admins subscribe isn't a good reason to wipe out that community. Only the ignorant and misinformed judge communities be it digital or physical by small pockets. Is America judged by the actions of the KKK and the Blank Panthers? Not really<mask> we acknowledge their existence and allow them to exist within the bounds of the law. [NEWLINE] [NEWLINE] Are the "bad" subreddits<mask> is used to judge reddit in an article? Perhaps,<mask> then do they acknowledge the good subreddits or the good things that have been accomplished due to this site in that same article? Doubt it<mask>, bad isn't interesting, bad is controversial.<mask> anybody who actually uses the site knows the good outnumbers the bad. [NEWLINE] [NEWLINE] Reddit shouldn't ban subreddits just<mask> they try to promote out of date, uninformed social views. They have just<mask> much a right to be exist an be around<mask> any other more widely accepted view subreddit. [NEWLINE] [NEWLINE] The website provides the platform, its majority users decide the content and<mask> you find a minority to be creating content that you don't agree with that you are seeing, you can change that for you.<mask> seeing<mask> that no law breaking is being done by perpetuating the hating of black people and<mask> Reddit is supposedly a "free" speech area, they should exist.<mask><mask> whether I subscribe to those ideas. [NEWLINE] [NEWLINE] I put free in ""<mask> any private company can censor<mask> they wish<mask> freedom of speech only applies to government laws and not private dealings.</s>
Label encoding: <s>Well when the website is being both touted as a platform for "free" speech, expression, content and is used by people around the world to spread information of thousands of viewpoints on millions of topics why should it be up to the website to determine what is allowed or what isn't if it doesn't break the law. [NEWLINE] [NEWLINE] Just because there are small communities that perpetuate ideas that I don't subscribe or the admins subscribe isn't a good reason to wipe out that community. Only the ignorant and misinformed judge communities be it digital or physical by small pockets. Is America judged by the actions of the KKK and the Blank Panthers? Not really but we acknowledge their existence and allow them to exist within the bounds of the law. [NEWLINE] [NEWLINE] Are the "bad" subreddits what is used to judge reddit in an article? Perhaps, but then do they acknowledge the good subreddits or the good things that have been accomplished due to this site in that same article? Doubt it because, bad isn't interesting, bad is controversial. But anybody who actually uses the site knows the good outnumbers the bad. [NEWLINE] [NEWLINE] Reddit shouldn't ban subreddits just because they try to promote out of date, uninformed social views. They have just as much a right to be exist an be around as any other more widely accepted view subreddit. [NEWLINE] [NEWLINE] The website provides the platform, its majority users decide the content and if you find a minority to be creating content that you don't agree with that you are seeing, you can change that for you. But seeing as that no law breaking is being done by perpetuating the hating of black people and if Reddit is supposedly a "free" speech area, they should exist. Regardless of whether I subscribe to those ideas. [NEWLINE] [NEWLINE] I put free in "" because any private company can censor as they wish since freedom of speech only applies to government laws and not private dealings.</s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have always firmly held the belief that marriage is something that celebrates heterosexual partnership [ENDQ] [NEWLINE] <mask><mask><mask> I'm curious<mask> to<mask> you feel marriage exclusively celebrates heterosexual partnerships,<mask> you can pinpoint<mask> you feel this way it might help in changing your view. [NEWLINE] [NEWLINE] [STARTQ] I would be extremely uncomfortable seeing men/women dancing/kissing people of the same sex at the wedding. [ENDQ] [NEWLINE] This is a personal feeling and something that I am unlikely to be able to change.<mask>, I would suggest that you consider<mask> hurt you would feel<mask> you went to a wedding and people felt it unpleasant to see you dance with your fiancée - it would feel awful. The easiest thing to do, honestly, is simply not to pay attention to them -<mask> I very much doubt you pay large attention to the dancing or kisses of straight couples around you at weddings. [NEWLINE] [NEWLINE] [STARTQ] one of these people is single, and like to try and seduce other women. [ENDQ] [NEWLINE] The simplest solution I can see to this is to speak to this person<mask> not make it in any way related to her orientation. Simply make it clear that you are hoping for a calm and tasteful wedding and that *nobody*<mask><mask> gender or orientation will be welcome<mask> they cannot respect the other guests. [NEWLINE] [NEWLINE] Something that might be worth thinking about is:<mask> are you getting married? Most people get married<mask> they want to have their relationship recognised by their loved ones. Now imagine<mask> you were invited to a wedding<mask> your fiancée was not - imagine<mask> disrespectful that would be. That is<mask> you are doing to these people - you are telling them that<mask> of something out of their control (their gender of their partner) their relationship is not'real' to you. You are saying that they do not count<mask> a real couple, that you do not respect them or their partner.</s><pad>
Label encoding: <s> [STARTQ] I have always firmly held the belief that marriage is something that celebrates heterosexual partnership [ENDQ] [NEWLINE] First of all I'm curious as to why you feel marriage exclusively celebrates heterosexual partnerships, if you can pinpoint why you feel this way it might help in changing your view. [NEWLINE] [NEWLINE] [STARTQ] I would be extremely uncomfortable seeing men/women dancing/kissing people of the same sex at the wedding. [ENDQ] [NEWLINE] This is a personal feeling and something that I am unlikely to be able to change. However, I would suggest that you consider how hurt you would feel if you went to a wedding and people felt it unpleasant to see you dance with your fiancée - it would feel awful. The easiest thing to do, honestly, is simply not to pay attention to them - since I very much doubt you pay large attention to the dancing or kisses of straight couples around you at weddings. [NEWLINE] [NEWLINE] [STARTQ] one of these people is single, and like to try and seduce other women. [ENDQ] [NEWLINE] The simplest solution I can see to this is to speak to this person but not make it in any way related to her orientation. Simply make it clear that you are hoping for a calm and tasteful wedding and that *nobody* regardless of gender or orientation will be welcome if they cannot respect the other guests. [NEWLINE] [NEWLINE] Something that might be worth thinking about is: why are you getting married? Most people get married because they want to have their relationship recognised by their loved ones. Now imagine if you were invited to a wedding but your fiancée was not - imagine how disrespectful that would be. That is what you are doing to these people - you are telling them that because of something out of their control (their gender of their partner) their relationship is not'real' to you. You are saying that they do not count as a real couple, that you do not respect them or their partner.</s><pad>
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We<mask> a society have agreed that I have no right whatsoever to the use of your body or any part thereof without your consent even<mask> my life depends on it. I cannot demand you donate a kidney to save my life, or part of your liver, or even blood plasma - I can't even demand you walk into danger to help me out of it. That's<mask> it is your body, your ultimate possession, and you have sovereign control over it. [ENDQ] [NEWLINE] <mask><mask> with this.  Specifically that "society has agreed" to this position. [NEWLINE] [NEWLINE] <mask> it is generally true that the *use* of a body without consent is not usually endorsed, I would posit that this has more to do with<mask> limited a scope this is.  It is<mask> not always true (see point 3). [NEWLINE] [NEWLINE] It is, after all, certainly *not* true that society gives *you* free reign to the use of your body.  Bans on suicide and euthanasia are the most obvious example. [NEWLINE] [NEWLINE] Furthermore, there are many many examples of limiting one's use of their own body directly to protect others.  This is moral foundation for many substance abuse laws, most prominently DUI laws.  Incarceration and committal to mental institutions can<mask> be similar. [NEWLINE] [NEWLINE] And of course<mask><mask>, pretty much every society bans late term abortions except under extreme conditions, which directly flies in the face of your argument.  Society, in effect, *does* directly give these fetuses the right to the use of their mother's body, generally without regard for the mother's continuing consent. [NEWLINE] [NEWLINE] <mask><mask> it is true that mandatory blood drives are not a thing<mask>, I feel one could make a fairly strong case that limiting one individual's rights to their own body to protect another individual's life is fairly consistent with other societal views.</s>
Label encoding: <s> [STARTQ] We as a society have agreed that I have no right whatsoever to the use of your body or any part thereof without your consent even if my life depends on it. I cannot demand you donate a kidney to save my life, or part of your liver, or even blood plasma - I can't even demand you walk into danger to help me out of it. That's because it is your body, your ultimate possession, and you have sovereign control over it. [ENDQ] [NEWLINE] I disagree with this.  Specifically that "society has agreed" to this position. [NEWLINE] [NEWLINE] While it is generally true that the *use* of a body without consent is not usually endorsed, I would posit that this has more to do with how limited a scope this is.  It is also not always true (see point 3). [NEWLINE] [NEWLINE] It is, after all, certainly *not* true that society gives *you* free reign to the use of your body.  Bans on suicide and euthanasia are the most obvious example. [NEWLINE] [NEWLINE] Furthermore, there are many many examples of limiting one's use of their own body directly to protect others.  This is moral foundation for many substance abuse laws, most prominently DUI laws.  Incarceration and committal to mental institutions can also be similar. [NEWLINE] [NEWLINE] And of course lastly, pretty much every society bans late term abortions except under extreme conditions, which directly flies in the face of your argument.  Society, in effect, *does* directly give these fetuses the right to the use of their mother's body, generally without regard for the mother's continuing consent. [NEWLINE] [NEWLINE] So while it is true that mandatory blood drives are not a thing yet, I feel one could make a fairly strong case that limiting one individual's rights to their own body to protect another individual's life is fairly consistent with other societal views.</s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In the end, you're asking people to pay more for something that they never see anyways, and they have to trust the supposedly better sources are actually better sources. Which<mask> you know anything about this, is not a great chance. [ENDQ] [NEWLINE] This is critical, and you're exactly right. <mask> with most causes, at some point it requires actual sacrifice on the part of those who claim to care, and it's at that point that you suddenly see a lot less people caring. [NEWLINE] [NEWLINE] I would still<mask><mask> paying more for meat is something that people are going to see<mask> a smaller sacrifice than giving up meat all together. [NEWLINE] [NEWLINE] <mask> it stands, the market for "good meat" is not a very competitive one.  There are relatively few people willing to pay a premium for<mask> they consider humane, and<mask> suppliers and stores can easily charge triple the amount they normally do, and get away with it. [NEWLINE] [NEWLINE] <mask>,<mask> it happens that a sizable percentage of the population starts to insist on humanely-treated animals for their meat, now there emerges some incentive for places to actually get competitive with their pricing. [NEWLINE] [NEWLINE] <mask><mask> that it's mostly horseshit all these claims of "cage-free" and<mask> on. [NEWLINE] [NEWLINE] I would actually counter this CMV by saying that the *very* best thing you can do is buy local.  I'm on a first-name basis with the guy I buy my eggs from.  I've met the chickens.  I know exactly<mask> they're treated, and I get a good price for them.  I've petted the cows that have become my burgers.  That's not<mask> possible for people living in New York city,<mask> the smaller the source you're getting your stuff from, the more control you have over it<mask> a consumer.  </s>
Label encoding: <s> [STARTQ] In the end, you're asking people to pay more for something that they never see anyways, and they have to trust the supposedly better sources are actually better sources. Which if you know anything about this, is not a great chance. [ENDQ] [NEWLINE] This is critical, and you're exactly right.  As with most causes, at some point it requires actual sacrifice on the part of those who claim to care, and it's at that point that you suddenly see a lot less people caring. [NEWLINE] [NEWLINE] I would still argue that paying more for meat is something that people are going to see as a smaller sacrifice than giving up meat all together. [NEWLINE] [NEWLINE] As it stands, the market for "good meat" is not a very competitive one.  There are relatively few people willing to pay a premium for what they consider humane, and so suppliers and stores can easily charge triple the amount they normally do, and get away with it. [NEWLINE] [NEWLINE] However, if it happens that a sizable percentage of the population starts to insist on humanely-treated animals for their meat, now there emerges some incentive for places to actually get competitive with their pricing. [NEWLINE] [NEWLINE] I agree that it's mostly horseshit all these claims of "cage-free" and so on. [NEWLINE] [NEWLINE] I would actually counter this CMV by saying that the *very* best thing you can do is buy local.  I'm on a first-name basis with the guy I buy my eggs from.  I've met the chickens.  I know exactly how they're treated, and I get a good price for them.  I've petted the cows that have become my burgers.  That's not as possible for people living in New York city, but the smaller the source you're getting your stuff from, the more control you have over it as a consumer.  </s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Secularism isn't really a bad view,<mask> you do have some strange reasons for coming to secularism<mask> a conclusion. [NEWLINE] [NEWLINE] I'd agree that<mask> there were a hypothetical world<mask> it was literally impossible to commit crimes, the laws defining those crimes would be pretty pointless. To take an example from our world, suppose your city had a law saying that exceeding the speed of light carries a fine of $2000 and a minimum 2 month jail sentence. This would be a silly and pointless law. (Legal law, obviously c is a natural law of some sort.) [NEWLINE] [NEWLINE] <mask>, reasoning given by religions,<mask> I understand it, is that their theological laws are important not<mask> they are not already covered by<mask> is possible in nature,<mask> for other reasons. Like the action itself is bad for other people, or would displease a deity.<mask> most religions define themselves by more than their prohibitions.<mask> they had no reason for those rules (like a rule for sobriety, and then all the alcohol is destroyed), then the religion would still be fine. It wouldn't cause their God to suddenly cease to exist or something. Even<mask> the world were changed<mask> that followers were incapable of displeasing their deity, it wouldn't mean that worship would have to stop. [NEWLINE] [NEWLINE] Here are some much simpler reasons to believe in secularism: [NEWLINE] [NEWLINE] * We don't know<mask> to judge objectively<mask> one religion is more right than another. [NEWLINE] * We don't know<mask> a God or gods exist.<mask> the reason for doing something is "God commands it" then we can't be sure whether a command was issued in the first place. [NEWLINE] * Even<mask> we did know a command was issued by God, we don't know whether following that command will be better or worse than the current situation without cross examining it with other moral reasoning.</s>
Label encoding: <s>Secularism isn't really a bad view, but you do have some strange reasons for coming to secularism as a conclusion. [NEWLINE] [NEWLINE] I'd agree that if there were a hypothetical world where it was literally impossible to commit crimes, the laws defining those crimes would be pretty pointless. To take an example from our world, suppose your city had a law saying that exceeding the speed of light carries a fine of $2000 and a minimum 2 month jail sentence. This would be a silly and pointless law. (Legal law, obviously c is a natural law of some sort.) [NEWLINE] [NEWLINE] However, reasoning given by religions, as I understand it, is that their theological laws are important not because they are not already covered by what is possible in nature, but for other reasons. Like the action itself is bad for other people, or would displease a deity. Additionally most religions define themselves by more than their prohibitions. If they had no reason for those rules (like a rule for sobriety, and then all the alcohol is destroyed), then the religion would still be fine. It wouldn't cause their God to suddenly cease to exist or something. Even if the world were changed so that followers were incapable of displeasing their deity, it wouldn't mean that worship would have to stop. [NEWLINE] [NEWLINE] Here are some much simpler reasons to believe in secularism: [NEWLINE] [NEWLINE] * We don't know how to judge objectively if one religion is more right than another. [NEWLINE] * We don't know if a God or gods exist. If the reason for doing something is "God commands it" then we can't be sure whether a command was issued in the first place. [NEWLINE] * Even if we did know a command was issued by God, we don't know whether following that command will be better or worse than the current situation without cross examining it with other moral reasoning.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Historically, it's been more common for men to be obligated to go to war under threat of imprisonment, torture, and/or execution. I say that counts<mask> oppression [ENDQ] [NEWLINE] Oppression by whom exactly? You have to have power over another group to oppress them. And I don't think there's anything<mask> a laughable argument that "historically" women have had power over men, *especially* the power to imprison, torture, or execute men<mask> they didn't go to war and/or die for women. "Historically" those prisons, torture facilities, execution facilities, and the laws that would have sent men there for refusing to go to war, we're **all** made by and run by men. [NEWLINE] [NEWLINE] <mask><mask> anything you're saying that men have "historically" oppressed themselves. Which<mask><mask> has an element of truth to it. Rather than view it<mask> a sadistic under-the-radar societal, or evolutionary (<mask> there is *no* scientific evidence for the effect you talk about.<mask> it was scientifically worded, it is just a shot in the dark scientifically) effect of female-on-male oppression, I would say the "disposability" of men that you talk about comes more from the "historical" tendency for men to want to (pardon my french) compare dick sizes, than it comes from anything on the female side. Men wanted to look more heroic to their fellow man, and<mask>'s more heroic than not caring about whether or not you die<mask><mask><mask> you keep your wife and family safe? And over time this barbaric pissing contest of whose manlier got ingrained in social systems to the point<mask> men<mask> legally obligated to go to war over women (<mask> men were making all the laws and social systems). It's<mask> simple<mask> that.</s>
Label encoding: <s> [STARTQ] Historically, it's been more common for men to be obligated to go to war under threat of imprisonment, torture, and/or execution. I say that counts as oppression [ENDQ] [NEWLINE] Oppression by whom exactly? You have to have power over another group to oppress them. And I don't think there's anything but a laughable argument that "historically" women have had power over men, *especially* the power to imprison, torture, or execute men if they didn't go to war and/or die for women. "Historically" those prisons, torture facilities, execution facilities, and the laws that would have sent men there for refusing to go to war, we're **all** made by and run by men. [NEWLINE] [NEWLINE] So if anything you're saying that men have "historically" oppressed themselves. Which I think has an element of truth to it. Rather than view it as a sadistic under-the-radar societal, or evolutionary ( though there is *no* scientific evidence for the effect you talk about. Though it was scientifically worded, it is just a shot in the dark scientifically) effect of female-on-male oppression, I would say the "disposability" of men that you talk about comes more from the "historical" tendency for men to want to (pardon my french) compare dick sizes, than it comes from anything on the female side. Men wanted to look more heroic to their fellow man, and what's more heroic than not caring about whether or not you die as long as you keep your wife and family safe? And over time this barbaric pissing contest of whose manlier got ingrained in social systems to the point where men where legally obligated to go to war over women ( because men were making all the laws and social systems). It's as simple as that.</s>
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0784, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Until there is, until you actually have that method and system figured out, it is completely useless to bring it up at all. [ENDQ] [NEWLINE] <mask> unless you've come up with a solution to a problem, bringing up that there is a problem is useless? I'm sorry,<mask> bullshit. We don't solve problems in real life by ignoring them or sweeping them under the rug, we solve them by gathering people together to think about the problem, define the problem, and find creative solutions. OWS for all its failings brought to the forefront of the American consciousness the gross imbalance of wealth in the system. And it's not that we CAN'T meet everyone's needs. There are currently enough foreclosed homes in America for a vast majority of homeless people to have a home: [URL] this points out that the raw data shows that there are 22 vacant homes for every homeless person. And half of the food we produce, we discard: [URL] and<mask> people are still going hungry. [NEWLINE] [NEWLINE] Now, I don't have a solution to the problem immediately at hand, mainly<mask> the people who own these vacant houses would be APPALLED at the idea of using them to house the homeless on their own dime, or worse, paying a dollar more in taxes to subsidize housing for the poor.<mask> I do think that capitalism falls apart<mask> an efficient means of distributing wealth<mask> goods and services become post-scarcity, and that is evidently<mask> we are seeing today.<mask> we have the ability to throw away food<mask> it doesn't look<mask> good<mask> other foods, or to have vacant houses with nobody to move into them, then we've passed the scarcity point of these things, and we apparently need a better system. And a better system won't come about by us just sitting and pretending that the problem doesn't exist.</s>
Label encoding: <s> [STARTQ] Until there is, until you actually have that method and system figured out, it is completely useless to bring it up at all. [ENDQ] [NEWLINE] So unless you've come up with a solution to a problem, bringing up that there is a problem is useless? I'm sorry, but bullshit. We don't solve problems in real life by ignoring them or sweeping them under the rug, we solve them by gathering people together to think about the problem, define the problem, and find creative solutions. OWS for all its failings brought to the forefront of the American consciousness the gross imbalance of wealth in the system. And it's not that we CAN'T meet everyone's needs. There are currently enough foreclosed homes in America for a vast majority of homeless people to have a home: [URL] this points out that the raw data shows that there are 22 vacant homes for every homeless person. And half of the food we produce, we discard: [URL] and yet people are still going hungry. [NEWLINE] [NEWLINE] Now, I don't have a solution to the problem immediately at hand, mainly because the people who own these vacant houses would be APPALLED at the idea of using them to house the homeless on their own dime, or worse, paying a dollar more in taxes to subsidize housing for the poor. But I do think that capitalism falls apart as an efficient means of distributing wealth when goods and services become post-scarcity, and that is evidently what we are seeing today. If we have the ability to throw away food because it doesn't look as good as other foods, or to have vacant houses with nobody to move into them, then we've passed the scarcity point of these things, and we apparently need a better system. And a better system won't come about by us just sitting and pretending that the problem doesn't exist.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you robbed a bank, a superior force would arrest, detain, and/or kill you. [ENDQ] [STARTQ] Totalitarian regimes have constantly been overthrown by violent revolution by the masses. [ENDQ] [STARTQ] Hitler was defeated by a superior force in the end. [ENDQ] [NEWLINE] The existence of a superior force and the possibility of it being overpowered doesn't have any bearing on the actions of the original force nor justify those actions. Setting aside the fact that some bank robbers get away with it, that there are still totalitarian regimes, and that Hitler very nearly did win the war... the external force's justifications for taking down the original force does in no way justify the inciting actions, the original "might". [NEWLINE] [NEWLINE] [STARTQ] I often get a lot of trouble for saying this,<mask> it was a good thing the United States conquered the Native Americans. The US is currently a powerhouse and devotes millions in foreign aid to alleviate suffering. I do not believe a tribal culture like the conquered natives could be capable of that. [ENDQ] [NEWLINE] And<mask> have<mask> been the costs of this nation? There were the millions of native americans who were slaughtered... we fought and died over slavery which lasted decades longer than in Europe... we've been in dozens of conflicts, including in Vietnam<mask> many of our boys went crazy and raped and murdered innocent civilians, including the past 13 years in the Middle East<mask> between drones and everything else we've killed thousands of people who may or may not be innocent or just trying to defend their homeland... we've had the recent CIA torture reports, the NSA scandal...<mask> much<mask> we try to appear to be the centre of western righteousness we've done plenty of wrong in our time. The power to do something does not mean that you're right to do it. The end doesn't necessarily justify the means. </s>
Label encoding: <s> [STARTQ] If you robbed a bank, a superior force would arrest, detain, and/or kill you. [ENDQ] [STARTQ] Totalitarian regimes have constantly been overthrown by violent revolution by the masses. [ENDQ] [STARTQ] Hitler was defeated by a superior force in the end. [ENDQ] [NEWLINE] The existence of a superior force and the possibility of it being overpowered doesn't have any bearing on the actions of the original force nor justify those actions. Setting aside the fact that some bank robbers get away with it, that there are still totalitarian regimes, and that Hitler very nearly did win the war... the external force's justifications for taking down the original force does in no way justify the inciting actions, the original "might". [NEWLINE] [NEWLINE] [STARTQ] I often get a lot of trouble for saying this, but it was a good thing the United States conquered the Native Americans. The US is currently a powerhouse and devotes millions in foreign aid to alleviate suffering. I do not believe a tribal culture like the conquered natives could be capable of that. [ENDQ] [NEWLINE] And what have also been the costs of this nation? There were the millions of native americans who were slaughtered... we fought and died over slavery which lasted decades longer than in Europe... we've been in dozens of conflicts, including in Vietnam where many of our boys went crazy and raped and murdered innocent civilians, including the past 13 years in the Middle East where between drones and everything else we've killed thousands of people who may or may not be innocent or just trying to defend their homeland... we've had the recent CIA torture reports, the NSA scandal... As much as we try to appear to be the centre of western righteousness we've done plenty of wrong in our time. The power to do something does not mean that you're right to do it. The end doesn't necessarily justify the means. </s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your argument is that there is [too much satire] on TV which means [people copy] leading to [more hate/ignorance] preventing discussion. [NEWLINE] [NEWLINE] 1. *‘Shows use excessive satire’* [NEWLINE] [NEWLINE] <mask><mask> there are more shows *with* satire,<mask> in **proportion to other shows** (even just other comedic shows) I really doubt the proportion is extreme<mask> you suggest. You suggest satire to be the prevailing (or at least one of the most significant) humour types,<mask> provide no evidence to support this. [NEWLINE] You later include social media to strengthen your argument<mask> again the above issue arises. [NEWLINE] [NEWLINE] [NEWLINE] 2. *‘Satire exposure leads to it being ingrained’* [NEWLINE] [NEWLINE] I’d probably agree with you; people copy<mask> they see on TV,<mask> Issue 1 arises again.<mask> much satire is seen/ingrained in proportion to other types of humour? [NEWLINE] [NEWLINE] [NEWLINE] 3. *‘It has the 2 negative side effects: [cool to hate] + [discussion prevention]’* [NEWLINE] [NEWLINE] You are implying these two phenomena have been exacerbated by excessive satire. Your argument loses structure here. You say we are surrounded by hate and stereotypes and<mask> accept it. It may be more noticeable,<mask> hard to<mask><mask> society is becoming less tolerant. Look at gender equality, race and LGBT movements. You may see more intolerance issues,<mask> we are more connected (i.e. more discussion)<mask> you shouldn't assume this means there actually are more. [NEWLINE] [NEWLINE] Here your argument relies on the **Golden Age Fallacy**, i.e. things used to be better back in the day. [NEWLINE] [NEWLINE] [NEWLINE] EDIT: Formatting (numbers should read 1,2,3) [NEWLINE] </s>
Label encoding: <s>Your argument is that there is [too much satire] on TV which means [people copy] leading to [more hate/ignorance] preventing discussion. [NEWLINE] [NEWLINE] 1. *‘Shows use excessive satire’* [NEWLINE] [NEWLINE] I agree there are more shows *with* satire, BUT in **proportion to other shows** (even just other comedic shows) I really doubt the proportion is extreme as you suggest. You suggest satire to be the prevailing (or at least one of the most significant) humour types, but provide no evidence to support this. [NEWLINE] You later include social media to strengthen your argument but again the above issue arises. [NEWLINE] [NEWLINE] [NEWLINE] 2. *‘Satire exposure leads to it being ingrained’* [NEWLINE] [NEWLINE] I’d probably agree with you; people copy what they see on TV, but Issue 1 arises again. How much satire is seen/ingrained in proportion to other types of humour? [NEWLINE] [NEWLINE] [NEWLINE] 3. *‘It has the 2 negative side effects: [cool to hate] + [discussion prevention]’* [NEWLINE] [NEWLINE] You are implying these two phenomena have been exacerbated by excessive satire. Your argument loses structure here. You say we are surrounded by hate and stereotypes and so accept it. It may be more noticeable, but hard to argue that society is becoming less tolerant. Look at gender equality, race and LGBT movements. You may see more intolerance issues, because we are more connected (i.e. more discussion) but you shouldn't assume this means there actually are more. [NEWLINE] [NEWLINE] Here your argument relies on the **Golden Age Fallacy**, i.e. things used to be better back in the day. [NEWLINE] [NEWLINE] [NEWLINE] EDIT: Formatting (numbers should read 1,2,3) [NEWLINE] </s>
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok,<mask> compare it to Poland. [NEWLINE] [NEWLINE] It's "free", yes. [NEWLINE] [NEWLINE] Most of older professors (and some younger) are stuck in communist state of mind. They treat you like parasite and they are doing<mask> they want. They don't like you? They will find a way to fail you. They can do anything they want, they are untouchable. [NEWLINE] [NEWLINE] Universities are under invested. We were working in laboratories older than me, mostly with equipment older than me. We were doing experiments in groups. Experiments<mask> you should learn<mask> to handle equipment. [NEWLINE] [NEWLINE] There is overload of everything. You're on University 5 days in week, sometimes from 7am to 8pm only on mandatory lectures and labs. Oh, and there exercises too.<mask> overall you have: on the end of semester 3 huge exams, 4-5 smaller "exams" on exercises, across semester you have many more. And I don't count here knowledge checks on every laboratory, it's normal, you must know<mask> you will be doing. [NEWLINE] [NEWLINE] <mask> yeah, I will take US college system over polish. [NEWLINE] [NEWLINE] Edit:<mask> you fail on exam (or in any other way) there are two options you have, depends on<mask> you failed: [NEWLINE] [NEWLINE] You can go to next semester, you must repeat it **and** you must pay for it (depends on<mask> it's range of 100 to ~200 euro) [NEWLINE] [NEWLINE] You can't go to the next semester, you must repeat it (or you must repeat **everything**) **and** you must pay for it. [NEWLINE] [NEWLINE] In many cases there is everytime coincidence and around ~10% of students don't pass one exam and they must pay for repeating it. Pure money for University. </s>
Label encoding: <s>Ok, so compare it to Poland. [NEWLINE] [NEWLINE] It's "free", yes. [NEWLINE] [NEWLINE] Most of older professors (and some younger) are stuck in communist state of mind. They treat you like parasite and they are doing what they want. They don't like you? They will find a way to fail you. They can do anything they want, they are untouchable. [NEWLINE] [NEWLINE] Universities are under invested. We were working in laboratories older than me, mostly with equipment older than me. We were doing experiments in groups. Experiments where you should learn how to handle equipment. [NEWLINE] [NEWLINE] There is overload of everything. You're on University 5 days in week, sometimes from 7am to 8pm only on mandatory lectures and labs. Oh, and there exercises too. So overall you have: on the end of semester 3 huge exams, 4-5 smaller "exams" on exercises, across semester you have many more. And I don't count here knowledge checks on every laboratory, it's normal, you must know what you will be doing. [NEWLINE] [NEWLINE] So yeah, I will take US college system over polish. [NEWLINE] [NEWLINE] Edit: when you fail on exam (or in any other way) there are two options you have, depends on what you failed: [NEWLINE] [NEWLINE] You can go to next semester, you must repeat it **and** you must pay for it (depends on what it's range of 100 to ~200 euro) [NEWLINE] [NEWLINE] You can't go to the next semester, you must repeat it (or you must repeat **everything**) **and** you must pay for it. [NEWLINE] [NEWLINE] In many cases there is everytime coincidence and around ~10% of students don't pass one exam and they must pay for repeating it. Pure money for University. </s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've always said that the progressive movement has very bad 'PR',<mask> in they make more enemies than friends.  Even the term 'White Privilege' is a hostile term<mask> it assumes that: [NEWLINE] [NEWLINE] 1. You will be treated better for just having European features [NEWLINE] [NEWLINE] 2. The reason others fail is<mask> of white people. [NEWLINE] [NEWLINE] <mask> basically you catch more bees with honey than with vinegar.  And Leftists have been trying to clean up the message to attract people,<mask> not a lot of people are buying it (ex: Hermaine from Harry Potter) [NEWLINE] [NEWLINE] [NEWLINE] <mask><mask> you are forgetting is that its not just<mask> the message is delivered, its<mask> its content. [NEWLINE] [NEWLINE] [NEWLINE] I'm Mexican American, not white at all, and I don't believe in 'white privilege' You have these ideas have some serious holes, and that people will continue to disagree with you<mask> of it (not<mask> they are offended) [NEWLINE] [NEWLINE] [NEWLINE] <mask><mask><mask> you need to explain<mask> the very real problem of racial profiling is just the white persons fault.  I was under the impression that this is everybody's fault<mask> I know immigrant communities are very suspicious of black youth,<mask> singling out white people is unfair. [NEWLINE] [NEWLINE] [NEWLINE] <mask> white privilege ignores the fact that white people are not even the most successful ethnic groups in America per capita:  East Asians and Jewish people are very well off in this supposed "white supremicist" nation [NEWLINE] [NEWLINE] [NEWLINE] <mask>, immigrants like my parents risk their lives to make it in here. <mask> this white privilege narrative doesn't make sense<mask> you put into account<mask> actual non-white immigrants feel about this nation.  They don't feel anythings holding you back here, the white privilege stuff they probably don't even know about. [NEWLINE] </s>
Label encoding: <s>I've always said that the progressive movement has very bad 'PR', as in they make more enemies than friends.  Even the term 'White Privilege' is a hostile term because it assumes that: [NEWLINE] [NEWLINE] 1. You will be treated better for just having European features [NEWLINE] [NEWLINE] 2. The reason others fail is because of white people. [NEWLINE] [NEWLINE] So basically you catch more bees with honey than with vinegar.  And Leftists have been trying to clean up the message to attract people, but not a lot of people are buying it (ex: Hermaine from Harry Potter) [NEWLINE] [NEWLINE] [NEWLINE] But what you are forgetting is that its not just how the message is delivered, its also its content. [NEWLINE] [NEWLINE] [NEWLINE] I'm Mexican American, not white at all, and I don't believe in 'white privilege' You have these ideas have some serious holes, and that people will continue to disagree with you because of it (not because they are offended) [NEWLINE] [NEWLINE] [NEWLINE] First of all you need to explain how the very real problem of racial profiling is just the white persons fault.  I was under the impression that this is everybody's fault because I know immigrant communities are very suspicious of black youth, so singling out white people is unfair. [NEWLINE] [NEWLINE] [NEWLINE] Also white privilege ignores the fact that white people are not even the most successful ethnic groups in America per capita:  East Asians and Jewish people are very well off in this supposed "white supremicist" nation [NEWLINE] [NEWLINE] [NEWLINE] Also, immigrants like my parents risk their lives to make it in here.  So this white privilege narrative doesn't make sense when you put into account how actual non-white immigrants feel about this nation.  They don't feel anythings holding you back here, the white privilege stuff they probably don't even know about. [NEWLINE] </s>
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In your example, the only way we can avoid the same contradiction is<mask> God watched the video after it occurred.<mask> watching the video = knowledge of events that will occur, and recording the video = creating the universe, this would mean that God did not know<mask> events would  occur at the time he created the universe. [NEWLINE] [NEWLINE] <mask> we assume that God did have knowledge of events before he created the universe and before they occurred, let's say that God can watch a video before the event happens. Now let's say that God watches the video of you throwing the ball at target B. Forget about free will and forget about omniscience for the moment.<mask> are the two possibilities of subsequent events<mask> God watches the video of you throwing a ball at B? There is either the possibility that you throw the ball at B, or that you throw the ball at A. That's it. There are no other choices.<mask> we determine that one option can't happen, then the other **must** occur. Let's say that you throw the ball at A. Then God watched a video of something that did not occur, i.e. he had knowledge of events that did not occur.<mask> we do not allow this<mask> a possibility, then you must throw the ball at B. You did not have the option to do otherwise. [NEWLINE] [NEWLINE] Now you could say that for God, time is not directional and<mask> causality is not directional. He would not have to record the video before or after the events occurred. I'll just use my imagination here to guess<mask> that would mean. It would mean that for God, everything would be instantaneous. He would simultaneously have knowledge of events<mask> they happened<mask> he was creating them. I'm not sure<mask> the full implications of that would be.</s>
Label encoding: <s>In your example, the only way we can avoid the same contradiction is if God watched the video after it occurred. Since watching the video = knowledge of events that will occur, and recording the video = creating the universe, this would mean that God did not know what events would  occur at the time he created the universe. [NEWLINE] [NEWLINE] Since we assume that God did have knowledge of events before he created the universe and before they occurred, let's say that God can watch a video before the event happens. Now let's say that God watches the video of you throwing the ball at target B. Forget about free will and forget about omniscience for the moment. What are the two possibilities of subsequent events if God watches the video of you throwing a ball at B? There is either the possibility that you throw the ball at B, or that you throw the ball at A. That's it. There are no other choices. If we determine that one option can't happen, then the other **must** occur. Let's say that you throw the ball at A. Then God watched a video of something that did not occur, i.e. he had knowledge of events that did not occur. Since we do not allow this as a possibility, then you must throw the ball at B. You did not have the option to do otherwise. [NEWLINE] [NEWLINE] Now you could say that for God, time is not directional and therefore causality is not directional. He would not have to record the video before or after the events occurred. I'll just use my imagination here to guess what that would mean. It would mean that for God, everything would be instantaneous. He would simultaneously have knowledge of events as they happened as he was creating them. I'm not sure what the full implications of that would be.</s>
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've been to a church<mask> they read each verse of the bible and apply its morals to modern times. They get a small part of the city to help out the rest of the city. They preach individuality. That place grew quick. [NEWLINE] [NEWLINE] I've<mask> been to a church<mask> the pastor got out a newspaper and spent half the sermon preaching about Obama. He had neat stuff to say sometimes and was very poetic. He had a lot of things to say about<mask> to act, learn, talk, vote, walk, listen, read, and think. It's lost most of its members now,<mask> its an older church with older people. Not much exciting about it. Imagine a church<mask> they tought you to listen, then told you<mask> to think, then paid you in entertainment/food/community. That'd grow quick. [NEWLINE] [NEWLINE] Being a Christian, I hate christianity<mask> much<mask> anyone else who hates simple thinking. I highly value being able to hear<mask> others say<mask><mask> maintaining your own perspective and experience. [NEWLINE] [NEWLINE] Except in rare cases, that's not<mask> people preach.<mask> they taught you<mask> to think for yourself using the moral guides the bible offers, Christianity wouldn't have the shit name it currently holds.<mask> people preach<mask> to think, not<mask>,<mask> they have interests to support. It isn't an inherent Christian thing, and it isn't<mask> makes Christianity useful (or even acceptable). [NEWLINE] [NEWLINE] Basically, Hitler could do this<mask> easily<mask> my preacher could, and it wouldn't matter which ocean he used to float his boat. The problem is improper methods of teaching philosophy/religion/science. [NEWLINE] [NEWLINE] [NEWLINE] Side note: You know<mask> would be cool?<mask> the ideas in the scientific method were more readily applied to philosophy. </s>
Label encoding: <s>I've been to a church where they read each verse of the bible and apply its morals to modern times. They get a small part of the city to help out the rest of the city. They preach individuality. That place grew quick. [NEWLINE] [NEWLINE] I've also been to a church where the pastor got out a newspaper and spent half the sermon preaching about Obama. He had neat stuff to say sometimes and was very poetic. He had a lot of things to say about how to act, learn, talk, vote, walk, listen, read, and think. It's lost most of its members now, but its an older church with older people. Not much exciting about it. Imagine a church where they tought you to listen, then told you what to think, then paid you in entertainment/food/community. That'd grow quick. [NEWLINE] [NEWLINE] Being a Christian, I hate christianity as much as anyone else who hates simple thinking. I highly value being able to hear what others say while also maintaining your own perspective and experience. [NEWLINE] [NEWLINE] Except in rare cases, that's not how people preach. If they taught you how to think for yourself using the moral guides the bible offers, Christianity wouldn't have the shit name it currently holds. But people preach what to think, not why, because they have interests to support. It isn't an inherent Christian thing, and it isn't what makes Christianity useful (or even acceptable). [NEWLINE] [NEWLINE] Basically, Hitler could do this as easily as my preacher could, and it wouldn't matter which ocean he used to float his boat. The problem is improper methods of teaching philosophy/religion/science. [NEWLINE] [NEWLINE] [NEWLINE] Side note: You know what would be cool? If the ideas in the scientific method were more readily applied to philosophy. </s>
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's interesting. I can only speak for my experience in the US, it could differ in other countries.<mask> is considered liberal in the states may not be in some parts of Europe. [NEWLINE] [NEWLINE] At the public research university I attended, most students and faculty were generally pro-GMO and pro-nuclear energy, or at least willing to consider their merits. [NEWLINE] [NEWLINE] To be fair, and<mask> I will elaborate on below, my "remotely educated" comment wasn't necessarily about the legitimate and well-thought out opposition to GMO's and nuclear energy themselves, which definitely does exist,<mask> rather about the demographic which opposes them for invalid reasons. [NEWLINE] [NEWLINE] I really don't know much about nuclear energy<mask> I can't really couch for or against it personally. [NEWLINE] [NEWLINE] There are definitely some legitimate concerns about GMO's<mask> you mentioned,<mask> there are<mask> lots of positives. Many of the potential issues are avoidable,<mask> it really is just a matter of proper regulation and continued research. [NEWLINE] [NEWLINE] The reason the opposition to GMO's was brought up, and the reason I compared it to the Tea Party on the right (a compaison which I perhaps should have more carefully explained) is that a large percentage of those opposed do not reject it for those legitimate concerns<mask> out of a natural product fetish and/or an excessive fear of corporate control over food. The main difference, of course, being that these movements are not nearly<mask> central to the modern liberal platform<mask> the tea party is for the conservative counterparts. [NEWLINE] [NEWLINE] **TL;DR** [NEWLINE] [NEWLINE] **There is legitimate opposition to GMO's,<mask> it is not<mask> OP was criticizing nor<mask> I was attempting to. The demographic in question does not represent liberalism nor do the issues make it hypocritical or equivalent to conservatism.**</s>
Label encoding: <s>That's interesting. I can only speak for my experience in the US, it could differ in other countries. What is considered liberal in the states may not be in some parts of Europe. [NEWLINE] [NEWLINE] At the public research university I attended, most students and faculty were generally pro-GMO and pro-nuclear energy, or at least willing to consider their merits. [NEWLINE] [NEWLINE] To be fair, and as I will elaborate on below, my "remotely educated" comment wasn't necessarily about the legitimate and well-thought out opposition to GMO's and nuclear energy themselves, which definitely does exist, but rather about the demographic which opposes them for invalid reasons. [NEWLINE] [NEWLINE] I really don't know much about nuclear energy so I can't really couch for or against it personally. [NEWLINE] [NEWLINE] There are definitely some legitimate concerns about GMO's as you mentioned, but there are also lots of positives. Many of the potential issues are avoidable, so it really is just a matter of proper regulation and continued research. [NEWLINE] [NEWLINE] The reason the opposition to GMO's was brought up, and the reason I compared it to the Tea Party on the right (a compaison which I perhaps should have more carefully explained) is that a large percentage of those opposed do not reject it for those legitimate concerns but out of a natural product fetish and/or an excessive fear of corporate control over food. The main difference, of course, being that these movements are not nearly as central to the modern liberal platform as the tea party is for the conservative counterparts. [NEWLINE] [NEWLINE] **TL;DR** [NEWLINE] [NEWLINE] **There is legitimate opposition to GMO's, but it is not what OP was criticizing nor what I was attempting to. The demographic in question does not represent liberalism nor do the issues make it hypocritical or equivalent to conservatism.**</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Imagine<mask> it would feel to see a player with better skills lose just<mask> their opponent got a pokeball with Kyogre in it. Or they got the hammer. Or the invincibility star. Or any of the other items that introduce an unfair advantage. [ENDQ] [NEWLINE] <mask> they really had better skills they could've dodged the hammer or the invincible player... [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> a key thing to realize here is that fun probably isn't the reason the majority of people play competitive matches. Whenever you do something competitively, you want that thing to have<mask> little luck involved<mask> possible. That's<mask> the only gambling games done professionally / competitively are the ones<mask> the players have the ability to apply some strategy (there are professional poker and blackjack players,<mask> not professional roulette players or slot machine players). [ENDQ] [NEWLINE] Yeah that makes sense to me.<mask> then you say [NEWLINE] [NEWLINE] [STARTQ] Smash with items on is a crapshoot. [ENDQ] [NEWLINE] <mask><mask> you're meaning to say is that Smash with items on is more comparable to a roulette wheel or slot machines than to poker or blackjack? That sounds kinda bogus to me... [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask> people not using the game the way it is intended, I don't know<mask> to tell you. This is kind of just<mask> humanity progresses. The inventor of basketball did not have it's current NBA incarnation in mind. And Henry Ford did not have the Ford Focus in mind<mask> he designed the Model T,<mask> aren't we all glad people changed things along the way<mask> that we have<mask> we have now? An inventor can create something,<mask> once they give it to the world, it belongs to the people. They can play Smash whichever way they want. [ENDQ] [NEWLINE] True. ∆</s>
Label encoding: <s> [STARTQ] Imagine how it would feel to see a player with better skills lose just because their opponent got a pokeball with Kyogre in it. Or they got the hammer. Or the invincibility star. Or any of the other items that introduce an unfair advantage. [ENDQ] [NEWLINE] If they really had better skills they could've dodged the hammer or the invincible player... [NEWLINE] [NEWLINE] [STARTQ] I think a key thing to realize here is that fun probably isn't the reason the majority of people play competitive matches. Whenever you do something competitively, you want that thing to have as little luck involved as possible. That's why the only gambling games done professionally / competitively are the ones where the players have the ability to apply some strategy (there are professional poker and blackjack players, but not professional roulette players or slot machine players). [ENDQ] [NEWLINE] Yeah that makes sense to me. But then you say [NEWLINE] [NEWLINE] [STARTQ] Smash with items on is a crapshoot. [ENDQ] [NEWLINE] So what you're meaning to say is that Smash with items on is more comparable to a roulette wheel or slot machines than to poker or blackjack? That sounds kinda bogus to me... [NEWLINE] [NEWLINE] [STARTQ] As far as people not using the game the way it is intended, I don't know what to tell you. This is kind of just how humanity progresses. The inventor of basketball did not have it's current NBA incarnation in mind. And Henry Ford did not have the Ford Focus in mind when he designed the Model T, but aren't we all glad people changed things along the way so that we have what we have now? An inventor can create something, but once they give it to the world, it belongs to the people. They can play Smash whichever way they want. [ENDQ] [NEWLINE] True. ∆</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, this is actually quite an interesting philosophical problem. At<mask> point are we no longer talking about the same 'person'. [NEWLINE] [NEWLINE] <mask> one of your good friends talks about /u/Mavericgamer and says all sorts of mostly true things about you, and I say, "oh, no, /u/Mavericgamer is a small, elderly Asian lady who spends her days talking to plum trees and sells coconut oil by the barrell", is it meaningful to say we are talking about the same person? Is the use of the same name enough to make this meaningfully the same? [NEWLINE] [NEWLINE] It's even worse<mask> you're not a monotheist in this debate,<mask> essentially you think that all the participants are talking about a non-existent being. Imagine, person A says "Well I know that unicorns are pink, can only be ridden by virgin girls, and are made up of fairy floss", and person B says "No, no, in my experience all unicorns are Silicon-based life forms and have no respect for virgin girls at all", and person C says "no, no, you are both wrong. You are talking about other creatures and everyone knows that unicorns are faerie creatures from an alternate dimension,<mask> they are carbon based after all". [NEWLINE] [NEWLINE] In<mask> sense are they talking about the same thing? Is the use of 'unicorn' enough to secure it?<mask> happens<mask> we decide that *different names* would be more accurate,<mask> we discovered three different creatures that fit these descriptions? [NEWLINE] [NEWLINE] Saying that things are 'the same'<mask> the same nominaliser is in play is, in my view, insufficient. There has to be a real degree of verisimilitude between the things compared.</s>
Label encoding: <s>Well, this is actually quite an interesting philosophical problem. At what point are we no longer talking about the same 'person'. [NEWLINE] [NEWLINE] If one of your good friends talks about /u/Mavericgamer and says all sorts of mostly true things about you, and I say, "oh, no, /u/Mavericgamer is a small, elderly Asian lady who spends her days talking to plum trees and sells coconut oil by the barrell", is it meaningful to say we are talking about the same person? Is the use of the same name enough to make this meaningfully the same? [NEWLINE] [NEWLINE] It's even worse if you're not a monotheist in this debate, because essentially you think that all the participants are talking about a non-existent being. Imagine, person A says "Well I know that unicorns are pink, can only be ridden by virgin girls, and are made up of fairy floss", and person B says "No, no, in my experience all unicorns are Silicon-based life forms and have no respect for virgin girls at all", and person C says "no, no, you are both wrong. You are talking about other creatures and everyone knows that unicorns are faerie creatures from an alternate dimension, but they are carbon based after all". [NEWLINE] [NEWLINE] In what sense are they talking about the same thing? Is the use of 'unicorn' enough to secure it? What happens if we decide that *different names* would be more accurate, because we discovered three different creatures that fit these descriptions? [NEWLINE] [NEWLINE] Saying that things are 'the same' because the same nominaliser is in play is, in my view, insufficient. There has to be a real degree of verisimilitude between the things compared.</s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe people pursue these degrees partly for the reason you listed: they have an unwarranted fear of mathematics and "hard" sciences. I firmly believe that most people are capable of understanding a topic and pursuing a career in virtually in field<mask><mask><mask> they apply themselves appropriately. The amount of time devoted to studying and learning the topic will vary between individuals,<mask> it is possible nonetheless. [NEWLINE] [NEWLINE] edit: I<mask> believe that people pursue these degrees<mask> they think it will be easy to get a job and simply don't do enough research/aren't aware of<mask> the job market actually is like. [NEWLINE] [NEWLINE] I<mask> acknowledge the idea that people will use their degrees to identify themselves and this can be somewhat damaging to their conscious self-esteem,<mask> this does not mean that they are *justified* in complaining.<mask> they don't have the will to further their education in their field<mask> that they can actually acquire a job in that field, then they weren't that concerned with the field in the first place and merely went through college to get a job. [NEWLINE] [NEWLINE] Personally,<mask><mask> that going into college straight after high school is a problem that needs to be addressed (I'm not sure<mask> the situation is in countries outside of the United States of America,<mask> please forgive me for any generalizations). These kids haven't fully developed and are expected to make a decision that is implicated to last the rest of their life.<mask> many teenagers do you think truly know<mask> they want to do with their lives<mask> they get out of high school?<mask> many people stick with their first choice of major upon entering college? We should do a lot more to educate younger generations in picking a career and furthering their knowledge in<mask> much work will be required to acquire such a degree/job.</s>
Label encoding: <s>I believe people pursue these degrees partly for the reason you listed: they have an unwarranted fear of mathematics and "hard" sciences. I firmly believe that most people are capable of understanding a topic and pursuing a career in virtually in field as long as they apply themselves appropriately. The amount of time devoted to studying and learning the topic will vary between individuals, but it is possible nonetheless. [NEWLINE] [NEWLINE] edit: I also believe that people pursue these degrees because they think it will be easy to get a job and simply don't do enough research/aren't aware of what the job market actually is like. [NEWLINE] [NEWLINE] I also acknowledge the idea that people will use their degrees to identify themselves and this can be somewhat damaging to their conscious self-esteem, but this does not mean that they are *justified* in complaining. If they don't have the will to further their education in their field so that they can actually acquire a job in that field, then they weren't that concerned with the field in the first place and merely went through college to get a job. [NEWLINE] [NEWLINE] Personally, I think that going into college straight after high school is a problem that needs to be addressed (I'm not sure what the situation is in countries outside of the United States of America, so please forgive me for any generalizations). These kids haven't fully developed and are expected to make a decision that is implicated to last the rest of their life. How many teenagers do you think truly know what they want to do with their lives when they get out of high school? How many people stick with their first choice of major upon entering college? We should do a lot more to educate younger generations in picking a career and furthering their knowledge in how much work will be required to acquire such a degree/job.</s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would agree that a religious exemption shouldn't be exclusive,<mask> I would<mask> like to say that "for any reason whatsoever" is much too broad. [NEWLINE] [NEWLINE] Consider the case of [conscientious objectors]( [URL] ) and military service. A conscientious objector is defined<mask> someone who has claimed the right to refuse military service on the grounds of "freedom of thought, conscience, and/or religion." Note that religion is only one of the possible ways to be a conscientious objector. Do you believe it's okay that you can claim to be a conscientious objector to avoid combat duties? I would<mask><mask> the mental toll it would take on a conscientious objector to be forced to serve in combat duty is not sufficient to outweigh a need for soldiers-<mask> we should grant an exemption for this personal conviction. [NEWLINE] [NEWLINE] Looking at religious beliefs objectively, without regard for the truth of any particular religion, that is all they are- strong personal convictions, which may be essential to a person's mental well-being.<mask> the question becomes, "Should we allow a person's strong personal convictions to factor in whether or not a law should apply equally?"<mask><mask> in rare cases we should, whether or not that belief is religious in nature,<mask> it is generally detrimental to force someone to act against their own convictions.<mask> it becomes<mask> detrimental that we cause them significant harm for no real gain, then perhaps we can violate the rule of law. I don't think this situation is very common (and certainly not in the case of Hobby Lobby)-<mask><mask>, off the top of my head, conscientious objection is about the only one I can think of.<mask> at least in principle, it should be possible to claim a religious exemption under the banner of strong personal conviction. </s>
Label encoding: <s>I would agree that a religious exemption shouldn't be exclusive, but I would also like to say that "for any reason whatsoever" is much too broad. [NEWLINE] [NEWLINE] Consider the case of [conscientious objectors]( [URL] ) and military service. A conscientious objector is defined as someone who has claimed the right to refuse military service on the grounds of "freedom of thought, conscience, and/or religion." Note that religion is only one of the possible ways to be a conscientious objector. Do you believe it's okay that you can claim to be a conscientious objector to avoid combat duties? I would argue that the mental toll it would take on a conscientious objector to be forced to serve in combat duty is not sufficient to outweigh a need for soldiers- so we should grant an exemption for this personal conviction. [NEWLINE] [NEWLINE] Looking at religious beliefs objectively, without regard for the truth of any particular religion, that is all they are- strong personal convictions, which may be essential to a person's mental well-being. So the question becomes, "Should we allow a person's strong personal convictions to factor in whether or not a law should apply equally?" I think in rare cases we should, whether or not that belief is religious in nature, because it is generally detrimental to force someone to act against their own convictions. If it becomes so detrimental that we cause them significant harm for no real gain, then perhaps we can violate the rule of law. I don't think this situation is very common (and certainly not in the case of Hobby Lobby)- in fact, off the top of my head, conscientious objection is about the only one I can think of. But at least in principle, it should be possible to claim a religious exemption under the banner of strong personal conviction. </s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I submit that every single female child should get an implant before puberty (until they can find one for males, in which case all males should get one too). [ENDQ] [NEWLINE] <mask>, puberty is a pretty critical time in development.  Anything that is hormonally based is going to cause serious health consequences.  Even things which are mechanical are going to be problematic<mask> the body changes and grows. [NEWLINE] [NEWLINE] <mask>,<mask> we're gong to be altering women's physiology we should definitely be altering men's too.  Its actually *much* less invasive to do a vasectomy than a tubal ligation.  Are you still in favor of this argument<mask> all male children will be getting vasectomies? [NEWLINE] [NEWLINE] EDIT: [NEWLINE] [NEWLINE] With this first (and incredibly serious) governmental invasion of bodily autonomy we've set the precedent for much, much more.  Obviously abortion is first to go.  Mandatory organ donation will probably soon follow.  Is this the only grounds society wants to use for its eugenics program?  Seems to me that health is at least<mask> salient<mask> wealth. <mask> those with medical concerns can't have families either, those with disabilities?  I don't think<mask>, even asthma, allergies or nearsightedness is really not ideal.  They all need to keep theirs in<mask> well.  Now assuming none of that happens.  Assuming we only use wealth<mask> an indicator..... [NEWLINE] [NEWLINE] Aboriginals almost universally fall into the lower economic ring.  Black people have a significant issue with that<mask> well and definitely recent immigrants.  Congratulations, you've just committed peaceful genocide.  Well, assuming everyone complies quietly it will be peaceful.  People are known for their tacit acceptance of dictatorship and genocide right? [NEWLINE] </s>
Label encoding: <s> [STARTQ] I submit that every single female child should get an implant before puberty (until they can find one for males, in which case all males should get one too). [ENDQ] [NEWLINE] Firstly, puberty is a pretty critical time in development.  Anything that is hormonally based is going to cause serious health consequences.  Even things which are mechanical are going to be problematic as the body changes and grows. [NEWLINE] [NEWLINE] Secondly, if we're gong to be altering women's physiology we should definitely be altering men's too.  Its actually *much* less invasive to do a vasectomy than a tubal ligation.  Are you still in favor of this argument if all male children will be getting vasectomies? [NEWLINE] [NEWLINE] EDIT: [NEWLINE] [NEWLINE] With this first (and incredibly serious) governmental invasion of bodily autonomy we've set the precedent for much, much more.  Obviously abortion is first to go.  Mandatory organ donation will probably soon follow.  Is this the only grounds society wants to use for its eugenics program?  Seems to me that health is at least as salient as wealth.  So those with medical concerns can't have families either, those with disabilities?  I don't think so, even asthma, allergies or nearsightedness is really not ideal.  They all need to keep theirs in as well.  Now assuming none of that happens.  Assuming we only use wealth as an indicator..... [NEWLINE] [NEWLINE] Aboriginals almost universally fall into the lower economic ring.  Black people have a significant issue with that as well and definitely recent immigrants.  Congratulations, you've just committed peaceful genocide.  Well, assuming everyone complies quietly it will be peaceful.  People are known for their tacit acceptance of dictatorship and genocide right? [NEWLINE] </s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yea, I might be suffering from a multitude of delusions. Or I could be looking at the reality of things. [NEWLINE] [NEWLINE] 1. Its really sad.<mask> must I compete with my fellow human being to survive in this social matrix?<mask><mask>, I did not "earn" it in the sense that I worked hard. No, I worked smart, and I gamed the system.<mask> is that really a crime? [NEWLINE] [NEWLINE] 2. I am only pointing to exceptions<mask> I am trying to say to you, all of you "<mask> is there exceptions in the first place?" Isn't the point of education to make more people like Steve Jobs, Bill Gates, Mark Zurkerberg?<mask> that is the point of education, they<mask> is it failing the millions of students out there? Educated people, create new systems and challange the status quo. Do you honestly believe the rich and powerful<mask> their power to be challenged? [NEWLINE] [NEWLINE] 2.2  That is<mask> most degree mills (colleges and universities) are set up that way. They get money, the governments get money, and you, the student, the one who is to learn end up a DEBT SLAVE, trying to pay off your debt with interest. [NEWLINE] [NEWLINE] 3.1 There was no damage inflicted.  I helped my buddy pass not by cheating no, I did it by actually helping him study believe it or not.  Some subjects I did not cheat in. [NEWLINE] [NEWLINE] 3.2 Ok, after you get out of grad school then<mask>? Think about it, after your grad school is over, no more gpa right?<mask> you get a job<mask> an astrophysicist or whatever you went to grad school for, then<mask>? </s>
Label encoding: <s>Yea, I might be suffering from a multitude of delusions. Or I could be looking at the reality of things. [NEWLINE] [NEWLINE] 1. Its really sad. Why must I compete with my fellow human being to survive in this social matrix? I agree, I did not "earn" it in the sense that I worked hard. No, I worked smart, and I gamed the system. But is that really a crime? [NEWLINE] [NEWLINE] 2. I am only pointing to exceptions because I am trying to say to you, all of you " why is there exceptions in the first place?" Isn't the point of education to make more people like Steve Jobs, Bill Gates, Mark Zurkerberg? If that is the point of education, they why is it failing the millions of students out there? Educated people, create new systems and challange the status quo. Do you honestly believe the rich and powerful what their power to be challenged? [NEWLINE] [NEWLINE] 2.2  That is why most degree mills (colleges and universities) are set up that way. They get money, the governments get money, and you, the student, the one who is to learn end up a DEBT SLAVE, trying to pay off your debt with interest. [NEWLINE] [NEWLINE] 3.1 There was no damage inflicted.  I helped my buddy pass not by cheating no, I did it by actually helping him study believe it or not.  Some subjects I did not cheat in. [NEWLINE] [NEWLINE] 3.2 Ok, after you get out of grad school then what? Think about it, after your grad school is over, no more gpa right? When you get a job as an astrophysicist or whatever you went to grad school for, then what? </s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, I wouldn't do this personally<mask> I don't mind walking. I don't even look for close parking spots. This discussion is purely hypothetical looking at the issue from an objective point of view. This post was inspired by a question I posted to /r/legaladvice and was told that it's a douche move. I didn't feel it was appropriate to discuss the douchiness of it there<mask> I decided to make this post. [NEWLINE] [NEWLINE] You go into a parking lot and are trying to find a place to park. There's one 15 spots away from the door and then there's one that's 3 spots away from the door. One of them is meant for expectant mothers<mask> there is no punishment for parking there<mask> you are not an expectant mother which makes the two spots equivalent in everything expect for distance from the door. Objectively, it's better to choose the closer one. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>First of all, I wouldn't do this personally because I don't mind walking. I don't even look for close parking spots. This discussion is purely hypothetical looking at the issue from an objective point of view. This post was inspired by a question I posted to /r/legaladvice and was told that it's a douche move. I didn't feel it was appropriate to discuss the douchiness of it there so I decided to make this post. [NEWLINE] [NEWLINE] You go into a parking lot and are trying to find a place to park. There's one 15 spots away from the door and then there's one that's 3 spots away from the door. One of them is meant for expectant mothers but there is no punishment for parking there if you are not an expectant mother which makes the two spots equivalent in everything expect for distance from the door. Objectively, it's better to choose the closer one. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is false. I can easily see this scene being pivotal for Theon's character. His tears were emphasized heavily. It's extremely presumptuous to assume that a scene, especially in a well written and coherent show like GOT, is pointless before we've gotten the chance to see<mask><mask> anything is impacted by the scene. In the preview for the next episode we see an interesting interaction between Sansa and Throb. [NEWLINE] [NEWLINE] <mask>, in whet world is being taped not character development. Not even in GOT, not even to the eternally tortured Sansa, is tape not a notable thing. Being tortured in different ways is an important part of Sansa's character, and those different ways of being tortured affect her in different ways. It's<mask> naive<mask> Sansa's old character yo thing that<mask> she's matured some and it's "time" for her revenge that she's supposed to stop getting tortured. Are we even watching the same show? There is no "time" for Sands to stop being tortured any more then there is the right "time" for Robb Stark to win the war and take the Iron Throne. Being Sansa is torture. The idea that that would change<mask> she is wed to *Ramsey Bolton* of all people is actually laughable to me. The reasons<mask> this tape scene is "too much"<mask> the GOT universe prides itself on not shying away from horror, reeks (no pun intended) of rationalism, of fishing for excuses. SANSA GOT RAPED.<mask> this fact got to you, you've simply found *that thing* which gets to you like baby murderer gets to other people. You got offended just like those other people "get offended." It's embarrassing.</s>
Label encoding: <s>This is false. I can easily see this scene being pivotal for Theon's character. His tears were emphasized heavily. It's extremely presumptuous to assume that a scene, especially in a well written and coherent show like GOT, is pointless before we've gotten the chance to see what if anything is impacted by the scene. In the preview for the next episode we see an interesting interaction between Sansa and Throb. [NEWLINE] [NEWLINE] Also, in whet world is being taped not character development. Not even in GOT, not even to the eternally tortured Sansa, is tape not a notable thing. Being tortured in different ways is an important part of Sansa's character, and those different ways of being tortured affect her in different ways. It's as naive as Sansa's old character yo thing that because she's matured some and it's "time" for her revenge that she's supposed to stop getting tortured. Are we even watching the same show? There is no "time" for Sands to stop being tortured any more then there is the right "time" for Robb Stark to win the war and take the Iron Throne. Being Sansa is torture. The idea that that would change when she is wed to *Ramsey Bolton* of all people is actually laughable to me. The reasons why this tape scene is "too much" when the GOT universe prides itself on not shying away from horror, reeks (no pun intended) of rationalism, of fishing for excuses. SANSA GOT RAPED. If this fact got to you, you've simply found *that thing* which gets to you like baby murderer gets to other people. You got offended just like those other people "get offended." It's embarrassing.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm still not sure<mask> people are entitled to in your framework, and<mask> that (and no more) is justified.  I don't think the entitlements/aggression framework is consistent and more useful than a consequentialist framework. [NEWLINE] [NEWLINE] [STARTQ] Supposing for a second that you are entitled to healthcare, it still wouldn't be aggression for me to deny sending you resources for your treatment. [ENDQ] [NEWLINE] Let me reword it.  The country is entitled to a royalty/commission on everyone's earnings, and everyone is entitled to use part of those earnings on healthcare (<mask> you want to be really pedantic, country is a proxy for the people in the country). <mask> is this not justified? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> would you be entitled to clean water or to healthcare?<mask> you're alone in a desert, through no fault of anyone's it's not<mask><mask> you have a moral greivance against anyone. [ENDQ] [NEWLINE] We don't live in a desert.  We live in a modern-day economy with innovations and institutions.  Some of those expand our possibilities and some restrict our possibilities. [NEWLINE] [NEWLINE] <mask> an example, suppose I live in a city<mask> all the water rights are bought up. <mask> those water companies didn't exist, I could survive by drawing water out of the stream. <mask> I have to pay the water companies, and maybe I don't have the money for it.  Is that aggression?  Am I entitled to the water? [NEWLINE] [NEWLINE] [STARTQ] I'm not depriving those 1000 civilians of anything. [ENDQ] [NEWLINE] (EDITED)<mask><mask> that's awfully narrow. <mask><mask> you're one of those 1000 civilians?  Can you then defend yourself by bombing the munitions depot?</s>
Label encoding: <s>I'm still not sure what people are entitled to in your framework, and why that (and no more) is justified.  I don't think the entitlements/aggression framework is consistent and more useful than a consequentialist framework. [NEWLINE] [NEWLINE] [STARTQ] Supposing for a second that you are entitled to healthcare, it still wouldn't be aggression for me to deny sending you resources for your treatment. [ENDQ] [NEWLINE] Let me reword it.  The country is entitled to a royalty/commission on everyone's earnings, and everyone is entitled to use part of those earnings on healthcare ( if you want to be really pedantic, country is a proxy for the people in the country).  Why is this not justified? [NEWLINE] [NEWLINE] [STARTQ] But why would you be entitled to clean water or to healthcare? If you're alone in a desert, through no fault of anyone's it's not as if you have a moral greivance against anyone. [ENDQ] [NEWLINE] We don't live in a desert.  We live in a modern-day economy with innovations and institutions.  Some of those expand our possibilities and some restrict our possibilities. [NEWLINE] [NEWLINE] As an example, suppose I live in a city where all the water rights are bought up.  If those water companies didn't exist, I could survive by drawing water out of the stream.  But I have to pay the water companies, and maybe I don't have the money for it.  Is that aggression?  Am I entitled to the water? [NEWLINE] [NEWLINE] [STARTQ] I'm not depriving those 1000 civilians of anything. [ENDQ] [NEWLINE] (EDITED) I think that's awfully narrow.  What if you're one of those 1000 civilians?  Can you then defend yourself by bombing the munitions depot?</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're overestimating the significance of the US's involvement there. The war would have ended either way. The next 50 years of history is all that would change. [NEWLINE] <mask> yes, it does. I'm actually in favor of the draft for more than a few reasons<mask> I have to disagree with you here. Humanity is all about caring and helping others out<mask> you can. Until the bombs start dropping and helping out means taking boys fresh out of highschool and shipping them across the sea to go kill another group of people and watch their friends die around them. That is<mask> shit gets a little more real and that's<mask> I believe the draft should be used for defense alone. I fully agree that you should fight for whats right and be willing to lay down your life.<mask><mask> you're not, then fuck it. The world is<mask> it is and an individual life honestly is more important than the greater good<mask> it's your life.<mask><mask> the time comes I'll be right here ready to go,<mask> I'm not going to judge anyone who jumps ship and decides living peacefully somewhere else is better than defending their home. It's not my place to control them, and I do not believe it is anyone's place<mask> it comes to your life.<mask> a person abhors violence and wanted to live their life without ever touching a gun then no one should be able to force them to.<mask> they want to move to another country instead, go ahead, you should always have freedom. [NEWLINE] [NEWLINE] Now<mask> you want to skip the draft<mask> continue to live in the country that's a problem to me.<mask> it comes down to fight or flight there should be no third option<mask> you chill out until it blows over.</s>
Label encoding: <s>I think you're overestimating the significance of the US's involvement there. The war would have ended either way. The next 50 years of history is all that would change. [NEWLINE] Also yes, it does. I'm actually in favor of the draft for more than a few reasons but I have to disagree with you here. Humanity is all about caring and helping others out when you can. Until the bombs start dropping and helping out means taking boys fresh out of highschool and shipping them across the sea to go kill another group of people and watch their friends die around them. That is when shit gets a little more real and that's why I believe the draft should be used for defense alone. I fully agree that you should fight for whats right and be willing to lay down your life. However if you're not, then fuck it. The world is what it is and an individual life honestly is more important than the greater good when it's your life. So if the time comes I'll be right here ready to go, but I'm not going to judge anyone who jumps ship and decides living peacefully somewhere else is better than defending their home. It's not my place to control them, and I do not believe it is anyone's place when it comes to your life. If a person abhors violence and wanted to live their life without ever touching a gun then no one should be able to force them to. If they want to move to another country instead, go ahead, you should always have freedom. [NEWLINE] [NEWLINE] Now if you want to skip the draft but continue to live in the country that's a problem to me. When it comes down to fight or flight there should be no third option where you chill out until it blows over.</s>
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> most of the data we have on this kind of stuff is skewed heavily.<mask> many people are honestly willing to admit they were denied a promotion or a raise<mask> they themselves suck? Its human nature to try to blame someone, and unfortunately a lot of groups (like feminism) lull people into the comfort zone of "its not your fault." Well, Bob got a raise and Sue didn't. Is that part of the glass ceiling? Well, sometimes Bob gets a raise<mask> he's "one of the boys." Sometimes Bob gets the raise<mask> Sue is a slacker and spends all day on facebook.<mask><mask> we collect statistics, it doesn't matter. Sue *feels* that she's been discriminated against. [NEWLINE] [NEWLINE] I think it was Scott Adams who had a pretty good analogy. Imagine a job interview for a high paying job, one man and one woman are applying. They are taken to a large hangar<mask> 100 fat guys drop their pants and proceed to present their asses for kissing. The woman says this is outrageous and starts calling up her friends and support group claiming sexual harassment, creating an online group to talk about it, etc etc. The guy is already halfway down the line, kissing every ass and slipping a business card into every crack. Who gets the job? The guy does - not<mask> he has a set of XY chromosomes,<mask><mask> he's willing to play the game. And he *doesn't* have the excuse of "They don't like be<mask> of the way I was born." [NEWLINE] [NEWLINE] Is there discrimination against [insert any group here]? Of course there is.<mask> is it keeping a whole gender from exceeding a certain level of pay? Thats sort of ridiculous to even think of. </s>
Label encoding: <s>I think most of the data we have on this kind of stuff is skewed heavily. How many people are honestly willing to admit they were denied a promotion or a raise because they themselves suck? Its human nature to try to blame someone, and unfortunately a lot of groups (like feminism) lull people into the comfort zone of "its not your fault." Well, Bob got a raise and Sue didn't. Is that part of the glass ceiling? Well, sometimes Bob gets a raise because he's "one of the boys." Sometimes Bob gets the raise because Sue is a slacker and spends all day on facebook. Yet when we collect statistics, it doesn't matter. Sue *feels* that she's been discriminated against. [NEWLINE] [NEWLINE] I think it was Scott Adams who had a pretty good analogy. Imagine a job interview for a high paying job, one man and one woman are applying. They are taken to a large hangar where 100 fat guys drop their pants and proceed to present their asses for kissing. The woman says this is outrageous and starts calling up her friends and support group claiming sexual harassment, creating an online group to talk about it, etc etc. The guy is already halfway down the line, kissing every ass and slipping a business card into every crack. Who gets the job? The guy does - not because he has a set of XY chromosomes, but because he's willing to play the game. And he *doesn't* have the excuse of "They don't like be because of the way I was born." [NEWLINE] [NEWLINE] Is there discrimination against [insert any group here]? Of course there is. But is it keeping a whole gender from exceeding a certain level of pay? Thats sort of ridiculous to even think of. </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wow, a nice case of arguing semantics and not the meat of the matter up there. You give out deltas easily my friend... [NEWLINE] [NEWLINE] <mask> for your original stance: any preference is a valid preference. The movement against slut shaming is just a bunch of women who don't really understand<mask> they're talking about. Basically they demand that men lower their standards<mask> those standards are too high and they don't want to live up to them. Which is funny<mask> you think about it objectively. [NEWLINE] [NEWLINE] *"<mask> dare you prefer women who can control their sexual urges? Women who are more likely to cheat on you are just<mask> good<mask> those who aren't, you oppressive shitlord! You mustn't prefer faithful types<mask> it hurts the egos of us unfaithful types!"* [NEWLINE] [NEWLINE] <mask>, yeah, it doesn't matter<mask> anyone tries to redefine the word "slut",<mask>'s important is the phenomenon behind the words. Men have the right to have high standards for female behaviour (just<mask> the other way around), and<mask> some people don't like this they can go pound sand. Nobody should be able to shame others into lowering their standards and accept behaviour that they don't want to accept. [NEWLINE] [NEWLINE] <mask>, there are actual statistics out there showing that the higher the previous partner count of a woman is, the more likely she will divorce her husband. It's an objective statistical fact that promiscuous women make worse long term partners than non-promiscuous women. It's only natural that sluts don't like this fact and do everything in their power to shame men into forgetting it. To their dismay it only works on low value males - men with intact spines hold on to their values.</s>
Label encoding: <s>Wow, a nice case of arguing semantics and not the meat of the matter up there. You give out deltas easily my friend... [NEWLINE] [NEWLINE] As for your original stance: any preference is a valid preference. The movement against slut shaming is just a bunch of women who don't really understand what they're talking about. Basically they demand that men lower their standards because those standards are too high and they don't want to live up to them. Which is funny if you think about it objectively. [NEWLINE] [NEWLINE] *" How dare you prefer women who can control their sexual urges? Women who are more likely to cheat on you are just as good as those who aren't, you oppressive shitlord! You mustn't prefer faithful types because it hurts the egos of us unfaithful types!"* [NEWLINE] [NEWLINE] So, yeah, it doesn't matter how anyone tries to redefine the word "slut", what's important is the phenomenon behind the words. Men have the right to have high standards for female behaviour (just as the other way around), and if some people don't like this they can go pound sand. Nobody should be able to shame others into lowering their standards and accept behaviour that they don't want to accept. [NEWLINE] [NEWLINE] Also, there are actual statistics out there showing that the higher the previous partner count of a woman is, the more likely she will divorce her husband. It's an objective statistical fact that promiscuous women make worse long term partners than non-promiscuous women. It's only natural that sluts don't like this fact and do everything in their power to shame men into forgetting it. To their dismay it only works on low value males - men with intact spines hold on to their values.</s>
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This sounds great in theory<mask> in actuality I find most people can't accept that. [ENDQ] [NEWLINE] Well, then they're not ready for sex. Pregnancy is always a risk. You take risks every day. You could die in a car accident on the way to work, your train could derail, you could get robbed coming home, whatever. [NEWLINE] [NEWLINE] You can either stay in your room all day, or take risks. The benefit of earning an income at work outweighs the risk of driving there. There is a chance you could be in an accident, it may not be a huge risk,<mask> it's there. Sex is no different. You want sex, it feels pleasurable,<mask> you have sex,<mask> the risk of pregnancy. The pleasure of sex is worth the risk. [NEWLINE] [NEWLINE] That's<mask> we have car insurance laws,<mask> you can face the consequences of your actions.<mask> you/someone else gets in an accident, and it's their fault, you/they must have insurance to pay for the damage they cause.<mask> liability insurance wasn't required,<mask> get it? You could cause an accident, and drive off scott free. The person who wasn't at fault would be stuck with a repair bill for something they didn't cause. [NEWLINE] [NEWLINE] We have child support laws in place for the same reason.<mask> I,<mask> a man get someone pregnant, I need to take responsibility for my actions. Just like the law requiring insurance offers a financial disincentive to get in accidents, child support laws are a disincentive to get women pregnant. [NEWLINE] [NEWLINE] <mask> I could just opt out of child support,<mask>'s to stop me from just knocking up every woman in town, and then bailing?</s>
Label encoding: <s> [STARTQ] This sounds great in theory but in actuality I find most people can't accept that. [ENDQ] [NEWLINE] Well, then they're not ready for sex. Pregnancy is always a risk. You take risks every day. You could die in a car accident on the way to work, your train could derail, you could get robbed coming home, whatever. [NEWLINE] [NEWLINE] You can either stay in your room all day, or take risks. The benefit of earning an income at work outweighs the risk of driving there. There is a chance you could be in an accident, it may not be a huge risk, but it's there. Sex is no different. You want sex, it feels pleasurable, so you have sex, despite the risk of pregnancy. The pleasure of sex is worth the risk. [NEWLINE] [NEWLINE] That's why we have car insurance laws, so you can face the consequences of your actions. If you/someone else gets in an accident, and it's their fault, you/they must have insurance to pay for the damage they cause. If liability insurance wasn't required, why get it? You could cause an accident, and drive off scott free. The person who wasn't at fault would be stuck with a repair bill for something they didn't cause. [NEWLINE] [NEWLINE] We have child support laws in place for the same reason. If I, as a man get someone pregnant, I need to take responsibility for my actions. Just like the law requiring insurance offers a financial disincentive to get in accidents, child support laws are a disincentive to get women pregnant. [NEWLINE] [NEWLINE] If I could just opt out of child support, what's to stop me from just knocking up every woman in town, and then bailing?</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've volunteered at homeless shelters before.  Not one of the people I worked with was unwilling to work.  The vast majority were victims of some sort of unfortunate circumstance, like having to choose between paying rent and paying hospital bills,<mask> they simply didn't have enough income for both.  And then, once they became homeless, there was a stigma against them, from people who think exactly the way you do.  There's an "Oh, you're homeless,<mask> you must be too lazy to hold down a job,<mask> we can't possibly hire you" mentality.  Couple that with the fact that the homeless people can't,<mask> a general rule, afford to get nice clean clothes for interviews and other such details, and you find that it's really rather difficult to go from being homeless back to having a home.   It's certainly not<mask> they're lazy. [NEWLINE] [NEWLINE] The same applies to most of the people on welfare.  Sure, there are a few people out there who sit around doing nothing and abuse the system,<mask> that's a really trivial minority.  Most really want to work,<mask> for much the same reason, can't get hired<mask> people think they must be lazy<mask> they aren't already working.  You're 16, and it sounds like you're relatively well off,<mask> employers understand that you don't already have a job<mask> you're busy with school.  They have a harder time justifying hiring that 35 year old guy who's spent the past two months on welfare,<mask><mask> he's spent the past two months on welfare he must have been lazing around. [NEWLINE] [NEWLINE] <mask> feel like I'm starting to repeat myself,<mask> I hope I'm getting my point across.</s>
Label encoding: <s>I've volunteered at homeless shelters before.  Not one of the people I worked with was unwilling to work.  The vast majority were victims of some sort of unfortunate circumstance, like having to choose between paying rent and paying hospital bills, because they simply didn't have enough income for both.  And then, once they became homeless, there was a stigma against them, from people who think exactly the way you do.  There's an "Oh, you're homeless, so you must be too lazy to hold down a job, so we can't possibly hire you" mentality.  Couple that with the fact that the homeless people can't, as a general rule, afford to get nice clean clothes for interviews and other such details, and you find that it's really rather difficult to go from being homeless back to having a home.   It's certainly not because they're lazy. [NEWLINE] [NEWLINE] The same applies to most of the people on welfare.  Sure, there are a few people out there who sit around doing nothing and abuse the system, but that's a really trivial minority.  Most really want to work, but for much the same reason, can't get hired because people think they must be lazy if they aren't already working.  You're 16, and it sounds like you're relatively well off, so employers understand that you don't already have a job because you're busy with school.  They have a harder time justifying hiring that 35 year old guy who's spent the past two months on welfare, because if he's spent the past two months on welfare he must have been lazing around. [NEWLINE] [NEWLINE] If feel like I'm starting to repeat myself, but I hope I'm getting my point across.</s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No they weren't.<mask> strong unions first started forming, the government beat them up and jailed them in order to suppress their speech.<mask> the civil rights movement got rolling, the government beat them up and jailed them in order to suppress their speech. The government has never hesitated to ignore the principle of free speech<mask> it comes to anything truly subversive. And<mask> I don't like the idea of beating people up, and disagree with those specific examples, I don't think they were being hypocrites. Truly unlimited free speech is too dangerous to allow. [ENDQ] [NEWLINE] <mask> you cherry-picked two examples, which are questionable (<mask> early Unions were intricately tied to mob activity, and many of the civil rights protests were acts of civil disobedience.), and you ignored the overwhelmingly peaceful transitions from Monarchy to Republic to Democracy, from Slavery to Capitaism to Socialism, from Religious to Humanist, etc.   There was bloodshed and suppression, there is no doubt. **<mask> : the liberalistic principle of freedom of speech provided peaceful transitions towards the principles of social progress in far more cases than it failed to.** [NEWLINE] [NEWLINE] Again,<mask> can you denounce the activities of the past and use that to justify them in the present and not understand the moral dilemma therein. [NEWLINE] [NEWLINE] [STARTQ] For example, suppose there's a group that believes all black people are criminals. This group is allowed to speak freely, becomes popular in some areas, and ends up with a bunch of leadership positions in the local governments of the Oklahoma panhandle.<mask><mask> it's perfectly reasonable to be worried about<mask> they'll do with their power, even<mask> most people don't agree with them. [ENDQ] [NEWLINE] That's not an issue of freedom of speech.  </s>
Label encoding: <s> [STARTQ] No they weren't. When strong unions first started forming, the government beat them up and jailed them in order to suppress their speech. When the civil rights movement got rolling, the government beat them up and jailed them in order to suppress their speech. The government has never hesitated to ignore the principle of free speech when it comes to anything truly subversive. And while I don't like the idea of beating people up, and disagree with those specific examples, I don't think they were being hypocrites. Truly unlimited free speech is too dangerous to allow. [ENDQ] [NEWLINE] So you cherry-picked two examples, which are questionable ( as early Unions were intricately tied to mob activity, and many of the civil rights protests were acts of civil disobedience.), and you ignored the overwhelmingly peaceful transitions from Monarchy to Republic to Democracy, from Slavery to Capitaism to Socialism, from Religious to Humanist, etc.   There was bloodshed and suppression, there is no doubt. ** However : the liberalistic principle of freedom of speech provided peaceful transitions towards the principles of social progress in far more cases than it failed to.** [NEWLINE] [NEWLINE] Again, how can you denounce the activities of the past and use that to justify them in the present and not understand the moral dilemma therein. [NEWLINE] [NEWLINE] [STARTQ] For example, suppose there's a group that believes all black people are criminals. This group is allowed to speak freely, becomes popular in some areas, and ends up with a bunch of leadership positions in the local governments of the Oklahoma panhandle. I think it's perfectly reasonable to be worried about what they'll do with their power, even if most people don't agree with them. [ENDQ] [NEWLINE] That's not an issue of freedom of speech.  </s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How do you know they don't actually find her beautiful? Beauty is inherently subjective.<mask> most people find similar things attractive there isn't an objective standard of beauty. Different people weigh very different things. A lot of people find the runway model look less attractive than other body types, for example. Compare Twiggy and Marilyn Monroe, they contrast pretty strongly in a number of traits<mask> both are beautiful. [NEWLINE] [NEWLINE] People are very, very bad at self assessment. We constantly think that we are far more skilled than we actually are. We rate our natural abilities based on peak performance rather than average performance. We invariably overestimate<mask> well we perform compared to peers. This is a question of *perception* and *perspective* more than anything else. We our own body poorly. Some people vastly overstate their appearance and others vastly understate their appearance.<mask> you want to see<mask> attractive you are, look at the people you date<mask> we are pretty good at pairing off with people of similar attractiveness. [NEWLINE] [NEWLINE] The person who posts on the internet who says "I don't look good" almost invariably believes that he or she is significantly less pretty than they actually are. On a hypothetical hot scale this 5 thinks that they are a 2.5. Most people who think that they are a 9 are really somewhere closer to a 6. The farther you get from "average" the more likely you are to be wrong. After all like 68% of us would have to be between a 4 and 6. [NEWLINE] [NEWLINE] <mask> people are saying isn't people trying to feed someone's inflated ego,<mask> bring someone's deflated view of themselves back in line. Is it often exaggerated? Yeah,<mask> that's often times necessary.</s>
Label encoding: <s>How do you know they don't actually find her beautiful? Beauty is inherently subjective. While most people find similar things attractive there isn't an objective standard of beauty. Different people weigh very different things. A lot of people find the runway model look less attractive than other body types, for example. Compare Twiggy and Marilyn Monroe, they contrast pretty strongly in a number of traits but both are beautiful. [NEWLINE] [NEWLINE] People are very, very bad at self assessment. We constantly think that we are far more skilled than we actually are. We rate our natural abilities based on peak performance rather than average performance. We invariably overestimate how well we perform compared to peers. This is a question of *perception* and *perspective* more than anything else. We our own body poorly. Some people vastly overstate their appearance and others vastly understate their appearance. If you want to see how attractive you are, look at the people you date because we are pretty good at pairing off with people of similar attractiveness. [NEWLINE] [NEWLINE] The person who posts on the internet who says "I don't look good" almost invariably believes that he or she is significantly less pretty than they actually are. On a hypothetical hot scale this 5 thinks that they are a 2.5. Most people who think that they are a 9 are really somewhere closer to a 6. The farther you get from "average" the more likely you are to be wrong. After all like 68% of us would have to be between a 4 and 6. [NEWLINE] [NEWLINE] What people are saying isn't people trying to feed someone's inflated ego, but bring someone's deflated view of themselves back in line. Is it often exaggerated? Yeah, but that's often times necessary.</s>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't believe there is a better sex [ENDQ] [NEWLINE] Feminism doesn't either,<mask> it has acknowledged a societal preference for one sex over another. [NEWLINE] [NEWLINE] [STARTQ] Your genitalia is not indicative of your role in society, your actions and decisions are [ENDQ] [NEWLINE] You can partly thank feminist thought for popularizing this sentiment,<mask> I'm sure making dick jokes has been a common human endeavor<mask> time immemorial. [NEWLINE] [NEWLINE] [STARTQ] feminism is a movement for women, not equality for all sexes [ENDQ] [NEWLINE] <mask> feminism is primarily concerned with the plight of women, it has resulted in theory and policies that benefit men, children, minorities, homosexuals, and transgenders alike. For example, the issue of the patriarchy (a much maligned term here on reddit) has influenced men<mask> the 70s to seek [male liberation]( [URL] %27s_liberation_movement) from traditional gender roles, much in the same way women sought to emancipate themselves from gender roles. (Men's liberation should not be confused with men's rights activism, an anti-feminist movement). [NEWLINE] [NEWLINE] [STARTQ] special interest group [ENDQ] [NEWLINE] Feminists are not a special interest group. There may be certain organizations that have a pro-feminist agenda,<mask> feminism<mask> a whole isn't an interest group.<mask> you have issues with certain organizations, then you should name them. For instance, gun-owners are not a special interest group,<mask> the National Rifle Association (NRA) is. [NEWLINE] [NEWLINE] <mask><mask> I'm on the NRA,<mask> don't we change that name too? After all, the NRA isn't just about teaching rifle marksmanship anymore. They are concerned with handguns, assault rifles, national background checks...</s>
Label encoding: <s> [STARTQ] I don't believe there is a better sex [ENDQ] [NEWLINE] Feminism doesn't either, but it has acknowledged a societal preference for one sex over another. [NEWLINE] [NEWLINE] [STARTQ] Your genitalia is not indicative of your role in society, your actions and decisions are [ENDQ] [NEWLINE] You can partly thank feminist thought for popularizing this sentiment, though I'm sure making dick jokes has been a common human endeavor since time immemorial. [NEWLINE] [NEWLINE] [STARTQ] feminism is a movement for women, not equality for all sexes [ENDQ] [NEWLINE] While feminism is primarily concerned with the plight of women, it has resulted in theory and policies that benefit men, children, minorities, homosexuals, and transgenders alike. For example, the issue of the patriarchy (a much maligned term here on reddit) has influenced men since the 70s to seek [male liberation]( [URL] %27s_liberation_movement) from traditional gender roles, much in the same way women sought to emancipate themselves from gender roles. (Men's liberation should not be confused with men's rights activism, an anti-feminist movement). [NEWLINE] [NEWLINE] [STARTQ] special interest group [ENDQ] [NEWLINE] Feminists are not a special interest group. There may be certain organizations that have a pro-feminist agenda, but feminism as a whole isn't an interest group. If you have issues with certain organizations, then you should name them. For instance, gun-owners are not a special interest group, but the National Rifle Association (NRA) is. [NEWLINE] [NEWLINE] So while I'm on the NRA, why don't we change that name too? After all, the NRA isn't just about teaching rifle marksmanship anymore. They are concerned with handguns, assault rifles, national background checks...</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They are always portrayed<mask> being right along side people that have actually done relevant things to contribute to society today. <mask><mask> that both culturally and musically they were not that big of a deal.  There are loads of better musicians and many more interesting people.  After all, they did start out<mask> an equivalent to today's boy bands; they were a figure head, something for girls to oogle at.  There are<mask> many conspiracies about The Beatles and I don't think there have been the same amount of cover bands for any other band in history (<mask> maybe The Dead,<mask> they were great musicians). [NEWLINE] [NEWLINE] I know this explanation is a bit scattered,<mask> it's late.  I mainly wanted to make this<mask> I wanted to get my point across.  I will further explain, in more lengthy detail, in comments. [NEWLINE] [NEWLINE] Change my view. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Label encoding: <s>They are always portrayed as being right along side people that have actually done relevant things to contribute to society today.  I think that both culturally and musically they were not that big of a deal.  There are loads of better musicians and many more interesting people.  After all, they did start out as an equivalent to today's boy bands; they were a figure head, something for girls to oogle at.  There are so many conspiracies about The Beatles and I don't think there have been the same amount of cover bands for any other band in history ( besides maybe The Dead, but they were great musicians). [NEWLINE] [NEWLINE] I know this explanation is a bit scattered, but it's late.  I mainly wanted to make this because I wanted to get my point across.  I will further explain, in more lengthy detail, in comments. [NEWLINE] [NEWLINE] Change my view. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My initial response, based on your description, is that I don't see<mask> having a future is enough to give something rights. A potential person is not the same<mask> an actual person.<mask> I haven't read the full article at this point. [ENDQ] [NEWLINE] The argument doesn't require that the thing in question is a person at all. The claim is that the morally relevant condition to have a right against being killed is that an organism has a future like ours. [NEWLINE] [NEWLINE] [STARTQ] <mask> is that different from arguing that a seed is a tree? [ENDQ] [NEWLINE] Well let's be careful here. I was posing the question to the person who claimed that, for instance, it is not permissible to kill a baby after birth<mask> that it was permissible to kill a fetus prior to birth. The question was,<mask> is the morally significant difference between the two. It is generally accepted that a fetus one day from being born and that half a be one day after being born have no significant moral difference.<mask> the regress can get started with no requirement to call the fetus a PERSON. The only requirement is that the fetus shares certain facts with people, and those facts confer the protection against being killed. [NEWLINE] [NEWLINE] And come to think of it, your example about the seed being a tree is only counterintuitive in a certain light. A seed having been planted in the ground does share something significant with a tree,<mask> does a sapling,<mask> does a sprout etc. [NEWLINE] [NEWLINE] <mask><mask> you are too caught up on the idea that anybody attacking abortion has the claim that the fetus is a full-blown person. They don't. All we have to maintain is that the fetus shares certain moral facts with people which provide the rights in question.</s>
Label encoding: <s> [STARTQ] My initial response, based on your description, is that I don't see how having a future is enough to give something rights. A potential person is not the same as an actual person. But I haven't read the full article at this point. [ENDQ] [NEWLINE] The argument doesn't require that the thing in question is a person at all. The claim is that the morally relevant condition to have a right against being killed is that an organism has a future like ours. [NEWLINE] [NEWLINE] [STARTQ] How is that different from arguing that a seed is a tree? [ENDQ] [NEWLINE] Well let's be careful here. I was posing the question to the person who claimed that, for instance, it is not permissible to kill a baby after birth but that it was permissible to kill a fetus prior to birth. The question was, what is the morally significant difference between the two. It is generally accepted that a fetus one day from being born and that half a be one day after being born have no significant moral difference. Thus the regress can get started with no requirement to call the fetus a PERSON. The only requirement is that the fetus shares certain facts with people, and those facts confer the protection against being killed. [NEWLINE] [NEWLINE] And come to think of it, your example about the seed being a tree is only counterintuitive in a certain light. A seed having been planted in the ground does share something significant with a tree, as does a sapling, as does a sprout etc. [NEWLINE] [NEWLINE] I think you are too caught up on the idea that anybody attacking abortion has the claim that the fetus is a full-blown person. They don't. All we have to maintain is that the fetus shares certain moral facts with people which provide the rights in question.</s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1) In that case, we'll be keeping some kids in school forever; some kids will never be able to get an A in every subject. [NEWLINE] [NEWLINE] Even kids who might do very well in later subjects won't be able to get an A in every subject.<mask> I was held in my freshman English class with the teacher who graded heavily on class participation<mask> I was super-shy, I would never have gotten to the next level, which I got an easy A in.<mask> my senior year English teacher who assigned books I hated and<mask> barely skimmed was at any other level I would never have been able to get past her class.<mask> even those, I passed with Bs. [NEWLINE] [NEWLINE] 2) The point is not to get everyone high school degrees,<mask> a degree is a marker of some level of knowledge.<mask> the system we have now *already guarantees that level of knowledge*. Anyone who gets a degree *already has the knowledge signified by it*.<mask> they didn't they would have actually FAILED their classes. [NEWLINE] [NEWLINE] A high school degree isn't intended to mean you have the knowledge required of a straight-A student; it's intended to mean you have the knowledge required of a straight-D student. Just<mask> it's not the highest possible bar doesn't mean it's worthless. [NEWLINE] [NEWLINE] 3)<mask> the curriculum is "properly" designed and the class is "properly" taught then there should be nobody who doesn't know everything in the class. The problem is this is<mask> difficult to do it's not seriously possible; you would have to structure the entire class to best teach every individual student. You'd basically have to have one student per teacher.</s>
Label encoding: <s>1) In that case, we'll be keeping some kids in school forever; some kids will never be able to get an A in every subject. [NEWLINE] [NEWLINE] Even kids who might do very well in later subjects won't be able to get an A in every subject. If I was held in my freshman English class with the teacher who graded heavily on class participation when I was super-shy, I would never have gotten to the next level, which I got an easy A in. If my senior year English teacher who assigned books I hated and therefore barely skimmed was at any other level I would never have been able to get past her class. But even those, I passed with Bs. [NEWLINE] [NEWLINE] 2) The point is not to get everyone high school degrees, because a degree is a marker of some level of knowledge. But the system we have now *already guarantees that level of knowledge*. Anyone who gets a degree *already has the knowledge signified by it*. If they didn't they would have actually FAILED their classes. [NEWLINE] [NEWLINE] A high school degree isn't intended to mean you have the knowledge required of a straight-A student; it's intended to mean you have the knowledge required of a straight-D student. Just because it's not the highest possible bar doesn't mean it's worthless. [NEWLINE] [NEWLINE] 3) If the curriculum is "properly" designed and the class is "properly" taught then there should be nobody who doesn't know everything in the class. The problem is this is so difficult to do it's not seriously possible; you would have to structure the entire class to best teach every individual student. You'd basically have to have one student per teacher.</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The damage to the building does not support an aircraft hitting it. The damage to the building - neat holes punch in 4 sequential brick walls - supports a projectile smaller than an airliner, and one with a really hard nose on it. The nose on a passenger jet is a soft cover, a cowling that covers the radar mounted on the nose of the plane. [NEWLINE] [NEWLINE] Eyewitnesses are famously inaccurate. I do not know<mask> happened,<mask> I can see the damage pattern of neatly punch holes deep into the structure, holes punched by a single projectile with a really hard nose in it, not debris. Are you aware of any of this? Something tells me you have not looked very far into it, evaluating the structural damage to the site. Oh and the part about the plane in question has engines out on the wings and there is no building punch-through on the sides of the main entrance hole<mask> the projectile hit, even<mask> you wanted to believe there were big airliners wings hitting the building. [NEWLINE] [NEWLINE] Low information? You're damned right<mask> you are talking about "eyewitnesses" and incapable to evaluate the damage site. [NEWLINE] [NEWLINE] Are you aware there is US GOV budget being paid to disinformation persons to go online and comment in discussion groups and "act like the people"<mask> placing and supporting planned information? It may sound hard to believe,<mask> it is right there out in the open. The military is not supposed to do disinformation and psyops on the people in the home country,<mask> they have budget for it. You ought to look up the term "psyop" official term for "psychological operation" to get an idea of this activity. It is an official term, not some screwy fiction.</s>
Label encoding: <s>The damage to the building does not support an aircraft hitting it. The damage to the building - neat holes punch in 4 sequential brick walls - supports a projectile smaller than an airliner, and one with a really hard nose on it. The nose on a passenger jet is a soft cover, a cowling that covers the radar mounted on the nose of the plane. [NEWLINE] [NEWLINE] Eyewitnesses are famously inaccurate. I do not know what happened, but I can see the damage pattern of neatly punch holes deep into the structure, holes punched by a single projectile with a really hard nose in it, not debris. Are you aware of any of this? Something tells me you have not looked very far into it, evaluating the structural damage to the site. Oh and the part about the plane in question has engines out on the wings and there is no building punch-through on the sides of the main entrance hole where the projectile hit, even if you wanted to believe there were big airliners wings hitting the building. [NEWLINE] [NEWLINE] Low information? You're damned right when you are talking about "eyewitnesses" and incapable to evaluate the damage site. [NEWLINE] [NEWLINE] Are you aware there is US GOV budget being paid to disinformation persons to go online and comment in discussion groups and "act like the people" while placing and supporting planned information? It may sound hard to believe, but it is right there out in the open. The military is not supposed to do disinformation and psyops on the people in the home country, but they have budget for it. You ought to look up the term "psyop" official term for "psychological operation" to get an idea of this activity. It is an official term, not some screwy fiction.</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I work at a biotech, and<mask> I can't speak for larger pharma's I can most certainly say this is not the case with all.<mask> you consider the profit motive a reason to not cure<mask> to manage symptoms, you ignore the fact that the competition is actually relatively stiff given the cost of innovation.<mask>, personalized medicine is becoming a bigger and bigger deal in oncology, and a drug company that has a good drug can make far more money selling diagnostics than it would selling treatments that ultimately result in dead patients. There is far more incentive to cure, than to treat,<mask> standards of care shift constantly. [NEWLINE] [NEWLINE] [NEWLINE] The FDA<mask><mask><mask><mask> has a monopoly on drug approval in the US, and the US is<mask> drug companies profit and<mask> create the market for research (single payer healthcares are less profitable and a drag on research spending. You can thank Europe and Canada for thottling drug research by roughly 50%).<mask> diseases start getting cured, the demand for drug approvals drops, and the FDA has to shrink, this is worth noting particularly after the questionable rejection of Aveo's renal cancer therapy which would upset IND's in that area. [NEWLINE] [NEWLINE] Corruption in the FDA runs deep, its a vicious cycle of large pharma shareholders regulating their own self interests. We really need to abolish the FDA, it holds back good drugs<mask> failing to stop bad drugs, and kills quite a large amount of people<mask><mask><mask>. [NEWLINE] [NEWLINE] Cancer is a particularly interesting case<mask> the layman feels more progress should be made,<mask> once you understand the complexities of it, its actually quite amazing the progress we've been making<mask> the oppressive costs to develop and gain approval for therapies. [NEWLINE] </s>
Label encoding: <s>I work at a biotech, and while I can't speak for larger pharma's I can most certainly say this is not the case with all. If you consider the profit motive a reason to not cure but to manage symptoms, you ignore the fact that the competition is actually relatively stiff given the cost of innovation. Additionally, personalized medicine is becoming a bigger and bigger deal in oncology, and a drug company that has a good drug can make far more money selling diagnostics than it would selling treatments that ultimately result in dead patients. There is far more incentive to cure, than to treat, as standards of care shift constantly. [NEWLINE] [NEWLINE] [NEWLINE] The FDA on the other hand has a monopoly on drug approval in the US, and the US is where drug companies profit and therefore create the market for research (single payer healthcares are less profitable and a drag on research spending. You can thank Europe and Canada for thottling drug research by roughly 50%). If diseases start getting cured, the demand for drug approvals drops, and the FDA has to shrink, this is worth noting particularly after the questionable rejection of Aveo's renal cancer therapy which would upset IND's in that area. [NEWLINE] [NEWLINE] Corruption in the FDA runs deep, its a vicious cycle of large pharma shareholders regulating their own self interests. We really need to abolish the FDA, it holds back good drugs while failing to stop bad drugs, and kills quite a large amount of people as a result. [NEWLINE] [NEWLINE] Cancer is a particularly interesting case where the layman feels more progress should be made, but once you understand the complexities of it, its actually quite amazing the progress we've been making despite the oppressive costs to develop and gain approval for therapies. [NEWLINE] </s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> are you? [ENDQ] [NEWLINE] It's really not<mask> bad<mask> it seems,<mask> that may be<mask> I skip most of my classes.<mask> I actually attended and paid full attention to all my class lectures, I would burn out extremely quickly.<mask>, I realized that I can learn much more efficiently by myself and I either just attend class occasionally to make sure I'm on track or just for the exams.<mask> a professor spends 1+ hour lecturing, I can learn in about 20-30min, sometimes quicker depending on the topic. Then it's just a matter of practicing or memorizing (whether it is a math-based or information-based class) [NEWLINE] [NEWLINE] [STARTQ] Like take that many credits in a semester?<mask> no. I mean<mask> you could who am I to say you can't<mask>, uhh, no. [ENDQ] [NEWLINE] My adviser is about to retire<mask> he doesn't really care about much anymore.<mask><mask> I ever want anything, like taking 27 or 45 credits, he just signs me off. [NEWLINE] [NEWLINE] A few days ago, I actually planned out my 45 credit courses and the times I will take them at, and I have considerable overlap. There are multiple cases in which I have two classes at the same time on the same day (again my adviser overrides this for me). Of course, I check with the professors beforehand to make sure attendance isn't required for both classes and that exams don't occur on the same day by simply asking for the syllabus. Or<mask> there is simply an exam conflict, I'll probably end up getting written permission from one of the professors to let me take the exam at a time different from the syllabus before the add/drop period ends </s>
Label encoding: <s> [STARTQ] What are you? [ENDQ] [NEWLINE] It's really not as bad as it seems, but that may be because I skip most of my classes. If I actually attended and paid full attention to all my class lectures, I would burn out extremely quickly. However, I realized that I can learn much more efficiently by myself and I either just attend class occasionally to make sure I'm on track or just for the exams. What a professor spends 1+ hour lecturing, I can learn in about 20-30min, sometimes quicker depending on the topic. Then it's just a matter of practicing or memorizing (whether it is a math-based or information-based class) [NEWLINE] [NEWLINE] [STARTQ] Like take that many credits in a semester? Because no. I mean if you could who am I to say you can't but, uhh, no. [ENDQ] [NEWLINE] My adviser is about to retire so he doesn't really care about much anymore. Thus if I ever want anything, like taking 27 or 45 credits, he just signs me off. [NEWLINE] [NEWLINE] A few days ago, I actually planned out my 45 credit courses and the times I will take them at, and I have considerable overlap. There are multiple cases in which I have two classes at the same time on the same day (again my adviser overrides this for me). Of course, I check with the professors beforehand to make sure attendance isn't required for both classes and that exams don't occur on the same day by simply asking for the syllabus. Or if there is simply an exam conflict, I'll probably end up getting written permission from one of the professors to let me take the exam at a time different from the syllabus before the add/drop period ends </s>
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Worse, these arguments require, by necessity, the outlook of a societal parasite. Rather than working to change things or make their own situation better, these people want to shame others for<mask> they themselves don't have.<mask><mask> screeching shrilly at someone that they should feel bad for being who they are, is somehow a morally superior stance. [ENDQ] [NEWLINE] The people who use these terms tend to be those concerned with social justice, i.e. "working to change things".<mask>, there's that. [NEWLINE] [NEWLINE] <mask>, the intent is not to shame anyone for being who they are,<mask> rather to draw attention to the problems faced by those who are systematically disadvantaged for who *they* are.<mask> a white person, my awareness of white privilege does not make me feel ashamed. It's not an individual guilt thing. It's more like "It's not fair that not everyone gets treated like this.<mask> can we do to change that, and<mask> can I make sure I'm not part of the problem." [NEWLINE] [NEWLINE] For example,<mask> I'm on the side of the road with a flat tire, at least one person will stop and try to help. Guaranteed. Happens every time. At the same time, I'm aware that<mask> I were black, this is much less likely the happen. This doesn't make me feel guilty about accepting help. I'm not gonna turn down help<mask> the person offering would probably not offer it to a black person. Instead, the next time I see a black person with a flat tire on the side of the road, I should ignore the white-person impulse to think "Oh god<mask><mask> I get robbed" and pull over to help them. </s>
Label encoding: <s> [STARTQ] Worse, these arguments require, by necessity, the outlook of a societal parasite. Rather than working to change things or make their own situation better, these people want to shame others for what they themselves don't have. As if screeching shrilly at someone that they should feel bad for being who they are, is somehow a morally superior stance. [ENDQ] [NEWLINE] The people who use these terms tend to be those concerned with social justice, i.e. "working to change things". So, there's that. [NEWLINE] [NEWLINE] Also, the intent is not to shame anyone for being who they are, but rather to draw attention to the problems faced by those who are systematically disadvantaged for who *they* are. As a white person, my awareness of white privilege does not make me feel ashamed. It's not an individual guilt thing. It's more like "It's not fair that not everyone gets treated like this. What can we do to change that, and how can I make sure I'm not part of the problem." [NEWLINE] [NEWLINE] For example, if I'm on the side of the road with a flat tire, at least one person will stop and try to help. Guaranteed. Happens every time. At the same time, I'm aware that if I were black, this is much less likely the happen. This doesn't make me feel guilty about accepting help. I'm not gonna turn down help because the person offering would probably not offer it to a black person. Instead, the next time I see a black person with a flat tire on the side of the road, I should ignore the white-person impulse to think "Oh god what if I get robbed" and pull over to help them. </s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This comes from a person who has seen family members kill themselves, and try to kill themselves. I was<mask> clinically depressed<mask> a teen due to a medical diagnosis. From<mask> I can tell of the issue; suicide is a decision a person makes<mask> they give up. [NEWLINE] [NEWLINE] I realize that is a HUGE oversimplification of a very very complicated issue,<mask> let me clarify my point. Suicide is related to mental illness yes,<mask> I understand it is a choice. Mentally Ill people have chemical imbalances in their brain,<mask> I don't think that makes them incapable of free will. They still actively chose to kill themselves in a specific way or fashion with all factors considered. [NEWLINE] [NEWLINE] A way I see it is; a drunk person is still liable for any crimes they did<mask> drunk,<mask><mask> there is an imbalance of chemicals in their brain. (<mask> I am unsure<mask> that is<mask> a person chooses to get inebriated,<mask> a mentally Ill person is born with it) [NEWLINE] [NEWLINE] <mask> they have chosen to kill themselves,<mask> don't they choose to actively improve their situation? Call me an optimist<mask> I sincerely believe that<mask> a person tries with the best of their ability, they can improve<mask> they live. Now a mentally ill person may not think like that at all.<mask> that doesn't change that they chose to die over choosing to strive for a better life. [NEWLINE] [NEWLINE] Suicide is weakness in my mind,<mask> it is a choice. And<mask> you have a choice between turning everything off, or 'beating the game', and you consciously choose to die, you are a quitter and that is weak. [NEWLINE] [NEWLINE] Change my view? [NEWLINE] [NEWLINE] -Edited for grammar</s>
Label encoding: <s>This comes from a person who has seen family members kill themselves, and try to kill themselves. I was also clinically depressed as a teen due to a medical diagnosis. From what I can tell of the issue; suicide is a decision a person makes when they give up. [NEWLINE] [NEWLINE] I realize that is a HUGE oversimplification of a very very complicated issue, but let me clarify my point. Suicide is related to mental illness yes, but I understand it is a choice. Mentally Ill people have chemical imbalances in their brain, but I don't think that makes them incapable of free will. They still actively chose to kill themselves in a specific way or fashion with all factors considered. [NEWLINE] [NEWLINE] A way I see it is; a drunk person is still liable for any crimes they did while drunk, even though there is an imbalance of chemicals in their brain. ( Although I am unsure if that is because a person chooses to get inebriated, while a mentally Ill person is born with it) [NEWLINE] [NEWLINE] Since they have chosen to kill themselves, why don't they choose to actively improve their situation? Call me an optimist but I sincerely believe that if a person tries with the best of their ability, they can improve how they live. Now a mentally ill person may not think like that at all. But that doesn't change that they chose to die over choosing to strive for a better life. [NEWLINE] [NEWLINE] Suicide is weakness in my mind, because it is a choice. And when you have a choice between turning everything off, or 'beating the game', and you consciously choose to die, you are a quitter and that is weak. [NEWLINE] [NEWLINE] Change my view? [NEWLINE] [NEWLINE] -Edited for grammar</s>
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are some words that people use to sound smart that are actually very useful words. "Plethora" is the word I like to use<mask> an example; we have a lot of words that have the same *basic* meaning of "plethora" (basically: a lot of, an excess of)<mask> a plethora is specific in the way that it is having<mask> much of something that it goes from being a good thing to a bad thing; It's the difference between, for instance, "Jamie ate a plethora of salmon and was full until his next meal" and "Jamie ate a plethora of salmon and died of mercury poison." [NEWLINE] [NEWLINE] Now, that's not the best example,<mask> it is one of the words that people love to use wrong, and it's all<mask> lost that important "<mask> much that it's bad for you" meaning and just became a synonym for "a fuckton of",<mask> there are good words that are similar in their specificity that go well<mask> trying to convey an idea. Hell, I bet that some of the words I used in that last paragraph would confuse younger readers. [NEWLINE] [NEWLINE] There is a time and a place for every word in your vocabulary, and just<mask> I wouldn't write "Fuck the goddamn cunt" in a children's book aimed at 5 year olds, nor either would I write "The square of the hypotenuse of a right triangle is equal to the sum of the square of its other two sides." [NEWLINE] [NEWLINE] <mask> on that note, I'm wondering<mask>, exactly, are you dealing with this sort of language that's inspiring this view? Are there any examples you have other than "preposterous" that irk you<mask>?</s>
Label encoding: <s>There are some words that people use to sound smart that are actually very useful words. "Plethora" is the word I like to use as an example; we have a lot of words that have the same *basic* meaning of "plethora" (basically: a lot of, an excess of) but a plethora is specific in the way that it is having so much of something that it goes from being a good thing to a bad thing; It's the difference between, for instance, "Jamie ate a plethora of salmon and was full until his next meal" and "Jamie ate a plethora of salmon and died of mercury poison." [NEWLINE] [NEWLINE] Now, that's not the best example, because it is one of the words that people love to use wrong, and it's all but lost that important " so much that it's bad for you" meaning and just became a synonym for "a fuckton of", but there are good words that are similar in their specificity that go well when trying to convey an idea. Hell, I bet that some of the words I used in that last paragraph would confuse younger readers. [NEWLINE] [NEWLINE] There is a time and a place for every word in your vocabulary, and just as I wouldn't write "Fuck the goddamn cunt" in a children's book aimed at 5 year olds, nor either would I write "The square of the hypotenuse of a right triangle is equal to the sum of the square of its other two sides." [NEWLINE] [NEWLINE] So on that note, I'm wondering where, exactly, are you dealing with this sort of language that's inspiring this view? Are there any examples you have other than "preposterous" that irk you so?</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see<mask> you are saying,<mask> my issues is this: [NEWLINE] [NEWLINE] <mask> happens<mask> the pet owner / parent / God is abusive, bad or evil? [NEWLINE] [NEWLINE] Does a dog that is not getting beat, and tormented by its owner have the right to bite the hand that feeds it? <mask> a child is being sexually abused, beaten, and scared does that child have the right to run away? [NEWLINE] [NEWLINE] You make the assumption that God is good and that whatever he does is<mask> is good for you.  Many people look at God's actions and say that he is more the abuser than the benevolent parent. [NEWLINE] [NEWLINE] My personal view is that God is literally worse than Hitler. (Not invoking Godwin's law or being figurative here either) <mask> ever Hitler did (and he did many things wrong) he only did finite punishment<mask> God (<mask> you believe in Hell) punishes people forever.  Under God's law you are required to accept substitutional punishment (Jesus dying for our sins), human sacrifice (Jesus dying for our sins), inherited crimes (you are responsible for your parent's crimes/sins, original sin).  Any of these taken alone are horrible and evil by our standards today.  I would not worship nor praise someone who truly follows or believes in this stuff, and<mask> such I find that even<mask> God existed he would not be worthy or praise or worship. [NEWLINE] [NEWLINE] Could it be possible that all of this is in our own interest, yes I guess it could. <mask>,<mask> a parent is beating and molesting their child to make them tough and show them his love, we still see that<mask> an evil act and remove the child from their God.</s>
Label encoding: <s>I see what you are saying, but my issues is this: [NEWLINE] [NEWLINE] What happens when the pet owner / parent / God is abusive, bad or evil? [NEWLINE] [NEWLINE] Does a dog that is not getting beat, and tormented by its owner have the right to bite the hand that feeds it?  If a child is being sexually abused, beaten, and scared does that child have the right to run away? [NEWLINE] [NEWLINE] You make the assumption that God is good and that whatever he does is what is good for you.  Many people look at God's actions and say that he is more the abuser than the benevolent parent. [NEWLINE] [NEWLINE] My personal view is that God is literally worse than Hitler. (Not invoking Godwin's law or being figurative here either)  What ever Hitler did (and he did many things wrong) he only did finite punishment where God ( if you believe in Hell) punishes people forever.  Under God's law you are required to accept substitutional punishment (Jesus dying for our sins), human sacrifice (Jesus dying for our sins), inherited crimes (you are responsible for your parent's crimes/sins, original sin).  Any of these taken alone are horrible and evil by our standards today.  I would not worship nor praise someone who truly follows or believes in this stuff, and as such I find that even IF God existed he would not be worthy or praise or worship. [NEWLINE] [NEWLINE] Could it be possible that all of this is in our own interest, yes I guess it could.  However, if a parent is beating and molesting their child to make them tough and show them his love, we still see that as an evil act and remove the child from their God.</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You appear to be justifying your own evil by saying, implicitly, that the use of evil against evil privilege is not evil. That is, in effect, calling people with privilege evil. [ENDQ] [NEWLINE] I have about<mask> much privilege<mask> you can have and I don't think of myself<mask> evil. I don't think privilege makes people evil,<mask><mask> privilege has a toxic effect on the way we think about the world and it needs to be addressed responsibly by the people that benefit from it.<mask> of that, yes, people with privilege often do bad things unintentionally. That doesn't make them bad people. [NEWLINE] [NEWLINE] [STARTQ] The fact that you're, very civilly, trying to present<mask> one person is wrong and horrible for doing &lt;thing&gt;<mask> privileged,<mask> you are not wrong and horrible for doing &lt;exact same thing&gt;<mask> not privileged/in defense of those who are not privileged, is exactly the problem. You are using the concept of privilege<mask> an excuse to attack others in the exact same way that you feel they are attacking you/those you defend. [ENDQ] [NEWLINE] The thing I am doing is not the exact same<mask> it *is not coming from privilege*, whereas the assumption that these things are okay to say, or that someone is a better judge of some community's problems than they are, or etc. are all enabled by privilege. Again, the disparity in power is the key difference. And<mask> I've said before, many of the more vitriolic comments come from people who have dealt with slurs their entire life. They're fed up, and<mask> it may not be the best strategy are you going to tell them their anger and bitterness isn't justified?</s>
Label encoding: <s> [STARTQ] You appear to be justifying your own evil by saying, implicitly, that the use of evil against evil privilege is not evil. That is, in effect, calling people with privilege evil. [ENDQ] [NEWLINE] I have about as much privilege as you can have and I don't think of myself as evil. I don't think privilege makes people evil, I think privilege has a toxic effect on the way we think about the world and it needs to be addressed responsibly by the people that benefit from it. Because of that, yes, people with privilege often do bad things unintentionally. That doesn't make them bad people. [NEWLINE] [NEWLINE] [STARTQ] The fact that you're, very civilly, trying to present how one person is wrong and horrible for doing &lt;thing&gt; while privileged, but you are not wrong and horrible for doing &lt;exact same thing&gt; while not privileged/in defense of those who are not privileged, is exactly the problem. You are using the concept of privilege as an excuse to attack others in the exact same way that you feel they are attacking you/those you defend. [ENDQ] [NEWLINE] The thing I am doing is not the exact same because it *is not coming from privilege*, whereas the assumption that these things are okay to say, or that someone is a better judge of some community's problems than they are, or etc. are all enabled by privilege. Again, the disparity in power is the key difference. And as I've said before, many of the more vitriolic comments come from people who have dealt with slurs their entire life. They're fed up, and while it may not be the best strategy are you going to tell them their anger and bitterness isn't justified?</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm seeing a lot of heat against FIFA<mask> of the unequal pay distributed to US woman soccer players compared to men. (See this article for more info: [URL] ) [NEWLINE] [NEWLINE] Now,<mask><mask><mask> that pay gaps are a problem in ~~general~~ some fields, sports is a different ballgame (no pun intended).<mask> such, the success of a franchise is dependent on the supporters who pay money for tickets, gear, etc.<mask> we were to look at attendance figures for the National Woman's Soccer League (NWSL) for 2014 compared to Major League Soccer (MLS), we see a huge disparity in people attending games. [NEWLINE] [NEWLINE] Of course, this translates over to other leagues. The NBA draws more people than the WNBA, both live and TV ratings. The revenue of the NFL is exponentially greater than the Independent Woman's Football League. [NEWLINE] [NEWLINE] I'm not here to provide solutions to this, or look into<mask> the mentality of America draws more into guy sports. Yes,<mask><mask> that ladies should be paid equally. ~~I<mask> think ladies have the ability to play all sports<mask> good<mask> men do,<mask> not better.~~ And of course, ladies have<mask> much of a right to play in any sport and play it professionally.<mask> (and yes here comes a South Park reference) "...you can't expect people to watch." My conclusion: the pay gap in sports is a consequence of economic revenue, rather than social injustice. [NEWLINE] [NEWLINE] Open to any replies and criticism. [NEWLINE] [NEWLINE] EDIT: This is awesome and I thank you all for your input. You all taught me that there is a physical difference in athleticism between men and women. And that the pay gap is exaggerated. </s>
Label encoding: <s>I'm seeing a lot of heat against FIFA because of the unequal pay distributed to US woman soccer players compared to men. (See this article for more info: [URL] ) [NEWLINE] [NEWLINE] Now, while I agree that pay gaps are a problem in ~~general~~ some fields, sports is a different ballgame (no pun intended). As such, the success of a franchise is dependent on the supporters who pay money for tickets, gear, etc. If we were to look at attendance figures for the National Woman's Soccer League (NWSL) for 2014 compared to Major League Soccer (MLS), we see a huge disparity in people attending games. [NEWLINE] [NEWLINE] Of course, this translates over to other leagues. The NBA draws more people than the WNBA, both live and TV ratings. The revenue of the NFL is exponentially greater than the Independent Woman's Football League. [NEWLINE] [NEWLINE] I'm not here to provide solutions to this, or look into why the mentality of America draws more into guy sports. Yes, I think that ladies should be paid equally. ~~I also think ladies have the ability to play all sports as good as men do, if not better.~~ And of course, ladies have as much of a right to play in any sport and play it professionally. But (and yes here comes a South Park reference) "...you can't expect people to watch." My conclusion: the pay gap in sports is a consequence of economic revenue, rather than social injustice. [NEWLINE] [NEWLINE] Open to any replies and criticism. [NEWLINE] [NEWLINE] EDIT: This is awesome and I thank you all for your input. You all taught me that there is a physical difference in athleticism between men and women. And that the pay gap is exaggerated. </s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] European players with a goon mentality don't last in the sport long enough to make it to the NHL. [ENDQ] [NEWLINE] Not really - fighting just isn't<mask> important in the European leagues and<mask> fewer people aspire to be goons. The NHL has evolved and the true "goon" is not really good enough for a roster spot any more,<mask> even 25 years ago the best goons were still good enough to make a roster spot without their fists. Bob Probert was *surprisingly* good at hockey.<mask> was Domi. They would easily make it in a European league even without fighting. [NEWLINE] [NEWLINE] [STARTQ] <mask> this were true,<mask> do you account for women who compete in boxing, MMA, rugby, etc? [ENDQ] [NEWLINE] Those sports are based around physicality, not a game. [NEWLINE] [NEWLINE] The argument of "checking makes the men's sport awesome<mask> makes the women's version suck" doesn't make any sense. You already see contact along the boards, its just the open ice hits that are against the rules - and those are the most fun to watch. Guys like Stevens and Kronwall are fan favorites<mask> its entertaining to watch. [NEWLINE] [NEWLINE] <mask> yes - it does promote a more "skillful" play in the sense that speed is rewarded more,<mask> thats like saying the no-dunking aspect of the WNBA (I know its not a rule...) makes the game better by showing off layup skills.<mask> you remove a very entertaining aspect to the game you're not only shooting yourself in the foot,<mask> you're probably struggling to put butts in seats<mask> well. [NEWLINE] [NEWLINE] And last I saw, women's hockey was doing worse than the WBNA in attendance... </s>
Label encoding: <s> [STARTQ] European players with a goon mentality don't last in the sport long enough to make it to the NHL. [ENDQ] [NEWLINE] Not really - fighting just isn't as important in the European leagues and therefore fewer people aspire to be goons. The NHL has evolved and the true "goon" is not really good enough for a roster spot any more, but even 25 years ago the best goons were still good enough to make a roster spot without their fists. Bob Probert was *surprisingly* good at hockey. So was Domi. They would easily make it in a European league even without fighting. [NEWLINE] [NEWLINE] [STARTQ] If this were true, how do you account for women who compete in boxing, MMA, rugby, etc? [ENDQ] [NEWLINE] Those sports are based around physicality, not a game. [NEWLINE] [NEWLINE] The argument of "checking makes the men's sport awesome but makes the women's version suck" doesn't make any sense. You already see contact along the boards, its just the open ice hits that are against the rules - and those are the most fun to watch. Guys like Stevens and Kronwall are fan favorites because its entertaining to watch. [NEWLINE] [NEWLINE] So yes - it does promote a more "skillful" play in the sense that speed is rewarded more, but thats like saying the no-dunking aspect of the WNBA (I know its not a rule...) makes the game better by showing off layup skills. When you remove a very entertaining aspect to the game you're not only shooting yourself in the foot, but you're probably struggling to put butts in seats as well. [NEWLINE] [NEWLINE] And last I saw, women's hockey was doing worse than the WBNA in attendance... </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If everyone took your view point no one would be here. We would stop having sex. Curl up in the fetal position and wait for death to come. Try to only worry about the problems you can fix today. Pulling your scope back to always thinking "fucking universe, never done anything for me, just gonna die eventually, there's no point" denies the whole other side of this thing called life... [NEWLINE] [NEWLINE] <mask> that's not<mask> life is, that's not<mask> life got here, that's not<mask> life continues. [NEWLINE] [NEWLINE] Think about this:<mask> will it be like<mask> you die? Well, you've been there before, for a lot longer too...<mask><mask> was it like to wake up after having never gone to sleep? This is<mask> you were born. Everyone is made this way. You can't exist without someone else, not just physically<mask> your emotions/thoughts/feelings are all reflected by culture/communication with others. [NEWLINE] [NEWLINE] You are a social creature. We live in one of the most connected times our race has ever been. Chase those connections, curiosity is the best trait for any person to have. rather than worrying about<mask> happens to them in the void afterwards, find interest in the tangible. You may live a happier life and can maybe make an impact on others along the way that will make them happier. [NEWLINE] [NEWLINE] <mask> this boils down to is your worth will only be truly defined by yourself. No material objects will fill this void of existence you feel. Best thing I've seen people fill it with is people that will make you feel like you matter<mask> they bring out the sides you like about yourself. Hopefully you'll do the same for them.</s>
Label encoding: <s>If everyone took your view point no one would be here. We would stop having sex. Curl up in the fetal position and wait for death to come. Try to only worry about the problems you can fix today. Pulling your scope back to always thinking "fucking universe, never done anything for me, just gonna die eventually, there's no point" denies the whole other side of this thing called life... [NEWLINE] [NEWLINE] But that's not what life is, that's not how life got here, that's not how life continues. [NEWLINE] [NEWLINE] Think about this: what will it be like when you die? Well, you've been there before, for a lot longer too... Because what was it like to wake up after having never gone to sleep? This is when you were born. Everyone is made this way. You can't exist without someone else, not just physically but your emotions/thoughts/feelings are all reflected by culture/communication with others. [NEWLINE] [NEWLINE] You are a social creature. We live in one of the most connected times our race has ever been. Chase those connections, curiosity is the best trait for any person to have. rather than worrying about what happens to them in the void afterwards, find interest in the tangible. You may live a happier life and can maybe make an impact on others along the way that will make them happier. [NEWLINE] [NEWLINE] What this boils down to is your worth will only be truly defined by yourself. No material objects will fill this void of existence you feel. Best thing I've seen people fill it with is people that will make you feel like you matter because they bring out the sides you like about yourself. Hopefully you'll do the same for them.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, you can't compare NHL and soccer, infinite subs vs. only 3 subs is a huge difference<mask> talking about different ways to settle the game. [NEWLINE] [NEWLINE] Before it goes to PKs, both teams have played 120 minutes with 8 players likely (could be red cards<mask> I'll assume none) playing all 120 minutes.  They are dog tired, and both teams just had 120 minutes of play with a chance to win the game.  You shouldn't be too concerned about the switch to the mini game<mask> each team literally had two hours in which they could have played differently and won the game. [NEWLINE] [NEWLINE] <mask> you watch player interviews afterwards with star players they all say the same thing.  "<mask> does it feel to lose to PKs?" "We had several chances to put this one away.  Frank hit the crossbar early in the second half, I drove one from distance I needed to curl better. [NEWLINE] [NEWLINE] Your other suggestions of separate play periods or rescheduling games are unrealistic<mask> it creates a terrible fan experience.  Hold on fans, we will be back in an hour just sit around in your extra uncomfortable seats in your suit (some countries people dress up real nice to go to matches). [NEWLINE] [NEWLINE] At the end of the match, both teams played almost identical football,<mask> one team was just that little bit better, and that was decided from the spot. [NEWLINE] [NEWLINE] Most importantly remember PKs are not a game of chance, and they involve at least 6 players from each team.  Those players and goalkeepers have watched a lot of footage coming into this game for<mask> their opponent handles PKs.  Some of them have even practiced against a specific expected opponent.</s>
Label encoding: <s>First of all, you can't compare NHL and soccer, infinite subs vs. only 3 subs is a huge difference when talking about different ways to settle the game. [NEWLINE] [NEWLINE] Before it goes to PKs, both teams have played 120 minutes with 8 players likely (could be red cards but I'll assume none) playing all 120 minutes.  They are dog tired, and both teams just had 120 minutes of play with a chance to win the game.  You shouldn't be too concerned about the switch to the mini game because each team literally had two hours in which they could have played differently and won the game. [NEWLINE] [NEWLINE] When you watch player interviews afterwards with star players they all say the same thing.  " How does it feel to lose to PKs?" "We had several chances to put this one away.  Frank hit the crossbar early in the second half, I drove one from distance I needed to curl better. [NEWLINE] [NEWLINE] Your other suggestions of separate play periods or rescheduling games are unrealistic because it creates a terrible fan experience.  Hold on fans, we will be back in an hour just sit around in your extra uncomfortable seats in your suit (some countries people dress up real nice to go to matches). [NEWLINE] [NEWLINE] At the end of the match, both teams played almost identical football, but one team was just that little bit better, and that was decided from the spot. [NEWLINE] [NEWLINE] Most importantly remember PKs are not a game of chance, and they involve at least 6 players from each team.  Those players and goalkeepers have watched a lot of footage coming into this game for how their opponent handles PKs.  Some of them have even practiced against a specific expected opponent.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I use the same reasoning<mask> your first point against suicide.<mask> your 2nd point,<mask> on face value, looks interesting doesn't explain the true picture. First, these geniuses came up with interesting ideas<mask> the pleasure of finding these outweighed any other despair they have had. Happiness is subjective. Recognition may be a nice perk to have,<mask> is not the sole motivation for Boltzman or Mendel in their work. Take the case of mathematicians like Grothendick, these guys don't give a damn whether the world takes them seriously. These guys are like isolated painters/artists, who derive pleasure in their work.<mask>, undermining the importance of happiness doesn't make sense. No good can come out of anyone who is unhappy. Show me one person who is genuinely unhappy, who will do something useful that has the potential to empower humanity. [NEWLINE] [NEWLINE] One reason you could use to your advantage is that, most unhappiness is transitory;<mask> you feel, you are not being recognized now and<mask> to end your life, you are just being stupid. For example,<mask> you are an A student and you somehow fail in an exam, you may feel you have failed the entire world and you may even desire to suicide.<mask> in the grand scheme of things, you won't even remember that exam,<mask> your work changes the world.<mask>, the only reason to undermine unhappiness is that you may(and most likely are) be overestimating it.<mask> it is extremely important, in order to be, a productive member of the society, to find something that is pleasurable to you. Only some form of individual pleasure produces something that makes this world better for humanity.</s>
Label encoding: <s>I use the same reasoning as your first point against suicide. But your 2nd point, though on face value, looks interesting doesn't explain the true picture. First, these geniuses came up with interesting ideas because the pleasure of finding these outweighed any other despair they have had. Happiness is subjective. Recognition may be a nice perk to have, but is not the sole motivation for Boltzman or Mendel in their work. Take the case of mathematicians like Grothendick, these guys don't give a damn whether the world takes them seriously. These guys are like isolated painters/artists, who derive pleasure in their work. So, undermining the importance of happiness doesn't make sense. No good can come out of anyone who is unhappy. Show me one person who is genuinely unhappy, who will do something useful that has the potential to empower humanity. [NEWLINE] [NEWLINE] One reason you could use to your advantage is that, most unhappiness is transitory; if you feel, you are not being recognized now and therefore to end your life, you are just being stupid. For example, if you are an A student and you somehow fail in an exam, you may feel you have failed the entire world and you may even desire to suicide. But in the grand scheme of things, you won't even remember that exam, when your work changes the world. So, the only reason to undermine unhappiness is that you may(and most likely are) be overestimating it. But it is extremely important, in order to be, a productive member of the society, to find something that is pleasurable to you. Only some form of individual pleasure produces something that makes this world better for humanity.</s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've been an atheist<mask> birth,<mask> I unsubbed from /r/atheism after about a week of being subscribed. Not<mask> I didn't like the content or attitude,<mask><mask> it wasn't a sub that I needed to be in. I'm from The Netherlands, and grew up without any antipathy towards the religious. Here the religious and non-religious live side-by-side in relative harmony. Whether or not you're religious just doesn't really matter to 99% of the people and/or institutions, politicians don't ever really talk about religion, it doesn't really ever cause any problems in schools; it just doesn't come up. I've never believed in a God or taken any holy scriptures<mask> anything more than stories,<mask> I hadn't even ever heard the word "atheist" until I started watching American television in my late teens.<mask> I started looking through /r/atheism I instantly recognized it<mask> a sub for<mask> I call (meaning no offense or disrespect) "American atheists". People living in places<mask> being an atheist means social exclusion, problems in love, possibly even discrimination in the field of employment and more. I can understand their frustration, considering<mask> it means to be an atheist<mask> they live is considerably different to<mask> it means to be one<mask> I live, and I share their disdain at the effects institutionalized religion has had and continues to have on the world,<mask> I simply cannot relate to their vitriol towards their local God-fearing folk,<mask> I found no pleasure in that sub. [NEWLINE] [NEWLINE] EDIT:<mask> I'm here I may<mask> well plug /r/Humanism<mask> the significantly preferrable religion-free sub.</s>
Label encoding: <s>I've been an atheist since birth, but I unsubbed from /r/atheism after about a week of being subscribed. Not because I didn't like the content or attitude, but because it wasn't a sub that I needed to be in. I'm from The Netherlands, and grew up without any antipathy towards the religious. Here the religious and non-religious live side-by-side in relative harmony. Whether or not you're religious just doesn't really matter to 99% of the people and/or institutions, politicians don't ever really talk about religion, it doesn't really ever cause any problems in schools; it just doesn't come up. I've never believed in a God or taken any holy scriptures as anything more than stories, but I hadn't even ever heard the word "atheist" until I started watching American television in my late teens. When I started looking through /r/atheism I instantly recognized it as a sub for what I call (meaning no offense or disrespect) "American atheists". People living in places where being an atheist means social exclusion, problems in love, possibly even discrimination in the field of employment and more. I can understand their frustration, considering what it means to be an atheist where they live is considerably different to what it means to be one where I live, and I share their disdain at the effects institutionalized religion has had and continues to have on the world, but I simply cannot relate to their vitriol towards their local God-fearing folk, so I found no pleasure in that sub. [NEWLINE] [NEWLINE] EDIT: While I'm here I may as well plug /r/Humanism as the significantly preferrable religion-free sub.</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In the interest of full disclosure, I am Christian,<mask> not your traditional one. That being said, this has nothing to do with my stance. [NEWLINE] [NEWLINE] My reasoning is simple: [NEWLINE] Throughout the western world, the Bible and Christianity have been far and away the most significant influences in culture, literature, art, philosophy, law, etc. That being said,<mask> can someone in the western world considered themselves educated without a basic familiarity and understanding of the Bible, its stories, and philosophy? It has nothing to do with teaching religion,<mask> examining the bible<mask> a piece of literature and philosophy. Such a class should be required of all students,<mask> it is their responsibility<mask> citizens that get in the voting booth to possess a rudimentary understanding of culture, philosophy, etc. [NEWLINE] [NEWLINE] Should other religious texts be taught, or atheism? Sure,<mask> only<mask> electives. For example the Koran,<mask> increasingly relevant, has not had nearly<mask> much influence<mask> the Bible and is simply not<mask> important to understanding the western world. Should I live in Saudi Arabia, the Koran should be mandatory and the Bible and elective. It's a simple matter deepening your understanding of the society you live in. [NEWLINE] [NEWLINE] Would this violate a separation of church and state? No,<mask> it's not an endorsement of any religion. It's a simple acknowledgement of the text's importance in western society. The point is not to teach a religion<mask> right or wrong,<mask> to examine it the same you would examine any other religion from an anthropological, historical, and philosophical perspective. [NEWLINE] [NEWLINE] EDIT: Deltas awarded to Hmkay and pporkpiehat. Both made very good responses<mask> give them a read.</s>
Label encoding: <s>In the interest of full disclosure, I am Christian, although not your traditional one. That being said, this has nothing to do with my stance. [NEWLINE] [NEWLINE] My reasoning is simple: [NEWLINE] Throughout the western world, the Bible and Christianity have been far and away the most significant influences in culture, literature, art, philosophy, law, etc. That being said, how can someone in the western world considered themselves educated without a basic familiarity and understanding of the Bible, its stories, and philosophy? It has nothing to do with teaching religion, but examining the bible as a piece of literature and philosophy. Such a class should be required of all students, as it is their responsibility as citizens that get in the voting booth to possess a rudimentary understanding of culture, philosophy, etc. [NEWLINE] [NEWLINE] Should other religious texts be taught, or atheism? Sure, but only as electives. For example the Koran, while increasingly relevant, has not had nearly as much influence as the Bible and is simply not as important to understanding the western world. Should I live in Saudi Arabia, the Koran should be mandatory and the Bible and elective. It's a simple matter deepening your understanding of the society you live in. [NEWLINE] [NEWLINE] Would this violate a separation of church and state? No, because it's not an endorsement of any religion. It's a simple acknowledgement of the text's importance in western society. The point is not to teach a religion as right or wrong, but to examine it the same you would examine any other religion from an anthropological, historical, and philosophical perspective. [NEWLINE] [NEWLINE] EDIT: Deltas awarded to Hmkay and pporkpiehat. Both made very good responses so give them a read.</s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Who cares? These measurements are not important. Who's walking toe-to-heel to measure out a mile? You use a map for that stuff, or a gps, or a car odometer (which is very realistic<mask> you're actually travelling that distance, people actually do have those with them alot) [ENDQ] [NEWLINE] Your argument was that metric is too arbitrary and imperial is based on readily available body parts or items,<mask> only a small number of imperial units actually are and the rest are just<mask> arbitrary. [NEWLINE] [NEWLINE] [STARTQ] <mask>, people are more likely to use cups<mask> measuring liquids and powders. [ENDQ] [NEWLINE] Whose cup? Your cup or my cup? Cup sizes vary widely, and<mask> you have a standard size imperial cup that's no less arbitrary than a standard metric measuring cup. [NEWLINE] [NEWLINE] [STARTQ] Except that it does, due to plate tectonics and earths shifting magnetic pole. [ENDQ] [NEWLINE] <mask><mask><mask> the magnetic pole has nothing to do with the geographic north pole, and that's<mask><mask> I said "the metre could be recreated to a high precision at any time." rather than "recreated exactly". The geographical distance involved changes less than a single standard bar contracts and expands due to temperature, which used to be the definition of the yard, foot and inch. [NEWLINE] [NEWLINE] [STARTQ] &gt;<mask> it's pretty telling that the current official definition of the inch is now 2.54cm. [ENDQ] [NEWLINE] [STARTQ] Or, you know, you could say that a centimeter is 0.393701 inches. It works either way. [ENDQ] [NEWLINE] Except it doesn't<mask> then the metre would be based on a single arbitrary lump of metal rather than a physical constant.</s>
Label encoding: <s> [STARTQ] Who cares? These measurements are not important. Who's walking toe-to-heel to measure out a mile? You use a map for that stuff, or a gps, or a car odometer (which is very realistic if you're actually travelling that distance, people actually do have those with them alot) [ENDQ] [NEWLINE] Your argument was that metric is too arbitrary and imperial is based on readily available body parts or items, but only a small number of imperial units actually are and the rest are just as arbitrary. [NEWLINE] [NEWLINE] [STARTQ] Also, people are more likely to use cups when measuring liquids and powders. [ENDQ] [NEWLINE] Whose cup? Your cup or my cup? Cup sizes vary widely, and if you have a standard size imperial cup that's no less arbitrary than a standard metric measuring cup. [NEWLINE] [NEWLINE] [STARTQ] Except that it does, due to plate tectonics and earths shifting magnetic pole. [ENDQ] [NEWLINE] First of all the magnetic pole has nothing to do with the geographic north pole, and that's also why I said "the metre could be recreated to a high precision at any time." rather than "recreated exactly". The geographical distance involved changes less than a single standard bar contracts and expands due to temperature, which used to be the definition of the yard, foot and inch. [NEWLINE] [NEWLINE] [STARTQ] &gt; Also it's pretty telling that the current official definition of the inch is now 2.54cm. [ENDQ] [NEWLINE] [STARTQ] Or, you know, you could say that a centimeter is 0.393701 inches. It works either way. [ENDQ] [NEWLINE] Except it doesn't because then the metre would be based on a single arbitrary lump of metal rather than a physical constant.</s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do you think that a student that received an Unsatisfactory mark should be allowed to proceed to the next level? [ENDQ] [NEWLINE] Earlier today I would have said no,<mask> [after reading this]( [URL].aspx?nfstatus=401&amp;nftoken=00000000-0000-0000-0000-000000000000&amp;nfstatusdescription=ERROR%3a+No+local+token) I've changed my mind in most cases. I don't think a student should just continue on the path with no intervention,<mask><mask><mask> they should be given extra help in the form of a tutor, or individual attention with a teacher or aide or parent. This is more or less feasible<mask> you have one or two "unsatisfactory" grades in a class of 30 than having 20 people get a grade lower than A. [NEWLINE] [NEWLINE] Typing this did remind me of the policy of my college language class - A C- was a passing grade, technically, and allowed you to move on.<mask>,<mask> you got more than one test grade of B- or below you had to meet with a tutor once a week until you demonstrated your proficiency. Would this be a possible solution to you?<mask> it seems pretty logical to me. [NEWLINE] [NEWLINE] Is your problem,<mask> we get down to it, really just the specificity/arbitrary-ness of the percent system of letter grades? Would you rather just have a teacher give a rating based on<mask> they know about a student on whether they should pass or not? I know you feel a student should have near perfect mastery of a subject in order to pass,<mask><mask> do you propose educators meet that goal? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Do you think that a student that received an Unsatisfactory mark should be allowed to proceed to the next level? [ENDQ] [NEWLINE] Earlier today I would have said no, but [after reading this]( [URL].aspx?nfstatus=401&amp;nftoken=00000000-0000-0000-0000-000000000000&amp;nfstatusdescription=ERROR%3a+No+local+token) I've changed my mind in most cases. I don't think a student should just continue on the path with no intervention, but I think they should be given extra help in the form of a tutor, or individual attention with a teacher or aide or parent. This is more or less feasible when you have one or two "unsatisfactory" grades in a class of 30 than having 20 people get a grade lower than A. [NEWLINE] [NEWLINE] Typing this did remind me of the policy of my college language class - A C- was a passing grade, technically, and allowed you to move on. However, if you got more than one test grade of B- or below you had to meet with a tutor once a week until you demonstrated your proficiency. Would this be a possible solution to you? Because it seems pretty logical to me. [NEWLINE] [NEWLINE] Is your problem, when we get down to it, really just the specificity/arbitrary-ness of the percent system of letter grades? Would you rather just have a teacher give a rating based on what they know about a student on whether they should pass or not? I know you feel a student should have near perfect mastery of a subject in order to pass, but how do you propose educators meet that goal? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> MRA's began to take active steps to dismantle this "man up" culture that is embedded in our society I can practically guarantee you that issues such<mask> the dismissive attitudes towards abused males would disappear. [ENDQ] [NEWLINE] I am sure you don't visit /r/mensrights,<mask> please do a search there on the phrase "man up". You will find that nobody there believes that is appropriate. I believe we all fight against that. I believe a man can be sensitive or overbearing<mask> his temperment dictates. Basically the masculine identity has been desovled and<mask> such, the idea of manning up is obsolete. The is no one thing to "man up" toward. There is no one kind of man now (was there ever?). [NEWLINE] [NEWLINE] [STARTQ] And<mask> exactly is this the fault of women? [ENDQ] [NEWLINE] It is not the sole fault of women,<mask> women willingly participate in behaviors that benefit them.<mask> shaming a man into doing her bidding seems appropriate at the time, some women will do<mask> by telling a man to man-up. Even current dating culture that shuns a man for not being<mask> women call confident or<mask> he doesn't make her feel like a woman by him acting like a man then he needs to man-up. [NEWLINE] [NEWLINE] Some women and some men benefit from<mask> feminists call patriarchy. Feminists tend to fight against the disadvantages they face from patriarchy<mask> ignore the aspects of patriarchy that disadvantage men. Feminists<mask> do not fight against the benefits they receive from<mask> they call patriarchy. You ask "<mask> exactly is this the fault of women?" Some women actively support the continuation of the negative aspects of patriarchy for men<mask> it has a benefit for women.</s>
Label encoding: <s> [STARTQ] If MRA's began to take active steps to dismantle this "man up" culture that is embedded in our society I can practically guarantee you that issues such as the dismissive attitudes towards abused males would disappear. [ENDQ] [NEWLINE] I am sure you don't visit /r/mensrights, but please do a search there on the phrase "man up". You will find that nobody there believes that is appropriate. I believe we all fight against that. I believe a man can be sensitive or overbearing as his temperment dictates. Basically the masculine identity has been desovled and as such, the idea of manning up is obsolete. The is no one thing to "man up" toward. There is no one kind of man now (was there ever?). [NEWLINE] [NEWLINE] [STARTQ] And how exactly is this the fault of women? [ENDQ] [NEWLINE] It is not the sole fault of women, but women willingly participate in behaviors that benefit them. If shaming a man into doing her bidding seems appropriate at the time, some women will do so by telling a man to man-up. Even current dating culture that shuns a man for not being what women call confident or if he doesn't make her feel like a woman by him acting like a man then he needs to man-up. [NEWLINE] [NEWLINE] Some women and some men benefit from what feminists call patriarchy. Feminists tend to fight against the disadvantages they face from patriarchy but ignore the aspects of patriarchy that disadvantage men. Feminists also do not fight against the benefits they receive from what they call patriarchy. You ask " how exactly is this the fault of women?" Some women actively support the continuation of the negative aspects of patriarchy for men because it has a benefit for women.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You assume that HOA's have perfect knowledge about the housing market &amp; perceived value on behalf of future residents.<mask> time goes on, vacant properties make a much bigger impact on residential value than any other aspect HOA regulations seek to enforce. [NEWLINE] [NEWLINE] MY in-laws HOA is<mask> nonsensically over-bearing, they HURT property values by being notoriously over-bearing. Nobody wants to live in that development, which was "under water" for 4-5 years. The only people moving in are retirees &amp; commuters who don't know the area. Its considered "over valued" locally.<mask> the HOA were a democratic government, we'd see headlines about 20% foreclosure &amp; vacancy rates. Other neighboring towns have had elections defined over much smaller economic failures. [NEWLINE] [NEWLINE] Alot of<mask> a HOA regulates...isn't black &amp; white. Hibiscus increase property values<mask> Sunflowers don't?<mask> of an agricultural association from post-ww2? Even<mask> your demographic is mostly immigrants from farming countries &amp; yuppies who deliberately moved "out of the city into the farmland"? Values change over time &amp; are more dynamic than a local tax assessor will realistically capture. [NEWLINE] [NEWLINE] Crab grass decreases property values? Okay -<mask><mask> does over the brown spots caused by bad fertilization &amp; pesticides applied casually by the HOA's individual households.<mask> they don't fine for this,<mask> they can't actually turn residents into green-thumbs &amp; they don't have a requirement to use landscapers in the HOA charter. Which nobody would agree to even<mask> they tried. [NEWLINE] </s>
Label encoding: <s>You assume that HOA's have perfect knowledge about the housing market &amp; perceived value on behalf of future residents. As time goes on, vacant properties make a much bigger impact on residential value than any other aspect HOA regulations seek to enforce. [NEWLINE] [NEWLINE] MY in-laws HOA is so nonsensically over-bearing, they HURT property values by being notoriously over-bearing. Nobody wants to live in that development, which was "under water" for 4-5 years. The only people moving in are retirees &amp; commuters who don't know the area. Its considered "over valued" locally. IF the HOA were a democratic government, we'd see headlines about 20% foreclosure &amp; vacancy rates. Other neighboring towns have had elections defined over much smaller economic failures. [NEWLINE] [NEWLINE] Alot of what a HOA regulates...isn't black &amp; white. Hibiscus increase property values but Sunflowers don't? Because of an agricultural association from post-ww2? Even when your demographic is mostly immigrants from farming countries &amp; yuppies who deliberately moved "out of the city into the farmland"? Values change over time &amp; are more dynamic than a local tax assessor will realistically capture. [NEWLINE] [NEWLINE] Crab grass decreases property values? Okay - but so does over the brown spots caused by bad fertilization &amp; pesticides applied casually by the HOA's individual households. But they don't fine for this, because they can't actually turn residents into green-thumbs &amp; they don't have a requirement to use landscapers in the HOA charter. Which nobody would agree to even when they tried. [NEWLINE] </s>
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem is that sending in ground forces wouldn't necessarily lower the civilian casualty rates. Members of HAMAS don't follow the rules of engagement, meaning that sending in ground troops would put Israeli soldiers at risk<mask> doing little to limit the civilian casualty rate. To HAMAS, a higher civilian casualty rate is beneficial<mask> it leads to more international opposition to Israel. This is<mask> they have no problem using civilians<mask> shield; it causes more civilian casualties<mask> turning the world against Israel. HAMAS knows that it doesn't have a chance of defeating Israel militarily,<mask><mask> they can force Israel to kill enough civilians then they believe that it can result in sanctions against Israel that would be the next best thing. [NEWLINE] [NEWLINE] <mask> for HAMAS physically stopping civilians from evacuating, I can't find the specific article,<mask> [this article mentions it without going into detail]( [URL] )<mask> well<mask> talking about<mask> HAMAS is using civilian deaths to hurt Israel. I will try to find the original source and post it. Regardless it seems like you agree that HAMAS tells Palestinian civilians not to evacuate knowing that they are in the line of fire from Israel. Even<mask> these people can run 500 feet away from their homes that they have been warned will be targeted it's better than them staying like HAMAS is advising them to do. [NEWLINE] [NEWLINE] It's a tricky situation.<mask> Israel does nothing then HAMAS will continue to fire rockets at civilian targets in Israel with no retaliation, making Israel look weak to other armed groups in the region.<mask> Israel does respond (which it has), then civilians are bound to die<mask><mask><mask> of HAMAS putting them directly in harms way<mask> a part of their overall strategy.</s>
Label encoding: <s>The problem is that sending in ground forces wouldn't necessarily lower the civilian casualty rates. Members of HAMAS don't follow the rules of engagement, meaning that sending in ground troops would put Israeli soldiers at risk while doing little to limit the civilian casualty rate. To HAMAS, a higher civilian casualty rate is beneficial as it leads to more international opposition to Israel. This is why they have no problem using civilians as shield; it causes more civilian casualties while turning the world against Israel. HAMAS knows that it doesn't have a chance of defeating Israel militarily, but if they can force Israel to kill enough civilians then they believe that it can result in sanctions against Israel that would be the next best thing. [NEWLINE] [NEWLINE] As for HAMAS physically stopping civilians from evacuating, I can't find the specific article, but [this article mentions it without going into detail]( [URL] ) as well as talking about how HAMAS is using civilian deaths to hurt Israel. I will try to find the original source and post it. Regardless it seems like you agree that HAMAS tells Palestinian civilians not to evacuate knowing that they are in the line of fire from Israel. Even if these people can run 500 feet away from their homes that they have been warned will be targeted it's better than them staying like HAMAS is advising them to do. [NEWLINE] [NEWLINE] It's a tricky situation. If Israel does nothing then HAMAS will continue to fire rockets at civilian targets in Israel with no retaliation, making Israel look weak to other armed groups in the region. If Israel does respond (which it has), then civilians are bound to die as a result of HAMAS putting them directly in harms way as a part of their overall strategy.</s>
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think that's the point. I'm a woman. I've had this situation happen, I went home with a man, we kissed, got undressed and we all know<mask> it was leading. Now I knew that<mask> I agreed to go home with him.<mask> sometime in between kissing and being on the bed I got a really, really bad feeling about this. I was extremely uncomfortable with the situation and not in the mood anymore. He seemed way too forceful for<mask> it was. I then voiced my doubts. He said sorry and tried to continue.<mask> he was still being too forceful I told him I had changed my mind and didn't want to. [NEWLINE] [NEWLINE] I started getting dressed, he appologised and said I shouldn't leave we could just watch TV. I agreed<mask> shortly after he started kissing me again. I was going to give him another chance<mask> again he started being very forceful. I said "You are doing it again. I told you I don't like<mask> you are doing. I'm going home." And then I went home. [NEWLINE] [NEWLINE] At no point would I ever say this was sexual assault. Not even close. I gave him consent in the beginning. Then I didn't anymore and yes he was a bit pushy trying to change my mind<mask> it was my choice<mask> I did change my mind or not.<mask> I said "fuck it" and just let him go ahead it would have been my choice. It is maybe not verbal consent<mask> there would have been nothing wrong with it. He just tried to convince me otherwise. Now i chose that I didn't want to sleep with him and I actually didn't and left. </s>
Label encoding: <s>I don't think that's the point. I'm a woman. I've had this situation happen, I went home with a man, we kissed, got undressed and we all know where it was leading. Now I knew that when I agreed to go home with him. But sometime in between kissing and being on the bed I got a really, really bad feeling about this. I was extremely uncomfortable with the situation and not in the mood anymore. He seemed way too forceful for what it was. I then voiced my doubts. He said sorry and tried to continue. When he was still being too forceful I told him I had changed my mind and didn't want to. [NEWLINE] [NEWLINE] I started getting dressed, he appologised and said I shouldn't leave we could just watch TV. I agreed but shortly after he started kissing me again. I was going to give him another chance when again he started being very forceful. I said "You are doing it again. I told you I don't like what you are doing. I'm going home." And then I went home. [NEWLINE] [NEWLINE] At no point would I ever say this was sexual assault. Not even close. I gave him consent in the beginning. Then I didn't anymore and yes he was a bit pushy trying to change my mind but it was my choice if I did change my mind or not. If I said "fuck it" and just let him go ahead it would have been my choice. It is maybe not verbal consent but there would have been nothing wrong with it. He just tried to convince me otherwise. Now i chose that I didn't want to sleep with him and I actually didn't and left. </s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you make a lot of good points here. My only problem with your post is this -- [NEWLINE] [NEWLINE] [STARTQ] <mask>, I'm going to call you out. You're not here to have your view changed. You're here to make a case for something that you think is<mask> obviously true that it's irrefutable. [ENDQ] [NEWLINE] -- and here's<mask>. [NEWLINE] [NEWLINE] My argument is somewhat open: you're correct. I find it surprising,<mask>, that<mask> someone who spoke of the danger of making assumptions you've automatically assumed things about me. An open argument does not mean that I do not want to have my view changed, or that I pompously believe my view to be correct and irrefutable. [NEWLINE] [NEWLINE] I'm a male who often questions his own sex. I often wonder<mask> it would be like to be a woman; there are days<mask> I wish I were a woman. I understand,<mask>, that this desire stems in part from a belief that AWW live easier lives -- that<mask> I had been born an AWW, in other words, my life would be easier. Certainly this post has felt very therapeutic and helped me work through some issues of my own.<mask> the cost of that is that my argument is open, well,<mask> be it. [NEWLINE] [NEWLINE] I take<mask> you wrote almost<mask> a personal jab, especially<mask> I had already awarded a delta before you even commented. [NEWLINE] [NEWLINE] Just<mask> I stated my view in the way that I did does not automatically make me an Internet troll just looking to stir things up. In my case, I had a very personal reason for asking and I'm quite glad that I've learned<mask> much from this. </s>
Label encoding: <s>I think you make a lot of good points here. My only problem with your post is this -- [NEWLINE] [NEWLINE] [STARTQ] Secondly, I'm going to call you out. You're not here to have your view changed. You're here to make a case for something that you think is so obviously true that it's irrefutable. [ENDQ] [NEWLINE] -- and here's why. [NEWLINE] [NEWLINE] My argument is somewhat open: you're correct. I find it surprising, however, that as someone who spoke of the danger of making assumptions you've automatically assumed things about me. An open argument does not mean that I do not want to have my view changed, or that I pompously believe my view to be correct and irrefutable. [NEWLINE] [NEWLINE] I'm a male who often questions his own sex. I often wonder what it would be like to be a woman; there are days when I wish I were a woman. I understand, however, that this desire stems in part from a belief that AWW live easier lives -- that if I had been born an AWW, in other words, my life would be easier. Certainly this post has felt very therapeutic and helped me work through some issues of my own. If the cost of that is that my argument is open, well, so be it. [NEWLINE] [NEWLINE] I take what you wrote almost as a personal jab, especially since I had already awarded a delta before you even commented. [NEWLINE] [NEWLINE] Just because I stated my view in the way that I did does not automatically make me an Internet troll just looking to stir things up. In my case, I had a very personal reason for asking and I'm quite glad that I've learned so much from this. </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Rights and freedoms are very subjective and arbitrary topics to discuss anyway. <mask> one might conceive of a right to own a firearm, another might use a different rhetoric and claim that the same exact situation is simply a lack of the right to live safely in an environment in which gun ownership is restricted and controlled. [NEWLINE] [NEWLINE] Rights speech is heavily critiqued for this arbitrary nature, and you can see this critique evident in the comments in this thread, which spend quite a bit of time debating<mask> exactly "freedoms" you enjoy in the UK vs. the US are.  More specifically, to use an example from this debate, we could frame the debate on public healthcare two ways.  You could<mask><mask> access to public healthcare isn't a freedom,<mask> you could<mask> claim that with public healthcare one has a freedom to obtain any needed healthcare without having to worry about a variety of consequences. [NEWLINE] [NEWLINE] With this train of thought in mind, your reasoning is more of a personal claim and not a general statement on the various "freedoms" people enjoy in either of the two nations. <mask> you see<mask> a freedom or right might be completely irrelevant to someone from a different socioeconomic or cultural background. [NEWLINE] [NEWLINE] I would<mask> like to say that I don't necessarily agree/disagree with your points about differences between the U.S. and the UK, I would just like to point out that we must consider different frames of mind<mask> considering whether there really is much of a difference in "freedoms" and "rights,"<mask> far<mask> the two countries go.  A legal system is, after all, directly influenced by it's populace's culture and economic history.</s>
Label encoding: <s>Rights and freedoms are very subjective and arbitrary topics to discuss anyway.  Where one might conceive of a right to own a firearm, another might use a different rhetoric and claim that the same exact situation is simply a lack of the right to live safely in an environment in which gun ownership is restricted and controlled. [NEWLINE] [NEWLINE] Rights speech is heavily critiqued for this arbitrary nature, and you can see this critique evident in the comments in this thread, which spend quite a bit of time debating what exactly "freedoms" you enjoy in the UK vs. the US are.  More specifically, to use an example from this debate, we could frame the debate on public healthcare two ways.  You could argue that access to public healthcare isn't a freedom, but you could also claim that with public healthcare one has a freedom to obtain any needed healthcare without having to worry about a variety of consequences. [NEWLINE] [NEWLINE] With this train of thought in mind, your reasoning is more of a personal claim and not a general statement on the various "freedoms" people enjoy in either of the two nations.  What you see as a freedom or right might be completely irrelevant to someone from a different socioeconomic or cultural background. [NEWLINE] [NEWLINE] I would also like to say that I don't necessarily agree/disagree with your points about differences between the U.S. and the UK, I would just like to point out that we must consider different frames of mind when considering whether there really is much of a difference in "freedoms" and "rights," so far as the two countries go.  A legal system is, after all, directly influenced by it's populace's culture and economic history.</s>
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Okay, I still hold<mask><mask>,<mask> I want to say I enjoyed this argument.** [NEWLINE] [NEWLINE] [STARTQ] Through an honest mistake, all the formula that she brought is lost or contaminated. [ENDQ] [NEWLINE] This is her responsibility to her child.<mask> she loses the formula, just like with a child she is responsible for<mask> happens. [NEWLINE] [NEWLINE] [STARTQ] She has food that she can eat,<mask> none that the baby can handle, and has a multi-day hike out of the woods. [ENDQ] [NEWLINE] Assuming she brought her phone, almost everyone does, she can call for help.<mask> not, and you can choose<mask> to take from this,<mask><mask> she should have to breastfeed<mask> it harms her in no way<mask> leaves the baby to die. Now this is different from the abortion argument<mask> the baby's life is more dependent on the mother's beliefs and less on her body. By the way I really like this argument you made. [NEWLINE] [NEWLINE] [STARTQ] <mask> now she has no access to formula, and could keep her baby nourished<mask> she<mask> chose. [ENDQ] [NEWLINE] Yes. She should<mask> it does no harm to her and the baby is dependent on her beliefs. [NEWLINE] [NEWLINE] [STARTQ] <mask> she comes out of the woods with a dead baby and says that she thought it would be infringing on her personal sovereignty<mask> she breastfed or premasticated food for the baby, would she be in the right? [ENDQ] [NEWLINE] No, it's infringing on the baby's rights. It doesn't harm her to breastfeed and the baby is suffering<mask> of it. It's *not* putting her in any potential harm. It<mask> isn't fair<mask> she's forcing her beliefs on others. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>**Okay, I still hold my opinion, but I want to say I enjoyed this argument.** [NEWLINE] [NEWLINE] [STARTQ] Through an honest mistake, all the formula that she brought is lost or contaminated. [ENDQ] [NEWLINE] This is her responsibility to her child. If she loses the formula, just like with a child she is responsible for what happens. [NEWLINE] [NEWLINE] [STARTQ] She has food that she can eat, but none that the baby can handle, and has a multi-day hike out of the woods. [ENDQ] [NEWLINE] Assuming she brought her phone, almost everyone does, she can call for help. If not, and you can choose what to take from this, I think she should have to breastfeed because it harms her in no way but leaves the baby to die. Now this is different from the abortion argument because the baby's life is more dependent on the mother's beliefs and less on her body. By the way I really like this argument you made. [NEWLINE] [NEWLINE] [STARTQ] So now she has no access to formula, and could keep her baby nourished if she so chose. [ENDQ] [NEWLINE] Yes. She should because it does no harm to her and the baby is dependent on her beliefs. [NEWLINE] [NEWLINE] [STARTQ] If she comes out of the woods with a dead baby and says that she thought it would be infringing on her personal sovereignty if she breastfed or premasticated food for the baby, would she be in the right? [ENDQ] [NEWLINE] No, it's infringing on the baby's rights. It doesn't harm her to breastfeed and the baby is suffering because of it. It's *not* putting her in any potential harm. It also isn't fair because she's forcing her beliefs on others. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, made to penetrate is not something that would result in the conception of a child 100% of the time. Aside from the fact that it is made to penetrate any orifice, it<mask> assumes that the woman is forcing penetration with the intend of conception (ie, not using birth control) [NEWLINE] [NEWLINE] These stats will<mask> include statutory rape, which<mask> not something I advocate in any form, they will include statistics<mask> the person in question would have consented,<mask> for the law preventing them from doing<mask><mask> of age. The vast difference in the forcing to penetrate and forcibly penetrating is the scope of the term.<mask> comparing direct genital rape, the stats reduce to the 1 in 5 and 1 in 71. The rest are forced oral actions, which<mask><mask> despicable, are irrelevant to the debate at hand, which focusses solely on the birthing of children. [NEWLINE] [NEWLINE] Same page, same report'the majority of male rape victims (93.3%) reported only male perpetrators.' Meaning 93.3% of genital rape (ie, not oral) in male cases, was male on male. [NEWLINE] [NEWLINE] <mask>,<mask> this debate is about forcibly fathering children, at least 27.8% of the time that would have been a medical miracle. [NEWLINE] 'More than one-quarter of male victims of completed rape (27.8%) were first raped<mask> they were 10 years old or younger (data not shown). With the exception of the youngest age category (i.e., age 10 or younger), the estimates for age at first completed rape for male victims in the other age groups were based upon numbers too small to calculate a reliable estimate and<mask> are not reported.'</s>
Label encoding: <s>Ok, made to penetrate is not something that would result in the conception of a child 100% of the time. Aside from the fact that it is made to penetrate any orifice, it also assumes that the woman is forcing penetration with the intend of conception (ie, not using birth control) [NEWLINE] [NEWLINE] These stats will also include statutory rape, which although not something I advocate in any form, they will include statistics where the person in question would have consented, but for the law preventing them from doing so because of age. The vast difference in the forcing to penetrate and forcibly penetrating is the scope of the term. When comparing direct genital rape, the stats reduce to the 1 in 5 and 1 in 71. The rest are forced oral actions, which although also despicable, are irrelevant to the debate at hand, which focusses solely on the birthing of children. [NEWLINE] [NEWLINE] Same page, same report'the majority of male rape victims (93.3%) reported only male perpetrators.' Meaning 93.3% of genital rape (ie, not oral) in male cases, was male on male. [NEWLINE] [NEWLINE] Also, as this debate is about forcibly fathering children, at least 27.8% of the time that would have been a medical miracle. [NEWLINE] 'More than one-quarter of male victims of completed rape (27.8%) were first raped when they were 10 years old or younger (data not shown). With the exception of the youngest age category (i.e., age 10 or younger), the estimates for age at first completed rape for male victims in the other age groups were based upon numbers too small to calculate a reliable estimate and therefore are not reported.'</s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Nope never said or implied that... [NEWLINE] [NEWLINE] I said Criminals are going to be criminals<mask><mask> Drugs being legal or not<mask> the legality of drugs isn't<mask> turned them into criminals [NEWLINE] [NEWLINE] I said that they would move on to things like prostitution, or going into<mask> ever drugs weren't declared legal. [NEWLINE] [NEWLINE] At no point do I infer that people are just criminals from birth<mask> instead that of the plethora of things that make a criminal the legality of drugs wasn't one of them. <mask> legalizing drugs isn't going to make them turn their back on a life of crime. <mask> drugs are legal, they still have all the same problems they always had and now, they don't have any income from drugs, it was these circumstances that turned them to crime in the first place. [NEWLINE] [NEWLINE] Yes in the beginning there would be far less people incarcerated... we covered that too,<mask> that doesn't come without a cost either, and you have to take that cost into account.  The Billions spent on the War on Drugs and incarcerating these people... wasn't just vanishing into thin air it was going into the pockets of hard working Americans in Law Enforcement, in DOC, and all the contractors who provided their supplies. <mask> the tax payer saves billions, that is billions less in the economy and more people on welfare and unemployment etc. [NEWLINE] [NEWLINE] <mask><mask> it may be an overall gain (I honestly don't know) there will be peoples whose lives would be ruined by it too and the money the government was spending to fund these things will no go into supporting these people<mask> they don't work.  That has to be taken into consideration too</s>
Label encoding: <s>Nope never said or implied that... [NEWLINE] [NEWLINE] I said Criminals are going to be criminals regardless of Drugs being legal or not because the legality of drugs isn't what turned them into criminals [NEWLINE] [NEWLINE] I said that they would move on to things like prostitution, or going into what ever drugs weren't declared legal. [NEWLINE] [NEWLINE] At no point do I infer that people are just criminals from birth but instead that of the plethora of things that make a criminal the legality of drugs wasn't one of them.  Thus legalizing drugs isn't going to make them turn their back on a life of crime.  When drugs are legal, they still have all the same problems they always had and now, they don't have any income from drugs, it was these circumstances that turned them to crime in the first place. [NEWLINE] [NEWLINE] Yes in the beginning there would be far less people incarcerated... we covered that too, but that doesn't come without a cost either, and you have to take that cost into account.  The Billions spent on the War on Drugs and incarcerating these people... wasn't just vanishing into thin air it was going into the pockets of hard working Americans in Law Enforcement, in DOC, and all the contractors who provided their supplies.  While the tax payer saves billions, that is billions less in the economy and more people on welfare and unemployment etc. [NEWLINE] [NEWLINE] So while it may be an overall gain (I honestly don't know) there will be peoples whose lives would be ruined by it too and the money the government was spending to fund these things will no go into supporting these people while they don't work.  That has to be taken into consideration too</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You'd<mask> have to aggregate over all benefits and costs.<mask> there are incentives put out for women, or any group, you may gain a few who otherwise would not have been there. The first problem is that<mask> the incentives are<mask> attracted them<mask> they otherwise would not have gone, they are by definition a marginal case and not really dedicated to that field,<mask> they will tend not to excel at it. [NEWLINE] [NEWLINE] The second problem is that<mask> incentives are made available, it adds a stigma to *all* women in that field, that nobody knows for sure<mask> they are truly there<mask> of their dedication to that field or<mask> they are there due to extra incentives. That is, it creates a *negative* incentive for women to be in that field<mask> they may now feel everybody thinks they are there due to the "special help" instead of legitimate dedications. It can tend to *remove* women from the field by tainting it, and it will remove women who would have been there had it not been for the incentives. [NEWLINE] [NEWLINE] It's<mask> important to note that all of these pressures can be true without any actual real-world behaviours or opinions. Merely the perceptions of the women themselves over<mask> they think, and<mask> they think other people think, are the largest direct effects. The effects of actually creating negative opinions or behaviours is a secondary effect which can<mask> feed the primary one. [NEWLINE] [NEWLINE] Yes, you need to balance the positive and negative, of course,<mask> the positive only works on a few and the negatives work on all,<mask> the negatives only need to be marginally small to more than offset the few positives. </s>
Label encoding: <s>You'd also have to aggregate over all benefits and costs. If there are incentives put out for women, or any group, you may gain a few who otherwise would not have been there. The first problem is that if the incentives are what attracted them when they otherwise would not have gone, they are by definition a marginal case and not really dedicated to that field, so they will tend not to excel at it. [NEWLINE] [NEWLINE] The second problem is that if incentives are made available, it adds a stigma to *all* women in that field, that nobody knows for sure if they are truly there because of their dedication to that field or if they are there due to extra incentives. That is, it creates a *negative* incentive for women to be in that field as they may now feel everybody thinks they are there due to the "special help" instead of legitimate dedications. It can tend to *remove* women from the field by tainting it, and it will remove women who would have been there had it not been for the incentives. [NEWLINE] [NEWLINE] It's also important to note that all of these pressures can be true without any actual real-world behaviours or opinions. Merely the perceptions of the women themselves over what they think, and what they think other people think, are the largest direct effects. The effects of actually creating negative opinions or behaviours is a secondary effect which can also feed the primary one. [NEWLINE] [NEWLINE] Yes, you need to balance the positive and negative, of course, but the positive only works on a few and the negatives work on all, so the negatives only need to be marginally small to more than offset the few positives. </s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Like I said in the title, I believe that<mask> you don't opt-in for organ donating, organs shouldn't be given to you in a time of need. Those who aren't willing to give shouldn't get. I<mask> feel that those who don't give to charity, should they find themselves in a bad situation one day, shouldn't be<mask><mask> much help. Simple karma<mask><mask><mask>. [NEWLINE] [NEWLINE] <mask> explaining<mask><mask> to others, most people I've argued with have completely disagreed with me without actually giving me any reasons<mask> to<mask>.<mask> you can, change my view. [NEWLINE] [NEWLINE] Edit: Those who can't donate their organs are exempt [NEWLINE] Edit 2: Sorry guys, I'm at school, I'll be home in 5 hours and I'll try to reply to all comments<mask> I get home [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Like I said in the title, I believe that If you don't opt-in for organ donating, organs shouldn't be given to you in a time of need. Those who aren't willing to give shouldn't get. I also feel that those who don't give to charity, should they find themselves in a bad situation one day, shouldn't be given that much help. Simple karma in my opinion. [NEWLINE] [NEWLINE] While explaining my opinion to others, most people I've argued with have completely disagreed with me without actually giving me any reasons as to why. If you can, change my view. [NEWLINE] [NEWLINE] Edit: Those who can't donate their organs are exempt [NEWLINE] Edit 2: Sorry guys, I'm at school, I'll be home in 5 hours and I'll try to reply to all comments when I get home [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> are we expected to treat food like they are another sentient being with feelings? [ENDQ] [NEWLINE] There seems to be a scientific consensus now that animals are<mask> conscious<mask> us. This is taken from [The Cambridge Declaration on Consciousness:]( [URL] #Cambridge_Declaration_on_Consciousness) [NEWLINE] [NEWLINE] [STARTQ] "The absence of a neocortex does not appear to preclude an organism from experiencing affective states. Convergent evidence indicates that non-human animals have the neuroanatomical, neurochemical, and neurophysiological substrates of conscious states along with the capacity to exhibit intentional behaviors.<mask>, the weight of evidence indicates that humans are not unique in possessing the neurological substrates that generate consciousness. Non-human animals, including all mammals and birds, and many other creatures, including octopuses,<mask> possess these neurological substrates." [ENDQ] [NEWLINE] [STARTQ] "Birds appear to offer, in their behaviour, neurophysiology, and neuroanatomy a striking case of parallel evolution of consciousness. Evidence of near human-like levels of consciousness has been most dramatically observed in African grey parrots. Mammalian and avian emotional networks and cognitive microcircuitries appear to be far more homologous than previously thought.<mask>, certain species of birds have been found to exhibit neural sleep patterns similar to those of mammals, including REM sleep and,<mask> was demonstrated in zebra finches, neurophysiological patterns previously thought to require a mammalian neocortex. Magpies in particular have been shown to exhibit striking similarities to humans, great apes, dolphins, and elephants in studies of mirror self-recognition." [ENDQ] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Why are we expected to treat food like they are another sentient being with feelings? [ENDQ] [NEWLINE] There seems to be a scientific consensus now that animals are as conscious as us. This is taken from [The Cambridge Declaration on Consciousness:]( [URL] #Cambridge_Declaration_on_Consciousness) [NEWLINE] [NEWLINE] [STARTQ] "The absence of a neocortex does not appear to preclude an organism from experiencing affective states. Convergent evidence indicates that non-human animals have the neuroanatomical, neurochemical, and neurophysiological substrates of conscious states along with the capacity to exhibit intentional behaviors. Consequently, the weight of evidence indicates that humans are not unique in possessing the neurological substrates that generate consciousness. Non-human animals, including all mammals and birds, and many other creatures, including octopuses, also possess these neurological substrates." [ENDQ] [NEWLINE] [STARTQ] "Birds appear to offer, in their behaviour, neurophysiology, and neuroanatomy a striking case of parallel evolution of consciousness. Evidence of near human-like levels of consciousness has been most dramatically observed in African grey parrots. Mammalian and avian emotional networks and cognitive microcircuitries appear to be far more homologous than previously thought. Moreover, certain species of birds have been found to exhibit neural sleep patterns similar to those of mammals, including REM sleep and, as was demonstrated in zebra finches, neurophysiological patterns previously thought to require a mammalian neocortex. Magpies in particular have been shown to exhibit striking similarities to humans, great apes, dolphins, and elephants in studies of mirror self-recognition." [ENDQ] [NEWLINE] </s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I worked for several years with a homelessness services organization, connecting people with emergency shelter beds. In my work I got to hear a lot of heartbreaking stories, and see a lot of wasted lives, too. [NEWLINE] [NEWLINE] **Things to keep in mind:** [Anywhere between 1:3 and 1:2 homeless mothers report having been physically abused]( [URL].pdf). Many people on the streets are fleeing domestic violence. [NEWLINE] [NEWLINE] [40% of homeless men have served in the armed forces]( [URL] ). Often they have invisible wounds, such<mask> PTSD; sometimes they are ineligible for any assistance<mask> they were dishonorably discharged (for things like smoking a joint, or loving another man). [NEWLINE] [NEWLINE] Lots of people I worked with were disabled seniors who had no family to help them, or else were abandoned by their families literally onto the street. [NEWLINE] [NEWLINE] On the other end of the spectrum, many [youth who are living on the streets]( [URL] ) are fleeing physical or sexual violence in the home, or have been abandoned by their families<mask> of their sexuality. And it is very, very common for foster youth to become homeless<mask> they "age out" of care at 18. [NEWLINE] [NEWLINE] And even for those who have spiraled down<mask> of addiction, its important to look at cause and effect. Some people become homeless<mask> of addiction, and some people become addicts<mask> of homelessness--miserable, depressed, humiliated, terrified, and looking for a way to just make the pain stop. It's not the best coping mechanism in the world, I'll admit.<mask> living on the street is no joke.</s>
Label encoding: <s>I worked for several years with a homelessness services organization, connecting people with emergency shelter beds. In my work I got to hear a lot of heartbreaking stories, and see a lot of wasted lives, too. [NEWLINE] [NEWLINE] **Things to keep in mind:** [Anywhere between 1:3 and 1:2 homeless mothers report having been physically abused]( [URL].pdf). Many people on the streets are fleeing domestic violence. [NEWLINE] [NEWLINE] [40% of homeless men have served in the armed forces]( [URL] ). Often they have invisible wounds, such as PTSD; sometimes they are ineligible for any assistance because they were dishonorably discharged (for things like smoking a joint, or loving another man). [NEWLINE] [NEWLINE] Lots of people I worked with were disabled seniors who had no family to help them, or else were abandoned by their families literally onto the street. [NEWLINE] [NEWLINE] On the other end of the spectrum, many [youth who are living on the streets]( [URL] ) are fleeing physical or sexual violence in the home, or have been abandoned by their families because of their sexuality. And it is very, very common for foster youth to become homeless when they "age out" of care at 18. [NEWLINE] [NEWLINE] And even for those who have spiraled down because of addiction, its important to look at cause and effect. Some people become homeless because of addiction, and some people become addicts because of homelessness--miserable, depressed, humiliated, terrified, and looking for a way to just make the pain stop. It's not the best coping mechanism in the world, I'll admit. But living on the street is no joke.</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Another aspect to consider is the increasing militarization of police, and advancements in technology.  I don't know the actual statistics of<mask> many bad cops there are, or<mask> it's gone up or down. <mask> even<mask> the percentage had gone down, the change in<mask> tools the police use<mask> affects<mask> much of a problem it is. [NEWLINE] [NEWLINE] Before tasers, there was a much bigger gap between the non-violent methods a cop could use, or going for their gun<mask> they were prepared to kill somebody.  Tasers were advertised<mask> a non-lethal option, to reduce deaths even<mask> things got to that extreme point<mask> a cop would have had to be prepared to kill before.  Instead, tasers have become a compliance tool, used on a much wider array of situations<mask> police would have (or should have) never considered reaching for a gun.  And now tasers have proven to be simply "less lethal" rather than non-lethal,<mask> they can still have fatal effects<mask> used on someone with health problems, or<mask> abused with excessive use. [NEWLINE] [NEWLINE] Other tools like LRADs are used<mask> crowd control.  In a crowd control situation, they could affect many non-violent protesters who would have complied with police directions.  Whether a non-violent protester should expect some risk of having some unruly people in the crowd and getting treated<mask> is a separate debate. <mask> the police should<mask> expect there to be innocents in the collateral damage, and<mask> they shouldn't employ tools like the LRAD's which can cause *permanent* hearing damage.</s>
Label encoding: <s>Another aspect to consider is the increasing militarization of police, and advancements in technology.  I don't know the actual statistics of how many bad cops there are, or if it's gone up or down.  But even if the percentage had gone down, the change in what tools the police use also affects how much of a problem it is. [NEWLINE] [NEWLINE] Before tasers, there was a much bigger gap between the non-violent methods a cop could use, or going for their gun when they were prepared to kill somebody.  Tasers were advertised as a non-lethal option, to reduce deaths even when things got to that extreme point where a cop would have had to be prepared to kill before.  Instead, tasers have become a compliance tool, used on a much wider array of situations where police would have (or should have) never considered reaching for a gun.  And now tasers have proven to be simply "less lethal" rather than non-lethal, as they can still have fatal effects when used on someone with health problems, or when abused with excessive use. [NEWLINE] [NEWLINE] Other tools like LRADs are used as crowd control.  In a crowd control situation, they could affect many non-violent protesters who would have complied with police directions.  Whether a non-violent protester should expect some risk of having some unruly people in the crowd and getting treated accordingly is a separate debate.  But the police should also expect there to be innocents in the collateral damage, and so they shouldn't employ tools like the LRAD's which can cause *permanent* hearing damage.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0904, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] People will not support laws that double or triple the price of milk and eggs. [ENDQ] [NEWLINE] That's bullshit. Just to give you a few examples: speeding, DUI, littering, jay-walking... [NEWLINE] [NEWLINE] [STARTQ] It is<mask> they should avoid financially supporting industries that normalize the consumption of animal products,<mask> those industries compete with vegan options [ENDQ] [NEWLINE] I don't want to be vegan,<mask> I'm sure it's not<mask> healthy<mask> most vegans paint it.<mask> not try to go for a world<mask> most people get vegetarian, or like I do, eat meat once or twice a week? You will never have my support for something I consider unhealty. [NEWLINE] [NEWLINE] [STARTQ] "I will buy cow milk,<mask> that is the normal option, and it is the cheapest option. I will not buy that expensive brand,<mask> I need to be thrifty". [ENDQ] [NEWLINE] I buy sheep milk, it has more nutrients, basically has the same taste, and I'm 100% sure the animal isn't at a factory-farm. About cow-milk, put a levy on farm-factory-milk and free-range-cows become an option for a company. [NEWLINE] [NEWLINE] [STARTQ] Is soy milk weird? No, I know many people who drink it, I'll buy that instead. [ENDQ] [NEWLINE] Soy milk has more or less the same price here, it's more expensive than generic milk,<mask> less than milk from big labels,<mask>... I buy it sometimes,<mask> it's not an alternative in the kitchen, nor does it taste well<mask> put into coffee.</s>
Label encoding: <s> [STARTQ] People will not support laws that double or triple the price of milk and eggs. [ENDQ] [NEWLINE] That's bullshit. Just to give you a few examples: speeding, DUI, littering, jay-walking... [NEWLINE] [NEWLINE] [STARTQ] It is why they should avoid financially supporting industries that normalize the consumption of animal products, because those industries compete with vegan options [ENDQ] [NEWLINE] I don't want to be vegan, because I'm sure it's not as healthy as most vegans paint it. Why not try to go for a world where most people get vegetarian, or like I do, eat meat once or twice a week? You will never have my support for something I consider unhealty. [NEWLINE] [NEWLINE] [STARTQ] "I will buy cow milk, because that is the normal option, and it is the cheapest option. I will not buy that expensive brand, because I need to be thrifty". [ENDQ] [NEWLINE] I buy sheep milk, it has more nutrients, basically has the same taste, and I'm 100% sure the animal isn't at a factory-farm. About cow-milk, put a levy on farm-factory-milk and free-range-cows become an option for a company. [NEWLINE] [NEWLINE] [STARTQ] Is soy milk weird? No, I know many people who drink it, I'll buy that instead. [ENDQ] [NEWLINE] Soy milk has more or less the same price here, it's more expensive than generic milk, but less than milk from big labels, so... I buy it sometimes, but it's not an alternative in the kitchen, nor does it taste well when put into coffee.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who says "Hi I'm African-American?"<mask> are you talking about people saying "I'm Irish-American"? [NEWLINE] [NEWLINE] [STARTQ] <mask> an American is of European heritage<mask> they don't know whereabouts in Europe,<mask> are they not referred to<mask> European-American? [ENDQ] [NEWLINE] 90% of European Americans know the country of their ancestry or at least identify with at least one of the countries. The same cannot be said for African Americans. That is the big difference. [NEWLINE] [NEWLINE] And Obama is referred to<mask> a Kenyan<mask><mask> the origin is known it's common to [refer to the country rather]( [URL] ) than the continent. [NEWLINE] [NEWLINE] In the end<mask> it's just words. White/black. European American/African American.<mask> does it really matter? European American never caught on and white has stuck<mask> talking about people of the European race. For blacks the color of their skin was used<mask> a derogatory term for them (Negro is black in Spanish). [NEWLINE] [NEWLINE] African-American became common in the 1980s to promote pride in heritage among ancestors of slaves. Irish-Americans, German-Americans, etc. had their pride in their heritage<mask> African-American became popular in the same vein. It's about pride not just in who you are<mask> in<mask> you came from. [NEWLINE] [NEWLINE] <mask>'s important to remember is that for generations America tried to deAfricanize slaves and their ancestors. The term African-American was pushed to reAfricanize themselves. [NEWLINE] [NEWLINE] I suppose it's good that<mask> many kids on Reddit are ignorant of this history. Hopefully it means we are moving on.</s>
Label encoding: <s>Who says "Hi I'm African-American?" Why are you talking about people saying "I'm Irish-American"? [NEWLINE] [NEWLINE] [STARTQ] If an American is of European heritage but they don't know whereabouts in Europe, why are they not referred to as European-American? [ENDQ] [NEWLINE] 90% of European Americans know the country of their ancestry or at least identify with at least one of the countries. The same cannot be said for African Americans. That is the big difference. [NEWLINE] [NEWLINE] And Obama is referred to as a Kenyan so when the origin is known it's common to [refer to the country rather]( [URL] ) than the continent. [NEWLINE] [NEWLINE] In the end though it's just words. White/black. European American/African American. What does it really matter? European American never caught on and white has stuck when talking about people of the European race. For blacks the color of their skin was used as a derogatory term for them (Negro is black in Spanish). [NEWLINE] [NEWLINE] African-American became common in the 1980s to promote pride in heritage among ancestors of slaves. Irish-Americans, German-Americans, etc. had their pride in their heritage so African-American became popular in the same vein. It's about pride not just in who you are but in where you came from. [NEWLINE] [NEWLINE] What's important to remember is that for generations America tried to deAfricanize slaves and their ancestors. The term African-American was pushed to reAfricanize themselves. [NEWLINE] [NEWLINE] I suppose it's good that so many kids on Reddit are ignorant of this history. Hopefully it means we are moving on.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am surprised that nobody, even the top-voted comment, mentions the thing that has *always* been the most important: applications. [NEWLINE] [NEWLINE] Sure, we may be in the cloud and always on the web now,<mask> it still comes down to Microsoft Office<mask> you're a Windows-using business type, and Adobe Creative Suite and ProTools<mask> you're a Mac-using creative type. And games. [NEWLINE] [NEWLINE] Few average PC users want to invest the time to switch to Google Docs or Open Office, nor do they want to accept the their limitations.<mask> for web browsers, yes, Linux has parity with Firefox and Chromium,<mask> will never have Internet Explorer or Safari. Users just don't like change. [NEWLINE] [NEWLINE] Your four bullets points are all quite irrelevant. [NEWLINE] [NEWLINE] * Energy efficiency is really more about hardware, and a new laptop will get you that (possible better on Windows with manufacturer-written proprietary power management drivers). [NEWLINE] [NEWLINE] * Performance is no longer an issue with most PCs, at least<mask> properly managed (i.e., no junkware and properly configured anti-virus). [NEWLINE] [NEWLINE] * Users don't want a learning experience, they want an enjoyable and/or productive experience. [NEWLINE] [NEWLINE] *<mask> percentage even of *current Linux* users actually examine their source? [NEWLINE] [NEWLINE] Linux is awesome<mask> you are a programmer or hobbyist (obviously excluding iOS and Windows development) and serves very well<mask> a professional web development platform. I've been a Linux user<mask> 1997, and it's been practically my sole desktop OS at home for a decade now.</s>
Label encoding: <s>I am surprised that nobody, even the top-voted comment, mentions the thing that has *always* been the most important: applications. [NEWLINE] [NEWLINE] Sure, we may be in the cloud and always on the web now, but it still comes down to Microsoft Office if you're a Windows-using business type, and Adobe Creative Suite and ProTools if you're a Mac-using creative type. And games. [NEWLINE] [NEWLINE] Few average PC users want to invest the time to switch to Google Docs or Open Office, nor do they want to accept the their limitations. As for web browsers, yes, Linux has parity with Firefox and Chromium, but will never have Internet Explorer or Safari. Users just don't like change. [NEWLINE] [NEWLINE] Your four bullets points are all quite irrelevant. [NEWLINE] [NEWLINE] * Energy efficiency is really more about hardware, and a new laptop will get you that (possible better on Windows with manufacturer-written proprietary power management drivers). [NEWLINE] [NEWLINE] * Performance is no longer an issue with most PCs, at least if properly managed (i.e., no junkware and properly configured anti-virus). [NEWLINE] [NEWLINE] * Users don't want a learning experience, they want an enjoyable and/or productive experience. [NEWLINE] [NEWLINE] * What percentage even of *current Linux* users actually examine their source? [NEWLINE] [NEWLINE] Linux is awesome if you are a programmer or hobbyist (obviously excluding iOS and Windows development) and serves very well as a professional web development platform. I've been a Linux user since 1997, and it's been practically my sole desktop OS at home for a decade now.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Would you require feminists to focus on the challenges of being a lesbian, of being a woman of an ethnic minority, etc.? [ENDQ] [NEWLINE] Yes [NEWLINE] [NEWLINE] [STARTQ] much like feminism primarily focuses on the challenges and discrimination faced by women,<mask><mask> ethnicity, sexual orientation etc. [ENDQ] [NEWLINE] This is incorrect. That was 2nd wave feminism. Third wave (modern) feminism came about specifically<mask> women of color and lesbians felt they didn't have a voice in the movement. Third wave feminism is all about intersectionality, or the crossing of identities to create a unique life experience based on those identities. For example, a white gay man faces a different set of problems, has an entirely different life experience, and is less oppressed than a black gay man. The intersectional basis of 3rd wave feminism focuses specifically on addressing the most underprivileged and attempts to no longer silence those whose voices were stifled during the 2nd wave. It's<mask> more about the patriarchy in general, and<mask> that negatively affects not just women<mask><mask> men and transgender people. [NEWLINE] You can't say that just<mask> people of color have their own movement and that LGBT people have their own movement that there is no need for their issues to be addressed in feminism, or in other movements in general. Everyone has a gender, sexual orientation, race/ethnicity, etc., and they intersect and interact every moment of our lives to create a unique life experience and form of oppression. To compartmentalize the movements based on these identities,<mask> our identities are not compartmentalized, is illogical and oppressive to underprivileged groups.</s>
Label encoding: <s> [STARTQ] Would you require feminists to focus on the challenges of being a lesbian, of being a woman of an ethnic minority, etc.? [ENDQ] [NEWLINE] Yes [NEWLINE] [NEWLINE] [STARTQ] much like feminism primarily focuses on the challenges and discrimination faced by women, regardless of ethnicity, sexual orientation etc. [ENDQ] [NEWLINE] This is incorrect. That was 2nd wave feminism. Third wave (modern) feminism came about specifically because women of color and lesbians felt they didn't have a voice in the movement. Third wave feminism is all about intersectionality, or the crossing of identities to create a unique life experience based on those identities. For example, a white gay man faces a different set of problems, has an entirely different life experience, and is less oppressed than a black gay man. The intersectional basis of 3rd wave feminism focuses specifically on addressing the most underprivileged and attempts to no longer silence those whose voices were stifled during the 2nd wave. It's also more about the patriarchy in general, and how that negatively affects not just women but also men and transgender people. [NEWLINE] You can't say that just because people of color have their own movement and that LGBT people have their own movement that there is no need for their issues to be addressed in feminism, or in other movements in general. Everyone has a gender, sexual orientation, race/ethnicity, etc., and they intersect and interact every moment of our lives to create a unique life experience and form of oppression. To compartmentalize the movements based on these identities, when our identities are not compartmentalized, is illogical and oppressive to underprivileged groups.</s>
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There isn't a strict dichotomy between introverts and extroverts. People have differing levels of introversion and extroversion.<mask> that's useless for trying to tell people about yourself easily. In an ideal world, people could convey exactly<mask> they fall on the extroversion spectrum,<mask><mask> it is, that's impractical.<mask>, people describe themselves<mask> introverts. This doesn't mean they dislike socialising<mask> much<mask> other introverts,<mask> is just a simple way to convey that they are not a big fan of social interactions [NEWLINE] [NEWLINE] People who describe themselves<mask> introverts could probable be sociable<mask> they want to,<mask> would likely find it draining, dull and simply do not want.<mask>, I suppose that you could<mask><mask> it is personal choice or preference,<mask> I am unsure<mask> this is needed to change your view.<mask> someone dislikes socialising,<mask> would you attempt to impose it on them? Surely personal choice should be respected.<mask> is this any less legitimate to you than autistic people? [NEWLINE] [NEWLINE] <mask> people complained about being "forced to talk", in<mask> context was this? Were people simply trying to strike up a conversation, or persisting in trying to talk to them<mask> they clearly didn't want to? In the former,<mask><mask> that there is not a responsibility to avoid striking up a conversation,<mask><mask> it was more the latter then<mask><mask> that it's fairly reasonable. You aren't being asked to put on "kid gloves", just to respect that some people may just not want to talk or socialise, and to leave them alone</s>
Label encoding: <s>There isn't a strict dichotomy between introverts and extroverts. People have differing levels of introversion and extroversion. But that's useless for trying to tell people about yourself easily. In an ideal world, people could convey exactly where they fall on the extroversion spectrum, but as it is, that's impractical. So, people describe themselves as introverts. This doesn't mean they dislike socialising as much as other introverts, but is just a simple way to convey that they are not a big fan of social interactions [NEWLINE] [NEWLINE] People who describe themselves as introverts could probable be sociable if they want to, but would likely find it draining, dull and simply do not want. So, I suppose that you could argue that it is personal choice or preference, but I am unsure why this is needed to change your view. If someone dislikes socialising, why would you attempt to impose it on them? Surely personal choice should be respected. Why is this any less legitimate to you than autistic people? [NEWLINE] [NEWLINE] When people complained about being "forced to talk", in what context was this? Were people simply trying to strike up a conversation, or persisting in trying to talk to them when they clearly didn't want to? In the former, I agree that there is not a responsibility to avoid striking up a conversation, but if it was more the latter then I think that it's fairly reasonable. You aren't being asked to put on "kid gloves", just to respect that some people may just not want to talk or socialise, and to leave them alone</s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The maturity level is different for many people;<mask><mask> that the median age<mask> one suddenly becomes "mature" is likely around 22-23. I would not argue to move consent to this age,<mask> I believe that 18 is a decent age for adulthood, and like you said, 30 year olds can act like kids.<mask>, after about 18 years one has had plenty of chances to mature and become responsible. [NEWLINE] [NEWLINE] <mask>,<mask> the majority of children under age 18 are not mature, one should act with the presumption that those under 18 are not mature enough to have sex. [NEWLINE] [NEWLINE] <mask> I would not move consent up, there is a much larger potential for harm to oneself with sex than with other responsibilities of an 18 year old, such<mask> voting. [NEWLINE] [NEWLINE] Of course, you could argue military service could be considered to have a larger potential for harm (And then there's the draft, which is another conversation entirely),<mask> there is much more preparation leading up to it once you decide upon it, there's training.<mask> I digress. [NEWLINE] [NEWLINE] <mask> I'm trying to say is that<mask> there is no legal obligation to not ask for sex from people in the 18-23 age range<mask> you are much older, there is the moral responsibility of such an intimate act that is very different for those experienced and inexperienced. [NEWLINE] [NEWLINE] A perhaps inaccurate parallel to this would be that teenagers and young adults are more likely to involuntarily have an orgasm<mask> raped than older adults are. This leads me to believe that the emotions are more raw, and<mask> the potential for harm is larger.</s>
Label encoding: <s>The maturity level is different for many people; I think that the median age when one suddenly becomes "mature" is likely around 22-23. I would not argue to move consent to this age, as I believe that 18 is a decent age for adulthood, and like you said, 30 year olds can act like kids. However, after about 18 years one has had plenty of chances to mature and become responsible. [NEWLINE] [NEWLINE] However, as the majority of children under age 18 are not mature, one should act with the presumption that those under 18 are not mature enough to have sex. [NEWLINE] [NEWLINE] Although I would not move consent up, there is a much larger potential for harm to oneself with sex than with other responsibilities of an 18 year old, such as voting. [NEWLINE] [NEWLINE] Of course, you could argue military service could be considered to have a larger potential for harm (And then there's the draft, which is another conversation entirely), but there is much more preparation leading up to it once you decide upon it, there's training. But I digress. [NEWLINE] [NEWLINE] What I'm trying to say is that although there is no legal obligation to not ask for sex from people in the 18-23 age range if you are much older, there is the moral responsibility of such an intimate act that is very different for those experienced and inexperienced. [NEWLINE] [NEWLINE] A perhaps inaccurate parallel to this would be that teenagers and young adults are more likely to involuntarily have an orgasm while raped than older adults are. This leads me to believe that the emotions are more raw, and therefore the potential for harm is larger.</s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here is a possibly novel idea: You can be a feminist *and* an egalitarian. [NEWLINE] [NEWLINE] Amazing right? There's actually and virtually nothing preventing you or I from being both a feminist and egalitarian. [NEWLINE] [NEWLINE] Here's the difference between them: [NEWLINE] [NEWLINE] Being an *egalitarian* means I support equality for everyone<mask><mask> gender, race, nationality, religious affiliation, sexual orientation, etc etc. [NEWLINE] [NEWLINE] Being a *feminist* means I acknowledge and understand the real, subtle and blatant, historical and ongoing, oppression of women and many forms of inequality<mask> a sole result of gender and its proportionality to inequality in relation to men. [NEWLINE] [NEWLINE] Nothing about them is mutually exclusive to the other. [NEWLINE] [NEWLINE] Just<mask> I can be a doctor *and* a cardiologist: [NEWLINE] [NEWLINE] Being a *doctor* means I practice medicine and help promote health in general for my patients and I hold a medical degree and license. [NEWLINE] [NEWLINE] Being a *cardiologist* means I have specialized in treating medical matters relating to the human heart and understand the heart specifically<mask> it relates to the health of my patients. [NEWLINE] [NEWLINE] <mask>,<mask> someone told me: hey you'd be a way better doctor<mask> you dropped your cardiologist title,<mask> then you could push much more strongly for the health of your patients in general! [NEWLINE] [NEWLINE] I'd say that was a silly and relatively trivial notion and to stop bothering me and let me get back to fixing hearts, fixing women's rights, supporting the general health of my patients, and supporting equal rights for all. [NEWLINE] [NEWLINE] Hopefully this helps change your view.</s>
Label encoding: <s>Here is a possibly novel idea: You can be a feminist *and* an egalitarian. [NEWLINE] [NEWLINE] Amazing right? There's actually and virtually nothing preventing you or I from being both a feminist and egalitarian. [NEWLINE] [NEWLINE] Here's the difference between them: [NEWLINE] [NEWLINE] Being an *egalitarian* means I support equality for everyone regardless of gender, race, nationality, religious affiliation, sexual orientation, etc etc. [NEWLINE] [NEWLINE] Being a *feminist* means I acknowledge and understand the real, subtle and blatant, historical and ongoing, oppression of women and many forms of inequality as a sole result of gender and its proportionality to inequality in relation to men. [NEWLINE] [NEWLINE] Nothing about them is mutually exclusive to the other. [NEWLINE] [NEWLINE] Just how I can be a doctor *and* a cardiologist: [NEWLINE] [NEWLINE] Being a *doctor* means I practice medicine and help promote health in general for my patients and I hold a medical degree and license. [NEWLINE] [NEWLINE] Being a *cardiologist* means I have specialized in treating medical matters relating to the human heart and understand the heart specifically as it relates to the health of my patients. [NEWLINE] [NEWLINE] Thus, if someone told me: hey you'd be a way better doctor if you dropped your cardiologist title, because then you could push much more strongly for the health of your patients in general! [NEWLINE] [NEWLINE] I'd say that was a silly and relatively trivial notion and to stop bothering me and let me get back to fixing hearts, fixing women's rights, supporting the general health of my patients, and supporting equal rights for all. [NEWLINE] [NEWLINE] Hopefully this helps change your view.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I don't intend on stocking up on food and renting a nuclear bunker,<mask> does Doomsday prepping differ fundamentally from any other hobby?  It gives the preppers something to focus on, and through prepping one may<mask> find himself learning various skills which may enrich their day-to-day life.  Hobbies are often wasteful, taking a private jet for sightseeing is using up a lot of fuel just for a little joyride,<mask> I don't really see anything wrong with it<mask> the "wasted" fuel went towards an individual's satisfaction. [NEWLINE] [NEWLINE] <mask><mask> your views of the other collapses are a little simplistic, after all we did come very close to a nuclear war with the Cuban missile crisis, and<mask> additional nations develop nuclear technology, less stable ones such<mask> North Korea may pose legitimate risk in the future.  All it takes is for one nation to fire nukes for the floodgates to be opened.  The economic collapse bit you mention doesn't sound right<mask> I don't know enough about the subject to dispute it.  The disease of a pandemic may not manifest itself in any notable symptoms until after spreading has occurred, unless we screened all travelers for health complications stopping the spread is not that simple.  Your argument of civil unrest not being an issue assumes that the police/military will take the side of the government rather being divided, which<mask><mask> is unreasonable,<mask><mask> these people will usually have family in the citizen population. [NEWLINE] [NEWLINE] There's<mask> issues which could cause a collapse which you haven't addressed, particularly something like an asteroid impact.</s>
Label encoding: <s>While I don't intend on stocking up on food and renting a nuclear bunker, how does Doomsday prepping differ fundamentally from any other hobby?  It gives the preppers something to focus on, and through prepping one may also find himself learning various skills which may enrich their day-to-day life.  Hobbies are often wasteful, taking a private jet for sightseeing is using up a lot of fuel just for a little joyride, but I don't really see anything wrong with it if the "wasted" fuel went towards an individual's satisfaction. [NEWLINE] [NEWLINE] I think your views of the other collapses are a little simplistic, after all we did come very close to a nuclear war with the Cuban missile crisis, and as additional nations develop nuclear technology, less stable ones such as North Korea may pose legitimate risk in the future.  All it takes is for one nation to fire nukes for the floodgates to be opened.  The economic collapse bit you mention doesn't sound right but I don't know enough about the subject to dispute it.  The disease of a pandemic may not manifest itself in any notable symptoms until after spreading has occurred, unless we screened all travelers for health complications stopping the spread is not that simple.  Your argument of civil unrest not being an issue assumes that the police/military will take the side of the government rather being divided, which I think is unreasonable, given that these people will usually have family in the citizen population. [NEWLINE] [NEWLINE] There's also issues which could cause a collapse which you haven't addressed, particularly something like an asteroid impact.</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll pick up OP's question then and add the following conditions: [NEWLINE] [NEWLINE] A man is dying of thirst, and you own evian and have a surplus of millions of gallons of water that will probably go stagnant and have to be discarded soon<mask> demand for it is<mask> low:<mask> you do have the wealth and the power to defend every drop of your asset from others. [NEWLINE] [NEWLINE] You have no previous relationship with the thirsty man, he is a stranger. There is no law in your region against charging different people different amounts to sell them your assets. You are literally<mask> wealthy and your family is<mask> well off that after fleecing this thirsty man of everything he owns and indenturing him to a lifetime of wage slavery to pay off his nine figure debt for one 21 oz bottle of water that will save his life today, you will most likely throw the pittances he offers you onto the side of a pile of money that you keep in a three cubic acre building at the edge of town and never interact with it again. [NEWLINE] [NEWLINE] <mask> you meet, you materially owe this thirsty man nothing and he owes you nothing and the direct health and wellfare of no people whatsoever<mask> him will be impacted by the decision that you make. This is<mask> a perfect-knowledge hypothetical<mask> all actors are rational and beholden to whatever value systems you assign them, and nobody lacks critical information about anything to make their decisions (contrast with real world beggars<mask> who knows<mask> the money you offer them will go to good or ill use? Real life is rarely a perfect-knowledge situation..)</s>
Label encoding: <s>I'll pick up OP's question then and add the following conditions: [NEWLINE] [NEWLINE] A man is dying of thirst, and you own evian and have a surplus of millions of gallons of water that will probably go stagnant and have to be discarded soon because demand for it is so low: but you do have the wealth and the power to defend every drop of your asset from others. [NEWLINE] [NEWLINE] You have no previous relationship with the thirsty man, he is a stranger. There is no law in your region against charging different people different amounts to sell them your assets. You are literally so wealthy and your family is so well off that after fleecing this thirsty man of everything he owns and indenturing him to a lifetime of wage slavery to pay off his nine figure debt for one 21 oz bottle of water that will save his life today, you will most likely throw the pittances he offers you onto the side of a pile of money that you keep in a three cubic acre building at the edge of town and never interact with it again. [NEWLINE] [NEWLINE] When you meet, you materially owe this thirsty man nothing and he owes you nothing and the direct health and wellfare of no people whatsoever besides him will be impacted by the decision that you make. This is also a perfect-knowledge hypothetical where all actors are rational and beholden to whatever value systems you assign them, and nobody lacks critical information about anything to make their decisions (contrast with real world beggars where who knows if the money you offer them will go to good or ill use? Real life is rarely a perfect-knowledge situation..)</s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] One of the things that led to my failed belief in Christianity is I could not justify<mask> I shouldn't<mask> believe<mask> the Quran said.<mask>,<mask> the New Testament and the Quran are not mutually compatible, I cannot believe both. [ENDQ] [NEWLINE] There is actually a lot more in common between Mormonism and Islam than there is between Christianity and Islam. [NEWLINE] [NEWLINE] * Both Islam and Mormonism were founded by one charismatic individual whose writings are to be considered the basis of the faith.  (The Bible by contrast is comprised of 66 complementary documents written by dozens of authors over hundreds of years.) [NEWLINE] [NEWLINE] * The founders of both Islam and Mormonism claimed to be the last in a long line of prophets. (Christianity is not based of of the writings of one 'founder',<mask> it is named after Jesus who did not claim to be a prophet,<mask> God himself.) [NEWLINE] [NEWLINE] * The founders of both Islam and Mormonism claimed that Christianity had become polluted, and that they had been given revelation by an angel (Gabriel and Moroni) that commanded them to establish their respective religions.(Christians for their part wonder<mask> that pollution might have occurred,<mask> all modern Bible translations are based off of Greek manuscripts that date (in some cases) back to the first century). [NEWLINE] [NEWLINE] * Both Islam and Mormonism are tied to the region of their founder.  Islam is undeniably 'Middle Eastern' in culture and practice, Mormonism is undeniably 'American' in culture and practice. (Christianity by contrast is not tied to any one culture, and<mask><mask> does not neatly fit into any culture.)  </s>
Label encoding: <s> [STARTQ] One of the things that led to my failed belief in Christianity is I could not justify why I shouldn't also believe what the Quran said. However, since the New Testament and the Quran are not mutually compatible, I cannot believe both. [ENDQ] [NEWLINE] There is actually a lot more in common between Mormonism and Islam than there is between Christianity and Islam. [NEWLINE] [NEWLINE] * Both Islam and Mormonism were founded by one charismatic individual whose writings are to be considered the basis of the faith.  (The Bible by contrast is comprised of 66 complementary documents written by dozens of authors over hundreds of years.) [NEWLINE] [NEWLINE] * The founders of both Islam and Mormonism claimed to be the last in a long line of prophets. (Christianity is not based of of the writings of one 'founder', but it is named after Jesus who did not claim to be a prophet, but God himself.) [NEWLINE] [NEWLINE] * The founders of both Islam and Mormonism claimed that Christianity had become polluted, and that they had been given revelation by an angel (Gabriel and Moroni) that commanded them to establish their respective religions.(Christians for their part wonder where that pollution might have occurred, since all modern Bible translations are based off of Greek manuscripts that date (in some cases) back to the first century). [NEWLINE] [NEWLINE] * Both Islam and Mormonism are tied to the region of their founder.  Islam is undeniably 'Middle Eastern' in culture and practice, Mormonism is undeniably 'American' in culture and practice. (Christianity by contrast is not tied to any one culture, and in fact does not neatly fit into any culture.)  </s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your ideas about prostitution aren't wrong. The problem is that criminalization hasn't solved these problems. There still are prostitutes in every city<mask><mask><mask><mask> of criminalization they are more vulnerable. They can't seek assistance from violent clients, open certain types of bank accounts, or organize to defend their safety and interests. They can't do this<mask> criminalization means that seeking help can harm them. [NEWLINE] [NEWLINE] Prostitution has to be legal to secure the safety of prostitutes and bring them out of the shadows.<mask><mask><mask> it happens in the dark, the criminal element will reign and our ability to help women abused by this system is limited.<mask> prostitution were legal then it would be a lot easier to access women in prostitution with aid. It would<mask> mean that they could organize brothels<mask> that they could hire security. It would mean that they could open bank accounts to save money to get out of the system. it would mean that laws could be imposed to mandate the wearing of condoms and STD checks (which would be a public health victory). It would mean that the criminal element is pushed about the legal industry. And it could<mask> lead to prostitutes speaking out, organizing, and creating blacklists of persons who abuse them. [NEWLINE] [NEWLINE] Criminalization hasn't solved any of the issues that mentioned. In pretty much all cases, it has made it harder to help women abused by the system. Decriminalization is a path to bringing prostitution out of the dark,<mask> that reforms can be made to make it a safer practice and remove large aspects of the criminal practice.</s>
Label encoding: <s>Your ideas about prostitution aren't wrong. The problem is that criminalization hasn't solved these problems. There still are prostitutes in every city but as a result of criminalization they are more vulnerable. They can't seek assistance from violent clients, open certain types of bank accounts, or organize to defend their safety and interests. They can't do this because criminalization means that seeking help can harm them. [NEWLINE] [NEWLINE] Prostitution has to be legal to secure the safety of prostitutes and bring them out of the shadows. So long as it happens in the dark, the criminal element will reign and our ability to help women abused by this system is limited. If prostitution were legal then it would be a lot easier to access women in prostitution with aid. It would also mean that they could organize brothels so that they could hire security. It would mean that they could open bank accounts to save money to get out of the system. it would mean that laws could be imposed to mandate the wearing of condoms and STD checks (which would be a public health victory). It would mean that the criminal element is pushed about the legal industry. And it could also lead to prostitutes speaking out, organizing, and creating blacklists of persons who abuse them. [NEWLINE] [NEWLINE] Criminalization hasn't solved any of the issues that mentioned. In pretty much all cases, it has made it harder to help women abused by the system. Decriminalization is a path to bringing prostitution out of the dark, so that reforms can be made to make it a safer practice and remove large aspects of the criminal practice.</s>
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0803, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In a world<mask> poverty is a larger contributor to crime than in-utero exposure to cocaine,<mask><mask> you're wrong.  I make that point primarily to show that<mask> we expect and/or look for can often be wrong. [NEWLINE] [NEWLINE] You're saying that media violence is contributing to real violence,<mask><mask> you look at historical trends violence has pretty much always been a form of entertainment.  There's the Roman coliseum, dog fights, cock fights, and plenty of more examples I can't think of right now.  That does still exist,<mask> I would<mask><mask> video games provide a simulated release that allows us to express our natural violent urges without actually hurting anything.  Wars have gotten bigger,<mask> that's<mask> weapons have gotten stronger.  And in a way, that has held back war too.  The cold war would have happened, and would have torn apart the world,<mask> nuclear weapons hadn't been around to assure mutual destruction and scare everyone away from a fight. [NEWLINE] [NEWLINE] Going back to my starting point, the economic situation is probably the biggest factor.  I was lucky enough to grow up in a very wealthy family.  We had a house, and food, and a dog, and a computer, just about everything, compared to far too many people.  I grew up rich enough to study two martial arts for fun, not<mask> I needed them to survive or defend myself.  Having the opportunity to blow off steam, to act violent, without hurting anyone or anything, has allowed me to keep from hurting others and myself.</s>
Label encoding: <s>In a world where poverty is a larger contributor to crime than in-utero exposure to cocaine, I think you're wrong.  I make that point primarily to show that what we expect and/or look for can often be wrong. [NEWLINE] [NEWLINE] You're saying that media violence is contributing to real violence, but if you look at historical trends violence has pretty much always been a form of entertainment.  There's the Roman coliseum, dog fights, cock fights, and plenty of more examples I can't think of right now.  That does still exist, but I would argue that video games provide a simulated release that allows us to express our natural violent urges without actually hurting anything.  Wars have gotten bigger, but that's because weapons have gotten stronger.  And in a way, that has held back war too.  The cold war would have happened, and would have torn apart the world, if nuclear weapons hadn't been around to assure mutual destruction and scare everyone away from a fight. [NEWLINE] [NEWLINE] Going back to my starting point, the economic situation is probably the biggest factor.  I was lucky enough to grow up in a very wealthy family.  We had a house, and food, and a dog, and a computer, just about everything, compared to far too many people.  I grew up rich enough to study two martial arts for fun, not because I needed them to survive or defend myself.  Having the opportunity to blow off steam, to act violent, without hurting anyone or anything, has allowed me to keep from hurting others and myself.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Note that, based on this definition, changing my view on only one of the two qualifiers would change<mask><mask> on the matter,<mask> existential nihilism requires both to be true. [ENDQ] [NEWLINE] I am going to assume that you are a physicalist and work from there: [NEWLINE] [NEWLINE] [NEWLINE] Let us begin by pointing out that it is arguably true<mask> physicalism is true then there is no such thing<mask> free will,<mask> there is no causal agent or "I". This seems to fly in the face of<mask> seems to be common sense notions of moral ability and moral responsibility. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] <mask> my mental processes are totally determined, I am totally determined to accept determinism.<mask><mask> my sole reason for believing in X is that I am causally determined to believe it I have no ground for holding to the judgement that it is true or false. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] There must be a genuine enduring I in order for anyone to think.<mask> there is one self who reflects on the premise "<mask> p than q" a second self that reflects on the premise "p" and a third self that reflects on the conclusion "q" than there is no enduring self that actually thinks through process and draws the conclusion.<mask> there is something or someone who stands at the center of the experience that holds the terms and relations together in a stream of consciousness. [NEWLINE] [NEWLINE] [NEWLINE] There are things like the laws of logic that are non physical and have meaning and value, I beg of you to seek them out and not give up on meaning beyond the physical world.</s>
Label encoding: <s> [STARTQ] Note that, based on this definition, changing my view on only one of the two qualifiers would change my opinion on the matter, as existential nihilism requires both to be true. [ENDQ] [NEWLINE] I am going to assume that you are a physicalist and work from there: [NEWLINE] [NEWLINE] [NEWLINE] Let us begin by pointing out that it is arguably true if physicalism is true then there is no such thing as free will, since there is no causal agent or "I". This seems to fly in the face of what seems to be common sense notions of moral ability and moral responsibility. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] If my mental processes are totally determined, I am totally determined to accept determinism. But if my sole reason for believing in X is that I am causally determined to believe it I have no ground for holding to the judgement that it is true or false. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] There must be a genuine enduring I in order for anyone to think. If there is one self who reflects on the premise " if p than q" a second self that reflects on the premise "p" and a third self that reflects on the conclusion "q" than there is no enduring self that actually thinks through process and draws the conclusion. So there is something or someone who stands at the center of the experience that holds the terms and relations together in a stream of consciousness. [NEWLINE] [NEWLINE] [NEWLINE] There are things like the laws of logic that are non physical and have meaning and value, I beg of you to seek them out and not give up on meaning beyond the physical world.</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's<mask> I imagine it working: [NEWLINE] [NEWLINE] Our phones broadcast which laws we consent to abide, creating a digital-legal space surrounding us.<mask> cops intervene in a situation they scan everyone's legal permissions to get the bullet points (augmented reality overlays might be helpful). [NEWLINE] [NEWLINE] You would only have the right to be protected against theft<mask> you agreed not to steal. You can be legally murdered<mask> you do not agree to the law against it. A basic package of protections is granted to all children, which they can modify at the age of majority. [NEWLINE] [NEWLINE] Anyone can propose a law, and all laws should be written in everyday language. There would be a ranked list to choose from with lag time<mask> changing<mask> you can't game the system. [NEWLINE] [NEWLINE] A set of conventions would emerge bottom-up from the people, instead of being imposed top-down from authorities or majorities. It would be more fair<mask> you only have to live by rules you personally agree to, and more resistant to abuse by entrenched powers<mask><mask><mask>. It solves the old problem of "tyranny of the majority." And we would finally have a government able to adapt rapidly to change. [NEWLINE] [NEWLINE] EDIT: There would have to be a higher tier of law that prevents pollution, toxins in food, asymmetric laws, etc.<mask> the more of it we could get to work within this framework, the better government would be,<mask><mask> [NEWLINE] [NEWLINE] EDIT2: There would have to be reddit style voting system to keep the worst laws from being available.</s>
Label encoding: <s>Here's how I imagine it working: [NEWLINE] [NEWLINE] Our phones broadcast which laws we consent to abide, creating a digital-legal space surrounding us. When cops intervene in a situation they scan everyone's legal permissions to get the bullet points (augmented reality overlays might be helpful). [NEWLINE] [NEWLINE] You would only have the right to be protected against theft if you agreed not to steal. You can be legally murdered if you do not agree to the law against it. A basic package of protections is granted to all children, which they can modify at the age of majority. [NEWLINE] [NEWLINE] Anyone can propose a law, and all laws should be written in everyday language. There would be a ranked list to choose from with lag time when changing so you can't game the system. [NEWLINE] [NEWLINE] A set of conventions would emerge bottom-up from the people, instead of being imposed top-down from authorities or majorities. It would be more fair because you only have to live by rules you personally agree to, and more resistant to abuse by entrenched powers as a result. It solves the old problem of "tyranny of the majority." And we would finally have a government able to adapt rapidly to change. [NEWLINE] [NEWLINE] EDIT: There would have to be a higher tier of law that prevents pollution, toxins in food, asymmetric laws, etc. But the more of it we could get to work within this framework, the better government would be, IMO [NEWLINE] [NEWLINE] EDIT2: There would have to be reddit style voting system to keep the worst laws from being available.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is it REALLY discrimination? This is<mask><mask><mask> it gets muddy. Yes, she is treating one group differently from another based on sexuality.<mask> are straight people being treated unfairly? [ENDQ] [NEWLINE] I would think they are. Fairness is not being established here. One group is being preferred over another for a trait that is uncontrollable. Imagine<mask> a store gave a discount to white persons<mask> every other race has to pay regular price. We would think this isn't fair,<mask><mask> should this same thought process not apply to this case? Both are dealing with giving preferred or better service. [NEWLINE] [NEWLINE] [STARTQ] They can still hire her<mask> a photographer for the regular price. I guess you could say "They aren't getting it<mask> cheap"<mask> of course they aren't, that's<mask> discounts work she can't give it to all her customers.<mask> your friend put a coupon in a newspaper for one week would you say she was discriminating against people who don't get the paper? [ENDQ] [NEWLINE] I would say no,<mask> the paper is reasonably accessible to all<mask> being gay and being in a same-sex relationship is not. [NEWLINE] [NEWLINE] [STARTQ] I found one definition of discrimination that<mask><mask> gets at my point: "the treatment of a person or particular group of people differently, in a way that is worse than the way people are usually treated." [URL]. [ENDQ] [NEWLINE] Having to pay for regular price versus a discounted price simply<mask> of my sexual orientation does imply getting the more worse deal of the two<mask> I have to pay full price. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Is it REALLY discrimination? This is where I think it gets muddy. Yes, she is treating one group differently from another based on sexuality. But are straight people being treated unfairly? [ENDQ] [NEWLINE] I would think they are. Fairness is not being established here. One group is being preferred over another for a trait that is uncontrollable. Imagine if a store gave a discount to white persons while every other race has to pay regular price. We would think this isn't fair, so why should this same thought process not apply to this case? Both are dealing with giving preferred or better service. [NEWLINE] [NEWLINE] [STARTQ] They can still hire her as a photographer for the regular price. I guess you could say "They aren't getting it as cheap" but of course they aren't, that's how discounts work she can't give it to all her customers. If your friend put a coupon in a newspaper for one week would you say she was discriminating against people who don't get the paper? [ENDQ] [NEWLINE] I would say no, since the paper is reasonably accessible to all while being gay and being in a same-sex relationship is not. [NEWLINE] [NEWLINE] [STARTQ] I found one definition of discrimination that I think gets at my point: "the treatment of a person or particular group of people differently, in a way that is worse than the way people are usually treated." [URL]. [ENDQ] [NEWLINE] Having to pay for regular price versus a discounted price simply because of my sexual orientation does imply getting the more worse deal of the two if I have to pay full price. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you mistake the meaning of the Confederate flag,<mask> do many southerners.  I will try to keep to a few small points: 1) Slavery, 2) Rebellion, and 3) Sovereignty. [NEWLINE] 1) This is probably the most oft overlooked issue. <mask> you were to analyze the slave registries of the period, you would find that a large portion of the Union owned slaves<mask> well.  There were even many instances of blacks owning slaves.  With this in mind, the American flag is every bit a symbol of slavery<mask> the Confederate. [NEWLINE] 2) This is more the meat and bones of the symbol.  The Civil War was not fought over slavery, much like the Afghanistan War was not fought over anything we were told.  In both instances, one group tried to impede on another group's rights, and the smaller group responded violently.  The Confederate flag is a symbol for standing up against a bigger, badder entity.  It was intended to embrace true democracy<mask> big government is frowned upon.  It would be an appropriate flag for today's issues. [NEWLINE] 3) The Confederacy formed a separate country.  Countries are sovereign and should be left to their own devices.  Being that the Union was trying to change the Confederacy, there was an obvious breach of that sovereignty, which ultimately led to the war. [NEWLINE] [NEWLINE] TL:DR  The Confederate flag is a sign of rebellion against oppression and is not an advocate of slavery.  (I do acknowledge that there are groups that think otherwise)</s>
Label encoding: <s>I think you mistake the meaning of the Confederate flag, as do many southerners.  I will try to keep to a few small points: 1) Slavery, 2) Rebellion, and 3) Sovereignty. [NEWLINE] 1) This is probably the most oft overlooked issue.  If you were to analyze the slave registries of the period, you would find that a large portion of the Union owned slaves as well.  There were even many instances of blacks owning slaves.  With this in mind, the American flag is every bit a symbol of slavery as the Confederate. [NEWLINE] 2) This is more the meat and bones of the symbol.  The Civil War was not fought over slavery, much like the Afghanistan War was not fought over anything we were told.  In both instances, one group tried to impede on another group's rights, and the smaller group responded violently.  The Confederate flag is a symbol for standing up against a bigger, badder entity.  It was intended to embrace true democracy where big government is frowned upon.  It would be an appropriate flag for today's issues. [NEWLINE] 3) The Confederacy formed a separate country.  Countries are sovereign and should be left to their own devices.  Being that the Union was trying to change the Confederacy, there was an obvious breach of that sovereignty, which ultimately led to the war. [NEWLINE] [NEWLINE] TL:DR  The Confederate flag is a sign of rebellion against oppression and is not an advocate of slavery.  (I do acknowledge that there are groups that think otherwise)</s>
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] &gt; In Effect, making it legal to buy elections, eroding democracy [ENDQ] [NEWLINE] [STARTQ] That's a silly statement. [ENDQ] [NEWLINE] <mask> you might indulge my random musings for a second:<mask><mask> that people are upset about the idea of "money buying elections",<mask> they don't stop to seriously consider<mask> the alternative would be,<mask> you just blanket-banned that without exception. [NEWLINE] [NEWLINE] Right now, in the US political system, anyone willing to pay for it can get a message out. Let's say we put a strict, severe cap on political campaign funding (<mask> don't make any other structural changes). Now<mask> do we decide who gets messages out?<mask> do we decide who gets access to politicians. My fear is that,<mask> we were to do this, it would fall back to more nepotism and networking connections. [NEWLINE] [NEWLINE] <mask> a concrete example. With unlimited campaign funding, a TV station airs whichever ad pays them<mask> it costs to air an ad. With highly restricted campaign funding, a TV station *is going to take a loss* on any political ad they run,<mask> campaign finance restrictions prevent the raising of enough money to pay<mask> TV ads cost.<mask> which ads get run on the TV station. You better bet it's some combination of a) whichever ad happens to correspond to the station owner's personal views; and b) whichever ad happens to be run by someone who has familial or business connections to the station owner. [NEWLINE] [NEWLINE] Money keeps people honest, in this sense at least. </s>
Label encoding: <s> [STARTQ] &gt; In Effect, making it legal to buy elections, eroding democracy [ENDQ] [NEWLINE] [STARTQ] That's a silly statement. [ENDQ] [NEWLINE] If you might indulge my random musings for a second: I think that people are upset about the idea of "money buying elections", but they don't stop to seriously consider what the alternative would be, if you just blanket-banned that without exception. [NEWLINE] [NEWLINE] Right now, in the US political system, anyone willing to pay for it can get a message out. Let's say we put a strict, severe cap on political campaign funding ( but don't make any other structural changes). Now how do we decide who gets messages out? How do we decide who gets access to politicians. My fear is that, if we were to do this, it would fall back to more nepotism and networking connections. [NEWLINE] [NEWLINE] As a concrete example. With unlimited campaign funding, a TV station airs whichever ad pays them what it costs to air an ad. With highly restricted campaign funding, a TV station *is going to take a loss* on any political ad they run, because campaign finance restrictions prevent the raising of enough money to pay what TV ads cost. So which ads get run on the TV station. You better bet it's some combination of a) whichever ad happens to correspond to the station owner's personal views; and b) whichever ad happens to be run by someone who has familial or business connections to the station owner. [NEWLINE] [NEWLINE] Money keeps people honest, in this sense at least. </s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>During the first week alone [32 million people]( [URL] /) downloaded a copy of a Game of Thrones episode. Even<mask> only 1% of those people keep the file they downloaded after they watch it it's still an amazing 320k - it's a pretty safe bet that possibly hundreds of years from now we'll still be able to watch that show. [NEWLINE] At no point in history has it ever been easier to ensure archival and preservation due to digital mediums, and<mask> by applying copyright protections the content makers are doing nothing more than ensuring their great works might be lost to the ages. [NEWLINE] [NEWLINE] It's not too hard to believe something like this might be lost - only 46 years ago millions of people *watched* the moon landing on TV,<mask> almost no one had the technical means to have their own recording of it. Here we are now with the [original tapes missing]( [URL] ) - sure we have a recording,<mask> think of<mask> much better recordings we might have<mask> (like today) people could have made their own high-end home recordings at least matching the quality of the original broadcast. [NEWLINE] A very similar story can be said about the [missing Dr. Who epsiodes]( [URL] )  - of which [some were recovered thanks to piracy]( [URL] #Private_collectors) [NEWLINE] [NEWLINE] 500+ years from now -<mask><mask> the only copies of some uncommon films going to be some pirated copy taken out of a private archive, of which copyright protection had to be illegally circumvented to make.</s>
Label encoding: <s>During the first week alone [32 million people]( [URL] /) downloaded a copy of a Game of Thrones episode. Even if only 1% of those people keep the file they downloaded after they watch it it's still an amazing 320k - it's a pretty safe bet that possibly hundreds of years from now we'll still be able to watch that show. [NEWLINE] At no point in history has it ever been easier to ensure archival and preservation due to digital mediums, and yet by applying copyright protections the content makers are doing nothing more than ensuring their great works might be lost to the ages. [NEWLINE] [NEWLINE] It's not too hard to believe something like this might be lost - only 46 years ago millions of people *watched* the moon landing on TV, but almost no one had the technical means to have their own recording of it. Here we are now with the [original tapes missing]( [URL] ) - sure we have a recording, but think of how much better recordings we might have if (like today) people could have made their own high-end home recordings at least matching the quality of the original broadcast. [NEWLINE] A very similar story can be said about the [missing Dr. Who epsiodes]( [URL] )  - of which [some were recovered thanks to piracy]( [URL] #Private_collectors) [NEWLINE] [NEWLINE] 500+ years from now - I think the only copies of some uncommon films going to be some pirated copy taken out of a private archive, of which copyright protection had to be illegally circumvented to make.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What do they get in return?  Tons and tons of money. [NEWLINE] [NEWLINE] 1.  Speaking gigs.  Booking a president or former president to give a speech is insanely expensive. [NEWLINE] [NEWLINE] 2.  Books.  Every ex-president writes a book.  And then sells millions and millions of copies. [NEWLINE] [NEWLINE] 3.  Insider Trading.  You're the President.  You have inside, intimate knowledge of<mask> laws are going to be passed and<mask>, access to all the largest names in business and the ability to help them (<mask> they help you...), and even the ability to directly vote to your pocketbook.  [They obviously use this to their advantage<mask> judged by stock returns]( [URL] ) [NEWLINE] [NEWLINE] 4. Retirement pension for life.  They get their annual salary for life<mask> they retire.  Know of another job<mask> 4 years gets you lifetime pension? <mask> you include the cost of their "retirement plan" they're paid much, much more than 400k. [NEWLINE] [NEWLINE] 5.  Low Risk.  All this is a low risk position too. <mask> a businessman/CEO you have a risk to lose it all.  For every Bill Gates you see there's 100 other attempted Bill Gates who lost their shirt. <mask> a politician you can't lose everything. [NEWLINE] [NEWLINE] Basically the 400k they get annually is a minor, minor part of their actual income.  There's a reason these guys come to Washington wealthy and leave filthy rich.   </s>
Label encoding: <s>What do they get in return?  Tons and tons of money. [NEWLINE] [NEWLINE] 1.  Speaking gigs.  Booking a president or former president to give a speech is insanely expensive. [NEWLINE] [NEWLINE] 2.  Books.  Every ex-president writes a book.  And then sells millions and millions of copies. [NEWLINE] [NEWLINE] 3.  Insider Trading.  You're the President.  You have inside, intimate knowledge of what laws are going to be passed and when, access to all the largest names in business and the ability to help them ( if they help you...), and even the ability to directly vote to your pocketbook.  [They obviously use this to their advantage as judged by stock returns]( [URL] ) [NEWLINE] [NEWLINE] 4. Retirement pension for life.  They get their annual salary for life when they retire.  Know of another job where 4 years gets you lifetime pension?  If you include the cost of their "retirement plan" they're paid much, much more than 400k. [NEWLINE] [NEWLINE] 5.  Low Risk.  All this is a low risk position too.  As a businessman/CEO you have a risk to lose it all.  For every Bill Gates you see there's 100 other attempted Bill Gates who lost their shirt.  As a politician you can't lose everything. [NEWLINE] [NEWLINE] Basically the 400k they get annually is a minor, minor part of their actual income.  There's a reason these guys come to Washington wealthy and leave filthy rich.   </s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There was a really interesting article written by a women due to a recent change in hometown's change in dress code. And<mask><mask> she bring sup some fair points. [URL] [NEWLINE] [NEWLINE] * Clothing can be ambigious<mask> to<mask> is "slutty" (for lack of a better word). Dress pants can be just<mask> tight and revealing<mask> some yoga pants. [NEWLINE] *<mask> is considered appropriate has always changed with the times. My mom remembers a time<mask> she was not allowed to wear pants at school<mask> it was "inappropriate" [NEWLINE] * It teaches women to be ashamed of their bodies. It's THEIR fault that their classmates are unable to concentrate. That they are punished for showing too much skin. [NEWLINE] * It takes blame away from male students for their behavior. They whole "she was asking for it mentality", instead of teaching them to treat all women with respect equally,<mask><mask> the clothes they wear. [NEWLINE] [NEWLINE] To your other points, it's inappropriate for a teacher to ask out a student<mask> the teacher is in a position of power. It's similar to your boss asking you out. It's not appropriate. [NEWLINE] [NEWLINE] By shaming girls with dress codes, you<mask> don't teach the male students<mask> to interract with females they potentially find attractive. You don't get a choice of who you work with,<mask><mask> should they not learn to deal with their sexual feelings and remain professional. Again it's the girl's fault, not the boys for their inability to be professional<mask> they fancy someone?</s>
Label encoding: <s>There was a really interesting article written by a women due to a recent change in hometown's change in dress code. And I think she bring sup some fair points. [URL] [NEWLINE] [NEWLINE] * Clothing can be ambigious as to what is "slutty" (for lack of a better word). Dress pants can be just as tight and revealing as some yoga pants. [NEWLINE] * What is considered appropriate has always changed with the times. My mom remembers a time when she was not allowed to wear pants at school because it was "inappropriate" [NEWLINE] * It teaches women to be ashamed of their bodies. It's THEIR fault that their classmates are unable to concentrate. That they are punished for showing too much skin. [NEWLINE] * It takes blame away from male students for their behavior. They whole "she was asking for it mentality", instead of teaching them to treat all women with respect equally, regardless of the clothes they wear. [NEWLINE] [NEWLINE] To your other points, it's inappropriate for a teacher to ask out a student because the teacher is in a position of power. It's similar to your boss asking you out. It's not appropriate. [NEWLINE] [NEWLINE] By shaming girls with dress codes, you also don't teach the male students how to interract with females they potentially find attractive. You don't get a choice of who you work with, so why should they not learn to deal with their sexual feelings and remain professional. Again it's the girl's fault, not the boys for their inability to be professional because they fancy someone?</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When you ask<mask> much I know of sociology, I will admit very little.  I am halfway through an introductory course. <mask> far it has been fairly simple<mask> I have<mask> flipped through the three books we have for the class and I never saw something that really caught my attention in the sense of being a very intriguing concept that I felt a person could not conclude with little effort. <mask>,<mask><mask> I know this is an incredibly small sample size, our class of forty seems to come to the same conclusion. <mask> for your point regarding philosophy, I believe (at least<mask><mask> me) that many of the concepts discussed in it are not exactly intuitively obvious.  Again, I am no expert in philosophy<mask> I have had a class on it and read a few books regarding it for fun.  Specifically, I feel like it is far easier for someone to grasp the order theory of society versus something like one of the philosophical concepts of George Hegel. He believed philosophy was like a river in the sense that certain thoughts arose and were swept away and no one thought on the eternal condition of mankind could ever be correct.  I feel like that is harder to grasp then sociology's "people form societies<mask> they are dependent on each other for each person's unique ability (from the order model)".  And again I would like to state that it is definitely nice to have some books and experts in the field<mask> I don't believe sociology is<mask> useful it needs to be required for students diplomas.</s>
Label encoding: <s>When you ask how much I know of sociology, I will admit very little.  I am halfway through an introductory course.  So far it has been fairly simple but I have also flipped through the three books we have for the class and I never saw something that really caught my attention in the sense of being a very intriguing concept that I felt a person could not conclude with little effort.  Also, even though I know this is an incredibly small sample size, our class of forty seems to come to the same conclusion.  As for your point regarding philosophy, I believe (at least according to me) that many of the concepts discussed in it are not exactly intuitively obvious.  Again, I am no expert in philosophy but I have had a class on it and read a few books regarding it for fun.  Specifically, I feel like it is far easier for someone to grasp the order theory of society versus something like one of the philosophical concepts of George Hegel. He believed philosophy was like a river in the sense that certain thoughts arose and were swept away and no one thought on the eternal condition of mankind could ever be correct.  I feel like that is harder to grasp then sociology's "people form societies because they are dependent on each other for each person's unique ability (from the order model)".  And again I would like to state that it is definitely nice to have some books and experts in the field but I don't believe sociology is so useful it needs to be required for students diplomas.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well,<mask> you have to consider is<mask> these people are often considered the ultimate force of good and<mask> they save them in the first place. [NEWLINE] [NEWLINE] These aren't ordinary people like you and me. They are<mask><mask> insanely optimistic. Even batman<mask> you get right down to it. [NEWLINE] [NEWLINE] They save the worst of humanity every time they fight it. They try<mask> hard<mask> they can to save everyone including the villains<mask> against all traditional logic and reasoning, they still see something in those villains worth saving. Some spark of humanity. Some special thing. Or they just can't bear to see even their worst enemy suffer a death not caused by old age. [NEWLINE] [NEWLINE] Shit, forget killing. I have heard that the justice league even thinks it's unethical to wipe a villain's memory<mask> that'll fundamentally alter their personality in a drastic way and none of those heroes can bring themselves to force something like that on anyone, even the most sinister people on earth. [NEWLINE] [NEWLINE] That's<mask> we don't like it<mask> heroes straight up kill. Not only is this supposed to be an escape from reality,<mask> these heroes aren't normal, hopeless<mask> fuck human beings. Each one of them is special in quite huge ways and they just can't bring themselves to kill every villain they come across. And they feel<mask> they start killing, they'll start becoming like they people they're trying to stop. [NEWLINE] [NEWLINE] That's<mask> you only see heroes killing in SHITTY ASS GRITTY REBOOTS.</s>
Label encoding: <s>Well, what you have to consider is why these people are often considered the ultimate force of good and why they save them in the first place. [NEWLINE] [NEWLINE] These aren't ordinary people like you and me. They are in fact insanely optimistic. Even batman when you get right down to it. [NEWLINE] [NEWLINE] They save the worst of humanity every time they fight it. They try as hard as they can to save everyone including the villains because against all traditional logic and reasoning, they still see something in those villains worth saving. Some spark of humanity. Some special thing. Or they just can't bear to see even their worst enemy suffer a death not caused by old age. [NEWLINE] [NEWLINE] Shit, forget killing. I have heard that the justice league even thinks it's unethical to wipe a villain's memory because that'll fundamentally alter their personality in a drastic way and none of those heroes can bring themselves to force something like that on anyone, even the most sinister people on earth. [NEWLINE] [NEWLINE] That's why we don't like it when heroes straight up kill. Not only is this supposed to be an escape from reality, but these heroes aren't normal, hopeless as fuck human beings. Each one of them is special in quite huge ways and they just can't bring themselves to kill every villain they come across. And they feel if they start killing, they'll start becoming like they people they're trying to stop. [NEWLINE] [NEWLINE] That's why you only see heroes killing in SHITTY ASS GRITTY REBOOTS.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Have you ever personally felt a strong appreciation for a person's internal qualities and personality (male, female, whatever),<mask> well<mask> a concern for their happiness and well-being? [ENDQ] [NEWLINE] Can I include myself? [NEWLINE] And then yes for some of my friends,<mask> I decided it. [NEWLINE] [NEWLINE] [STARTQ] Whether or not you have personally felt this, does that preclude others from feeling it or from you feeling it in the future? [ENDQ] [NEWLINE] <mask>? Love? [NEWLINE] You expect me to answer something which haven't been defined? [NEWLINE] [NEWLINE] [STARTQ] <mask> would you explain acts of complete selflessness, such<mask> sacrificing yourself to save a family friend, or a lover?<mask> biological function would drive someone to cease their own life for the benefit of another? [ENDQ] [NEWLINE] Good point. [NEWLINE] These acts are emotionally driven ;The emotions can be manipulated : a person can make an other person act for her (the first) before herself(the second). [NEWLINE] It's my only logical answer<mask> you have other i'll be glad to hear it [NEWLINE] [NEWLINE] [STARTQ] Is self-reporting data not legitimate? I can tell you I'm feeling sad, do you automatically assume I am lying<mask> you cannot validate, or is it reasonable to assume that<mask> there is no contradicting evidence that we should believe people about their own reports of their internal thoughts/feelings? [ENDQ] [NEWLINE] Feelings are real<mask><mask> realty is subjective... [NEWLINE] You are just saying there are no proof? [NEWLINE] [NEWLINE] Carry on with a reasoning then! [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Have you ever personally felt a strong appreciation for a person's internal qualities and personality (male, female, whatever), as well as a concern for their happiness and well-being? [ENDQ] [NEWLINE] Can I include myself? [NEWLINE] And then yes for some of my friends, but I decided it. [NEWLINE] [NEWLINE] [STARTQ] Whether or not you have personally felt this, does that preclude others from feeling it or from you feeling it in the future? [ENDQ] [NEWLINE] What? Love? [NEWLINE] You expect me to answer something which haven't been defined? [NEWLINE] [NEWLINE] [STARTQ] How would you explain acts of complete selflessness, such as sacrificing yourself to save a family friend, or a lover? What biological function would drive someone to cease their own life for the benefit of another? [ENDQ] [NEWLINE] Good point. [NEWLINE] These acts are emotionally driven ;The emotions can be manipulated : a person can make an other person act for her (the first) before herself(the second). [NEWLINE] It's my only logical answer if you have other i'll be glad to hear it [NEWLINE] [NEWLINE] [STARTQ] Is self-reporting data not legitimate? I can tell you I'm feeling sad, do you automatically assume I am lying since you cannot validate, or is it reasonable to assume that if there is no contradicting evidence that we should believe people about their own reports of their internal thoughts/feelings? [ENDQ] [NEWLINE] Feelings are real but since realty is subjective... [NEWLINE] You are just saying there are no proof? [NEWLINE] [NEWLINE] Carry on with a reasoning then! [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry,<mask> that tree-ring study does absolutley nothing to refute anthropogenic climate change. Read this: [URL] [NEWLINE] [NEWLINE] "**'Our study doesn't go against anthropogenic global warming in any way,'** said Robert Wilson, a paleoclimatologist at the University of St. Andrews in Scotland and a co-author of the study"<mask> from that argicle: "Wilson, Schmidt and the vast majority of climate scientists agree: human-caused warming of the entire globe now overwhelms those subtle, regional heat redistributions." [NEWLINE] [NEWLINE] Currently there is scientific consensus that climate change is anthropogenic and extremely real.<mask> there may be some discussion<mask> to<mask> severe the effects of climate change are, there are no studies that effectively dismiss global warming. Here: [URL] is a good article that shows the arguments skeptics use simply don't hold up. [NEWLINE] [NEWLINE] From the article, "We carefully studied issues raised by skeptics: biases from urban heating (we duplicated our results using rural data alone), from data selection (prior groups selected fewer than 20 percent of the available temperature stations; we used virtually 100 percent), from poor station quality (we separately analyzed good stations and poor ones) and from human intervention and data adjustment (our work is completely automated and hands-off). **In our papers we demonstrate that none of these potentially troublesome effects unduly biased our conclusions**." [NEWLINE] [NEWLINE] I've<mask> to see a convincing argument<mask> to<mask> the earth isn't warming.</s>
Label encoding: <s>I'm sorry, but that tree-ring study does absolutley nothing to refute anthropogenic climate change. Read this: [URL] [NEWLINE] [NEWLINE] "**'Our study doesn't go against anthropogenic global warming in any way,'** said Robert Wilson, a paleoclimatologist at the University of St. Andrews in Scotland and a co-author of the study" Also from that argicle: "Wilson, Schmidt and the vast majority of climate scientists agree: human-caused warming of the entire globe now overwhelms those subtle, regional heat redistributions." [NEWLINE] [NEWLINE] Currently there is scientific consensus that climate change is anthropogenic and extremely real. While there may be some discussion as to how severe the effects of climate change are, there are no studies that effectively dismiss global warming. Here: [URL] is a good article that shows the arguments skeptics use simply don't hold up. [NEWLINE] [NEWLINE] From the article, "We carefully studied issues raised by skeptics: biases from urban heating (we duplicated our results using rural data alone), from data selection (prior groups selected fewer than 20 percent of the available temperature stations; we used virtually 100 percent), from poor station quality (we separately analyzed good stations and poor ones) and from human intervention and data adjustment (our work is completely automated and hands-off). **In our papers we demonstrate that none of these potentially troublesome effects unduly biased our conclusions**." [NEWLINE] [NEWLINE] I've yet to see a convincing argument as to why the earth isn't warming.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You still haven't really articulated *<mask> * it's good. You've articulated that it is good,<mask> now in<mask> way. [NEWLINE] [NEWLINE] You explained that you couldn't get EU citizenship or even permission to travel the schengen area, which is a shame,<mask> it's<mask> rather anecdotal. Furthermore, yeah, a lot of people won't like Stockholm just like they won't like Dallas and<mask> they like neither of these places then they'll just move on. In *both* cases they'll just move on<mask><mask> you moved to Stockholm then you'll have gotten your EU certificates and ability to rome freely and work freely. Just like<mask> you find yourself able to work in Dallas you'll be able to work in LA or Chicago. I don't see the difference, apart from in your rather unfortunate case. [NEWLINE] [NEWLINE] And yeah, the US gets a lot of immigrants. It gets more immigration than any other country in the world. That doesn't mean it's the best place for immigrants. It is indicative of one thing. That immigrants go there. It might be an excellent place for immigrants and it probably is. It almost certainly is,<mask> it's not indicative of it being the best place. It's indicative of it being a place which a lot of immigrants can and do make it to. A lot of immigrants<mask> enter the EU from all over in comparable (<mask> smaller) numbers than enter the USA,<mask> this doesn't mean it is the best for immigrants. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>You still haven't really articulated * how * it's good. You've articulated that it is good, but now in what way. [NEWLINE] [NEWLINE] You explained that you couldn't get EU citizenship or even permission to travel the schengen area, which is a shame, but it's also rather anecdotal. Furthermore, yeah, a lot of people won't like Stockholm just like they won't like Dallas and if they like neither of these places then they'll just move on. In *both* cases they'll just move on because if you moved to Stockholm then you'll have gotten your EU certificates and ability to rome freely and work freely. Just like if you find yourself able to work in Dallas you'll be able to work in LA or Chicago. I don't see the difference, apart from in your rather unfortunate case. [NEWLINE] [NEWLINE] And yeah, the US gets a lot of immigrants. It gets more immigration than any other country in the world. That doesn't mean it's the best place for immigrants. It is indicative of one thing. That immigrants go there. It might be an excellent place for immigrants and it probably is. It almost certainly is, but it's not indicative of it being the best place. It's indicative of it being a place which a lot of immigrants can and do make it to. A lot of immigrants also enter the EU from all over in comparable ( but smaller) numbers than enter the USA, but this doesn't mean it is the best for immigrants. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People are ignorant. Its just a fact.<mask><mask> democracy instills this idea in people that they are always right and that they can do anything. The fact is they cant. People are good at different things. Im not going to walk into a conference of biologists and demand that my completely uneducated opinion be respected there. People spend their entire lives studying things.<mask> i was in school i realized "the more i learn about this the less i know about this"<mask> i was hanging around doctors that knew things in such detail that it made my knowledge seem like a shallow survey of a deep ocean of knowledge.<mask> step one step outside of their expertise and they were right there with me in the shallow water.<mask> let people woth shallow understandings decide on policy that affects billions of people? [NEWLINE] [NEWLINE] Maybe your idea that layman have figured things out in a field could be a concession that is addressed in the constitution i mentioned. And i obviously wouldnt have anyone live under this without consent. [NEWLINE] [NEWLINE] <mask> that brings us to another important point: scale. A lot of other political philosophies, including anarchy, would work fine on a small scale.<mask> only 50 people are living together then communism is just fine. Go to certain music festivals and its complete voluntary anarchy for 5000 people and its completely fine.<mask> under a small scale, democracy isnt that much better than a lot of other philosophies. This has nothing to do with my other points<mask> it is important to keep in mind. </s>
Label encoding: <s>People are ignorant. Its just a fact. I think democracy instills this idea in people that they are always right and that they can do anything. The fact is they cant. People are good at different things. Im not going to walk into a conference of biologists and demand that my completely uneducated opinion be respected there. People spend their entire lives studying things. When i was in school i realized "the more i learn about this the less i know about this" because i was hanging around doctors that knew things in such detail that it made my knowledge seem like a shallow survey of a deep ocean of knowledge. But step one step outside of their expertise and they were right there with me in the shallow water. Why let people woth shallow understandings decide on policy that affects billions of people? [NEWLINE] [NEWLINE] Maybe your idea that layman have figured things out in a field could be a concession that is addressed in the constitution i mentioned. And i obviously wouldnt have anyone live under this without consent. [NEWLINE] [NEWLINE] But that brings us to another important point: scale. A lot of other political philosophies, including anarchy, would work fine on a small scale. If only 50 people are living together then communism is just fine. Go to certain music festivals and its complete voluntary anarchy for 5000 people and its completely fine. So under a small scale, democracy isnt that much better than a lot of other philosophies. This has nothing to do with my other points but it is important to keep in mind. </s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You said this: [NEWLINE] [NEWLINE] [STARTQ] I do not believe that the victim is at fault for the attackers crimes. [ENDQ] [NEWLINE] Which he agreed with,<mask> I do, and all reasonable people would. <mask> you<mask> said this: [NEWLINE] [NEWLINE] [STARTQ] I do not believe that the way a person dresses,<mask> they act, or<mask> much they drink contributes to them being sexually assaulted. [ENDQ] [NEWLINE] Which is actually patently false.  Getting blackout drunk<mask> alone in a shitty bar in a section of town the cops don't dare to go, for instance, absolutely will alter the probability of getting assaulted. [NEWLINE] [NEWLINE] <mask> most people will conclude that saying "<mask> a person does influences the outcome," is the same<mask> assigning some measure of blame.  And<mask> they'll shout "victim blaming!" [NEWLINE] [NEWLINE] <mask> of course you had to cover your ass and say that<mask> a person acts has no bearing on<mask> happens to them.  Or maybe you may do truly believe that. [NEWLINE] [NEWLINE] <mask> that is untrue.  The choices a person makes, do influence<mask> happens to them. [NEWLINE] [NEWLINE] That idea SHOULD not be offensive to anyone,<mask> there is a vast difference between events that are CAUSALLY related, and events that are MORALLY related. [NEWLINE] [NEWLINE] <mask> it seems most people cannot distinguish between the two, and view them<mask> identical, and<mask> any discussion at all of contributing factors is called victim blaming,<mask> the fact the person voicing such an opinion is not assigning any blame at all.</s>
Label encoding: <s>You said this: [NEWLINE] [NEWLINE] [STARTQ] I do not believe that the victim is at fault for the attackers crimes. [ENDQ] [NEWLINE] Which he agreed with, as I do, and all reasonable people would.  But you also said this: [NEWLINE] [NEWLINE] [STARTQ] I do not believe that the way a person dresses, how they act, or how much they drink contributes to them being sexually assaulted. [ENDQ] [NEWLINE] Which is actually patently false.  Getting blackout drunk while alone in a shitty bar in a section of town the cops don't dare to go, for instance, absolutely will alter the probability of getting assaulted. [NEWLINE] [NEWLINE] But most people will conclude that saying " what a person does influences the outcome," is the same as assigning some measure of blame.  And so they'll shout "victim blaming!" [NEWLINE] [NEWLINE] So of course you had to cover your ass and say that how a person acts has no bearing on what happens to them.  Or maybe you may do truly believe that. [NEWLINE] [NEWLINE] But that is untrue.  The choices a person makes, do influence what happens to them. [NEWLINE] [NEWLINE] That idea SHOULD not be offensive to anyone, because there is a vast difference between events that are CAUSALLY related, and events that are MORALLY related. [NEWLINE] [NEWLINE] But it seems most people cannot distinguish between the two, and view them as identical, and thus any discussion at all of contributing factors is called victim blaming, despite the fact the person voicing such an opinion is not assigning any blame at all.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The way Google has been handling the mobile market, ever<mask> Android became the dominant OS for cell phones, is really all you need to know to realize that Google's "Don't be evil" motto is a joke. [NEWLINE] [NEWLINE] [Here's]( [URL] ) a [few]( [URL] ) [examples]( [URL].com/en/articles/2013/april/rivals-claim-googles-deceptive-use-of-android-has-been-anti-competitive/) for you. I've saved the best for last: [Here you go]( [URL] /). [NEWLINE] [NEWLINE] A short search can provide more (I recommend [duckduckgo]( [URL] )). [NEWLINE] Essentially<mask> google are doing is stopping development on the open source versions of tools, one at a time, branching off a closed source branch, and only providing those improvements (which are some serious features, usually)<mask> you agree to play by Google's rules, and use everything they provide. For instance,<mask> you want to use Google Maps in its latest incarnation, rather than a seriously outdated open source version, you now<mask> have to accept that you can't make your own mail client for your phone, you have to use Google's. Google are abusing their market position for financial gain, and I'm sorry,<mask> that is *not cool*. [NEWLINE] [NEWLINE] EDIT: [NEWLINE] More links (not all about Android): [NEWLINE] [URL] [NEWLINE] [URL] [NEWLINE] [URL] / [NEWLINE] [URL] [NEWLINE] [URL] / [NEWLINE] [URL] /</s>
Label encoding: <s>The way Google has been handling the mobile market, ever since Android became the dominant OS for cell phones, is really all you need to know to realize that Google's "Don't be evil" motto is a joke. [NEWLINE] [NEWLINE] [Here's]( [URL] ) a [few]( [URL] ) [examples]( [URL].com/en/articles/2013/april/rivals-claim-googles-deceptive-use-of-android-has-been-anti-competitive/) for you. I've saved the best for last: [Here you go]( [URL] /). [NEWLINE] [NEWLINE] A short search can provide more (I recommend [duckduckgo]( [URL] )). [NEWLINE] Essentially what google are doing is stopping development on the open source versions of tools, one at a time, branching off a closed source branch, and only providing those improvements (which are some serious features, usually) if you agree to play by Google's rules, and use everything they provide. For instance, if you want to use Google Maps in its latest incarnation, rather than a seriously outdated open source version, you now also have to accept that you can't make your own mail client for your phone, you have to use Google's. Google are abusing their market position for financial gain, and I'm sorry, but that is *not cool*. [NEWLINE] [NEWLINE] EDIT: [NEWLINE] More links (not all about Android): [NEWLINE] [URL] [NEWLINE] [URL] [NEWLINE] [URL] / [NEWLINE] [URL] [NEWLINE] [URL] / [NEWLINE] [URL] /</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0845, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This really isn't medically accurate<mask>. There are all sorts of congenital intersex conditions. Take congenital androgen insensitivity syndrome,<mask> someone who is genetically male will develop essentially<mask> a normal female (except be unable to reproduce). [ENDQ] [NEWLINE] That makes them a male with a congenital defect, not a female. It doesn't change their gender anymore than Vitiligo changes someone's race. [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It's a very forced choice one has to make then. Choosing<mask> to respond behaviorally and medically to profound physical or hormonal abnormalities is a precondition to living a tolerably normal life, not a lifestyle decision. [ENDQ] [NEWLINE] Societal norms aren't Divine Mandate.<mask> you're talking about is someone making a choice to fit into society, not adherence to God's will, and that's perfectly fine- people are free to live<mask> they choose,<mask> again, it doesn't make it any less of a sin. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [STARTQ] One could<mask><mask> in cases of profound gender dysphoria, there's a disconnect between the way the brain and body have developed. You can even experimentally create gender dysphoria and doctors used to do it routinely in cases<mask> babies were born with ambiguous genitalia by arbitrarily assigning them a gender with disastrous consequences. [ENDQ] [NEWLINE] [NEWLINE] I'll ask<mask> again, "are you classifying someone being Transgender<mask> having a birth defect"?</s>
Label encoding: <s> [STARTQ] This really isn't medically accurate though. There are all sorts of congenital intersex conditions. Take congenital androgen insensitivity syndrome, where someone who is genetically male will develop essentially as a normal female (except be unable to reproduce). [ENDQ] [NEWLINE] That makes them a male with a congenital defect, not a female. It doesn't change their gender anymore than Vitiligo changes someone's race. [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It's a very forced choice one has to make then. Choosing how to respond behaviorally and medically to profound physical or hormonal abnormalities is a precondition to living a tolerably normal life, not a lifestyle decision. [ENDQ] [NEWLINE] Societal norms aren't Divine Mandate. What you're talking about is someone making a choice to fit into society, not adherence to God's will, and that's perfectly fine- people are free to live however they choose, but again, it doesn't make it any less of a sin. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [STARTQ] One could argue that in cases of profound gender dysphoria, there's a disconnect between the way the brain and body have developed. You can even experimentally create gender dysphoria and doctors used to do it routinely in cases where babies were born with ambiguous genitalia by arbitrarily assigning them a gender with disastrous consequences. [ENDQ] [NEWLINE] [NEWLINE] I'll ask yet again, "are you classifying someone being Transgender as having a birth defect"?</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We live in a country of 300 million people with nearly 300 million guns,<mask> do we do about that? Do we disarm everyone? Do we force everyone to register them or lock them in a local law enforcement vault? This would undoubtedly lead to many tens of thousands of deaths. [ENDQ] [NEWLINE] Well, the rest of the world is a pretty good indicator of<mask> you guys can start.  Offer free disposal services, licence for ownership, registration for each gun, heavy fines for non-compliance. [NEWLINE] [NEWLINE] <mask> on earth would that lead to tens of thousands of deaths? [NEWLINE] [NEWLINE] [STARTQ] Gun sales increase every year, and "gun killings" have decreased 39%<mask> 1993. Not to mention "Gun crimes that weren’t fatal fell by 69%" [ENDQ] [NEWLINE] You missed an important line there "paralleling a broader drop in violent crimes committed with or without guns."  In other words, crime is dropping,<mask><mask> the involvement of guns. [NEWLINE] [NEWLINE] [STARTQ] Not to mention that<mask> accounting for race, even with our guns, we have a rate of gun violence equal or lower than certain parts of Europe. Europe of course being a happier, more homogenous place for the most part. [ENDQ] [NEWLINE] Erm, which ones? <mask><mask> [this]( [URL] ), the closest European country in terms of firearm related violence is Greece, and that's still less than half of the rate in the US.  And hardly a happer, more homogenous place...</s>
Label encoding: <s> [STARTQ] We live in a country of 300 million people with nearly 300 million guns, what do we do about that? Do we disarm everyone? Do we force everyone to register them or lock them in a local law enforcement vault? This would undoubtedly lead to many tens of thousands of deaths. [ENDQ] [NEWLINE] Well, the rest of the world is a pretty good indicator of where you guys can start.  Offer free disposal services, licence for ownership, registration for each gun, heavy fines for non-compliance. [NEWLINE] [NEWLINE] How on earth would that lead to tens of thousands of deaths? [NEWLINE] [NEWLINE] [STARTQ] Gun sales increase every year, and "gun killings" have decreased 39% since 1993. Not to mention "Gun crimes that weren’t fatal fell by 69%" [ENDQ] [NEWLINE] You missed an important line there "paralleling a broader drop in violent crimes committed with or without guns."  In other words, crime is dropping, regardless of the involvement of guns. [NEWLINE] [NEWLINE] [STARTQ] Not to mention that when accounting for race, even with our guns, we have a rate of gun violence equal or lower than certain parts of Europe. Europe of course being a happier, more homogenous place for the most part. [ENDQ] [NEWLINE] Erm, which ones?  According to [this]( [URL] ), the closest European country in terms of firearm related violence is Greece, and that's still less than half of the rate in the US.  And hardly a happer, more homogenous place...</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't quite see<mask> those two things are contradictory. Humans choose their ideologies in accordance with their benefit. Would you not agree that it's far rarer to see a socialist billionaire? Or a capitalist miner? And yes, those exist,<mask> the former<mask> the billionaire has been conditioned either by evolution or his upbringing to take pleasure in acts that are selfless or pro-equality and the latter<mask> the miner either is a masochist or believes that the world is supposed to be that way and enforcing/accepting this reality makes him feel better or<mask> the miner believes in karma and reincarnation. [NEWLINE] [NEWLINE] <mask>, even<mask> this statement doesn't satisfy the condition in your first paragraph, I'll restate the miner hypothesis, examplify it with American conservativism amongst the lower class white men, and ask you<mask> these people are acting in self interest in your terms<mask> opposed to the tautology you provided. [NEWLINE] [NEWLINE] My thesis<mask> is that people choose<mask> we might call ideology on the basis of<mask> pleases them, be it pleasurable in their direct self-interest or due to evolutionary/upbringing-related(religion/culture) absurdity. This will occasionally contradict their direct self-interest. Example already provided above; supporting argument is that a world<mask> everyone acted only to further their own material gains and comfort (and of their kin) is a world<mask> Ayn Rand would not have thought he needed to write a book.</s><pad>
Label encoding: <s>I don't quite see how those two things are contradictory. Humans choose their ideologies in accordance with their benefit. Would you not agree that it's far rarer to see a socialist billionaire? Or a capitalist miner? And yes, those exist, but the former because the billionaire has been conditioned either by evolution or his upbringing to take pleasure in acts that are selfless or pro-equality and the latter because the miner either is a masochist or believes that the world is supposed to be that way and enforcing/accepting this reality makes him feel better or because the miner believes in karma and reincarnation. [NEWLINE] [NEWLINE] So, even if this statement doesn't satisfy the condition in your first paragraph, I'll restate the miner hypothesis, examplify it with American conservativism amongst the lower class white men, and ask you how these people are acting in self interest in your terms as opposed to the tautology you provided. [NEWLINE] [NEWLINE] My thesis therefore is that people choose what we might call ideology on the basis of what pleases them, be it pleasurable in their direct self-interest or due to evolutionary/upbringing-related(religion/culture) absurdity. This will occasionally contradict their direct self-interest. Example already provided above; supporting argument is that a world where everyone acted only to further their own material gains and comfort (and of their kin) is a world where Ayn Rand would not have thought he needed to write a book.</s><pad>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You shouldn't be in here.<mask> an analogy is going to get you this riled up, it seems that you will only serve to get yourself offended.<mask><mask> r/aww might be a better place. Analogies are supposed to represent a part of an argument in order to represent an idea, not be the direct stand in for whatever is being discussed. There will always be differences, and<mask> that is all you are going to pick apart rather than understand that it isnt<mask>'s being said, you lose any ability to bring anything meaningful to the conversation. [ENDQ] [NEWLINE] Analogies should at least mostly fit. The details are<mask> makes the analogy too.<mask> YOU can't deal with that then you need to re-attend English 101 and learn<mask> to form a functioning analogy<mask> you clearly missed that week of study. [NEWLINE] [NEWLINE] [STARTQ] That being said,<mask><mask> this idea with the cakes is a very complex one in the same manner that op does. We have too many people deciding that their way is the only acceptable one, and<mask> they are pushing beliefs, religious or otherwise on everyone else. Everyone pretends to take the moral high ground,<mask> have never taken a step back to see that they are being hypocrites. [ENDQ] [NEWLINE] It doesn't make you a hypocrite to believe that people shouldn't be allowed to oppress people. We're not talking about difference of opinion here. We're talking about entire groups of people being marginalized.</s>
Label encoding: <s> [STARTQ] You shouldn't be in here. If an analogy is going to get you this riled up, it seems that you will only serve to get yourself offended. I think r/aww might be a better place. Analogies are supposed to represent a part of an argument in order to represent an idea, not be the direct stand in for whatever is being discussed. There will always be differences, and if that is all you are going to pick apart rather than understand that it isnt what's being said, you lose any ability to bring anything meaningful to the conversation. [ENDQ] [NEWLINE] Analogies should at least mostly fit. The details are what makes the analogy too. If YOU can't deal with that then you need to re-attend English 101 and learn how to form a functioning analogy because you clearly missed that week of study. [NEWLINE] [NEWLINE] [STARTQ] That being said, I think this idea with the cakes is a very complex one in the same manner that op does. We have too many people deciding that their way is the only acceptable one, and thus they are pushing beliefs, religious or otherwise on everyone else. Everyone pretends to take the moral high ground, but have never taken a step back to see that they are being hypocrites. [ENDQ] [NEWLINE] It doesn't make you a hypocrite to believe that people shouldn't be allowed to oppress people. We're not talking about difference of opinion here. We're talking about entire groups of people being marginalized.</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, you can go into<mask> -<mask>,<mask> -<mask> all day. Obviously, it's generally not a good idea to argue on the basis of "Your honor, the defense council is a doodoo-head." Yours wasn't exactly a contentious original statement,<mask> I'm trying to give an example of<mask> insults could potentially be advantageous. Having an established beef with someone can be a good PR move.<mask> you try to get a good volley going and they come back with something that puts you to shame, it's not going to work, and you'll look like an asshole.<mask> they refuse to play the game and call you out on using ad hominem, you'll look like an asshole.<mask>,<mask> both of you go into the discussion with a good sense of humor and some sort of mutual understanding, you might be able to benefit from having a rivalry of sorts. Obviously you shouldn't insult everyone who you debate against publicly in hopes that a spirited rivalry will arise,<mask> in some situations, it can help both of you. Jon Stewart and Bill O'Reilly, Biggie and Tupac, and trash talk in boxing / pro wrestling are examples<mask> insults help both sides.<mask> you're morally opposed to insulting someone<mask> you think it's childish, then don't do it,<mask> given the right context, insults can get more people interested in a discussion, which can be helpful<mask> getting people to care is a major objective.</s>
Label encoding: <s>Well, you can go into what - if, what - if all day. Obviously, it's generally not a good idea to argue on the basis of "Your honor, the defense council is a doodoo-head." Yours wasn't exactly a contentious original statement, but I'm trying to give an example of where insults could potentially be advantageous. Having an established beef with someone can be a good PR move. If you try to get a good volley going and they come back with something that puts you to shame, it's not going to work, and you'll look like an asshole. If they refuse to play the game and call you out on using ad hominem, you'll look like an asshole. But, if both of you go into the discussion with a good sense of humor and some sort of mutual understanding, you might be able to benefit from having a rivalry of sorts. Obviously you shouldn't insult everyone who you debate against publicly in hopes that a spirited rivalry will arise, but in some situations, it can help both of you. Jon Stewart and Bill O'Reilly, Biggie and Tupac, and trash talk in boxing / pro wrestling are examples where insults help both sides. If you're morally opposed to insulting someone because you think it's childish, then don't do it, but given the right context, insults can get more people interested in a discussion, which can be helpful if getting people to care is a major objective.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0724, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A zero-sum game is a situation in which a participant's gain (or loss)  is exactly balanced by the losses (or gains)  of the other participant. For exemple, a soccer game is a zero sum game:<mask> a team wins the other team loses. [NEWLINE] [NEWLINE] <mask><mask>, non-zero-sum describes a situation in which the interacting parties' aggregate gains and losses are either less than or more than zero. [NEWLINE] [NEWLINE] Many worldwide human endeavours rely on cooperation (non-zero sum games) like  scientific and technological development. A sports event<mask> everyone roots for the same outcome is impossible. [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>A zero-sum game is a situation in which a participant's gain (or loss)  is exactly balanced by the losses (or gains)  of the other participant. For exemple, a soccer game is a zero sum game: when a team wins the other team loses. [NEWLINE] [NEWLINE] In contrast, non-zero-sum describes a situation in which the interacting parties' aggregate gains and losses are either less than or more than zero. [NEWLINE] [NEWLINE] Many worldwide human endeavours rely on cooperation (non-zero sum games) like  scientific and technological development. A sports event where everyone roots for the same outcome is impossible. [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is pretty troubling<mask> careless and irresponsible a lot of companies and individuals are with the personal data they are trusted with. I wish more was done about holding them accountable<mask> things like that happen. I will never use facebook and am very careful about<mask> much of my personal information I put online. I do my best to keep all my devices and all my online accounts secure from unauthorized access.<mask> at the end of the day security breaches happen no matter<mask> careful we are, that's just something we all have to live with and stay mindful of I guess. [NEWLINE] [NEWLINE] Here is the thing about<mask> the gov. is up to<mask> : with all this data they are storing, even<mask> right now they don't have time or resources to single out individual people, doesn't mean that will always be the case. You never know<mask> laws and technology are going to change in the near future. [NEWLINE] [NEWLINE] They might not be able to legally look into individual people's activities,<mask> OTOH I wouldn't have thought they would be able to demand access to ALL of Verizon's phone calls either for example.<mask><mask> tomorrow these agencies decide this is<mask> they need to start doing in order to protect us from terrorism or whatever?<mask> things are all done in secret,<mask>'s stopping them? I wouldn't put it past those in power to implement something like this<mask> they had the capability and a secret OK from some judge. It's a slippery slope.</s>
Label encoding: <s>It is pretty troubling how careless and irresponsible a lot of companies and individuals are with the personal data they are trusted with. I wish more was done about holding them accountable when things like that happen. I will never use facebook and am very careful about how much of my personal information I put online. I do my best to keep all my devices and all my online accounts secure from unauthorized access. But at the end of the day security breaches happen no matter how careful we are, that's just something we all have to live with and stay mindful of I guess. [NEWLINE] [NEWLINE] Here is the thing about what the gov. is up to though : with all this data they are storing, even if right now they don't have time or resources to single out individual people, doesn't mean that will always be the case. You never know how laws and technology are going to change in the near future. [NEWLINE] [NEWLINE] They might not be able to legally look into individual people's activities, but OTOH I wouldn't have thought they would be able to demand access to ALL of Verizon's phone calls either for example. What if tomorrow these agencies decide this is what they need to start doing in order to protect us from terrorism or whatever? When things are all done in secret, what's stopping them? I wouldn't put it past those in power to implement something like this if they had the capability and a secret OK from some judge. It's a slippery slope.</s>
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I wrote an earlier post,<mask><mask><mask> this still misses<mask> much of<mask> the government does isn't determined by the Constitution,<mask> by the statutes and cases that have accumulated over time. [NEWLINE] [NEWLINE] The First Amendment doesn't *necessarily* protect campaign spending<mask> speech, or make that speech paramount over the Congress's Constitutional authority to reasonably regulate elections.  The Supreme Court made that choice in 1976.  (Buckley v. Valeo.) [NEWLINE] [NEWLINE] The Second Amendment doesn't *necessarily* give an individual right to personal firearms.  The Supreme Court decided this in 2008. (DC v. Heller.) [NEWLINE] [NEWLINE] The Fourth Amendment doesn't *necessarily* give the government the right to intercept electronic documents in the hands of third parties.  The Congress has just failed to draft an exception to common law principles of waiver that existed before the U.S. was formed (which they applied to, eg, mail couriers.) [NEWLINE] [NEWLINE] The U.S. doesn't lack universal healthcare<mask> of the 10th Amendment -- there is a lack of political will.  The U.S. unquestionably has the right to levy taxes and use the taxes to pay for universal healthcare; we don't have that<mask> not enough people want it strongly enough. [NEWLINE] [NEWLINE] In all these cases, the issue isn't the skeletal framework of the Constitution, it's the accumulation of laws and precedents that have come in the 250 years<mask>.  </s>
Label encoding: <s>I wrote an earlier post, but I think this still misses how much of what the government does isn't determined by the Constitution, but by the statutes and cases that have accumulated over time. [NEWLINE] [NEWLINE] The First Amendment doesn't *necessarily* protect campaign spending as speech, or make that speech paramount over the Congress's Constitutional authority to reasonably regulate elections.  The Supreme Court made that choice in 1976.  (Buckley v. Valeo.) [NEWLINE] [NEWLINE] The Second Amendment doesn't *necessarily* give an individual right to personal firearms.  The Supreme Court decided this in 2008. (DC v. Heller.) [NEWLINE] [NEWLINE] The Fourth Amendment doesn't *necessarily* give the government the right to intercept electronic documents in the hands of third parties.  The Congress has just failed to draft an exception to common law principles of waiver that existed before the U.S. was formed (which they applied to, eg, mail couriers.) [NEWLINE] [NEWLINE] The U.S. doesn't lack universal healthcare because of the 10th Amendment -- there is a lack of political will.  The U.S. unquestionably has the right to levy taxes and use the taxes to pay for universal healthcare; we don't have that because not enough people want it strongly enough. [NEWLINE] [NEWLINE] In all these cases, the issue isn't the skeletal framework of the Constitution, it's the accumulation of laws and precedents that have come in the 250 years since.  </s>
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No offense,<mask> I totally disagree with you about less intelligent animals feeling less suffering. For example, say a child has a low iq and developmental disabilities; does that mean<mask> you cut their arm off they dont feel the same pain<mask> you or i do? [NEWLINE] [NEWLINE] I might be misunderstanding,<mask> is your general argument that our breeding choices have to do with the nature of the animal(s)?<mask><mask>,<mask><mask> it's unfair to say that cows lack empathy or emotional compatibility. In western societies few of us are even around cows and<mask> we've been born we've been socialized to believe cows are food and dogs are pets.<mask> take for example india - my parents friends all used to keep cows for milk and developed bonds with them in the same way you would any pet. And they werent the mainstream religion,<mask> you cant use that<mask> an excuse. [NEWLINE] [NEWLINE] <mask><mask> its quite clear that all animals feel pain. Watch a dairy cow have her newborn calf taken away for becoming veal and she'll start making loud disgruntling noises. Whether they openly share that same connection with humans doesnt matter. Before the chopping block,<mask><mask> animal, they all feel pain. And i would like to think that none less than the other. [NEWLINE] [NEWLINE] I apologize<mask> i said anything to offend you, or<mask> my spelling sucks or this is choppy. Wrote it on my phone. Haha just<mask><mask> :)</s>
Label encoding: <s>No offense, but I totally disagree with you about less intelligent animals feeling less suffering. For example, say a child has a low iq and developmental disabilities; does that mean when you cut their arm off they dont feel the same pain as you or i do? [NEWLINE] [NEWLINE] I might be misunderstanding, but is your general argument that our breeding choices have to do with the nature of the animal(s)? If so, I think it's unfair to say that cows lack empathy or emotional compatibility. In western societies few of us are even around cows and since we've been born we've been socialized to believe cows are food and dogs are pets. But take for example india - my parents friends all used to keep cows for milk and developed bonds with them in the same way you would any pet. And they werent the mainstream religion, so you cant use that as an excuse. [NEWLINE] [NEWLINE] I think its quite clear that all animals feel pain. Watch a dairy cow have her newborn calf taken away for becoming veal and she'll start making loud disgruntling noises. Whether they openly share that same connection with humans doesnt matter. Before the chopping block, regardless of animal, they all feel pain. And i would like to think that none less than the other. [NEWLINE] [NEWLINE] I apologize if i said anything to offend you, or if my spelling sucks or this is choppy. Wrote it on my phone. Haha just my opinion :)</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It seems to me your chances of getting the flu are higher WITH the shot<mask> you are absolutely exposing yourself to the flu versus maybe not being exposed at all. [ENDQ] [NEWLINE] Well,<mask><mask> the [CDC]( [URL] ) you can't get sick from the flu shot<mask> "Flu vaccines that are administered with a needle are currently made in two ways: the vaccine is made either with a) flu vaccine viruses that have been ‘inactivated’ and are<mask> not infectious, or b) with no flu vaccine viruses at all (which is the case for recombinant influenza vaccine). In randomized, blinded studies,<mask> some people got flu shots and others got saltwater shots, the only differences in symptoms was increased soreness in the arm and redness at the injection site among people who got the flu shot. There were no differences in terms of body aches, fever, cough, runny nose or sore throat."<mask><mask> you get sick after getting a flu shot, it's not<mask> the shot gave you the flu. It's<mask> your immune system was weakened by the shot and you were exposed to something else (maybe even another strain of the flu that wasn't covered by the vaccine) which was able to attack your immune system in its weakened state.<mask>,<mask> you're not exposed to the flu<mask> you don't get the vaccine, then<mask> do people who don't get the vaccine catch it? </s>
Label encoding: <s> [STARTQ] It seems to me your chances of getting the flu are higher WITH the shot since you are absolutely exposing yourself to the flu versus maybe not being exposed at all. [ENDQ] [NEWLINE] Well, according to the [CDC]( [URL] ) you can't get sick from the flu shot because "Flu vaccines that are administered with a needle are currently made in two ways: the vaccine is made either with a) flu vaccine viruses that have been ‘inactivated’ and are therefore not infectious, or b) with no flu vaccine viruses at all (which is the case for recombinant influenza vaccine). In randomized, blinded studies, where some people got flu shots and others got saltwater shots, the only differences in symptoms was increased soreness in the arm and redness at the injection site among people who got the flu shot. There were no differences in terms of body aches, fever, cough, runny nose or sore throat." So if you get sick after getting a flu shot, it's not because the shot gave you the flu. It's because your immune system was weakened by the shot and you were exposed to something else (maybe even another strain of the flu that wasn't covered by the vaccine) which was able to attack your immune system in its weakened state. Also, if you're not exposed to the flu if you don't get the vaccine, then how do people who don't get the vaccine catch it? </s>
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I read an excellent argument this morning which has really put a lot of doubt in my mind about the use and demand for trigger warnings. [URL] [NEWLINE] [NEWLINE] It's pretty long and goes into a whole lot more than trigger warning, gender politics and micro aggressions and racial insensitivity.<mask><mask><mask> it might change your view. And then I guess you'd have to send a delta to the author, haha. [NEWLINE] [NEWLINE] My personal philosophy is summed up in this proverb I'll paraphrase: "A foolish man walks out of his house and steps on a thorn. Angry that he's in pain, he demands that this is unacceptable and the whole world should be covered in a layer of leather. A wise man walks out of his house and steps on a thorn. He takes leather and attaches it to his feet<mask> that he's protected from thorns wherever he goes." [NEWLINE] [NEWLINE] I don't think it's the world's responsibility to make you not feel bad. Disasters are going to happen, people are going to rob stores, people are going to die, rape is going to make the news, planes are going to crash. The world has some tough shit in it. You need to make yourself some shoes and figure out<mask> you're going to deal with the world. [NEWLINE] [NEWLINE] <mask>, avoidance is not the best way to overcome trauma. Avoidance is a symptom of ptsd, it's not a good coping mechanism. </s><pad>
Label encoding: <s>I read an excellent argument this morning which has really put a lot of doubt in my mind about the use and demand for trigger warnings. [URL] [NEWLINE] [NEWLINE] It's pretty long and goes into a whole lot more than trigger warning, gender politics and micro aggressions and racial insensitivity. But I think it might change your view. And then I guess you'd have to send a delta to the author, haha. [NEWLINE] [NEWLINE] My personal philosophy is summed up in this proverb I'll paraphrase: "A foolish man walks out of his house and steps on a thorn. Angry that he's in pain, he demands that this is unacceptable and the whole world should be covered in a layer of leather. A wise man walks out of his house and steps on a thorn. He takes leather and attaches it to his feet so that he's protected from thorns wherever he goes." [NEWLINE] [NEWLINE] I don't think it's the world's responsibility to make you not feel bad. Disasters are going to happen, people are going to rob stores, people are going to die, rape is going to make the news, planes are going to crash. The world has some tough shit in it. You need to make yourself some shoes and figure out how you're going to deal with the world. [NEWLINE] [NEWLINE] Also, avoidance is not the best way to overcome trauma. Avoidance is a symptom of ptsd, it's not a good coping mechanism. </s><pad>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As far as technological advancements go the porn industry really helped out [digital cameras]( [URL] ;q=cache:LZwhuhLAt3kJ:webpages.dcu.ie/~mcdonpi/30-digital.pdf+sex+in+consumer+culture&amp;hl=pl&amp;pid=bl&amp;srcid=ADGEESghw9nhnGfWnsQyTGk0mpFdEAbw1WvDoqE7w8sO0FJi34ELk6vHMu_sEhpYHEzrRTN8XJKqIt4FSyx_I5mYR-bhD4hzJbdddlnyCqMDh1usRGFx8XfFbaDlPHOz1At3l46gkSgM&amp;sig=AHIEtbT-w65cQhsYtalZWg5kdBAHw-EtTw)  and is normally at the front of any [new technology]( [URL].technology/index.html?hpt=Sbin) and IIRC one of the first books to be widely distributed by the printing press was basically medieval porn.  [google glass]( [URL] ) even had a porn app by the weekend it was released.  </s>
Label encoding: <s>As far as technological advancements go the porn industry really helped out [digital cameras]( [URL] ;q=cache:LZwhuhLAt3kJ:webpages.dcu.ie/~mcdonpi/30-digital.pdf+sex+in+consumer+culture&amp;hl=pl&amp;pid=bl&amp;srcid=ADGEESghw9nhnGfWnsQyTGk0mpFdEAbw1WvDoqE7w8sO0FJi34ELk6vHMu_sEhpYHEzrRTN8XJKqIt4FSyx_I5mYR-bhD4hzJbdddlnyCqMDh1usRGFx8XfFbaDlPHOz1At3l46gkSgM&amp;sig=AHIEtbT-w65cQhsYtalZWg5kdBAHw-EtTw)  and is normally at the front of any [new technology]( [URL].technology/index.html?hpt=Sbin) and IIRC one of the first books to be widely distributed by the printing press was basically medieval porn.  [google glass]( [URL] ) even had a porn app by the weekend it was released.  </s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> throws me is:<mask> does all that gender reinforcement not take? [ENDQ] [NEWLINE] We're not sure. Children's identities tend to form by "the age of reason," which is about 5 or 6. This corresponds with many stories of transgender people knowing they were different/male/female at a very young age.<mask> with all biological processes, this isn't an overnight realization,<mask> it's not a stretch to presume these children experienced a lot of confusion prior to and then after recognizing their gender and it's opposition with their sex and the group they're told/expected to associate with. [NEWLINE] [NEWLINE] <mask> your statement reveals<mask> is a belief that children are not autonomous human beings, that they and their personality is completely dictated by their parents and environment. By extension, we could assume that all homosexual and bisexual people are<mask><mask> their parents "trained them to be gay." That you only have your sexuality and your gender identity<mask> it was engrained in you. [NEWLINE] [NEWLINE] This relates black to the previous redditor's fantastic description of<mask> it's like growing up with gender dysphoria.<mask>, on some level,<mask> you were correct, the only reason you identify<mask> male is<mask> every morning you wake up and recognize that you have a penis which prompts you to recall all of your conditioning that you are male. Instead, your environment only reinforced your already held gender identity which is the opposite of<mask> transgender people experience.</s>
Label encoding: <s> [STARTQ] What throws me is: how does all that gender reinforcement not take? [ENDQ] [NEWLINE] We're not sure. Children's identities tend to form by "the age of reason," which is about 5 or 6. This corresponds with many stories of transgender people knowing they were different/male/female at a very young age. As with all biological processes, this isn't an overnight realization, so it's not a stretch to presume these children experienced a lot of confusion prior to and then after recognizing their gender and it's opposition with their sex and the group they're told/expected to associate with. [NEWLINE] [NEWLINE] What your statement reveals though is a belief that children are not autonomous human beings, that they and their personality is completely dictated by their parents and environment. By extension, we could assume that all homosexual and bisexual people are so because their parents "trained them to be gay." That you only have your sexuality and your gender identity because it was engrained in you. [NEWLINE] [NEWLINE] This relates black to the previous redditor's fantastic description of what it's like growing up with gender dysphoria. But, on some level, if you were correct, the only reason you identify as male is because every morning you wake up and recognize that you have a penis which prompts you to recall all of your conditioning that you are male. Instead, your environment only reinforced your already held gender identity which is the opposite of what transgender people experience.</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Only by switching to the artificial Standard. I actually have a hard time speaking with someone from London or from anywhere in the south that ain't Columbia or Charleston, SC and I've had a lot of other English speakers question my own speech<mask> of my own dialect (southern influenced New York English), I don't speak in my natural dialect<mask> talking to coworkers or people from outta the city, only my close friends or family'll hear me talk<mask> natural, and that's the thing, humans are good at code switching<mask>, given enough time, the differences will've become<mask> great that even that won't work anymore. [NEWLINE] [NEWLINE] The time-depth isn't<mask> there, American English started diverging in the 17th century and converged again during the Georgian and Victorian eras (<mask> non-rhoticity became a feature of Coastal Southern speech, New England speech &amp; New York speech (New York speech could've been considered the standard for American English up till WWII<mask> prestige shifted toward the mid-west). [NEWLINE] [NEWLINE] Give it a couple generations<mask> it seems the Englishes are diverging again rather rapidly (the break-off of California &amp; Pacific Northwestern English from the Western group springs to mind, especially Californian English phonemics with it's reintroduction of front-rounded vowels (not seen in a standard English<mask> West-Saxon (Wessexian) Old English).</s>
Label encoding: <s>Only by switching to the artificial Standard. I actually have a hard time speaking with someone from London or from anywhere in the south that ain't Columbia or Charleston, SC and I've had a lot of other English speakers question my own speech because of my own dialect (southern influenced New York English), I don't speak in my natural dialect when talking to coworkers or people from outta the city, only my close friends or family'll hear me talk as natural, and that's the thing, humans are good at code switching but, given enough time, the differences will've become so great that even that won't work anymore. [NEWLINE] [NEWLINE] The time-depth isn't yet there, American English started diverging in the 17th century and converged again during the Georgian and Victorian eras ( when non-rhoticity became a feature of Coastal Southern speech, New England speech &amp; New York speech (New York speech could've been considered the standard for American English up till WWII when prestige shifted toward the mid-west). [NEWLINE] [NEWLINE] Give it a couple generations since it seems the Englishes are diverging again rather rapidly (the break-off of California &amp; Pacific Northwestern English from the Western group springs to mind, especially Californian English phonemics with it's reintroduction of front-rounded vowels (not seen in a standard English since West-Saxon (Wessexian) Old English).</s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Raising a child is the most economical use of resources you can possibly have. Generally speaking, you are creating a better version of yourself, not to mention you still contribute a little bit to society<mask> raising a kid,<mask> everything that kid does in the future is<mask> of your efforts.<mask><mask><mask> efficiency goes, Having kids is like having compound interest on everything you do, and it keeps adding up until the human race is extinct. [ENDQ] [NEWLINE] This assumes a whole lot about your kid not being an insufferable shit to everyone all the time who soaks up resources and gives nothing back, ultimately dying before procreating him or herself.  Granted, you have a good deal of influence over this,<mask> plenty of people who are conscientious/loving parents have kids who grow up to be inexplicably shitty, and who go on to ruin their own lives (and sometimes the lives of others too).  Unlucky genetics unquestionably plays a role, it's not scientifically justifiable anymore to believe in 100% nurture.  There's<mask> the fact that parents aren't the only influences on kids, eventually they go off on their own and you can't control<mask> the world inflicts on them.  I see having kids less<mask> an investment than<mask> a high stakes gamble<mask> you're allowed to stack the deck a little. [NEWLINE] [NEWLINE] That said, I like your username, the struggle is real.</s>
Label encoding: <s> [STARTQ] Raising a child is the most economical use of resources you can possibly have. Generally speaking, you are creating a better version of yourself, not to mention you still contribute a little bit to society while raising a kid, but everything that kid does in the future is because of your efforts. As far as efficiency goes, Having kids is like having compound interest on everything you do, and it keeps adding up until the human race is extinct. [ENDQ] [NEWLINE] This assumes a whole lot about your kid not being an insufferable shit to everyone all the time who soaks up resources and gives nothing back, ultimately dying before procreating him or herself.  Granted, you have a good deal of influence over this, but plenty of people who are conscientious/loving parents have kids who grow up to be inexplicably shitty, and who go on to ruin their own lives (and sometimes the lives of others too).  Unlucky genetics unquestionably plays a role, it's not scientifically justifiable anymore to believe in 100% nurture.  There's also the fact that parents aren't the only influences on kids, eventually they go off on their own and you can't control what the world inflicts on them.  I see having kids less as an investment than as a high stakes gamble where you're allowed to stack the deck a little. [NEWLINE] [NEWLINE] That said, I like your username, the struggle is real.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>100% true.   Anyone who thinks that civilians with weaponry are going to defeat the army directly is delusional.  <mask> you dont have to win outright to achieve your goal. [NEWLINE] [NEWLINE] <mask> the foreign examples show, you only have to harass, bleed, and resist long enough for the political situation to demand the army's withdrawl. [NEWLINE] [NEWLINE] A domestic movement would have it much easier.  A significant portion of our soldiers are going to have big problems with firing on American citizens on American soil.   Forcing the governments hand in this case is only step 1.   The most likely step 2 is a fracturing of the Army.   Now its not just civilians with small arms against the Leviathan... its a full civil war, with the all-powerful US military fractured and against itself.<mask> the people are truly against the government, and the army is even partially fractured... the calculus on outcome becomes a lot more in doubt. [NEWLINE] The harder the government cracks down on the rebellion, the more sympathizers they create.  Every bomb enrages a new group of previously neutral bystanders.  Every dead soldier makes another family ask<mask> the fuck did my boy have to die? [NEWLINE] [NEWLINE] Jimbo and his 20 survivalists arent going to do anything,<mask><mask> a significant chunk of the population takes up arms, the government is going to be in a no-win situation.  </s>
Label encoding: <s>100% true.   Anyone who thinks that civilians with weaponry are going to defeat the army directly is delusional.   But you dont have to win outright to achieve your goal. [NEWLINE] [NEWLINE] As the foreign examples show, you only have to harass, bleed, and resist long enough for the political situation to demand the army's withdrawl. [NEWLINE] [NEWLINE] A domestic movement would have it much easier.  A significant portion of our soldiers are going to have big problems with firing on American citizens on American soil.   Forcing the governments hand in this case is only step 1.   The most likely step 2 is a fracturing of the Army.   Now its not just civilians with small arms against the Leviathan... its a full civil war, with the all-powerful US military fractured and against itself. If the people are truly against the government, and the army is even partially fractured... the calculus on outcome becomes a lot more in doubt. [NEWLINE] The harder the government cracks down on the rebellion, the more sympathizers they create.  Every bomb enrages a new group of previously neutral bystanders.  Every dead soldier makes another family ask why the fuck did my boy have to die? [NEWLINE] [NEWLINE] Jimbo and his 20 survivalists arent going to do anything, but if a significant chunk of the population takes up arms, the government is going to be in a no-win situation.  </s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Economics<mask> a discipline does not and cannot discuss the idea of value. Value here, I define<mask> the "innate goodness" of something.<mask> Economics does do is it describe<mask> prices come about. The mistake you are making is that you assume that something's price reflects its value. It need not. Consider now that the "price" of work is effectively a salary. [NEWLINE] [NEWLINE] Teachers may do an enormous amount of social "good" and<mask> many people may think that they are valuable (perhaps more<mask> than a CEO -<mask> that's probably<mask> it's harder to see<mask> a CEO does).<mask> in terms of supply, there is an awful lot of supply for teachers - far more<mask> than for good CEOs. Willingness to pay is certainly an important factor on the demand side,<mask> your analysis ignores the supply side, which /u/hacksoncode explained. People may well value teachers highly,<mask> they do not have to pay them lots of money<mask> there are lots of teachers out there.<mask> teacher's salaries are low<mask> this does not neccesarily reflect the personal valuation of the individual. Again, I stress, prices do not reflect valuations,<mask> rather prices reflect the interaction of supply and demand. [NEWLINE] [NEWLINE] tl;dr Willingness to pay is a demand side factor,<mask> you must consider the supply side too. Value does not equal price.</s>
Label encoding: <s>Economics as a discipline does not and cannot discuss the idea of value. Value here, I define as the "innate goodness" of something. What Economics does do is it describe how prices come about. The mistake you are making is that you assume that something's price reflects its value. It need not. Consider now that the "price" of work is effectively a salary. [NEWLINE] [NEWLINE] Teachers may do an enormous amount of social "good" and so many people may think that they are valuable (perhaps more so than a CEO - but that's probably because it's harder to see what a CEO does). But in terms of supply, there is an awful lot of supply for teachers - far more so than for good CEOs. Willingness to pay is certainly an important factor on the demand side, but your analysis ignores the supply side, which /u/hacksoncode explained. People may well value teachers highly, but they do not have to pay them lots of money because there are lots of teachers out there. So teacher's salaries are low but this does not neccesarily reflect the personal valuation of the individual. Again, I stress, prices do not reflect valuations, but rather prices reflect the interaction of supply and demand. [NEWLINE] [NEWLINE] tl;dr Willingness to pay is a demand side factor, but you must consider the supply side too. Value does not equal price.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> lawsuits cause the city to lose tons of money,<mask> shouldn't they be able to lower overall compensation? [ENDQ] [NEWLINE] You can, you just can't do it retroactively.  Clawbacks are incredibly rare, and basically only exist for malfeasance or repayment of benefits specifically tied to something you personally didn't do.  A clawback<mask> "someone else fucked up unrelated to you" is an absurd contractual term, that might be considered unconscionable. [NEWLINE] [NEWLINE] <mask><mask> you want to lower their salaries for next year, that's totally legal.  Just be prepared for them to go on strike. [NEWLINE] [NEWLINE] The other reason not to do it I mentioned in a comment to someone else: it would cost a *fortune* in legal costs. [NEWLINE] [NEWLINE] [STARTQ] <mask> such a clause were written in, you would have to join literally every officer on the force<mask> a real party in interest to every lawsuit<mask> the municipality might have to indemnify an officer. And then probably have to indemnify them each against the legal costs of their joinder. [ENDQ] [NEWLINE] Basically,<mask> you sue an NYPD officer, which has like 40,000 officers on staff, you now have each and every officer on the hook to possibly lose money. <mask> each of them now is entitled to appear in court, with their own lawyer, and fight you.  That would be insane and crazy expensive.  </s>
Label encoding: <s> [STARTQ] If lawsuits cause the city to lose tons of money, why shouldn't they be able to lower overall compensation? [ENDQ] [NEWLINE] You can, you just can't do it retroactively.  Clawbacks are incredibly rare, and basically only exist for malfeasance or repayment of benefits specifically tied to something you personally didn't do.  A clawback because "someone else fucked up unrelated to you" is an absurd contractual term, that might be considered unconscionable. [NEWLINE] [NEWLINE] But if you want to lower their salaries for next year, that's totally legal.  Just be prepared for them to go on strike. [NEWLINE] [NEWLINE] The other reason not to do it I mentioned in a comment to someone else: it would cost a *fortune* in legal costs. [NEWLINE] [NEWLINE] [STARTQ] if such a clause were written in, you would have to join literally every officer on the force as a real party in interest to every lawsuit where the municipality might have to indemnify an officer. And then probably have to indemnify them each against the legal costs of their joinder. [ENDQ] [NEWLINE] Basically, if you sue an NYPD officer, which has like 40,000 officers on staff, you now have each and every officer on the hook to possibly lose money.  So each of them now is entitled to appear in court, with their own lawyer, and fight you.  That would be insane and crazy expensive.  </s>
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree completely with your last paragraph and would like to apologize<mask> I have been discussing with a lot of people the notion that to defend freedom of speech one is obligated to republish the images that instigate attacks, and incorrectly attributed this view to you in our discussion. [NEWLINE] [NEWLINE] I strongly believe in freedom of expression and feel that<mask><mask><mask> of this it is my duty to exercise by telling people<mask> I feel that their actions have an overall negative effect. I don't want to tell people that they can't print whatever they want,<mask> I will sure<mask> hell tell people that<mask><mask> they shouldn't<mask> they are being divisive and promoting<mask> I see<mask> bad trends. The thing I take issue with in your post is: [NEWLINE] [STARTQ] <mask> it would never occur to me to ask people to stop expressing these views simply to make me more comfortable... [ENDQ] [NEWLINE] Freedom of speech is meant to defend your right to say<mask> you will, not defend your speech from criticism.<mask> someone says something that I would consider racist, sexist, or generally awful, they have a right to do<mask>, and I in turn have a right to tell them that their speech is harmful. I can't and wouldn't want to forcefully suppress them,<mask> I can tell them that I don't support<mask> they say, think they are wrong, and I don't think they should say it for a variety of reasons. </s>
Label encoding: <s>I agree completely with your last paragraph and would like to apologize as I have been discussing with a lot of people the notion that to defend freedom of speech one is obligated to republish the images that instigate attacks, and incorrectly attributed this view to you in our discussion. [NEWLINE] [NEWLINE] I strongly believe in freedom of expression and feel that as a result of this it is my duty to exercise by telling people when I feel that their actions have an overall negative effect. I don't want to tell people that they can't print whatever they want, but I will sure as hell tell people that I think they shouldn't because they are being divisive and promoting what I see as bad trends. The thing I take issue with in your post is: [NEWLINE] [STARTQ] But it would never occur to me to ask people to stop expressing these views simply to make me more comfortable... [ENDQ] [NEWLINE] Freedom of speech is meant to defend your right to say what you will, not defend your speech from criticism. If someone says something that I would consider racist, sexist, or generally awful, they have a right to do so, and I in turn have a right to tell them that their speech is harmful. I can't and wouldn't want to forcefully suppress them, but I can tell them that I don't support what they say, think they are wrong, and I don't think they should say it for a variety of reasons. </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Sort of - by showing that you were inconsistent in the case of rape, they were intending to change your view that more abortions should be allowed, instead you went the other direction, which is your prerogative,<mask> not encouraging to those trying to change your view. [ENDQ] [NEWLINE] Ive been answering nearly every comment with an open mind, and like Ive said my view has been changed on the general topic<mask> far and some people were engaging me with conversation conducive to altering my feelings on the matter. [NEWLINE] [NEWLINE] Edit:I've answered all the comments with an open mind,<mask> I felt they were at least not disrespectful or making strange claims of body parts and cows. [NEWLINE] [NEWLINE] You should keep rule 3 in mind by the way,<mask> you are literally telling me that I wont change my mind<mask><mask> I TELL YOU that I am having some turmoil about the fucking topic. Are you sure youre a mod and not on the wrong account? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> you are unwilling to change the view that abortion of any kind at any point is murdering a human being, there's not a lot to discuss. [ENDQ] [NEWLINE] Remember having an open mind doesnt mean your a mind is a door mat, I would think it is acceptable to discuss my own view of their presentation instead of laying down and letting their ideas wash over me. [NEWLINE] [NEWLINE] The appearance of resistance is not the proof of ignorance.</s>
Label encoding: <s> [STARTQ] Sort of - by showing that you were inconsistent in the case of rape, they were intending to change your view that more abortions should be allowed, instead you went the other direction, which is your prerogative, but not encouraging to those trying to change your view. [ENDQ] [NEWLINE] Ive been answering nearly every comment with an open mind, and like Ive said my view has been changed on the general topic so far and some people were engaging me with conversation conducive to altering my feelings on the matter. [NEWLINE] [NEWLINE] Edit:I've answered all the comments with an open mind, when I felt they were at least not disrespectful or making strange claims of body parts and cows. [NEWLINE] [NEWLINE] You should keep rule 3 in mind by the way, as you are literally telling me that I wont change my mind WHILE I TELL YOU that I am having some turmoil about the fucking topic. Are you sure youre a mod and not on the wrong account? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] If you are unwilling to change the view that abortion of any kind at any point is murdering a human being, there's not a lot to discuss. [ENDQ] [NEWLINE] Remember having an open mind doesnt mean your a mind is a door mat, I would think it is acceptable to discuss my own view of their presentation instead of laying down and letting their ideas wash over me. [NEWLINE] [NEWLINE] The appearance of resistance is not the proof of ignorance.</s>
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Something a little more like<mask> they have in the UK might be a direction to move toward: [NEWLINE] [NEWLINE] "Today only a small proportion of officers are authorised to use firearms. Latest Home Office figures show there were just 6,653 officers authorised to use firearms in England and Wales - about 5% of the total number.  None of which implies, of course, that the British police are somehow gun-free.  Each police force has its own firearms unit. Police armed response vehicles have been deployed<mask> 1991." [NEWLINE] [NEWLINE] (Source: [URL] ) [NEWLINE] [NEWLINE] <mask> in America, it might be different,<mask> of the high volume of legal and illegal weapons on the street,<mask> that doesn't mean one can't imagine a reduction in guns for the average cop (who would<mask> be trained to engage with the community more). [NEWLINE] [NEWLINE] And<mask><mask>, in the case of someone firing a weapon in the street, we want cops to be able to effect an arrest / stop that perp. <mask> in most of those scenarios, there are<mask> MANY more interventions that should have happened BEFORE the person went on a rampage (or whatever). [NEWLINE] [NEWLINE] And you can imagine<mask>, a well-trained, lightly armed cop could stop and just chat with a person of interest, and really communicate with them, versus the climate of having an armed officer just strolling around, not really interacting with the public.</s>
Label encoding: <s>Something a little more like what they have in the UK might be a direction to move toward: [NEWLINE] [NEWLINE] "Today only a small proportion of officers are authorised to use firearms. Latest Home Office figures show there were just 6,653 officers authorised to use firearms in England and Wales - about 5% of the total number.  None of which implies, of course, that the British police are somehow gun-free.  Each police force has its own firearms unit. Police armed response vehicles have been deployed since 1991." [NEWLINE] [NEWLINE] (Source: [URL] ) [NEWLINE] [NEWLINE] But in America, it might be different, because of the high volume of legal and illegal weapons on the street, but that doesn't mean one can't imagine a reduction in guns for the average cop (who would also be trained to engage with the community more). [NEWLINE] [NEWLINE] And I agree, in the case of someone firing a weapon in the street, we want cops to be able to effect an arrest / stop that perp.  But in most of those scenarios, there are also MANY more interventions that should have happened BEFORE the person went on a rampage (or whatever). [NEWLINE] [NEWLINE] And you can imagine where, a well-trained, lightly armed cop could stop and just chat with a person of interest, and really communicate with them, versus the climate of having an armed officer just strolling around, not really interacting with the public.</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many open source projects are made by people who want to use the software. A programmer might need software for some purpose to do something faster. By making it open source they can get advice from others on<mask> to develop it and improvements. They get an enhanced product for free. They get more beta testers for free. [NEWLINE] [NEWLINE] You<mask> get to be involved in a community and get praise for<mask> you have done. [NEWLINE] [NEWLINE] [STARTQ] Not only that,<mask> you are potentially ruining somebody else's business by developing a open source alternative, which is to say, you're kind of disrespecting the effort and work they've put on the product they're trying to sell by not even competing for the money. [ENDQ] [NEWLINE] <mask> someone is making a product that's already out there for free then they don't deserve money. No one deserves your money. You have to earn it by providing value to the user. [NEWLINE] [NEWLINE] [STARTQ] <mask>, can you imagine<mask> everyone went open source? There would be no money on software anymore, only infrastructure. Videogames, for example, wouldn't be<mask> they are today. There would be no big push for innovation<mask> there would be no money to be made. [ENDQ] [NEWLINE] Big business can do things open source can't<mask> it has more money and more programmers. High quality video games and products are always going to be in demand, open source products won't displace them.</s>
Label encoding: <s>Many open source projects are made by people who want to use the software. A programmer might need software for some purpose to do something faster. By making it open source they can get advice from others on how to develop it and improvements. They get an enhanced product for free. They get more beta testers for free. [NEWLINE] [NEWLINE] You also get to be involved in a community and get praise for what you have done. [NEWLINE] [NEWLINE] [STARTQ] Not only that, but you are potentially ruining somebody else's business by developing a open source alternative, which is to say, you're kind of disrespecting the effort and work they've put on the product they're trying to sell by not even competing for the money. [ENDQ] [NEWLINE] If someone is making a product that's already out there for free then they don't deserve money. No one deserves your money. You have to earn it by providing value to the user. [NEWLINE] [NEWLINE] [STARTQ] Also, can you imagine if everyone went open source? There would be no money on software anymore, only infrastructure. Videogames, for example, wouldn't be what they are today. There would be no big push for innovation because there would be no money to be made. [ENDQ] [NEWLINE] Big business can do things open source can't as it has more money and more programmers. High quality video games and products are always going to be in demand, open source products won't displace them.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Alcohol: Oranges have a lot more varied uses than apples. Screwdrivers, mimosas, fuzzy navels and tequila sunrises all use orange juice.  Orange liquors are staples in a huge variety of mixed drinks, including sidecars and margaritas.  Blood orange bitters is great in a lot of drinks.  And it's used<mask> a garnish in things like old fashioneds, or with a beer.  And the zest is used in brewing. Apples mostly just have cider.  No contest - the orange is way more versatile. [NEWLINE] [NEWLINE] Food: Orange adds a great burst of brightness to a lot of food.  It goes in salads.  It's in duck l'orange. You specifically called out pork,<mask> I use orange zest<mask> a key component in one of my spice rubs that's great on pork.  And orange sauce is used all the time in american Chinese food, both on chicken and on pork.  And you can do an orange pie.  It's like a key lime pie,<mask> with oranges.  And orange spice cake is amazing. [NEWLINE] [NEWLINE] Oranges are great<mask> they have a whole additional class of uses.  For apples, you use the fruit or the juice.  For oranges, you get the fruit, the juice, and the zest/peel.  </s>
Label encoding: <s>Alcohol: Oranges have a lot more varied uses than apples. Screwdrivers, mimosas, fuzzy navels and tequila sunrises all use orange juice.  Orange liquors are staples in a huge variety of mixed drinks, including sidecars and margaritas.  Blood orange bitters is great in a lot of drinks.  And it's used as a garnish in things like old fashioneds, or with a beer.  And the zest is used in brewing. Apples mostly just have cider.  No contest - the orange is way more versatile. [NEWLINE] [NEWLINE] Food: Orange adds a great burst of brightness to a lot of food.  It goes in salads.  It's in duck l'orange. You specifically called out pork, but I use orange zest as a key component in one of my spice rubs that's great on pork.  And orange sauce is used all the time in american Chinese food, both on chicken and on pork.  And you can do an orange pie.  It's like a key lime pie, but with oranges.  And orange spice cake is amazing. [NEWLINE] [NEWLINE] Oranges are great because they have a whole additional class of uses.  For apples, you use the fruit or the juice.  For oranges, you get the fruit, the juice, and the zest/peel.  </s>
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You will not succeed in college<mask> you do not do the reading, especially, *especially*<mask> you're after a liberal arts degree of any kind.* The value of any course worth taking is primarily encoded in the things you will read and the discussions that come out of it. It primes you to be receptive to ideas in subsequent courses. [NEWLINE] [NEWLINE] You do karate: you can think of a good book<mask> learning new kata for your brain. The learning encoded in a book doesn't itself provide value,<mask> opens up a new space through which your mind (rather than your body) can move. You're learning new intellectual actions from people with more experience in the field than you. [NEWLINE] [NEWLINE] Which means that not everything you read has equivalent value. The cheap paperbacks in the grocery store are throwaway reading; they contain nothing new for your brain to chew on. Some books are like martial arts charlatans -- they purport to have great new ideas,<mask> most of it's just bullshido. A good nonfiction title,<mask><mask><mask><mask>, can expose you to a giant box of new ideas -- sometimes, ideas you may not understand or be ready for. [NEWLINE] [NEWLINE] \* - Note that<mask> you may skip the reading and still get the grades, you're better-served by lighting your tuition money on fire to cook bacon<mask> you do<mask>.</s>
Label encoding: <s>You will not succeed in college if you do not do the reading, especially, *especially* if you're after a liberal arts degree of any kind.* The value of any course worth taking is primarily encoded in the things you will read and the discussions that come out of it. It primes you to be receptive to ideas in subsequent courses. [NEWLINE] [NEWLINE] You do karate: you can think of a good book as learning new kata for your brain. The learning encoded in a book doesn't itself provide value, but opens up a new space through which your mind (rather than your body) can move. You're learning new intellectual actions from people with more experience in the field than you. [NEWLINE] [NEWLINE] Which means that not everything you read has equivalent value. The cheap paperbacks in the grocery store are throwaway reading; they contain nothing new for your brain to chew on. Some books are like martial arts charlatans -- they purport to have great new ideas, but most of it's just bullshido. A good nonfiction title, on the other hand, can expose you to a giant box of new ideas -- sometimes, ideas you may not understand or be ready for. [NEWLINE] [NEWLINE] \* - Note that while you may skip the reading and still get the grades, you're better-served by lighting your tuition money on fire to cook bacon if you do so.</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't believe that the British public are knowledgeable/qualified to even comment, let alone have any emotional response to the current US political standoff over the budget, which has resulted in the current shutdown. I say this<mask> a British person myself, who has taken the time to familiarise himself with US politics, including the constitution, bill of rights, the independant governing of the individual states, and<mask> fundamentally important this all is<mask> part of a democratic society. My experience is that people in the UK not only ignorantly assume that the USA works the same way Britain does,<mask> that the people have the exact same rights, priorities and philosophies<mask> we do here. They don't understand<mask> is going on and will believe whatever information is given to them, right or wrong, and start getting all upset about it. This to me is utterly ridiculous, for people who have no concept of the situation to be getting angry about it. CMV. [NEWLINE] [NEWLINE] EDIT: Resolved. Delta has been awarded. View has been changed. You can all go home now and move on, we're done here. Thankyou<mask> much to all the awesome, well thought out replies. To the people who have come here intending on throwing insults at me: you will simply be reported to the mods. I won't argue with you or engage you or give you any satisfaction whatsoever,<mask> save it.</s>
Label encoding: <s>I don't believe that the British public are knowledgeable/qualified to even comment, let alone have any emotional response to the current US political standoff over the budget, which has resulted in the current shutdown. I say this as a British person myself, who has taken the time to familiarise himself with US politics, including the constitution, bill of rights, the independant governing of the individual states, and how fundamentally important this all is as part of a democratic society. My experience is that people in the UK not only ignorantly assume that the USA works the same way Britain does, but that the people have the exact same rights, priorities and philosophies as we do here. They don't understand what is going on and will believe whatever information is given to them, right or wrong, and start getting all upset about it. This to me is utterly ridiculous, for people who have no concept of the situation to be getting angry about it. CMV. [NEWLINE] [NEWLINE] EDIT: Resolved. Delta has been awarded. View has been changed. You can all go home now and move on, we're done here. Thankyou so much to all the awesome, well thought out replies. To the people who have come here intending on throwing insults at me: you will simply be reported to the mods. I won't argue with you or engage you or give you any satisfaction whatsoever, so save it.</s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree. This thread has made me realize that I need to narrow my statement. *Look at that my view changed!*<mask><mask> it does come down to the style of music. I guess my frustration came from the fact that currently, composers get almost no credit,<mask> singers get almost all of it. Band members, especially those other than lead guitar get fucked in the same way,<mask> they play a huge role<mask> well. Perhaps I overreacted by saying composers always deserve more credit. [NEWLINE] [NEWLINE] The song that made me think of this was Clarity by Zedd (I'm a little girl over that song right now and I'm generally very neutral about popular music.) In all these heavily engineered songs,<mask><mask> the composer/producer really is more responsible for the catchiness of the song. [NEWLINE] [NEWLINE] I'd go further<mask>. I really think my claim is true in most cases. Sometimes, the singer really stands out, and then the singer becomes more important.<mask> generally this is not the case. [NEWLINE] [NEWLINE] <mask><mask> generally, the problem is that the singer commands the most attention,<mask> people think s/he is the main force of the song.<mask> people will tend to undervalue the music behind the singer. It's even easier to undervalue it<mask> you're not seeing the musicians on stage (<mask> it is with any recorded music).</s>
Label encoding: <s>I agree. This thread has made me realize that I need to narrow my statement. *Look at that my view changed!* I think it does come down to the style of music. I guess my frustration came from the fact that currently, composers get almost no credit, while singers get almost all of it. Band members, especially those other than lead guitar get fucked in the same way, when they play a huge role as well. Perhaps I overreacted by saying composers always deserve more credit. [NEWLINE] [NEWLINE] The song that made me think of this was Clarity by Zedd (I'm a little girl over that song right now and I'm generally very neutral about popular music.) In all these heavily engineered songs, I think the composer/producer really is more responsible for the catchiness of the song. [NEWLINE] [NEWLINE] I'd go further though. I really think my claim is true in most cases. Sometimes, the singer really stands out, and then the singer becomes more important. But generally this is not the case. [NEWLINE] [NEWLINE] I think generally, the problem is that the singer commands the most attention, so people think s/he is the main force of the song. Thus people will tend to undervalue the music behind the singer. It's even easier to undervalue it when you're not seeing the musicians on stage ( as it is with any recorded music).</s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One of my favorite quotes from Rome "... you know nothing of violence.  A pack of mangy dogs can kill a lion." [NEWLINE] [NEWLINE] Democracy is a stable form of government<mask> physical force against it is pointless.  All things being equal the minority would always lose in a fight and<mask> rule by the majority is physically unassailable.  The brilliance of western democracies was constitutional protections to limit the power of even the majority to some reasonable boundaries that all can agree to and<mask>, even<mask> the majority are in favor of a proposition, the constitutional values will sway popular opinion. [NEWLINE] [NEWLINE] My point... It is important that Bill gates doesn't get more political impact than 100 million other people<mask>, eventually, those 100 million other people would storm his house, kill his guards, and mount his head on a pike. [NEWLINE] [NEWLINE] Just look at<mask> well feudal societies worked out in the past, and<mask> well Russia is working out today. [NEWLINE] [NEWLINE] The brilliance of western society is that we are all in this together and we all get to choose the country's course together. [NEWLINE] [NEWLINE] Finally... Bill Gates is a great man. <mask> there was a person alive today who<mask><mask> could be king and do good it would be Bill Gates. <mask> I probably wouldn't be in the mob at his gates... <mask><mask> Kim Kardashian were making decisions for millions of others....</s>
Label encoding: <s>One of my favorite quotes from Rome "... you know nothing of violence.  A pack of mangy dogs can kill a lion." [NEWLINE] [NEWLINE] Democracy is a stable form of government because physical force against it is pointless.  All things being equal the minority would always lose in a fight and so rule by the majority is physically unassailable.  The brilliance of western democracies was constitutional protections to limit the power of even the majority to some reasonable boundaries that all can agree to and thus, even if the majority are in favor of a proposition, the constitutional values will sway popular opinion. [NEWLINE] [NEWLINE] My point... It is important that Bill gates doesn't get more political impact than 100 million other people because, eventually, those 100 million other people would storm his house, kill his guards, and mount his head on a pike. [NEWLINE] [NEWLINE] Just look at how well feudal societies worked out in the past, and how well Russia is working out today. [NEWLINE] [NEWLINE] The brilliance of western society is that we are all in this together and we all get to choose the country's course together. [NEWLINE] [NEWLINE] Finally... Bill Gates is a great man.  If there was a person alive today who I think could be king and do good it would be Bill Gates.  So I probably wouldn't be in the mob at his gates...  If however Kim Kardashian were making decisions for millions of others....</s>
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0927, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No matter<mask> costly, you're not depriving them of anything. At most you could crack a connection to e.g. Netflix and download films without paying for them. You would be depriving them of bandwidth.<mask> that's not usually<mask> it goes. [ENDQ] [NEWLINE] By you (the generalized "you") downloading and using a piece of software, you are implicitly stating that it has value to you. <mask> you download that software for free<mask><mask><mask> the creators wishes, you are depriving them of the profits they should be able to reasonably expect to receive from someone using their creation. [NEWLINE] [NEWLINE] People frequently argue back at this in the form of "well, I wouldn't have downloaded it and used it<mask> it cost money"<mask> I don't personally see that<mask> valid.  The software DOES cost money, and you have simply taken it upon yourself to decide for someone else<mask> their risk and effort are worth. <mask> the creator sets the price too high then they won't see sufficient returns,<mask><mask> the creator must set the price too low to succeed, perhaps they simply will do something else with their time in the future. <mask>, more importantly, that is THEIR decision to make. [NEWLINE] [NEWLINE] Piracy is<mask> a low injury crime,<mask> just<mask> it doesn't hurt someone "all that much" doesn't make it justifiable.</s>
Label encoding: <s> [STARTQ] No matter how costly, you're not depriving them of anything. At most you could crack a connection to e.g. Netflix and download films without paying for them. You would be depriving them of bandwidth. But that's not usually how it goes. [ENDQ] [NEWLINE] By you (the generalized "you") downloading and using a piece of software, you are implicitly stating that it has value to you.  When you download that software for free in spite of the creators wishes, you are depriving them of the profits they should be able to reasonably expect to receive from someone using their creation. [NEWLINE] [NEWLINE] People frequently argue back at this in the form of "well, I wouldn't have downloaded it and used it if it cost money" but I don't personally see that as valid.  The software DOES cost money, and you have simply taken it upon yourself to decide for someone else what their risk and effort are worth.  If the creator sets the price too high then they won't see sufficient returns, but if the creator must set the price too low to succeed, perhaps they simply will do something else with their time in the future.  But, more importantly, that is THEIR decision to make. [NEWLINE] [NEWLINE] Piracy is indeed a low injury crime, but just because it doesn't hurt someone "all that much" doesn't make it justifiable.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Obviously I don't agree with that. [ENDQ] [NEWLINE] Okay,<mask> is your *right to not be killed* distinct from whether people acknowledge it or not? And distinct from whether people provide you with things like education or attorneys? [NEWLINE] [NEWLINE] [STARTQ] Morality is separate from rights and laws [ENDQ] [NEWLINE] Well, I certainly think it is separate from laws,<mask><mask><mask> moral facts are true or false irrespective of human opinion (in other words, *natural*).<mask> I do think laws are supposed to match up to morality<mask> best we can make them.<mask> morality is emphatically not separate from rights. Rights are normative by nature. Just read any of the philosophical literature on them. For example, the [Internet Encyclopedia of Philosophy]( [URL] /) article on rights begins<mask> follows: "*Human rights are certain moral guarantees.*" [NEWLINE] [NEWLINE] [STARTQ] <mask> all of them are human constructs. [ENDQ] [NEWLINE] <mask> think moral rights are mere human constructs? This is one view,<mask> it isn't the majority view amongst philosophers (most philosophers are moral realists), and it would commit you to some pretty counter-intuitive beliefs (like that<mask> the Nazi's had won WW2, took over the world, and brainwashed everyone into believing they were right, then there would be nothing wrong with massacring Jews,<mask> morality is just a human construct, right?<mask> clearly this would be wrong.)</s>
Label encoding: <s> [STARTQ] Obviously I don't agree with that. [ENDQ] [NEWLINE] Okay, so is your *right to not be killed* distinct from whether people acknowledge it or not? And distinct from whether people provide you with things like education or attorneys? [NEWLINE] [NEWLINE] [STARTQ] Morality is separate from rights and laws [ENDQ] [NEWLINE] Well, I certainly think it is separate from laws, because I think moral facts are true or false irrespective of human opinion (in other words, *natural*). Though I do think laws are supposed to match up to morality as best we can make them. But morality is emphatically not separate from rights. Rights are normative by nature. Just read any of the philosophical literature on them. For example, the [Internet Encyclopedia of Philosophy]( [URL] /) article on rights begins as follows: "*Human rights are certain moral guarantees.*" [NEWLINE] [NEWLINE] [STARTQ] although all of them are human constructs. [ENDQ] [NEWLINE] Why think moral rights are mere human constructs? This is one view, but it isn't the majority view amongst philosophers (most philosophers are moral realists), and it would commit you to some pretty counter-intuitive beliefs (like that if the Nazi's had won WW2, took over the world, and brainwashed everyone into believing they were right, then there would be nothing wrong with massacring Jews, since morality is just a human construct, right? Yet clearly this would be wrong.)</s>
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is<mask> I'm getting at, and<mask> I don't understand.<mask> is there this underlying influence of the term "oppression" on the definition? Racism is degrading, yes, and can most definitely be harmful<mask> used by an oppressor.<mask><mask> makes being a minority and the use of racism mutually exclusive? In the example I gave from /r/nottheonion, the subject was denying white males access to a social activist meeting. [NEWLINE] [NEWLINE] [NEWLINE] You are lumping the term oppression into the definition of racism, and that's<mask> we are disagreeing. Yes, the two can both exist,<mask> in some cases I feel like they can exist without one another. In the example, a minority was discriminating against another race, and prohibiting social cooperation from an entire demographic. I understand that the magnitude between this example and black oppression<mask> a whole is distinctly different,<mask> I feel like this is a prime example of racism/predjudice/discrimination on behalf of a minority. [NEWLINE] [NEWLINE] [NEWLINE] In this instance, I see this<mask> racist,<mask> I don't see it<mask> oppressive,<mask> it doesn't carry<mask> much weight<mask> a racist societal/political system would be.<mask> aren't oppression and racism seperate terms?<mask> are you defining them<mask> synonymous. And wouldn't that invalidate the idea that minorities can't be racist?</s>
Label encoding: <s>This is what I'm getting at, and what I don't understand. Why is there this underlying influence of the term "oppression" on the definition? Racism is degrading, yes, and can most definitely be harmful if used by an oppressor. But what makes being a minority and the use of racism mutually exclusive? In the example I gave from /r/nottheonion, the subject was denying white males access to a social activist meeting. [NEWLINE] [NEWLINE] [NEWLINE] You are lumping the term oppression into the definition of racism, and that's where we are disagreeing. Yes, the two can both exist, but in some cases I feel like they can exist without one another. In the example, a minority was discriminating against another race, and prohibiting social cooperation from an entire demographic. I understand that the magnitude between this example and black oppression as a whole is distinctly different, but I feel like this is a prime example of racism/predjudice/discrimination on behalf of a minority. [NEWLINE] [NEWLINE] [NEWLINE] In this instance, I see this as racist, but I don't see it as oppressive, because it doesn't carry as much weight as a racist societal/political system would be. So aren't oppression and racism seperate terms? Why are you defining them as synonymous. And wouldn't that invalidate the idea that minorities can't be racist?</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People are capable of forming well structured arguments,<mask> there might be major flaws in them without philosophical study. I'm not arguing that philosophers own these arguments, I am arguing that a philosopher is going to be better educated on past and current arguments and is more able to establish the strength of propositions. [NEWLINE] [NEWLINE] I'll use myself<mask> an example. I entered a debate subreddit a long<mask> ago *thinking* I had good arguments. I found out rather quickly that I was an idiot on philosophical arguments regarding religion. Through reading professionals, taking a university level course in it, and studying older arguments I am now slightly less of an idiot (still not<mask> well informed<mask> many<mask> ). The same,<mask><mask>, can be said for many people in many different arguments. All of the people you know can probably create good sounding arguments,<mask> put them in a room for 20 minutes with an opposing philosopher and they would likely find out that their argument has been done before and has been thoroughly shown to be bad. [NEWLINE] [NEWLINE] <mask> yes, anyone can make an argument. Not everyone can make a well informed and good quality argument that doesn't have any holes in it. That's<mask> philosophy comes in, they poke holes in each other ideas and improve upon them with the possibility of them having major effects on policy or individuals. [NEWLINE] [NEWLINE] Edit: Not -&gt; now.</s>
Label encoding: <s>People are capable of forming well structured arguments, but there might be major flaws in them without philosophical study. I'm not arguing that philosophers own these arguments, I am arguing that a philosopher is going to be better educated on past and current arguments and is more able to establish the strength of propositions. [NEWLINE] [NEWLINE] I'll use myself as an example. I entered a debate subreddit a long while ago *thinking* I had good arguments. I found out rather quickly that I was an idiot on philosophical arguments regarding religion. Through reading professionals, taking a university level course in it, and studying older arguments I am now slightly less of an idiot (still not as well informed as many though ). The same, I think, can be said for many people in many different arguments. All of the people you know can probably create good sounding arguments, but put them in a room for 20 minutes with an opposing philosopher and they would likely find out that their argument has been done before and has been thoroughly shown to be bad. [NEWLINE] [NEWLINE] So yes, anyone can make an argument. Not everyone can make a well informed and good quality argument that doesn't have any holes in it. That's where philosophy comes in, they poke holes in each other ideas and improve upon them with the possibility of them having major effects on policy or individuals. [NEWLINE] [NEWLINE] Edit: Not -&gt; now.</s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not some hippy liberal christian, I'm a serious southern baptist Sunday school teacher.<mask><mask> that after examining the Bible, there is no argument that being transgender is wrong.<mask>, there are only three main prongs of attack, all of which are incorrect. [NEWLINE] [NEWLINE] The first prong of attack is the homosexuality argument.<mask>,<mask> someone really is the opposite gender, then it would by definition not be homosexual. [NEWLINE] [NEWLINE] The second prong of attack is the rule against cross dressing.<mask>,<mask> someone really is the opposite gender, it's not cross dressing. [NEWLINE] [NEWLINE] The third prong of attack is against physical mutilation of the body.<mask><mask> there are other things wrong with this argument.<mask>, that someone is transgender does not imply that they will or have to'mutilate' their body. They may be happier<mask> they do,<mask> being transgender does not entail it happening. [NEWLINE] [NEWLINE] None of these imply that being transgender it's self is in any way wrong. It is always something else that commonly goes along with transgender issues that makes it wrong. [NEWLINE] [NEWLINE] Edit: This argument depends upon a non-biological definition of gender.<mask> gender is biological, then the attacks make a lot more sense.<mask>, this raises the question, "Can we define gender<mask> biological based on the Bible?" </s><pad>
Label encoding: <s>I'm not some hippy liberal christian, I'm a serious southern baptist Sunday school teacher. I think that after examining the Bible, there is no argument that being transgender is wrong. Indeed, there are only three main prongs of attack, all of which are incorrect. [NEWLINE] [NEWLINE] The first prong of attack is the homosexuality argument. However, if someone really is the opposite gender, then it would by definition not be homosexual. [NEWLINE] [NEWLINE] The second prong of attack is the rule against cross dressing. However, if someone really is the opposite gender, it's not cross dressing. [NEWLINE] [NEWLINE] The third prong of attack is against physical mutilation of the body. I think there are other things wrong with this argument. However, that someone is transgender does not imply that they will or have to'mutilate' their body. They may be happier if they do, but being transgender does not entail it happening. [NEWLINE] [NEWLINE] None of these imply that being transgender it's self is in any way wrong. It is always something else that commonly goes along with transgender issues that makes it wrong. [NEWLINE] [NEWLINE] Edit: This argument depends upon a non-biological definition of gender. If gender is biological, then the attacks make a lot more sense. However, this raises the question, "Can we define gender as biological based on the Bible?" </s><pad>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask><mask> we're kind of talking through each other here. I am not advocating that we try to decouple law from ethics. Quite the opposite<mask><mask>. [NEWLINE] [NEWLINE] I'm simply advocating that we should abhor the view that laws should be passed on the same basis that ethical precepts are accepted by a society, even<mask> that society agrees on those precepts. In other words, laws are not the same<mask> a moral or ethical code,<mask> unlike a code that one chooses or even agrees with others to follow, the law implies giving the state the power to do something about it, and that makes it an inherently different animal. [NEWLINE] [NEWLINE] Rather than thinking of laws<mask> comparable to precepts, we should think of them first and foremost<mask> a *tool* that is being used by the state itself to accomplish some end for the interest of society. [NEWLINE] [NEWLINE] So in other words, I am arguing that in passing a law about any given ethical precept, the consideration *should always be* "I want to pass x law<mask> it will accomplish goal y"<mask> opposed to "I want to pass to pass x law<mask> I believe in precept x." This seems like a pedantic distinction,<mask> it's a very important one, and one that far too many people seem to be on the wrong side of<mask> arguing for a given piece of legislation. </s>
Label encoding: <s>So I think we're kind of talking through each other here. I am not advocating that we try to decouple law from ethics. Quite the opposite in fact. [NEWLINE] [NEWLINE] I'm simply advocating that we should abhor the view that laws should be passed on the same basis that ethical precepts are accepted by a society, even if that society agrees on those precepts. In other words, laws are not the same as a moral or ethical code, because unlike a code that one chooses or even agrees with others to follow, the law implies giving the state the power to do something about it, and that makes it an inherently different animal. [NEWLINE] [NEWLINE] Rather than thinking of laws as comparable to precepts, we should think of them first and foremost as a *tool* that is being used by the state itself to accomplish some end for the interest of society. [NEWLINE] [NEWLINE] So in other words, I am arguing that in passing a law about any given ethical precept, the consideration *should always be* "I want to pass x law because it will accomplish goal y" as opposed to "I want to pass to pass x law because I believe in precept x." This seems like a pedantic distinction, but it's a very important one, and one that far too many people seem to be on the wrong side of when arguing for a given piece of legislation. </s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is just a rephrasing of Pascal's Wager, OP, which has been thoroughly debunked for ages. [NEWLINE] [NEWLINE] First - it ignores the opportunity cost of god-belief;<mask> could that monk have done with their life<mask> they hadn't spent it cloistered in honor<mask> their god?  Your wager presupposes its conclusion by pretending wrongful god-belief has no down side. [NEWLINE] [NEWLINE] Second - it ignores the possibility of choosing the wrong god.  Follow Yahweh and it turns out Odin exists?  You may well be worse off than were you an honest unbeliever. [NEWLINE] [NEWLINE] Third - it treats belief<mask> something you choose,<mask> quite simply put, it isn't.  Nobody chooses<mask> they believe, they only choose<mask> to react to their state of belief. [NEWLINE] [NEWLINE] Fourth - it assumes the god in question is stupid, and will willingly reward someone for going through the motions of belief<mask> hopes of, well, reward.  Any god worth their salt would look at someone making that kind<mask> calculated choice<mask> an untrustworthy, mercenary jerk and treat them<mask>, whereas the well-meaning unbeliever, being decent and expecting no reward, would be someone worth rewarding. [NEWLINE] [NEWLINE] There are pages and pages of such flaws in the Wager; it stands<mask> one of the prime examples of bad philosophy.</s>
Label encoding: <s>This is just a rephrasing of Pascal's Wager, OP, which has been thoroughly debunked for ages. [NEWLINE] [NEWLINE] First - it ignores the opportunity cost of god-belief; what could that monk have done with their life if they hadn't spent it cloistered in honor if their god?  Your wager presupposes its conclusion by pretending wrongful god-belief has no down side. [NEWLINE] [NEWLINE] Second - it ignores the possibility of choosing the wrong god.  Follow Yahweh and it turns out Odin exists?  You may well be worse off than were you an honest unbeliever. [NEWLINE] [NEWLINE] Third - it treats belief as something you choose, when quite simply put, it isn't.  Nobody chooses what they believe, they only choose how to react to their state of belief. [NEWLINE] [NEWLINE] Fourth - it assumes the god in question is stupid, and will willingly reward someone for going through the motions of belief if hopes of, well, reward.  Any god worth their salt would look at someone making that kind if calculated choice as an untrustworthy, mercenary jerk and treat them accordingly, whereas the well-meaning unbeliever, being decent and expecting no reward, would be someone worth rewarding. [NEWLINE] [NEWLINE] There are pages and pages of such flaws in the Wager; it stands as one of the prime examples of bad philosophy.</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The problems to which they're suggesting solutions aren't going to go away,<mask> these alternate solutions will be arrived at regardless. [ENDQ] [NEWLINE] Ideally, yes,<mask> they may not come about<mask> fast.<mask> we know they are wrong, then we have an added necessity to come up with something better. [NEWLINE] [NEWLINE] [STARTQ] I would<mask> still like an explanation of<mask> you think it's better for us to (a) define suggestions in opposition to the Tea Party, with the risk that some of the bad Tea Party suggestions will be implemented than to (b) define suggestions in the absence of the Tea Party. [ENDQ] [NEWLINE] <mask><mask> of this like trying to prosecute a criminal. It may be that the criminal is guilty and it's not too difficult to prove,<mask> that doesn't give a prosecutor permission to half-ass the case. That's<mask> the defense attorney is there,<mask> the prosecution is forced to come up with an argument that can hold up against scrutiny. The same goal is ultimately reached<mask> there would have been<mask> there wasn't an opposition,<mask> now the prosecution is forced to give an argument that is much more clear and indisputable- more reflective of the truth of the matter. [NEWLINE] [NEWLINE] Finding out<mask> is right becomes easier<mask> you can clearly identify<mask>'s wrong, and<mask><mask> the tea party forces us to do that. </s>
Label encoding: <s> [STARTQ] The problems to which they're suggesting solutions aren't going to go away, so these alternate solutions will be arrived at regardless. [ENDQ] [NEWLINE] Ideally, yes, but they may not come about as fast. If we know they are wrong, then we have an added necessity to come up with something better. [NEWLINE] [NEWLINE] [STARTQ] I would also still like an explanation of why you think it's better for us to (a) define suggestions in opposition to the Tea Party, with the risk that some of the bad Tea Party suggestions will be implemented than to (b) define suggestions in the absence of the Tea Party. [ENDQ] [NEWLINE] I think of this like trying to prosecute a criminal. It may be that the criminal is guilty and it's not too difficult to prove, but that doesn't give a prosecutor permission to half-ass the case. That's why the defense attorney is there, so the prosecution is forced to come up with an argument that can hold up against scrutiny. The same goal is ultimately reached as there would have been if there wasn't an opposition, but now the prosecution is forced to give an argument that is much more clear and indisputable- more reflective of the truth of the matter. [NEWLINE] [NEWLINE] Finding out what is right becomes easier when you can clearly identify what's wrong, and I think the tea party forces us to do that. </s>
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I just cannot see<mask> it benefits the child<mask> the religious ways, morals are not explained and are taught in an environment<mask> there is little questioning. [ENDQ] [NEWLINE] Raised Catholic here and schooled by Catholic institutions until college. There definitely wasn't an environment of little questioning, even<mask> being taught be nuns. I'm not saying that this is the experience without exception across all Catholic institutions.<mask>,<mask> it's done right, Catholic education is supposed to be an extension of scholasticism which is based upon reconciling faith with reason and giving heavy attention and emphasis to critical thinking. [NEWLINE] [NEWLINE] Furthermore, a sacramental rites of initiation are meant to reaffirm the voluntary nature of Catholic belief. I'm aware of the criticisms of baptizing infants, it may even be problematic to claim volition<mask> a child usually goes through communion, I won't address those particulars here.<mask>, Confirmation generally comes at an age<mask> the individual is certainly capable of thinking for oneself.<mask>, again<mask> done right, there is usually a process of examining one's faith leading into confirmation.<mask><mask>, from that point forward these sacraments are symbolically reaffirmed every time during different parts of the Mass. I get that you may still disagree,<mask> having reflected critically upon my personal experience I find your above criticism to be entirely anathema.</s>
Label encoding: <s> [STARTQ] I just cannot see how it benefits the child if the religious ways, morals are not explained and are taught in an environment where there is little questioning. [ENDQ] [NEWLINE] Raised Catholic here and schooled by Catholic institutions until college. There definitely wasn't an environment of little questioning, even when being taught be nuns. I'm not saying that this is the experience without exception across all Catholic institutions. However, if it's done right, Catholic education is supposed to be an extension of scholasticism which is based upon reconciling faith with reason and giving heavy attention and emphasis to critical thinking. [NEWLINE] [NEWLINE] Furthermore, a sacramental rites of initiation are meant to reaffirm the voluntary nature of Catholic belief. I'm aware of the criticisms of baptizing infants, it may even be problematic to claim volition when a child usually goes through communion, I won't address those particulars here. But, Confirmation generally comes at an age when the individual is certainly capable of thinking for oneself. Also, again if done right, there is usually a process of examining one's faith leading into confirmation. In fact, from that point forward these sacraments are symbolically reaffirmed every time during different parts of the Mass. I get that you may still disagree, but having reflected critically upon my personal experience I find your above criticism to be entirely anathema.</s>
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> redditors think "a business" is in America. [NEWLINE] [NEWLINE] There are<mask> many forms of businesses, ranging from sole proprietorships to national public corporations to national close partnerships, and they all should be treated differently. [NEWLINE] [NEWLINE] Your example fails to take into account the situation of a sole proprietor, which is a business on any reasonable observation, and<mask> the "owner" and the "business" are indistinguishable.<mask> the idea that the property belongs to the business and not the owner is nonsensical. [NEWLINE] [NEWLINE] In a sole proprietorship, no "business" exists for tax purposes. Depending on the city and state, you don't need any licenses to operate - for example - a web development service or a photography service. Their income is taxed on their personal income tax statement and nowhere else. The individual is entirely liable for any debts of the business. These kinds of businesses don't even need to advertise publicly and may only spread by word of mouth.<mask> the "business" were involved in a court case, the business brand name would not be used<mask> one of the parties to the case,<mask> the individual him/herself. [NEWLINE] [NEWLINE] In this situation, the individual is the business, and for the government to compel them to provide a service (or even sell property) seems to be way out of line.</s>
Label encoding: <s>I'm not sure what redditors think "a business" is in America. [NEWLINE] [NEWLINE] There are so many forms of businesses, ranging from sole proprietorships to national public corporations to national close partnerships, and they all should be treated differently. [NEWLINE] [NEWLINE] Your example fails to take into account the situation of a sole proprietor, which is a business on any reasonable observation, and yet the "owner" and the "business" are indistinguishable. Thus the idea that the property belongs to the business and not the owner is nonsensical. [NEWLINE] [NEWLINE] In a sole proprietorship, no "business" exists for tax purposes. Depending on the city and state, you don't need any licenses to operate - for example - a web development service or a photography service. Their income is taxed on their personal income tax statement and nowhere else. The individual is entirely liable for any debts of the business. These kinds of businesses don't even need to advertise publicly and may only spread by word of mouth. If the "business" were involved in a court case, the business brand name would not be used as one of the parties to the case, but the individual him/herself. [NEWLINE] [NEWLINE] In this situation, the individual is the business, and for the government to compel them to provide a service (or even sell property) seems to be way out of line.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, it's not magic. Humans are pretty versatile and find labor in most places. Technology shouldn't be seen<mask> making human labor redundant,<mask> instead<mask> freeing up humans to go do something else. This follows the pattern already established by history. [ENDQ] [NEWLINE] Sure,<mask> it's not<mask> simple<mask> that.  I'm all for people leaving shitty manual labor jobs and leading fulfilling lives<mask> potters and weavers and Ren Fair bards,<mask> money to live on has to come from somewhere<mask> you work out<mask> to make pottery and/or Ren Fair bardery earn cash.  We're getting away from the original question,<mask>. [NEWLINE] [NEWLINE] [STARTQ] It would take machines doing everything amazingly at a trivial cost... [ENDQ] [NEWLINE] Nope, it just requires machines doing *most* things<mask> well or better than a human for less money than it costs to pay the human.  And every year, we get closer to the point<mask> almost all non-creative jobs (creativity being really the only thing humans do better than machines) can be done by a machine. [NEWLINE] [NEWLINE] Machines doing more work means freeing up humans to do things that machines can't do, which is great, except that<mask> technology improves there will be a growing number of humans who can't do *anything* that a machine can't do.</s>
Label encoding: <s> [STARTQ] Firstly, it's not magic. Humans are pretty versatile and find labor in most places. Technology shouldn't be seen as making human labor redundant, but instead as freeing up humans to go do something else. This follows the pattern already established by history. [ENDQ] [NEWLINE] Sure, though it's not as simple as that.  I'm all for people leaving shitty manual labor jobs and leading fulfilling lives as potters and weavers and Ren Fair bards, but money to live on has to come from somewhere while you work out how to make pottery and/or Ren Fair bardery earn cash.  We're getting away from the original question, though. [NEWLINE] [NEWLINE] [STARTQ] It would take machines doing everything amazingly at a trivial cost... [ENDQ] [NEWLINE] Nope, it just requires machines doing *most* things as well or better than a human for less money than it costs to pay the human.  And every year, we get closer to the point where almost all non-creative jobs (creativity being really the only thing humans do better than machines) can be done by a machine. [NEWLINE] [NEWLINE] Machines doing more work means freeing up humans to do things that machines can't do, which is great, except that as technology improves there will be a growing number of humans who can't do *anything* that a machine can't do.</s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "Hate speech is speech that attacks a person or group on the basis of attributes" [ENDQ] [NEWLINE] Well, gosh, I don't really want to wade into the specifics of<mask> to draw the line between hate speech and just being mean,<mask> that can be a gnarly problem,<mask> you should at least quote the entire sentence from wikipieda: [NEWLINE] [NEWLINE] [STARTQ] speech that attacks a person or group on the basis of attributes such<mask> gender, ethnic origin, religion, race, disability, or sexual orientation. [ENDQ] [NEWLINE] I don't think this is a great definition, due to the vague use of "such<mask> ",<mask> *any* time you hate something, its going to "on the basis of attributes".<mask> your abridged version is certainly too broad. [NEWLINE] [NEWLINE] <mask> you need to do is argue<mask> "an interest in my little pony" should be given the same level of social and/or legal protection<mask> one's religion, sexual orientation, race, etc... [NEWLINE] [NEWLINE] <mask> for the response you got, it was clearly meant<mask> a joke. Its up to you to decide<mask> it was funny, tasteful, or appropriate<mask> I don't think you were supposed to take it seriously. I'm not even sure<mask> the guy that posted it could tell<mask> *you* were being serious.</s>
Label encoding: <s> [STARTQ] "Hate speech is speech that attacks a person or group on the basis of attributes" [ENDQ] [NEWLINE] Well, gosh, I don't really want to wade into the specifics of where to draw the line between hate speech and just being mean, as that can be a gnarly problem, but you should at least quote the entire sentence from wikipieda: [NEWLINE] [NEWLINE] [STARTQ] speech that attacks a person or group on the basis of attributes such as gender, ethnic origin, religion, race, disability, or sexual orientation. [ENDQ] [NEWLINE] I don't think this is a great definition, due to the vague use of "such as ", but *any* time you hate something, its going to "on the basis of attributes". So your abridged version is certainly too broad. [NEWLINE] [NEWLINE] What you need to do is argue why "an interest in my little pony" should be given the same level of social and/or legal protection as one's religion, sexual orientation, race, etc... [NEWLINE] [NEWLINE] As for the response you got, it was clearly meant as a joke. Its up to you to decide if it was funny, tasteful, or appropriate but I don't think you were supposed to take it seriously. I'm not even sure if the guy that posted it could tell if *you* were being serious.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> you're referring to is the problem of growing income inequality. Correct me<mask> I'm wrong<mask> you see corporations<mask> a main source of income inequality<mask> they are seeing large profits<mask> may not distribute the profits to the employees (or to a limited number of employees).<mask> this maybe in some sense be true, corporate money has to go somewhere. It doesn't just stay in the bank account of the CEO. For the most part, it is used to pay for stuff like R&amp;D, advertisement, material costs, and labor. All these things are good for the economy and good for individual people. [NEWLINE] [NEWLINE] Rather, I'd<mask><mask> it is large sums of money in individual hands that is causing problems. Yes, individuals can be, and often times are good employment drivers by making good investments in new technologies/companies.<mask>,<mask> they choose to spend their money on luxury items or hold their money, then it is essentially doing nothing.<mask>, money breeds money. A million dollar investment can yield much larger returns than a $10k investment. In this sense, upward mobility is very difficult and income inequality is increased. [NEWLINE] [NEWLINE] <mask>, the corporate tax is not a good tool to balance income inequality. Businesses drive employment, allowing people to live on their own rather than depend on government handouts. </s>
Label encoding: <s>So what you're referring to is the problem of growing income inequality. Correct me if I'm wrong but you see corporations as a main source of income inequality because they are seeing large profits but may not distribute the profits to the employees (or to a limited number of employees). While this maybe in some sense be true, corporate money has to go somewhere. It doesn't just stay in the bank account of the CEO. For the most part, it is used to pay for stuff like R&amp;D, advertisement, material costs, and labor. All these things are good for the economy and good for individual people. [NEWLINE] [NEWLINE] Rather, I'd argue that it is large sums of money in individual hands that is causing problems. Yes, individuals can be, and often times are good employment drivers by making good investments in new technologies/companies. However, if they choose to spend their money on luxury items or hold their money, then it is essentially doing nothing. Also, money breeds money. A million dollar investment can yield much larger returns than a $10k investment. In this sense, upward mobility is very difficult and income inequality is increased. [NEWLINE] [NEWLINE] Thus, the corporate tax is not a good tool to balance income inequality. Businesses drive employment, allowing people to live on their own rather than depend on government handouts. </s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You don't need "empathy." An argument that can only be made from a particular point of view isn't a very good argument. [NEWLINE] [NEWLINE] There is a lot of evidence that your race plays a huge role in your likelihood to succeed educationally in America.<mask>, you need to recognize all of the structural elements. Some types of systematic descrimination are very clear such<mask> racial barriers to employment, internment, and unequal protection and benefits. These were very clear in the '60s<mask> still very prominent today. [NEWLINE] [NEWLINE] Justice is still delved out unevenly. Money for public schools is still completely uneven, based upon the local property taxes in most areas. [NEWLINE] [NEWLINE] <mask> less talked about, and probably much more important, is the way that a culture's values shapes the way someone thinks and acts. Class values are very entrenched and there is a reason that there is very little social mobility anywhere in the developed world. [NEWLINE] [NEWLINE] Think of discrimination like this:<mask> someone like a teacher thinks you such<mask> they associate you with your race, it would seem like it is *easier* to make yourself an exception,<mask><mask> the standards have been lowered for you. The opposite is actually true. Your standards have been increased, and you will need to demonstrate beyond all doubt that you embody different traits than is prescribed by your race.</s>
Label encoding: <s>You don't need "empathy." An argument that can only be made from a particular point of view isn't a very good argument. [NEWLINE] [NEWLINE] There is a lot of evidence that your race plays a huge role in your likelihood to succeed educationally in America. However, you need to recognize all of the structural elements. Some types of systematic descrimination are very clear such as racial barriers to employment, internment, and unequal protection and benefits. These were very clear in the '60s but still very prominent today. [NEWLINE] [NEWLINE] Justice is still delved out unevenly. Money for public schools is still completely uneven, based upon the local property taxes in most areas. [NEWLINE] [NEWLINE] But less talked about, and probably much more important, is the way that a culture's values shapes the way someone thinks and acts. Class values are very entrenched and there is a reason that there is very little social mobility anywhere in the developed world. [NEWLINE] [NEWLINE] Think of discrimination like this: When someone like a teacher thinks you such because they associate you with your race, it would seem like it is *easier* to make yourself an exception, as if the standards have been lowered for you. The opposite is actually true. Your standards have been increased, and you will need to demonstrate beyond all doubt that you embody different traits than is prescribed by your race.</s>
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> we're arguing semantics? [ENDQ] [NEWLINE] No, I was just asking for<mask> you mean. [NEWLINE] [NEWLINE] [STARTQ] I certainly feel more compelled to serve the interests of my home nation rather than interests of a foreign nation. [ENDQ] [NEWLINE] This makes sense,<mask> it is in your selfish interest. I would feel the same even<mask> I lived in North Korea<mask> I don't want to die. [NEWLINE] [NEWLINE] [STARTQ] That doesn't mean any and all foreign interests are unimportant to me<mask> the hometown pull is certainly stronger. [ENDQ] [NEWLINE] The strangers that live near you effect you more then strangers who live far away. This is<mask> I want your definition of patriotism,<mask> selfishness isn't synonymous with patriotism to me. [NEWLINE] [NEWLINE] [STARTQ] I know I am<mask> I am today<mask> forces outside my control allowed me to be. Someone else built the roads and schools I used to achieve<mask> I have and<mask><mask> it's ok to be proud of that and return the favor in some small way. [ENDQ] [NEWLINE] Return the favor to whom? The guy who laid the asphalt? Walmart has provided me with countless goods for cheap and I don't feel I owe them any more then<mask> I've already paid them. [NEWLINE] [NEWLINE] [STARTQ] I simply have not had anything given to me by any other nation to nearly the same degree. [ENDQ] [NEWLINE] <mask> you weren't born there.</s>
Label encoding: <s> [STARTQ] So we're arguing semantics? [ENDQ] [NEWLINE] No, I was just asking for what you mean. [NEWLINE] [NEWLINE] [STARTQ] I certainly feel more compelled to serve the interests of my home nation rather than interests of a foreign nation. [ENDQ] [NEWLINE] This makes sense, as it is in your selfish interest. I would feel the same even if I lived in North Korea because I don't want to die. [NEWLINE] [NEWLINE] [STARTQ] That doesn't mean any and all foreign interests are unimportant to me but the hometown pull is certainly stronger. [ENDQ] [NEWLINE] The strangers that live near you effect you more then strangers who live far away. This is why I want your definition of patriotism, because selfishness isn't synonymous with patriotism to me. [NEWLINE] [NEWLINE] [STARTQ] I know I am where I am today because forces outside my control allowed me to be. Someone else built the roads and schools I used to achieve what I have and I think it's ok to be proud of that and return the favor in some small way. [ENDQ] [NEWLINE] Return the favor to whom? The guy who laid the asphalt? Walmart has provided me with countless goods for cheap and I don't feel I owe them any more then what I've already paid them. [NEWLINE] [NEWLINE] [STARTQ] I simply have not had anything given to me by any other nation to nearly the same degree. [ENDQ] [NEWLINE] Because you weren't born there.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0897, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, after reading this comment and your comment to the OP's reply, I'd like to change YOUR view that there is no misogyny within the common last-name-taking system. [NEWLINE] [NEWLINE] Individually, it may not be based in misogynistic thoughts.<mask> the fact of doing it the way it's currently done over most of the world certainly follows from a paternalistic system that we<mask> a society are still struggling to fully set aside. [NEWLINE] [NEWLINE] Just look at<mask> you said: [NEWLINE] [NEWLINE] [STARTQ] the dominant partner should be the one to take the name. [ENDQ] [NEWLINE] I believe you meant, the one whose name is used by both men should be the dominant partners' name.<mask>...<mask> do you believe a gay partnership must have a dominant partner?<mask> is the dominant partner the one whose name should be used... And<mask> the man's name is always used in a straight marriage, does that mean the man is implicitly the dominant partner? [NEWLINE] [NEWLINE] This isn't to invalidate your point about sharing a name making a statement and showing unity. My argument is about whose original name is used and<mask>.<mask> you fell in love with a woman who had a very strong feminist streak and said, "I'm only getting married<mask> the man agrees to take *my* name," would you agree to it?<mask> not<mask> not?</s>
Label encoding: <s>So, after reading this comment and your comment to the OP's reply, I'd like to change YOUR view that there is no misogyny within the common last-name-taking system. [NEWLINE] [NEWLINE] Individually, it may not be based in misogynistic thoughts. But the fact of doing it the way it's currently done over most of the world certainly follows from a paternalistic system that we as a society are still struggling to fully set aside. [NEWLINE] [NEWLINE] Just look at what you said: [NEWLINE] [NEWLINE] [STARTQ] the dominant partner should be the one to take the name. [ENDQ] [NEWLINE] I believe you meant, the one whose name is used by both men should be the dominant partners' name. So... Why do you believe a gay partnership must have a dominant partner? Why is the dominant partner the one whose name should be used... And if the man's name is always used in a straight marriage, does that mean the man is implicitly the dominant partner? [NEWLINE] [NEWLINE] This isn't to invalidate your point about sharing a name making a statement and showing unity. My argument is about whose original name is used and why. If you fell in love with a woman who had a very strong feminist streak and said, "I'm only getting married if the man agrees to take *my* name," would you agree to it? If not why not?</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In THIS language right NOW the word faggot has been stripped of all meaning other than a derogatory insult. [ENDQ] [NEWLINE] You've solved your own issue with this sentence; It has been stripped of all meaning - (usually) including the association with homosexuality.<mask> I call someone a "fucker", I am using the word in a derogatory manner, which is goes against the meaning of the word (ie. someone who fucks -<mask> can that possibly be an insult to say "that guy has a lot of sex"?). [NEWLINE] [NEWLINE] It's<mask> the word has shock value... it has stigma attached to it. [NEWLINE] [NEWLINE] It's<mask> rappers use "nigger"<mask> much. Obviously<mask> they're black, using "nigger"<mask> a pejorative is only insulting themselves,<mask><mask> do they do it?<mask> they've disassociated the word from its literal meaning, and are using it<mask> of the stigma attached. Stigma gets through; stigma punctuates the expletive.<mask> I say "that darned Johnson kid", it lacks the punch of saying "that fucking little shit Johnson kid". [NEWLINE] [NEWLINE] Faggot is being used the same way. Shock value and stigma - nothing to do with actual homosexuality anymore (at least, in the cases<mask> it's not being used<mask> a hate term). [NEWLINE] </s>
Label encoding: <s> [STARTQ] In THIS language right NOW the word faggot has been stripped of all meaning other than a derogatory insult. [ENDQ] [NEWLINE] You've solved your own issue with this sentence; It has been stripped of all meaning - (usually) including the association with homosexuality. If I call someone a "fucker", I am using the word in a derogatory manner, which is goes against the meaning of the word (ie. someone who fucks - how can that possibly be an insult to say "that guy has a lot of sex"?). [NEWLINE] [NEWLINE] It's because the word has shock value... it has stigma attached to it. [NEWLINE] [NEWLINE] It's why rappers use "nigger" so much. Obviously if they're black, using "nigger" as a pejorative is only insulting themselves, so why do they do it? Because they've disassociated the word from its literal meaning, and are using it because of the stigma attached. Stigma gets through; stigma punctuates the expletive. If I say "that darned Johnson kid", it lacks the punch of saying "that fucking little shit Johnson kid". [NEWLINE] [NEWLINE] Faggot is being used the same way. Shock value and stigma - nothing to do with actual homosexuality anymore (at least, in the cases where it's not being used as a hate term). [NEWLINE] </s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with saying things like "corporations should do..." is that it fails to recognize the core raison d'etre of most corporations, which is to make money. Those corporations that do not make money and those corporations who are not<mask> worried about making money are less likely to make enough money to grow and prosper, especially<mask> hard times come.<mask>, those corporations that keep a tighter hold on their cash or more likely to be able to weather storms, whereas those who are not<mask> tightfisted are less likely to do<mask>. This is not to say that all charitable corporations are bound to fail,<mask> that those that are less carefree with their money are more likely to survive and flourish<mask> rough times come. There are charitable organizations that may survive,<mask> they certainly have a harder time doing<mask>. This in turn means that you have more corporations surviving in the long term<mask> those corporations focus solely on making money for their owners/investors than<mask> those corporations focus on other things,<mask> increasing the amount of seemingly "evil and selfish corporations" who are in it solely for the money. It is not that all corporations are solely there to make money,<mask> that those who are focused solely on that end are better able to address challenges<mask> they arise, and are<mask> more likely to survive in the long term.</s>
Label encoding: <s>The problem with saying things like "corporations should do..." is that it fails to recognize the core raison d'etre of most corporations, which is to make money. Those corporations that do not make money and those corporations who are not as worried about making money are less likely to make enough money to grow and prosper, especially when hard times come. Thus, those corporations that keep a tighter hold on their cash or more likely to be able to weather storms, whereas those who are not so tightfisted are less likely to do so. This is not to say that all charitable corporations are bound to fail, but that those that are less carefree with their money are more likely to survive and flourish when rough times come. There are charitable organizations that may survive, but they certainly have a harder time doing so. This in turn means that you have more corporations surviving in the long term when those corporations focus solely on making money for their owners/investors than when those corporations focus on other things, thus increasing the amount of seemingly "evil and selfish corporations" who are in it solely for the money. It is not that all corporations are solely there to make money, but that those who are focused solely on that end are better able to address challenges as they arise, and are therefore more likely to survive in the long term.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sex doesn't need to lead to babies, this is the 21st century.  You seem to think that having sex means assuming "responsibility" for having a child, that's just bogus.  Only<mask> a mother is unwilling to have an abortion is procreation sometimes unavoidable. <mask> a woman engages in sex and is unwilling to have an abortion, maybe she should assume "responsibility" for the life she insists on creating. [NEWLINE] [NEWLINE] <mask> is insisting on linking sex and babies different than insisting on linking skydiving and falling to one's death?  We have parachutes,<mask> you refuse to use one then I suppose death is the result.  Would it not be reasonable, then, to simply not skydive<mask> you're morally against using a parachute? [NEWLINE] [NEWLINE] [STARTQ] The only reason people like OP are arguing for a change is<mask> they don't want to take responsibility for all possible outcomes of engaging in sex. [ENDQ] [NEWLINE] This is low.  You don't know OP, you only know<mask> information OP provides.  You don't know me either, you don't even know whether or not I'm a man.  Would you find it surprising to learn that financial abortions were first championed in US courts by a woman?  Would it surprise you to learn that mine is considered to be the feminist position?</s>
Label encoding: <s>Sex doesn't need to lead to babies, this is the 21st century.  You seem to think that having sex means assuming "responsibility" for having a child, that's just bogus.  Only if a mother is unwilling to have an abortion is procreation sometimes unavoidable.  If a woman engages in sex and is unwilling to have an abortion, maybe she should assume "responsibility" for the life she insists on creating. [NEWLINE] [NEWLINE] How is insisting on linking sex and babies different than insisting on linking skydiving and falling to one's death?  We have parachutes, if you refuse to use one then I suppose death is the result.  Would it not be reasonable, then, to simply not skydive if you're morally against using a parachute? [NEWLINE] [NEWLINE] [STARTQ] The only reason people like OP are arguing for a change is because they don't want to take responsibility for all possible outcomes of engaging in sex. [ENDQ] [NEWLINE] This is low.  You don't know OP, you only know what information OP provides.  You don't know me either, you don't even know whether or not I'm a man.  Would you find it surprising to learn that financial abortions were first championed in US courts by a woman?  Would it surprise you to learn that mine is considered to be the feminist position?</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I didn't imply that the claims have the same probability [ENDQ] [NEWLINE] You used it to support the claim that **<mask> the victim can't provide credible evidence for their claim, then they can't have a rational expectation of being believed with high probability.** It seems to me that this implies that the claims have similar probability—otherwise a person could very rationally expect to be believed<mask> claiming to have been raped even<mask> not expecting to be believed<mask> claiming to have been kidnapped by aliens. It especially seems that way<mask> your next post was all about<mask> people SHOULD factor the prior expectation of something's likelihood into their estimates of its probability in any particular case. [NEWLINE] [NEWLINE] <mask>, I do see<mask> you meant. I would have added the word "necessarily" into that sentence I bolded to clarify<mask> the subsequent sentence was meant to support. [NEWLINE] [NEWLINE] That might sound like nitpicking,<mask> personally I feel it's important to be<mask> careful<mask> possible<mask> you're laying out a logical argument, and by a different token,<mask><mask> talking about something that could be sensitive to someone<mask> taken the wrong way (*e.g.* a potential implication that the personal story she's just told is unbelievable). And I do still think that read either a little carelessly or very carefully your first sentence had the implication I originally suggested.</s>
Label encoding: <s> [STARTQ] I didn't imply that the claims have the same probability [ENDQ] [NEWLINE] You used it to support the claim that ** If the victim can't provide credible evidence for their claim, then they can't have a rational expectation of being believed with high probability.** It seems to me that this implies that the claims have similar probability—otherwise a person could very rationally expect to be believed when claiming to have been raped even while not expecting to be believed when claiming to have been kidnapped by aliens. It especially seems that way when your next post was all about how people SHOULD factor the prior expectation of something's likelihood into their estimates of its probability in any particular case. [NEWLINE] [NEWLINE] However, I do see what you meant. I would have added the word "necessarily" into that sentence I bolded to clarify what the subsequent sentence was meant to support. [NEWLINE] [NEWLINE] That might sound like nitpicking, but personally I feel it's important to be as careful as possible when you're laying out a logical argument, and by a different token, also when talking about something that could be sensitive to someone if taken the wrong way (*e.g.* a potential implication that the personal story she's just told is unbelievable). And I do still think that read either a little carelessly or very carefully your first sentence had the implication I originally suggested.</s>
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The shift in support for slavery had more to do with the civil war than a change in culture [ENDQ] [NEWLINE] Hardly explains the ban on slavery in Britain. Further, it doesn't explain<mask> individuals change their morality, which leads to societal change. [NEWLINE] [NEWLINE] [STARTQ] No one would agree to a moral that only benefits a few people. [ENDQ] [NEWLINE] You've missed my point. You argued that morality is based purely in self-interest.<mask><mask> this is true, then logically people would not extend morality to all people,<mask> it is against their self-interest. Your faulty premises then ruin your whole argument -<mask> you can no longer explain *any* moral change in society that goes against self-interest e.g. slavery, dissenters in Germany etc. [NEWLINE] [NEWLINE] <mask> you don't think we should extend morality to animals based on self-interest.<mask><mask> we have seen that your self-interest argument for morality is contradictory, it invalidates your view on animals. Morality can and should be extended to animals,<mask> one can reason that they are deserving of moral treatment. [NEWLINE] [NEWLINE] [STARTQ] Kind of like<mask> Americans at the time were probably taught to not care about killing Germans [ENDQ] [NEWLINE] People aren't just taught morality. People can come to their own conclusions - that it was justified to kill Germans<mask> of<mask> they represented.</s>
Label encoding: <s> [STARTQ] The shift in support for slavery had more to do with the civil war than a change in culture [ENDQ] [NEWLINE] Hardly explains the ban on slavery in Britain. Further, it doesn't explain why individuals change their morality, which leads to societal change. [NEWLINE] [NEWLINE] [STARTQ] No one would agree to a moral that only benefits a few people. [ENDQ] [NEWLINE] You've missed my point. You argued that morality is based purely in self-interest. But if this is true, then logically people would not extend morality to all people, as it is against their self-interest. Your faulty premises then ruin your whole argument - as you can no longer explain *any* moral change in society that goes against self-interest e.g. slavery, dissenters in Germany etc. [NEWLINE] [NEWLINE] So you don't think we should extend morality to animals based on self-interest. But because we have seen that your self-interest argument for morality is contradictory, it invalidates your view on animals. Morality can and should be extended to animals, because one can reason that they are deserving of moral treatment. [NEWLINE] [NEWLINE] [STARTQ] Kind of like how Americans at the time were probably taught to not care about killing Germans [ENDQ] [NEWLINE] People aren't just taught morality. People can come to their own conclusions - that it was justified to kill Germans because of what they represented.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is not accurate. [NEWLINE] [NEWLINE] [STARTQ] Before contemporary feminist movement was less than 10 years old, feminist thinkers began to talk about the way in which patriarchy was harmful to men. Without changing our fierce critique of male domination feminist politics expanded to include the recognition that patriarchy stripped men of certain rights, lmposing on them a sexist masculine identity. [ENDQ] [NEWLINE]... [NEWLINE] [NEWLINE] [STARTQ] Feminists who called for a recognition of men<mask> comrades in struggle never received mass media attention. Our theoretical work critiquing the demonization of men<mask> the enemy did not change the perspectives of women who were anti-male. And it was reaction to negative representations of manhood that led to the development of a men's movement that was anti-female. Writing about the "men's liberation movement" I called attention to the opportunism undergirding this movement: [ENDQ] [NEWLINE] [STARTQ] "These men identified themselves<mask> victims of sexism, working to liberate men. They identified rigid sex roles<mask> the primary source of their victimization, and,<mask> they wanted to change the notion of masculinity, they were not particularly concerned with their sexist exploitation and oppression of women." [ENDQ] [NEWLINE] [STARTQ] In many ways the men's movement mirrored the most negative aspects of the women's movement. [ENDQ] [NEWLINE] -- Bell Hooks, *Feminism is for Everybody*, pp 68-69</s>
Label encoding: <s>This is not accurate. [NEWLINE] [NEWLINE] [STARTQ] Before contemporary feminist movement was less than 10 years old, feminist thinkers began to talk about the way in which patriarchy was harmful to men. Without changing our fierce critique of male domination feminist politics expanded to include the recognition that patriarchy stripped men of certain rights, lmposing on them a sexist masculine identity. [ENDQ] [NEWLINE]... [NEWLINE] [NEWLINE] [STARTQ] Feminists who called for a recognition of men as comrades in struggle never received mass media attention. Our theoretical work critiquing the demonization of men as the enemy did not change the perspectives of women who were anti-male. And it was reaction to negative representations of manhood that led to the development of a men's movement that was anti-female. Writing about the "men's liberation movement" I called attention to the opportunism undergirding this movement: [ENDQ] [NEWLINE] [STARTQ] "These men identified themselves as victims of sexism, working to liberate men. They identified rigid sex roles as the primary source of their victimization, and, though they wanted to change the notion of masculinity, they were not particularly concerned with their sexist exploitation and oppression of women." [ENDQ] [NEWLINE] [STARTQ] In many ways the men's movement mirrored the most negative aspects of the women's movement. [ENDQ] [NEWLINE] -- Bell Hooks, *Feminism is for Everybody*, pp 68-69</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Disorders are defined and classified by the distress and/or harm they cause to themselves or others. [NEWLINE] [NEWLINE] At this point, homosexuals do not harm others with their orientation, and for the most part it does not cause themselves distress to be homosexual (once they are 'out' and accepting of<mask> they are). [NEWLINE] [NEWLINE] For a pedophile, they have the potential to harm others (the children), and<mask> they are not acting on those feelings and urges it is likely causing them some level of distress. <mask> either we would want to forcibly give treatment (<mask> they attempted to indulge in their urges), or they would likely want treatment to help them cope with the attraction. [NEWLINE] [NEWLINE] It's really that simple. [NEWLINE] [NEWLINE] edit: The main point to take away from my post is that in our society, we don't allow pedophiles to act out on their urges, and I would<mask><mask> not being able to act on your sexual urges would definitely cause distress.<mask> either pedophiles ARE acting out on their urges (which is basically rape in our society,<mask> children can't give consent), or they are NOT, in which case I would imagine they would be feeling distress and would like help. To my understanding, that's enough to classify it<mask> a disorder (<mask> I dislike the negative connotations that disorder has).</s>
Label encoding: <s>Disorders are defined and classified by the distress and/or harm they cause to themselves or others. [NEWLINE] [NEWLINE] At this point, homosexuals do not harm others with their orientation, and for the most part it does not cause themselves distress to be homosexual (once they are 'out' and accepting of what they are). [NEWLINE] [NEWLINE] For a pedophile, they have the potential to harm others (the children), and if they are not acting on those feelings and urges it is likely causing them some level of distress.  So either we would want to forcibly give treatment ( if they attempted to indulge in their urges), or they would likely want treatment to help them cope with the attraction. [NEWLINE] [NEWLINE] It's really that simple. [NEWLINE] [NEWLINE] edit: The main point to take away from my post is that in our society, we don't allow pedophiles to act out on their urges, and I would argue that not being able to act on your sexual urges would definitely cause distress. So either pedophiles ARE acting out on their urges (which is basically rape in our society, since children can't give consent), or they are NOT, in which case I would imagine they would be feeling distress and would like help. To my understanding, that's enough to classify it as a disorder ( though I dislike the negative connotations that disorder has).</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I would have to ask you, is<mask> is nigger still a racist word<mask> society doesn't consider it a problem to be black?  Calling someone a nigger is obviously different than saying "You are black."  Calling someone a homophobic slur doesn't just mean "You are gay" or<mask> have you. <mask> I come out I assume people would say that I am bi and that makes sense. <mask> you use a slur, you are saying "You are gay and that is a problem/you are a lesser human than me." [NEWLINE] [NEWLINE] I<mask> agree that words themselves aren't the biggest problem.  Obviously, hate and misunderstanding are.  Slurs in casual discourse is still a problem,<mask><mask> of the connotations of the word,<mask> the user's intentions. [NEWLINE] [NEWLINE] [STARTQ] I used the word "faggot" in this post,<mask> does my sexual orientation dictate whether it's a good reason? Does it become unacceptable<mask> I'm not LGBT? Is it okay<mask> I am? [ENDQ] [NEWLINE] <mask> I didn't make myself clear, this post is mostly referring to casual conversations. <mask> you say "He called me a faggot" or "Is it okay to say faggot" that's different.  Use of it<mask> an insult is the problem.  </s>
Label encoding: <s>What I would have to ask you, is how is nigger still a racist word when society doesn't consider it a problem to be black?  Calling someone a nigger is obviously different than saying "You are black."  Calling someone a homophobic slur doesn't just mean "You are gay" or what have you.  When I come out I assume people would say that I am bi and that makes sense.  When you use a slur, you are saying "You are gay and that is a problem/you are a lesser human than me." [NEWLINE] [NEWLINE] I also agree that words themselves aren't the biggest problem.  Obviously, hate and misunderstanding are.  Slurs in casual discourse is still a problem, however because of the connotations of the word, despite the user's intentions. [NEWLINE] [NEWLINE] [STARTQ] I used the word "faggot" in this post, but does my sexual orientation dictate whether it's a good reason? Does it become unacceptable if I'm not LGBT? Is it okay if I am? [ENDQ] [NEWLINE] If I didn't make myself clear, this post is mostly referring to casual conversations.  If you say "He called me a faggot" or "Is it okay to say faggot" that's different.  Use of it as an insult is the problem.  </s>
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0796, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 21 foot rule [ENDQ] [NEWLINE] The crux of your argument is the 21 foot rule. This is probably<mask> some police outside of the US are trained in martial arts -- it's highly effective in a lot of very close range, and somewhat more controlled scenarios (an officer knows<mask> he is about to apprehend someone). After that it falls apart. [NEWLINE] [NEWLINE] [STARTQ] I<mask> believe guns are a completely inefficient tool for anything other than combat with intent to kill [ENDQ] [NEWLINE] No kidding.<mask> you don't have a justifiable reason to kill someone, you can't use a gun. [NEWLINE] [NEWLINE] [STARTQ] idea that we need to guns to stop muggers and rapists is a fallacy we use to excuse our love affair with weaponry. [ENDQ] [NEWLINE] Loaded statement, troll meter has spiked. I'm not going to expect someone to endure rape or permanent bodily injury just<mask> the perpetrator can not get shot. Absolutely sickening you would brush that off<mask> readily. [NEWLINE] [NEWLINE] [STARTQ] investing in home security that prevents home invaders from actually entering the building, and options like tasers which can still be misused<mask> are less likely to cause loss of life. [ENDQ] [NEWLINE] Expecting people to just have thousands of dollars lying around for home security features? And tasers, those are often banned in the same places<mask> concealed carry is banned. </s>
Label encoding: <s> [STARTQ] 21 foot rule [ENDQ] [NEWLINE] The crux of your argument is the 21 foot rule. This is probably why some police outside of the US are trained in martial arts -- it's highly effective in a lot of very close range, and somewhat more controlled scenarios (an officer knows when he is about to apprehend someone). After that it falls apart. [NEWLINE] [NEWLINE] [STARTQ] I also believe guns are a completely inefficient tool for anything other than combat with intent to kill [ENDQ] [NEWLINE] No kidding. If you don't have a justifiable reason to kill someone, you can't use a gun. [NEWLINE] [NEWLINE] [STARTQ] idea that we need to guns to stop muggers and rapists is a fallacy we use to excuse our love affair with weaponry. [ENDQ] [NEWLINE] Loaded statement, troll meter has spiked. I'm not going to expect someone to endure rape or permanent bodily injury just so the perpetrator can not get shot. Absolutely sickening you would brush that off so readily. [NEWLINE] [NEWLINE] [STARTQ] investing in home security that prevents home invaders from actually entering the building, and options like tasers which can still be misused but are less likely to cause loss of life. [ENDQ] [NEWLINE] Expecting people to just have thousands of dollars lying around for home security features? And tasers, those are often banned in the same places where concealed carry is banned. </s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>EDIT: Nvm, I'm pretty sure you're just trolling after reading your responses. [NEWLINE] [NEWLINE] [STARTQ] Reason 1: Laziness - A lot of technology today encourages laziness.<mask>, this takes it to a whole new level. [ENDQ] [NEWLINE] <mask><mask>?<mask><mask> is working harder than you have to a virtue?<mask> people can consume more news from a monotone robot voice, isn't that better than no news?<mask> it's hilarious that you would complain about getting a weather report from a talking box<mask> opposed to a phone, computer, or television - a different kind of talking box. And most people don't have internal barometers to tell them the chance of rain for the next 24 hours. [NEWLINE] [NEWLINE] [STARTQ] Reason 2: Reliance on Technology [ENDQ] [NEWLINE] We're already reliant on our phones. This technology is nothing new except it sits on a table.<mask> anything this is just a backup, making you less reliant on existing technology. [NEWLINE] [NEWLINE] [STARTQ] Reason 3: Invasion of Privacy [ENDQ] [NEWLINE] Without more on the kind of technology, this fear is unfounded. Even<mask> it's always listening, whatever. Corporations' ability to aggregate all the data that comes out of your mouth and distill it into useful information is decades behind<mask> it would need to be to do anything useful, let alone harmful. </s>
Label encoding: <s>EDIT: Nvm, I'm pretty sure you're just trolling after reading your responses. [NEWLINE] [NEWLINE] [STARTQ] Reason 1: Laziness - A lot of technology today encourages laziness. However, this takes it to a whole new level. [ENDQ] [NEWLINE] So what? Since when is working harder than you have to a virtue? If people can consume more news from a monotone robot voice, isn't that better than no news? Also it's hilarious that you would complain about getting a weather report from a talking box as opposed to a phone, computer, or television - a different kind of talking box. And most people don't have internal barometers to tell them the chance of rain for the next 24 hours. [NEWLINE] [NEWLINE] [STARTQ] Reason 2: Reliance on Technology [ENDQ] [NEWLINE] We're already reliant on our phones. This technology is nothing new except it sits on a table. If anything this is just a backup, making you less reliant on existing technology. [NEWLINE] [NEWLINE] [STARTQ] Reason 3: Invasion of Privacy [ENDQ] [NEWLINE] Without more on the kind of technology, this fear is unfounded. Even if it's always listening, whatever. Corporations' ability to aggregate all the data that comes out of your mouth and distill it into useful information is decades behind where it would need to be to do anything useful, let alone harmful. </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. Read this [article]( [URL],9171,2136864,00.html). It's long,<mask> it's worthwhile. It's about the costs of American health care. I know you're not proposing the US system,<mask> the US system is both a market-driven and first-world system,<mask> it's worth your time to think about<mask>'s wrong with it. [NEWLINE] [NEWLINE] 2. Health care is not a discretionary good,<mask> it's less responsive to free markets than other goods.<mask> you need health you are going to pay<mask> you must to get it and<mask> not you are going to suffer pain, economic loss, or death.<mask> the forces that allow for competition are stifled. Market incentive is decreased in oligopolies, which is essentially<mask> the health business looks like. [NEWLINE] [NEWLINE] 3. In my experience, Americans who support their own health care system think people should never be forced to pay for a health system they don't need (libertarian view of government),<mask> they are<mask> those wealthy enough to have decent coverage,<mask> shielding them from the true costs of their own health care systems. [NEWLINE] [NEWLINE] Edit: sorry,<mask> pointed out the article is behind a paywall. It wasn't<mask> I read it. I'll try and find another link somewhere.</s>
Label encoding: <s>1. Read this [article]( [URL],9171,2136864,00.html). It's long, but it's worthwhile. It's about the costs of American health care. I know you're not proposing the US system, but the US system is both a market-driven and first-world system, so it's worth your time to think about what's wrong with it. [NEWLINE] [NEWLINE] 2. Health care is not a discretionary good, so it's less responsive to free markets than other goods. If you need health you are going to pay what you must to get it and if not you are going to suffer pain, economic loss, or death. So the forces that allow for competition are stifled. Market incentive is decreased in oligopolies, which is essentially what the health business looks like. [NEWLINE] [NEWLINE] 3. In my experience, Americans who support their own health care system think people should never be forced to pay for a health system they don't need (libertarian view of government), but they are also those wealthy enough to have decent coverage, thus shielding them from the true costs of their own health care systems. [NEWLINE] [NEWLINE] Edit: sorry, as pointed out the article is behind a paywall. It wasn't when I read it. I'll try and find another link somewhere.</s>
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When the technology arrives, it will be...<mask> AI hasnt gotten there<mask>... [NEWLINE] [NEWLINE] Im a mechanical engineer who works in a factory with robots, granted its not Watson state-of-the-art tech,<mask> they fuck up a lot more often than most people would think. Its just not<mask> simple<mask> "automate ALL of the "street-sweeper" jobs!!"...<mask> you automate it, u'll need on-site maintenance crews to fix it<mask> it breaks, which is a whole hell of a lot more expensive than a 16 yr old at min wage (unless ur in denmark or something), not to mention the initial capital costs would be staggering compared to current initial capital required to build a place like wendy's...<mask> back to the part<mask> they break down: 16 year olds dont stop working<mask> of a broken ball bearing or bug in the code, whereas automation does and<mask> now u are talking about lost profit during downtime required to fix the robots and customer dissatisfaction which effects future profits... [NEWLINE] [NEWLINE] There are tons of other things too,<mask><mask><mask> just the aforementioned is enough to deter implementation of automation in fast food, at least for the time being...one day we will have a bunch of I Robots running around,<mask> today is not that day nor is tomorrow</s>
Label encoding: <s>When the technology arrives, it will be... but AI hasnt gotten there yet... [NEWLINE] [NEWLINE] Im a mechanical engineer who works in a factory with robots, granted its not Watson state-of-the-art tech, but they fuck up a lot more often than most people would think. Its just not as simple as "automate ALL of the "street-sweeper" jobs!!"... if you automate it, u'll need on-site maintenance crews to fix it when it breaks, which is a whole hell of a lot more expensive than a 16 yr old at min wage (unless ur in denmark or something), not to mention the initial capital costs would be staggering compared to current initial capital required to build a place like wendy's... but back to the part where they break down: 16 year olds dont stop working because of a broken ball bearing or bug in the code, whereas automation does and so now u are talking about lost profit during downtime required to fix the robots and customer dissatisfaction which effects future profits... [NEWLINE] [NEWLINE] There are tons of other things too, but i think just the aforementioned is enough to deter implementation of automation in fast food, at least for the time being...one day we will have a bunch of I Robots running around, but today is not that day nor is tomorrow</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Trends do not imply that it occurs in this specific situation. [NEWLINE] I was judging by my personal experiences in hearing simmilar things IRL. [NEWLINE] [NEWLINE] 1: You can say the same thing, differently. And perception of the idea depends on<mask> you present it. [NEWLINE] 2: Yeah, judging by trends observed on my own. [NEWLINE] 3: Great. She's sexist. "<mask> he meant was that I did not dress<mask> provocatively or in<mask> revealing clothing<mask> my other female coworkers" This is pretty much implying "this male wanted me to dress sexy". [NEWLINE] 5: Most CEO's are male.<mask>... that confirms sexism? [NEWLINE] 6: Being different, doesn't mean unequal. Or<mask> did it go. [NEWLINE] [NEWLINE] [NEWLINE] I'm implying that people tend to play the discrimination card, instead of looking for reasons elsewhere. Be that sexism/racism/nazism. [NEWLINE] I perfectly understand<mask> is discrimination,<mask>,<mask> i was told that "We have a rule against "your people" ". Yeah,<mask><mask>. [NEWLINE] [NEWLINE] I will look closer to claims of sexism/racism<mask> i am not biased,<mask> rather quite critical of these situations. [NEWLINE] [NEWLINE] [NEWLINE] P.S. Not in a fully clear state of mind, might have missed a few logical links.</s>
Label encoding: <s>Trends do not imply that it occurs in this specific situation. [NEWLINE] I was judging by my personal experiences in hearing simmilar things IRL. [NEWLINE] [NEWLINE] 1: You can say the same thing, differently. And perception of the idea depends on how you present it. [NEWLINE] 2: Yeah, judging by trends observed on my own. [NEWLINE] 3: Great. She's sexist. " What he meant was that I did not dress as provocatively or in as revealing clothing as my other female coworkers" This is pretty much implying "this male wanted me to dress sexy". [NEWLINE] 5: Most CEO's are male. So... that confirms sexism? [NEWLINE] 6: Being different, doesn't mean unequal. Or how did it go. [NEWLINE] [NEWLINE] [NEWLINE] I'm implying that people tend to play the discrimination card, instead of looking for reasons elsewhere. Be that sexism/racism/nazism. [NEWLINE] I perfectly understand what is discrimination, as, when i was told that "We have a rule against "your people" ". Yeah, but yet. [NEWLINE] [NEWLINE] I will look closer to claims of sexism/racism because i am not biased, but rather quite critical of these situations. [NEWLINE] [NEWLINE] [NEWLINE] P.S. Not in a fully clear state of mind, might have missed a few logical links.</s>
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's just giving corpulent people a free pass<mask> it comes to not holding themselves accountable which seems kind of absurd...<mask> is someone who knows they're going to be 400lbs in a train seat any different from someone who is at a healthy size and carrying 3 personal bags, two of which they are holding in their lap<mask> that your space is clearly intruded upon and your comfort is reduced significantly? Using a train example<mask> you can carry on more luggage,<mask> it's the same idea. Lady with a big purse on her lap or<mask> have you. [NEWLINE] [NEWLINE] At some point, a person needs to be responsible for their immediate area, whether that radius is taken up by bags or corpulence or smell or nothing at all. It's THAT PERSON'S responsibility,<mask> a member of our society, to at least have some base-line courtesy. And honestly, most people do.<mask> you have extra bags, you're mindful of<mask> you carry them. MOST overweight people are very, very aware of themselves and the area they take up, and act<mask>, which is considerate.<mask> blaming the airline for not asking for girth measurements<mask> you purchase tickets<mask> that you can shift responsibility onto them<mask> 33A is now 33A-and-most-of-B seems wrong.</s>
Label encoding: <s>That's just giving corpulent people a free pass when it comes to not holding themselves accountable which seems kind of absurd... how is someone who knows they're going to be 400lbs in a train seat any different from someone who is at a healthy size and carrying 3 personal bags, two of which they are holding in their lap so that your space is clearly intruded upon and your comfort is reduced significantly? Using a train example because you can carry on more luggage, but it's the same idea. Lady with a big purse on her lap or what have you. [NEWLINE] [NEWLINE] At some point, a person needs to be responsible for their immediate area, whether that radius is taken up by bags or corpulence or smell or nothing at all. It's THAT PERSON'S responsibility, as a member of our society, to at least have some base-line courtesy. And honestly, most people do. If you have extra bags, you're mindful of how you carry them. MOST overweight people are very, very aware of themselves and the area they take up, and act accordingly, which is considerate. But blaming the airline for not asking for girth measurements when you purchase tickets so that you can shift responsibility onto them when 33A is now 33A-and-most-of-B seems wrong.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A police officers job is not to keep peoples safe. The officers job is to enforce the law. I am willing to wager that there was a tremendous amount of police presence at the parade. The NYPD force is massive and many were required in order to facilitate this event.<mask> is the harm in the officer in the video dancing for 15 seconds? The NYPD are not the Queens Guards. They are not required to stand still, march several times, and remain focused for hours. The odds of this officer spotting a terrorist plan unfolding and stopping it during those 15 seconds is **highly** unlikely.<mask>, police typically travel with partners incase a fight breaks out and they need assistance. Chances are that his partner was nearby. [NEWLINE] [NEWLINE] Furthermore, there has been very bad media attention on police lately. They have been exercising the broken windows police model. This means that obscene amounts of discretion are awarded to the police. They can arrest, attack, and shoot people with greater reliance on their own judgment. This creates a sense of us vs them.  The officer in the video is exercising the community policing model. This is<mask> the officers are integrated<mask> members of a community you can talk to and rely on. This is much better for police image, relations with public, and crime control in society overall. </s>
Label encoding: <s>A police officers job is not to keep peoples safe. The officers job is to enforce the law. I am willing to wager that there was a tremendous amount of police presence at the parade. The NYPD force is massive and many were required in order to facilitate this event. What is the harm in the officer in the video dancing for 15 seconds? The NYPD are not the Queens Guards. They are not required to stand still, march several times, and remain focused for hours. The odds of this officer spotting a terrorist plan unfolding and stopping it during those 15 seconds is **highly** unlikely. Also, police typically travel with partners incase a fight breaks out and they need assistance. Chances are that his partner was nearby. [NEWLINE] [NEWLINE] Furthermore, there has been very bad media attention on police lately. They have been exercising the broken windows police model. This means that obscene amounts of discretion are awarded to the police. They can arrest, attack, and shoot people with greater reliance on their own judgment. This creates a sense of us vs them.  The officer in the video is exercising the community policing model. This is when the officers are integrated as members of a community you can talk to and rely on. This is much better for police image, relations with public, and crime control in society overall. </s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of rap that gets popular either on the whole or within certain groups, is nowhere near<mask> difficult<mask> that song. I picked a couple of songs out at random.<mask> in essence<mask><mask> the average rap song is easier to perform than the vocal delivery of the average hardrock, metal or equivalent. [NEWLINE] [NEWLINE] And you know<mask>, there's nothing particularly wrong with that. People like<mask> they like. I personally think the faster and harder a rap song is, the less it actually seems to say. Much of the slower easier-to-perform rap is the stuff with meaningful lyrics. [NEWLINE] [NEWLINE] And<mask><mask> you should keep it mind, it is other people telling me to do this, I am not a rapper.<mask> yes I could pick this up and learn it in a few days to a week.<mask> you haven't picked up a guitar before you would not be able to learn a whole song (even something easy in a week). That is the point I'm treading around. In music any particular skill is not always equal in difficulty, even<mask> taste is subjective. [NEWLINE] [NEWLINE] <mask> don't take it like I'm saying "rap is not a musical skill" or some shit like that, any vocal delivery to music is, it just varies in delivery and execution.</s>
Label encoding: <s>A lot of rap that gets popular either on the whole or within certain groups, is nowhere near as difficult as that song. I picked a couple of songs out at random. but in essence I think the average rap song is easier to perform than the vocal delivery of the average hardrock, metal or equivalent. [NEWLINE] [NEWLINE] And you know what, there's nothing particularly wrong with that. People like what they like. I personally think the faster and harder a rap song is, the less it actually seems to say. Much of the slower easier-to-perform rap is the stuff with meaningful lyrics. [NEWLINE] [NEWLINE] And I think you should keep it mind, it is other people telling me to do this, I am not a rapper. But yes I could pick this up and learn it in a few days to a week. If you haven't picked up a guitar before you would not be able to learn a whole song (even something easy in a week). That is the point I'm treading around. In music any particular skill is not always equal in difficulty, even if taste is subjective. [NEWLINE] [NEWLINE] So don't take it like I'm saying "rap is not a musical skill" or some shit like that, any vocal delivery to music is, it just varies in delivery and execution.</s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Isn't it awfully convenient<mask> it<mask> means that you don't have to pay people for their work? [ENDQ] [NEWLINE] <mask> a visual artist, I consider it a terrible burden to have a copyright lawsuit dangling above my head, just<mask> I might subconsciously have remembered something and made a similar image, or simply<mask> I had a similar idea. I now have to spend time, money and energy covering my ass for lawsuits. Conversely, copyright doesn't protect me<mask> I don't have the resources to sniff out the entire internet and take down websites that happen to host crappy copies of my stuff. Copyright is a very awkward tool, and it's no longer relevant in an age<mask> it literally costs cents to copy the entire oeuvre of any artists work. It simply is unenforceable<mask> the volume is too big. Instead of trying to control the production of culture and reduce it to 19th century levels, we should embrace the technological changes (and the advantages they bring) and find better ways to get money to artists (after all, most artists *already* have to do another job to make their income regular - people living of the copyright of their own works alone are very rare). And all the money we're now wasting on copyright administration and litigation can be spent on artists too.</s>
Label encoding: <s> [STARTQ] Isn't it awfully convenient how it also means that you don't have to pay people for their work? [ENDQ] [NEWLINE] As a visual artist, I consider it a terrible burden to have a copyright lawsuit dangling above my head, just because I might subconsciously have remembered something and made a similar image, or simply because I had a similar idea. I now have to spend time, money and energy covering my ass for lawsuits. Conversely, copyright doesn't protect me because I don't have the resources to sniff out the entire internet and take down websites that happen to host crappy copies of my stuff. Copyright is a very awkward tool, and it's no longer relevant in an age where it literally costs cents to copy the entire oeuvre of any artists work. It simply is unenforceable because the volume is too big. Instead of trying to control the production of culture and reduce it to 19th century levels, we should embrace the technological changes (and the advantages they bring) and find better ways to get money to artists (after all, most artists *already* have to do another job to make their income regular - people living of the copyright of their own works alone are very rare). And all the money we're now wasting on copyright administration and litigation can be spent on artists too.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So I've read everyone else's response to OP's question and it just doesn't seem like everyone even remotely on the same page<mask><mask> he is talking about. the top response is from a guy who didn't drink in college. OP's age requirement are college students. [NEWLINE] [NEWLINE] I am definitely not the best person to ask<mask> I drink a lot,<mask> I've been the "drinks too much in one night" guy in college more than once. I was in a fraternity in college. Pledged<mask> a freshman, was a member til graduation 4.5 years later. The pre-game started at 8. Ran til 10, then we partied til 2, post game til 4. Most Fridays and Saturdays are blurry at a minimum, and at worst you forget everything from 9pm. [NEWLINE] [NEWLINE] I honestly don't think anyone here is going to change your mind.<mask> I was first initiated, we had a party, and for 2 weekends in a row, I blackout out hard. It wasn't until someone said something to me in person that I became concerned and changed, and personally<mask><mask> that's<mask>'s going to force you to change. Strangers on the internet are not going to coerce you into drinking less. The people you drink with will<mask>. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>So I've read everyone else's response to OP's question and it just doesn't seem like everyone even remotely on the same page as what he is talking about. the top response is from a guy who didn't drink in college. OP's age requirement are college students. [NEWLINE] [NEWLINE] I am definitely not the best person to ask because I drink a lot, but I've been the "drinks too much in one night" guy in college more than once. I was in a fraternity in college. Pledged as a freshman, was a member til graduation 4.5 years later. The pre-game started at 8. Ran til 10, then we partied til 2, post game til 4. Most Fridays and Saturdays are blurry at a minimum, and at worst you forget everything from 9pm. [NEWLINE] [NEWLINE] I honestly don't think anyone here is going to change your mind. When I was first initiated, we had a party, and for 2 weekends in a row, I blackout out hard. It wasn't until someone said something to me in person that I became concerned and changed, and personally I think that's what's going to force you to change. Strangers on the internet are not going to coerce you into drinking less. The people you drink with will though. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] God is actually doing things that we don't/can't understand<mask> are actually good for us<mask><mask> they appear to be things that cause suffering. [ENDQ] [NEWLINE] The moment you say that, you are throwing all of morality out of the window.<mask> can we say ANYTHING is truly good or evil<mask> we don't know the true way to determine whether something is good or evil? [NEWLINE] [NEWLINE] <mask><mask> Alice were a mass murderer and torturer, you'd have to admit that you have no idea<mask> Alice is a good or bad person,<mask> God (<mask><mask> the Bible)<mask> engages in mass murder and torture. [NEWLINE] [NEWLINE] [STARTQ] They ONLY way you can be OKAY with this is that<mask> you have faith that 'god knows<mask> he is doing and he is doing it in your best interest.' With out that assumption (taken on faith) the whole argument crumbles. [ENDQ] [NEWLINE] <mask> you appeal to "faith" then you admit that anyone can appeal to "faith" to justify *anything*, right?<mask> then it's not a real justification.<mask> it's not only false,<mask> dangerous,<mask> Alice can use the "faith" justification to rationalize her mass murder and torture. [NEWLINE] [NEWLINE] The only way to avoid this is to deny that "faith" is coherent.</s>
Label encoding: <s> [STARTQ] God is actually doing things that we don't/can't understand but are actually good for us even though they appear to be things that cause suffering. [ENDQ] [NEWLINE] The moment you say that, you are throwing all of morality out of the window. How can we say ANYTHING is truly good or evil if we don't know the true way to determine whether something is good or evil? [NEWLINE] [NEWLINE] So if Alice were a mass murderer and torturer, you'd have to admit that you have no idea if Alice is a good or bad person, because God ( according to the Bible) also engages in mass murder and torture. [NEWLINE] [NEWLINE] [STARTQ] They ONLY way you can be OKAY with this is that if you have faith that 'god knows what he is doing and he is doing it in your best interest.' With out that assumption (taken on faith) the whole argument crumbles. [ENDQ] [NEWLINE] When you appeal to "faith" then you admit that anyone can appeal to "faith" to justify *anything*, right? So then it's not a real justification. So it's not only false, but dangerous, because Alice can use the "faith" justification to rationalize her mass murder and torture. [NEWLINE] [NEWLINE] The only way to avoid this is to deny that "faith" is coherent.</s>
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd like to ask<mask> you want your view changed on this matter.<mask><mask><mask>, knowing that the crimes committed by the Nazis were the result of human nature is the single most humbling thought. That on some level, we're all capable of that, well that means that we must proceed with extreme cautiousness in our decisions. [NEWLINE] Refusing "humanity" to these crimes is really avoiding the tough questions,<mask><mask>. [NEWLINE] [NEWLINE] <mask><mask> others have explained,<mask><mask> your thinking is flawed. It's not, to me, human nature in its most basic state (which is something that's extremely arguable). It's a very specific, extreme form of hate, of tyranny, or<mask> you want to qualify it. It's<mask> we try our hardest to hold back that suddenly went free. [NEWLINE] [NEWLINE] <mask> to change your views: I do not believe you can call it a display of human nature. You can call it a display of human nature's flaw,<mask> you want, from the start (massive vote for the wrong candidate in a difficult context) till the end (deaths…). Human nature cannot be limited to killing and rage, otherwise<mask> others said we'd be in a constant state of warfare. The simple fact that we are a social creature contradicts this.</s>
Label encoding: <s>I'd like to ask why you want your view changed on this matter. In my opinion, knowing that the crimes committed by the Nazis were the result of human nature is the single most humbling thought. That on some level, we're all capable of that, well that means that we must proceed with extreme cautiousness in our decisions. [NEWLINE] Refusing "humanity" to these crimes is really avoiding the tough questions, imo. [NEWLINE] [NEWLINE] However as others have explained, I think your thinking is flawed. It's not, to me, human nature in its most basic state (which is something that's extremely arguable). It's a very specific, extreme form of hate, of tyranny, or however you want to qualify it. It's what we try our hardest to hold back that suddenly went free. [NEWLINE] [NEWLINE] So to change your views: I do not believe you can call it a display of human nature. You can call it a display of human nature's flaw, if you want, from the start (massive vote for the wrong candidate in a difficult context) till the end (deaths…). Human nature cannot be limited to killing and rage, otherwise as others said we'd be in a constant state of warfare. The simple fact that we are a social creature contradicts this.</s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0872, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd say it depends on<mask> you consider a "better" music generation. [NEWLINE] [NEWLINE] <mask><mask><mask> accessibility, I'd agree that we live in a generation<mask> its incredibly easy to find and enjoy new music.  A band can very easily get their music out there without any support of a label. [NEWLINE] [NEWLINE] <mask><mask><mask> creativity, I don't think our generation is more or less creative than past ones - it's all relative. [NEWLINE] [NEWLINE] <mask><mask><mask> profitability?  Now here's the big one. <mask><mask> that<mask> time goes on, it is becoming harder and harder to be profitable playing the music that *you* want to play.  Imagine a band like Pink Floyd coming out today, in terms of creativity.  Would that band be on the radio and on TV?  Hell no.  The only thing that garners huge success these days is very produced pop music.  Think about the number of *bands* who came out in the last 10 years that are still successful now.  I mean big time success.  Are there even 10?  5? [NEWLINE] [NEWLINE] Solo pop acts are dominant.  People not-really-singing at live shows are dominant. <mask> overall,<mask><mask> the industry<mask> a whole is<mask><mask> getting worse.</s>
Label encoding: <s>I'd say it depends on what you consider a "better" music generation. [NEWLINE] [NEWLINE] As far as accessibility, I'd agree that we live in a generation where its incredibly easy to find and enjoy new music.  A band can very easily get their music out there without any support of a label. [NEWLINE] [NEWLINE] As far as creativity, I don't think our generation is more or less creative than past ones - it's all relative. [NEWLINE] [NEWLINE] As far as profitability?  Now here's the big one.  I think that as time goes on, it is becoming harder and harder to be profitable playing the music that *you* want to play.  Imagine a band like Pink Floyd coming out today, in terms of creativity.  Would that band be on the radio and on TV?  Hell no.  The only thing that garners huge success these days is very produced pop music.  Think about the number of *bands* who came out in the last 10 years that are still successful now.  I mean big time success.  Are there even 10?  5? [NEWLINE] [NEWLINE] Solo pop acts are dominant.  People not-really-singing at live shows are dominant.  So overall, I think the industry as a whole is in fact getting worse.</s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A *lot* of analysis cannot be done in math-land. SAT is an NP-Complete problem,<mask> the average instance of SAT should be very difficult to solve.<mask> we have observed empirically that SAT solvers are **extremely** effective at solving many of the instances that we actually care about.<mask>? This is poorly understood. This means that the process of developing a better SAT solver is hugely empirical. Your algorithms are all still going to be exponential in the worst case<mask> they might give you huge wins in real life. [NEWLINE] [NEWLINE] Program Analysis is almost entirely made up of these sorts of situations. In general, basically anything you ever want to know about a program is undecidable.<mask> we have the entire field of static analysis that works on answering these questions. BDDs are one example of a data structure that works pretty well for a lot program analysis problems...<mask> is still exponential<mask> we only stay in pure math land. [NEWLINE] [NEWLINE] People love to throw out that Dijkstra quote<mask> in my experience almost none of the people who do are actually CS researchers. The world isn't quite the same<mask> it was<mask> Dijkstra was a big shot. [NEWLINE] [NEWLINE] Source: I'm a CS researcher who does a lot of work in Program Analysis. </s>
Label encoding: <s>A *lot* of analysis cannot be done in math-land. SAT is an NP-Complete problem, so the average instance of SAT should be very difficult to solve. But we have observed empirically that SAT solvers are **extremely** effective at solving many of the instances that we actually care about. Why? This is poorly understood. This means that the process of developing a better SAT solver is hugely empirical. Your algorithms are all still going to be exponential in the worst case but they might give you huge wins in real life. [NEWLINE] [NEWLINE] Program Analysis is almost entirely made up of these sorts of situations. In general, basically anything you ever want to know about a program is undecidable. Yet we have the entire field of static analysis that works on answering these questions. BDDs are one example of a data structure that works pretty well for a lot program analysis problems... but is still exponential if we only stay in pure math land. [NEWLINE] [NEWLINE] People love to throw out that Dijkstra quote but in my experience almost none of the people who do are actually CS researchers. The world isn't quite the same as it was when Dijkstra was a big shot. [NEWLINE] [NEWLINE] Source: I'm a CS researcher who does a lot of work in Program Analysis. </s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0776, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0862, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0632, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This argument is from a [letter]( [URL] ) to Robert E Lee from John Dalberg Acton (he's famous for "power corrupts, absolute power corrupts absolutely"). [NEWLINE] [NEWLINE] [STARTQ] Without presuming to decide the purely legal question, on which it seems evident to me from Madison's and Hamilton's papers that the Fathers of the Constitution were not agreed, I saw in State Rights the only availing check upon the absolutism of the sovereign will, and secession filled me with hope, not<mask> the destruction<mask><mask> the redemption of Democracy. The institutions of your Republic have not exercised on the old world the salutary and liberating influence which ought to have belonged to them, by reason of those defects and abuses of principle which the Confederate Constitution was expressly and wisely calculated to remedy. I believed that the example of that great Reform would have blessed all the races of mankind by establishing true freedom purged of the native dangers and disorders of Republics.<mask> I deemed that you were fighting the battles of our liberty, our progress, and our civilization; and I mourn for the stake which was lost at Richmond more deeply than I rejoice over that which was saved at Waterloo. [ENDQ] [NEWLINE] <mask> ban flying a flag that is a symbol of liberty, progress and western civilization? [NEWLINE] </s>
Label encoding: <s>This argument is from a [letter]( [URL] ) to Robert E Lee from John Dalberg Acton (he's famous for "power corrupts, absolute power corrupts absolutely"). [NEWLINE] [NEWLINE] [STARTQ] Without presuming to decide the purely legal question, on which it seems evident to me from Madison's and Hamilton's papers that the Fathers of the Constitution were not agreed, I saw in State Rights the only availing check upon the absolutism of the sovereign will, and secession filled me with hope, not as the destruction but as the redemption of Democracy. The institutions of your Republic have not exercised on the old world the salutary and liberating influence which ought to have belonged to them, by reason of those defects and abuses of principle which the Confederate Constitution was expressly and wisely calculated to remedy. I believed that the example of that great Reform would have blessed all the races of mankind by establishing true freedom purged of the native dangers and disorders of Republics. Therefore I deemed that you were fighting the battles of our liberty, our progress, and our civilization; and I mourn for the stake which was lost at Richmond more deeply than I rejoice over that which was saved at Waterloo. [ENDQ] [NEWLINE] Why ban flying a flag that is a symbol of liberty, progress and western civilization? [NEWLINE] </s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well,<mask> a high school math teacher I will tell you that<mask><mask> the existing high school math courses are pretty packed full already (I'm working on my curriculum map for Algebra and Geometry next year and I don't have an inch of wiggle room.) [NEWLINE] [NEWLINE] [NEWLINE] I would<mask><mask><mask> you need a lot longer than 2 weeks to achieve even the limited outcomes you're advocating.  To really get statistics right, you need to build up a lot of concepts in probability, which is VERY unintuitive for most students and given very short shrift in most secondary math curriculum.  Even<mask> you aren't doing t-tests by hand or anything like that, to even build up the intuition needed to understand something<mask> basic<mask> a "confidence interval" is going to take you at least a couple months (think about it-- you need at least a basic understanding of probability distributions, the law of large numbers, standard deviation-- none of which are simple ideas!).  I would<mask><mask> a semester course is the minimum to really get across the basic intuitions. <mask> you<mask> want kids to be able to do something like basic tests in Excel (which I would argue is important to helping the subject come to life), you're looking at up to a year.</s>
Label encoding: <s>Well, as a high school math teacher I will tell you that I think the existing high school math courses are pretty packed full already (I'm working on my curriculum map for Algebra and Geometry next year and I don't have an inch of wiggle room.) [NEWLINE] [NEWLINE] [NEWLINE] I would also argue that you need a lot longer than 2 weeks to achieve even the limited outcomes you're advocating.  To really get statistics right, you need to build up a lot of concepts in probability, which is VERY unintuitive for most students and given very short shrift in most secondary math curriculum.  Even if you aren't doing t-tests by hand or anything like that, to even build up the intuition needed to understand something as basic as a "confidence interval" is going to take you at least a couple months (think about it-- you need at least a basic understanding of probability distributions, the law of large numbers, standard deviation-- none of which are simple ideas!).  I would argue that a semester course is the minimum to really get across the basic intuitions.  If you also want kids to be able to do something like basic tests in Excel (which I would argue is important to helping the subject come to life), you're looking at up to a year.</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ]...all it tells me is that patriarchy is such a norm that many women help perpetuate it. [ENDQ] [NEWLINE] <mask> does it tell you that?<mask> do you draw that conclusion? [NEWLINE] [NEWLINE] It seems to me you're arguing there's no way women could prefer to vote for men *unless* they're victims of the patriarchy.<mask><mask>, it sounds like your starting point is the assumption that the patriarchy exists and has a strong influence, and<mask> every example of men being favored over women is proof of such. Are there any possible instances of men being favored for a certain task or role that are *not* attributable to the patriarchy? Conversely, are there any examples of women being favored for certain tasks or roles that are<mask> not the result of the patriarchy? [NEWLINE] [NEWLINE] [STARTQ]...sitcoms don't accurately represent US culture. I'm not sure<mask> you are assuming they do. [ENDQ] [NEWLINE] I'm not assuming they do. I'm suggesting they might. [NEWLINE] [NEWLINE] In the age of Ralph Kramden and Archie Bunker, the man was clearly viewed<mask> the head of the household.<mask> [census data confirms]( [URL].pdf)<mask>'s reflected in today's popular culture: the prevalence of female-headed households has grown much faster than that of men. [NEWLINE] </s>
Label encoding: <s> [STARTQ]...all it tells me is that patriarchy is such a norm that many women help perpetuate it. [ENDQ] [NEWLINE] Why does it tell you that? How do you draw that conclusion? [NEWLINE] [NEWLINE] It seems to me you're arguing there's no way women could prefer to vote for men *unless* they're victims of the patriarchy. In fact, it sounds like your starting point is the assumption that the patriarchy exists and has a strong influence, and therefore every example of men being favored over women is proof of such. Are there any possible instances of men being favored for a certain task or role that are *not* attributable to the patriarchy? Conversely, are there any examples of women being favored for certain tasks or roles that are also not the result of the patriarchy? [NEWLINE] [NEWLINE] [STARTQ]...sitcoms don't accurately represent US culture. I'm not sure why you are assuming they do. [ENDQ] [NEWLINE] I'm not assuming they do. I'm suggesting they might. [NEWLINE] [NEWLINE] In the age of Ralph Kramden and Archie Bunker, the man was clearly viewed as the head of the household. But [census data confirms]( [URL].pdf) what's reflected in today's popular culture: the prevalence of female-headed households has grown much faster than that of men. [NEWLINE] </s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0707, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> people are dying across the world due to civil war and starvation, the average person on the street knowing next to nothing about our political system, and the presence of hundreds of other actually important topics of conversation available,<mask> the hell is the big deal about games? [ENDQ] [NEWLINE] <mask> you're<mask> busy thinking about serious issues then<mask> did you have the intellectual time to post this question to CMV? Or even think about having your view changed? [NEWLINE] [NEWLINE] <mask> brains are a big place and you can't spend every moment of the day working on the intractable problems of the world - they're intractable<mask> they're too big to charge through intellectually. You have a hell of a lot of room in your head for random crap. [NEWLINE] [NEWLINE] Sports may get undue attention relative to serious stuff,<mask> that's<mask> they're *fun.* Humans get serious biochemical reward feedback for stuff they deem fun,<mask> they're going to spend<mask> much time having fun<mask> they can possibly manage. Meaning they'll occupy it with sports, videogames, and reddit<mask> much<mask> possible,<mask><mask> the problems in the world. [NEWLINE] [NEWLINE] They don't know any better, honestly. They're having fun and in the day-to-day of most people that's enough.</s>
Label encoding: <s> [STARTQ] When people are dying across the world due to civil war and starvation, the average person on the street knowing next to nothing about our political system, and the presence of hundreds of other actually important topics of conversation available, what the hell is the big deal about games? [ENDQ] [NEWLINE] If you're so busy thinking about serious issues then how did you have the intellectual time to post this question to CMV? Or even think about having your view changed? [NEWLINE] [NEWLINE] Because brains are a big place and you can't spend every moment of the day working on the intractable problems of the world - they're intractable because they're too big to charge through intellectually. You have a hell of a lot of room in your head for random crap. [NEWLINE] [NEWLINE] Sports may get undue attention relative to serious stuff, but that's because they're *fun.* Humans get serious biochemical reward feedback for stuff they deem fun, so they're going to spend as much time having fun as they can possibly manage. Meaning they'll occupy it with sports, videogames, and reddit as much as possible, regardless of the problems in the world. [NEWLINE] [NEWLINE] They don't know any better, honestly. They're having fun and in the day-to-day of most people that's enough.</s>
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> framework of argument would change your view,<mask> I'll go cultural.  Even<mask> I could prove they aren't much better off than whites, those facts don't put them in the "privellege/oppression" model. [NEWLINE] [NEWLINE] They don't fit<mask> most of them purposely migrated across the pacific ocean to start/work for a business.  Blacks had the opposite experience, Mexican's and latinos just have to hop a train and many are seen<mask> escaping corrupt countries that perhaps the US screwed with. <mask> already there is much less sadness around the story of an Asian in the US. [NEWLINE] [NEWLINE] <mask> I'd<mask><mask> they are "oppressed" in media a lot.  I'd<mask><mask> latinos should be more represented<mask> they are a larger ethnic group in the US than blacks,<mask> that's irrelevant.  It is very hard to cast an asian without that being the reason they were cast.  Walking Dead did it, which was cool,<mask> 9 out of 10 times, the Asian guy is supposed to be an Asian guy and have Asian characteristics of pride and insight. [NEWLINE] [NEWLINE] This probably isn't a well stated enough argument to change your view,<mask> its the only framework I can see working.</s>
Label encoding: <s>I'm not sure what framework of argument would change your view, but I'll go cultural.  Even if I could prove they aren't much better off than whites, those facts don't put them in the "privellege/oppression" model. [NEWLINE] [NEWLINE] They don't fit because most of them purposely migrated across the pacific ocean to start/work for a business.  Blacks had the opposite experience, Mexican's and latinos just have to hop a train and many are seen as escaping corrupt countries that perhaps the US screwed with.  So already there is much less sadness around the story of an Asian in the US. [NEWLINE] [NEWLINE] But I'd argue that they are "oppressed" in media a lot.  I'd argue that latinos should be more represented since they are a larger ethnic group in the US than blacks, but that's irrelevant.  It is very hard to cast an asian without that being the reason they were cast.  Walking Dead did it, which was cool, but 9 out of 10 times, the Asian guy is supposed to be an Asian guy and have Asian characteristics of pride and insight. [NEWLINE] [NEWLINE] This probably isn't a well stated enough argument to change your view, but its the only framework I can see working.</s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0870, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A first date is an interview, to put it bluntly. It's a screening process. Unless you're just looking for flings and one-night-stands, the point is to gauge longer-term compatibility. [ENDQ] [NEWLINE] <mask><mask> you have a different view of a date than many people. Everything about a date is totally arbitrary and dates only happen<mask> "that's the way things are done" en route to finding a partner. [NEWLINE] [NEWLINE] You don't talk about politics or religion on a first date for the same reason you don't ask someone about their sexual appetite or kinks on a first date:<mask> it is considered impolite and its not<mask> things are done. A first date is the date that society has ordained<mask> the date<mask> you examine a person superficially. It is impolite to ask someone you just met personal questions<mask> you can do it<mask> you want,<mask> it won't actually get you anywhere in most cases. [NEWLINE] [NEWLINE] Sure taking your time to slowly enter someone's trust and get to know them better isn't the most efficient way of doing things,<mask><mask> you want efficiency<mask> don't you just mail out a survey? [NEWLINE] [NEWLINE] Summary:<mask><mask> you misunderstand the purpose of a first date. </s>
Label encoding: <s> [STARTQ] A first date is an interview, to put it bluntly. It's a screening process. Unless you're just looking for flings and one-night-stands, the point is to gauge longer-term compatibility. [ENDQ] [NEWLINE] I think you have a different view of a date than many people. Everything about a date is totally arbitrary and dates only happen because "that's the way things are done" en route to finding a partner. [NEWLINE] [NEWLINE] You don't talk about politics or religion on a first date for the same reason you don't ask someone about their sexual appetite or kinks on a first date: because it is considered impolite and its not how things are done. A first date is the date that society has ordained as the date where you examine a person superficially. It is impolite to ask someone you just met personal questions so you can do it if you want, but it won't actually get you anywhere in most cases. [NEWLINE] [NEWLINE] Sure taking your time to slowly enter someone's trust and get to know them better isn't the most efficient way of doing things, but if you want efficiency why don't you just mail out a survey? [NEWLINE] [NEWLINE] Summary: I think you misunderstand the purpose of a first date. </s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> Exactly. I was going to make my own post,<mask> I would like to add. [NEWLINE] [NEWLINE] Imagine a new medical device was made that performed some essential function perfectly (perhaps it is an artificial organ),<mask> that it required some component that is only scarcely available (like a rare earth metal). Maybe you could only make 10k of these a year based on current mining yield. All other devices are inferior in at least one regard. [NEWLINE] [NEWLINE] Clearly, you cannot promise the device to all who could benefit<mask> that means, say, 2 million initial patients with 100k per year additional.<mask> it cannot be universally guaranteed. [NEWLINE] [NEWLINE] <mask>, suppose a Steve-Jobs-type has the wherewithal to get this device 100% on his own. It would be immoral to restrict him from doing<mask>,<mask> it forces a less-optimal health choice on him. [NEWLINE] [NEWLINE] <mask>, given extreme parameters, it is clear that some healthcare may be available without universal coverage,<mask> the gut-reaction "unfairness" of the striation of care is not justified. The determination<mask> to whether care should be universal or not can only debated in the context of available resources, and the position that all care should be universal is nearly indefensible. </s>
Label encoding: <s> Exactly. I was going to make my own post, but I would like to add. [NEWLINE] [NEWLINE] Imagine a new medical device was made that performed some essential function perfectly (perhaps it is an artificial organ), but that it required some component that is only scarcely available (like a rare earth metal). Maybe you could only make 10k of these a year based on current mining yield. All other devices are inferior in at least one regard. [NEWLINE] [NEWLINE] Clearly, you cannot promise the device to all who could benefit if that means, say, 2 million initial patients with 100k per year additional. So it cannot be universally guaranteed. [NEWLINE] [NEWLINE] However, suppose a Steve-Jobs-type has the wherewithal to get this device 100% on his own. It would be immoral to restrict him from doing so, as it forces a less-optimal health choice on him. [NEWLINE] [NEWLINE] So, given extreme parameters, it is clear that some healthcare may be available without universal coverage, so the gut-reaction "unfairness" of the striation of care is not justified. The determination as to whether care should be universal or not can only debated in the context of available resources, and the position that all care should be universal is nearly indefensible. </s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] wouldn't more exercise<mask> cause you to eat more? [ENDQ] [NEWLINE] Something that sticks with me from Cycling is "<mask> you carb up before the hill, you will climb the hill faster &amp; ultimately burn more calories". [NEWLINE] [NEWLINE] "More exercise" changes your *metabolism*. No two people retain/burn/excrete the exact same # of calories from a given food. Simple calorie counting isn't very accurate for any given individuals' weight changes. [NEWLINE] [NEWLINE] I realize you're view was already changed. Just wanted to throw that out there. I went for a year unable to do (normal) cardio<mask> of a back injury. I walked my dog 1 hr a day, every day, and watched<mask> I ate -<mask> I still went from a "normal" 175 to an alarming 205. After I healed, Cardio brought that weight back down in a couple months,<mask><mask> I no longer 'diet' myself. The amount of dieting it would require for 45 minute walks to be enough exercise is significant.<mask> 45 minutes of higher-intensity exercise might make a world of difference with much less dietary restrictions. [NEWLINE] [NEWLINE] More power to you, good luck, sorry<mask> this isn't really an "argument" reply. </s>
Label encoding: <s> [STARTQ] wouldn't more exercise also cause you to eat more? [ENDQ] [NEWLINE] Something that sticks with me from Cycling is " if you carb up before the hill, you will climb the hill faster &amp; ultimately burn more calories". [NEWLINE] [NEWLINE] "More exercise" changes your *metabolism*. No two people retain/burn/excrete the exact same # of calories from a given food. Simple calorie counting isn't very accurate for any given individuals' weight changes. [NEWLINE] [NEWLINE] I realize you're view was already changed. Just wanted to throw that out there. I went for a year unable to do (normal) cardio because of a back injury. I walked my dog 1 hr a day, every day, and watched what I ate - but I still went from a "normal" 175 to an alarming 205. After I healed, Cardio brought that weight back down in a couple months, even though I no longer 'diet' myself. The amount of dieting it would require for 45 minute walks to be enough exercise is significant. But 45 minutes of higher-intensity exercise might make a world of difference with much less dietary restrictions. [NEWLINE] [NEWLINE] More power to you, good luck, sorry if this isn't really an "argument" reply. </s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The use of "privilege" in a debate or philosophical evaluation of a concept/idea, from<mask> I've seen, has two facets. [NEWLINE] [NEWLINE] * One party in the discussion is attempting to use "privilege"<mask> a way to invalidate or discredit their opponent. [NEWLINE] [NEWLINE] This use is,<mask> you say, inexcusable. It's a dirty tactic and honestly I'd consider it to fit the logical fallacy "[Poisoning the Well]( [URL] )". Being privileged does not invalidate any contributions you make to a discussion. [NEWLINE] [NEWLINE] That being said... [NEWLINE] [NEWLINE] *<mask> you are privileged, there are concepts that won't be understood naturally. The important part is to recognize that privilege, realize<mask> it is like to have it versus not have it, and have the perspective to understand issues from an unprivileged viewpoint. [NEWLINE] [NEWLINE] This becomes an issue,<mask><mask> a society it's only very recently that the debate about privilege has even come up. It's not about guilt, it's about being able to see privilege<mask> a bias, and being able to separate that bias to gain an alternate perspective. This is compounded by people who misuse this discussion tool<mask> a way to attack their opponents for something they cannot help.</s>
Label encoding: <s>The use of "privilege" in a debate or philosophical evaluation of a concept/idea, from what I've seen, has two facets. [NEWLINE] [NEWLINE] * One party in the discussion is attempting to use "privilege" as a way to invalidate or discredit their opponent. [NEWLINE] [NEWLINE] This use is, as you say, inexcusable. It's a dirty tactic and honestly I'd consider it to fit the logical fallacy "[Poisoning the Well]( [URL] )". Being privileged does not invalidate any contributions you make to a discussion. [NEWLINE] [NEWLINE] That being said... [NEWLINE] [NEWLINE] * If you are privileged, there are concepts that won't be understood naturally. The important part is to recognize that privilege, realize what it is like to have it versus not have it, and have the perspective to understand issues from an unprivileged viewpoint. [NEWLINE] [NEWLINE] This becomes an issue, because as a society it's only very recently that the debate about privilege has even come up. It's not about guilt, it's about being able to see privilege as a bias, and being able to separate that bias to gain an alternate perspective. This is compounded by people who misuse this discussion tool as a way to attack their opponents for something they cannot help.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> makes this brand of feminism worse or better than the tumblr style of feminism?<mask> about the armchair feminist?<mask> about the topless through the Vatican feminist (rad-fem<mask><mask> they are called)? [ENDQ] [NEWLINE] I would say it's about on-par with the tumblr style of feminism,<mask> at least the tumblr feminism at least sometimes gets it right,<mask> being<mask> extreme that they miss the point. I would identify 'tumblr feminism'<mask> being something like 'Buzzfeed feminism'<mask> on the opposite side of the spectrum.<mask> Buzzfeed is too simplistic, tumblr is too misadventurously radical. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> makes that particular brand of feminism worse or better? Buzzfeed is competing with a lot of different feminist groups, from the more conservative league of women voters, to the more radical like rad fem or tumblr. [ENDQ] [NEWLINE] The same thing that makes any strand of thought worse or better- coherence, reasonableness, ability to be backed up by facts/data/statistics. I would<mask><mask> Buzzfeed Feminism is often more based in sentiment or the 'trendiness' of feminism than<mask><mask>, analysis or a genuine concern for feminist issues. [NEWLINE] </s>
Label encoding: <s> [STARTQ] What makes this brand of feminism worse or better than the tumblr style of feminism? What about the armchair feminist? What about the topless through the Vatican feminist (rad-fem I think they are called)? [ENDQ] [NEWLINE] I would say it's about on-par with the tumblr style of feminism, although at least the tumblr feminism at least sometimes gets it right, despite being so extreme that they miss the point. I would identify 'tumblr feminism' as being something like 'Buzzfeed feminism' but on the opposite side of the spectrum. Where Buzzfeed is too simplistic, tumblr is too misadventurously radical. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] What makes that particular brand of feminism worse or better? Buzzfeed is competing with a lot of different feminist groups, from the more conservative league of women voters, to the more radical like rad fem or tumblr. [ENDQ] [NEWLINE] The same thing that makes any strand of thought worse or better- coherence, reasonableness, ability to be backed up by facts/data/statistics. I would argue that Buzzfeed Feminism is often more based in sentiment or the 'trendiness' of feminism than in fact, analysis or a genuine concern for feminist issues. [NEWLINE] </s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In a heat death scenario there is no available thermodynamic energy that can be harvested to perform work, in particular to fuel consciousness. Unless you have a way to preserve consciousness that requires zero energy transfer, then the only thing you can fall back on is quantum fluctuations spontaneously providing the energy needed.<mask>, were sufficient energy to arise from that event, then you would no longer be in a heat death scenario. Imagine the big bang<mask> such an event -- it's a bit out of the realm of our current understanding. Perhaps a big bang is an inevitability given empty space and time,<mask> that the universe itself operates in a cycle. In order to'survive' across these cycles, it would be necessary to preserve order (i.e. consciousness) for an arbitrary time in heat death (physically impossible). The only sense of continuity are the physical laws of the universe themselves,<mask> they will dictate the phenomena and formations of the various fields<mask> they transition from low to high entropy.<mask> you can alter or define these laws, you can escape mortality by ensuring the evolution of your consciousness. Essentially becoming god. Perhaps this has already happened. Maybe you are god. Other than that, I'd say there is no chance to survive. </s>
Label encoding: <s>In a heat death scenario there is no available thermodynamic energy that can be harvested to perform work, in particular to fuel consciousness. Unless you have a way to preserve consciousness that requires zero energy transfer, then the only thing you can fall back on is quantum fluctuations spontaneously providing the energy needed. Though, were sufficient energy to arise from that event, then you would no longer be in a heat death scenario. Imagine the big bang as such an event -- it's a bit out of the realm of our current understanding. Perhaps a big bang is an inevitability given empty space and time, so that the universe itself operates in a cycle. In order to'survive' across these cycles, it would be necessary to preserve order (i.e. consciousness) for an arbitrary time in heat death (physically impossible). The only sense of continuity are the physical laws of the universe themselves, as they will dictate the phenomena and formations of the various fields as they transition from low to high entropy. If you can alter or define these laws, you can escape mortality by ensuring the evolution of your consciousness. Essentially becoming god. Perhaps this has already happened. Maybe you are god. Other than that, I'd say there is no chance to survive. </s>
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Whether they are biologically unfit or they are simply not taught the skills is not really important here.<mask> we are to take the general population of the US and teach them anything it's going to need to be dumbed down to the point that it will be worthless. My daughter is a HS senior and her economics class was barely economic's vocabulary. [NEWLINE] [NEWLINE] Many, many, many children are dumb. Whether they are dumb<mask> they aren't biologically able or<mask> our culture doesn't value education or<mask> they're household doesn't value education isn't really the issue. the fact that they are dumb is. [NEWLINE] [NEWLINE] [STARTQ] The 15% figure for full literacy, equivalent to a university undergraduate level, is consistent with the notion that the "average" American reads at a 7th or 8th grade level which is<mask> consistent with recommendations, guidelines, and norms of readability for medication directions, product information, and popular...  Wikipedia on literacy in the US. [ENDQ] [NEWLINE] [Here's a tip on writing for an 8th]( [URL] /) grade reading level. Which will only get you to half of the population. Want to reach the bottom half? Try explaining thinking critically only using terms a 5th grader can understand. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Whether they are biologically unfit or they are simply not taught the skills is not really important here. If we are to take the general population of the US and teach them anything it's going to need to be dumbed down to the point that it will be worthless. My daughter is a HS senior and her economics class was barely economic's vocabulary. [NEWLINE] [NEWLINE] Many, many, many children are dumb. Whether they are dumb because they aren't biologically able or because our culture doesn't value education or because they're household doesn't value education isn't really the issue. the fact that they are dumb is. [NEWLINE] [NEWLINE] [STARTQ] The 15% figure for full literacy, equivalent to a university undergraduate level, is consistent with the notion that the "average" American reads at a 7th or 8th grade level which is also consistent with recommendations, guidelines, and norms of readability for medication directions, product information, and popular...  Wikipedia on literacy in the US. [ENDQ] [NEWLINE] [Here's a tip on writing for an 8th]( [URL] /) grade reading level. Which will only get you to half of the population. Want to reach the bottom half? Try explaining thinking critically only using terms a 5th grader can understand. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I cant speak for everyone,<mask> the thing that drives me to do more, achieve<mask> much<mask> possible is providing opportunities for my children, specifically in education. I am fairly bitter I couldnt afford to go to school<mask> I wanted, and I couldnt afford to get a PhD. Sucks,<mask> instead of bitching I am determined to make sure all my kids get to go to whatever school they want, and study<mask> they want with their only concern<mask> there being to better themselves. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] From a purely utilitarian POV, this desire to give this opportunity to my kids makes me much much more productive. Its one of the reason I want to start my own business, its<mask> I work<mask> much<mask> I do. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Now,<mask> all people were gaurunteed the right to a quality education, I might be signing a different tune,<mask> until that happens I am gonna bust my ass to provide this for them. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] There is<mask> the idea of tradition, and pride in property. This is directed at the family farms, being passed down from generation to generation.<mask> I dont personally get into that kind of stuff,<mask><mask> the fact that some people do should be respected</s>
Label encoding: <s>I cant speak for everyone, but the thing that drives me to do more, achieve as much as possible is providing opportunities for my children, specifically in education. I am fairly bitter I couldnt afford to go to school where I wanted, and I couldnt afford to get a PhD. Sucks, but instead of bitching I am determined to make sure all my kids get to go to whatever school they want, and study what they want with their only concern while there being to better themselves. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] From a purely utilitarian POV, this desire to give this opportunity to my kids makes me much much more productive. Its one of the reason I want to start my own business, its why I work as much as I do. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] Now, if all people were gaurunteed the right to a quality education, I might be signing a different tune, but until that happens I am gonna bust my ass to provide this for them. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] There is also the idea of tradition, and pride in property. This is directed at the family farms, being passed down from generation to generation. While I dont personally get into that kind of stuff, I think the fact that some people do should be respected</s>
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is no reason to shit on 3D<mask> it is used in the ways you said, the real difference is that it opens our possibilities to make a better visual experience.  Does it mean that it doesn't get overused in a gimmicky sort of way? Not at all.  Does that something gets used poorly mean it cannot be a great tool? no. [NEWLINE] [NEWLINE] [NEWLINE] I used to bartend, and I learned to make cocktails in a very traditional sort of way, and I hated the idea of flavored vodka, basically<mask> I believed that I could make a better tasting cocktail using skill over the shortcut of pre-flavored alcohol, that was probably made<mask> 16 year olds would like it.  I realized that later on that yes it is annoying<mask> you see bartenders make lazy drinks,<mask> its not the vodka's fault, its the lazy people who decided to use lazy methods instead of create actual quality drinks, and that there is a use for these flavored vodkas in making things that would not be possible without them; 3d is the same way. [NEWLINE] [NEWLINE] [NEWLINE] TL;DR: Blaming something for it being used poorly is pointless,<mask> it being inanimate cannot force its own usage.</s><pad>
Label encoding: <s>There is no reason to shit on 3D because it is used in the ways you said, the real difference is that it opens our possibilities to make a better visual experience.  Does it mean that it doesn't get overused in a gimmicky sort of way? Not at all.  Does that something gets used poorly mean it cannot be a great tool? no. [NEWLINE] [NEWLINE] [NEWLINE] I used to bartend, and I learned to make cocktails in a very traditional sort of way, and I hated the idea of flavored vodka, basically because I believed that I could make a better tasting cocktail using skill over the shortcut of pre-flavored alcohol, that was probably made so 16 year olds would like it.  I realized that later on that yes it is annoying when you see bartenders make lazy drinks, but its not the vodka's fault, its the lazy people who decided to use lazy methods instead of create actual quality drinks, and that there is a use for these flavored vodkas in making things that would not be possible without them; 3d is the same way. [NEWLINE] [NEWLINE] [NEWLINE] TL;DR: Blaming something for it being used poorly is pointless, since it being inanimate cannot force its own usage.</s><pad>
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have the same habits<mask> you, OP. I do this same thing at the end of virtually every day --<mask> not? I've gotten my work done, I've taken care of my responsibilities, I've earned some time to relax. [NEWLINE] [NEWLINE] Now<mask> someone who<mask> partakes in the same capacity you do, I<mask> know<mask> easy it can be for this to become more than just something you look forward to. Anytime I'm high, it's impeding me from being proactive. More and more<mask> I smoke weed nightly, I come to expect it. My day is based on it. Whatever I do is all done for the purpose of getting to reward myself with weed at the end of the day. [NEWLINE] [NEWLINE] In short, smoking weed at the end of every day has different implications than rewarding yourself with, say, a bowl of ice cream (which might,<mask><mask>, accompany the former). I can't remember<mask> comedian or show I heard it from,<mask> there's a quote out there that goes, "Smoking pot makes you okay with being bored." There's no real argument *against* this habit<mask> just knowing that you could always be doing other things<mask> you're smoking weed.</s>
Label encoding: <s>I have the same habits as you, OP. I do this same thing at the end of virtually every day -- why not? I've gotten my work done, I've taken care of my responsibilities, I've earned some time to relax. [NEWLINE] [NEWLINE] Now as someone who also partakes in the same capacity you do, I also know how easy it can be for this to become more than just something you look forward to. Anytime I'm high, it's impeding me from being proactive. More and more as I smoke weed nightly, I come to expect it. My day is based on it. Whatever I do is all done for the purpose of getting to reward myself with weed at the end of the day. [NEWLINE] [NEWLINE] In short, smoking weed at the end of every day has different implications than rewarding yourself with, say, a bowl of ice cream (which might, in fact, accompany the former). I can't remember what comedian or show I heard it from, but there's a quote out there that goes, "Smoking pot makes you okay with being bored." There's no real argument *against* this habit besides just knowing that you could always be doing other things when you're smoking weed.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm with you that good and bad are experiences that are had, I believe that this is a reasonable starting point (axioms) for a completely logical ethics.<mask> on the face of it, increasing the amount of good in the world by plugging people in to good experience machines would seem like a sensible idea. [NEWLINE] [NEWLINE] <mask>, this neglects much greater goods that may exist in the future. We know that matter experiences something<mask> it is arranged in the correct way, this is self-evident (we're made of such matter). It's<mask> evident that the number of ways that matter can be arranged increases exponentially<mask> more matter is added. Finally, we know that more complex brains can have more complex experiences. [NEWLINE] [NEWLINE] <mask><mask><mask><mask> it's reasonable to assume that sometime in the future we'll figure out<mask> to build more complex minds than ours out of the dumb matter that makes up most of the universe, and that these will have far better experiences than not only any human,<mask> the entire human race combined. [NEWLINE] [NEWLINE] Anything that stands in the way of this is immoral<mask> it denies the future greater good, it's selfish to only consider mankind<mask> there's a whole universe to awaken.</s>
Label encoding: <s>I'm with you that good and bad are experiences that are had, I believe that this is a reasonable starting point (axioms) for a completely logical ethics. So on the face of it, increasing the amount of good in the world by plugging people in to good experience machines would seem like a sensible idea. [NEWLINE] [NEWLINE] However, this neglects much greater goods that may exist in the future. We know that matter experiences something if it is arranged in the correct way, this is self-evident (we're made of such matter). It's also evident that the number of ways that matter can be arranged increases exponentially as more matter is added. Finally, we know that more complex brains can have more complex experiences. [NEWLINE] [NEWLINE] So in my opinion it's reasonable to assume that sometime in the future we'll figure out how to build more complex minds than ours out of the dumb matter that makes up most of the universe, and that these will have far better experiences than not only any human, but the entire human race combined. [NEWLINE] [NEWLINE] Anything that stands in the way of this is immoral because it denies the future greater good, it's selfish to only consider mankind when there's a whole universe to awaken.</s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Did the Greeks go into the battle thinking they could win? [ENDQ] [NEWLINE] Well,<mask> I wrote that comment originally I was just thinking of the movie,<mask> afterward I went back to look at the battle. The Greek troops at Thermopylae came from different city-states, each with their own ideologies.<mask> was Leonidas the de-facto leader?<mask> Sparta's ideology was to value war and war preparation over all else. This differed from the ideologies of other Greek city-states, which included literacy and democracy. [NEWLINE] [NEWLINE] They did go into that battle thinking they could win,<mask><mask> the Persians found the path to outflank the Greeks, it became clear that the battle was lost. The Spartans and soldiers from a few other city-states stayed behind to battle to the death.<mask> one group and not all of them?<mask> not everybody run? Military strategy, the fighting of a rear-guard action to protect the main battle force, could be a legitimate reason for the force remaining,<mask><mask> that particular group? [NEWLINE] [NEWLINE] Even<mask> you can address some of the higher-level questions in terms of self interest, you have to understand values to explain the details of<mask> happens.</s>
Label encoding: <s> [STARTQ] Did the Greeks go into the battle thinking they could win? [ENDQ] [NEWLINE] Well, when I wrote that comment originally I was just thinking of the movie, but afterward I went back to look at the battle. The Greek troops at Thermopylae came from different city-states, each with their own ideologies. Why was Leonidas the de-facto leader? Because Sparta's ideology was to value war and war preparation over all else. This differed from the ideologies of other Greek city-states, which included literacy and democracy. [NEWLINE] [NEWLINE] They did go into that battle thinking they could win, but when the Persians found the path to outflank the Greeks, it became clear that the battle was lost. The Spartans and soldiers from a few other city-states stayed behind to battle to the death. Why one group and not all of them? Why not everybody run? Military strategy, the fighting of a rear-guard action to protect the main battle force, could be a legitimate reason for the force remaining, but why that particular group? [NEWLINE] [NEWLINE] Even if you can address some of the higher-level questions in terms of self interest, you have to understand values to explain the details of what happens.</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was at the Chicago protest.  It was about the Tarp bailouts and<mask> tax payers shouldn't be paying for corporate mistakes on Wall Street.  [Maybe you didn't realize that the US government gave over 400 billion to private companies<mask> they wouldn't go bankrupt....]( [URL] ).  It later had a lot of smaller movements that added to it.  The mainstream press took advantage of this and kept saying no one knew<mask> it was about.  All you had to do was open up a web browser and open the web pages dedicated to each major protest location and it had all the information you needed about the protests.  This included protest times, meeting notes, and financial information<mask> the protests were pretty well organized and [some of them made enough money to do more than stand in the streets outside of banking institutions]( [URL] /). [NEWLINE] [NEWLINE] After a long drawn out protest and a cold winter the protest broke up and a lot of people moved on.  The occupy groups that wanted to keep protesting moved into more specific areas to continue working on issues of social inequality and more.  [Occupy Sandy]( [URL] /) is a good example of<mask> one Occupy group did after the main protests were over.</s><pad><pad>
Label encoding: <s>I was at the Chicago protest.  It was about the Tarp bailouts and how tax payers shouldn't be paying for corporate mistakes on Wall Street.  [Maybe you didn't realize that the US government gave over 400 billion to private companies so they wouldn't go bankrupt....]( [URL] ).  It later had a lot of smaller movements that added to it.  The mainstream press took advantage of this and kept saying no one knew what it was about.  All you had to do was open up a web browser and open the web pages dedicated to each major protest location and it had all the information you needed about the protests.  This included protest times, meeting notes, and financial information as the protests were pretty well organized and [some of them made enough money to do more than stand in the streets outside of banking institutions]( [URL] /). [NEWLINE] [NEWLINE] After a long drawn out protest and a cold winter the protest broke up and a lot of people moved on.  The occupy groups that wanted to keep protesting moved into more specific areas to continue working on issues of social inequality and more.  [Occupy Sandy]( [URL] /) is a good example of what one Occupy group did after the main protests were over.</s><pad><pad>
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be clear, I'm not an apologist for rude people.  I<mask> don't think it's the responsibility of every or any black person to explain white privilege to you. <mask> people keep yelling at you for something you believe, maybe take a look at some of their reasoned arguments rather than their shrill tone? [NEWLINE] [NEWLINE] All boats rise together, brother in arms, I'm right there with you on socialism,<mask> arguing which movement is more important is silly - they're all important, and we all have our part to play in making the world a better place.  I've felt that sentiment (Your opinions are not valued) and it sucks, and maybe those are just people you don't jibe with.  Don't confuse people with the ideology.  People are fucking nasty and mean, and, you know, sometimes they have good reasons for being nasty and mean.  Malcolm X or some Black Panthers might not like me personally for being white,<mask> I understand and respect their opinions and see<mask> they're coming from.  I'm honestly not that up to date on black intellectuals, or<mask> that's a fair characterization of Malcolm,<mask> pardon my ignorance,<mask> still, ya know?</s>
Label encoding: <s>To be clear, I'm not an apologist for rude people.  I also don't think it's the responsibility of every or any black person to explain white privilege to you.  If people keep yelling at you for something you believe, maybe take a look at some of their reasoned arguments rather than their shrill tone? [NEWLINE] [NEWLINE] All boats rise together, brother in arms, I'm right there with you on socialism, but arguing which movement is more important is silly - they're all important, and we all have our part to play in making the world a better place.  I've felt that sentiment (Your opinions are not valued) and it sucks, and maybe those are just people you don't jibe with.  Don't confuse people with the ideology.  People are fucking nasty and mean, and, you know, sometimes they have good reasons for being nasty and mean.  Malcolm X or some Black Panthers might not like me personally for being white, but I understand and respect their opinions and see where they're coming from.  I'm honestly not that up to date on black intellectuals, or if that's a fair characterization of Malcolm, so pardon my ignorance, but still, ya know?</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fun Fact: Hitler made it *mandatory* for all citizens to have firearms carry permits in a 1928 law, which<mask> required a second, different permit for the acquisition of a gun. These permits were good for 1 year. Then, in the 1938 German Weapons Act: [NEWLINE] [NEWLINE] * These restrictions only applied to handguns. The transfer and sale of shotguns, long guns and ammo was completely deregulated. [NEWLINE] [NEWLINE] * The legal age for possession was lowered from 20 to 18. [NEWLINE] [NEWLINE] * Permits became valid for 3 years, increased from just 1 year. [NEWLINE] [NEWLINE] * Prior to the 1938 law, govt workers were exempt from the need for an acquisition permit. This exemption was expanded to include hunting permit holders and NSDAP members. This means that they could buys, sell and transfer *any* firearm without restriction. [NEWLINE] [NEWLINE] Take away guns, throw them at people, turn them into art... it doesn't matter. Dictators could care less about the Beretta in someone's nightstand. [NEWLINE] [NEWLINE] [Gun legislation in Germany]( [URL] #Hitler.27s_partial_relaxation_of_gun_control_on_government_workers_in_Nazi_Germany)</s>
Label encoding: <s>Fun Fact: Hitler made it *mandatory* for all citizens to have firearms carry permits in a 1928 law, which also required a second, different permit for the acquisition of a gun. These permits were good for 1 year. Then, in the 1938 German Weapons Act: [NEWLINE] [NEWLINE] * These restrictions only applied to handguns. The transfer and sale of shotguns, long guns and ammo was completely deregulated. [NEWLINE] [NEWLINE] * The legal age for possession was lowered from 20 to 18. [NEWLINE] [NEWLINE] * Permits became valid for 3 years, increased from just 1 year. [NEWLINE] [NEWLINE] * Prior to the 1938 law, govt workers were exempt from the need for an acquisition permit. This exemption was expanded to include hunting permit holders and NSDAP members. This means that they could buys, sell and transfer *any* firearm without restriction. [NEWLINE] [NEWLINE] Take away guns, throw them at people, turn them into art... it doesn't matter. Dictators could care less about the Beretta in someone's nightstand. [NEWLINE] [NEWLINE] [Gun legislation in Germany]( [URL] #Hitler.27s_partial_relaxation_of_gun_control_on_government_workers_in_Nazi_Germany)</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Tell me<mask> registers you need to use in<mask> order to save a file to your system, and<mask> pointers need to reference<mask> memory locations. [ENDQ] [NEWLINE] That would take a long<mask> to explain,<mask> it happens to be something I could tell you about. I would start with the Win32 API and calls like CreateFile, ReadFile, WriteFile. We can cover that in any depth you like. [NEWLINE] [NEWLINE] A better analogy, in my case, would be car knowledge. I own a car, and I drive it. I know the basics of<mask> an internal combustion engine works, I know the basics of electrical circuits, I know about some of the components of the car and their functions - I know about the battery, the transmission, the alternator... [NEWLINE] [NEWLINE] That being said,<mask> my car were to break down, my knowledge of its workings is rudimentary. I cannot fix anything<mask> the most obvious problems - such<mask> a flat tire, or empty battery. This makes me self-conscious about<mask> little I know about cars, and I consider myself under-educated about the topic.<mask><mask> people who know comparably little about something else should be aware of that and feel appropriately, too.</s>
Label encoding: <s> [STARTQ] Tell me what registers you need to use in what order to save a file to your system, and what pointers need to reference what memory locations. [ENDQ] [NEWLINE] That would take a long while to explain, but it happens to be something I could tell you about. I would start with the Win32 API and calls like CreateFile, ReadFile, WriteFile. We can cover that in any depth you like. [NEWLINE] [NEWLINE] A better analogy, in my case, would be car knowledge. I own a car, and I drive it. I know the basics of how an internal combustion engine works, I know the basics of electrical circuits, I know about some of the components of the car and their functions - I know about the battery, the transmission, the alternator... [NEWLINE] [NEWLINE] That being said, if my car were to break down, my knowledge of its workings is rudimentary. I cannot fix anything but the most obvious problems - such as a flat tire, or empty battery. This makes me self-conscious about what little I know about cars, and I consider myself under-educated about the topic. I think people who know comparably little about something else should be aware of that and feel appropriately, too.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not really the point I'm trying to make. [NEWLINE] [NEWLINE] Extremes temper each other out. Social consensus encompasses those group of ideas which *have not* been discarded<mask> bad. Social consensus generally tends to favor a middle of the road approach between the two extremes in the social discourse spectrum. [NEWLINE] [NEWLINE] Radicals on one side provide a stark contrast to radicals on the other. Which allows social discourse to discard either extreme and arrive at a middle of the road consensus<mask> to<mask> constitutes reasonable and unreasonable. [NEWLINE] [NEWLINE] We only appreciate the ideas we do<mask> "good" ideas<mask> we've had "bad" ideas to contrast them to. Everyone is guilty of this. You don't even realize<mask> much your concept of "good vs. bad" is dependent on the conflict between the two sides. [NEWLINE] [NEWLINE] You aren't partial to some magical moral certainty that the rest of society lacks. These wonderful progressive ideals you take for granted are something you only hold<mask> you have the last several thousand years of humans arguing back and forth and re-tuning morality to fall back on.<mask> you existed in a more horrible time you would think more horrible things. Everyone (everyone) is a product of their environment.</s>
Label encoding: <s>That's not really the point I'm trying to make. [NEWLINE] [NEWLINE] Extremes temper each other out. Social consensus encompasses those group of ideas which *have not* been discarded as bad. Social consensus generally tends to favor a middle of the road approach between the two extremes in the social discourse spectrum. [NEWLINE] [NEWLINE] Radicals on one side provide a stark contrast to radicals on the other. Which allows social discourse to discard either extreme and arrive at a middle of the road consensus as to what constitutes reasonable and unreasonable. [NEWLINE] [NEWLINE] We only appreciate the ideas we do as "good" ideas because we've had "bad" ideas to contrast them to. Everyone is guilty of this. You don't even realize how much your concept of "good vs. bad" is dependent on the conflict between the two sides. [NEWLINE] [NEWLINE] You aren't partial to some magical moral certainty that the rest of society lacks. These wonderful progressive ideals you take for granted are something you only hold because you have the last several thousand years of humans arguing back and forth and re-tuning morality to fall back on. If you existed in a more horrible time you would think more horrible things. Everyone (everyone) is a product of their environment.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It is a result of this last argument that some members of the MRM treat all feminists with a similar combative stance,<mask><mask> the diverse individual beliefs held by all feminists,<mask> all feminists are believed to share some responsibility for existing measures that were lobbied for under the guise of generic feminism.<mask> recently stated by reddit /r/Feminism moderator, demmian, [ENDQ] [NEWLINE] Heh.<mask> I knew my defense of feminism would be caught in an article, I would have at least better capitalized my words. [NEWLINE] [NEWLINE] <mask> seriously,<mask> do<mask> many MRAs profess a liking for rationality,<mask> have such a big blind spot<mask> it comes to the irrational group blaming against feminists? Should this defense be expressed in terms of [logic]( [URL] ), instead of a legal term, such<mask> collective guilt? [NEWLINE] [NEWLINE] gww and typhonblue &amp; co love to avoid this issue by claiming that all feminists share on the power of feminist organizations.<mask> this very kind of relation is easily dismissed by the same MRAs<mask> they eagerly point out that not all men have political power,<mask> power positions being held predominantly by men (an even greater evidence of political/economic power).</s>
Label encoding: <s> [STARTQ] It is a result of this last argument that some members of the MRM treat all feminists with a similar combative stance, regardless of the diverse individual beliefs held by all feminists, as all feminists are believed to share some responsibility for existing measures that were lobbied for under the guise of generic feminism. As recently stated by reddit /r/Feminism moderator, demmian, [ENDQ] [NEWLINE] Heh. If I knew my defense of feminism would be caught in an article, I would have at least better capitalized my words. [NEWLINE] [NEWLINE] But seriously, how do so many MRAs profess a liking for rationality, yet have such a big blind spot when it comes to the irrational group blaming against feminists? Should this defense be expressed in terms of [logic]( [URL] ), instead of a legal term, such as collective guilt? [NEWLINE] [NEWLINE] gww and typhonblue &amp; co love to avoid this issue by claiming that all feminists share on the power of feminist organizations. Yet this very kind of relation is easily dismissed by the same MRAs when they eagerly point out that not all men have political power, despite power positions being held predominantly by men (an even greater evidence of political/economic power).</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would rather find an alternative to abortion.<mask>,<mask><mask> that's<mask> i say i would do, i am not a woman, and i have never been in the position of helping someone decide such a thing,<mask> in reality i don't know. [NEWLINE] [NEWLINE] <mask> i do know<mask>, is that illegal abortions happen all the time. With not a lot of cleanliness or security. I<mask> thing that a mistake shouldn't change a person's life completely. It is not my right to judge someone on something stupid that they did, we all fuck up, some big time, others not<mask> much.<mask>, my stance is the same<mask> with prostitution, legalise it. I'd probably never use either,<mask> that's not to say that i have the right to tell a consenting adult<mask> to live their life.<mask>, most pregnancies end naturally, and people don't even realise it. And legalisation would bring with it much needed healthcare and psychological treatment for those involved. [NEWLINE] [NEWLINE] I'm someone who values options, let the option stand,<mask><mask><mask> the people involved are well informed about the consequences, and not strongarmed/scared into doing something they don't want to.</s>
Label encoding: <s>I would rather find an alternative to abortion. However, even though that's what i say i would do, i am not a woman, and i have never been in the position of helping someone decide such a thing, so in reality i don't know. [NEWLINE] [NEWLINE] What i do know however, is that illegal abortions happen all the time. With not a lot of cleanliness or security. I also thing that a mistake shouldn't change a person's life completely. It is not my right to judge someone on something stupid that they did, we all fuck up, some big time, others not so much. So, my stance is the same as with prostitution, legalise it. I'd probably never use either, but that's not to say that i have the right to tell a consenting adult how to live their life. Besides, most pregnancies end naturally, and people don't even realise it. And legalisation would bring with it much needed healthcare and psychological treatment for those involved. [NEWLINE] [NEWLINE] I'm someone who values options, let the option stand, as long as the people involved are well informed about the consequences, and not strongarmed/scared into doing something they don't want to.</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] [STARTQ] Most films designed with 3D in mind are filled with cliche pop-out moments, such<mask> a character pointing out of the screen towards the audience, or a baseball being hit out of the screen. [ENDQ] That's true.<mask> then, this is the case for most films designed with [cinematic technology X] in mind. Films making early use of sound were desperate to include songs, jokes, and other things that people would actually want to hear aloud; films making early use of surround sound were heavy on bass and multi-channel soundscapes; films making use of technicolour made damned sure that they had actually colourful things on screen; early films comprised solely of CGI (i.e. Toy Story) stuck to subject matter that would be obstructively hard to film in live-action; and films that will make use of the 48 FPS standard in the future will be more bold in their digital effects given the greater scope the format provides for seamless integration. The desirability of the latter thing is a whole other debate,<mask>, and I don't want to get into it here -- I'm just offering it up<mask> a further example. [NEWLINE] [NEWLINE] Thanks.</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] [STARTQ] Most films designed with 3D in mind are filled with cliche pop-out moments, such as a character pointing out of the screen towards the audience, or a baseball being hit out of the screen. [ENDQ] That's true. But then, this is the case for most films designed with [cinematic technology X] in mind. Films making early use of sound were desperate to include songs, jokes, and other things that people would actually want to hear aloud; films making early use of surround sound were heavy on bass and multi-channel soundscapes; films making use of technicolour made damned sure that they had actually colourful things on screen; early films comprised solely of CGI (i.e. Toy Story) stuck to subject matter that would be obstructively hard to film in live-action; and films that will make use of the 48 FPS standard in the future will be more bold in their digital effects given the greater scope the format provides for seamless integration. The desirability of the latter thing is a whole other debate, though, and I don't want to get into it here -- I'm just offering it up as a further example. [NEWLINE] [NEWLINE] Thanks.</s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. facts provide security more than feelings do for many men. [NEWLINE] [NEWLINE] 2. i dont think its unfair. its not fair that a woman knows 100% that shes the mother of her child<mask> a man doesnt. even<mask> your<mask> was cheating and fathers a child with another woman, *you are not on the hook for those children.* [NEWLINE] [NEWLINE] 3. you would be surprised<mask> many cheaters go bareback.<mask>, getting pregnant<mask> cheating is not a'mistake' forgetting to water your flowers<mask> you got busy doing the housework is a mistake. having unprotected sex with someone else and subsequently getting impregnated<mask> in an exclusive relationship is a series of conscious choices. calling those choices a mistake is minimalistic. given those choices it would be a lot harder to 'work through it' than you imagine. [NEWLINE] [NEWLINE] 4. Verbal agreements are made all the time and broken all the time and are for the most part non enforceable. that is<mask> contracts exist. this particular contract (marriage) has a hefty burden of a father<mask> it turns out a kid is born into his wedlock that is not his. that is his reason and is not unprompted.</s>
Label encoding: <s>1. facts provide security more than feelings do for many men. [NEWLINE] [NEWLINE] 2. i dont think its unfair. its not fair that a woman knows 100% that shes the mother of her child when a man doesnt. even if your SO was cheating and fathers a child with another woman, *you are not on the hook for those children.* [NEWLINE] [NEWLINE] 3. you would be surprised how many cheaters go bareback. Also, getting pregnant while cheating is not a'mistake' forgetting to water your flowers because you got busy doing the housework is a mistake. having unprotected sex with someone else and subsequently getting impregnated while in an exclusive relationship is a series of conscious choices. calling those choices a mistake is minimalistic. given those choices it would be a lot harder to 'work through it' than you imagine. [NEWLINE] [NEWLINE] 4. Verbal agreements are made all the time and broken all the time and are for the most part non enforceable. that is why contracts exist. this particular contract (marriage) has a hefty burden of a father if it turns out a kid is born into his wedlock that is not his. that is his reason and is not unprompted.</s>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1) Training time is much longer. Bigger investment up front should mean bigger rewards. [ENDQ] [NEWLINE] <mask> this isn't really the way things happen. Someone can get an MBA with much less work than medical school. They can go on to be leaders of corporations and make a lot more than a doctor. [NEWLINE] [NEWLINE] [STARTQ] 2) Much greater precision of craft. Let's be honest. We've all survived shitty teachers. Surviving a shitty doctor is actually uncertain. [ENDQ] [NEWLINE] Not really. I had a doctor misdiagnose an infection in my throat. Ended up needing to go to the ER later,<mask> I survived. A bad teacher can create some pretty awful circumstances that can ruin a person's potential. [NEWLINE] [NEWLINE] [STARTQ] 3) Level of responsibility. [ENDQ] [NEWLINE] Taking care of and guiding the minds of future generations is pretty important. [NEWLINE] [NEWLINE] [STARTQ] 4) Scarcity of aptitude. Say<mask> you will about sports salaries,<mask> does the star athlete make<mask> much<mask> it's the most important job? [ENDQ] [NEWLINE] <mask> we paid the best athletes $45k/year, do you think we'd still have<mask> great of athletes playing sports? Increasing salaries will encourage better and better teacher candidates. [NEWLINE] </s>
Label encoding: <s> [STARTQ] 1) Training time is much longer. Bigger investment up front should mean bigger rewards. [ENDQ] [NEWLINE] But this isn't really the way things happen. Someone can get an MBA with much less work than medical school. They can go on to be leaders of corporations and make a lot more than a doctor. [NEWLINE] [NEWLINE] [STARTQ] 2) Much greater precision of craft. Let's be honest. We've all survived shitty teachers. Surviving a shitty doctor is actually uncertain. [ENDQ] [NEWLINE] Not really. I had a doctor misdiagnose an infection in my throat. Ended up needing to go to the ER later, but I survived. A bad teacher can create some pretty awful circumstances that can ruin a person's potential. [NEWLINE] [NEWLINE] [STARTQ] 3) Level of responsibility. [ENDQ] [NEWLINE] Taking care of and guiding the minds of future generations is pretty important. [NEWLINE] [NEWLINE] [STARTQ] 4) Scarcity of aptitude. Say what you will about sports salaries, but does the star athlete make so much because it's the most important job? [ENDQ] [NEWLINE] If we paid the best athletes $45k/year, do you think we'd still have as great of athletes playing sports? Increasing salaries will encourage better and better teacher candidates. [NEWLINE] </s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Race is a human construct. This has been known for over 100 years. Your understanding is a Victorian age relic. [NEWLINE] [NEWLINE] There is more genetic diversity between "races" than within them. That means "race" is not a biologically meaningful term. It does not demarcate groups of people in any reliable or useful way, other than modern ignorant social attitudes. Biologists may look at geography or ethnicity, or study other possible meaningful variations,<mask> "race" is not a technical term in biology. The concept didn't even exist in a presently recognizable way until very recently in human history, the 19th century. Try telling a Frenchmen and an Englishman that they were the same "race" a few hundred years ago and they would've taken turns kicking your ass. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [NEWLINE] By the way, about "race" [NEWLINE] [STARTQ] First used to refer to speakers of a common language and then to denote national affiliations, in the 17th century, people began to use the term to relate to observable physical traits. [ENDQ] [URL] (human_classification)#Social_constructions [NEWLINE] [NEWLINE] <mask> that whole business of accents not implying race kinda goes out the window.</s>
Label encoding: <s>Race is a human construct. This has been known for over 100 years. Your understanding is a Victorian age relic. [NEWLINE] [NEWLINE] There is more genetic diversity between "races" than within them. That means "race" is not a biologically meaningful term. It does not demarcate groups of people in any reliable or useful way, other than modern ignorant social attitudes. Biologists may look at geography or ethnicity, or study other possible meaningful variations, but "race" is not a technical term in biology. The concept didn't even exist in a presently recognizable way until very recently in human history, the 19th century. Try telling a Frenchmen and an Englishman that they were the same "race" a few hundred years ago and they would've taken turns kicking your ass. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [NEWLINE] By the way, about "race" [NEWLINE] [STARTQ] First used to refer to speakers of a common language and then to denote national affiliations, in the 17th century, people began to use the term to relate to observable physical traits. [ENDQ] [URL] (human_classification)#Social_constructions [NEWLINE] [NEWLINE] So that whole business of accents not implying race kinda goes out the window.</s>
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>can you give me an example of any card game that treated those who purchased no-cards vs. those who purchased a lot of cards equally? [NEWLINE] [NEWLINE] I have never recieved a new pack of YGO-cards<mask> I won against an opponent 10 times, I had to go to the store, buy a new pack and hope for the best. [NEWLINE] [NEWLINE] <mask> I didn't expand my library than that meant I was stuck with all the strategies I currently had. [NEWLINE] [NEWLINE] Of course payers will have an edge and there is nothing wrong with that. That is the nature of CCG/TCG's. [NEWLINE] [NEWLINE] <mask> you want some solace<mask>, it all comes back to the time vs. value conversation. [NEWLINE] [NEWLINE] <mask> your time is more worth to you than your money, you are more likely to spend some bucks for progress than the other way around. [NEWLINE] [NEWLINE] and<mask> you want another solace, the game is already free to play and of superb-quality,<mask> you dish out 50$, it's<mask><mask> you bought a new game at retail price for which you can get a decent value out of,<mask> you calculate $-spent/h-played.</s>
Label encoding: <s>can you give me an example of any card game that treated those who purchased no-cards vs. those who purchased a lot of cards equally? [NEWLINE] [NEWLINE] I have never recieved a new pack of YGO-cards because I won against an opponent 10 times, I had to go to the store, buy a new pack and hope for the best. [NEWLINE] [NEWLINE] If I didn't expand my library than that meant I was stuck with all the strategies I currently had. [NEWLINE] [NEWLINE] Of course payers will have an edge and there is nothing wrong with that. That is the nature of CCG/TCG's. [NEWLINE] [NEWLINE] If you want some solace though, it all comes back to the time vs. value conversation. [NEWLINE] [NEWLINE] If your time is more worth to you than your money, you are more likely to spend some bucks for progress than the other way around. [NEWLINE] [NEWLINE] and if you want another solace, the game is already free to play and of superb-quality, if you dish out 50$, it's as if you bought a new game at retail price for which you can get a decent value out of, if you calculate $-spent/h-played.</s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Asexual is merely an extreme. There are many more people with a low sex drive who are not considered abnormal. With the nurturing instinct, there are<mask> many social pressures that make people who don't really want kids have them anyway. These pressures are starting to diminish in first world countries,<mask> we are starting to see a lot more people with a low level of nurturing instinct choosing to be child free. This is actually quite common now the social pressures are being stripped away and certainly less than'most people'. This is a process which has not<mask> ended and the future will bring more acceptance of the child free lifestyle. [NEWLINE] [NEWLINE] The reward feelings that the nurturing instinct provides can only go<mask> far in some cases to mitigate against the problems that children cause. The sort of things that the child free list<mask> serious problems<mask> parents say they 'just deal with'. Part of the function of the nurturing instinct is to lessen the downsides of having kids to avoid parents dashing their children against rocks, which would be bad for the species.<mask> someone has a low level of nurturing instinct and you add other problems you mention, such<mask> an unstable relationship, that can tip the person over the edge.</s>
Label encoding: <s>Asexual is merely an extreme. There are many more people with a low sex drive who are not considered abnormal. With the nurturing instinct, there are also many social pressures that make people who don't really want kids have them anyway. These pressures are starting to diminish in first world countries, so we are starting to see a lot more people with a low level of nurturing instinct choosing to be child free. This is actually quite common now the social pressures are being stripped away and certainly less than'most people'. This is a process which has not yet ended and the future will bring more acceptance of the child free lifestyle. [NEWLINE] [NEWLINE] The reward feelings that the nurturing instinct provides can only go so far in some cases to mitigate against the problems that children cause. The sort of things that the child free list as serious problems but parents say they 'just deal with'. Part of the function of the nurturing instinct is to lessen the downsides of having kids to avoid parents dashing their children against rocks, which would be bad for the species. If someone has a low level of nurturing instinct and you add other problems you mention, such as an unstable relationship, that can tip the person over the edge.</s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm interested, were they football players or hockey players? And<mask> old were you<mask> it happened? [ENDQ] [NEWLINE] I was 20, and they were 20 +/- 2.  They weren't sports players, or psychotics, they were just lowlifes who had dropped out of highschool and thought that selling weed was their legit job,<mask> of course they could never make any money at it<mask> they'd always smoke it all. [NEWLINE] [NEWLINE] Yeah, I don't talk to them anymore.  The people I hang out with now are almost all extremely well educated, and to whatever extent they're rape apologists, it's much less in-your-face.  It's still there<mask>,<mask> it might be a lot less obvious to you<mask> you're a dude. [NEWLINE] [NEWLINE]...And that's<mask> I don't think it's really helpful to avoid addressing the problem by simply saying, "Well, those guys are<mask> ridiculous that nobody who isn't 100% awful would put up with that."  Those guys had their good points and their bad points, just like everybody does, and they have friends and other people who they interact with--just like everybody does. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I'm interested, were they football players or hockey players? And how old were you when it happened? [ENDQ] [NEWLINE] I was 20, and they were 20 +/- 2.  They weren't sports players, or psychotics, they were just lowlifes who had dropped out of highschool and thought that selling weed was their legit job, but of course they could never make any money at it because they'd always smoke it all. [NEWLINE] [NEWLINE] Yeah, I don't talk to them anymore.  The people I hang out with now are almost all extremely well educated, and to whatever extent they're rape apologists, it's much less in-your-face.  It's still there though, but it might be a lot less obvious to you if you're a dude. [NEWLINE] [NEWLINE]...And that's why I don't think it's really helpful to avoid addressing the problem by simply saying, "Well, those guys are so ridiculous that nobody who isn't 100% awful would put up with that."  Those guys had their good points and their bad points, just like everybody does, and they have friends and other people who they interact with--just like everybody does. [NEWLINE] </s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We all can be hateful in our words,<mask> we mostly never mean harm in it. Look at Reddit, we are still full of people who will make remakes similar to "op is a fag" among other things [ENDQ] [NEWLINE] [NEWLINE] Yes, reddit, like 4chan, is full of people who are in effect very hateful,<mask> they try their hardest to justify that behavior<mask> not hateful. I don't buy it for a second. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask>, I very superbly doubt--especially in Reddit's liberal demographic--that people who say that actually are homophobic and/or hate marriage equality. [ENDQ] [NEWLINE] [NEWLINE] They are homophobic. They may not be willing to admit it,<mask> they are. Using homophobic slurs is the easiest way to identify yourself<mask> a homophobe. [NEWLINE] [NEWLINE] [NEWLINE] <mask>, supporting marriage equality does not mean a person is not homophobic. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> having a legit Nazism thread on /pol/ makes all of 4chan bad, having subreddits like /r/picsofdeadkids[1] makes all of Reddit bad [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] Allowing all those terrible subs does reflect poorly on all of reddit. </s>
Label encoding: <s> [STARTQ] We all can be hateful in our words, but we mostly never mean harm in it. Look at Reddit, we are still full of people who will make remakes similar to "op is a fag" among other things [ENDQ] [NEWLINE] [NEWLINE] Yes, reddit, like 4chan, is full of people who are in effect very hateful, but they try their hardest to justify that behavior as not hateful. I don't buy it for a second. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] However, I very superbly doubt--especially in Reddit's liberal demographic--that people who say that actually are homophobic and/or hate marriage equality. [ENDQ] [NEWLINE] [NEWLINE] They are homophobic. They may not be willing to admit it, but they are. Using homophobic slurs is the easiest way to identify yourself as a homophobe. [NEWLINE] [NEWLINE] [NEWLINE] Also, supporting marriage equality does not mean a person is not homophobic. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] If having a legit Nazism thread on /pol/ makes all of 4chan bad, having subreddits like /r/picsofdeadkids[1] makes all of Reddit bad [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] Allowing all those terrible subs does reflect poorly on all of reddit. </s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0712, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most of the things you cite are largely continuations of trends begun before Bush took office, especially with the Clinton administration.  Clinton was<mask> tapping everyone's phones (Echelon), bombing people for no reason (Iraq), screwing up international relations (China and Japan), and militarizing the police (Waco). [NEWLINE] [NEWLINE] <mask> anything Bush the Lesser will likely be remembered<mask> the last Reaganesque President.  Reagan, Bush, Clinton, and Bush the Lesser all publicly supported an ideal of government that was nominally smaller, less regulated, with less support for the welfare state, and working with private industry<mask> opposed to earlier more strongly centralized models. [NEWLINE] [NEWLINE] Obama has largely broken with this particular rhetoric.  Instead he's framed himself<mask> the new Progressive much like politicians of the early 20th century. [NEWLINE] [NEWLINE] <mask> Bush's system of government is seen by history<mask> negative, then Reagan will most likely get the blame. <mask> positive, the Reagan will get all the credit and Bush will probably be seen<mask> the worst Reaganesque president. <mask> nobody is likely to see him<mask> the straw that broke the camels back like people see Buchanan.</s>
Label encoding: <s>Most of the things you cite are largely continuations of trends begun before Bush took office, especially with the Clinton administration.  Clinton was also tapping everyone's phones (Echelon), bombing people for no reason (Iraq), screwing up international relations (China and Japan), and militarizing the police (Waco). [NEWLINE] [NEWLINE] If anything Bush the Lesser will likely be remembered as the last Reaganesque President.  Reagan, Bush, Clinton, and Bush the Lesser all publicly supported an ideal of government that was nominally smaller, less regulated, with less support for the welfare state, and working with private industry as opposed to earlier more strongly centralized models. [NEWLINE] [NEWLINE] Obama has largely broken with this particular rhetoric.  Instead he's framed himself as the new Progressive much like politicians of the early 20th century. [NEWLINE] [NEWLINE] If Bush's system of government is seen by history as negative, then Reagan will most likely get the blame.  If positive, the Reagan will get all the credit and Bush will probably be seen as the worst Reaganesque president.  But nobody is likely to see him as the straw that broke the camels back like people see Buchanan.</s>
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't fear<mask> follows death, nor even<mask> initiates the transition,<mask> only the actual "crossing over". [NEWLINE] [NEWLINE] <mask> there is an afterlife, then I can neither prove nor effect it from here; there is no use dreading<mask> I have no control over. <mask> there is not, then I will not exist to care. <mask>, between this state of being and whichever follows is a point of transition the experience of which can not be wholly anticipated.  There is only one more possibility, and its foundation is the perception of that instant. [NEWLINE] [NEWLINE] Suppose that perception is immutable<mask> life and time are not.  My perception of time would continue<mask><mask> nothing changed<mask> my body ceases, and outside of "me" time would continue to flow normally. <mask> would be my final nanosecond to you, to me would be an eternity to shape with anything I may imagine.  Then in that transition I would become a universe. [NEWLINE] [NEWLINE] In that case,<mask> my afterlife is unpleasant then it is my own fault.  Be careful<mask> you contemplate ideas such<mask> Hell. <mask><mask> you would be the one to create it?</s>
Label encoding: <s>I don't fear what follows death, nor even what initiates the transition, but only the actual "crossing over". [NEWLINE] [NEWLINE] If there is an afterlife, then I can neither prove nor effect it from here; there is no use dreading what I have no control over.  If there is not, then I will not exist to care.  However, between this state of being and whichever follows is a point of transition the experience of which can not be wholly anticipated.  There is only one more possibility, and its foundation is the perception of that instant. [NEWLINE] [NEWLINE] Suppose that perception is immutable but life and time are not.  My perception of time would continue as if nothing changed while my body ceases, and outside of "me" time would continue to flow normally.  What would be my final nanosecond to you, to me would be an eternity to shape with anything I may imagine.  Then in that transition I would become a universe. [NEWLINE] [NEWLINE] In that case, if my afterlife is unpleasant then it is my own fault.  Be careful when you contemplate ideas such as Hell.  What if you would be the one to create it?</s>
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some others have made good point that's I won't repeat. The view you've represented here of the mental health system is extraordinarily narrow. [NEWLINE] [NEWLINE] <mask> about marriage counseling, for example?<mask> is that diagnosing or ruining someone's life,<mask> you've alluded to? Clients come in with very clear problems in their relationship and communication patterns and they work to resolve those. After that they move on with their lives in one way or another. No diagnostic codes. No hospitalizations. None of the stuff you've listed. [NEWLINE] [NEWLINE] Even outside of marriage counseling, not every therapist is interested in giving a formal diagnosis to their clients (outside of for their own treatment planning purposes). Not every client needs hospitalization and many clients make very positive changes in their lives. Going to therapy isn't always the result of being broken, it's often used to make improvements on an otherwise okay life. Based on your CMV, it appears you are a victim to the stigma of the mental health system;<mask><mask> it's only for "crazy" people and that it's just awash in medications and hospitalizations. You have neglected to see a huge chunk of the industry. </s>
Label encoding: <s>Some others have made good point that's I won't repeat. The view you've represented here of the mental health system is extraordinarily narrow. [NEWLINE] [NEWLINE] What about marriage counseling, for example? How is that diagnosing or ruining someone's life, as you've alluded to? Clients come in with very clear problems in their relationship and communication patterns and they work to resolve those. After that they move on with their lives in one way or another. No diagnostic codes. No hospitalizations. None of the stuff you've listed. [NEWLINE] [NEWLINE] Even outside of marriage counseling, not every therapist is interested in giving a formal diagnosis to their clients (outside of for their own treatment planning purposes). Not every client needs hospitalization and many clients make very positive changes in their lives. Going to therapy isn't always the result of being broken, it's often used to make improvements on an otherwise okay life. Based on your CMV, it appears you are a victim to the stigma of the mental health system; assuming that it's only for "crazy" people and that it's just awash in medications and hospitalizations. You have neglected to see a huge chunk of the industry. </s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> you reach the age of 50, we will kill you in a humane way and then we will eat your corpse." I would find that to be a good deal. [ENDQ] [NEWLINE] I'm going to take a wild guess and assume that 50 seems pretty far away.<mask> you are healthy at 50, you might think differently. Dying at 50 means many won't see their grandchildren grow into adults. You can still do just about anything at 50; there is<mask> much life left to experience. [NEWLINE] [NEWLINE] [STARTQ] And now we get to lamb meat. These guys are killed at 1 month to 1 year, which is in the best case less then 1/10th of their lives. They don't get to enjoy their lives. [ENDQ] [NEWLINE] There is a big difference between a lamb and a human. The life of a lamb will consist of eating, pooping, and reproducing. They have no goals or aspirations, they get no joy from being with their young, they don't accumulate and use knowledge and experience and pass that along to other lambs.<mask> a lamb were to live 10x longer, it will just be 10 more years of eating and pooping.</s>
Label encoding: <s> [STARTQ] But when you reach the age of 50, we will kill you in a humane way and then we will eat your corpse." I would find that to be a good deal. [ENDQ] [NEWLINE] I'm going to take a wild guess and assume that 50 seems pretty far away. If you are healthy at 50, you might think differently. Dying at 50 means many won't see their grandchildren grow into adults. You can still do just about anything at 50; there is so much life left to experience. [NEWLINE] [NEWLINE] [STARTQ] And now we get to lamb meat. These guys are killed at 1 month to 1 year, which is in the best case less then 1/10th of their lives. They don't get to enjoy their lives. [ENDQ] [NEWLINE] There is a big difference between a lamb and a human. The life of a lamb will consist of eating, pooping, and reproducing. They have no goals or aspirations, they get no joy from being with their young, they don't accumulate and use knowledge and experience and pass that along to other lambs. If a lamb were to live 10x longer, it will just be 10 more years of eating and pooping.</s>
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If legal, the sport would almost take a back seat to the PE arms race you set in motion. You have to assume things will simply get more sophisticated over time. [NEWLINE] [NEWLINE] A professional Golfer from the 1920's would be absolutely crushed by a modern golfer markedly worse at the game, simply owing to the clubs and ball used. [NEWLINE] [NEWLINE] This is like that only far more complicated and sticky. [NEWLINE] [NEWLINE] No moral argument, just a slippery slope argument with a side of practicality. Some sports would become who could roid up the best. [NEWLINE] [NEWLINE] It should be legal (and it is, except Baseball),<mask> I don't think Leagues would have it in their interest.<mask> the Owners would be the ones picking up the tabs for the suite of medical supplies, Doctors, etc to make it happen<mask> sanctioned. That's gonna get expensive. Plus Liability. Plus PR nightmare of Helen Lovejoys. Not to mention State Athletic Commissions safety concerns. [NEWLINE] [NEWLINE] All in all...I wouldn't mind a league of steroid hulks doing whatever,<mask> I can think of a raft of good reasons nobody would sanction it or back it financially.</s>
Label encoding: <s>If legal, the sport would almost take a back seat to the PE arms race you set in motion. You have to assume things will simply get more sophisticated over time. [NEWLINE] [NEWLINE] A professional Golfer from the 1920's would be absolutely crushed by a modern golfer markedly worse at the game, simply owing to the clubs and ball used. [NEWLINE] [NEWLINE] This is like that only far more complicated and sticky. [NEWLINE] [NEWLINE] No moral argument, just a slippery slope argument with a side of practicality. Some sports would become who could roid up the best. [NEWLINE] [NEWLINE] It should be legal (and it is, except Baseball), but I don't think Leagues would have it in their interest. Besides the Owners would be the ones picking up the tabs for the suite of medical supplies, Doctors, etc to make it happen if sanctioned. That's gonna get expensive. Plus Liability. Plus PR nightmare of Helen Lovejoys. Not to mention State Athletic Commissions safety concerns. [NEWLINE] [NEWLINE] All in all...I wouldn't mind a league of steroid hulks doing whatever, but I can think of a raft of good reasons nobody would sanction it or back it financially.</s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point about God is that you cannot prove His existence one way or another. It requires faith. [NEWLINE] [NEWLINE] <mask> you want to believe then allow yourself to do<mask>. It is<mask> logical a position<mask> Atheism which in its absolute rejection of the divine requires a spiritual leap akin to Faith. [NEWLINE] [NEWLINE] [STARTQ] <mask> do I know I have the right God? [ENDQ] [NEWLINE] You cannot. You can only do<mask> seems right to you. [NEWLINE] [NEWLINE] [STARTQ] <mask> does the physical world reconcile with scripture (genesis,<mask> read literal, appears to deny evolution)? [ENDQ] [NEWLINE] Evolution is a theory. One I happen to believe is accurate,<mask> you cannot take science<mask> absolute truth. Almost all science is open to change<mask> new evidence emerges.<mask>, scripture can be read<mask> an allegory rather than a literal truth. Like art, the interpretation is for you to make not for another to enforce. [NEWLINE] [NEWLINE] [STARTQ] <mask> can someone who created the universe care about me individually? [ENDQ] [NEWLINE] Omnipotence? Or perhaps you are the only person in the Universe and the rest of us are merely playing parts around you? Again, this is up to you. [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s>The point about God is that you cannot prove His existence one way or another. It requires faith. [NEWLINE] [NEWLINE] If you want to believe then allow yourself to do so. It is as logical a position as Atheism which in its absolute rejection of the divine requires a spiritual leap akin to Faith. [NEWLINE] [NEWLINE] [STARTQ] How do I know I have the right God? [ENDQ] [NEWLINE] You cannot. You can only do what seems right to you. [NEWLINE] [NEWLINE] [STARTQ] How does the physical world reconcile with scripture (genesis, when read literal, appears to deny evolution)? [ENDQ] [NEWLINE] Evolution is a theory. One I happen to believe is accurate, but you cannot take science as absolute truth. Almost all science is open to change if new evidence emerges. Besides, scripture can be read as an allegory rather than a literal truth. Like art, the interpretation is for you to make not for another to enforce. [NEWLINE] [NEWLINE] [STARTQ] How can someone who created the universe care about me individually? [ENDQ] [NEWLINE] Omnipotence? Or perhaps you are the only person in the Universe and the rest of us are merely playing parts around you? Again, this is up to you. [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I just want to refute most of your reasons for button superiority. [NEWLINE] [NEWLINE] Easily replaceable yes.<mask><mask> easily lost. In my life of zippered coats I've never had a problem, within a year of my buttoned peacoat two of three button fell off and needed resewed. [NEWLINE] [NEWLINE] Catching genitalia.<mask> true<mask> very rare. I've never had an issue with this in zippers. Not a substantial concern. [NEWLINE] [NEWLINE] Easy use. Not easier than most zippers. I button every day for my work shirt. It is much slower and more difficult to do than any zippered clothing. [NEWLINE] [NEWLINE] Never seems rusted zipper. Zippers are usually pretty durable too. [NEWLINE] [NEWLINE] Silent. Pointless. I don't need my jeans to be silent for everyday use. I just don't. [NEWLINE] [NEWLINE] Come undone? Maybe you've just experienced very cheap zippers. I have not had hardly any of these zipper problems. [NEWLINE] [NEWLINE] I'm not gonna cannibalize my clothing for a poker game. Chips for playing any game are dirt cheap<mask> ruining clothes is a waste of time and resources. </s>
Label encoding: <s>I just want to refute most of your reasons for button superiority. [NEWLINE] [NEWLINE] Easily replaceable yes. But also easily lost. In my life of zippered coats I've never had a problem, within a year of my buttoned peacoat two of three button fell off and needed resewed. [NEWLINE] [NEWLINE] Catching genitalia. Also true but very rare. I've never had an issue with this in zippers. Not a substantial concern. [NEWLINE] [NEWLINE] Easy use. Not easier than most zippers. I button every day for my work shirt. It is much slower and more difficult to do than any zippered clothing. [NEWLINE] [NEWLINE] Never seems rusted zipper. Zippers are usually pretty durable too. [NEWLINE] [NEWLINE] Silent. Pointless. I don't need my jeans to be silent for everyday use. I just don't. [NEWLINE] [NEWLINE] Come undone? Maybe you've just experienced very cheap zippers. I have not had hardly any of these zipper problems. [NEWLINE] [NEWLINE] I'm not gonna cannibalize my clothing for a poker game. Chips for playing any game are dirt cheap so ruining clothes is a waste of time and resources. </s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I fully accept that there is no objective higher authority that can "correctly" prescribe meaning, meaning is purely a function of usage.<mask> my argument is fully consistent with meaning<mask> a function of usage - I just think that without imposing common rules of the road for<mask> usage conveys meaning, it will slowly become more difficult to communicate with one another<mask> the various links between usage and meaning become more diverse. [NEWLINE] [NEWLINE] <mask><mask> that traffic rules are a good analogy. There is no "right" answer about whether the signal to "go" should be red or green, or whether the "go" light should be at the top or the bottom, or whether it is better to drive on the left side of the road or the right side of the road, or whether it is better for stop signs to be squares or octagons....<mask><mask> *is* important is choosing one and sticking with it,<mask> society benefits from having a clear and universal set of rules about<mask> to effectively "communicate" with one another<mask> we are driving. I don't want to have each driver on the road thinking that he can invent or interpret his own "language".</s>
Label encoding: <s>I fully accept that there is no objective higher authority that can "correctly" prescribe meaning, meaning is purely a function of usage. But my argument is fully consistent with meaning as a function of usage - I just think that without imposing common rules of the road for how usage conveys meaning, it will slowly become more difficult to communicate with one another as the various links between usage and meaning become more diverse. [NEWLINE] [NEWLINE] I think that traffic rules are a good analogy. There is no "right" answer about whether the signal to "go" should be red or green, or whether the "go" light should be at the top or the bottom, or whether it is better to drive on the left side of the road or the right side of the road, or whether it is better for stop signs to be squares or octagons.... But what *is* important is choosing one and sticking with it, because society benefits from having a clear and universal set of rules about how to effectively "communicate" with one another when we are driving. I don't want to have each driver on the road thinking that he can invent or interpret his own "language".</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's stay on topic here; pregnancy and abortion aren't the same<mask> the civil war.<mask> abortion can really be a woman claiming ownership of her body. It's understood that the child is depending on it, and that it will die without her support. [NEWLINE] [NEWLINE] The question here is<mask> pregnancy has to be a permanent state, once it happens? Obviously, there are plenty of ways that pregnancy can happen against the mother's wishes. There are plenty of pregnancies that will result in the death of the mother, sometimes both the mother and the child. In those cases, it seems shallow to shout that the mother had it coming and refuse to give he a procedure that could save at least one life. [NEWLINE] [NEWLINE] Of course, your argument would have more weight<mask> people had sex solely for procreation.<mask> that's not the case anymore; casual, recreational sex is culturally accepted, and a lot of old anti-abortion rhetoric needs to be thrown out to recognize this new societal view. Now, that argument just makes you look a bit self-righteous: "<mask> you losers had abstained from sex like I do, you wouldn't be in this situation"</s>
Label encoding: <s>Let's stay on topic here; pregnancy and abortion aren't the same as the civil war. But abortion can really be a woman claiming ownership of her body. It's understood that the child is depending on it, and that it will die without her support. [NEWLINE] [NEWLINE] The question here is why pregnancy has to be a permanent state, once it happens? Obviously, there are plenty of ways that pregnancy can happen against the mother's wishes. There are plenty of pregnancies that will result in the death of the mother, sometimes both the mother and the child. In those cases, it seems shallow to shout that the mother had it coming and refuse to give he a procedure that could save at least one life. [NEWLINE] [NEWLINE] Of course, your argument would have more weight if people had sex solely for procreation. But that's not the case anymore; casual, recreational sex is culturally accepted, and a lot of old anti-abortion rhetoric needs to be thrown out to recognize this new societal view. Now, that argument just makes you look a bit self-righteous: " if you losers had abstained from sex like I do, you wouldn't be in this situation"</s>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, I openly said that I would have no problem with a men's only period too,<mask> I'm not advocating for discrimination.<mask>, I'm not sure I'd consider it a punishment to be told to come back in a half-hour to the gym; there's a huge difference between oppression and not getting everything you want. Equality isn't possible<mask> we pretend that everyone's starting from square one. [NEWLINE] [NEWLINE] <mask>, about objectification, obviously women objectify men too,<mask> much more rarely does that objectification result in<mask> it can for women, anything from catcalling to rape to getting the shit beaten out of you. Louis CK got it right: the greatest threat to women is men, and the greatest threat to men is heart disease. Put another way, by Margaret Atwood:<mask> a room full of men were asked<mask> they were afraid of women, they said, "they might laugh at me." [NEWLINE] <mask> a room full of women were asked<mask> they were afraid of men, they said, "they might kill me." The objectification may go both ways,<mask> the potential consequences don't. They just don't.</s>
Label encoding: <s>First off, I openly said that I would have no problem with a men's only period too, so I'm not advocating for discrimination. Secondly, I'm not sure I'd consider it a punishment to be told to come back in a half-hour to the gym; there's a huge difference between oppression and not getting everything you want. Equality isn't possible if we pretend that everyone's starting from square one. [NEWLINE] [NEWLINE] Also, about objectification, obviously women objectify men too, but much more rarely does that objectification result in what it can for women, anything from catcalling to rape to getting the shit beaten out of you. Louis CK got it right: the greatest threat to women is men, and the greatest threat to men is heart disease. Put another way, by Margaret Atwood: when a room full of men were asked why they were afraid of women, they said, "they might laugh at me." [NEWLINE] When a room full of women were asked why they were afraid of men, they said, "they might kill me." The objectification may go both ways, but the potential consequences don't. They just don't.</s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Religion compares to none of these. [ENDQ] [NEWLINE] <mask><mask> it's actually worse. The deathcount due to religion is much higher than that of tobacco.  Manipulative and abusive interactions with people happen in religious institutions I would say even more<mask> they are covered up and considered righteous. [NEWLINE] [NEWLINE] [STARTQ] None of these are beliefs, they are lifestyles. [ENDQ] [NEWLINE] Based on beliefs. [NEWLINE] [NEWLINE] [STARTQ] Religion is not a choice like your other examples - not a conscious one any way [ENDQ] [NEWLINE] We disagree there.  Maybe we can debate free will here,<mask><mask> you think one does not choose<mask> to believe in then one<mask> doesn't choose<mask> to be addicted to,<mask> to want to have sex with and<mask> lifestyle to lead,<mask> the point is moot. [NEWLINE] [NEWLINE] [STARTQ] To my knowledge all of the various scriptures explicitly state that it is the job of the parent to rear the child into the fold. [ENDQ] [NEWLINE] Fortunately religious institutions have ways to reinterpret their scripture to fit the times,<mask> we don't have slavery, selling of daughters, slaughter of babies and torture by religious groups,<mask> I am not that worried about that.  </s>
Label encoding: <s> [STARTQ] Religion compares to none of these. [ENDQ] [NEWLINE] I think it's actually worse. The deathcount due to religion is much higher than that of tobacco.  Manipulative and abusive interactions with people happen in religious institutions I would say even more because they are covered up and considered righteous. [NEWLINE] [NEWLINE] [STARTQ] None of these are beliefs, they are lifestyles. [ENDQ] [NEWLINE] Based on beliefs. [NEWLINE] [NEWLINE] [STARTQ] Religion is not a choice like your other examples - not a conscious one any way [ENDQ] [NEWLINE] We disagree there.  Maybe we can debate free will here, but if you think one does not choose what to believe in then one also doesn't choose what to be addicted to, what to want to have sex with and what lifestyle to lead, so the point is moot. [NEWLINE] [NEWLINE] [STARTQ] To my knowledge all of the various scriptures explicitly state that it is the job of the parent to rear the child into the fold. [ENDQ] [NEWLINE] Fortunately religious institutions have ways to reinterpret their scripture to fit the times, so we don't have slavery, selling of daughters, slaughter of babies and torture by religious groups, so I am not that worried about that.  </s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] only has a set amount of time before that drug can be manufactured by other companies. [ENDQ] [NEWLINE] Its called a drug patent. You can patent the chemical formula for normal patent duration. [NEWLINE] [NEWLINE] [STARTQ] <mask> with this first company spending millions, sometimes billions, to create these medications, they need a way to get that money back before cheaper, generic versions flood the market, and end the return on the investment. [ENDQ] [NEWLINE] Note that you do not *need* to recoup the costs of drug research. Its just the predominant economic model of drug creation in the United States. There are several alternative ways you could do drug R&amp;D without having the state sponsored limited duration monopoly model the US uses. [NEWLINE] [NEWLINE] [STARTQ] there is a reason that advertising is allowed, no doubt the doing of some lobbyist group. [ENDQ] [NEWLINE] Well its free speech right now, and the fact the viewing audience accepts this content<mask> acceptable. There have been several campaigns to ban the advertising of fast food toys on TV<mask> there is public dissonance over it, for a counterexample. Not sure on the success of those campaigns<mask>,<mask> I never really watch TV...</s>
Label encoding: <s> [STARTQ] only has a set amount of time before that drug can be manufactured by other companies. [ENDQ] [NEWLINE] Its called a drug patent. You can patent the chemical formula for normal patent duration. [NEWLINE] [NEWLINE] [STARTQ] So with this first company spending millions, sometimes billions, to create these medications, they need a way to get that money back before cheaper, generic versions flood the market, and end the return on the investment. [ENDQ] [NEWLINE] Note that you do not *need* to recoup the costs of drug research. Its just the predominant economic model of drug creation in the United States. There are several alternative ways you could do drug R&amp;D without having the state sponsored limited duration monopoly model the US uses. [NEWLINE] [NEWLINE] [STARTQ] there is a reason that advertising is allowed, no doubt the doing of some lobbyist group. [ENDQ] [NEWLINE] Well its free speech right now, and the fact the viewing audience accepts this content as acceptable. There have been several campaigns to ban the advertising of fast food toys on TV because there is public dissonance over it, for a counterexample. Not sure on the success of those campaigns though, since I never really watch TV...</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm transgender and bisexual. [NEWLINE] [NEWLINE] You're operating under the assumption that LGBTQ+ people want to change. You acknowledge that some people are "proud to be gay,"<mask> only<mask> "they have no other choice." [NEWLINE] [NEWLINE] <mask> I could take a magic pill that would make me straight and cisgender, I wouldn't do it. Not<mask> being trans and queer doesn't suck sometimes,<mask> it really does.<mask> it's part of who I am. It makes things harder,<mask> it<mask> makes my life<mask> much more genuine,<mask> that makes any sense. [NEWLINE] [NEWLINE] I<mask> wonder<mask> you're more inclined to remove the "gay gene" than to remove homophobia. We're already on an upward trend<mask> it comes to eliminating homophobia,<mask><mask> not make the wish to continue that trend rather than get rid of the victims of the problem? [NEWLINE] [NEWLINE] <mask> the solution to intolerance was always to get rid of the people who weren't being tolerated, than intolerant people would win. You hate black people? Gay people? Trans people? Jewish people? Great, just make them go away. Being hateful would suddenly become effective. </s>
Label encoding: <s>I'm transgender and bisexual. [NEWLINE] [NEWLINE] You're operating under the assumption that LGBTQ+ people want to change. You acknowledge that some people are "proud to be gay," but only because "they have no other choice." [NEWLINE] [NEWLINE] If I could take a magic pill that would make me straight and cisgender, I wouldn't do it. Not because being trans and queer doesn't suck sometimes, because it really does. But it's part of who I am. It makes things harder, but it also makes my life so much more genuine, if that makes any sense. [NEWLINE] [NEWLINE] I also wonder why you're more inclined to remove the "gay gene" than to remove homophobia. We're already on an upward trend when it comes to eliminating homophobia, so why not make the wish to continue that trend rather than get rid of the victims of the problem? [NEWLINE] [NEWLINE] If the solution to intolerance was always to get rid of the people who weren't being tolerated, than intolerant people would win. You hate black people? Gay people? Trans people? Jewish people? Great, just make them go away. Being hateful would suddenly become effective. </s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0764, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> exactly should Israel do instead? These are a people who have claimed to want Israel destroyed throughout this whole ordeal<mask> the wars started decades ago. Do you think lifting the blockade would result in anything other than more deaths? [ENDQ] [NEWLINE] Israel needs to reach a permanent negotiated settlement with the Palestinians. Not a stop gap, not a ceasefire, something that really solves the underlying issues taking place in the region. Israel and Palestine will never have peace until both parties are willing to agree to something that is fair. That unfortunately requires painful concessions on both sides. [NEWLINE] [NEWLINE] Israel has to balance its short term interests with its long term ones. Israel wants peace,<mask> can't have it until they end the occupation. Israel won't even consider ending the occupation unless the rockets stop<mask> that wont happen until there is peace! It's a catch-22 here! Fair or not, the ball is in the Israeli court.<mask> Israel seriously wants a peace that will last, there are parties on the Palestinian side that they can work with and support. This policy of rampant death and destruction in Gaza works against those interests and puts peace farther away every time. [NEWLINE] </s>
Label encoding: <s> [STARTQ] What exactly should Israel do instead? These are a people who have claimed to want Israel destroyed throughout this whole ordeal since the wars started decades ago. Do you think lifting the blockade would result in anything other than more deaths? [ENDQ] [NEWLINE] Israel needs to reach a permanent negotiated settlement with the Palestinians. Not a stop gap, not a ceasefire, something that really solves the underlying issues taking place in the region. Israel and Palestine will never have peace until both parties are willing to agree to something that is fair. That unfortunately requires painful concessions on both sides. [NEWLINE] [NEWLINE] Israel has to balance its short term interests with its long term ones. Israel wants peace, but can't have it until they end the occupation. Israel won't even consider ending the occupation unless the rockets stop but that wont happen until there is peace! It's a catch-22 here! Fair or not, the ball is in the Israeli court. If Israel seriously wants a peace that will last, there are parties on the Palestinian side that they can work with and support. This policy of rampant death and destruction in Gaza works against those interests and puts peace farther away every time. [NEWLINE] </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Artist here. [NEWLINE] [NEWLINE] It is my firm opinion that anything that,<mask> purposely affected in any intended manner, expresses something else than its original purpose, it can be considered art. [NEWLINE] [NEWLINE] <mask><mask> this is the widest definition I can think of. [NEWLINE] [NEWLINE] That way things that happen by accident are not art, unless the person who tries to sell them<mask> art. In that fringe case, the narrative is part of the art piece. [NEWLINE] [NEWLINE] Random splats are covered by this rule. Pollock may have splattered paint without ever touching the canvas,<mask> he directed the general direction<mask> the paint would fall, he didn't just buy a bucket of paint and a canvas, left them in his garage, and the paint magically escape the bucket by itself and went to the canvas. [NEWLINE] [NEWLINE] A kid's doodles are art only in the way the kid intends them to.<mask> the kid intends his drawings to be a jet fighter and a house,<mask> drew them<mask> crappy that they look like a dinosaur planting lilies, and someone tries to sell the drawings<mask> the latter rather than the former, it's not art.</s>
Label encoding: <s>Artist here. [NEWLINE] [NEWLINE] It is my firm opinion that anything that, when purposely affected in any intended manner, expresses something else than its original purpose, it can be considered art. [NEWLINE] [NEWLINE] I think this is the widest definition I can think of. [NEWLINE] [NEWLINE] That way things that happen by accident are not art, unless the person who tries to sell them as art. In that fringe case, the narrative is part of the art piece. [NEWLINE] [NEWLINE] Random splats are covered by this rule. Pollock may have splattered paint without ever touching the canvas, but he directed the general direction where the paint would fall, he didn't just buy a bucket of paint and a canvas, left them in his garage, and the paint magically escape the bucket by itself and went to the canvas. [NEWLINE] [NEWLINE] A kid's doodles are art only in the way the kid intends them to. If the kid intends his drawings to be a jet fighter and a house, but drew them so crappy that they look like a dinosaur planting lilies, and someone tries to sell the drawings as the latter rather than the former, it's not art.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask><mask> the thing people are fans of<mask> they've only seen the TV shows or watched the movies is the character, the art, and the story-telling.  All of which is actually very similar between comics, tv, and movies compared to other forms of entertainment like video games (<mask> you have the visuals,<mask> not the same story telling) or novels (no visuals, just story-telling). [NEWLINE] [NEWLINE] I am somewhere in between.  I've read a few comics and found out I am not much of a fan of the medium.  They just seem<mask> flat to me.  Not my flavor. <mask> I love the old animated batman series, am generally pretty happy with the movies, and am a huge fan of the arkham series of games, I<mask> really liked batman beyond (which has no comic source)<mask> it maintains the feel of the characters from the original animated series. [NEWLINE] [NEWLINE] In general,<mask><mask><mask> the characters are right, and the feel is right, it shouldn't matter<mask> the medium is, just be happy with people loving the same thing you do.</s>
Label encoding: <s>I would argue that the thing people are fans of when they've only seen the TV shows or watched the movies is the character, the art, and the story-telling.  All of which is actually very similar between comics, tv, and movies compared to other forms of entertainment like video games ( where you have the visuals, but not the same story telling) or novels (no visuals, just story-telling). [NEWLINE] [NEWLINE] I am somewhere in between.  I've read a few comics and found out I am not much of a fan of the medium.  They just seem so flat to me.  Not my flavor.  But I love the old animated batman series, am generally pretty happy with the movies, and am a huge fan of the arkham series of games, I also really liked batman beyond (which has no comic source) because it maintains the feel of the characters from the original animated series. [NEWLINE] [NEWLINE] In general, as long as the characters are right, and the feel is right, it shouldn't matter what the medium is, just be happy with people loving the same thing you do.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>IMO, the main reason people see more value in Science + Engineering vs. other majors is the sheer difficulty of attempting to study any practical aspects on your own. In this day and age, I can learn just about any amount of theory I want through online textbooks and journal articles. I can practice coding or attempt to solve math/accounting problems, etc. [NEWLINE] [NEWLINE] <mask> 99.9% of people can't do,<mask>, is have access to a lab (organic chemistry, biology, chemistry, etc.) outside of university. Theoretically, even<mask> I had the money to buy the glassware, heat wells, reflux condensers, etc. required, I still wouldn't have the permits to access to the most of the chemicals/cell cultures (legally). Nor would I be able to find a pharmaceutical or research lab willing to take me on<mask> an intern/trainee (given the current degree requirements for those fields). [NEWLINE] [NEWLINE] Edit:<mask> I guess it's a matter of mobility? A scientist could conceivably become a writer or accountant much more easily than the opposite, given no additional schooling.</s>
Label encoding: <s>IMO, the main reason people see more value in Science + Engineering vs. other majors is the sheer difficulty of attempting to study any practical aspects on your own. In this day and age, I can learn just about any amount of theory I want through online textbooks and journal articles. I can practice coding or attempt to solve math/accounting problems, etc. [NEWLINE] [NEWLINE] What 99.9% of people can't do, however, is have access to a lab (organic chemistry, biology, chemistry, etc.) outside of university. Theoretically, even if I had the money to buy the glassware, heat wells, reflux condensers, etc. required, I still wouldn't have the permits to access to the most of the chemicals/cell cultures (legally). Nor would I be able to find a pharmaceutical or research lab willing to take me on as an intern/trainee (given the current degree requirements for those fields). [NEWLINE] [NEWLINE] Edit: So I guess it's a matter of mobility? A scientist could conceivably become a writer or accountant much more easily than the opposite, given no additional schooling.</s>
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In addition to seconding<mask> Glory2Hypnotoad says,<mask><mask><mask> that it takes a very serious kind of push to make someone jump the Trans ship,<mask> to speak.  Someone who feels like an average woman feels,<mask> who has a man's genitals, might not feel<mask> inclined to be trans,<mask> women and men aren't all that distant in personality on an average level.  It's more often the people who feel intensely drawn to overly feminine behavior, almost trope-ishly feminine behavior, who are more likely to become trans. [NEWLINE] [NEWLINE] I'm a trope-ishly feminine guy.  I like baking adorable cupcakes, being taken out on dates, wearing pretty outfits, dressing up, I value romance way more than sex, want to be proposed to, fantasize about planning my wedding, cry at silly animes, all that.  I feel that<mask> I were less intensely this way,<mask> I just generally felt feminine,<mask> without all the very cliche feminine things, that I just wouldn't identify<mask> feminine in any way. [NEWLINE] [NEWLINE] Anything under pressure will need a lot of force to escape.</s>
Label encoding: <s>In addition to seconding what Glory2Hypnotoad says, I think also that it takes a very serious kind of push to make someone jump the Trans ship, so to speak.  Someone who feels like an average woman feels, but who has a man's genitals, might not feel so inclined to be trans, because women and men aren't all that distant in personality on an average level.  It's more often the people who feel intensely drawn to overly feminine behavior, almost trope-ishly feminine behavior, who are more likely to become trans. [NEWLINE] [NEWLINE] I'm a trope-ishly feminine guy.  I like baking adorable cupcakes, being taken out on dates, wearing pretty outfits, dressing up, I value romance way more than sex, want to be proposed to, fantasize about planning my wedding, cry at silly animes, all that.  I feel that if I were less intensely this way, if I just generally felt feminine, but without all the very cliche feminine things, that I just wouldn't identify as feminine in any way. [NEWLINE] [NEWLINE] Anything under pressure will need a lot of force to escape.</s>
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I fail to see any negative consequences in any way<mask> a man gets a woman pregnant. Can you find any? I mean<mask> stops a guy from getting a girl pregnant on a Monday, walking away from that situation, and being in a pick-up bar again by Tuesday?   I don't see anything that stops that behavior. [NEWLINE] [NEWLINE] This is an important question.<mask> a taxpayer, I would just be paying for these children who have a lack of resources. I would be paying<mask> that men could have sex with<mask> many women<mask> they wanted<mask> there is no consequences for this behavior. You don't think that men would take advantage of this system. [NEWLINE] [NEWLINE] And<mask>,<mask> you said, the system is unfair for women, wouldn't you just be taking a system that you think is unfair for men and converting that into a system that unfair for women?  Per your view, one party is free<mask> free can be and the other party is forced to make a choice.  That doesn't seem at all fair to me. Is your perspective just creating a new system that would be unfair for one demographic?</s>
Label encoding: <s>I fail to see any negative consequences in any way if a man gets a woman pregnant. Can you find any? I mean what stops a guy from getting a girl pregnant on a Monday, walking away from that situation, and being in a pick-up bar again by Tuesday?   I don't see anything that stops that behavior. [NEWLINE] [NEWLINE] This is an important question. As a taxpayer, I would just be paying for these children who have a lack of resources. I would be paying so that men could have sex with as many women as they wanted because there is no consequences for this behavior. You don't think that men would take advantage of this system. [NEWLINE] [NEWLINE] And if, as you said, the system is unfair for women, wouldn't you just be taking a system that you think is unfair for men and converting that into a system that unfair for women?  Per your view, one party is free as free can be and the other party is forced to make a choice.  That doesn't seem at all fair to me. Is your perspective just creating a new system that would be unfair for one demographic?</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I noticed<mask> the women were gauging the men's attractiveness they were extremely critical. I swear the women would mark average looking guys<mask> 2's or 3's (a 2 in my view would be Abe Vigoda or Clint Howard). It was really brutal. The show made no mention of this seemingly critical aspect to the assessment of attractiveness. [ENDQ] My take was that women are not really adept at gauging attractiveness from a picture alone and<mask> don't seem to be aware of it. Of course, there are obvious exceptions<mask> otherwise teen hear-throb magazines wouldn't exist. [NEWLINE] [NEWLINE] Basically pulling this out my arse,<mask> couldn't that be<mask> women biologically are wired to find the best possible partner to reproduce with,<mask> men are biologically wired to copulate with<mask> many females<mask> possible? [NEWLINE] [NEWLINE] <mask>,<mask><mask> the point was about female objectification in *public* society. [NEWLINE] [NEWLINE] One example could be the shoe shop next to my gym. They specifically sell shoes to women,<mask> there are a total of 7 pictures of completely naked women - who are not wearing shoes. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] What I noticed when the women were gauging the men's attractiveness they were extremely critical. I swear the women would mark average looking guys as 2's or 3's (a 2 in my view would be Abe Vigoda or Clint Howard). It was really brutal. The show made no mention of this seemingly critical aspect to the assessment of attractiveness. [ENDQ] My take was that women are not really adept at gauging attractiveness from a picture alone and yet don't seem to be aware of it. Of course, there are obvious exceptions because otherwise teen hear-throb magazines wouldn't exist. [NEWLINE] [NEWLINE] Basically pulling this out my arse, but couldn't that be because women biologically are wired to find the best possible partner to reproduce with, while men are biologically wired to copulate with as many females as possible? [NEWLINE] [NEWLINE] Also, I think the point was about female objectification in *public* society. [NEWLINE] [NEWLINE] One example could be the shoe shop next to my gym. They specifically sell shoes to women, but there are a total of 7 pictures of completely naked women - who are not wearing shoes. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The OP seems to want to create a legal double standard based on difficult<mask> not impossible to prove assumptions. [NEWLINE] [NEWLINE] [STARTQ] Furthermore, adolescent boys **are often** physically bigger and stronger than older women,<mask> feelings of coercion in that respect are not<mask> pronounced<mask> the genders were switched. [ENDQ] [NEWLINE] This sounds a lot like victim blaming, which should trip some red flags. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> punishing them<mask> much<mask> older men who prey on young girls is<mask> not appropriate. [ENDQ] [NEWLINE] Honestly, I'm really concerned that this is an acceptable reason to hold adult women to different standards than adult men based on... someone's (it's not clear who) nebulous "feelings of coercion".  Men and women's actions should not be  legally equal<mask> of<mask> people perceive genders?  I can't tell<mask> this is a pro-women or anti-women argument. [NEWLINE] [NEWLINE] [NEWLINE] <mask> is the end goal of a lessor punishment on women that commit statutory rape?  Some kind of gender adjusted pretend "fairness" for committing the same action<mask> the accused happens to have a particular set of genitals?</s>
Label encoding: <s>The OP seems to want to create a legal double standard based on difficult if not impossible to prove assumptions. [NEWLINE] [NEWLINE] [STARTQ] Furthermore, adolescent boys **are often** physically bigger and stronger than older women, so feelings of coercion in that respect are not as pronounced if the genders were switched. [ENDQ] [NEWLINE] This sounds a lot like victim blaming, which should trip some red flags. [NEWLINE] [NEWLINE] [STARTQ] I think punishing them as much as older men who prey on young girls is also not appropriate. [ENDQ] [NEWLINE] Honestly, I'm really concerned that this is an acceptable reason to hold adult women to different standards than adult men based on... someone's (it's not clear who) nebulous "feelings of coercion".  Men and women's actions should not be  legally equal because of how people perceive genders?  I can't tell if this is a pro-women or anti-women argument. [NEWLINE] [NEWLINE] [NEWLINE] What is the end goal of a lessor punishment on women that commit statutory rape?  Some kind of gender adjusted pretend "fairness" for committing the same action because the accused happens to have a particular set of genitals?</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This makes perfect sense in any type of creative or design-oriented atmosphere.<mask><mask> this is<mask><mask> you need creative thinkers with good ideas, it effectively makes your pool of possible employees scarce instead of abundant. This easily makes high wages justifiable,<mask> you can't get that same brain from just anyone.<mask> it doesn't seem to work in a lot of jobs (food service, customer service, etc)<mask> slight variances that are only due to uniqueness (not skill or qualifications) are negligible.<mask><mask> this was the original problem for OP and,<mask> this definitely works in his country, anyone forced to pay maternity leave (for workers that generally do the same level of work<mask> any other workers) will probably have to bite the bullet legally without a principle to justify it personally. At least in the US, it is illegal to offer women lower salaries than men with the same qualifications,<mask> the trade-off would not be an option in the states. Sadly, this may be a part of<mask>'s perpetuating gender discrimination for entry-level jobs: that it simply costs more for businesses to hire women.</s>
Label encoding: <s>This makes perfect sense in any type of creative or design-oriented atmosphere. I think this is because when you need creative thinkers with good ideas, it effectively makes your pool of possible employees scarce instead of abundant. This easily makes high wages justifiable, since you can't get that same brain from just anyone. But it doesn't seem to work in a lot of jobs (food service, customer service, etc) where slight variances that are only due to uniqueness (not skill or qualifications) are negligible. I think this was the original problem for OP and, while this definitely works in his country, anyone forced to pay maternity leave (for workers that generally do the same level of work as any other workers) will probably have to bite the bullet legally without a principle to justify it personally. At least in the US, it is illegal to offer women lower salaries than men with the same qualifications, so the trade-off would not be an option in the states. Sadly, this may be a part of what's perpetuating gender discrimination for entry-level jobs: that it simply costs more for businesses to hire women.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your idea seems to be more based on the belief that it would fix math education than that formal logic is really useful to everybody. [ENDQ] [NEWLINE] Well... sort of, yeah.<mask><mask> that there are a lot of people who could solve math problems competently<mask> feel that they are incapable<mask> they fail to see some of the simple logical rules that govern proofs and other things.<mask><mask> these difficulties come from the way math is taught in schools, even from an early age, and that introducing formal logic might assuage the panicky feeling that high school / college students get<mask> they have to learn 'hard' math. [NEWLINE] [NEWLINE] [STARTQ] The benefits of it will be pretty much to future STEM majors. [ENDQ] [NEWLINE] Maybe the STEM fields would progress even faster,<mask>,<mask> everyone knew logic and was better at reasoning deductively at an early age. These 'average' people might go on to actually realize their potential more completely and do above-average things with mathematics<mask> their already-STEM-inclined classmates would do even *greater* things<mask>. In other words,<mask><mask> it could benefit everyone.</s>
Label encoding: <s> [STARTQ] Your idea seems to be more based on the belief that it would fix math education than that formal logic is really useful to everybody. [ENDQ] [NEWLINE] Well... sort of, yeah. I think that there are a lot of people who could solve math problems competently but feel that they are incapable because they fail to see some of the simple logical rules that govern proofs and other things. I think these difficulties come from the way math is taught in schools, even from an early age, and that introducing formal logic might assuage the panicky feeling that high school / college students get when they have to learn 'hard' math. [NEWLINE] [NEWLINE] [STARTQ] The benefits of it will be pretty much to future STEM majors. [ENDQ] [NEWLINE] Maybe the STEM fields would progress even faster, though, if everyone knew logic and was better at reasoning deductively at an early age. These 'average' people might go on to actually realize their potential more completely and do above-average things with mathematics while their already-STEM-inclined classmates would do even *greater* things yet. In other words, I think it could benefit everyone.</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I would rather bite the bullet paying for care<mask> I need it than having huge increases in taxes to pay for everyone else's.<mask>, I have heard good things from the other side, that a universal healthcare system does work<mask> I have trouble believing that it is any better than paying for my own. [ENDQ] [NEWLINE] <mask> you have trouble believing universal healthcare is better than you paying for it, I would direct you to: [NEWLINE] [NEWLINE] * [**Official OECD statistics about healthcare outcomes in the OECD nations**]( [URL].pdf). You may not be sure which is better.<mask> questions such<mask> child mortality, median age, and life expectancy at birth are pretty straight-forward ways to answer this question. [NEWLINE] [NEWLINE] * [**This CDC Report**]( [URL].pdf), which outlines that medical costs are responsible for the largest share of bankruptcies int he United States. [NEWLINE] [NEWLINE] <mask> basically, one way leads to a massive number of bankruptcies, and the other way leads to the best macro-level healthcare outcomes in the world.  I don't exactly see<mask> one would be indifferent between these two outcomes.  </s>
Label encoding: <s> [STARTQ] I would rather bite the bullet paying for care when I need it than having huge increases in taxes to pay for everyone else's. However, I have heard good things from the other side, that a universal healthcare system does work but I have trouble believing that it is any better than paying for my own. [ENDQ] [NEWLINE] If you have trouble believing universal healthcare is better than you paying for it, I would direct you to: [NEWLINE] [NEWLINE] * [**Official OECD statistics about healthcare outcomes in the OECD nations**]( [URL].pdf). You may not be sure which is better. but questions such as child mortality, median age, and life expectancy at birth are pretty straight-forward ways to answer this question. [NEWLINE] [NEWLINE] * [**This CDC Report**]( [URL].pdf), which outlines that medical costs are responsible for the largest share of bankruptcies int he United States. [NEWLINE] [NEWLINE] So basically, one way leads to a massive number of bankruptcies, and the other way leads to the best macro-level healthcare outcomes in the world.  I don't exactly see why one would be indifferent between these two outcomes.  </s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your post talks of a "punctual mistake"<mask> you almost certainly mean "punctuation mistake." (And even<mask> you meant to say that the person isn't punctual, "punctual mistake" wouldn't be the correct way to describe it either.) Muphry's Law strikes again! [NEWLINE] [NEWLINE] True, you didn't need to quadruple-check this post and the job applicant should have quadruple-checked their letter.<mask> an indicator to the applicant's attention to detail, it's pretty damning. [NEWLINE] [NEWLINE] <mask> attention to detail is not the most important job requirement for every position.<mask> does a newspaper have an editing department?<mask> a journalist needs to deliver copy fast, which invariably results in mistakes. And that's fine: speed beats correctness in this case. [NEWLINE] [NEWLINE] I'm checking resumes for a Technical Writer position daily, and<mask> a typo or spelling mistake counts very heavily against a candidate, it doesn't automatically discount them, even in this situation.<mask> the person has an otherwise impeccable CV, it stands to reason that they simply made a mistake. Everybody makes mistakes. [NEWLINE] </s>
Label encoding: <s>Your post talks of a "punctual mistake" when you almost certainly mean "punctuation mistake." (And even if you meant to say that the person isn't punctual, "punctual mistake" wouldn't be the correct way to describe it either.) Muphry's Law strikes again! [NEWLINE] [NEWLINE] True, you didn't need to quadruple-check this post and the job applicant should have quadruple-checked their letter. As an indicator to the applicant's attention to detail, it's pretty damning. [NEWLINE] [NEWLINE] But attention to detail is not the most important job requirement for every position. Why does a newspaper have an editing department? Because a journalist needs to deliver copy fast, which invariably results in mistakes. And that's fine: speed beats correctness in this case. [NEWLINE] [NEWLINE] I'm checking resumes for a Technical Writer position daily, and while a typo or spelling mistake counts very heavily against a candidate, it doesn't automatically discount them, even in this situation. If the person has an otherwise impeccable CV, it stands to reason that they simply made a mistake. Everybody makes mistakes. [NEWLINE] </s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't want to date or marry someone who is anti-abortion,<mask> that's not<mask> I believe in. I want someone who would support her friend's decision to abort,<mask> she herself wouldn't make that same choice<mask> it was her, just<mask> I would do. [ENDQ] [NEWLINE] <mask><mask><mask> it's an odd distinction to make, my point still stands:<mask><mask><mask> you know her stance upfront and you choose to date her<mask> of that existing stance, then<mask><mask> that's OK. And you have already conceded that changing her mind for your own sake is not<mask> you intend. [NEWLINE] [NEWLINE] The only potential remaining issue I could see is the effect on the women who you reject based on these criteria. From studies of the effects of race preferences on dating sites, we know that,<mask> mostly non-white partners are frequently turned down<mask> of their non-whiteness, this does affect them negatively over time. [NEWLINE] [NEWLINE] <mask> I guess this would only apply to your situation<mask> men were routinely turning down women<mask> of their pro-choice stance, which probably isn't the case.</s>
Label encoding: <s> [STARTQ] I don't want to date or marry someone who is anti-abortion, because that's not what I believe in. I want someone who would support her friend's decision to abort, while she herself wouldn't make that same choice if it was her, just as I would do. [ENDQ] [NEWLINE] While I think it's an odd distinction to make, my point still stands: as long as you know her stance upfront and you choose to date her because of that existing stance, then I think that's OK. And you have already conceded that changing her mind for your own sake is not what you intend. [NEWLINE] [NEWLINE] The only potential remaining issue I could see is the effect on the women who you reject based on these criteria. From studies of the effects of race preferences on dating sites, we know that, where mostly non-white partners are frequently turned down because of their non-whiteness, this does affect them negatively over time. [NEWLINE] [NEWLINE] But I guess this would only apply to your situation if men were routinely turning down women because of their pro-choice stance, which probably isn't the case.</s>
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know it seems silly to put a warning like that on a hair dryer,<mask> a lot of people think that appliances with Ground Fault Circuit Interrupters (that big, boxy plug on hair dryers) will automatically cut the power<mask> the appliance falls into water.  And yeah, that's<mask> a properly functioning GFCI is designed to do.<mask> a GFCI can break, or wear out, or be defective,<mask> you still shouldn't risk getting a hair dryer wet. <mask> the warning label is still important.  A GFCI may save your life in an emergency,<mask> ideally, you should never put yourself in a situation<mask> you need to rely on it.  That's true of most safety equipment. [NEWLINE] [NEWLINE] People tend to behave much less safely<mask> they *think* technology is protecting them - for instance, people tend to behave much less cautiously<mask> they bike<mask> wearing a helmet. <mask> you still need to remind them that they need to keep following the same common-sense safety precautions even<mask> the product has some kind of emergency safety feature.</s>
Label encoding: <s>I know it seems silly to put a warning like that on a hair dryer, but a lot of people think that appliances with Ground Fault Circuit Interrupters (that big, boxy plug on hair dryers) will automatically cut the power if the appliance falls into water.  And yeah, that's what a properly functioning GFCI is designed to do. But a GFCI can break, or wear out, or be defective, so you still shouldn't risk getting a hair dryer wet.  So the warning label is still important.  A GFCI may save your life in an emergency, but ideally, you should never put yourself in a situation where you need to rely on it.  That's true of most safety equipment. [NEWLINE] [NEWLINE] People tend to behave much less safely when they *think* technology is protecting them - for instance, people tend to behave much less cautiously when they bike while wearing a helmet.  So you still need to remind them that they need to keep following the same common-sense safety precautions even when the product has some kind of emergency safety feature.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I’m a therapist.  The things we work on in therapy are only a very small sliver of who we are.  People have repetition compulsions, anger, emotional outbursts with unclear origin.  Just<mask> we don’t know<mask> they come from-- doesn’t mean they don’t exist. [NEWLINE] [NEWLINE] <mask> we sleep we lose consciousness. Does that make us any less alive until we wake up? [NEWLINE] [NEWLINE] The way you framed your question: there is<mask> a controlling entity that controls existence.  People study it in physics: Energy is neither created nor destroyed.  That which is in motion, stays in motion.  Molecules and compounds that make us exist continue to exist. [NEWLINE] [NEWLINE] People continue in some form.  Maybe there is a chance another entity will manifest from your elements. Just<mask> it wasn’t aware of origin doesn’t make it any less connected... sure, it probably won’t happen,<mask> it could. [NEWLINE] [NEWLINE] And<mask> you recognize that possibility: You believe in a chance of an afterlife.</s>
Label encoding: <s>I’m a therapist.  The things we work on in therapy are only a very small sliver of who we are.  People have repetition compulsions, anger, emotional outbursts with unclear origin.  Just because we don’t know where they come from-- doesn’t mean they don’t exist. [NEWLINE] [NEWLINE] When we sleep we lose consciousness. Does that make us any less alive until we wake up? [NEWLINE] [NEWLINE] The way you framed your question: there is indeed a controlling entity that controls existence.  People study it in physics: Energy is neither created nor destroyed.  That which is in motion, stays in motion.  Molecules and compounds that make us exist continue to exist. [NEWLINE] [NEWLINE] People continue in some form.  Maybe there is a chance another entity will manifest from your elements. Just because it wasn’t aware of origin doesn’t make it any less connected... sure, it probably won’t happen, but it could. [NEWLINE] [NEWLINE] And if you recognize that possibility: You believe in a chance of an afterlife.</s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The people involved with them a strong interest in keeping their organizations<mask> cost-efficient<mask> possible than the average person does. [ENDQ] [NEWLINE] yes, that's<mask> I'm convinced capitalism is the problem. [NEWLINE] [NEWLINE] [STARTQ] I never suggested obliging everyone to be a vegan. [ENDQ] [NEWLINE] Oh, no, don't take it personally. It's just that the little vegans I know are very vocal about it, and really condemn other people for not agreeing with them. I'm sure not every vegan is like that, and I don't know/it doesn't matter<mask> you are or not. [NEWLINE] [NEWLINE] [STARTQ] which is<mask> I'm confused by your tendency to bring meat consumption up in this discussion [ENDQ] [NEWLINE] <mask> many vegetarians, aren't strict vegetarians. Many of them eat meat once a week or once every two weeks,<mask> it's a pain in the ass to explain that to other people,<mask> they stick to a simple label. Vegans<mask><mask><mask><mask> are taking things to extremes (no figs. no free-range-eggs) and/<mask> are consequent in it.</s><pad>
Label encoding: <s> [STARTQ] The people involved with them a strong interest in keeping their organizations as cost-efficient as possible than the average person does. [ENDQ] [NEWLINE] yes, that's why I'm convinced capitalism is the problem. [NEWLINE] [NEWLINE] [STARTQ] I never suggested obliging everyone to be a vegan. [ENDQ] [NEWLINE] Oh, no, don't take it personally. It's just that the little vegans I know are very vocal about it, and really condemn other people for not agreeing with them. I'm sure not every vegan is like that, and I don't know/it doesn't matter if you are or not. [NEWLINE] [NEWLINE] [STARTQ] which is why I'm confused by your tendency to bring meat consumption up in this discussion [ENDQ] [NEWLINE] because many vegetarians, aren't strict vegetarians. Many of them eat meat once a week or once every two weeks, but it's a pain in the ass to explain that to other people, so they stick to a simple label. Vegans on the other hand are taking things to extremes (no figs. no free-range-eggs) and/ but are consequent in it.</s><pad>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you find yourself defending that kind of despicable behaviour, you need to get a reality check. It adds nothing to society. We are all worse off for its tolerance. [ENDQ] [NEWLINE] The question becomes who draws the line, and who enforces it? <mask> you take away these kinds of freedoms, you're left with shithole countries like Afghanistan<mask> gays are slaughtered in the streets, women *can't* get abortions, and women are beat daily. <mask> you start revoking freedoms, you allow for a radical group (in America that would be Christian fundamentalists, etc) to take power in a vacuum and implement laws against anything their ideology disagrees with. That's<mask> there are free speech laws, to protect the ideas you disagree with,<mask> the majority are often worn, and<mask> the majority allow the minority to be wrong, too. [NEWLINE] [NEWLINE] I personally think Klan rallies and anti-abortion pickets or abhorrent,<mask> these people have the legal right to have their say, the legal right to have their voices heard - no matter<mask> offensive their views are to you. </s>
Label encoding: <s> [STARTQ] If you find yourself defending that kind of despicable behaviour, you need to get a reality check. It adds nothing to society. We are all worse off for its tolerance. [ENDQ] [NEWLINE] The question becomes who draws the line, and who enforces it?  When you take away these kinds of freedoms, you're left with shithole countries like Afghanistan where gays are slaughtered in the streets, women *can't* get abortions, and women are beat daily.  If you start revoking freedoms, you allow for a radical group (in America that would be Christian fundamentalists, etc) to take power in a vacuum and implement laws against anything their ideology disagrees with. That's why there are free speech laws, to protect the ideas you disagree with, because the majority are often worn, and because the majority allow the minority to be wrong, too. [NEWLINE] [NEWLINE] I personally think Klan rallies and anti-abortion pickets or abhorrent, but these people have the legal right to have their say, the legal right to have their voices heard - no matter how offensive their views are to you. </s>
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. My face looks better with a beard. That is a matter of preference. [NEWLINE] [NEWLINE] 2. Initially I started<mask> I was too lazy to shave, then I got used to having facial hair, then it felt weird without any facial hair. [NEWLINE] [NEWLINE] 3. I<mask> dont like long gandalf style beards, I dont think most people do,<mask> a well-groomed, short beard looks nice on most people. There is nothing gross about it, unless there are food particles and/or dead things in there, it shouldn't elicit a 'gross' response. [NEWLINE] [NEWLINE] 4. A lot of time insecurities about beards stem from ones own incapability of growing a proper beard. I don't know, maybe its a subconscious dislike of beards<mask> you don't like your own. [NEWLINE] [NEWLINE] 4.<mask> there is the common stigma of beards being associated with pride and "manliness" in some cases its true<mask> men are the only ones that can grow beards,<mask> most of the time its not a primary reason for having a beard. </s>
Label encoding: <s>1. My face looks better with a beard. That is a matter of preference. [NEWLINE] [NEWLINE] 2. Initially I started because I was too lazy to shave, then I got used to having facial hair, then it felt weird without any facial hair. [NEWLINE] [NEWLINE] 3. I also dont like long gandalf style beards, I dont think most people do, but a well-groomed, short beard looks nice on most people. There is nothing gross about it, unless there are food particles and/or dead things in there, it shouldn't elicit a 'gross' response. [NEWLINE] [NEWLINE] 4. A lot of time insecurities about beards stem from ones own incapability of growing a proper beard. I don't know, maybe its a subconscious dislike of beards since you don't like your own. [NEWLINE] [NEWLINE] 4. Also there is the common stigma of beards being associated with pride and "manliness" in some cases its true as men are the only ones that can grow beards, but most of the time its not a primary reason for having a beard. </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Unfortunately there may come a point, were your life will get fucked up. Maybe it's an accident, maybe something else. Maybe financial problems, or a psychosis, everything is possible. Who will stay with you, support you,<mask> they get nothing in return? Friends? Even<mask> you have really good friends it would surprise me. Probably they will leave quiet, some sooner, the best ones later.<mask> they will leave,<mask> they have their own lifes that can exist without you. A<mask>? Maybe, maybe not.<mask> you are very lucky then maybe they do.<mask> often this kind of bond is not strong enough. Your parents will stay. They put up with all your shit in your childhood (You said they did care for you,<mask> I'll assume it was a healthy relationship between you and them). You are<mask> will stay from them<mask> they die. You are simply the most important thing for them.<mask>, you will always can rely on them.<mask> you are not able to love them for<mask> they have already done for you, love them for<mask> they might do.</s>
Label encoding: <s>Unfortunately there may come a point, were your life will get fucked up. Maybe it's an accident, maybe something else. Maybe financial problems, or a psychosis, everything is possible. Who will stay with you, support you, when they get nothing in return? Friends? Even if you have really good friends it would surprise me. Probably they will leave quiet, some sooner, the best ones later. But they will leave, because they have their own lifes that can exist without you. A SO? Maybe, maybe not. When you are very lucky then maybe they do. But often this kind of bond is not strong enough. Your parents will stay. They put up with all your shit in your childhood (You said they did care for you, so I'll assume it was a healthy relationship between you and them). You are what will stay from them when they die. You are simply the most important thing for them. So, you will always can rely on them. If you are not able to love them for what they have already done for you, love them for what they might do.</s>
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> about churches that want to ban interracial marriages? Should we let them do it<mask> it's their faith? [ENDQ] [NEWLINE] Which denominations prohibit interracial marriage? I can't find any, unless they're pretty obscure. Is this meant<mask> a hypothetical? [NEWLINE] [NEWLINE] [STARTQ] You are proposing to have separate churches for the gays and the non-gays, essentially providing the same service for the different groups of people. [ENDQ] [NEWLINE] Don't we already have separate churches providing the same services for different groups of people? This is<mask> we have separate Catholic, Baptist, Episcopalian, etc. churches. Different churches appeal to different beliefs and different styles of worship. [NEWLINE] [NEWLINE] [STARTQ] Except gays won't be allowed in most churches, creating a serious problem in terms of equality. [ENDQ] [NEWLINE] I'm pretty sure gays are already allowed in most churches. [NEWLINE] [NEWLINE] [STARTQ] This "separate<mask> equal" approach has been a huge failure in the past. [ENDQ] [NEWLINE] Separate<mask> equal applies to public accommodations. A church isn't providing a public accommodation. A church is basically a private group or club.</s>
Label encoding: <s> [STARTQ] What about churches that want to ban interracial marriages? Should we let them do it since it's their faith? [ENDQ] [NEWLINE] Which denominations prohibit interracial marriage? I can't find any, unless they're pretty obscure. Is this meant as a hypothetical? [NEWLINE] [NEWLINE] [STARTQ] You are proposing to have separate churches for the gays and the non-gays, essentially providing the same service for the different groups of people. [ENDQ] [NEWLINE] Don't we already have separate churches providing the same services for different groups of people? This is why we have separate Catholic, Baptist, Episcopalian, etc. churches. Different churches appeal to different beliefs and different styles of worship. [NEWLINE] [NEWLINE] [STARTQ] Except gays won't be allowed in most churches, creating a serious problem in terms of equality. [ENDQ] [NEWLINE] I'm pretty sure gays are already allowed in most churches. [NEWLINE] [NEWLINE] [STARTQ] This "separate but equal" approach has been a huge failure in the past. [ENDQ] [NEWLINE] Separate but equal applies to public accommodations. A church isn't providing a public accommodation. A church is basically a private group or club.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean, for all you know, maybe Rodman was faced with the odd situation of having KJU want to be his friend, and he thinks that maybe<mask> he's nice to him he can kind of sway him in a more positive direction<mask> KJU probably looks up to him<mask> some kind of role model due to his celebrity status.  I'm not saying that's the case,<mask> essentially your whole objection is that he's friends with him and you don't like it.  You don't have even a remote clue<mask> his intentions are,<mask> it's impossible to judge such a situation. [NEWLINE] [NEWLINE] Personally<mask>,<mask> I were a celebrity and it kind of fell into my lap that KJU wants to be my best friend, I would sure<mask> hell jump on that opportunity. <mask> he idolizes me, perhaps my gentle criticisms won't fall on deaf ears.  "Hey, my celebrity idol thinks somewhat badly of me<mask> my people are starving...I want him to like me,<mask> maybe I should increase rations." [NEWLINE] [NEWLINE] Who knows.</s>
Label encoding: <s>I mean, for all you know, maybe Rodman was faced with the odd situation of having KJU want to be his friend, and he thinks that maybe if he's nice to him he can kind of sway him in a more positive direction since KJU probably looks up to him as some kind of role model due to his celebrity status.  I'm not saying that's the case, but essentially your whole objection is that he's friends with him and you don't like it.  You don't have even a remote clue what his intentions are, so it's impossible to judge such a situation. [NEWLINE] [NEWLINE] Personally though, if I were a celebrity and it kind of fell into my lap that KJU wants to be my best friend, I would sure as hell jump on that opportunity.  If he idolizes me, perhaps my gentle criticisms won't fall on deaf ears.  "Hey, my celebrity idol thinks somewhat badly of me because my people are starving...I want him to like me, so maybe I should increase rations." [NEWLINE] [NEWLINE] Who knows.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are many great opportunities for interacting with others and making a positive impact on the world that<mask> you are really able to retire in your 30s, you're likely to be interested in, would be challenging, and you could feel good about. Playing games is fun<mask> a break,<mask> most people find that it eventually palls. [NEWLINE] [NEWLINE] <mask> we'll start at the beginning;<mask> would you do,<mask> there were no constraints, and you could pick any type of activity you can imagine? Would your answer change<mask> you were going to be doing it for a decade straight?<mask> the answer isn't "Play WoW,"<mask> not quit your current job and pursue one somewhere else, or in something you enjoy more? Worst case, you can find a job in a similar type of work that is less demanding; at a smaller company, part time, or even volunteering doing something related for a nonprofit. You can play video games in your spare time,<mask><mask><mask> you'd find that giving up on work for the next 60 years wouldn't be<mask> much fun<mask> you imagine.</s><pad>
Label encoding: <s>There are many great opportunities for interacting with others and making a positive impact on the world that if you are really able to retire in your 30s, you're likely to be interested in, would be challenging, and you could feel good about. Playing games is fun as a break, but most people find that it eventually palls. [NEWLINE] [NEWLINE] So we'll start at the beginning; what would you do, if there were no constraints, and you could pick any type of activity you can imagine? Would your answer change if you were going to be doing it for a decade straight? If the answer isn't "Play WoW," why not quit your current job and pursue one somewhere else, or in something you enjoy more? Worst case, you can find a job in a similar type of work that is less demanding; at a smaller company, part time, or even volunteering doing something related for a nonprofit. You can play video games in your spare time, but I think you'd find that giving up on work for the next 60 years wouldn't be as much fun as you imagine.</s><pad>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, I've seen the argument that the fetus is not a human being,<mask> it's just a clump of cells that has the potential to become a baby.<mask><mask> with this<mask> well,<mask> fundamentally speaking,<mask> nothing is altered within the woman's body, the fetus *will* grow, and *will* be born. Nothing will halt this other than external actions (barring complications and<mask> on). I don't understand<mask> some people do not consider that "alive" or "human". [ENDQ] [NEWLINE] Yes, a fetus is both alive and human,<mask> that's not the issue. The same can be said for tumors. The question is whether the fetus can already be considered a person and has the rights associated with that. I don't think it does, especially early in its development,<mask> it doesn't even have a brain to speak of. It really is just a clump of cells, at this point. The fact that it might become a person later (<mask> all goes well) doesn't change the fact that it isn't now. </s>
Label encoding: <s> [STARTQ] Also, I've seen the argument that the fetus is not a human being, as it's just a clump of cells that has the potential to become a baby. I disagree with this as well, as fundamentally speaking, if nothing is altered within the woman's body, the fetus *will* grow, and *will* be born. Nothing will halt this other than external actions (barring complications and so on). I don't understand how some people do not consider that "alive" or "human". [ENDQ] [NEWLINE] Yes, a fetus is both alive and human, but that's not the issue. The same can be said for tumors. The question is whether the fetus can already be considered a person and has the rights associated with that. I don't think it does, especially early in its development, when it doesn't even have a brain to speak of. It really is just a clump of cells, at this point. The fact that it might become a person later ( if all goes well) doesn't change the fact that it isn't now. </s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> gender dysphoria was better treated through counseling to stay the gender they were assigned at birth, then the medical community would do that. [ENDQ] [NEWLINE] Former GLBT therapist here, and this is<mask> I came here to say. Gender dysphoria is pretty unique among mental health conditions in that it goes away<mask> the person transitions. You have dysphoria, you transition, BOOM you don't have dysphoria anymore. For people who<mask> undergo gender confirmation surgery, it's basically unheard of for them to have any regrets. This is in stark contrast to body dysmorphic disorder (being obsessively unhappy with a perceived flaw with your body);<mask> someone gets plastic surgery<mask> of BDD, they are usually NOT happy with the results. [NEWLINE] [NEWLINE] That in itself is enough to tell me that transitioning genders really does fix the underlying problem.<mask> we were just conceding to people's delusion that they're a different sex, we'd expect them to act like people with BDD after surgery.<mask> instead, we see that transitioning *cures their mental illness.* That's pretty powerful.</s><pad><pad><pad>
Label encoding: <s> [STARTQ] If gender dysphoria was better treated through counseling to stay the gender they were assigned at birth, then the medical community would do that. [ENDQ] [NEWLINE] Former GLBT therapist here, and this is what I came here to say. Gender dysphoria is pretty unique among mental health conditions in that it goes away when the person transitions. You have dysphoria, you transition, BOOM you don't have dysphoria anymore. For people who also undergo gender confirmation surgery, it's basically unheard of for them to have any regrets. This is in stark contrast to body dysmorphic disorder (being obsessively unhappy with a perceived flaw with your body); if someone gets plastic surgery because of BDD, they are usually NOT happy with the results. [NEWLINE] [NEWLINE] That in itself is enough to tell me that transitioning genders really does fix the underlying problem. If we were just conceding to people's delusion that they're a different sex, we'd expect them to act like people with BDD after surgery. But instead, we see that transitioning *cures their mental illness.* That's pretty powerful.</s><pad><pad><pad>
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> we were to get rid of all empathy, we would be more successful beings, individually. [ENDQ] [NEWLINE] Yes, individually we would be better,<mask> obviously, we would be worse off<mask> a society. The quality of relationships would be seriously affected, and I'm predicting a drop in global fertility rates simply<mask> more people are single. Growing up would be tough<mask> your parents didn't empathize with you, and<mask> it's been shown that psychopaths, serial killers, whatever messed up people nowadays typically have had a poor childhood and neglected by parents. Keep in mind that this would be the case<mask> empathy was non-existent. [NEWLINE] [NEWLINE] [STARTQ] I feel that<mask> observing one's specific life, empathy does nothing<mask> hinder one from becoming the best they can be [ENDQ] [NEWLINE] I would<mask><mask> empathy can help someone in their specific life,<mask> in dealing with any other person, empathy can help improve that relationship, and<mask> improving their life...unless you're a hermit living in the woods with no human interaction, in which case empathy is probably a foreign concept to you anyways.</s>
Label encoding: <s> [STARTQ] If we were to get rid of all empathy, we would be more successful beings, individually. [ENDQ] [NEWLINE] Yes, individually we would be better, but obviously, we would be worse off as a society. The quality of relationships would be seriously affected, and I'm predicting a drop in global fertility rates simply because more people are single. Growing up would be tough if your parents didn't empathize with you, and indeed it's been shown that psychopaths, serial killers, whatever messed up people nowadays typically have had a poor childhood and neglected by parents. Keep in mind that this would be the case if empathy was non-existent. [NEWLINE] [NEWLINE] [STARTQ] I feel that when observing one's specific life, empathy does nothing but hinder one from becoming the best they can be [ENDQ] [NEWLINE] I would argue that empathy can help someone in their specific life, because in dealing with any other person, empathy can help improve that relationship, and thus improving their life...unless you're a hermit living in the woods with no human interaction, in which case empathy is probably a foreign concept to you anyways.</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>With that being said, I would go<mask> far to say that institutional racism, racial discrimination, and racism aren't the same thing. Each of those modifiers to the phrase should add meaning to the definition. Racial discrimination being the errant recognition of differences in race, whether that be detrimental or beneficial; Racism being the application of said racial discrimination; institutional racism being the systemic existence of racist ideologies from a societal/broader scope. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] discrimination against racial minorities is a far more serious problem on a national level (<mask> opposed to an individual level of an unfortunate white kid in a poor black neighbourhood). [ENDQ] [NEWLINE] [NEWLINE] I whole-heartedly agree on the seriousness of black oppression, and the value in identifying and seperating the impacts of black vs. the idea of white oppression,<mask> that sharply, and dangerously minimizes the impacts of racism at it's roots. The root of which being hate. I feel like identifying hate and eliminating it is<mask> is important, and the exclusion of Mustafa's actions from<mask> would be considered racist is reductive and not socially progressive.</s>
Label encoding: <s>With that being said, I would go as far to say that institutional racism, racial discrimination, and racism aren't the same thing. Each of those modifiers to the phrase should add meaning to the definition. Racial discrimination being the errant recognition of differences in race, whether that be detrimental or beneficial; Racism being the application of said racial discrimination; institutional racism being the systemic existence of racist ideologies from a societal/broader scope. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] discrimination against racial minorities is a far more serious problem on a national level ( as opposed to an individual level of an unfortunate white kid in a poor black neighbourhood). [ENDQ] [NEWLINE] [NEWLINE] I whole-heartedly agree on the seriousness of black oppression, and the value in identifying and seperating the impacts of black vs. the idea of white oppression, but that sharply, and dangerously minimizes the impacts of racism at it's roots. The root of which being hate. I feel like identifying hate and eliminating it is what is important, and the exclusion of Mustafa's actions from what would be considered racist is reductive and not socially progressive.</s>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It was not a matter of ignorance.  Just<mask> a lot of other people do something, it does not mean it should be done.  It is common to appeal to the actions of the masses,<mask> it results in a fallacious argument ([argumentum ad populum]( [URL] )).  The modern societies that supposedly welcome this behavior likely do<mask><mask> of the desire for many people nowadays to be more "welcoming" to those who are different,<mask>, in actuality, most of them are really doing<mask> to feel that they personally are better people.  Many philosophies related to egoism argue, very compellingly, that every action a person takes is for their own good. Even donating to charity is only done<mask> the wealthy want to feel the "warm fuzzy feeling" after they feel like they have helped someone less fortunate.  I feel that pushing for the equality of the "other" gender is the same thing.  People want to feel the warm fuzzies from feeling like they are improving the lives of the oppressed individuals who do not have a gender identification.</s>
Label encoding: <s>It was not a matter of ignorance.  Just because a lot of other people do something, it does not mean it should be done.  It is common to appeal to the actions of the masses, but it results in a fallacious argument ([argumentum ad populum]( [URL] )).  The modern societies that supposedly welcome this behavior likely do so because of the desire for many people nowadays to be more "welcoming" to those who are different, when, in actuality, most of them are really doing so to feel that they personally are better people.  Many philosophies related to egoism argue, very compellingly, that every action a person takes is for their own good. Even donating to charity is only done because the wealthy want to feel the "warm fuzzy feeling" after they feel like they have helped someone less fortunate.  I feel that pushing for the equality of the "other" gender is the same thing.  People want to feel the warm fuzzies from feeling like they are improving the lives of the oppressed individuals who do not have a gender identification.</s>
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Again: "No" would've been a perfectly acceptable answer. They could've said "No"<mask> I asked<mask> they were still open. They could've said "No"<mask> I told there it would be a large party.<mask> I told them we didn't want a limited menu, and said we were leaving then, they could've said "Have a nice night". They chose to, not only say "Yes",<mask> to do<mask> over and over again. [ENDQ] [NEWLINE] It sounds like the waitstaff wanted you gone,<mask> were required by the manager to seat and serve you.<mask> again, you didn't break any rules,<mask> basic empathy would generally alert you to the fact that the waitstaff didn't want to be there, and would probably be fired<mask> they asked you to leave.<mask> you're putting people in a position that requires them to choose between being inconvenienced or being fired, then you're being a dick. You just are. You're entitled to be a dick,<mask> that doesn't mean there's nothing wrong with it. </s>
Label encoding: <s> [STARTQ] Again: "No" would've been a perfectly acceptable answer. They could've said "No" when I asked if they were still open. They could've said "No" when I told there it would be a large party. When I told them we didn't want a limited menu, and said we were leaving then, they could've said "Have a nice night". They chose to, not only say "Yes", but to do so over and over again. [ENDQ] [NEWLINE] It sounds like the waitstaff wanted you gone, but were required by the manager to seat and serve you. So again, you didn't break any rules, but basic empathy would generally alert you to the fact that the waitstaff didn't want to be there, and would probably be fired if they asked you to leave. If you're putting people in a position that requires them to choose between being inconvenienced or being fired, then you're being a dick. You just are. You're entitled to be a dick, but that doesn't mean there's nothing wrong with it. </s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] it's going to be quite a hassle sending feminine products to the front-lines, [ENDQ] [NEWLINE] Put all women on birth control (religious issues being the exception, not the rule) and don't send women on long term missions or to slightly safer areas<mask> the risk of extending the mission is less. Yes, it's still not equal,<mask> it's more equal and there is a biological reason for it. [NEWLINE] [NEWLINE] [STARTQ] <mask> no combat is involved, someone still has to take care of the families at home [ENDQ] [NEWLINE] At<mask> age is the military service?<mask> it's like Germany used to be,<mask> many 18 year olds have families to take care of?<mask> they do, that should be the exception, not the rule. [NEWLINE] [NEWLINE] Edit:<mask>, to say men will rape women<mask> they are bored doesn't give them any credit. Yes, rapes in the military do happen,<mask> they should address the issue way better, not hide it or try to ignore it. Do you think gays shouldn't be allowed to join the military?<mask>'s the difference?</s>
Label encoding: <s> [STARTQ] it's going to be quite a hassle sending feminine products to the front-lines, [ENDQ] [NEWLINE] Put all women on birth control (religious issues being the exception, not the rule) and don't send women on long term missions or to slightly safer areas where the risk of extending the mission is less. Yes, it's still not equal, but it's more equal and there is a biological reason for it. [NEWLINE] [NEWLINE] [STARTQ] If no combat is involved, someone still has to take care of the families at home [ENDQ] [NEWLINE] At what age is the military service? If it's like Germany used to be, how many 18 year olds have families to take care of? If they do, that should be the exception, not the rule. [NEWLINE] [NEWLINE] Edit: Also, to say men will rape women because they are bored doesn't give them any credit. Yes, rapes in the military do happen, but they should address the issue way better, not hide it or try to ignore it. Do you think gays shouldn't be allowed to join the military? What's the difference?</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be fair, I rarely eat meat and<mask> I do it is a very small amount of it. [NEWLINE] [NEWLINE] Generally my beef is with the idea of eating carnivores.  There is a huge loss of energy<mask> eating meat that eats meat.  There is<mask> a high risk of pollutants that can build up from the chain of eating meat that eats meat.  Carnivores are<mask> typically leaner and gamey. [NEWLINE] [NEWLINE] Dogs are omnivores with a bias to carnivorous.  My understanding of dog farms is that the dogs are on a strict vegetarian diet (which requires nutrient supplements in order to give the animal their complete amino acids<mask> without them the dog is sickened and I would hardly think eating a sick dog is good). [NEWLINE] [NEWLINE] Spending the money and energy on raising a dog that is healthy enough for human consumption would possibly make sense<mask> they were delicious<mask> anecdotal evidence from people I know....it isn't.  It gets hidden in stews. [NEWLINE] [NEWLINE] Overall, the whole thing doesn't make sense to me.  </s>
Label encoding: <s>To be fair, I rarely eat meat and when I do it is a very small amount of it. [NEWLINE] [NEWLINE] Generally my beef is with the idea of eating carnivores.  There is a huge loss of energy when eating meat that eats meat.  There is also a high risk of pollutants that can build up from the chain of eating meat that eats meat.  Carnivores are also typically leaner and gamey. [NEWLINE] [NEWLINE] Dogs are omnivores with a bias to carnivorous.  My understanding of dog farms is that the dogs are on a strict vegetarian diet (which requires nutrient supplements in order to give the animal their complete amino acids because without them the dog is sickened and I would hardly think eating a sick dog is good). [NEWLINE] [NEWLINE] Spending the money and energy on raising a dog that is healthy enough for human consumption would possibly make sense if they were delicious but anecdotal evidence from people I know....it isn't.  It gets hidden in stews. [NEWLINE] [NEWLINE] Overall, the whole thing doesn't make sense to me.  </s>
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> dealing with racists, I tend to use the strategy outlined, just continually ask them to explain their views and watch<mask> they get more and more moderate. [ENDQ] [NEWLINE] I once got an extreme conservative who was arguing against universal healthcare to support universal healthcare in the same conversation by using that strategy. [NEWLINE] [NEWLINE] She was saying<mask> taxpayer money shouldn't go to other people's medical bills who can't pay for it themselves and<mask> it will decrease the quality of care and the usual talking points against universal healthcare. Then I said "don't you agree that certain things need to be funded by the government, like roads and infrastructure?" "Yes, of course." "Me too. Personally, I<mask> think things like schools should continue to be funded in this way too. I'm pro our current 'universal education' system." "Oh yes, me too for sure." "Yes, right? Well I<mask> personally just think this should carry over to health care<mask> well for universal health care just like we have universal education." "Hm, yes, I totally agree." </s>
Label encoding: <s> [STARTQ] When dealing with racists, I tend to use the strategy outlined, just continually ask them to explain their views and watch as they get more and more moderate. [ENDQ] [NEWLINE] I once got an extreme conservative who was arguing against universal healthcare to support universal healthcare in the same conversation by using that strategy. [NEWLINE] [NEWLINE] She was saying how taxpayer money shouldn't go to other people's medical bills who can't pay for it themselves and how it will decrease the quality of care and the usual talking points against universal healthcare. Then I said "don't you agree that certain things need to be funded by the government, like roads and infrastructure?" "Yes, of course." "Me too. Personally, I also think things like schools should continue to be funded in this way too. I'm pro our current 'universal education' system." "Oh yes, me too for sure." "Yes, right? Well I also personally just think this should carry over to health care as well for universal health care just like we have universal education." "Hm, yes, I totally agree." </s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0701, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand<mask> pulling something off the wall - like a dolphin on your ankle or three stars on your arm - can convey this.<mask> I have a number of extremely well-thought out tattoos that are all rooted in my personal aesthetic and each one takes on its own meaning<mask> I grow older. [NEWLINE] [NEWLINE] Tattoos will do this naturally anyway:<mask> you<mask> a person grow, they remind you of different parts of your life,<mask> you were into,<mask> you've changed, the people you loved, etc.<mask> anything<mask><mask> getting a tattoo<mask> you like the way it looks - provided you haven't just walked into the parlour that day,<mask> actually taken the time to think about it - is a far better way of doing it than attempting to attach bullshit meaning<mask> people think you're "deep." [NEWLINE] [NEWLINE] Can't tell you<mask> much I hate programmes like LA Ink, with people going, "Oh, well I wanted to get a bee<mask> my Grandmother's best friend saved her life once, and her middle initial is B." </s>
Label encoding: <s>I understand why pulling something off the wall - like a dolphin on your ankle or three stars on your arm - can convey this. But I have a number of extremely well-thought out tattoos that are all rooted in my personal aesthetic and each one takes on its own meaning as I grow older. [NEWLINE] [NEWLINE] Tattoos will do this naturally anyway: as you as a person grow, they remind you of different parts of your life, what you were into, how you've changed, the people you loved, etc. If anything I think getting a tattoo because you like the way it looks - provided you haven't just walked into the parlour that day, but actually taken the time to think about it - is a far better way of doing it than attempting to attach bullshit meaning so people think you're "deep." [NEWLINE] [NEWLINE] Can't tell you how much I hate programmes like LA Ink, with people going, "Oh, well I wanted to get a bee because my Grandmother's best friend saved her life once, and her middle initial is B." </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know that in Latin cultures people have like 12 names. [NEWLINE] [NEWLINE] I mean it does get complicated,<mask> I don't think it's a bad thing. [NEWLINE] [NEWLINE] Actually over time the population is going to continue to grow and maybe it would be better to have a unique name with hyphens<mask> that you could be identified easier. That is<mask> you want to be. [NEWLINE] [NEWLINE] I'm pretty sure Michael Smith's have a hard time being found on Facebook,<mask> maybe<mask> they were Miachel Smith-Johnson-Rodriguez ect. it wouldn't be<mask> hard. [NEWLINE] [NEWLINE] I get<mask> you feel that there are negatives,<mask> there are positives to it<mask> well. [NEWLINE] [NEWLINE] I don't think it should be assumed that eventually you will have to choose to drop off names.<mask> you want to you can,<mask><mask> you don't you can just continue to expand. People see names differently, and for some the more unique the better.<mask><mask> we each should have our own ability to decide, and I'm fine with either.</s>
Label encoding: <s>I know that in Latin cultures people have like 12 names. [NEWLINE] [NEWLINE] I mean it does get complicated, but I don't think it's a bad thing. [NEWLINE] [NEWLINE] Actually over time the population is going to continue to grow and maybe it would be better to have a unique name with hyphens so that you could be identified easier. That is if you want to be. [NEWLINE] [NEWLINE] I'm pretty sure Michael Smith's have a hard time being found on Facebook, but maybe if they were Miachel Smith-Johnson-Rodriguez ect. it wouldn't be so hard. [NEWLINE] [NEWLINE] I get why you feel that there are negatives, but there are positives to it as well. [NEWLINE] [NEWLINE] I don't think it should be assumed that eventually you will have to choose to drop off names. If you want to you can, but if you don't you can just continue to expand. People see names differently, and for some the more unique the better. I think we each should have our own ability to decide, and I'm fine with either.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are certain mental disorders that smoking appears to help. <mask> some people with [schizophrenia]( [URL].short) and [ADHD]( [URL].pdf) tend to use cigarettes to self-medicate,<mask> it helps them focus or feel more normal. <mask><mask> some of the symptoms that nicotine can help alleviate can severely degrade a person's quality of life, I don't find it surprising that someone would be willing to put up with the long-term negative effects of smoking in order to feel better now. [NEWLINE] [NEWLINE] Now they have to actually *try* smoking first and it's not like the benefits of smoking are often touted,<mask> once someone discovers that it can help them focus<mask> nothing else they have access to does it can be very appealing (not everyone has access to doctors or prescriptions,<mask> cigarettes are easy).  For some of the symptoms of schizophrenia, cigarettes probably have *fewer* side effects than the normal prescribed drugs, plus there's no fear of being labeled mentally ill for purchasing a pack of cigarettes.  </s>
Label encoding: <s>There are certain mental disorders that smoking appears to help.  So some people with [schizophrenia]( [URL].short) and [ADHD]( [URL].pdf) tend to use cigarettes to self-medicate, because it helps them focus or feel more normal.  Given that some of the symptoms that nicotine can help alleviate can severely degrade a person's quality of life, I don't find it surprising that someone would be willing to put up with the long-term negative effects of smoking in order to feel better now. [NEWLINE] [NEWLINE] Now they have to actually *try* smoking first and it's not like the benefits of smoking are often touted, but once someone discovers that it can help them focus when nothing else they have access to does it can be very appealing (not everyone has access to doctors or prescriptions, but cigarettes are easy).  For some of the symptoms of schizophrenia, cigarettes probably have *fewer* side effects than the normal prescribed drugs, plus there's no fear of being labeled mentally ill for purchasing a pack of cigarettes.  </s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> then I thought,<mask><mask> I owned a screen print shop and Westboro Baptist Church wanted me to print up a bunch of homophobic shirts I'd tell them to get the hell out. [ENDQ] [NEWLINE] I would bet that any judge will allow you to refuse *any* hate-based messages,<mask><mask> which (otherwise protected) group they came from. And<mask> the WBC simply wanted a T-shirt with their church logo, you couldn't turn them down for ideological reasons. [NEWLINE] [NEWLINE] <mask> calculating costs vs. benefits of having minority protections,<mask><mask> that<mask> protecting some groups we don't agree with, is a price worth paying.<mask> we can stipulate that in order to qualify their messages may not be hate-based. [NEWLINE] [NEWLINE] [STARTQ] Or<mask> I was a tattoo artist and someone wanted a neo-nazi tattoo should I have to do that<mask> well? [ENDQ] [NEWLINE] No,<mask> neo-nazis are not a protected group like race, sexual orientation or religion,<mask> you could refuse to do any work for them.</s>
Label encoding: <s> [STARTQ] But then I thought, what if I owned a screen print shop and Westboro Baptist Church wanted me to print up a bunch of homophobic shirts I'd tell them to get the hell out. [ENDQ] [NEWLINE] I would bet that any judge will allow you to refuse *any* hate-based messages, regardless of which (otherwise protected) group they came from. And if the WBC simply wanted a T-shirt with their church logo, you couldn't turn them down for ideological reasons. [NEWLINE] [NEWLINE] When calculating costs vs. benefits of having minority protections, I think that also protecting some groups we don't agree with, is a price worth paying. But we can stipulate that in order to qualify their messages may not be hate-based. [NEWLINE] [NEWLINE] [STARTQ] Or if I was a tattoo artist and someone wanted a neo-nazi tattoo should I have to do that as well? [ENDQ] [NEWLINE] No, because neo-nazis are not a protected group like race, sexual orientation or religion, so you could refuse to do any work for them.</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay then,<mask> you accept this is true: [NEWLINE] [NEWLINE] [STARTQ] I don't have any illusions that anyone, rich or poor, is acting in anything<mask> their own interests [ENDQ] [NEWLINE] Then you accept the rich will not run the government in a way that benefits the poor, they will run it in a way that benefits them. We will see an increase in tax breaks for the rich and more leniency on companies. After all, they are after their own interests, and<mask> they say, the rich get richer, right? One really only has to look to the industrial revolution to see<mask> this plays out, and it isn't well for anyone else. Furthermore it is bad for the economy. [A recent study came out from the IMF]( [URL] ): [NEWLINE] [NEWLINE] &gt;“<mask> the income share of the top 20% increases, then GDP growth actually declines over the medium term, suggesting that the benefits do not trickle down.<mask><mask>, an increase in the income share of the bottom 20% is associated with higher GDP growth,” </s>
Label encoding: <s>Okay then, so you accept this is true: [NEWLINE] [NEWLINE] [STARTQ] I don't have any illusions that anyone, rich or poor, is acting in anything but their own interests [ENDQ] [NEWLINE] Then you accept the rich will not run the government in a way that benefits the poor, they will run it in a way that benefits them. We will see an increase in tax breaks for the rich and more leniency on companies. After all, they are after their own interests, and as they say, the rich get richer, right? One really only has to look to the industrial revolution to see how this plays out, and it isn't well for anyone else. Furthermore it is bad for the economy. [A recent study came out from the IMF]( [URL] ): [NEWLINE] [NEWLINE] &gt;“ If the income share of the top 20% increases, then GDP growth actually declines over the medium term, suggesting that the benefits do not trickle down. In contrast, an increase in the income share of the bottom 20% is associated with higher GDP growth,” </s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] `you are humanising animals` [ENDQ] [NEWLINE] <mask><mask><mask> he is? Animals often have human qualities, after all, humans are animals. You are not addressing his argument that animals can clearly show consent to activities that they like and deny consent to activities they don't like. [NEWLINE] [NEWLINE] [STARTQ] `have a PHD in animal behaviour` [ENDQ] [NEWLINE] You don't need a PhD, you only need basic knowledge of the animal language and mating rituals.<mask> the person doesn't have that, he/she is likely to violate one or many animal abuse laws and get punished. [NEWLINE] [NEWLINE] [STARTQ] `This isn't even to mention the list of zoonotic diseases which could be transmitted or created<mask><mask><mask> of such copulations.` [ENDQ] [NEWLINE] I let you know that sex with humans creates and transmit about 16 times more disease than<mask> you could get from non-human animals.<mask> you can use zoonosis to be against bestiality, unless you are<mask> against human+human sex<mask> it is far more dangerous than animal+human sex.</s>
Label encoding: <s> [STARTQ] `you are humanising animals` [ENDQ] [NEWLINE] So what if he is? Animals often have human qualities, after all, humans are animals. You are not addressing his argument that animals can clearly show consent to activities that they like and deny consent to activities they don't like. [NEWLINE] [NEWLINE] [STARTQ] `have a PHD in animal behaviour` [ENDQ] [NEWLINE] You don't need a PhD, you only need basic knowledge of the animal language and mating rituals. If the person doesn't have that, he/she is likely to violate one or many animal abuse laws and get punished. [NEWLINE] [NEWLINE] [STARTQ] `This isn't even to mention the list of zoonotic diseases which could be transmitted or created as a result of such copulations.` [ENDQ] [NEWLINE] I let you know that sex with humans creates and transmit about 16 times more disease than what you could get from non-human animals. So you can use zoonosis to be against bestiality, unless you are also against human+human sex because it is far more dangerous than animal+human sex.</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>7 times out of<mask><mask>? I mean that statistic to me sounds like someone saying you are seven times more likely to die in a car accident<mask> you are driving/in a car. Its a non starter. OFC someone who is aborting a child would have less chance of dieing in child birth<mask> they would never be going through it in the first place. [NEWLINE] [NEWLINE] I don't agree with calling the changes a deformation<mask> it feels like painting the person who agreed to have sex<mask> a victim of an unknown outcome (talking about adults here, children can be forgiving for not realizing the changes). [NEWLINE] [NEWLINE] And for the killing that which is not alive<mask><mask> i can be after some thinking about it convinced to change my time of<mask> i consider life from first heart beat to first signs of brain activity.<mask> thats still the 12th week. [NEWLINE] [NEWLINE] <mask> i did not know miscarriages<mask> that high. Thats a verry sad realization.<mask> i do not see<mask> the percent of miscarriage has to do with abortions. </s>
Label encoding: <s>7 times out of what though? I mean that statistic to me sounds like someone saying you are seven times more likely to die in a car accident if you are driving/in a car. Its a non starter. OFC someone who is aborting a child would have less chance of dieing in child birth since they would never be going through it in the first place. [NEWLINE] [NEWLINE] I don't agree with calling the changes a deformation since it feels like painting the person who agreed to have sex as a victim of an unknown outcome (talking about adults here, children can be forgiving for not realizing the changes). [NEWLINE] [NEWLINE] And for the killing that which is not alive i think i can be after some thinking about it convinced to change my time of what i consider life from first heart beat to first signs of brain activity. But thats still the 12th week. [NEWLINE] [NEWLINE] Also i did not know miscarriages as that high. Thats a verry sad realization. but i do not see how the percent of miscarriage has to do with abortions. </s>
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Replicating an experiment could be viewed<mask> science<mask> only up until a certain point.<mask> something is occurring for the thousandth time and the outcome is never really in question then its really more of a demonstration of a scientific phenomenon rather than true scientific inquiry. [NEWLINE] [NEWLINE] In case you try claiming this distinction is arbitrary, you can formalize this notion using Bayesian inference.<mask> the number of observations of something is low than each additional observation lends increased certainty that a true phenomenon has been observed and<mask> early replications meaningfully add to science by reducing uncertainty. After a large number of replication the effect of each additional iteration of the experiment is negligible (in terms of reducing uncertainty) and its more of a demonstration of science. [NEWLINE] [NEWLINE] Flicking on a light switch and having the lights come one illustrates a large number of scientific principles<mask> I'm not going to give you credit for testing the hypothesis that closing a circuit allows electricity to flow through tungsten filaments and electric flow through those filaments would produce light. We all already knew that would happen.</s>
Label encoding: <s>Replicating an experiment could be viewed as science but only up until a certain point. When something is occurring for the thousandth time and the outcome is never really in question then its really more of a demonstration of a scientific phenomenon rather than true scientific inquiry. [NEWLINE] [NEWLINE] In case you try claiming this distinction is arbitrary, you can formalize this notion using Bayesian inference. When the number of observations of something is low than each additional observation lends increased certainty that a true phenomenon has been observed and thus early replications meaningfully add to science by reducing uncertainty. After a large number of replication the effect of each additional iteration of the experiment is negligible (in terms of reducing uncertainty) and its more of a demonstration of science. [NEWLINE] [NEWLINE] Flicking on a light switch and having the lights come one illustrates a large number of scientific principles but I'm not going to give you credit for testing the hypothesis that closing a circuit allows electricity to flow through tungsten filaments and electric flow through those filaments would produce light. We all already knew that would happen.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's less that people explicitly state "She totally deserved it" -<mask> that does happen - and more that they imply it with their words and<mask> they do. [NEWLINE] [NEWLINE] For example, in November 2010 an eleven year old girl was sexually assaulted by eighteen older men. The New York Times article about the attack added this: *"They said she dressed older than her age, wearing makeup and fashions more appropriate to a woman in her 20s. She would hang out with teenage boys at a playground, some said."* [article about the article here]( [URL] /). [NEWLINE] [NEWLINE] And you have to understand, this isn't a one-off thing. Women who report being raped are often asked<mask> they were wearing<mask> the incident occurred. Women who dress sexy might be accused of "leading on" the assailant - and then have it turned around "Well that's<mask> happens<mask> you dress like a whore". [NEWLINE] [NEWLINE] I know I had more examples of this sort of thing,<mask> I can't seem to find the articles again.</s>
Label encoding: <s>It's less that people explicitly state "She totally deserved it" - although that does happen - and more that they imply it with their words and what they do. [NEWLINE] [NEWLINE] For example, in November 2010 an eleven year old girl was sexually assaulted by eighteen older men. The New York Times article about the attack added this: *"They said she dressed older than her age, wearing makeup and fashions more appropriate to a woman in her 20s. She would hang out with teenage boys at a playground, some said."* [article about the article here]( [URL] /). [NEWLINE] [NEWLINE] And you have to understand, this isn't a one-off thing. Women who report being raped are often asked what they were wearing when the incident occurred. Women who dress sexy might be accused of "leading on" the assailant - and then have it turned around "Well that's what happens when you dress like a whore". [NEWLINE] [NEWLINE] I know I had more examples of this sort of thing, but I can't seem to find the articles again.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Really, this can be taken a step further,<mask> a fertilized egg that fails to implant will be passed during the woman's normal period.<mask> it is determined that I took an Advil (which has the side effect of thinning the blood, and<mask> may increase the risk of miscarriage) the day before my period showed up,<mask> not having any way to know<mask> I've got a fertilized egg in there, by OP's definition I've just ended (albeit unknowingly) a human life. Should I be prosecuted for manslaughter?<mask> I have technically just committed it. To monitor for this crime, should I start submitting my used tampons to the state for testing? [NEWLINE] [NEWLINE] Medicine is not black and white. Plenty of day-to-day activities can increase the risk of miscarriage. Your pregnancy do's and don'ts will vary depending on your physician.<mask> the doctors can't agree on<mask> is safe for the baby,<mask> are women supposed to know<mask> to do to avoid being charged with a criminal act?</s>
Label encoding: <s>Really, this can be taken a step further, because a fertilized egg that fails to implant will be passed during the woman's normal period. If it is determined that I took an Advil (which has the side effect of thinning the blood, and so may increase the risk of miscarriage) the day before my period showed up, despite not having any way to know if I've got a fertilized egg in there, by OP's definition I've just ended (albeit unknowingly) a human life. Should I be prosecuted for manslaughter? Because I have technically just committed it. To monitor for this crime, should I start submitting my used tampons to the state for testing? [NEWLINE] [NEWLINE] Medicine is not black and white. Plenty of day-to-day activities can increase the risk of miscarriage. Your pregnancy do's and don'ts will vary depending on your physician. If the doctors can't agree on what is safe for the baby, how are women supposed to know what to do to avoid being charged with a criminal act?</s>
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe that in cases of genocide, WMD use (including chemical and biological weapons), or acts that cause grievous human suffering we have a moral obligation to intervene. [ENDQ] [NEWLINE] By "we" I assume you mean the U.S.A.<mask> do we (the rest of the world) had a moral obligation to intervene<mask> U.S.A. dropped two nuclear bombs on japan? [NEWLINE] [NEWLINE] <mask><mask> could a country with a high number of human rights violations be a policing agent in the world? (for a short list see here : [URL] ) [NEWLINE] [NEWLINE] Without talking about specific countries, let's ask this simple question:<mask> two states have different definitions of<mask> is a human rights violation and<mask> is moral -<mask> should all that work out? [NEWLINE] [NEWLINE] A simple artificial example: the West might think that women in Islamic countries are oppressed,<mask> they wear burkas. Middle east might think that western women are oppressed,<mask> they are over-sexualized in the media and advertising. It is a bit relative.</s>
Label encoding: <s> [STARTQ] I believe that in cases of genocide, WMD use (including chemical and biological weapons), or acts that cause grievous human suffering we have a moral obligation to intervene. [ENDQ] [NEWLINE] By "we" I assume you mean the U.S.A. So do we (the rest of the world) had a moral obligation to intervene when U.S.A. dropped two nuclear bombs on japan? [NEWLINE] [NEWLINE] Also how could a country with a high number of human rights violations be a policing agent in the world? (for a short list see here : [URL] ) [NEWLINE] [NEWLINE] Without talking about specific countries, let's ask this simple question: if two states have different definitions of what is a human rights violation and what is moral - how should all that work out? [NEWLINE] [NEWLINE] A simple artificial example: the West might think that women in Islamic countries are oppressed, because they wear burkas. Middle east might think that western women are oppressed, because they are over-sexualized in the media and advertising. It is a bit relative.</s>
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A society run by Kim Kardashian, Paris Hilton, and Donald Trump is not a society I'd want to live in. [NEWLINE] [NEWLINE] Admittedly, a society run by Bill Gates and J.K. Rowling would probably be pretty awesome,<mask> there might be some merit to that. [NEWLINE] [NEWLINE] All joking aside, most rich people aren't rich just<mask> they're smarter than poor people; it's<mask> they're more aggressive than poor people, and know enough to listen to people who are smarter than them; Vince McMahon, billionaire chairman of the WWE will be the first to admit he doesn't think he's a smart guy,<mask> he surrounds himself (or did) with smart people who give him ideas,<mask> he makes the final decision after hearing and considering all the smart ideas. His favorite phrase is "Chocolate or Vanilla?" meaning that no idea is necessarily any better or worse, just that he has to make the decision of<mask> flavor the company is trying this time. [NEWLINE] [NEWLINE] I'd imagine that many other billionaires are from the same cloth.</s>
Label encoding: <s>A society run by Kim Kardashian, Paris Hilton, and Donald Trump is not a society I'd want to live in. [NEWLINE] [NEWLINE] Admittedly, a society run by Bill Gates and J.K. Rowling would probably be pretty awesome, so there might be some merit to that. [NEWLINE] [NEWLINE] All joking aside, most rich people aren't rich just because they're smarter than poor people; it's because they're more aggressive than poor people, and know enough to listen to people who are smarter than them; Vince McMahon, billionaire chairman of the WWE will be the first to admit he doesn't think he's a smart guy, but he surrounds himself (or did) with smart people who give him ideas, but he makes the final decision after hearing and considering all the smart ideas. His favorite phrase is "Chocolate or Vanilla?" meaning that no idea is necessarily any better or worse, just that he has to make the decision of what flavor the company is trying this time. [NEWLINE] [NEWLINE] I'd imagine that many other billionaires are from the same cloth.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you are confusing the brain and the mind. Consciousness, self-awareness, thoughts, etc. are all products of our mind. Our brains establish a capacity for our mind,<mask>, like any muscle - it can be flexed.<mask> we exercise our mind we can increase the capacity. A person with a severe form of mental retardation will have a malformed brain.<mask><mask><mask>, he/she will have a much lower mental capacity. Some theories you should look into: [NEWLINE] [NEWLINE] 1. Neurolaw. This is a field in criminology that studies gray matter in the pre-frontal cortex. Some criminologists hypothesize that the amount of gray matter in the brain correlates to a propensity to commit crime. [NEWLINE] [NEWLINE] 2. Weinstein's Cyst - A mans personality began to shift and he killed his wife. It was determined that a cyst was growing on his brain and this changed his neurological activity.<mask>, we are unsure<mask> this caused his change in personality and led to his random crime.</s>
Label encoding: <s>I think you are confusing the brain and the mind. Consciousness, self-awareness, thoughts, etc. are all products of our mind. Our brains establish a capacity for our mind, however, like any muscle - it can be flexed. If we exercise our mind we can increase the capacity. A person with a severe form of mental retardation will have a malformed brain. As a result, he/she will have a much lower mental capacity. Some theories you should look into: [NEWLINE] [NEWLINE] 1. Neurolaw. This is a field in criminology that studies gray matter in the pre-frontal cortex. Some criminologists hypothesize that the amount of gray matter in the brain correlates to a propensity to commit crime. [NEWLINE] [NEWLINE] 2. Weinstein's Cyst - A mans personality began to shift and he killed his wife. It was determined that a cyst was growing on his brain and this changed his neurological activity. However, we are unsure if this caused his change in personality and led to his random crime.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand<mask> a straw purchase is,<mask><mask><mask> we are arguing this (really irrelevant) point without considering<mask> the other is trying to say. I understand that<mask><mask> the absolute letter of the law, cutpeach is incorrect and you are right. I have understood that<mask> you first replied to me.<mask> basically<mask> I am trying to say is that the situation created within straw purchasing is similar enough to lawful diverted to unlawful that it supports cutpeach's point. His point is that the ability for an individual to go out and buy a gun lawfully, has<mask> given the opportunity for gun ownership to take place unlawfully and it happens often. This is<mask> straw purchasing essentially is; someone going and buying a gun<mask> they have the legal right to, for someone who may (<mask> you said depending on that decision) or may not have the legal right to do the same. [NEWLINE] [NEWLINE] And of course the perception of an outside observer defines whether is is legal or illegal,<mask> it distinguishes which of your 3 scenarios took place.</s>
Label encoding: <s>I understand what a straw purchase is, but I think we are arguing this (really irrelevant) point without considering what the other is trying to say. I understand that according to the absolute letter of the law, cutpeach is incorrect and you are right. I have understood that since you first replied to me. But basically what I am trying to say is that the situation created within straw purchasing is similar enough to lawful diverted to unlawful that it supports cutpeach's point. His point is that the ability for an individual to go out and buy a gun lawfully, has also given the opportunity for gun ownership to take place unlawfully and it happens often. This is what straw purchasing essentially is; someone going and buying a gun because they have the legal right to, for someone who may ( as you said depending on that decision) or may not have the legal right to do the same. [NEWLINE] [NEWLINE] And of course the perception of an outside observer defines whether is is legal or illegal, because it distinguishes which of your 3 scenarios took place.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In response to the objectification, I feel like the guys get just<mask> objectified<mask> the females. Now the female objectification can be more obnoxious and everything,<mask> it seems like girls tend to get a free pass on doing the same to guys. I always hear stuff about Ronaldo and just the other day in /r/soccer there was a discussion about Alexis Sanchez not wearing shirts and<mask> that's a great thing. All the girls I know who follow soccer participate in this. Finally, David Beckham. [NEWLINE] [NEWLINE] I don't think anyone would say those 3 aren't three of the best players around (obviously Beckham is past his prime),<mask> I don't think anyone would disagree that Alex Morgan is one of the best in the women's game either.<mask><mask> attractive people are going to be recognized for their attractiveness, whether they're athletes or news anchors or Target employees. [NEWLINE] [NEWLINE] You do raise some other interesting points, I just wanted to say that I see objectification<mask> being an issue on both sides. </s>
Label encoding: <s>In response to the objectification, I feel like the guys get just as objectified as the females. Now the female objectification can be more obnoxious and everything, but it seems like girls tend to get a free pass on doing the same to guys. I always hear stuff about Ronaldo and just the other day in /r/soccer there was a discussion about Alexis Sanchez not wearing shirts and how that's a great thing. All the girls I know who follow soccer participate in this. Finally, David Beckham. [NEWLINE] [NEWLINE] I don't think anyone would say those 3 aren't three of the best players around (obviously Beckham is past his prime), but I don't think anyone would disagree that Alex Morgan is one of the best in the women's game either. I think attractive people are going to be recognized for their attractiveness, whether they're athletes or news anchors or Target employees. [NEWLINE] [NEWLINE] You do raise some other interesting points, I just wanted to say that I see objectification as being an issue on both sides. </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0921, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] By the Married Women's Property Acts a woman has complete control over all property acquired or inherited by her in any way, free from any claim on the part of her husband. With cynical injustice she is left in possession of all her old claims on her husband's property, and the latest charter of female privilege, the Statute of 1895, gives her claims regardless even of her adultery. [ENDQ] [NEWLINE] [URL] #Matrimonial_Privileges_Of_Women. [NEWLINE] [NEWLINE] This equality(and "finally women were not property") sure came handy. [NEWLINE] [NEWLINE] [STARTQ] Under the married women property act a husband has no jurisdiction over his wife's property and income. Under the income tax he is responsible for her taxes.<mask> the taxes are not paid, the husband, not the wife, is imprisoned. Mrs. Wilks refused to pay her income taxes--$185--and her husband was locked up. He will spend the rest of his life in prison unless the wife pays or the laws are changed. [ENDQ] [NEWLINE] [URL] </s>
Label encoding: <s> [STARTQ] By the Married Women's Property Acts a woman has complete control over all property acquired or inherited by her in any way, free from any claim on the part of her husband. With cynical injustice she is left in possession of all her old claims on her husband's property, and the latest charter of female privilege, the Statute of 1895, gives her claims regardless even of her adultery. [ENDQ] [NEWLINE] [URL] #Matrimonial_Privileges_Of_Women. [NEWLINE] [NEWLINE] This equality(and "finally women were not property") sure came handy. [NEWLINE] [NEWLINE] [STARTQ] Under the married women property act a husband has no jurisdiction over his wife's property and income. Under the income tax he is responsible for her taxes. If the taxes are not paid, the husband, not the wife, is imprisoned. Mrs. Wilks refused to pay her income taxes--$185--and her husband was locked up. He will spend the rest of his life in prison unless the wife pays or the laws are changed. [ENDQ] [NEWLINE] [URL] </s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think it is reasonable to accept any country<mask> world police. [NEWLINE] [NEWLINE] The nature of geopolitics makes it impossible to consistently enforce any set of rules on intervention. For example, Saudi Arabia would not be a target for the United States simply<mask> of the important resource ties,<mask> Saudi Arabia's horrible stance on women. North Korea cannot be targeted<mask> of their relationship with China. Israel cannot be targeted<mask> they are close allies of the U.S. [NEWLINE] [NEWLINE] Any state or organization with access to nuclear weapons basically becomes exempted from any possibility of intervention. This,<mask><mask><mask>, is one of the main reasons we see countries like Iran pursuing nuclear weapons: to maintain sovereignty. [NEWLINE] [NEWLINE] These two points make possible essentially only arbitrary interventions without any consistent means of application. [NEWLINE] [NEWLINE] <mask> these problems could be solved and a consistent set of rules for intervention (such<mask> violations of the Universal Declaration of Human Rights) could be created, then I would accept any country or organization that could do the job<mask> world police.</s>
Label encoding: <s>I don't think it is reasonable to accept any country as world police. [NEWLINE] [NEWLINE] The nature of geopolitics makes it impossible to consistently enforce any set of rules on intervention. For example, Saudi Arabia would not be a target for the United States simply because of the important resource ties, despite Saudi Arabia's horrible stance on women. North Korea cannot be targeted because of their relationship with China. Israel cannot be targeted because they are close allies of the U.S. [NEWLINE] [NEWLINE] Any state or organization with access to nuclear weapons basically becomes exempted from any possibility of intervention. This, in my opinion, is one of the main reasons we see countries like Iran pursuing nuclear weapons: to maintain sovereignty. [NEWLINE] [NEWLINE] These two points make possible essentially only arbitrary interventions without any consistent means of application. [NEWLINE] [NEWLINE] If these problems could be solved and a consistent set of rules for intervention (such as violations of the Universal Declaration of Human Rights) could be created, then I would accept any country or organization that could do the job as world police.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> are you able to enforce the "contract" you have with the government-provided law enforcement now? Can you make them care about your self-interests? Can you make them protect you from others?<mask> will you do<mask> they break that unearned contract with you<mask> still come to collect their payments? [ENDQ] [NEWLINE] Corruption happens,<mask> we have systems in place (laws) to dissuade corruption and to punish it<mask> it happens.<mask> corruption becomes<mask> prevalent that the populace can't stand it anymore, that's<mask> you get revolutions. [NEWLINE] [NEWLINE] Under voluntaryism, there is no protection from corruption AT ALL. In voluntaryism, the market will eventually only have one security provider,<mask> the lack of laws, regulations, and enforcements, the unfettered capitalism will make them drive all competition into the dirt. They will screw over their clients<mask> there will be nothing to stop them and no reason to stop. At least we elect a government. A corporation doesn't have to answer to voters.</s><pad>
Label encoding: <s> [STARTQ] How are you able to enforce the "contract" you have with the government-provided law enforcement now? Can you make them care about your self-interests? Can you make them protect you from others? What will you do when they break that unearned contract with you but still come to collect their payments? [ENDQ] [NEWLINE] Corruption happens, but we have systems in place (laws) to dissuade corruption and to punish it when it happens. When corruption becomes so prevalent that the populace can't stand it anymore, that's when you get revolutions. [NEWLINE] [NEWLINE] Under voluntaryism, there is no protection from corruption AT ALL. In voluntaryism, the market will eventually only have one security provider, because the lack of laws, regulations, and enforcements, the unfettered capitalism will make them drive all competition into the dirt. They will screw over their clients because there will be nothing to stop them and no reason to stop. At least we elect a government. A corporation doesn't have to answer to voters.</s><pad>
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Saying that nicotine is bad for you<mask> it effects your heart rate and blood pressure is like saying alcohol is bad for you. Nicotine acts<mask> a supplement for the naturally occurring hormones in your body, mimicking things like acetylcholine and norepinephrine.<mask> taken in small quantities say a single cigarettes worth, it's not necessarily bad for you.<mask>, you can overdo it like you could most things in life causing all sorts of things to happen to your body like increased heart rate, sweating, and nausea.<mask>, small doses of straight nicotine don't or haven't been shown to cause any sort of lasting negative effects on your body. And just FYI, those ecigs that give you a dose of nicotine through vapor are only being debated on<mask> of the likelihood that people will overdo it<mask> nicotine is readily and easily available to them with literally no strings attached.  Source: I just spent the last month learning about this in my physiology and anatomy course at the university of Colorado. </s>
Label encoding: <s>Saying that nicotine is bad for you because it effects your heart rate and blood pressure is like saying alcohol is bad for you. Nicotine acts as a supplement for the naturally occurring hormones in your body, mimicking things like acetylcholine and norepinephrine. If taken in small quantities say a single cigarettes worth, it's not necessarily bad for you. However, you can overdo it like you could most things in life causing all sorts of things to happen to your body like increased heart rate, sweating, and nausea. However, small doses of straight nicotine don't or haven't been shown to cause any sort of lasting negative effects on your body. And just FYI, those ecigs that give you a dose of nicotine through vapor are only being debated on because of the likelihood that people will overdo it when nicotine is readily and easily available to them with literally no strings attached.  Source: I just spent the last month learning about this in my physiology and anatomy course at the university of Colorado. </s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My point is that it wouldnt matter<mask> we were 200% efficient at converting solar to electricity. You would still need to keep a coal plant open for that one cloudy day. [ENDQ] <mask>, I see<mask> Germany is doing and my whole argument is that they are making a big mistake. I'm not arguing that solar and wind cant be done,<mask> that it makes no sense to do them. [NEWLINE] [NEWLINE] There are two problems with that.. first: coal plants are not that fast to start or stop.. gas plants are better suited for sudden demand changes. [NEWLINE] [NEWLINE] And second that solar and wind complement each other quite well (see that recent Frauenhofer study). [NEWLINE] [NEWLINE] <mask> lets not forget about the fact, that peak load from AC-units, etc. coincide quite well with peak solar output. [NEWLINE] [NEWLINE] Solar and wind are not the only two renewable energy sources btw.. add biogas, wood chip, etc. to the mix and you can safely get rid of the coal plants.</s>
Label encoding: <s> [STARTQ] My point is that it wouldnt matter if we were 200% efficient at converting solar to electricity. You would still need to keep a coal plant open for that one cloudy day. [ENDQ] Also, I see what Germany is doing and my whole argument is that they are making a big mistake. I'm not arguing that solar and wind cant be done, but that it makes no sense to do them. [NEWLINE] [NEWLINE] There are two problems with that.. first: coal plants are not that fast to start or stop.. gas plants are better suited for sudden demand changes. [NEWLINE] [NEWLINE] And second that solar and wind complement each other quite well (see that recent Frauenhofer study). [NEWLINE] [NEWLINE] Also lets not forget about the fact, that peak load from AC-units, etc. coincide quite well with peak solar output. [NEWLINE] [NEWLINE] Solar and wind are not the only two renewable energy sources btw.. add biogas, wood chip, etc. to the mix and you can safely get rid of the coal plants.</s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*"Violence is not merely killing another. It is violence<mask> we use a sharp word,<mask> we make a gesture to brush away a person,<mask> we obey<mask> there is fear.<mask> violence isn’t merely organized butchery in the name of God, in the name of society or country. Violence is much more subtle, much deeper, and we are inquiring into the very depths of violence.<mask> you call yourself an Indian or a Muslim or a Christian or a European, or anything else, you are being violent. Do you know<mask> it is violent?<mask> you are* **separating yourself from the rest of mankind**. *<mask> you separate yourself by belief, by nationality, by tradition, it breeds violence.<mask> a man who is seeking to understand violence does not belong to any country, to any religion, to any political party or system; he is concerned with the total understanding of mankind.* [NEWLINE] [NEWLINE] **J. Krishnamurti - Freedom from the Known**</s>
Label encoding: <s>*"Violence is not merely killing another. It is violence when we use a sharp word, when we make a gesture to brush away a person, when we obey because there is fear. So violence isn’t merely organized butchery in the name of God, in the name of society or country. Violence is much more subtle, much deeper, and we are inquiring into the very depths of violence. When you call yourself an Indian or a Muslim or a Christian or a European, or anything else, you are being violent. Do you know why it is violent? Because you are* **separating yourself from the rest of mankind**. * When you separate yourself by belief, by nationality, by tradition, it breeds violence. So a man who is seeking to understand violence does not belong to any country, to any religion, to any political party or system; he is concerned with the total understanding of mankind.* [NEWLINE] [NEWLINE] **J. Krishnamurti - Freedom from the Known**</s>
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have issues with the OP's view,<mask> this is not one of them. [NEWLINE] [NEWLINE] You're urging him to take "morality" into account.<mask> morality? Whose morality? The OP is using<mask>'s called a "utilitarian" system of morality, wherein actions are judged by their net influence on the pleasure and displeasure of all people. He's made one or two mistakes in my view,<mask> his system of morality is well established and frankly quite reasonable. [NEWLINE] [NEWLINE] <mask><mask> with the notion of "human rights," and other theories of moral objectivism. It's all just teleologic nonsense at its core; we are a particularly advanced animal species which arose from the chaos of natural processes.<mask> should we assume that there's some invisible, clearly defined code of "rights" which we all automatically deserve by virtue of our species? [NEWLINE] [NEWLINE] Morality is<mask> we make it, and<mask> constructed systems of right and wrong go, utilitarianism may be the most fair and justifiable.</s>
Label encoding: <s>I have issues with the OP's view, but this is not one of them. [NEWLINE] [NEWLINE] You're urging him to take "morality" into account. What morality? Whose morality? The OP is using what's called a "utilitarian" system of morality, wherein actions are judged by their net influence on the pleasure and displeasure of all people. He's made one or two mistakes in my view, but his system of morality is well established and frankly quite reasonable. [NEWLINE] [NEWLINE] I disagree with the notion of "human rights," and other theories of moral objectivism. It's all just teleologic nonsense at its core; we are a particularly advanced animal species which arose from the chaos of natural processes. Why should we assume that there's some invisible, clearly defined code of "rights" which we all automatically deserve by virtue of our species? [NEWLINE] [NEWLINE] Morality is what we make it, and as constructed systems of right and wrong go, utilitarianism may be the most fair and justifiable.</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I used to think like that. Some posts are really dumb<mask><mask><mask> and have no space here.<mask> guess<mask>, they don't get many upvotes. It is new and growing subreddit,<mask> sometimes you will find some bad quality posts in front page. [NEWLINE] [NEWLINE] [Bacon is disgusting. CMV]( [URL] /) [NEWLINE] 8 upvotes and 10 downvotes. Bang, problem solved. [NEWLINE] [NEWLINE] Upvotes exist for a reason. I believe that the top topics now deserve to be there. Submissions that<mask><mask> that is a wast of space don't have a score bigger than 7. Downvote it and move along. Or go even further, click on hide. [NEWLINE] [NEWLINE] <mask>,<mask> this sub get out of control (<mask> it is ok right now), the mods could enable the downvote (people downvote anyway,<mask> now it's easier to everyone bury the crap). [NEWLINE] [NEWLINE] **TL;RD: Bad posts exists,<mask> they are badly upvoted at all** [NEWLINE] </s>
Label encoding: <s>I used to think like that. Some posts are really dumb in my opinion and have no space here. But guess what, they don't get many upvotes. It is new and growing subreddit, so sometimes you will find some bad quality posts in front page. [NEWLINE] [NEWLINE] [Bacon is disgusting. CMV]( [URL] /) [NEWLINE] 8 upvotes and 10 downvotes. Bang, problem solved. [NEWLINE] [NEWLINE] Upvotes exist for a reason. I believe that the top topics now deserve to be there. Submissions that I think that is a wast of space don't have a score bigger than 7. Downvote it and move along. Or go even further, click on hide. [NEWLINE] [NEWLINE] However, if this sub get out of control ( but it is ok right now), the mods could enable the downvote (people downvote anyway, but now it's easier to everyone bury the crap). [NEWLINE] [NEWLINE] **TL;RD: Bad posts exists, but they are badly upvoted at all** [NEWLINE] </s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Those movements are completely incomparable and its a poor argument to put forth. [NEWLINE] [NEWLINE] Have a look at that first link. That is a sound and completely reasonable concept. [NEWLINE] [NEWLINE] Oh no, the texts available around the subject aren't in-line with<mask> we think is ok. Have a look back through the history of Feminism. Some of the texts that have been used to explore the concepts are abominable. This is very important. Mein Kampf is required reading in all sorts of courses. Not<mask> they want you to follow the philosophy<mask> to consider it and form your own opinion which is<mask> that text served to do for me<mask> I read it. Both groups have toxic elements and some really valuable thinking<mask> well. [NEWLINE] [NEWLINE] The hypocrisy of people is<mask> shits me up the wall. Mysandry and mysoginy are both very real and exist within all gender equality groups. Maintain the same standard for both or you don't help at all. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Those movements are completely incomparable and its a poor argument to put forth. [NEWLINE] [NEWLINE] Have a look at that first link. That is a sound and completely reasonable concept. [NEWLINE] [NEWLINE] Oh no, the texts available around the subject aren't in-line with what we think is ok. Have a look back through the history of Feminism. Some of the texts that have been used to explore the concepts are abominable. This is very important. Mein Kampf is required reading in all sorts of courses. Not because they want you to follow the philosophy but to consider it and form your own opinion which is what that text served to do for me when I read it. Both groups have toxic elements and some really valuable thinking as well. [NEWLINE] [NEWLINE] The hypocrisy of people is what shits me up the wall. Mysandry and mysoginy are both very real and exist within all gender equality groups. Maintain the same standard for both or you don't help at all. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> the best analogies for my case is the fat guy that keeps going to the gym,<mask> can't drop a pound,<mask> he always "rewards himself with a burger / fries / cake" [ENDQ] [NEWLINE] Tell me something,<mask> does he buy those things?<mask> happens to the short-order cook<mask> the fat guy can't afford a burger?<mask><mask> there's 10 fat guys that can no longer afford to eat there? [NEWLINE] [NEWLINE] Welfare money doesn't stop<mask> it gets to the recipient's pocket, *it gets spent*. The best way to get out of poverty is to get a job, right?<mask><mask> does one go about getting a job<mask> there are no customers to support employers? [NEWLINE] [NEWLINE] There is no economic argument to be made here<mask> people with no money are excluded from the economy. It doesn't matter<mask> they are spending their money on,<mask><mask><mask> they are spending it. Welfare isn't a drain on the economy, it props it up.</s>
Label encoding: <s> [STARTQ] I think the best analogies for my case is the fat guy that keeps going to the gym, but can't drop a pound, because he always "rewards himself with a burger / fries / cake" [ENDQ] [NEWLINE] Tell me something, where does he buy those things? What happens to the short-order cook when the fat guy can't afford a burger? What if there's 10 fat guys that can no longer afford to eat there? [NEWLINE] [NEWLINE] Welfare money doesn't stop when it gets to the recipient's pocket, *it gets spent*. The best way to get out of poverty is to get a job, right? But how does one go about getting a job when there are no customers to support employers? [NEWLINE] [NEWLINE] There is no economic argument to be made here because people with no money are excluded from the economy. It doesn't matter what they are spending their money on, so long as they are spending it. Welfare isn't a drain on the economy, it props it up.</s>
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Brazil has exactly that: "Voto Nulo" and "Voto em Branco". [NEWLINE] [NEWLINE] In Elections for President, people care,<mask> celebrities don't get elected there,<mask> lots of misinformed people are forced to vote - and they don't vote "Nulo", their vote ends up with a candidate that appeals to the most misinformed. [NEWLINE] [NEWLINE] In almost all elections for the legislative in Brazil, lots and lots of widely recognized candidates win, with no experience in politics, nor economics, nor anything useful actually. [NEWLINE] [NEWLINE] Compulsory voting is a bad system not only<mask> of these consequences we see in Brazil,<mask><mask><mask> it is, by principle, an anti-liberty idea... think about it, you are forcing people to participate in a system they might not agree with. They lose a whole day of work or leisure, which might be important for someone who does not care about elections -<mask> care about making money that day, or sleeping, or drinking with friends.</s>
Label encoding: <s>Brazil has exactly that: "Voto Nulo" and "Voto em Branco". [NEWLINE] [NEWLINE] In Elections for President, people care, so celebrities don't get elected there, however lots of misinformed people are forced to vote - and they don't vote "Nulo", their vote ends up with a candidate that appeals to the most misinformed. [NEWLINE] [NEWLINE] In almost all elections for the legislative in Brazil, lots and lots of widely recognized candidates win, with no experience in politics, nor economics, nor anything useful actually. [NEWLINE] [NEWLINE] Compulsory voting is a bad system not only because of these consequences we see in Brazil, but also because it is, by principle, an anti-liberty idea... think about it, you are forcing people to participate in a system they might not agree with. They lose a whole day of work or leisure, which might be important for someone who does not care about elections - but care about making money that day, or sleeping, or drinking with friends.</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Without reading any other responses and<mask> being quite depressed myself, life is an experience.<mask> you don't enjoy life, that has no bearing on the billions of other lives existing across the planet. Assuming you're a male, would you value life<mask> you had the most beautiful female in the world yearning for you? Personally, in my depressed state, I can safely say that would change my life. It may not be *permanent*,<mask> life is never about permanence. It's about retention of the highest quality of life. Mentally, I don't have any quality of life,<mask> I use the fact that I have internet to push for everyone to have better quality of life. [NEWLINE] [NEWLINE] We could definitely just all kill ourselves,<mask> we've got, whatever, billions of years or<mask> telling us to fuck some more life across the planet. Our drives secure your philosophical opinion<mask> insubstantial. Can life be meaningless and worthy of suicide<mask> we truly never reach that point of thought?</s>
Label encoding: <s>Without reading any other responses and also being quite depressed myself, life is an experience. If you don't enjoy life, that has no bearing on the billions of other lives existing across the planet. Assuming you're a male, would you value life if you had the most beautiful female in the world yearning for you? Personally, in my depressed state, I can safely say that would change my life. It may not be *permanent*, but life is never about permanence. It's about retention of the highest quality of life. Mentally, I don't have any quality of life, so I use the fact that I have internet to push for everyone to have better quality of life. [NEWLINE] [NEWLINE] We could definitely just all kill ourselves, but we've got, whatever, billions of years or so telling us to fuck some more life across the planet. Our drives secure your philosophical opinion as insubstantial. Can life be meaningless and worthy of suicide if we truly never reach that point of thought?</s>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Reading fiction can absolutely be worth your time, and there is much to get from good literature.  A good fiction piece provides an opportunity to imagine the experience of someone else - encounter experiences of characters who you may or may not have anything in common with.  Sure, you may learn some new scienc<mask> you read a chemistry text<mask> you can immerse yourself in being a mad chemist in a good fiction story.  You can read a biography about someone famous or you can read a story and connect with the underdog and learn a few new life lessons. [NEWLINE] [NEWLINE] Fiction is taught in school<mask> stories are one of the oldest traditions humans have to pass down wisdom and knowledge.  I would challenge you to read some of the great literature of modern time...Mark Twain, Kurt Vonnegut (spelling?), Dickens, toltsky, Fitzgerald, Austen.  The stories are a way to understand humanity in new ways and to learn something from someone's else's perspective.  </s>
Label encoding: <s>Reading fiction can absolutely be worth your time, and there is much to get from good literature.  A good fiction piece provides an opportunity to imagine the experience of someone else - encounter experiences of characters who you may or may not have anything in common with.  Sure, you may learn some new scienc if you read a chemistry text but you can immerse yourself in being a mad chemist in a good fiction story.  You can read a biography about someone famous or you can read a story and connect with the underdog and learn a few new life lessons. [NEWLINE] [NEWLINE] Fiction is taught in school because stories are one of the oldest traditions humans have to pass down wisdom and knowledge.  I would challenge you to read some of the great literature of modern time...Mark Twain, Kurt Vonnegut (spelling?), Dickens, toltsky, Fitzgerald, Austen.  The stories are a way to understand humanity in new ways and to learn something from someone's else's perspective.  </s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks, I appreciate it. I would like you to be able to enjoy BBT and feel for the characters like I do,<mask> it's likely that our tastes are just different. [NEWLINE] [NEWLINE] Out of curiosity, can you name some comedy shows you do enjoy?<mask> do you feel about Seinfeld? [NEWLINE] [NEWLINE] I name Seinfeld<mask> it's a highly popular show which I personally find unfunny. I'm not amused by its comedy style - at one point, I even saw Seinfeld perform live, and was unimpressed. I thought the guy who performed before him to warm up the audience was much better. Then, adding to my dislike of his comedy style, I don't empathize with, or really *like*, any of the characters on the show. Seinfeld is snooty, Kramer is crazy, George is despicable, Elaine's a bitch. [NEWLINE] [NEWLINE] That being said, the Soup Nazi episode was funny. [NEWLINE] [NEWLINE] Perhaps it's similar for you with BBT?</s>
Label encoding: <s>Thanks, I appreciate it. I would like you to be able to enjoy BBT and feel for the characters like I do, but it's likely that our tastes are just different. [NEWLINE] [NEWLINE] Out of curiosity, can you name some comedy shows you do enjoy? How do you feel about Seinfeld? [NEWLINE] [NEWLINE] I name Seinfeld because it's a highly popular show which I personally find unfunny. I'm not amused by its comedy style - at one point, I even saw Seinfeld perform live, and was unimpressed. I thought the guy who performed before him to warm up the audience was much better. Then, adding to my dislike of his comedy style, I don't empathize with, or really *like*, any of the characters on the show. Seinfeld is snooty, Kramer is crazy, George is despicable, Elaine's a bitch. [NEWLINE] [NEWLINE] That being said, the Soup Nazi episode was funny. [NEWLINE] [NEWLINE] Perhaps it's similar for you with BBT?</s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What about thin women who are<mask> skinny they're flat chested, or they don't have a butt, etc. in a society that celebrates curves -albeit on an hourglass frame-<mask><mask> lots of women can be made self conscious by this manner of being "too thin". [NEWLINE] [NEWLINE] Not to mention the arguments of eating disorders that others have already brought up. One of my friends who has spent the past year in treatment for bulimia is very sensitive to "skinny shaming". She's very self conscious of her frame now. [NEWLINE] [NEWLINE] Additional: the fact that you summarize your entire argument in the word "jelly" kinda makes me feel like you are simplifying the complex spectrum of emotions we feel,<mask> well<mask> mass body perception. It's not just a contrast of "jelly" people and skinny people. My sister was shamed for being heavier at one point in her life and shamed for being too thin at another. The world is not black and white. </s>
Label encoding: <s>What about thin women who are so skinny they're flat chested, or they don't have a butt, etc. in a society that celebrates curves -albeit on an hourglass frame- I think lots of women can be made self conscious by this manner of being "too thin". [NEWLINE] [NEWLINE] Not to mention the arguments of eating disorders that others have already brought up. One of my friends who has spent the past year in treatment for bulimia is very sensitive to "skinny shaming". She's very self conscious of her frame now. [NEWLINE] [NEWLINE] Additional: the fact that you summarize your entire argument in the word "jelly" kinda makes me feel like you are simplifying the complex spectrum of emotions we feel, as well as mass body perception. It's not just a contrast of "jelly" people and skinny people. My sister was shamed for being heavier at one point in her life and shamed for being too thin at another. The world is not black and white. </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I do think the case for religiously (<mask> opposed to ethnically) motivated violence can be made, it's a long, hard, messy statistical slog.  I might have a crack at it<mask> I get home<mask> this CMV is still going. [NEWLINE] [NEWLINE] <mask> more relevantly,<mask><mask> most atheists (myself included) take the view that it isn't religion *per se* that causes violence,<mask> a broader category of the rejection of skepticism and free inquiry.  Christopher Hitchens puts it well<mask> [NEWLINE] [NEWLINE] [STARTQ] No society has gone the way of gulags or concentration camps by following the path of Spinoza and Einstein and Jefferson and Thomas Paine. [ENDQ] [NEWLINE] It's not that religion exogenously causes violence, it's that *non serviam* is the only robust solution to violence, and the rejection of religious authority is part of that.  This is to say, atheism is certainly not a sufficient condition for peace,<mask> it is a necessary one.</s>
Label encoding: <s>While I do think the case for religiously ( as opposed to ethnically) motivated violence can be made, it's a long, hard, messy statistical slog.  I might have a crack at it when I get home if this CMV is still going. [NEWLINE] [NEWLINE] But more relevantly, I think most atheists (myself included) take the view that it isn't religion *per se* that causes violence, but a broader category of the rejection of skepticism and free inquiry.  Christopher Hitchens puts it well as [NEWLINE] [NEWLINE] [STARTQ] No society has gone the way of gulags or concentration camps by following the path of Spinoza and Einstein and Jefferson and Thomas Paine. [ENDQ] [NEWLINE] It's not that religion exogenously causes violence, it's that *non serviam* is the only robust solution to violence, and the rejection of religious authority is part of that.  This is to say, atheism is certainly not a sufficient condition for peace, but it is a necessary one.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> do you mean by cocktails?<mask> they have Manhattans on the menu, then they damn well better be able to make them.<mask><mask> they have long islands and whiskey sours, then they may not be that kind of bar. I used to live in a college town<mask> there were mixed drinks available,<mask> no one who wasn't a twat who watched too much Mad Men ordered one of your "classic cocktails." [NEWLINE] [NEWLINE] The point is, it's all very contextual to the bar. Some bars have clientele that want these sorts of drinks,<mask> they would be remiss not to make them. Other bars, most people don't know a thing about that sort of drink and the only ones that try to order them are neckbeards that think it makes them classy. There's no point in those bartenders knowing<mask> to make gimlets, cosmopolitans, and manhattans<mask> no one asks. Maybe you're just in the wrong bar.</s>
Label encoding: <s>So what do you mean by cocktails? If they have Manhattans on the menu, then they damn well better be able to make them. But if they have long islands and whiskey sours, then they may not be that kind of bar. I used to live in a college town where there were mixed drinks available, but no one who wasn't a twat who watched too much Mad Men ordered one of your "classic cocktails." [NEWLINE] [NEWLINE] The point is, it's all very contextual to the bar. Some bars have clientele that want these sorts of drinks, so they would be remiss not to make them. Other bars, most people don't know a thing about that sort of drink and the only ones that try to order them are neckbeards that think it makes them classy. There's no point in those bartenders knowing how to make gimlets, cosmopolitans, and manhattans because no one asks. Maybe you're just in the wrong bar.</s>
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> very few of them commit suicide.<mask><mask> most terminally ill people who have nothing<mask> suffering left keep on living. [ENDQ] [NEWLINE] Playing devil's advocate here slightly,<mask> your argument seems to be<mask><mask> committing suicide is easy. Have you considered that suicide rates might be low<mask> : [NEWLINE] [NEWLINE] a) It's a terrifying thing to take your own life. Some people may want to be dead,<mask> not want to *die*. Killing yourself is a messy, painful business and of course there's always a chance that you might mess it up and suffer an extreme amount of pain [NEWLINE] [NEWLINE] b) You know that doing<mask> would cause a tremendous amount of pain and suffering for your loved ones [NEWLINE] [NEWLINE] c)<mask> you live in a religious environment (particularly Catholic) you may believe that committing suicide would send you to eternal torture in hell [NEWLINE] [NEWLINE] d) They quite simply may not be able to physically commit suicide without help.<mask> could a severely handicapped person go about committing suicide? </s>
Label encoding: <s> [STARTQ] But very few of them commit suicide. In fact most terminally ill people who have nothing but suffering left keep on living. [ENDQ] [NEWLINE] Playing devil's advocate here slightly, but your argument seems to be assuming that committing suicide is easy. Have you considered that suicide rates might be low because : [NEWLINE] [NEWLINE] a) It's a terrifying thing to take your own life. Some people may want to be dead, but not want to *die*. Killing yourself is a messy, painful business and of course there's always a chance that you might mess it up and suffer an extreme amount of pain [NEWLINE] [NEWLINE] b) You know that doing so would cause a tremendous amount of pain and suffering for your loved ones [NEWLINE] [NEWLINE] c) If you live in a religious environment (particularly Catholic) you may believe that committing suicide would send you to eternal torture in hell [NEWLINE] [NEWLINE] d) They quite simply may not be able to physically commit suicide without help. How could a severely handicapped person go about committing suicide? </s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think OP actually has a point that film isn't primarily a story-telling medium the way, for example, novels are.  In a book, you are reading a story about a guy in a bar.  In a movie, you are seeing a specific guy, who looks a specific way, in a specific bar. <mask> there's an ontology to it that writing doesn't have. [NEWLINE] [NEWLINE] A good example from this year is *Boyhood.*  I would consider it a great film,<mask><mask> makes it great isn't the story it tells.  It's the ideas in it.  And Linklater communicates those ideas using characters and dialogue,<mask> I have the feeling that<mask> you read boyhood<mask> a story it wouldn't be very good at all. [NEWLINE] [NEWLINE] Now, you can use film to tell a story and it can be effective,<mask> that doesn't mean that's<mask> film is best at or that's<mask> all movies are trying to do. [NEWLINE] </s>
Label encoding: <s>I think OP actually has a point that film isn't primarily a story-telling medium the way, for example, novels are.  In a book, you are reading a story about a guy in a bar.  In a movie, you are seeing a specific guy, who looks a specific way, in a specific bar.  So there's an ontology to it that writing doesn't have. [NEWLINE] [NEWLINE] A good example from this year is *Boyhood.*  I would consider it a great film, but what makes it great isn't the story it tells.  It's the ideas in it.  And Linklater communicates those ideas using characters and dialogue, but I have the feeling that if you read boyhood as a story it wouldn't be very good at all. [NEWLINE] [NEWLINE] Now, you can use film to tell a story and it can be effective, but that doesn't mean that's what film is best at or that's what all movies are trying to do. [NEWLINE] </s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you think out great-great-great-great grandchildren will be happy<mask> the only remaining animal species are cats, dogs, cows, horses, pigs, chicken, goats, sheep, rabbits, ants and cockroaches? [NEWLINE] [NEWLINE] Relevant xkcd: [URL] / [NEWLINE] [NEWLINE] Losing a species is *very* hard to reverse, and we're already losing species at a terrible rate. It's quite likely that our far future descendants will wish we saved more,<mask> we should do<mask><mask> it's still possible. Especially those that are especially cute or interesting or unique (again,<mask> our descendants will<mask> value them for those qualities). [NEWLINE] [NEWLINE] I wish the dodo was still around,<mask> well<mask> the American megafauna, and even the dinosaurs - they're *interesting*. They give examples of<mask> different life can be, of all the ways an ecosystem can work. [NEWLINE] [NEWLINE] (For similar reasons, I mourn the [loss of many ancient texts]( [URL] ))</s><pad><pad>
Label encoding: <s>Do you think out great-great-great-great grandchildren will be happy if the only remaining animal species are cats, dogs, cows, horses, pigs, chicken, goats, sheep, rabbits, ants and cockroaches? [NEWLINE] [NEWLINE] Relevant xkcd: [URL] / [NEWLINE] [NEWLINE] Losing a species is *very* hard to reverse, and we're already losing species at a terrible rate. It's quite likely that our far future descendants will wish we saved more, so we should do so while it's still possible. Especially those that are especially cute or interesting or unique (again, because our descendants will also value them for those qualities). [NEWLINE] [NEWLINE] I wish the dodo was still around, as well as the American megafauna, and even the dinosaurs - they're *interesting*. They give examples of how different life can be, of all the ways an ecosystem can work. [NEWLINE] [NEWLINE] (For similar reasons, I mourn the [loss of many ancient texts]( [URL] ))</s><pad><pad>
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Anyone with<mask>'s considered an unusual or less desirable way of living is going to get treated negatively by society. Not driving a Lamborghini is not consider to be unusual in society. [NEWLINE] [NEWLINE] <mask> we are supposed to want to have kids. It's something expected by parents, who ask<mask> they don't have grandchildren. It's something expected by different levels of government, who give parents some benefits not afforded to non-parents. [NEWLINE] [NEWLINE] Other examples of people that are considered "unusual" and who might need their own support groups: atheists, people who want 10+ kids, asexual people, transgender people, and vegans. Someone belonging to one of those  groups might experience people looking down on them,<mask> they may want to find other people like them<mask> they feel they are alone. For example, I'm transgender, and don't know many other trans people<mask> I live.<mask> being active on the trans subreddit makes me feel like I'm less alone. </s>
Label encoding: <s>Anyone with what's considered an unusual or less desirable way of living is going to get treated negatively by society. Not driving a Lamborghini is not consider to be unusual in society. [NEWLINE] [NEWLINE] But we are supposed to want to have kids. It's something expected by parents, who ask why they don't have grandchildren. It's something expected by different levels of government, who give parents some benefits not afforded to non-parents. [NEWLINE] [NEWLINE] Other examples of people that are considered "unusual" and who might need their own support groups: atheists, people who want 10+ kids, asexual people, transgender people, and vegans. Someone belonging to one of those  groups might experience people looking down on them, so they may want to find other people like them if they feel they are alone. For example, I'm transgender, and don't know many other trans people where I live. So being active on the trans subreddit makes me feel like I'm less alone. </s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So,<mask> I can sum up the thread. [NEWLINE] [NEWLINE] There are several "valid" reasons for not peeing in the shower: [NEWLINE] [NEWLINE] #1)  Pee destroys pipes (false) [NEWLINE] #2)  Most showers don't drain ever (false) [NEWLINE] #3)  Pee is full of bacteria (false) [NEWLINE] #4)  5 yr olds will prefer to pee into the shower which is not running rather than into the toilet (possibly true... I guess) [NEWLINE] #5) <mask> someone pees in the shower, they are likely to then poo in the shower<mask> both of those actions are identical in function and result (false). [NEWLINE] #6)  Pee is highly toxic and opens a portal to a Lovercraft domain in flying brain worms can enter your bathroom (true?) [NEWLINE] [NEWLINE] [NEWLINE] Then<mask><mask><mask><mask> we have: [NEWLINE] - The first hand experience of every single person who has ever peed<mask> standing under running warm water</s>
Label encoding: <s>So, if I can sum up the thread. [NEWLINE] [NEWLINE] There are several "valid" reasons for not peeing in the shower: [NEWLINE] [NEWLINE] #1)  Pee destroys pipes (false) [NEWLINE] #2)  Most showers don't drain ever (false) [NEWLINE] #3)  Pee is full of bacteria (false) [NEWLINE] #4)  5 yr olds will prefer to pee into the shower which is not running rather than into the toilet (possibly true... I guess) [NEWLINE] #5)  If someone pees in the shower, they are likely to then poo in the shower because both of those actions are identical in function and result (false). [NEWLINE] #6)  Pee is highly toxic and opens a portal to a Lovercraft domain in flying brain worms can enter your bathroom (true?) [NEWLINE] [NEWLINE] [NEWLINE] Then on the other hand we have: [NEWLINE] - The first hand experience of every single person who has ever peed when standing under running warm water</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] right with the law running into burning buildings. Firefighters like cops act<mask> agents of the state and it makes sense to demand a level of quality [ENDQ] [NEWLINE] Are you asserting that someone has to be licensed to provide aid in an emergency? [NEWLINE] [NEWLINE] I don't violate any laws whether I run into a burning building, provide aid to someone in a car accident<mask> waiting for EMT's to arrive, or even detain someone who is in the process of committing a felony... [NEWLINE] [NEWLINE] We don't have laws preventing "normal people" from helping their communities in times of need...  we don't have laws like that for a reason...<mask> our individual desire to help our communities is a net positive and frankly shouldn't be regulated. [NEWLINE] [NEWLINE] <mask> we screw up<mask> helping?  Sure we're liable...<mask> to be prohibited by law from helping in the first place<mask> people think the state should have a monopoly on helping our fellow citizens in a time of need is silly.  </s>
Label encoding: <s> [STARTQ] right with the law running into burning buildings. Firefighters like cops act as agents of the state and it makes sense to demand a level of quality [ENDQ] [NEWLINE] Are you asserting that someone has to be licensed to provide aid in an emergency? [NEWLINE] [NEWLINE] I don't violate any laws whether I run into a burning building, provide aid to someone in a car accident while waiting for EMT's to arrive, or even detain someone who is in the process of committing a felony... [NEWLINE] [NEWLINE] We don't have laws preventing "normal people" from helping their communities in times of need...  we don't have laws like that for a reason... because our individual desire to help our communities is a net positive and frankly shouldn't be regulated. [NEWLINE] [NEWLINE] If we screw up while helping?  Sure we're liable... but to be prohibited by law from helping in the first place because people think the state should have a monopoly on helping our fellow citizens in a time of need is silly.  </s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for enlightening me about<mask> kind of person I am. [NEWLINE] [NEWLINE] Rosemary is not marijuana and<mask> the high school kids may experience a mental placebo high, you are not selling them the product they paid for<mask> you're a swindler. [NEWLINE] [NEWLINE] A restaurant has a certain supply of meat, it's all still the same kind of meat,<mask> yes some cuts are ofcourse better than others. The menu says 'prime rib'. You are still getting a prime rib<mask> you order it well done. Nowhere on the menu does it say: 'prime rib of this and this quality'. Nobody is making fraudulent claims. You do not notice the difference, the other customer does<mask> it's quite sensible to present you with the less favorable cut instead of either: throwing it away, or: serving it to a customer who *does* notice<mask> they ordered medium-rare. [NEWLINE] [NEWLINE] I don't see fraud here and I don't see a problem.</s>
Label encoding: <s>Thank you for enlightening me about what kind of person I am. [NEWLINE] [NEWLINE] Rosemary is not marijuana and although the high school kids may experience a mental placebo high, you are not selling them the product they paid for so you're a swindler. [NEWLINE] [NEWLINE] A restaurant has a certain supply of meat, it's all still the same kind of meat, but yes some cuts are ofcourse better than others. The menu says 'prime rib'. You are still getting a prime rib if you order it well done. Nowhere on the menu does it say: 'prime rib of this and this quality'. Nobody is making fraudulent claims. You do not notice the difference, the other customer does so it's quite sensible to present you with the less favorable cut instead of either: throwing it away, or: serving it to a customer who *does* notice because they ordered medium-rare. [NEWLINE] [NEWLINE] I don't see fraud here and I don't see a problem.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Obviously<mask> I become a parent,<mask> other methods are equally effective or more<mask> then I would happily forgo smacking. I'm just basing its occasional necessity on my own childhood,<mask> something my stubbornness and commitment to being difficult meant that it was used<mask> a last resort. [NEWLINE] [NEWLINE] It really depends on the child, and a good parent should be able to differentiate. For example my sister was only rarely smacked to my knowledge -<mask> I may not remember correctly -<mask> she was much less difficult, whereas I was smacked occasionally. [NEWLINE] [NEWLINE] <mask><mask> I would<mask> say I certainly was a happy child,<mask> not necessarily well behaved,<mask> through the same parenting system my sister was very well behaved and<mask> happy, again making me think it depends on the kid. Again,<mask> only anecdotal, I<mask> know many people who were very badly behaved kids who weren't smacking, and weren't happy children,<mask> I<mask> know many who were model children.</s>
Label encoding: <s>Obviously when I become a parent, if other methods are equally effective or more so then I would happily forgo smacking. I'm just basing its occasional necessity on my own childhood, where something my stubbornness and commitment to being difficult meant that it was used as a last resort. [NEWLINE] [NEWLINE] It really depends on the child, and a good parent should be able to differentiate. For example my sister was only rarely smacked to my knowledge - though I may not remember correctly - because she was much less difficult, whereas I was smacked occasionally. [NEWLINE] [NEWLINE] In addition I would also say I certainly was a happy child, though not necessarily well behaved, though through the same parenting system my sister was very well behaved and also happy, again making me think it depends on the kid. Again, while only anecdotal, I also know many people who were very badly behaved kids who weren't smacking, and weren't happy children, while I also know many who were model children.</s>
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.2179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The problem is<mask> someone assumes that the ideal has already been reached<mask> it has not,<mask> then they are blind to actual racism and might end up committing racism. [ENDQ] [STARTQ] <mask> are people who believe the world is post-racial or point to factual discrepancies between race are not wholly ignorant.<mask> they are ignorant of in the former case is that the world is not<mask> free of racism, and in the latter case that factual discrepancies that are correlated with race are not caused by race. [ENDQ] [NEWLINE] I still sort of disagree with this. <mask><mask> that more than being ignorant of discrepancies correlated with race, they (often willfully) ignorant of the discrepancies which ARE caused by race, that is to say, institutional racism. [NEWLINE] [NEWLINE] [STARTQ] Some ability for rational thought is better than no ability for rational thought upon which to fight racist cognitive structures. [ENDQ] [NEWLINE] ∆ I'll buy that. <mask> do I do the thingie? [NEWLINE] [NEWLINE] EDIT: Did the thingie!</s><pad>
Label encoding: <s> [STARTQ] The problem is when someone assumes that the ideal has already been reached when it has not, because then they are blind to actual racism and might end up committing racism. [ENDQ] [STARTQ] So are people who believe the world is post-racial or point to factual discrepancies between race are not wholly ignorant. What they are ignorant of in the former case is that the world is not yet free of racism, and in the latter case that factual discrepancies that are correlated with race are not caused by race. [ENDQ] [NEWLINE] I still sort of disagree with this.  I think that more than being ignorant of discrepancies correlated with race, they (often willfully) ignorant of the discrepancies which ARE caused by race, that is to say, institutional racism. [NEWLINE] [NEWLINE] [STARTQ] Some ability for rational thought is better than no ability for rational thought upon which to fight racist cognitive structures. [ENDQ] [NEWLINE] ∆ I'll buy that.  How do I do the thingie? [NEWLINE] [NEWLINE] EDIT: Did the thingie!</s><pad>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Black people in the US have been historically denied knowledge of their heritage prior to America, and many have been here for longer than 5 generations.<mask><mask> we accept this criteria, it seems like the term African American is a bit strange. And<mask><mask> your criteria makes some amount of sense. [NEWLINE] [NEWLINE] This is part of<mask> I personally think that African American is a confusing and strange terminology. It blurs the line between black people who have been in this country for ages, and recent immigrants. The implication seems to be that black people are perpetually foreigners,<mask><mask> black people have contributed<mask> much or more to America<mask> any other group of people. [NEWLINE] [NEWLINE] I am a bit divided about it<mask>,<mask> there are studies showing [weird attitude differences based on whether someone is described<mask> "black" or "African American."]( [URL] /)<mask> a black person,<mask><mask> black is the more logical description.<mask> I'm not willing to give up a job over it.</s>
Label encoding: <s>Black people in the US have been historically denied knowledge of their heritage prior to America, and many have been here for longer than 5 generations. So if we accept this criteria, it seems like the term African American is a bit strange. And I think your criteria makes some amount of sense. [NEWLINE] [NEWLINE] This is part of why I personally think that African American is a confusing and strange terminology. It blurs the line between black people who have been in this country for ages, and recent immigrants. The implication seems to be that black people are perpetually foreigners, even though black people have contributed as much or more to America as any other group of people. [NEWLINE] [NEWLINE] I am a bit divided about it though, because there are studies showing [weird attitude differences based on whether someone is described as "black" or "African American."]( [URL] /) As a black person, I think black is the more logical description. But I'm not willing to give up a job over it.</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Even<mask> the fetus' brain was fully formed, sentient, and with decades of human experiences already imprinted, the rights of the mother still win out. [ENDQ] [NEWLINE] Ok, let's assume that the fetus inside the womb is a person, with which be both strongly disagree. [NEWLINE] [NEWLINE] <mask> the mother was for instance a rape victim, I would agree. I don't think we would force someone to donate their blood / organs / bodily fluids, even<mask> it is clearly the only way to save a person's life. We simply have no right to impose such a burden. [NEWLINE] [NEWLINE] <mask> the mother can be held responsible for the pregnancy, I don't know<mask> you could justify abortion rights in this hypothetical.<mask> you are responsible for the creation of a person, you should be held responsible for the person's well-being. You should not be able to do whatever you want with the person, simply<mask> you chose to create that person inside your own body.</s>
Label encoding: <s> [STARTQ] Even if the fetus' brain was fully formed, sentient, and with decades of human experiences already imprinted, the rights of the mother still win out. [ENDQ] [NEWLINE] Ok, let's assume that the fetus inside the womb is a person, with which be both strongly disagree. [NEWLINE] [NEWLINE] If the mother was for instance a rape victim, I would agree. I don't think we would force someone to donate their blood / organs / bodily fluids, even if it is clearly the only way to save a person's life. We simply have no right to impose such a burden. [NEWLINE] [NEWLINE] If the mother can be held responsible for the pregnancy, I don't know how you could justify abortion rights in this hypothetical. If you are responsible for the creation of a person, you should be held responsible for the person's well-being. You should not be able to do whatever you want with the person, simply because you chose to create that person inside your own body.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I suggest you buy some low quality meat and some really high quality meat.  Cook a well done version of both and a medium rare version of both.  Try them WITHOUT any sauces or gravy.  The difference between low quality well done and high quality well done will be much harder to taste than that of medium rare. <mask> medium rare really shines with high quality restaurants / meat.  I was much like you until I actually tried a piece of medium rare steak from a fancy restaurant- I was blown away by the flavor and richness and angry I've been missing out on this for<mask> long! [NEWLINE] [NEWLINE] Cooking a steak well done removes a lot of the steak's natural juices and flavor and the meat ends up tougher.  You may think you've had good well done steaks,<mask> that's<mask> you're used to well done steaks.  A really nice piece done medium rare will be almost like butter it melts in your mouth.  </s><pad>
Label encoding: <s>I suggest you buy some low quality meat and some really high quality meat.  Cook a well done version of both and a medium rare version of both.  Try them WITHOUT any sauces or gravy.  The difference between low quality well done and high quality well done will be much harder to taste than that of medium rare.  Also medium rare really shines with high quality restaurants / meat.  I was much like you until I actually tried a piece of medium rare steak from a fancy restaurant- I was blown away by the flavor and richness and angry I've been missing out on this for so long! [NEWLINE] [NEWLINE] Cooking a steak well done removes a lot of the steak's natural juices and flavor and the meat ends up tougher.  You may think you've had good well done steaks, but that's because you're used to well done steaks.  A really nice piece done medium rare will be almost like butter it melts in your mouth.  </s><pad>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Keep in mind that it would vary from anarchist to anarchist,<mask><mask><mask> is that the community would decide the rules, this is evidenced in the Russian Civil War<mask> peasants organized their towns to defend themselves against both the Whites and Reds. [NEWLINE] And no they won't all agree to live by the same rules and customs, societies would vary, and those who don't want anarchy would be free to organize a government<mask><mask><mask> they don't impose it on people who don't want a government. [NEWLINE] [NEWLINE] <mask> people prefer their new rulers, then<mask><mask><mask>, let them be,<mask> they are ALWAYS in a position to overthrow the new leader,<mask> he stopped handing out food,<mask> being the first to start blackmailing them, then the people can organize against them. The worse scenario<mask>, can easily be defeated<mask> evidenced by the Color and Singing Revolutions. [NEWLINE] [NEWLINE] And<mask><mask>, look at the Free Territory of Ukraine with 7 million. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Keep in mind that it would vary from anarchist to anarchist, but my opinion is that the community would decide the rules, this is evidenced in the Russian Civil War where peasants organized their towns to defend themselves against both the Whites and Reds. [NEWLINE] And no they won't all agree to live by the same rules and customs, societies would vary, and those who don't want anarchy would be free to organize a government as long as they don't impose it on people who don't want a government. [NEWLINE] [NEWLINE] If people prefer their new rulers, then in my opinion, let them be, but they are ALWAYS in a position to overthrow the new leader, if he stopped handing out food, thus being the first to start blackmailing them, then the people can organize against them. The worse scenario however, can easily be defeated as evidenced by the Color and Singing Revolutions. [NEWLINE] [NEWLINE] And lastly, look at the Free Territory of Ukraine with 7 million. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you do some searching, you will find cases of women who were "brain dead" for many months<mask> were still capable of giving birth to a healthy child. Dead people don't give birth. The definition of death has changed over time. The current brain death concept comes from about the 1950's<mask> I recall properly. There are some good arguments for it and some against it<mask> I feel that caution is called for<mask> in doubt, which I am. [NEWLINE] [NEWLINE] There are more certain definitions of death --<mask> none of them permit the taking of a still-beating heart from the body of a person still breathing, still metabolizing, and carrying on many bodily functions, which brain death permits. [NEWLINE] [NEWLINE] <mask> I cannot hold that brain death is certainly true death and<mask> I could donate organs that are able to be given from a living human being or from a cold corpse<mask> not organs like the heart which require the person to be questionably still living.</s>
Label encoding: <s>If you do some searching, you will find cases of women who were "brain dead" for many months but were still capable of giving birth to a healthy child. Dead people don't give birth. The definition of death has changed over time. The current brain death concept comes from about the 1950's if I recall properly. There are some good arguments for it and some against it but I feel that caution is called for when in doubt, which I am. [NEWLINE] [NEWLINE] There are more certain definitions of death -- but none of them permit the taking of a still-beating heart from the body of a person still breathing, still metabolizing, and carrying on many bodily functions, which brain death permits. [NEWLINE] [NEWLINE] Therefore I cannot hold that brain death is certainly true death and thus I could donate organs that are able to be given from a living human being or from a cold corpse but not organs like the heart which require the person to be questionably still living.</s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I do agree to an extent.<mask><mask><mask> a massive mistake a lot of people make is to not think deeply enough about the time period in which the painting was released<mask> well<mask> looking at<mask> the artist contemporaries were doing at the time. [NEWLINE] [NEWLINE] The first example you provided is a work by Piet Mondrian circa 1930ish. [NEWLINE] <mask> now we are very used to seeing such bold lines and flat colours such<mask> those often used in advertising, the web and print, Work like this was a massive juxtaposition (important wanky art word) to the more realistic styles like impressionism and post-impressionism that came before. [NEWLINE] [NEWLINE] Sometimes ground breaking art can be<mask> simple<mask> looking at<mask> everyone else is doing and then doing the opposite. You don't have to like it<mask> you have to respect some of it for challenging the norms, it is just sometimes you have to look a little deeper to understand<mask> the norms<mask> at the time.</s>
Label encoding: <s>I do agree to an extent. However I think a massive mistake a lot of people make is to not think deeply enough about the time period in which the painting was released as well as looking at what the artist contemporaries were doing at the time. [NEWLINE] [NEWLINE] The first example you provided is a work by Piet Mondrian circa 1930ish. [NEWLINE] While now we are very used to seeing such bold lines and flat colours such as those often used in advertising, the web and print, Work like this was a massive juxtaposition (important wanky art word) to the more realistic styles like impressionism and post-impressionism that came before. [NEWLINE] [NEWLINE] Sometimes ground breaking art can be as simple as looking at what everyone else is doing and then doing the opposite. You don't have to like it but you have to respect some of it for challenging the norms, it is just sometimes you have to look a little deeper to understand what the norms where at the time.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me rephrase<mask> I was trying to say<mask> it helps<mask> I feel I did not make it concise<mask> possible :) [NEWLINE] [NEWLINE] A)<mask> you can learn to program BASIC, I feel you can learn Python.  Both can be vary easy. [NEWLINE] [NEWLINE] B)<mask> you are going to learn either BASIC or Python, I feel Python is preferable<mask> you can build upon that easy initial foundation to create rather complex programs. [NEWLINE] [NEWLINE] C)<mask> a person has vary limited exposure to computers, I feel a graphical based system can be a good introduction to the core concepts of functional programming.  I am referring to people such<mask> children or your technology illiterate parents or grandparents. [NEWLINE] [NEWLINE] <mask> you are interested and know a bit about python, these sample programs can help you understand the basics of the language: [NEWLINE] [URL] [NEWLINE] <mask><mask> you will agree they are more similar to modern programming syntax,<mask> are<mask> easy to understand like BASIC. </s>
Label encoding: <s>Let me rephrase what I was trying to say if it helps as I feel I did not make it concise as possible :) [NEWLINE] [NEWLINE] A) If you can learn to program BASIC, I feel you can learn Python.  Both can be vary easy. [NEWLINE] [NEWLINE] B) If you are going to learn either BASIC or Python, I feel Python is preferable as you can build upon that easy initial foundation to create rather complex programs. [NEWLINE] [NEWLINE] C) If a person has vary limited exposure to computers, I feel a graphical based system can be a good introduction to the core concepts of functional programming.  I am referring to people such as children or your technology illiterate parents or grandparents. [NEWLINE] [NEWLINE] If you are interested and know a bit about python, these sample programs can help you understand the basics of the language: [NEWLINE] [URL] [NEWLINE] I think you will agree they are more similar to modern programming syntax, but are also easy to understand like BASIC. </s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you familiar with the concept of the [Uncanny Valley?]( [URL].svg) Essentially, something that appears close to human,<mask> slightly off in some way will appear<mask> increasingly unnerving to people. ([See here for a demonstration.]( [URL] )) [NEWLINE] [NEWLINE] Now, think about clowns with this in mind. They look generally human,<mask> there are things that are just...not quite right, by the standards of a normal person. They can have false eyebrow markings over there real ones, which makes the clown's true emotive state difficult to read. They have a big, garish smile drawn over their real mouth, which again may or may not correspond with their true expression. [NEWLINE] [NEWLINE] Clowns by definition try to display an exaggerated sense of comical happiness,<mask> this can come off<mask> unnerving<mask> it is something artificial, and not necessarily demonstrative of the emotional state of the human being under the makeup. </s>
Label encoding: <s>Are you familiar with the concept of the [Uncanny Valley?]( [URL].svg) Essentially, something that appears close to human, but slightly off in some way will appear as increasingly unnerving to people. ([See here for a demonstration.]( [URL] )) [NEWLINE] [NEWLINE] Now, think about clowns with this in mind. They look generally human, but there are things that are just...not quite right, by the standards of a normal person. They can have false eyebrow markings over there real ones, which makes the clown's true emotive state difficult to read. They have a big, garish smile drawn over their real mouth, which again may or may not correspond with their true expression. [NEWLINE] [NEWLINE] Clowns by definition try to display an exaggerated sense of comical happiness, but this can come off as unnerving because it is something artificial, and not necessarily demonstrative of the emotional state of the human being under the makeup. </s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your source says nothing about execution, just that they were holding him unlawfully. This is<mask> an isolated (<mask><mask> I know) incident involving one man whereas your original post claims Lincoln executed multiple civilians. Not saying your source is not credible<mask> you did embellish a bit. [NEWLINE] [NEWLINE] <mask><mask><mask> suspending habeas corpus, the Supreme Court ruled twice that Lincoln was in the wrong.<mask> think of this:<mask> there was a US citizen giving aid to enemies of the country during a time of war would you want him free? In terms of the law, Lincoln was in the wrong. In terms of doing<mask> was necessary to stop (read: win) the war, he did<mask> he needed to. [NEWLINE] [NEWLINE] <mask> Merryman did could<mask> be considered a form of treason, justifying his execution<mask> there was one and it just wasn't mentioned in your source. (I highly doubt a fact that significant would be left out,<mask>.) </s><pad>
Label encoding: <s>Your source says nothing about execution, just that they were holding him unlawfully. This is also an isolated ( insofar as I know) incident involving one man whereas your original post claims Lincoln executed multiple civilians. Not saying your source is not credible but you did embellish a bit. [NEWLINE] [NEWLINE] As far as suspending habeas corpus, the Supreme Court ruled twice that Lincoln was in the wrong. But think of this: if there was a US citizen giving aid to enemies of the country during a time of war would you want him free? In terms of the law, Lincoln was in the wrong. In terms of doing what was necessary to stop (read: win) the war, he did what he needed to. [NEWLINE] [NEWLINE] What Merryman did could also be considered a form of treason, justifying his execution if there was one and it just wasn't mentioned in your source. (I highly doubt a fact that significant would be left out, though.) </s><pad>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>On the whole<mask><mask>,<mask><mask><mask> your statement is too broad. There are plenty of covers done by popular bands who have changed enough of the source material to warrant a new look. (Pennywise redid "Stand by Me" - virtually everything that Me First and The Gimme Gimmes have played) Even not in the punk genre there's plenty of others. Stevie Ray Vaughan reinterpreted Hendrix, Jeff Buckley's rendition of "Hallelujah" etc.<mask> their versions arguably stand out from the original. [NEWLINE] [NEWLINE] <mask><mask> that<mask> you're talking pop songs you're right,<mask><mask> you're talking *musicianship* you might be off. Taking other peoples songs and reworking them in your own unique style means that you ought to be given credit for that. (I'll add<mask>, that<mask> I hear one more person say that John Mayer wrote Bold<mask> Love I'll go on a shooting spree)</s>
Label encoding: <s>On the whole I agree, but I think your statement is too broad. There are plenty of covers done by popular bands who have changed enough of the source material to warrant a new look. (Pennywise redid "Stand by Me" - virtually everything that Me First and The Gimme Gimmes have played) Even not in the punk genre there's plenty of others. Stevie Ray Vaughan reinterpreted Hendrix, Jeff Buckley's rendition of "Hallelujah" etc. where their versions arguably stand out from the original. [NEWLINE] [NEWLINE] I think that when you're talking pop songs you're right, but when you're talking *musicianship* you might be off. Taking other peoples songs and reworking them in your own unique style means that you ought to be given credit for that. (I'll add though, that if I hear one more person say that John Mayer wrote Bold as Love I'll go on a shooting spree)</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Here is the problem with this. You're imposing a human concept of justice onto God [ENDQ] [NEWLINE] <mask> can we even know then, that God is just,<mask> we can't use any human standards to evaluate him (good or bad)? [NEWLINE] [NEWLINE] This seems just an appeal to force, i.e. he has the power,<mask> he gets to do whatever he wants? [NEWLINE] [NEWLINE] [STARTQ] God does not do things<mask> they are just. Things are just<mask> God does them. [ENDQ] [NEWLINE] You seem to assume an answer to the [Euthypro dilemma]( [URL].php?title=Euthyphro): "Is that which is good commanded by God<mask> it's good, or is it good<mask> God commands it?", albeit slightly modified to apply to justice.<mask> things are just simply by virtue of God doing them, then he could simply perform any arbitrary (seemingly reprehensible) action, and simply declare it just.</s>
Label encoding: <s> [STARTQ] Here is the problem with this. You're imposing a human concept of justice onto God [ENDQ] [NEWLINE] How can we even know then, that God is just, if we can't use any human standards to evaluate him (good or bad)? [NEWLINE] [NEWLINE] This seems just an appeal to force, i.e. he has the power, so he gets to do whatever he wants? [NEWLINE] [NEWLINE] [STARTQ] God does not do things because they are just. Things are just because God does them. [ENDQ] [NEWLINE] You seem to assume an answer to the [Euthypro dilemma]( [URL].php?title=Euthyphro): "Is that which is good commanded by God because it's good, or is it good because God commands it?", albeit slightly modified to apply to justice. If things are just simply by virtue of God doing them, then he could simply perform any arbitrary (seemingly reprehensible) action, and simply declare it just.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, and the problem is that the only thing you can do is not vote for that party at the next election.<mask> that doesn't change anything<mask> all the parties act like this. [NEWLINE] [NEWLINE] That makes me really glad to live in Switzerland<mask> I have the opportunity to vote not only for people who I have to hope that they represent<mask><mask><mask><mask> have a say in the decision of actual matters. [NEWLINE] [NEWLINE] [STARTQ] You can't force people to care,<mask><mask><mask> having to vote makes me pay more attention to local politics than I might otherwise. [ENDQ] [NEWLINE] This is interesting. Don't you think you'd be equally involved with politics<mask> it weren't compulsory just<mask> you feel it's your obligation<mask> a cititzen to form an opinion?<mask> it's definitely like that for me and it leads me to invest a couple of hours to inform myself prior to every vote even<mask> I'm not really interested in or affected by the topic.</s>
Label encoding: <s>Yes, and the problem is that the only thing you can do is not vote for that party at the next election. But that doesn't change anything when all the parties act like this. [NEWLINE] [NEWLINE] That makes me really glad to live in Switzerland where I have the opportunity to vote not only for people who I have to hope that they represent my opinion but also have a say in the decision of actual matters. [NEWLINE] [NEWLINE] [STARTQ] You can't force people to care, but I think having to vote makes me pay more attention to local politics than I might otherwise. [ENDQ] [NEWLINE] This is interesting. Don't you think you'd be equally involved with politics if it weren't compulsory just because you feel it's your obligation as a cititzen to form an opinion? Because it's definitely like that for me and it leads me to invest a couple of hours to inform myself prior to every vote even if I'm not really interested in or affected by the topic.</s>
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To say "just semantics" is a person's way to say the bickering is pointless. Imagine two kids arguing "You broke my toy." "No, I just cracked it." To one kid the crack is an important flaw that wasn't there before<mask> to the other kid it's still all in one piece<mask><mask>'s the big deal. Now in one case of a brand new favorite toy you might easily side with one kid<mask><mask><mask> was old and beat up and already had multiple cracks? [NEWLINE] [NEWLINE] Depending on the situation<mask> someone tells you it is semantics you probably need to shift the subject slightly to get them to care about something they don't or vice versa. [NEWLINE] [NEWLINE] Tell me<mask> you have an example<mask> replacing "semantics" with "I don't care" doesn't seem accurate to you. Obviously it's annoying<mask> someone dismisses something you care about<mask> it's nothing to pull your hair out over! </s>
Label encoding: <s>To say "just semantics" is a person's way to say the bickering is pointless. Imagine two kids arguing "You broke my toy." "No, I just cracked it." To one kid the crack is an important flaw that wasn't there before but to the other kid it's still all in one piece so what's the big deal. Now in one case of a brand new favorite toy you might easily side with one kid but what if was old and beat up and already had multiple cracks? [NEWLINE] [NEWLINE] Depending on the situation if someone tells you it is semantics you probably need to shift the subject slightly to get them to care about something they don't or vice versa. [NEWLINE] [NEWLINE] Tell me if you have an example where replacing "semantics" with "I don't care" doesn't seem accurate to you. Obviously it's annoying when someone dismisses something you care about but it's nothing to pull your hair out over! </s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well this is going off on a bit of a tangent,<mask> I don't worry about candidates getting locked up<mask> a particularly onerous threat. Politicians are, almost by definition, people who will say whatever is popular that they think will get them elected. Politicians should and do reflect the popular will. Locking up one politician merely creates a vacuum into which the next opportunist will step.<mask><mask>, locking up a politician who does<mask><mask> represent the popular will would almost certainly bolster the popularity of whatever the contentious views might be.<mask><mask><mask> people retain their right to vote, then not voting is a sign that people are generally happy with the outcomes of elections.<mask> times were to take a significant turn for the worse? Then voting turnout would spike dramatically. It is the right to vote and the opportunity to vote that must remain intact<mask> the ability to correct whatever dangerous course we might get sent down is going to remain intact.</s>
Label encoding: <s>Well this is going off on a bit of a tangent, but I don't worry about candidates getting locked up as a particularly onerous threat. Politicians are, almost by definition, people who will say whatever is popular that they think will get them elected. Politicians should and do reflect the popular will. Locking up one politician merely creates a vacuum into which the next opportunist will step. In fact, locking up a politician who does in fact represent the popular will would almost certainly bolster the popularity of whatever the contentious views might be. As long as people retain their right to vote, then not voting is a sign that people are generally happy with the outcomes of elections. If times were to take a significant turn for the worse? Then voting turnout would spike dramatically. It is the right to vote and the opportunity to vote that must remain intact if the ability to correct whatever dangerous course we might get sent down is going to remain intact.</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm no art aficionado by any stretch,<mask><mask><mask> the idea of exploring the history of the artist and<mask> the piece was made is important for just about any work.  It is often, at least for me, difficult to understand the significance of a piece I might see in a museum without knowing of it's context. [NEWLINE] [NEWLINE] For instance the first painting linked by OP seems simple, boring, and easily reproduced.  Part of the reason it seems<mask>, for lack of a better word, "obvious" is<mask> it's style was part of a minimalist movement that influenced art and architecture for decades afterward.  It doesn't seem special<mask> that style has been deeply incorporated into popular fashion.  Without paintings like that the style may have never evolved, having a ripple effect in our surroundings today.  That's<mask> you'll find it hanging in a museum,<mask> you would never know it by just looking at it.</s>
Label encoding: <s>I'm no art aficionado by any stretch, but I think the idea of exploring the history of the artist and how the piece was made is important for just about any work.  It is often, at least for me, difficult to understand the significance of a piece I might see in a museum without knowing of it's context. [NEWLINE] [NEWLINE] For instance the first painting linked by OP seems simple, boring, and easily reproduced.  Part of the reason it seems so, for lack of a better word, "obvious" is because it's style was part of a minimalist movement that influenced art and architecture for decades afterward.  It doesn't seem special because that style has been deeply incorporated into popular fashion.  Without paintings like that the style may have never evolved, having a ripple effect in our surroundings today.  That's why you'll find it hanging in a museum, but you would never know it by just looking at it.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your first sentence is missing at least one word, and your overall view seems very disjointed.<mask>, your third sentence refers to the actions of the company, not the founder. Which one is it? Or did Kellogg somehow do this through the company? [NEWLINE] [NEWLINE] <mask> your issue is with Kellogg's actions, or<mask> Kellogg acted through the company,<mask> should we punish the Kellogg Company? A corporation is an abstract entity, incapable of possessing any form of intent. Unless support for (insert act/view here) is written directly into its bylaws or whatever, there's literally no connection between the current company and the founder's views. Hell, neither the CEO nor the chairman of the board (<mask><mask><mask> I can tell) is related to Kellogg in any way, and none of the major institutional stockholders are affiliated with the man either (again,<mask><mask><mask> I can tell from a quick glance).</s>
Label encoding: <s>Your first sentence is missing at least one word, and your overall view seems very disjointed. Also, your third sentence refers to the actions of the company, not the founder. Which one is it? Or did Kellogg somehow do this through the company? [NEWLINE] [NEWLINE] If your issue is with Kellogg's actions, or how Kellogg acted through the company, why should we punish the Kellogg Company? A corporation is an abstract entity, incapable of possessing any form of intent. Unless support for (insert act/view here) is written directly into its bylaws or whatever, there's literally no connection between the current company and the founder's views. Hell, neither the CEO nor the chairman of the board ( as far as I can tell) is related to Kellogg in any way, and none of the major institutional stockholders are affiliated with the man either (again, as far as I can tell from a quick glance).</s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I'm more of a utilitarian<mask> it comes to animal ethics,<mask> here's a basic argument for rights. [NEWLINE] [NEWLINE] <mask> a being has the capacity for certain experiences (pain, pleasure, awareness) then that being has a right to have those experiences taken into account. The being qualifies for basic rights which protect it from the actions of moral agents who wish to harm it. [NEWLINE] [NEWLINE] Animals are capable of experiencing harms, and are desirous not have harm inflicted on them (either through being confined in a tiny stall or having pleasurable experiences [e.g. eating, living] cut short by slaughter). [NEWLINE] [NEWLINE] Again, those who argue for animal rights have various systems of thought. Tom Regan says animals who show the capacity for certain virtues (awareness, ability to experience pain, a regard for their welfare) qualify<mask> a Subject Of A Life and<mask> are afforded some sort of Kantian inherent value and subsequent rights.</s><pad>
Label encoding: <s>Well I'm more of a utilitarian when it comes to animal ethics, but here's a basic argument for rights. [NEWLINE] [NEWLINE] If a being has the capacity for certain experiences (pain, pleasure, awareness) then that being has a right to have those experiences taken into account. The being qualifies for basic rights which protect it from the actions of moral agents who wish to harm it. [NEWLINE] [NEWLINE] Animals are capable of experiencing harms, and are desirous not have harm inflicted on them (either through being confined in a tiny stall or having pleasurable experiences [e.g. eating, living] cut short by slaughter). [NEWLINE] [NEWLINE] Again, those who argue for animal rights have various systems of thought. Tom Regan says animals who show the capacity for certain virtues (awareness, ability to experience pain, a regard for their welfare) qualify as a Subject Of A Life and thus are afforded some sort of Kantian inherent value and subsequent rights.</s><pad>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are plenty of people who will end relationships based on one's ability/preference to have or not have children. /r/childfree has countless stories about Childfree people of both/all genders who were in relationships or even married to people who thought they were on the fence or that not wanting children was a "phase they were going through" who have ended their relationships<mask> of that. [NEWLINE] [NEWLINE] Likewise, there are plenty of people who will not enter a relationship with someone who does(n't) want children<mask> wanting/not wanting children is a pretty big deal. [NEWLINE] [NEWLINE] <mask><mask> that there are some cases<mask> people will pursue other options<mask> their partner cannot biologically have children for one reason or the other and there are a lot of things we can do with technology now that we couldn't before.<mask><mask> no one would end a relationship over one of the most life changing choices you can make is pretty narrow minded. </s><pad><pad><pad>
Label encoding: <s>There are plenty of people who will end relationships based on one's ability/preference to have or not have children. /r/childfree has countless stories about Childfree people of both/all genders who were in relationships or even married to people who thought they were on the fence or that not wanting children was a "phase they were going through" who have ended their relationships because of that. [NEWLINE] [NEWLINE] Likewise, there are plenty of people who will not enter a relationship with someone who does(n't) want children because wanting/not wanting children is a pretty big deal. [NEWLINE] [NEWLINE] I agree that there are some cases where people will pursue other options if their partner cannot biologically have children for one reason or the other and there are a lot of things we can do with technology now that we couldn't before. Assuming that no one would end a relationship over one of the most life changing choices you can make is pretty narrow minded. </s><pad><pad><pad>
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes,<mask><mask> there is a strong link.  People can have various motives for getting married, such<mask> the desire to raise a family, the desire for increased financial security, the convenience of being able to share a house, or even such things<mask> making your parents happy or becoming more respectable to your neighbors,<mask> people always prefer to marry someone to whom they are sexually attracted,<mask> at all possible.  Is true that you do not have to get married to have sex,<mask> marriage is a means of reserving a given person<mask> your permanent sexual partner.  And<mask> someone is tremendously valued<mask> a sexual partner, then that is something you would want to do,<mask> you could.  Lust creates a feeling of greed or possessiveness. <mask> the sexual attraction is strong enough, then you do not just want to have sex, you want to possess that desirable partner. <mask>, you will want to marry.</s>
Label encoding: <s>Yes, I think there is a strong link.  People can have various motives for getting married, such as the desire to raise a family, the desire for increased financial security, the convenience of being able to share a house, or even such things as making your parents happy or becoming more respectable to your neighbors, but people always prefer to marry someone to whom they are sexually attracted, if at all possible.  Is true that you do not have to get married to have sex, but marriage is a means of reserving a given person as your permanent sexual partner.  And if someone is tremendously valued as a sexual partner, then that is something you would want to do, if you could.  Lust creates a feeling of greed or possessiveness.  If the sexual attraction is strong enough, then you do not just want to have sex, you want to possess that desirable partner.  Hence, you will want to marry.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0726, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> is it then? [ENDQ] [NEWLINE] <mask> you don't know<mask> it is<mask> you know you don't agree with it? <mask> don't you treat feminism<mask> a topic and pretend that you have to write a paper about it and do some actual research and find out for yourself.  I'm tired of this conversation.  Go [here]( [URL] /)<mask> you want.<mask> I suspect you won't. [NEWLINE] [NEWLINE] [STARTQ] Actually, it kind of does.<mask> can any rational person claim that the prejudice against men having hair that doesn't conform to one of a handful of hair styles is misogyny<mask> it makes them look more like women (even<mask> that hair style isn't one that women wear), and concurrently defend the social demand that men spend time every day making their faces look less like that of an adult male, and more like that of a female? [ENDQ] [NEWLINE] Can you re-phrase that like i'm 5?</s>
Label encoding: <s> [STARTQ] What is it then? [ENDQ] [NEWLINE] So you don't know what it is but you know you don't agree with it?  Why don't you treat feminism as a topic and pretend that you have to write a paper about it and do some actual research and find out for yourself.  I'm tired of this conversation.  Go [here]( [URL] /) if you want. but I suspect you won't. [NEWLINE] [NEWLINE] [STARTQ] Actually, it kind of does. How can any rational person claim that the prejudice against men having hair that doesn't conform to one of a handful of hair styles is misogyny because it makes them look more like women (even when that hair style isn't one that women wear), and concurrently defend the social demand that men spend time every day making their faces look less like that of an adult male, and more like that of a female? [ENDQ] [NEWLINE] Can you re-phrase that like i'm 5?</s>
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I just finished watching House, MD last night and<mask><mask> it's a great example against your view, and I'm sure there are others similar.  It's a good premise, and the character of house MD is conceptualize pretty well<mask> mediocre writing, exemplified by the fact that every character other then house chokes on the same lines every other episode.  Hugh Laurie is a wonder to watch<mask><mask> his character like all the others basically plays out the same three situations for the entire series.  All the other actors are fine<mask> you recognize there tropes<mask> sounding rehashed every time they say them(you push people away, you like puzzles, you don't want to face xyz). Most every moment Hugh Laurie is on screen with house feels fresh, even<mask> it isn't, and finally Id say he takes these larger than life characterizations and makes them feel real, even<mask> they blatantly aren't.</s>
Label encoding: <s>I just finished watching House, MD last night and I think it's a great example against your view, and I'm sure there are others similar.  It's a good premise, and the character of house MD is conceptualize pretty well but mediocre writing, exemplified by the fact that every character other then house chokes on the same lines every other episode.  Hugh Laurie is a wonder to watch even though his character like all the others basically plays out the same three situations for the entire series.  All the other actors are fine but you recognize there tropes as sounding rehashed every time they say them(you push people away, you like puzzles, you don't want to face xyz). Most every moment Hugh Laurie is on screen with house feels fresh, even when it isn't, and finally Id say he takes these larger than life characterizations and makes them feel real, even when they blatantly aren't.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to agree that for the most part rape culture<mask> used by most feminists I hear from is completely overblown, and for the most part does not exist. [NEWLINE] [NEWLINE] That does not mean<mask> that rape culture does not exist. It absolutely does, and it is found in prisons, and even to a lesser extent with regards to men in general. I have heard the exact same people for years who would bemoan the terrible rape culture in our society then go on to say they hope some criminal gets raped in prison.<mask><mask> anyone being raped is reprehensible, no matter who that person is. The overwhelming majority of people in this country agree it is reprehensible<mask><mask><mask> you are talking about a female, or to a lesser degree free males.<mask> it comes to men in prison most people do not give a fuck, and even encourage it in many cases. That is the very definition of a rape culture.</s>
Label encoding: <s>I have to agree that for the most part rape culture as used by most feminists I hear from is completely overblown, and for the most part does not exist. [NEWLINE] [NEWLINE] That does not mean however that rape culture does not exist. It absolutely does, and it is found in prisons, and even to a lesser extent with regards to men in general. I have heard the exact same people for years who would bemoan the terrible rape culture in our society then go on to say they hope some criminal gets raped in prison. I think anyone being raped is reprehensible, no matter who that person is. The overwhelming majority of people in this country agree it is reprehensible as long as you are talking about a female, or to a lesser degree free males. When it comes to men in prison most people do not give a fuck, and even encourage it in many cases. That is the very definition of a rape culture.</s>
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Regarding Einstein being religious, yes, childhood indoctrination and one's environment is very powerful. That's kind of the problem. [NEWLINE] [NEWLINE] Your scenario fails to define a higher power. And that lack of definition is<mask> bothers me about religious belief. People use God<mask> a catch all for knowledge we don't have, and this includes using God<mask> a<mask>. It's possible to find a biological and scientific reason<mask> to<mask> we exist. We don't know<mask> the universe happened<mask>,<mask> we might one day. My argument is that it's better not to defend people who just stay with the catch all instead of trying to promote a deeper introspection. [NEWLINE] [NEWLINE] Our material composition doesn't exclude the possibility of a higher power?<mask> does that mean? Do you mean aliens seeded us?<mask><mask><mask><mask>, there's plenty of evidence for abiogenesis and evolution with no higher power fitting in that picture at all.</s>
Label encoding: <s>Regarding Einstein being religious, yes, childhood indoctrination and one's environment is very powerful. That's kind of the problem. [NEWLINE] [NEWLINE] Your scenario fails to define a higher power. And that lack of definition is what bothers me about religious belief. People use God as a catch all for knowledge we don't have, and this includes using God as a why. It's possible to find a biological and scientific reason as to why we exist. We don't know why the universe happened yet, but we might one day. My argument is that it's better not to defend people who just stay with the catch all instead of trying to promote a deeper introspection. [NEWLINE] [NEWLINE] Our material composition doesn't exclude the possibility of a higher power? What does that mean? Do you mean aliens seeded us? On the other hand, there's plenty of evidence for abiogenesis and evolution with no higher power fitting in that picture at all.</s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Murder is actually the unlawful taking of a human being's life. There are many cases<mask> a deliberate homicide is not murder,<mask> is called self-defense. [ENDQ] [NEWLINE] This is correct.  Let me re-phrase:<mask> someone should initiate force against me and take my life without my consent, that is nothing<mask> not murder.  By refusing to pay taxes, I would not have initiated force against others. [NEWLINE] [NEWLINE] I would like to clarify<mask>,<mask> I asked in my comment: *Do I own my life only in virtue of the rest of society granting it to me?* [NEWLINE] [NEWLINE] <mask><mask>, then this means that<mask> the will of the majority dictates that all Jews be exterminated, then I<mask> a Jew must submit my life,<mask> I have no ownership of it. [NEWLINE] [NEWLINE] <mask> not, then there must be some essential difference between me owning my life, and me owning my labour.</s>
Label encoding: <s> [STARTQ] Murder is actually the unlawful taking of a human being's life. There are many cases where a deliberate homicide is not murder, but is called self-defense. [ENDQ] [NEWLINE] This is correct.  Let me re-phrase: if someone should initiate force against me and take my life without my consent, that is nothing if not murder.  By refusing to pay taxes, I would not have initiated force against others. [NEWLINE] [NEWLINE] I would like to clarify though, as I asked in my comment: *Do I own my life only in virtue of the rest of society granting it to me?* [NEWLINE] [NEWLINE] If so, then this means that if the will of the majority dictates that all Jews be exterminated, then I as a Jew must submit my life, since I have no ownership of it. [NEWLINE] [NEWLINE] If not, then there must be some essential difference between me owning my life, and me owning my labour.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not about being ashamed, it's about not wanting to open yourself up to harassment or being obviously ogled. It's a very uncomfortable feeling and it happens enough to me, and I'm sure plenty other women, in just regular situations like being fully clothed in a grocery store and I'd just rather not. [NEWLINE] [NEWLINE] And yes, there are certain types of women I would feel uncomfortable undressing around<mask> I have only met like<mask>, two maybe in my entire life. And with those two women,<mask> I told them don't speak to me like that, they listened and apologized. I say that to men and most times it causes more problems and I'm the bitch in the situation. [NEWLINE] [NEWLINE] Edit: I would<mask> be uncomfortable with this<mask> it were and always has been the norm. I'd probably never use a gym shower, dressing room, or use a bathroom that wasn't single stall.</s>
Label encoding: <s>It's not about being ashamed, it's about not wanting to open yourself up to harassment or being obviously ogled. It's a very uncomfortable feeling and it happens enough to me, and I'm sure plenty other women, in just regular situations like being fully clothed in a grocery store and I'd just rather not. [NEWLINE] [NEWLINE] And yes, there are certain types of women I would feel uncomfortable undressing around but I have only met like what, two maybe in my entire life. And with those two women, when I told them don't speak to me like that, they listened and apologized. I say that to men and most times it causes more problems and I'm the bitch in the situation. [NEWLINE] [NEWLINE] Edit: I would also be uncomfortable with this if it were and always has been the norm. I'd probably never use a gym shower, dressing room, or use a bathroom that wasn't single stall.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Should we not be allowed to disagree with this licensing board's decision about the vaccine? And, I really hate to say it,<mask> aren't we giving the licensing board an awful lot of power? There is money involved in the vaccines, and surely the companies that produce them would benefit from their vaccine being pushed into the market, no? I realize I'm starting to sound like a conspiracy theorist,<mask> I'm really just trying to see whether it makes sense for a concerned citizen to be able to refuse a vaccination. [NEWLINE] [NEWLINE] I guess the problem with the peanut analogy is that peanuts can be tried. You can have one peanut, or a part of a peanut, before you shove a whole handful of them into your mouth. Even<mask> you are allergic, you probably won't die<mask> you can try them out. This isn't the case with vaccinations; sometimes it will be too late once the vaccine is administered. </s>
Label encoding: <s>Should we not be allowed to disagree with this licensing board's decision about the vaccine? And, I really hate to say it, but aren't we giving the licensing board an awful lot of power? There is money involved in the vaccines, and surely the companies that produce them would benefit from their vaccine being pushed into the market, no? I realize I'm starting to sound like a conspiracy theorist, but I'm really just trying to see whether it makes sense for a concerned citizen to be able to refuse a vaccination. [NEWLINE] [NEWLINE] I guess the problem with the peanut analogy is that peanuts can be tried. You can have one peanut, or a part of a peanut, before you shove a whole handful of them into your mouth. Even if you are allergic, you probably won't die because you can try them out. This isn't the case with vaccinations; sometimes it will be too late once the vaccine is administered. </s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And<mask> we aren't. [NEWLINE] [NEWLINE] Monogamy is pair bonding arising from oxytocin,  vasopressin, and some other fun chemicals.  Chemicals  we can find plenty of research around.   Monogamy is not going against biology, nor necessarily is polygamy.  Depending on oxytocin distribution in the individual you'll find their preferences to differ. [NEWLINE] [NEWLINE] Typically I do believe we're a monogamous species.  Too much exists in our biological social structure (not cultural)  that suggests monogamy is the status quo.   Thing is this differs between genders<mask> well.  It goes a lot deeper than a single chemical and reproduction, <mask> those chemicals pay a huge role in many things.  There is a ton of studies it there,  particularly with voles that shows variant oxytocin distribution standards in polygamous and monogamous voles.  </s>
Label encoding: <s>And yet we aren't. [NEWLINE] [NEWLINE] Monogamy is pair bonding arising from oxytocin,  vasopressin, and some other fun chemicals.  Chemicals  we can find plenty of research around.   Monogamy is not going against biology, nor necessarily is polygamy.  Depending on oxytocin distribution in the individual you'll find their preferences to differ. [NEWLINE] [NEWLINE] Typically I do believe we're a monogamous species.  Too much exists in our biological social structure (not cultural)  that suggests monogamy is the status quo.   Thing is this differs between genders as well.  It goes a lot deeper than a single chemical and reproduction,  but those chemicals pay a huge role in many things.  There is a ton of studies it there,  particularly with voles that shows variant oxytocin distribution standards in polygamous and monogamous voles.  </s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to say, its a little low not giving a "guest" option. [NEWLINE] [NEWLINE] I am gay, living with my gf, neither of us are loud out there people. (<mask> your issue with "putting it in my face" is another topic). [NEWLINE] [NEWLINE] Wedding invitation from her cousin comes around and she only gets an invitation for her... no spot for "guest" or "+1." [NEWLINE] [NEWLINE] We were pretty upset by it. I mean, we are adults. Any other adult gets a guest option on their invitation. [NEWLINE] [NEWLINE] Well come to find out, it wasnt just us. They did this to a lot of people. [NEWLINE] [NEWLINE] You do<mask> you want.<mask> its pretty rude<mask><mask>. We were not the only people upset by this, and there was a bit of drama around it at the wedding (<mask><mask> my gf). </s><pad><pad>
Label encoding: <s>I have to say, its a little low not giving a "guest" option. [NEWLINE] [NEWLINE] I am gay, living with my gf, neither of us are loud out there people. ( although your issue with "putting it in my face" is another topic). [NEWLINE] [NEWLINE] Wedding invitation from her cousin comes around and she only gets an invitation for her... no spot for "guest" or "+1." [NEWLINE] [NEWLINE] We were pretty upset by it. I mean, we are adults. Any other adult gets a guest option on their invitation. [NEWLINE] [NEWLINE] Well come to find out, it wasnt just us. They did this to a lot of people. [NEWLINE] [NEWLINE] You do what you want. But its pretty rude IMO. We were not the only people upset by this, and there was a bit of drama around it at the wedding ( according to my gf). </s><pad><pad>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your presumption is entirely false and discredits human capacity to think long term. [ENDQ] [NEWLINE] Well, I somewhat exaggerated the claim. It's not that people will never pay for public goods in the free market. It's that fewer public goods will be paid for than is socially optimal. There is tons of literature on this in economics, both theoretical and empirical.<mask> all of society operated on the NPR model, we'd all be worse off. [NEWLINE] [NEWLINE] And then there are common goods like the environment, which<mask> /u/fludru points out, is subject to the tragedy of the commons. Suppose I can make $100 profit by pumping pollution into the air, and I regard the harm to my health from breathing the polluted air<mask> only costing $1.<mask><mask> 1000 people make the same calculation, we'll all suffer $1000 worth of harm from the pollution, and everyone ends up worse off. </s>
Label encoding: <s> [STARTQ] Your presumption is entirely false and discredits human capacity to think long term. [ENDQ] [NEWLINE] Well, I somewhat exaggerated the claim. It's not that people will never pay for public goods in the free market. It's that fewer public goods will be paid for than is socially optimal. There is tons of literature on this in economics, both theoretical and empirical. If all of society operated on the NPR model, we'd all be worse off. [NEWLINE] [NEWLINE] And then there are common goods like the environment, which as /u/fludru points out, is subject to the tragedy of the commons. Suppose I can make $100 profit by pumping pollution into the air, and I regard the harm to my health from breathing the polluted air as only costing $1. But if 1000 people make the same calculation, we'll all suffer $1000 worth of harm from the pollution, and everyone ends up worse off. </s>
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you are defining greed<mask> any sort of behaviour that isn't completely altruistic,<mask><mask> it's you who is being disingenuous [ENDQ] [NEWLINE] <mask><mask> you're going to completely ignore the established meaning of words<mask> far<mask> to give them their opposite value, then we can't even have a discussion. [NEWLINE] [NEWLINE] <mask> I rape someone<mask> I want to have sex<mask> they don't is that altruistic?<mask> I'm a trader who embezzels millions from a pension fund, is that altruistic? [NEWLINE] [NEWLINE] It is an invalid argument to discredit any study concerning humans or animals. [NEWLINE] [NEWLINE] [STARTQ] There are examples of stateless societies in which greed is not encouraged and rewarded by the system people live in. The vast majority are prehistoric cultures. [ENDQ] [NEWLINE] Okay, then that still falls under the spectrum of "any study about humans",<mask> you're aguing with OP's hypothetical, not my argument.</s>
Label encoding: <s> [STARTQ] If you are defining greed as any sort of behaviour that isn't completely altruistic, I think it's you who is being disingenuous [ENDQ] [NEWLINE] But if you're going to completely ignore the established meaning of words so far as to give them their opposite value, then we can't even have a discussion. [NEWLINE] [NEWLINE] If I rape someone because I want to have sex but they don't is that altruistic? If I'm a trader who embezzels millions from a pension fund, is that altruistic? [NEWLINE] [NEWLINE] It is an invalid argument to discredit any study concerning humans or animals. [NEWLINE] [NEWLINE] [STARTQ] There are examples of stateless societies in which greed is not encouraged and rewarded by the system people live in. The vast majority are prehistoric cultures. [ENDQ] [NEWLINE] Okay, then that still falls under the spectrum of "any study about humans", so you're aguing with OP's hypothetical, not my argument.</s>
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*Most animals that we eat would be extinct or lesser in numbers by huge factors, purely based on numbers, there are more and most live longer (<mask> slaughtered at maturity). Pigs for instance, big blobs of walking food, they can't survive in nature,<mask> they do exist<mask> of humans and we take care of them, until we eat them. [NEWLINE] [NEWLINE] *I don't consider natural things to be either moral or immoral. Humans are part of nature, most are pro saving endangered species for instance. Animals kill and eat other animals in the wilderness all the time. Just<mask> we could survive on plants and seeds doesn't mean it's most optimal. Sports would be fairly more difficult without being able to eat meat for instance. [NEWLINE] [NEWLINE] <mask> next time you go buy a nice pork chop, think like this. His entire species wouldn't exist<mask> you wouldn't be there buying it. Hope this helps.</s>
Label encoding: <s>*Most animals that we eat would be extinct or lesser in numbers by huge factors, purely based on numbers, there are more and most live longer ( when slaughtered at maturity). Pigs for instance, big blobs of walking food, they can't survive in nature, but they do exist because of humans and we take care of them, until we eat them. [NEWLINE] [NEWLINE] *I don't consider natural things to be either moral or immoral. Humans are part of nature, most are pro saving endangered species for instance. Animals kill and eat other animals in the wilderness all the time. Just because we could survive on plants and seeds doesn't mean it's most optimal. Sports would be fairly more difficult without being able to eat meat for instance. [NEWLINE] [NEWLINE] So next time you go buy a nice pork chop, think like this. His entire species wouldn't exist if you wouldn't be there buying it. Hope this helps.</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Quoting OP: "To address the gross factor, I know there are plenty of people grossed out by this idea, and thats fine.<mask> you don't want to pee in the shower then don't. Nobody is going to force you or anything." [NEWLINE] [NEWLINE] This suggests that you too believe that the idea need not be adopted on a mainstream basis due to grossness. [NEWLINE] [NEWLINE] <mask> you allow for that fact that even some people may be grossed out by shower peeing, it is safe to NOT encourage your kids to pee there. Most social norms are based on this principle of not grossing out others or putting our interests before theirs. [NEWLINE] [NEWLINE] For instance you may even prefer to pee in your coffee cup and then wash it thoroughly.<mask> it is safe to assume that a good number of people will not like that idea.<mask> expecting it to become a social norm is a bit too much.</s>
Label encoding: <s>Quoting OP: "To address the gross factor, I know there are plenty of people grossed out by this idea, and thats fine. If you don't want to pee in the shower then don't. Nobody is going to force you or anything." [NEWLINE] [NEWLINE] This suggests that you too believe that the idea need not be adopted on a mainstream basis due to grossness. [NEWLINE] [NEWLINE] If you allow for that fact that even some people may be grossed out by shower peeing, it is safe to NOT encourage your kids to pee there. Most social norms are based on this principle of not grossing out others or putting our interests before theirs. [NEWLINE] [NEWLINE] For instance you may even prefer to pee in your coffee cup and then wash it thoroughly. But it is safe to assume that a good number of people will not like that idea. Hence expecting it to become a social norm is a bit too much.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] African Americans only exist in modern times<mask> we brought them across the ocean<mask> slaves. [ENDQ] [NEWLINE] Relevance? [NEWLINE] [NEWLINE] [STARTQ] Most livestock are<mask> domesticated<mask> to be dependent on us for their survival. [ENDQ] [NEWLINE] I probably phrased that badly, I didn't mean just a dependence for survival.  Sort of a dependence on one another<mask> a species in a social way. [NEWLINE] [NEWLINE] [STARTQ] <mask>?<mask> makes it worse? It may be more emotionally difficult to kill and eat something you love,<mask><mask> should it be "worse"? [ENDQ] [NEWLINE] Imagine in the future we create two kinds of robots, one that integrates into our society with the intelligence of a young child, lives in our homes, becomes a part of our families, and one that basically does mindless menial labor in factories.  We have to destroy one and upgrade it into a new unit.  Which is worse? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] African Americans only exist in modern times because we brought them across the ocean as slaves. [ENDQ] [NEWLINE] Relevance? [NEWLINE] [NEWLINE] [STARTQ] Most livestock are so domesticated as to be dependent on us for their survival. [ENDQ] [NEWLINE] I probably phrased that badly, I didn't mean just a dependence for survival.  Sort of a dependence on one another as a species in a social way. [NEWLINE] [NEWLINE] [STARTQ] Why? What makes it worse? It may be more emotionally difficult to kill and eat something you love, but why should it be "worse"? [ENDQ] [NEWLINE] Imagine in the future we create two kinds of robots, one that integrates into our society with the intelligence of a young child, lives in our homes, becomes a part of our families, and one that basically does mindless menial labor in factories.  We have to destroy one and upgrade it into a new unit.  Which is worse? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, actually.<mask> someone claims that the Easter Bunny doesn't exist and it is a serious scientific conversation, then they have to prove that claim. The same with Santa Claus. This isn't hard either. We know<mask> the Easter Bunny was created and we know the<mask> the story of Santa Claus was created.<mask> the person trying to prove these negative claims has an easy time. [NEWLINE] [NEWLINE] I doubt your last claim would make it into a serious conversation<mask>.<mask><mask> it did, you would still have to prove it. There is not requirement to engage in conversation with someone. I don't engage with people who claim that the USA hasn't landed on the moon.<mask> someone made a claim like  "there is not a man in a big red suit who every year mindcontrols parents all over the world to get their children presents" I woudln't engage them to begin with. </s>
Label encoding: <s>Yes, actually. If someone claims that the Easter Bunny doesn't exist and it is a serious scientific conversation, then they have to prove that claim. The same with Santa Claus. This isn't hard either. We know where the Easter Bunny was created and we know the where the story of Santa Claus was created. So the person trying to prove these negative claims has an easy time. [NEWLINE] [NEWLINE] I doubt your last claim would make it into a serious conversation though. But if it did, you would still have to prove it. There is not requirement to engage in conversation with someone. I don't engage with people who claim that the USA hasn't landed on the moon. If someone made a claim like  "there is not a man in a big red suit who every year mindcontrols parents all over the world to get their children presents" I woudln't engage them to begin with. </s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know,<mask><mask> your reasons could (and will,<mask><mask> ) work against you. <mask> you're going out of your way to remain ignorant about things, you're setting yourself up to have ill-informed ideas about them. [NEWLINE] [NEWLINE] You don't live in a vacuum; you *will* hear bits and pieces about these things no matter<mask> hard you try to isolate yourself, and<mask> you're forcing your own ignorance beyond<mask> you pick up by accident, you will end up forming your own opinion based on probably a whole mess of misconceptions, uninformed opinions, incomplete and out of context information, etc. [NEWLINE] [NEWLINE] Far better for people to first learn about the things they are going to have an opinion about.  Knowledge is<mask> took us, and continues to take us away from the "isms" that you fear.  Ignorance breeds those attitudes more than anything else.</s>
Label encoding: <s>I don't know, I think your reasons could (and will, imho ) work against you.  If you're going out of your way to remain ignorant about things, you're setting yourself up to have ill-informed ideas about them. [NEWLINE] [NEWLINE] You don't live in a vacuum; you *will* hear bits and pieces about these things no matter how hard you try to isolate yourself, and if you're forcing your own ignorance beyond what you pick up by accident, you will end up forming your own opinion based on probably a whole mess of misconceptions, uninformed opinions, incomplete and out of context information, etc. [NEWLINE] [NEWLINE] Far better for people to first learn about the things they are going to have an opinion about.  Knowledge is what took us, and continues to take us away from the "isms" that you fear.  Ignorance breeds those attitudes more than anything else.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is true,<mask> it<mask> means that a non-European person learning English<mask> has an immediate foothold in other Romance and Germanic languages and vice versa. The more globalized we become, the more we want dominant languages to be diverse and generalizable. [ENDQ] [NEWLINE] I guess it's impossible for natural world language to avoid giving certain regions an advantage over others,<mask> I still hesitate to approve of that. [NEWLINE] [NEWLINE] [STARTQ] This is true to an extent. Anglish certainly has a distinctly epic feel.<mask> a massive and diverse vocabulary has the benefit of added specificity. And that benefit doesn't just apply to art and literature<mask> to logic, science, law, and philosophy. [ENDQ] [NEWLINE] &amp;#8710; That's a good point.  One of the nice things about English is the ability to draw specific distinctions<mask> you want to, and speak more ambiguously<mask> you don't.</s>
Label encoding: <s> [STARTQ] This is true, but it also means that a non-European person learning English also has an immediate foothold in other Romance and Germanic languages and vice versa. The more globalized we become, the more we want dominant languages to be diverse and generalizable. [ENDQ] [NEWLINE] I guess it's impossible for natural world language to avoid giving certain regions an advantage over others, but I still hesitate to approve of that. [NEWLINE] [NEWLINE] [STARTQ] This is true to an extent. Anglish certainly has a distinctly epic feel. But a massive and diverse vocabulary has the benefit of added specificity. And that benefit doesn't just apply to art and literature but to logic, science, law, and philosophy. [ENDQ] [NEWLINE] &amp;#8710; That's a good point.  One of the nice things about English is the ability to draw specific distinctions when you want to, and speak more ambiguously when you don't.</s>
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're argument basically boils down to the fact that you won't pay for something<mask> you don't feel that it's worth the price. In this instance the "price" is your time and attention. The issue I have is that you're still enjoying the content. I don't see<mask><mask> you're doing is any different than sneaking into a movie. You definitely haven't shown an argument that it isn't wrong. You've just basically said that<mask> you use it, you're not<mask> wrong<mask> others,<mask> you turn it off<mask> it's a content provider that you really like. [NEWLINE] [NEWLINE] You're indie game comparison doesn't really work, either. Presumably you aren't pirating the games that you aren't buying,<mask> it's not really the same<mask> circumventing the price to get to the content.<mask> you are pirating them,<mask><mask> that's just<mask> wrong.</s>
Label encoding: <s>You're argument basically boils down to the fact that you won't pay for something if you don't feel that it's worth the price. In this instance the "price" is your time and attention. The issue I have is that you're still enjoying the content. I don't see how what you're doing is any different than sneaking into a movie. You definitely haven't shown an argument that it isn't wrong. You've just basically said that when you use it, you're not as wrong as others, because you turn it off if it's a content provider that you really like. [NEWLINE] [NEWLINE] You're indie game comparison doesn't really work, either. Presumably you aren't pirating the games that you aren't buying, So it's not really the same as circumventing the price to get to the content. If you are pirating them, I think that's just as wrong.</s>
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't beleive it is possible [ENDQ] [NEWLINE] Not possible?<mask> about... World War II for example? And worse things are imaginable.  And<mask> you linked is mostly beside the point. The general trend might be, - for us, now - towards less violence. This does not mean anything close to the claim that that all possible societies with more technology are happier.<mask> you didn't answer my question. [NEWLINE] [NEWLINE] I'm glad you've started using "to me" and "in my mind". <mask> you can't just impose your concepts of worth on others - especially at their expense.  Or at least,<mask> you think it's okay to, I<mask> disagree with you there. [NEWLINE] [NEWLINE] [STARTQ] This makes the creator more capable than the other, and<mask> better. [ENDQ] [NEWLINE] No, it does not follow. This is very explicitly "might is right".  </s>
Label encoding: <s> [STARTQ] I don't beleive it is possible [ENDQ] [NEWLINE] Not possible? How about... World War II for example? And worse things are imaginable.  And what you linked is mostly beside the point. The general trend might be, - for us, now - towards less violence. This does not mean anything close to the claim that that all possible societies with more technology are happier. So you didn't answer my question. [NEWLINE] [NEWLINE] I'm glad you've started using "to me" and "in my mind".  But you can't just impose your concepts of worth on others - especially at their expense.  Or at least, if you think it's okay to, I also disagree with you there. [NEWLINE] [NEWLINE] [STARTQ] This makes the creator more capable than the other, and thus better. [ENDQ] [NEWLINE] No, it does not follow. This is very explicitly "might is right".  </s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I would like to see my child-free peers in a more favorable light. [ENDQ] [NEWLINE] <mask> you showed them this post, and let them know the terms in which you think of them, it may be a problem that solves itself. [NEWLINE] [NEWLINE] I mean, is it really that difficult for a person to simply say "Hey! There's a life choice that literally has absolutely nothing to do with me! I'll just keep on truckin' by!" [NEWLINE] [NEWLINE] Or to acknowledge that such complicated and convoluted rationalizations have nearly nothing to do with the object or person you perceive and everything to do with feeding your own sense of self satisfaction? [NEWLINE] [NEWLINE] Or to recognize that at no point in your life has your shit smelled like roses,<mask> maybe you should be a little forgiving about the stink you *perceive* wafting off of others? [NEWLINE] [NEWLINE] <mask>... is it?  </s>
Label encoding: <s> [STARTQ] I would like to see my child-free peers in a more favorable light. [ENDQ] [NEWLINE] If you showed them this post, and let them know the terms in which you think of them, it may be a problem that solves itself. [NEWLINE] [NEWLINE] I mean, is it really that difficult for a person to simply say "Hey! There's a life choice that literally has absolutely nothing to do with me! I'll just keep on truckin' by!" [NEWLINE] [NEWLINE] Or to acknowledge that such complicated and convoluted rationalizations have nearly nothing to do with the object or person you perceive and everything to do with feeding your own sense of self satisfaction? [NEWLINE] [NEWLINE] Or to recognize that at no point in your life has your shit smelled like roses, therefore maybe you should be a little forgiving about the stink you *perceive* wafting off of others? [NEWLINE] [NEWLINE] So... is it?  </s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe that no one should be able to pay for school [ENDQ] [NEWLINE] You do nothing to illustrate<mask> the education gap will somehow close. Your solution is to make it<mask> that nobody is allowed to put money into their education. All this accomplishes is making all school systems equally poor, ergo they are all equally bad. [NEWLINE] [NEWLINE] This punishes the best, most capable teachers by cutting the majority of their wages, meaning they will most likely have to leave teaching altogether, and punishes the rich by limiting their education opportunities just for the sake of fairness.<mask> does anyone benefit from this? [NEWLINE] [NEWLINE] [STARTQ] I believe that all schools in all stages of education should be free [ENDQ] [NEWLINE] Yeah,<mask> does everyone else.<mask><mask>, everything should be free. Unfortunately, it costs resources. Education costs are hugely inflated,<mask> making them illegal is not a fix, it's a handicap.</s>
Label encoding: <s> [STARTQ] I believe that no one should be able to pay for school [ENDQ] [NEWLINE] You do nothing to illustrate how the education gap will somehow close. Your solution is to make it so that nobody is allowed to put money into their education. All this accomplishes is making all school systems equally poor, ergo they are all equally bad. [NEWLINE] [NEWLINE] This punishes the best, most capable teachers by cutting the majority of their wages, meaning they will most likely have to leave teaching altogether, and punishes the rich by limiting their education opportunities just for the sake of fairness. How does anyone benefit from this? [NEWLINE] [NEWLINE] [STARTQ] I believe that all schools in all stages of education should be free [ENDQ] [NEWLINE] Yeah, so does everyone else. In fact, everything should be free. Unfortunately, it costs resources. Education costs are hugely inflated, but making them illegal is not a fix, it's a handicap.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] we can safely conclude they aren't *solely* used for exploiting people. [ENDQ] [NEWLINE] The same could very well be true for Scientology in a century or two. [NEWLINE] [NEWLINE] [STARTQ] Considering we know a fair amount about the historical context of their creation, and have documentation following their evolution, we can further conclude they were not created with the intent of manipulating people. [ENDQ] [NEWLINE] <mask><mask> much do we really know?  Do you know who began the religion of Christianity, who was the first little leader?  I don't.  I don't think anyone does. [NEWLINE] [NEWLINE] [STARTQ] Scientology itself (<mask> we can't necessarily fault him with<mask> it was used) has at no point in time not exploited it's followers. [ENDQ] [NEWLINE] Over ~60 years.  Were the first sixty years of Christianity any better?  Hell, are there even decent records of the first sixty years of Christianity?</s>
Label encoding: <s> [STARTQ] we can safely conclude they aren't *solely* used for exploiting people. [ENDQ] [NEWLINE] The same could very well be true for Scientology in a century or two. [NEWLINE] [NEWLINE] [STARTQ] Considering we know a fair amount about the historical context of their creation, and have documentation following their evolution, we can further conclude they were not created with the intent of manipulating people. [ENDQ] [NEWLINE] But how much do we really know?  Do you know who began the religion of Christianity, who was the first little leader?  I don't.  I don't think anyone does. [NEWLINE] [NEWLINE] [STARTQ] Scientology itself ( because we can't necessarily fault him with how it was used) has at no point in time not exploited it's followers. [ENDQ] [NEWLINE] Over ~60 years.  Were the first sixty years of Christianity any better?  Hell, are there even decent records of the first sixty years of Christianity?</s>
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you are discussing a concept that hasn't been named before, then creating a word, or redefining an old one helps with clarity. In almost every other circumstances it is used to muddy the waters or introduce fuzziness that will benefit the definer. [NEWLINE] [NEWLINE] Im looking specifically at your last example,<mask> would you ignore the botanical definition? And the answer would be to put<mask> you are importing into a category with lighter tariffs. [NEWLINE] [NEWLINE] Theres a reason that legally and philosophically, every argument begins with a clarification and agreement on the definitions of terms.<mask> you have to redefine a word to win, your argument sucks. [NEWLINE] [NEWLINE] Finally, the experts in the field usually have the best definitions,<mask> they have spent years, possibly decades making those definitions accurate. An appeal to general language is either laziness (too hard to learn the language!) or obfuscation. </s>
Label encoding: <s>If you are discussing a concept that hasn't been named before, then creating a word, or redefining an old one helps with clarity. In almost every other circumstances it is used to muddy the waters or introduce fuzziness that will benefit the definer. [NEWLINE] [NEWLINE] Im looking specifically at your last example, why would you ignore the botanical definition? And the answer would be to put what you are importing into a category with lighter tariffs. [NEWLINE] [NEWLINE] Theres a reason that legally and philosophically, every argument begins with a clarification and agreement on the definitions of terms. If you have to redefine a word to win, your argument sucks. [NEWLINE] [NEWLINE] Finally, the experts in the field usually have the best definitions, because they have spent years, possibly decades making those definitions accurate. An appeal to general language is either laziness (too hard to learn the language!) or obfuscation. </s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Now, just a question, would it not be important to<mask> have information whether or not your anatomy has changed<mask> you've gone MTF? I mean, it would be medically important to figure out<mask> you still have your prostate, right? [ENDQ] [NEWLINE] Yes,<mask> the importance of this information depends on the procedure.  Ultimately, I have to take responsibility for my own health, and I should be at least<mask> educated about my self-specific health issues<mask> the physicians who are treating me.  This is true for everyone,<mask> especially transsexuals - we are medically unusual,<mask> we need to be our own advocates first and foremost. [NEWLINE] [NEWLINE] I do have a prostate (afaik all modern MTF sex reassignment surgeries leave the prostate intact),<mask> keep in mind that castration and estrogen therapy mean that prostate cancer and enlarged prostate problems are a vanishingly remote concern.</s>
Label encoding: <s> [STARTQ] Now, just a question, would it not be important to also have information whether or not your anatomy has changed since you've gone MTF? I mean, it would be medically important to figure out if you still have your prostate, right? [ENDQ] [NEWLINE] Yes, but the importance of this information depends on the procedure.  Ultimately, I have to take responsibility for my own health, and I should be at least as educated about my self-specific health issues as the physicians who are treating me.  This is true for everyone, but especially transsexuals - we are medically unusual, so we need to be our own advocates first and foremost. [NEWLINE] [NEWLINE] I do have a prostate (afaik all modern MTF sex reassignment surgeries leave the prostate intact), but keep in mind that castration and estrogen therapy mean that prostate cancer and enlarged prostate problems are a vanishingly remote concern.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this will just come down to semantics<mask> lets start with the definition of words: [NEWLINE] [NEWLINE] sport  [spawrt, spohrt]  Show IPA [NEWLINE] noun [NEWLINE] 1. an athletic activity requiring skill or physical prowess and often of a [NEWLINE] competitive nature,<mask> racing, baseball, tennis, golf, bowling, wrestling, [NEWLINE] boxing, hunting, fishing, etc. [NEWLINE] [source]( [URL] ) [NEWLINE] [NEWLINE] I've erased the other definitions<mask><mask><mask> this is the one that applies. [NEWLINE] [NEWLINE] I'll ignore that the source decided to include golf<mask> an example (ha). [NEWLINE] [NEWLINE] <mask> that definition above suffices, golf certainly requires skills and is competitive.  I suspect it  is only the physical prowess that you are defining sport by,<mask> by definition alone it can be skill **or** physical prowess, thereby, making golf a sport. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I think this will just come down to semantics so lets start with the definition of words: [NEWLINE] [NEWLINE] sport  [spawrt, spohrt]  Show IPA [NEWLINE] noun [NEWLINE] 1. an athletic activity requiring skill or physical prowess and often of a [NEWLINE] competitive nature, as racing, baseball, tennis, golf, bowling, wrestling, [NEWLINE] boxing, hunting, fishing, etc. [NEWLINE] [source]( [URL] ) [NEWLINE] [NEWLINE] I've erased the other definitions as I think this is the one that applies. [NEWLINE] [NEWLINE] I'll ignore that the source decided to include golf as an example (ha). [NEWLINE] [NEWLINE] If that definition above suffices, golf certainly requires skills and is competitive.  I suspect it  is only the physical prowess that you are defining sport by, but by definition alone it can be skill **or** physical prowess, thereby, making golf a sport. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you want to have your "300-level" discussion amongst other people who clearly understand the modified rules of debate, then sure. The problem is<mask> someone posts some idiot quote to twitter like "Dear white folks. You're racist. I'm not. End of story" and then *doesn't* explain<mask> the hell they are trying to say. It's a big part of the reason that "SJW" is the slur that it is. You would think "SJW" would be a positive thing,<mask><mask><mask> you'll be hard pressed to find people who agree without identifying themselves<mask> one already. Instead the label is associated with ignorant young women that don't know<mask> to communicate with people that don't already agree with them. [NEWLINE] [NEWLINE] And I *do* hope we're not talking about SRS<mask> a 300-level arena for sociology discussion.</s>
Label encoding: <s>If you want to have your "300-level" discussion amongst other people who clearly understand the modified rules of debate, then sure. The problem is when someone posts some idiot quote to twitter like "Dear white folks. You're racist. I'm not. End of story" and then *doesn't* explain what the hell they are trying to say. It's a big part of the reason that "SJW" is the slur that it is. You would think "SJW" would be a positive thing, but I think you'll be hard pressed to find people who agree without identifying themselves as one already. Instead the label is associated with ignorant young women that don't know how to communicate with people that don't already agree with them. [NEWLINE] [NEWLINE] And I *do* hope we're not talking about SRS as a 300-level arena for sociology discussion.</s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] For all Christians' pro-life pretentions, no half-baked theosophy will ever rationalize progeria, or miscarriage/stillbirth. Given an omnipotent, omniscient, omnibenevolent creator, the fall of man happened<mask> God consciously chose to make it happen. No way around it. [ENDQ] [NEWLINE] This doesn't really seem to be relevant to your view,<mask> I just want to mention: you have brought up "the problem of evil," which is a well-known atheist attempt to disprove Christianity. The LDS Church, of which I'm a member, has an extremely reasonable answer to the problems of suffering and evil. I'm willing to try and explain them<mask> you're interested. [NEWLINE] [NEWLINE] The LDS Church is an outlier among Christian religions,<mask> I'm certain that other Christian churches<mask> have responses that they believe are reasonable.</s><pad>
Label encoding: <s> [STARTQ] For all Christians' pro-life pretentions, no half-baked theosophy will ever rationalize progeria, or miscarriage/stillbirth. Given an omnipotent, omniscient, omnibenevolent creator, the fall of man happened because God consciously chose to make it happen. No way around it. [ENDQ] [NEWLINE] This doesn't really seem to be relevant to your view, but I just want to mention: you have brought up "the problem of evil," which is a well-known atheist attempt to disprove Christianity. The LDS Church, of which I'm a member, has an extremely reasonable answer to the problems of suffering and evil. I'm willing to try and explain them if you're interested. [NEWLINE] [NEWLINE] The LDS Church is an outlier among Christian religions, but I'm certain that other Christian churches also have responses that they believe are reasonable.</s><pad>
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think the only variable is skin colour here. Blacks and whites live in different kinds of neighbourhoods and live different kinds of lives, and are<mask> subject to different levels of policing.<mask> you were looking at the UK and comparing the arrest rate for marijuana use for white middle class and working class groups you'd probably see a similar pattern, and that's not to do with racism<mask> both groups are white. [NEWLINE] [NEWLINE] I'm not saying racism does not exist,<mask> I would certainly question whether it is<mask> pervasive<mask> some believe it to be,and<mask> part it plays in the fortunes of blacks. Saying "Racism is<mask> pervasive," is completely unqualified<mask> a statement: It's hard to measure and<mask> hard to disprove one way or the other. And most will jump to it<mask> an explanation without exhuasting other possibilities, or even considering them.</s>
Label encoding: <s>I don't think the only variable is skin colour here. Blacks and whites live in different kinds of neighbourhoods and live different kinds of lives, and are thus subject to different levels of policing. If you were looking at the UK and comparing the arrest rate for marijuana use for white middle class and working class groups you'd probably see a similar pattern, and that's not to do with racism because both groups are white. [NEWLINE] [NEWLINE] I'm not saying racism does not exist, but I would certainly question whether it is as pervasive as some believe it to be,and what part it plays in the fortunes of blacks. Saying "Racism is so pervasive," is completely unqualified as a statement: It's hard to measure and thus hard to disprove one way or the other. And most will jump to it as an explanation without exhuasting other possibilities, or even considering them.</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Even that would require quite a lot of labor, and be open to an incredible amount of legislation. Say I report you. They now need to have a camera in your car, or mine (assume at least one), and analyze the video. They open themselves up to a lot of expensive lawsuits by levying higher or lower premiums, especially ones that are not backed by statistical models<mask> current ones are. Consider, say, reporting of crimes by the public. Even something that carries the force of law requires things like a trial, often with a jury, defense and prosecuting attorneys, and<mask> on.<mask> traffic? Something like safe merging distance is open to an incredible amount of individual interpretation and judgement.<mask><mask> the cost of implementing and defending such a system would far outweigh any small premium benefits a safe driver might see. It's a simple case of penny wise, pound foolish.</s>
Label encoding: <s>Even that would require quite a lot of labor, and be open to an incredible amount of legislation. Say I report you. They now need to have a camera in your car, or mine (assume at least one), and analyze the video. They open themselves up to a lot of expensive lawsuits by levying higher or lower premiums, especially ones that are not backed by statistical models as current ones are. Consider, say, reporting of crimes by the public. Even something that carries the force of law requires things like a trial, often with a jury, defense and prosecuting attorneys, and so on. But traffic? Something like safe merging distance is open to an incredible amount of individual interpretation and judgement. I think the cost of implementing and defending such a system would far outweigh any small premium benefits a safe driver might see. It's a simple case of penny wise, pound foolish.</s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Correct me<mask> I'm wrong,<mask> you seem to be taking physics to be the scientific study of everything - quite simply, it's not. [<mask> it studies *applies* to everything in the natural world]( [URL] ),<mask> that doesn't mean it *is* the study of everything or else there would be no need for other sciences. Physics in itself tells us relatively little about<mask> a living organism functions, or<mask> chemicals react with each other, or<mask> the mind forms the thoughts it does, or any number of other topics covered by other fields. Whether you consider those things meaningful to the search for truth is one thing,<mask> "physics" is not a synonym for "all scientific disciplines" and it shouldn't be used<mask> one.<mask>, are you saying that ultimate truth can be found in the sciences, or in the study of physics specifically?</s>
Label encoding: <s>Correct me if I'm wrong, but you seem to be taking physics to be the scientific study of everything - quite simply, it's not. [ What it studies *applies* to everything in the natural world]( [URL] ), but that doesn't mean it *is* the study of everything or else there would be no need for other sciences. Physics in itself tells us relatively little about how a living organism functions, or how chemicals react with each other, or how the mind forms the thoughts it does, or any number of other topics covered by other fields. Whether you consider those things meaningful to the search for truth is one thing, but "physics" is not a synonym for "all scientific disciplines" and it shouldn't be used as one. So, are you saying that ultimate truth can be found in the sciences, or in the study of physics specifically?</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0874, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I fully agree that the scientific method is the way to go and your arguments are close to changing part of my view already. [NEWLINE] [NEWLINE] I have to admit I'm not very well read on all the popular conspiracy theories,<mask> I'd be surprised<mask> absolutely all of them have been absolutely disproved by the scientific method. There are many aspects to each conspiracy and there are a lot of them. I doubt all these aspects of every conspiracy theory has been scientifically refuted. [NEWLINE] [NEWLINE] <mask> you're speaking a lot of their/them. This is the stereotypical "crazy" view that I don't defend, I want to defend the genuine parts of each conspiracy theory. [NEWLINE] [NEWLINE] Take [URL] / for instance, over 2000 verified architects and engineers that oppose the official 9/11 story. I'm quite sure they all know very well about the scientific method and do their best to follow it.</s><pad>
Label encoding: <s>I fully agree that the scientific method is the way to go and your arguments are close to changing part of my view already. [NEWLINE] [NEWLINE] I have to admit I'm not very well read on all the popular conspiracy theories, but I'd be surprised if absolutely all of them have been absolutely disproved by the scientific method. There are many aspects to each conspiracy and there are a lot of them. I doubt all these aspects of every conspiracy theory has been scientifically refuted. [NEWLINE] [NEWLINE] Also you're speaking a lot of their/them. This is the stereotypical "crazy" view that I don't defend, I want to defend the genuine parts of each conspiracy theory. [NEWLINE] [NEWLINE] Take [URL] / for instance, over 2000 verified architects and engineers that oppose the official 9/11 story. I'm quite sure they all know very well about the scientific method and do their best to follow it.</s><pad>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is the perfect example of<mask> OP is talking about.  It's about time people got some pride in the human race<mask> a whole instead of just their section of the human race.  Be proud of your country, sure. <mask> perhaps it's time to realise that we should be looking after everyone, not just ourselves. <mask> should being born in a different place mean that you should not care about people born in a different place<mask> much? [NEWLINE] [NEWLINE] Patriotism puts barriers up between people<mask> that each group only consider themselves and not the effects they are having on others.  I love my country<mask> a home,<mask><mask><mask> the world would be much better off<mask> we started thinking of ourselves<mask> a community that works together to advance the human race - instead of a bunch of separate communities competing to be the richest at the expense of all the others.</s>
Label encoding: <s>This is the perfect example of what OP is talking about.  It's about time people got some pride in the human race as a whole instead of just their section of the human race.  Be proud of your country, sure.  But perhaps it's time to realise that we should be looking after everyone, not just ourselves.  Why should being born in a different place mean that you should not care about people born in a different place as much? [NEWLINE] [NEWLINE] Patriotism puts barriers up between people so that each group only consider themselves and not the effects they are having on others.  I love my country as a home, but I think the world would be much better off if we started thinking of ourselves as a community that works together to advance the human race - instead of a bunch of separate communities competing to be the richest at the expense of all the others.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know this is a little old,<mask> I finally have the time to answer some of these. [NEWLINE] [NEWLINE] <mask> is "a few people making very arbitrary decisions" much different from the founding fathers of the united states writing the constitution? [NEWLINE] [NEWLINE] Ideally I would take a group of informed individuals and put them in a room and we would discuss it at length until a system agreeable to most could be implemented. [NEWLINE] [NEWLINE] I believe this would not just result in another democratic republic<mask> we have two things that the founding fathers did not. [NEWLINE] [NEWLINE] We have computers and all that entails, the internet, virtually instantaneous communication and emotionless third parties. [NEWLINE] [NEWLINE] We<mask> have results from the "experiment in democracy" we have seen which problems occur with which systems, e.g. gerrymandering and vote spoiling. and would be able to make allowances for them.</s>
Label encoding: <s>I know this is a little old, but I finally have the time to answer some of these. [NEWLINE] [NEWLINE] How is "a few people making very arbitrary decisions" much different from the founding fathers of the united states writing the constitution? [NEWLINE] [NEWLINE] Ideally I would take a group of informed individuals and put them in a room and we would discuss it at length until a system agreeable to most could be implemented. [NEWLINE] [NEWLINE] I believe this would not just result in another democratic republic because we have two things that the founding fathers did not. [NEWLINE] [NEWLINE] We have computers and all that entails, the internet, virtually instantaneous communication and emotionless third parties. [NEWLINE] [NEWLINE] We also have results from the "experiment in democracy" we have seen which problems occur with which systems, e.g. gerrymandering and vote spoiling. and would be able to make allowances for them.</s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I definitely understand your point. [NEWLINE] [NEWLINE] Easiest<mask> most unfair would be to tax everything at a given percent. [NEWLINE] [NEWLINE] Most complicated<mask> fairest would be to zoom into each individual case and calculate<mask><mask> hundreds of inputs. [NEWLINE] [NEWLINE] A hybrid of both is<mask> we currently have.<mask> tax brackets come in and allowances/deductions etc. [NEWLINE] [NEWLINE] I would love to have a single calculation,<mask> I just can't see it happen in the current tax system. It would have to be built up from scratch and given the diversity of the populace and the necessity for fairness, that would be a very hard thing to design. [NEWLINE] [NEWLINE] Chances are that whatever system one would come up with, it would still be complicated cause even<mask> the calculation by itself would be single, tax deductions would not. [NEWLINE] [NEWLINE] <mask> yes, I am theoretically all for it.</s>
Label encoding: <s>I definitely understand your point. [NEWLINE] [NEWLINE] Easiest but most unfair would be to tax everything at a given percent. [NEWLINE] [NEWLINE] Most complicated but fairest would be to zoom into each individual case and calculate according to hundreds of inputs. [NEWLINE] [NEWLINE] A hybrid of both is what we currently have. So tax brackets come in and allowances/deductions etc. [NEWLINE] [NEWLINE] I would love to have a single calculation, but I just can't see it happen in the current tax system. It would have to be built up from scratch and given the diversity of the populace and the necessity for fairness, that would be a very hard thing to design. [NEWLINE] [NEWLINE] Chances are that whatever system one would come up with, it would still be complicated cause even if the calculation by itself would be single, tax deductions would not. [NEWLINE] [NEWLINE] But yes, I am theoretically all for it.</s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0849, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Hollywood. [ENDQ] [NEWLINE] I'm just going to throw out there that<mask> comics and movies share much of their visual languages<mask> a medium, comics have inherently different strengths with regard to pacing and linearity.<mask> I'm reading a comic, I can skip back and forth relatively easily to be able to keep up with<mask> happened fifty pages ago,<mask><mask> I'm watching a movie, I don't really have that luxury. [NEWLINE] [NEWLINE] With regard to your friend, I can't really comment without being able to talk to him myself. [NEWLINE] [NEWLINE] All I can really say is that there are a lot of good comics out there. Even in Capeville it's not just about the superpowers. [NEWLINE] [NEWLINE] <mask> you haven't already, give the "Superman is the worst character" thread a gander, some of the entry level comments will knock your socks off.</s>
Label encoding: <s> [STARTQ] Hollywood. [ENDQ] [NEWLINE] I'm just going to throw out there that while comics and movies share much of their visual languages as a medium, comics have inherently different strengths with regard to pacing and linearity. If I'm reading a comic, I can skip back and forth relatively easily to be able to keep up with what happened fifty pages ago, but if I'm watching a movie, I don't really have that luxury. [NEWLINE] [NEWLINE] With regard to your friend, I can't really comment without being able to talk to him myself. [NEWLINE] [NEWLINE] All I can really say is that there are a lot of good comics out there. Even in Capeville it's not just about the superpowers. [NEWLINE] [NEWLINE] If you haven't already, give the "Superman is the worst character" thread a gander, some of the entry level comments will knock your socks off.</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I grow a beard for a couple of different reasons 1. I don't feel like shaving every single day to keep a smooth face<mask> growing is easier. 2.<mask> it starts dropping into the teens and twenties a beard helps keep my face warm.<mask> I guess it is just practical for me to do. [NEWLINE] [NEWLINE] <mask> people are attracted to them? Maybe<mask> growing a beard takes take which shows dedication or commitment? I guess you would have to ask a woman<mask>.<mask><mask>,<mask> having facial hair is less common, in younger people especially, it could be seen<mask> a sign of maturity or confidence to go against the normal culture. [NEWLINE] [NEWLINE] <mask> is<mask> great about them is that everyone can be clean shaven including men, women and children,<mask> not everyone can grow a beard. You can even grow and style it into a beard all your own. </s>
Label encoding: <s>I grow a beard for a couple of different reasons 1. I don't feel like shaving every single day to keep a smooth face so growing is easier. 2. when it starts dropping into the teens and twenties a beard helps keep my face warm. So I guess it is just practical for me to do. [NEWLINE] [NEWLINE] Why people are attracted to them? Maybe because growing a beard takes take which shows dedication or commitment? I guess you would have to ask a woman why. In addition, since having facial hair is less common, in younger people especially, it could be seen as a sign of maturity or confidence to go against the normal culture. [NEWLINE] [NEWLINE] What is so great about them is that everyone can be clean shaven including men, women and children, but not everyone can grow a beard. You can even grow and style it into a beard all your own. </s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you are confusing racial profiling with poor police work. [NEWLINE] [NEWLINE] [STARTQ] The data you use to dictate which race to profile and scrutinize with more force is based on statistics in which you were profiling to begin with. [ENDQ] [NEWLINE] You are<mask><mask> you are profiling without using statistic. <mask> you are using statistic to do this,<mask> you focus on the most common racial group, they will be deterred from committing these same crimes.  And the numbers will change.  And you should adjust your profiling. [NEWLINE] [NEWLINE] [STARTQ] <mask> you trained an officer to profile against blacks for 20 years, and then all of a sudden it swayed, and whites were 5 times<mask> likely to commit murder over a black male. [ENDQ] [NEWLINE] You would train them to prioritize whites<mask> the number suggest to do<mask>.  Poor training and poor police work is a different thing than racial profiling.</s>
Label encoding: <s>I think you are confusing racial profiling with poor police work. [NEWLINE] [NEWLINE] [STARTQ] The data you use to dictate which race to profile and scrutinize with more force is based on statistics in which you were profiling to begin with. [ENDQ] [NEWLINE] You are assuming that you are profiling without using statistic.  If you are using statistic to do this, as you focus on the most common racial group, they will be deterred from committing these same crimes.  And the numbers will change.  And you should adjust your profiling. [NEWLINE] [NEWLINE] [STARTQ] If you trained an officer to profile against blacks for 20 years, and then all of a sudden it swayed, and whites were 5 times as likely to commit murder over a black male. [ENDQ] [NEWLINE] You would train them to prioritize whites when the number suggest to do so.  Poor training and poor police work is a different thing than racial profiling.</s>
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Depending on the nature of the case, it is possible that legal fees will be paid by the losing party. That very much varies by statute<mask>. [NEWLINE] [NEWLINE] <mask>, even in cases<mask> this is true, the person bringing the claim will have to pay their own legal fees<mask> they lose, which is a pretty strong disincentive to bring frivolous lawsuits.<mask> in most cases that go to trial, both parties will legitimately believe they are right, and they have chosen to take the matter to court to prove it at risk of losing. The role of the court is to resolve their disagreement by making a decision<mask> an authority that is binding. [NEWLINE] [NEWLINE] Bottom line: It is unlikely that people will risk bringing baseless lawsuits<mask> they will bear the costs of their own legal fees<mask> they lose. Cases will generally only go to trial<mask> there is a legitimate question of responsibility.</s>
Label encoding: <s>Depending on the nature of the case, it is possible that legal fees will be paid by the losing party. That very much varies by statute however. [NEWLINE] [NEWLINE] Also, even in cases where this is true, the person bringing the claim will have to pay their own legal fees if they lose, which is a pretty strong disincentive to bring frivolous lawsuits. Therefore in most cases that go to trial, both parties will legitimately believe they are right, and they have chosen to take the matter to court to prove it at risk of losing. The role of the court is to resolve their disagreement by making a decision as an authority that is binding. [NEWLINE] [NEWLINE] Bottom line: It is unlikely that people will risk bringing baseless lawsuits when they will bear the costs of their own legal fees when they lose. Cases will generally only go to trial if there is a legitimate question of responsibility.</s>
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>School shootings average about [15-20 per year in the US]( [URL] ). [NEWLINE] [NEWLINE] There are around [16,000 nonfatal, accidental firearm injuries]( [URL].pdf) every year. [NEWLINE] [NEWLINE] [NEWLINE] There are around [680 unintentional firearm deaths]( [URL].pdf) every year. [NEWLINE] [NEWLINE] There is [a strong, proven correlation between availability of firearms and firearm injuries and death]( [URL] /). The rate of both accidental and intentional shootings increase along with gun ownership. [NEWLINE] [NEWLINE] <mask> the available facts suggest that requiring a gun in every single classroom would kill and injure several times<mask> many students every year<mask> are killed today by deliberate school shootings. Whether intentional school shootings would be prevented by armed teachers<mask> you claim or not, the fact is that arming teachers would likely result in **a large increase in total deaths and injuries** among both teachers and students. </s>
Label encoding: <s>School shootings average about [15-20 per year in the US]( [URL] ). [NEWLINE] [NEWLINE] There are around [16,000 nonfatal, accidental firearm injuries]( [URL].pdf) every year. [NEWLINE] [NEWLINE] [NEWLINE] There are around [680 unintentional firearm deaths]( [URL].pdf) every year. [NEWLINE] [NEWLINE] There is [a strong, proven correlation between availability of firearms and firearm injuries and death]( [URL] /). The rate of both accidental and intentional shootings increase along with gun ownership. [NEWLINE] [NEWLINE] Therefore the available facts suggest that requiring a gun in every single classroom would kill and injure several times as many students every year as are killed today by deliberate school shootings. Whether intentional school shootings would be prevented by armed teachers as you claim or not, the fact is that arming teachers would likely result in **a large increase in total deaths and injuries** among both teachers and students. </s>
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>(I'm going to start by saying I've read the 1st book<mask> not seen past the show's pilot.) [NEWLINE] [NEWLINE] Based on your responses to other comments, it seems to me that you and I are in disagreement about<mask> the protagonist of a story is. You make the claim that<mask> Cersei and the Lannisters cause the main conflicts in the story (such<mask> injuring Bran and seating Joff on the Iron Throne.<mask>, the way I, and apparently most others (heh, Others) ITT regard the protagonist<mask> the main character dealing with the story's conflict<mask> it is caused and evolves. By this definition, Ned, Sansa, John Snow, Arya, King Baratheon, or a whole host of others (hehehe, Others) dealing with Lannister (and Targaryen) **antagonism**.</s>
Label encoding: <s>(I'm going to start by saying I've read the 1st book but not seen past the show's pilot.) [NEWLINE] [NEWLINE] Based on your responses to other comments, it seems to me that you and I are in disagreement about what the protagonist of a story is. You make the claim that since Cersei and the Lannisters cause the main conflicts in the story (such as injuring Bran and seating Joff on the Iron Throne. However, the way I, and apparently most others (heh, Others) ITT regard the protagonist as the main character dealing with the story's conflict as it is caused and evolves. By this definition, Ned, Sansa, John Snow, Arya, King Baratheon, or a whole host of others (hehehe, Others) dealing with Lannister (and Targaryen) **antagonism**.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For point 2.  We don't really have a problem making food for people. Transporting food is the problem(logistical or financial).  Unless you live somewhere<mask> that is an issue your extra mouth doesn't take food out of another's. It's no more selfish than feeding yourself. [NEWLINE] [NEWLINE] <mask> for point 1. That voice in your head telling you to have a child is<mask> connects us to all other sentient life.  We exist to reproduce.  It's<mask> we lived to take domain over the earth (and moon) in a darwinian sense you are acting selflessly sacrificing time energy and possibly health to broaden the human gene pool. (this<mask> answers point three for me<mask> well.  Unless you believe humans are truly special in the eye of some deity then we can be punished for being the animals we are.) [NEWLINE] [NEWLINE] </s>
Label encoding: <s>For point 2.  We don't really have a problem making food for people. Transporting food is the problem(logistical or financial).  Unless you live somewhere where that is an issue your extra mouth doesn't take food out of another's. It's no more selfish than feeding yourself. [NEWLINE] [NEWLINE] As for point 1. That voice in your head telling you to have a child is what connects us to all other sentient life.  We exist to reproduce.  It's how we lived to take domain over the earth (and moon) in a darwinian sense you are acting selflessly sacrificing time energy and possibly health to broaden the human gene pool. (this also answers point three for me as well.  Unless you believe humans are truly special in the eye of some deity then we can be punished for being the animals we are.) [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you're using this example<mask> a way in which Jews beg for attention, which is a little... awkward. I would suggest you be careful<mask> often you say stuff like this, [ENDQ] [NEWLINE] This is one of the points he is making. You can say things against other ethnic groups,<mask> the minute you say something against the Jewish group, you are warned to stop or else! [NEWLINE] [NEWLINE] [STARTQ] it's a sentiment that was very popular not<mask> long ago, and is pretty strongly associated with Holocaust denial. [ENDQ] [NEWLINE] He's said repeatedly that he isn't,<mask> you still through that out like it is a threat. That is part of<mask> is ingrained into Jewish heritage, like many blacks automatically go for the race card<mask> anything is said about black people<mask> a whole. Granted you shouldn't make broad accusations in most cases,<mask> sometimes they fit.</s>
Label encoding: <s> [STARTQ] you're using this example as a way in which Jews beg for attention, which is a little... awkward. I would suggest you be careful how often you say stuff like this, [ENDQ] [NEWLINE] This is one of the points he is making. You can say things against other ethnic groups, but the minute you say something against the Jewish group, you are warned to stop or else! [NEWLINE] [NEWLINE] [STARTQ] it's a sentiment that was very popular not so long ago, and is pretty strongly associated with Holocaust denial. [ENDQ] [NEWLINE] He's said repeatedly that he isn't, but you still through that out like it is a threat. That is part of what is ingrained into Jewish heritage, like many blacks automatically go for the race card when anything is said about black people as a whole. Granted you shouldn't make broad accusations in most cases, but sometimes they fit.</s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The only time I ever forgot to check was<mask> a kid. Made a mistake once, learned from it and problem solved forever. [NEWLINE] [NEWLINE] The toilet seat debate is such a non-issue I'm surprised it's actually an issue for some people. It's<mask> obviously about more (sexism/feminism/man vs woman) it's ridiculous we don't just come right out and say "I am a man, roar to the toilet seat, up be you" or "I am a woman, roar to the toilet seat, provide me comfort in period time". [NEWLINE] [NEWLINE] Change the toilet seat to whatever position you need whenever you enter the bathroom, then be a lazy fuck and and leave it<mask> you used it.<mask> you happen to be the one who has to put it up more often, good for you, train those biceps hard.</s>
Label encoding: <s>The only time I ever forgot to check was as a kid. Made a mistake once, learned from it and problem solved forever. [NEWLINE] [NEWLINE] The toilet seat debate is such a non-issue I'm surprised it's actually an issue for some people. It's so obviously about more (sexism/feminism/man vs woman) it's ridiculous we don't just come right out and say "I am a man, roar to the toilet seat, up be you" or "I am a woman, roar to the toilet seat, provide me comfort in period time". [NEWLINE] [NEWLINE] Change the toilet seat to whatever position you need whenever you enter the bathroom, then be a lazy fuck and and leave it as you used it. If you happen to be the one who has to put it up more often, good for you, train those biceps hard.</s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From a strictly moral standpoint,<mask><mask>. That would be a better way of adjudicating things. [NEWLINE] [NEWLINE] From a practical standpoint, considering that people will lie, information after the fact will be incomplete, and peoples' motivations are often very complicated and not even personally understood, it's virtually impossible to have a system of laws and justice based on intentions. [NEWLINE] [NEWLINE] You just can't ever know exactly<mask> someone wanted to do something. It's wide open to abuse either way - trust the defendant, and you give a lot of room for violent criminals to claim accidents (and plan crimes to look like deniable accidents). Trust the prosecution based on'reasonable belief' and you get overzealous attorneys making a reasonable-sounding case that intentions were much worse. For purposes of *actually* policing and adjudicating people, you need a system based on facts.</s>
Label encoding: <s>From a strictly moral standpoint, I agree. That would be a better way of adjudicating things. [NEWLINE] [NEWLINE] From a practical standpoint, considering that people will lie, information after the fact will be incomplete, and peoples' motivations are often very complicated and not even personally understood, it's virtually impossible to have a system of laws and justice based on intentions. [NEWLINE] [NEWLINE] You just can't ever know exactly why someone wanted to do something. It's wide open to abuse either way - trust the defendant, and you give a lot of room for violent criminals to claim accidents (and plan crimes to look like deniable accidents). Trust the prosecution based on'reasonable belief' and you get overzealous attorneys making a reasonable-sounding case that intentions were much worse. For purposes of *actually* policing and adjudicating people, you need a system based on facts.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I'm outside and I want to check my email/look up directions/etc. on a laptop [ENDQ] [NEWLINE] <mask> is checking email outside *absolutely necessary* at that given moment? Does the fate and integrity of your job record rely on your instant email response? Is it not possible to wait until you return to a desk setting to check your email? [NEWLINE] [NEWLINE] A GPS can<mask> look up directions. [NEWLINE] [NEWLINE] [STARTQ] <mask>, it would suck a lot more for me to lose my $1000 dollar laptop with around $400 in software than it would to lose my couple-hundred-dollar smartphone. [ENDQ] [NEWLINE] It seems easier to lose a smartphone than a laptop, given<mask> small they are and are more accessible to one handed carelessness. Given the same amount of pampering, surely the smartphone is more prone to theft, loss, etc.</s>
Label encoding: <s> [STARTQ] If I'm outside and I want to check my email/look up directions/etc. on a laptop [ENDQ] [NEWLINE] But is checking email outside *absolutely necessary* at that given moment? Does the fate and integrity of your job record rely on your instant email response? Is it not possible to wait until you return to a desk setting to check your email? [NEWLINE] [NEWLINE] A GPS can also look up directions. [NEWLINE] [NEWLINE] [STARTQ] Also, it would suck a lot more for me to lose my $1000 dollar laptop with around $400 in software than it would to lose my couple-hundred-dollar smartphone. [ENDQ] [NEWLINE] It seems easier to lose a smartphone than a laptop, given how small they are and are more accessible to one handed carelessness. Given the same amount of pampering, surely the smartphone is more prone to theft, loss, etc.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0822, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Excellent point.  The irony of this entire conversation is that I am actually hugely supportive of women (and men) choosing for themselves<mask> to do about work, child care, parenting, etc. <mask><mask> it is a serious shame that there is still a stigma about going back to work. <mask><mask><mask><mask>, there is<mask> a definite stigma about being a SAHM.  Unfortunately, it goes both ways. [NEWLINE] [NEWLINE] Of course, it would be ideal<mask> someone could say, "That's great that you are staying home, I am excited for you.  I am going back to work, and I am really excited about that too." <mask>, that is probably a bit of a pipe dream.  I am sure I have been in similar situations and taken the easy way out (said the less true<mask> more acceptable thing).</s>
Label encoding: <s>Excellent point.  The irony of this entire conversation is that I am actually hugely supportive of women (and men) choosing for themselves what to do about work, child care, parenting, etc.  I think it is a serious shame that there is still a stigma about going back to work.  On the other hand, there is also a definite stigma about being a SAHM.  Unfortunately, it goes both ways. [NEWLINE] [NEWLINE] Of course, it would be ideal if someone could say, "That's great that you are staying home, I am excited for you.  I am going back to work, and I am really excited about that too."  But, that is probably a bit of a pipe dream.  I am sure I have been in similar situations and taken the easy way out (said the less true but more acceptable thing).</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll take a stab at putting it in layman's terms. Some decisions in the brain are made on a quantum level. This means that they truly are random. These small decisions which would change every time you made them, even with the same past history, can change<mask> you'll react to future things in a butterfly effect. Picking snickers over milky way, a quantum random choice, can ultimately have unforseeable effects later on. [NEWLINE] [NEWLINE] <mask>,<mask> it's not *exactly* free will (<mask> it's more free will subject to quantum whims), it does mean that the future is not deterministic based on a mathematical equation laid down at the dawn of time. There is not one single outcome any universe laid down with these same rules will come to. Due to quantum level choices, it will be different every time.</s>
Label encoding: <s>I'll take a stab at putting it in layman's terms. Some decisions in the brain are made on a quantum level. This means that they truly are random. These small decisions which would change every time you made them, even with the same past history, can change how you'll react to future things in a butterfly effect. Picking snickers over milky way, a quantum random choice, can ultimately have unforseeable effects later on. [NEWLINE] [NEWLINE] So, while it's not *exactly* free will ( as it's more free will subject to quantum whims), it does mean that the future is not deterministic based on a mathematical equation laid down at the dawn of time. There is not one single outcome any universe laid down with these same rules will come to. Due to quantum level choices, it will be different every time.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1) You don't "mine a bitcoin" per se, you "mine a block" which is ~10 minutes worth of Bitcoin transactions, and you are given a "block reward" which at the moment is 25 bitcoins IIRC.<mask> you successfully mine a block, your results for mining the block are broadcast to the other nodes and they can review the work to instantly accept or reject it<mask> being legitimate. [NEWLINE] [NEWLINE] 2) Many of the security checks built into the system rely upon consensus. It would take more than a single node broadcasting / accepting fraudulent transactions for other nodes to agree. This is<mask> the "51% attack" comes into play and<mask> it is important for the total hash rate of the Bitcoin network to be very high - this makes it incredibly difficult for any one entity to gain 51% control of the nodes on the network.</s>
Label encoding: <s>1) You don't "mine a bitcoin" per se, you "mine a block" which is ~10 minutes worth of Bitcoin transactions, and you are given a "block reward" which at the moment is 25 bitcoins IIRC. When you successfully mine a block, your results for mining the block are broadcast to the other nodes and they can review the work to instantly accept or reject it as being legitimate. [NEWLINE] [NEWLINE] 2) Many of the security checks built into the system rely upon consensus. It would take more than a single node broadcasting / accepting fraudulent transactions for other nodes to agree. This is where the "51% attack" comes into play and why it is important for the total hash rate of the Bitcoin network to be very high - this makes it incredibly difficult for any one entity to gain 51% control of the nodes on the network.</s>
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand, and<mask><mask> most people agree with you. <mask><mask><mask><mask> with you. [NEWLINE] [NEWLINE] Here's<mask> hangs me up. <mask> is some supernatural being even part of the equation?  The only reason we have the opinion that we don't know there is no god is<mask> of primitive superstitions which evolved and unscrupulous men who learned to take advantage of others through their fear.  There are infinite alternative explanations for<mask> set off the big bang, the vast majority of which we will never hear about and have never even occurred to anyone.  Most of these have the exact same likelyhood<mask> a supernatural instigator, some have more. [NEWLINE] [NEWLINE] To me, the abrahamic gods are equally likely<mask> a leprichaun named Semus.  The fact that more people believe in god is irrelevent.</s>
Label encoding: <s>I understand, and I think most people agree with you.  In fact I agree with you. [NEWLINE] [NEWLINE] Here's what hangs me up.  Why is some supernatural being even part of the equation?  The only reason we have the opinion that we don't know there is no god is because of primitive superstitions which evolved and unscrupulous men who learned to take advantage of others through their fear.  There are infinite alternative explanations for what set off the big bang, the vast majority of which we will never hear about and have never even occurred to anyone.  Most of these have the exact same likelyhood as a supernatural instigator, some have more. [NEWLINE] [NEWLINE] To me, the abrahamic gods are equally likely as a leprichaun named Semus.  The fact that more people believe in god is irrelevent.</s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My point is precisely that it's not always good to treat everyone "the same" - you sometimes need to account for the difficulties that people may face<mask> they don't have the privileges that you do.<mask> you're not even aware of your privilege, you can't do this. [NEWLINE] [NEWLINE] Example: yesterday a woman friend of mine was venting to me<mask> some guy randomly offered to pay for the snack she was buying at the grocery store.<mask> I don’t check my male privilege, it'd be easy for me to assume that she was overreacting to a guy who was just being nice. In reality, it's more likely that this guy was invading her personal space and giving other signs of being a creep, and she probably has to deal with this kind of thing on a regular basis and is sick of it. </s>
Label encoding: <s>My point is precisely that it's not always good to treat everyone "the same" - you sometimes need to account for the difficulties that people may face because they don't have the privileges that you do. If you're not even aware of your privilege, you can't do this. [NEWLINE] [NEWLINE] Example: yesterday a woman friend of mine was venting to me because some guy randomly offered to pay for the snack she was buying at the grocery store. If I don’t check my male privilege, it'd be easy for me to assume that she was overreacting to a guy who was just being nice. In reality, it's more likely that this guy was invading her personal space and giving other signs of being a creep, and she probably has to deal with this kind of thing on a regular basis and is sick of it. </s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry TanithRosenbaum, your submission has been removed: [NEWLINE] [NEWLINE] [STARTQ] Submission Rule E\. "Only post<mask> you are willing to have a conversation with those who reply to you, and are available to do<mask> within 3 hours after posting.<mask> you haven't replied within this time, your post will be removed." [See the wiki for more information.]( [URL] #wiki_rule_e). [ENDQ] [NEWLINE] <mask> you would like to appeal, please respond to some of the arguments people have made, and then [message the moderators by clicking this link.]( [URL] ;subject=Removed+Submission+Rule+E+Post+Appeal&amp;message=TanithRosenbaum+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] /\))</s>
Label encoding: <s>Sorry TanithRosenbaum, your submission has been removed: [NEWLINE] [NEWLINE] [STARTQ] Submission Rule E\. "Only post if you are willing to have a conversation with those who reply to you, and are available to do so within 3 hours after posting. If you haven't replied within this time, your post will be removed." [See the wiki for more information.]( [URL] #wiki_rule_e). [ENDQ] [NEWLINE] If you would like to appeal, please respond to some of the arguments people have made, and then [message the moderators by clicking this link.]( [URL] ;subject=Removed+Submission+Rule+E+Post+Appeal&amp;message=TanithRosenbaum+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] /\))</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Its come up<mask> often<mask> my dad's divorce. Which is to say, it hasn't. Dad's still divorced, aunt's still gay, nothing new to report on that front.<mask> are they going to say? "Hey Becky, you still gay?" "Yup" "Dangit" [NEWLINE] [NEWLINE] My family is pretty geographically dispersed and we don't all get together often. I can count on one hand the number of times my dad, aunt, grandma and I have all been in the same place at the same time. There hasn't been a lot of opportunity to just talk about this stuff in the first place. And<mask>, at this point my aunt is well into her 60s. She's a grown woman who can take care of herself, she doesn't need me meddling in her relationship with her mother. </s>
Label encoding: <s>Its come up as often as my dad's divorce. Which is to say, it hasn't. Dad's still divorced, aunt's still gay, nothing new to report on that front. What are they going to say? "Hey Becky, you still gay?" "Yup" "Dangit" [NEWLINE] [NEWLINE] My family is pretty geographically dispersed and we don't all get together often. I can count on one hand the number of times my dad, aunt, grandma and I have all been in the same place at the same time. There hasn't been a lot of opportunity to just talk about this stuff in the first place. And besides, at this point my aunt is well into her 60s. She's a grown woman who can take care of herself, she doesn't need me meddling in her relationship with her mother. </s>
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This argument can easily be carried to the point of absurdity.  The fact that something will eventually become a person does not make it a person with feelings and experiences.  A flock of chickens left for millions of years might well evolve intelligence<mask> nobody considers that<mask> deciding whether or not to eat them. [NEWLINE] [NEWLINE] <mask> we are going to place significant value on the potential lives of people, then it should follow that it is immoral not to constantly have all the unprotected sex that you possibly can in order to generate<mask> many people<mask> possible.  There is no moral difference between contraception and abortion.  the only difference is that one takes place before two haploid cells become a single diploid cell, and other takes place after. <mask> is it about this event, this genetic roll of the dice, that makes the new cell more significant?</s>
Label encoding: <s>This argument can easily be carried to the point of absurdity.  The fact that something will eventually become a person does not make it a person with feelings and experiences.  A flock of chickens left for millions of years might well evolve intelligence but nobody considers that when deciding whether or not to eat them. [NEWLINE] [NEWLINE] If we are going to place significant value on the potential lives of people, then it should follow that it is immoral not to constantly have all the unprotected sex that you possibly can in order to generate as many people as possible.  There is no moral difference between contraception and abortion.  the only difference is that one takes place before two haploid cells become a single diploid cell, and other takes place after.  What is it about this event, this genetic roll of the dice, that makes the new cell more significant?</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My second delta was a rather long post discussing the nature of the human brain and demonstrating that modern science does have some understanding of<mask> the brain works, and that through this understanding we can<mask><mask> have some meaningful effect on it via medication. [NEWLINE] [NEWLINE] <mask><mask> the problem here is being able to discern someone who needs a thorough explanation of an issue from a troll. Some people hand out deltas relatively quickly, and I'm generally too late for that sort of discussion. Other people aren't looking to actually have their views changed,<mask> it doesn't matter<mask> you'll tell them. This means that the people who could have their views changed,<mask> only the data was presented to them clearly and completely, don't get the sort of attention here that they might need, and<mask> they get it, it's not terribly visible. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>My second delta was a rather long post discussing the nature of the human brain and demonstrating that modern science does have some understanding of how the brain works, and that through this understanding we can in fact have some meaningful effect on it via medication. [NEWLINE] [NEWLINE] I think the problem here is being able to discern someone who needs a thorough explanation of an issue from a troll. Some people hand out deltas relatively quickly, and I'm generally too late for that sort of discussion. Other people aren't looking to actually have their views changed, so it doesn't matter what you'll tell them. This means that the people who could have their views changed, if only the data was presented to them clearly and completely, don't get the sort of attention here that they might need, and when they get it, it's not terribly visible. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If the unemployed person is actively working to not get work or even try, then yes fuck them and let's make a hate sub for them.<mask> this is about the Tess Monsters of the world who actively make themselves sick and weak and encourage others to follow or not be disgusted by their fucking disgusting, horrible bodies and attitudes. I was a 320+ pound guy who decided that I didn't want to die young from my weight or be kept from doing things (or force people to accommodate me) that I wanted to do. I am still a big guy<mask> it's coming off through discipline and the fact that I HATE BEING FAT. I don't care about people that don't want to, I hate them for not taking responsibility and trying to make it other people's faults and trying to warp society to literally fit them in. </s>
Label encoding: <s>If the unemployed person is actively working to not get work or even try, then yes fuck them and let's make a hate sub for them. But this is about the Tess Monsters of the world who actively make themselves sick and weak and encourage others to follow or not be disgusted by their fucking disgusting, horrible bodies and attitudes. I was a 320+ pound guy who decided that I didn't want to die young from my weight or be kept from doing things (or force people to accommodate me) that I wanted to do. I am still a big guy but it's coming off through discipline and the fact that I HATE BEING FAT. I don't care about people that don't want to, I hate them for not taking responsibility and trying to make it other people's faults and trying to warp society to literally fit them in. </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I didn't really think through all the later implications, simply the fact that the principle of incertitude make initial (or stopping, by stopping I mean "at a given time") conditions impossible to know, from a physical point of view. [ENDQ] [NEWLINE] I'm not sure<mask> that matters. A deterministic system is deterministic<mask><mask> whether entities from within the system can predict its behavior. [NEWLINE] [NEWLINE] In general, a machine cannot simulate itself faster than it runs,<mask> even<mask> we "could" know absolutely everything about initial and current conditions, reliably computing events T seconds from now would almost systemically take more than T seconds. In other words, even<mask> there were no uncertainty principle and no quantum physics, predicting the future precisely would *still* be physically impossible. Would that suffice to make the universe not deterministic? [NEWLINE] </s>
Label encoding: <s> [STARTQ] I didn't really think through all the later implications, simply the fact that the principle of incertitude make initial (or stopping, by stopping I mean "at a given time") conditions impossible to know, from a physical point of view. [ENDQ] [NEWLINE] I'm not sure why that matters. A deterministic system is deterministic regardless of whether entities from within the system can predict its behavior. [NEWLINE] [NEWLINE] In general, a machine cannot simulate itself faster than it runs, so even if we "could" know absolutely everything about initial and current conditions, reliably computing events T seconds from now would almost systemically take more than T seconds. In other words, even if there were no uncertainty principle and no quantum physics, predicting the future precisely would *still* be physically impossible. Would that suffice to make the universe not deterministic? [NEWLINE] </s>
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0780, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See, this sounds great on paper,<mask> 90% of kids don't give a shit about anything more than<mask>'s directly in front of them.<mask> you say you're teaching kids to think for themselves,<mask>'s actually going to happen is they will just not think *at all*. The teenagers and young adults who just blindly follow religion<mask> it's<mask> their parents taught them, will be replaced by teenagers and young adults who never bothered to think too much about it,<mask> they didn't care. [NEWLINE] [NEWLINE] It's very idealistic to believe we'd all be more enlightened<mask> we didn't push religion down our children's throats. I contend that no one would be any'smarter' in that scenario. Mind you I'm not arguing the world wouldn't be better off without religion- that's a completely different argument.</s>
Label encoding: <s>See, this sounds great on paper, but 90% of kids don't give a shit about anything more than what's directly in front of them. When you say you're teaching kids to think for themselves, what's actually going to happen is they will just not think *at all*. The teenagers and young adults who just blindly follow religion because it's what their parents taught them, will be replaced by teenagers and young adults who never bothered to think too much about it, because they didn't care. [NEWLINE] [NEWLINE] It's very idealistic to believe we'd all be more enlightened if we didn't push religion down our children's throats. I contend that no one would be any'smarter' in that scenario. Mind you I'm not arguing the world wouldn't be better off without religion- that's a completely different argument.</s>
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You make a lot of good points. And<mask><mask> Chomsky of a few decades ago was probably worthy of this eloquent defense.<mask> he's grown more strident and less scholarly over time, with no apparent lessening of his prestige and influence. [NEWLINE] [NEWLINE] I already posted this in response to another post; Christopher Hitchens take down of Chomsky, using his "scholarship" about 9/11<mask> a very strong case in point. To be honest, I forgot that I'd seen this a few years ago, I believe it was<mask> my view of Chomsky began to change. It's illuminating. Saying 9/11 was no greater an atrocity than Clinton's use of air strikes in the Sudan, which Chomsky said, is exactly<mask> ridiculous and unfounded<mask> anything you are likely to hear on Fox News: [NEWLINE] [NEWLINE] [URL] </s>
Label encoding: <s>You make a lot of good points. And I think Chomsky of a few decades ago was probably worthy of this eloquent defense. But he's grown more strident and less scholarly over time, with no apparent lessening of his prestige and influence. [NEWLINE] [NEWLINE] I already posted this in response to another post; Christopher Hitchens take down of Chomsky, using his "scholarship" about 9/11 as a very strong case in point. To be honest, I forgot that I'd seen this a few years ago, I believe it was when my view of Chomsky began to change. It's illuminating. Saying 9/11 was no greater an atrocity than Clinton's use of air strikes in the Sudan, which Chomsky said, is exactly as ridiculous and unfounded as anything you are likely to hear on Fox News: [NEWLINE] [NEWLINE] [URL] </s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And I "should" have a majillion dollars. And I "should" be Spider-Man. [NEWLINE] [NEWLINE] We can talk all day about<mask> should be. And I *agree*.<mask> reality doesn't care. The only reason we have rights is<mask> we all agree we should, and no one more powerful is currently taking them away. Smokers *shouldn't* get cancer. I should be able to smoke<mask> I want and not get lung cancer.<mask> I will<mask> I keep smoking. [NEWLINE] [NEWLINE] At any rate, the first point is moot,<mask> there isn't actually any correlation between attire and rape. Most rapes are committed by family members.<mask> the point then becomes yes, people can wear whatever they want,<mask> it isn't hurting anything. [NEWLINE] [NEWLINE] For the second two, my point stands.</s>
Label encoding: <s>And I "should" have a majillion dollars. And I "should" be Spider-Man. [NEWLINE] [NEWLINE] We can talk all day about what should be. And I *agree*. But reality doesn't care. The only reason we have rights is because we all agree we should, and no one more powerful is currently taking them away. Smokers *shouldn't* get cancer. I should be able to smoke if I want and not get lung cancer. But I will if I keep smoking. [NEWLINE] [NEWLINE] At any rate, the first point is moot, because there isn't actually any correlation between attire and rape. Most rapes are committed by family members. So the point then becomes yes, people can wear whatever they want, because it isn't hurting anything. [NEWLINE] [NEWLINE] For the second two, my point stands.</s>
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The way I like to think about it is that historically oppressed groups must be afforded more rhetorical tools to make their case<mask> they are already lacking many things that the historically oppressive group has (X privilege). [NEWLINE] [NEWLINE] That is<mask> it is okay to say "Yes all women"<mask> not okay to say "Yes all white people." [NEWLINE] [NEWLINE] <mask> an afterthought, I don't live in fear of crime and I'm white.<mask> "Yes all white people" is just incorrect in this particular case. Furthermore, I'm not sure that "all women live in fear of being raped."<mask> feminist groups generally say is that all women have been victims of cat-calling or denigration based on their appearance by both strangers and relative strangers.<mask> I understand it correctly, that is<mask> "Yes All Women" usually refers to.</s>
Label encoding: <s>The way I like to think about it is that historically oppressed groups must be afforded more rhetorical tools to make their case since they are already lacking many things that the historically oppressive group has (X privilege). [NEWLINE] [NEWLINE] That is why it is okay to say "Yes all women" but not okay to say "Yes all white people." [NEWLINE] [NEWLINE] As an afterthought, I don't live in fear of crime and I'm white. So "Yes all white people" is just incorrect in this particular case. Furthermore, I'm not sure that "all women live in fear of being raped." What feminist groups generally say is that all women have been victims of cat-calling or denigration based on their appearance by both strangers and relative strangers. If I understand it correctly, that is what "Yes All Women" usually refers to.</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0698, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>* No one argues that chewing tobacco will give you lung cancer<mask> no one could<mask><mask> it's good for you. I can make that argument with marijuana. [NEWLINE] [NEWLINE] * "Cannabis use typically begins during adolescence and early adulthood, a period<mask> cannabinoid receptors are still abundant in white matter pathways across the brain." &amp; "Delaying the age at which regular use begins may minimize the severity of microstructural impairment." I'd be interested to see the full article, do you know the age spread on their population? [NEWLINE] [NEWLINE] * That last one was a stretch, I fully agree that exercise is the most cost effective (and smart) thing for these benefits.<mask>,<mask><mask> we both know that I was referring to other drugs (illegal &amp; over the counter) and other material quick fixes. [NEWLINE] </s>
Label encoding: <s>* No one argues that chewing tobacco will give you lung cancer but no one could argue that it's good for you. I can make that argument with marijuana. [NEWLINE] [NEWLINE] * "Cannabis use typically begins during adolescence and early adulthood, a period when cannabinoid receptors are still abundant in white matter pathways across the brain." &amp; "Delaying the age at which regular use begins may minimize the severity of microstructural impairment." I'd be interested to see the full article, do you know the age spread on their population? [NEWLINE] [NEWLINE] * That last one was a stretch, I fully agree that exercise is the most cost effective (and smart) thing for these benefits. However, I think we both know that I was referring to other drugs (illegal &amp; over the counter) and other material quick fixes. [NEWLINE] </s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] there are myriads of possible explanations for our world, and evidence does not necessarily lead us to the truth (see 3-second-analogy). [ENDQ] [NEWLINE] There is no evidence to believe our memories are only 3 seconds long,<mask><mask> it is a possibility. Just like it may be possible for an inter-dimensional duck to have created the universe, there is no reason to believe it. And that is the same<mask> god. You can get rid of contradictions<mask> it is possible, you can change/interpret passages to fit with<mask> we now know about the universe, and change them again<mask> /<mask> our understanding grows,<mask> that still doesn't provide any good reason to believe it. From the sounds of it, you are basically renouncing the importance of evidence, which is at least honest of you. </s>
Label encoding: <s> [STARTQ] there are myriads of possible explanations for our world, and evidence does not necessarily lead us to the truth (see 3-second-analogy). [ENDQ] [NEWLINE] There is no evidence to believe our memories are only 3 seconds long, even though it is a possibility. Just like it may be possible for an inter-dimensional duck to have created the universe, there is no reason to believe it. And that is the same as god. You can get rid of contradictions so it is possible, you can change/interpret passages to fit with what we now know about the universe, and change them again if / when our understanding grows, but that still doesn't provide any good reason to believe it. From the sounds of it, you are basically renouncing the importance of evidence, which is at least honest of you. </s>
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0679, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're missing the key idea of medicating people with learning disabilities and attention deficiencies: having them take medication does not give them an extra advantage, it simply puts them up on the level that everybody else is already on. It levels the playing field. For many people with ADHD, focusing on their work without medication is not just unpleasant, it is impossible.<mask><mask><mask><mask>,<mask> you give these same drugs to someone who has not been diagnosed with attention deficit, it *does* give them a decided advantage. That *would* make college easier for you, and that wouldn't be fair to everyone else, including those with learning disabilities. [NEWLINE] [NEWLINE] That said, you may be dealing with a mild form of undiagnosed ADD. See a psychologist<mask> you think you might want to get pharmaceutical help for that.</s>
Label encoding: <s>You're missing the key idea of medicating people with learning disabilities and attention deficiencies: having them take medication does not give them an extra advantage, it simply puts them up on the level that everybody else is already on. It levels the playing field. For many people with ADHD, focusing on their work without medication is not just unpleasant, it is impossible. On the other hand, when you give these same drugs to someone who has not been diagnosed with attention deficit, it *does* give them a decided advantage. That *would* make college easier for you, and that wouldn't be fair to everyone else, including those with learning disabilities. [NEWLINE] [NEWLINE] That said, you may be dealing with a mild form of undiagnosed ADD. See a psychologist if you think you might want to get pharmaceutical help for that.</s>
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would say you are wrong about there not being objectively correct answers to moral questions. I am a utilitarian and constructivist. Ethics is built from logic, and everyone (or anything) who is rationally thinking using logic will arrive at the same conclusion<mask> to<mask> is moral. It is very difficult to be able to do that, and not very many people even try, which is<mask> there is not a huge amount of conscious.<mask>, I am not saying that<mask> I believe is the truth, only that it is the closest thing to the truth that I know.<mask> to say there there is no objective truth about ethics is to say that multiple different things can be correct. These different things are<mask> correct even<mask> they contradict each other (which is<mask> makes them different). This is a logically nonsensical statement.</s>
Label encoding: <s>I would say you are wrong about there not being objectively correct answers to moral questions. I am a utilitarian and constructivist. Ethics is built from logic, and everyone (or anything) who is rationally thinking using logic will arrive at the same conclusion as to what is moral. It is very difficult to be able to do that, and not very many people even try, which is why there is not a huge amount of conscious. Also, I am not saying that what I believe is the truth, only that it is the closest thing to the truth that I know. But to say there there is no objective truth about ethics is to say that multiple different things can be correct. These different things are therefore correct even as they contradict each other (which is what makes them different). This is a logically nonsensical statement.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi, [NEWLINE] [NEWLINE] Agreed that polar bears are a much more efficient way to raise money for environmentalists. This is really my point<mask><mask> : it's marketing, nothing bad for people except for those who have a polar bear among their friends, or who, more seriously, depend on polar bears, to eat or to make money (study polar bears, make pictures of polar bears, etc.).<mask> that, zero impact. [NEWLINE] [NEWLINE] <mask> for the IKEA example, there are tons of new species every day, and tons of old species that disappear. It is<mask><mask>, in your example, you lived close to IKEA and they dumped tons of spare parts every day in the dump. You would probably not keep your spare parts,<mask> you know there is a natural supply of spare parts every day.</s>
Label encoding: <s>Hi, [NEWLINE] [NEWLINE] Agreed that polar bears are a much more efficient way to raise money for environmentalists. This is really my point in fact : it's marketing, nothing bad for people except for those who have a polar bear among their friends, or who, more seriously, depend on polar bears, to eat or to make money (study polar bears, make pictures of polar bears, etc.). Besides that, zero impact. [NEWLINE] [NEWLINE] As for the IKEA example, there are tons of new species every day, and tons of old species that disappear. It is as if, in your example, you lived close to IKEA and they dumped tons of spare parts every day in the dump. You would probably not keep your spare parts, as you know there is a natural supply of spare parts every day.</s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> people who have previously scored well, are going to do better in the minds of the judges, you can't change it. [ENDQ] [NEWLINE] And you attribute this to biased judging rather than the top scorers are simply the best performers? I know that figure skating has overhauled their judging criteria for the past two Olympics. They now require skaters to perform certain elements and have set deductions for falls and imperfect landings on jumps<mask> assigning point values to a jump<mask><mask> who performs it. The points are increased in the second half of the long performances to reward athletes who are able to perform harder maneuvers towards the end of their performance<mask> they are beginning to tire. There is<mask> more than one judge and the top and bottom scores are disregarded. Your concern is not lost on the governing bodies of these sports.</s>
Label encoding: <s> [STARTQ] because people who have previously scored well, are going to do better in the minds of the judges, you can't change it. [ENDQ] [NEWLINE] And you attribute this to biased judging rather than the top scorers are simply the best performers? I know that figure skating has overhauled their judging criteria for the past two Olympics. They now require skaters to perform certain elements and have set deductions for falls and imperfect landings on jumps while assigning point values to a jump regardless of who performs it. The points are increased in the second half of the long performances to reward athletes who are able to perform harder maneuvers towards the end of their performance when they are beginning to tire. There is also more than one judge and the top and bottom scores are disregarded. Your concern is not lost on the governing bodies of these sports.</s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The point isn't to let people off the hook, it's to put them on the hook at all times. [ENDQ] [NEWLINE] No one can effectively live on the hook at all times. This promotes anxiety and depression<mask> accepted and outright hostility<mask> opposed. You want these people to help others.<mask> you destroy the base from which they operate through guilt tripping and shaming them in a way they cannot escape, they will be unable or unwilling to do<mask>. [NEWLINE] [NEWLINE] The problem is the disadvantages that certain groups face. The problem is *not* that other groups don't face these disadvantages. Framing the discussion in the second sense,<mask> any framing based on privilege does, focuses our efforts in the wrong direction. [NEWLINE] [NEWLINE] Equality achieved by bringing everyone down to the lowest common denominator is well worth opposing.</s>
Label encoding: <s> [STARTQ] The point isn't to let people off the hook, it's to put them on the hook at all times. [ENDQ] [NEWLINE] No one can effectively live on the hook at all times. This promotes anxiety and depression when accepted and outright hostility when opposed. You want these people to help others. If you destroy the base from which they operate through guilt tripping and shaming them in a way they cannot escape, they will be unable or unwilling to do so. [NEWLINE] [NEWLINE] The problem is the disadvantages that certain groups face. The problem is *not* that other groups don't face these disadvantages. Framing the discussion in the second sense, as any framing based on privilege does, focuses our efforts in the wrong direction. [NEWLINE] [NEWLINE] Equality achieved by bringing everyone down to the lowest common denominator is well worth opposing.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Obviously, the pirate is not taking money,<mask><mask> you are pirating some media, it is safe to assume you are doing<mask><mask> you enjoy that media. <mask> you enjoy that media, you'd be willing to pay for it. [NEWLINE] [NEWLINE] <mask> there is a movie I really like, like the Avengers,<mask> an example, I'd be willing to buy it<mask> I could own my own copy.  Let's say I'd spend $10 just to keep it simple.  I find out I can download a pirated version of the movie, with no discernible difference in quality vs the original.  Now, rather than my money making its way to the creators of the movie, I have spent $0.  Now the creators have made less money.  Effectively taking their money. </s>
Label encoding: <s>Obviously, the pirate is not taking money, but if you are pirating some media, it is safe to assume you are doing so because you enjoy that media.  If you enjoy that media, you'd be willing to pay for it. [NEWLINE] [NEWLINE] If there is a movie I really like, like the Avengers, as an example, I'd be willing to buy it so I could own my own copy.  Let's say I'd spend $10 just to keep it simple.  I find out I can download a pirated version of the movie, with no discernible difference in quality vs the original.  Now, rather than my money making its way to the creators of the movie, I have spent $0.  Now the creators have made less money.  Effectively taking their money. </s>
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The following are of course only case by case solutions<mask> nothing stops the triad from having a marriage ritual not ratified by the government and agreeing that in all private *and* social events they will introduce each other<mask> their spouses. That at least would stop any feelings of inequality. [NEWLINE] [NEWLINE] Regarding your own N formation, wouldn't it make sense to at least form two married pairs? Or could CPS still take your children away for perceived 'unsavory' practices?<mask> that would be a whole different issue. [NEWLINE] [NEWLINE] <mask> yes. Polyamorous marriages would need the design of a new system (<mask> opposed to the simple fix that made same-sex marriage possible) and for that non-dyad polyamorists interested in poly marriage need to gain much more political acumen. And donation money.</s>
Label encoding: <s>The following are of course only case by case solutions but nothing stops the triad from having a marriage ritual not ratified by the government and agreeing that in all private *and* social events they will introduce each other as their spouses. That at least would stop any feelings of inequality. [NEWLINE] [NEWLINE] Regarding your own N formation, wouldn't it make sense to at least form two married pairs? Or could CPS still take your children away for perceived 'unsavory' practices? Because that would be a whole different issue. [NEWLINE] [NEWLINE] But yes. Polyamorous marriages would need the design of a new system ( as opposed to the simple fix that made same-sex marriage possible) and for that non-dyad polyamorists interested in poly marriage need to gain much more political acumen. And donation money.</s>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is a way in which science differs from religion even for people who have a reverential attitude toward science, which is that any specific scientific conclusion, whether it is considered to be a law of nature or any other type of conclusion about<mask> the world works, is subject to revision in light of new evidence and/or logical analysis. <mask>, science is not dogmatic.  Religion, in comparision, makes assertions which it is unwilling to change or reconsider in any way.  It is dogmatic.  Of course, some religions are more dogmatic than others, and to some extent, religion and science can resemble each other<mask> you consider the more liberal or flexible approach that some people take to religion, or for that matter, the more dogmatic approach that some people take to science.</s>
Label encoding: <s>There is a way in which science differs from religion even for people who have a reverential attitude toward science, which is that any specific scientific conclusion, whether it is considered to be a law of nature or any other type of conclusion about how the world works, is subject to revision in light of new evidence and/or logical analysis.  Thus, science is not dogmatic.  Religion, in comparision, makes assertions which it is unwilling to change or reconsider in any way.  It is dogmatic.  Of course, some religions are more dogmatic than others, and to some extent, religion and science can resemble each other if you consider the more liberal or flexible approach that some people take to religion, or for that matter, the more dogmatic approach that some people take to science.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0729, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ok let's start off with a quick thing, the spread of riches, white people are richer. Forbe's [sic] list of billionaires shows this pretty clearly. [ENDQ] [NEWLINE] <mask><mask> that's using historical bias<mask><mask> opposed to saying that whites have better opportunities than black. There are many deprived white families<mask> there are black. Sure, many people have inherited wealth, or have accumulated it,<mask> then again many have not. [NEWLINE] [NEWLINE] I understand (and may be incorrect) that the term white privilege means that<mask><mask> any other criteria, in any situation<mask> discrimination is exercised, the white person is preferable, just due to their skin colour. [NEWLINE] [NEWLINE] <mask> examining past wealth and a list of wealthy people is<mask> the question, no? [NEWLINE] [NEWLINE] Please do correct me<mask> I'm wrong!</s><pad>
Label encoding: <s> [STARTQ] Ok let's start off with a quick thing, the spread of riches, white people are richer. Forbe's [sic] list of billionaires shows this pretty clearly. [ENDQ] [NEWLINE] I think that's using historical bias though as opposed to saying that whites have better opportunities than black. There are many deprived white families as there are black. Sure, many people have inherited wealth, or have accumulated it, but then again many have not. [NEWLINE] [NEWLINE] I understand (and may be incorrect) that the term white privilege means that regardless of any other criteria, in any situation where discrimination is exercised, the white person is preferable, just due to their skin colour. [NEWLINE] [NEWLINE] Therefore examining past wealth and a list of wealthy people is besides the question, no? [NEWLINE] [NEWLINE] Please do correct me if I'm wrong!</s><pad>
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think it's the nature of the medical procedure that was the issue in *Hobby Lobby* (could be totally wrong on that,<mask> ). The argument was, in essence, whether Hobby Lobby had to pay for the objected-to medicine. The court said that making them pay for something against their beliefs violated RFRA<mask> there are less restrictive ways the government could ensure the medicine is distributed (e.g. gov't provided meds, employee paying for it, etc.). [NEWLINE] [NEWLINE] <mask><mask><mask><mask> was the issue, then the expense of the procedure doesn't factor in.<mask>, part of Ginsburg's dissent, IIRC, was the high costs of IUDs, which was given the ol' "psh". (&lt;- highly legal terminology there)</s>
Label encoding: <s>I don't think it's the nature of the medical procedure that was the issue in *Hobby Lobby* (could be totally wrong on that, though ). The argument was, in essence, whether Hobby Lobby had to pay for the objected-to medicine. The court said that making them pay for something against their beliefs violated RFRA because there are less restrictive ways the government could ensure the medicine is distributed (e.g. gov't provided meds, employee paying for it, etc.). [NEWLINE] [NEWLINE] ASSUMING that was the issue, then the expense of the procedure doesn't factor in. Indeed, part of Ginsburg's dissent, IIRC, was the high costs of IUDs, which was given the ol' "psh". (&lt;- highly legal terminology there)</s>
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the main misconception you have is that the people of the U.S would be fighting the U.S. military. Any armed revolt in the U.S. would be primarily between the people and various law enforcement agencies. The  Federal Government trying to use conventional military units to squash a revolt would cause a disintegration of the very military they are trying to use. The men and women serving would stop cooperating once they see their home towns turned into war zones. This could turn into a huge circle-jerk<mask> suffice it to say its not<mask> simple<mask> the people vs the U.S. Army. [NEWLINE] [NEWLINE] Edit:<mask><mask> we are 50 years or less from major armed conflict in the U.S.<mask><mask><mask> the ideological divide in this country continues to grow. </s><pad>
Label encoding: <s>I think the main misconception you have is that the people of the U.S would be fighting the U.S. military. Any armed revolt in the U.S. would be primarily between the people and various law enforcement agencies. The  Federal Government trying to use conventional military units to squash a revolt would cause a disintegration of the very military they are trying to use. The men and women serving would stop cooperating once they see their home towns turned into war zones. This could turn into a huge circle-jerk but suffice it to say its not as simple as the people vs the U.S. Army. [NEWLINE] [NEWLINE] Edit: IMO we are 50 years or less from major armed conflict in the U.S. as long as the ideological divide in this country continues to grow. </s><pad>
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, even this curmudgeon sometimes likes watching those portions of the Olympics that represent phenomenal individual acheivement. That brings me to something which may be more germane to you question or post. Americans like individual heros/role models and individual stories of acheivement, and have a harder time with chaotic team sports<mask> one has to be a die hard fan to even recognize the players or have a clue<mask> their contributions are. You can tune into something like gymnastics, or figure skating after a decade away from caring, and get a feel for raw talent on display within a couple hours. Not<mask> with soccer. I'm certainly NOT saying soccer doesn't take loads of talent.<mask>, it might not be<mask> readily apparent or<mask> wonderous to the extremely casual observer.</s>
Label encoding: <s>Well, even this curmudgeon sometimes likes watching those portions of the Olympics that represent phenomenal individual acheivement. That brings me to something which may be more germane to you question or post. Americans like individual heros/role models and individual stories of acheivement, and have a harder time with chaotic team sports where one has to be a die hard fan to even recognize the players or have a clue what their contributions are. You can tune into something like gymnastics, or figure skating after a decade away from caring, and get a feel for raw talent on display within a couple hours. Not so with soccer. I'm certainly NOT saying soccer doesn't take loads of talent. But, it might not be as readily apparent or as wonderous to the extremely casual observer.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me take a mathematical approach to this question.  Yes, everyone knows this. <mask> do you see everyone committing suicide?  No,<mask> your choices are<mask> follows: end it now and get it overwith, or don't, and keep your option.  Options have value, and<mask> you can't see the future, you don't know the future value of the option right now. <mask> its certainly greater than zero. <mask><mask> you are arguing that it is less than zero,<mask> options are never worth less than a single outcome of that option.  Now that we have a mathematical argument, I just want to say that nothing in life is easy,<mask> that's<mask> hard work pays off, both in terms of outcome and your own sense of self worth.  </s>
Label encoding: <s>Let me take a mathematical approach to this question.  Yes, everyone knows this.  But do you see everyone committing suicide?  No, because your choices are as follows: end it now and get it overwith, or don't, and keep your option.  Options have value, and since you can't see the future, you don't know the future value of the option right now.  But its certainly greater than zero.  I think you are arguing that it is less than zero, but options are never worth less than a single outcome of that option.  Now that we have a mathematical argument, I just want to say that nothing in life is easy, but that's why hard work pays off, both in terms of outcome and your own sense of self worth.  </s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I kind of agree with this<mask> you have to remember that "gay" has had different meanings throughout history. It originally meant you were joyous then some<mask> it turned into a word to describe homosexuals. Many words that are considered to be "bad" words had different meanings at one point.<mask> you call someone a dick or a cunt you don't actually think that person is genitalia.<mask> I called someone gay<mask> I was younger, did I actually think that person was a homosexual? [NEWLINE] [NEWLINE] It is different<mask> you call someone a derogatory word to their face such<mask> faggot and queer. You are actually making an attempt to bully that person. <mask> calling someone gay jokingly hurts no one. [NEWLINE] [NEWLINE] The real problem in my eyes is that we are too easily offended.</s>
Label encoding: <s>I kind of agree with this but you have to remember that "gay" has had different meanings throughout history. It originally meant you were joyous then some how it turned into a word to describe homosexuals. Many words that are considered to be "bad" words had different meanings at one point. When you call someone a dick or a cunt you don't actually think that person is genitalia. When I called someone gay when I was younger, did I actually think that person was a homosexual? [NEWLINE] [NEWLINE] It is different when you call someone a derogatory word to their face such as faggot and queer. You are actually making an attempt to bully that person.  While calling someone gay jokingly hurts no one. [NEWLINE] [NEWLINE] The real problem in my eyes is that we are too easily offended.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well obviously not all people are alike I still for the most part disagree with your statement. I'm one of those guys who has always been called a 'nice guy' and for a lot of my teenage years was like the guys being talked about her. [NEWLINE] [NEWLINE] <mask><mask> for most 'nice guys' it's not like some pre-planned "I'm gonna just be really nice, then she'll have to like me!!" thing. It's really just a  combination of lack of social skills and low self esteem, often caused by the aforementioned lack of social skills. Most guys spend a lot of time wishing they could me more forward, ask the girl out, make a move ect...<mask> get to nervous or shy and just end up being too nice<mask> their only way to interact. </s>
Label encoding: <s>Well obviously not all people are alike I still for the most part disagree with your statement. I'm one of those guys who has always been called a 'nice guy' and for a lot of my teenage years was like the guys being talked about her. [NEWLINE] [NEWLINE] I think for most 'nice guys' it's not like some pre-planned "I'm gonna just be really nice, then she'll have to like me!!" thing. It's really just a  combination of lack of social skills and low self esteem, often caused by the aforementioned lack of social skills. Most guys spend a lot of time wishing they could me more forward, ask the girl out, make a move ect... But get to nervous or shy and just end up being too nice as their only way to interact. </s>
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Ecator, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Ecator+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry Ecator, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Ecator+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> I can change your view,<mask> I don't have much life experience with this kind of thing, I am going to college this year and never had to experience real poverty. The only experience I ever had with welfare was watching my adult friends or neighbors look for work  and find nothing and start to have to worry<mask> they will be able to feed their kids or have a place to sleep. The argument against welfare is a strong one,<mask> lots of people who apply are just using it<mask> an excuse not to get a job.<mask> for the people who do need to in order to get by<mask><mask> it is worth it. To me it seems like throwing the baby out with the bathwater to cut welfare<mask> some people work hard and don't need it.</s>
Label encoding: <s>I don't know if I can change your view, as I don't have much life experience with this kind of thing, I am going to college this year and never had to experience real poverty. The only experience I ever had with welfare was watching my adult friends or neighbors look for work  and find nothing and start to have to worry if they will be able to feed their kids or have a place to sleep. The argument against welfare is a strong one, as lots of people who apply are just using it as an excuse not to get a job. But for the people who do need to in order to get by I think it is worth it. To me it seems like throwing the baby out with the bathwater to cut welfare because some people work hard and don't need it.</s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It might not be the question,<mask> just<mask> you choose. In most places you almost have to go out of your way to make a post that has mass down votes. In this sub, you really need to almost try to have something really down voted. [NEWLINE] [NEWLINE] Not every sub is the same. There are places<mask> you can ask things the are places<mask> that' frowned upon. [NEWLINE] [NEWLINE] I mean you almost have gauge the setting<mask> much<mask> decide<mask> you're going to ask your question. You can almost ask any question in a way that can offend and then same question in a way that won't offend.<mask> old were you<mask> you knew you were gay. vs.<mask> did you adopt the gay lifestyle? IT could just be<mask> you are doing the asking. </s>
Label encoding: <s>It might not be the question, but just where you choose. In most places you almost have to go out of your way to make a post that has mass down votes. In this sub, you really need to almost try to have something really down voted. [NEWLINE] [NEWLINE] Not every sub is the same. There are places where you can ask things the are places where that' frowned upon. [NEWLINE] [NEWLINE] I mean you almost have gauge the setting as much as decide how you're going to ask your question. You can almost ask any question in a way that can offend and then same question in a way that won't offend. How old were you when you knew you were gay. vs. When did you adopt the gay lifestyle? IT could just be how you are doing the asking. </s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Interesting points. [NEWLINE] [NEWLINE] To the first, I'm thinking, we may now be parsing<mask> a person CAN do legally, and<mask> h/she  SHOULD do.<mask><mask><mask>, just<mask> something CAN be done, it is not necessarily something that SHOULD be done. I object<mask> a person engages in behavior for which s/he will not bear the cost. [NEWLINE] [NEWLINE] To the second, I am NOT in favor of more government regulation. I AM in favor of people bearing the costs of their own decisions. Do you think people who make the effort to live a healthy lifestyle--healthy eating, reasonable exercise, no smoking or excessive drinking, etc.--should pay higher insurance premiums and taxes to "cover the costs" of people who, legally or illegally, harm themselves willingly?</s>
Label encoding: <s>Interesting points. [NEWLINE] [NEWLINE] To the first, I'm thinking, we may now be parsing what a person CAN do legally, and what h/she  SHOULD do. In my opinion, just because something CAN be done, it is not necessarily something that SHOULD be done. I object when a person engages in behavior for which s/he will not bear the cost. [NEWLINE] [NEWLINE] To the second, I am NOT in favor of more government regulation. I AM in favor of people bearing the costs of their own decisions. Do you think people who make the effort to live a healthy lifestyle--healthy eating, reasonable exercise, no smoking or excessive drinking, etc.--should pay higher insurance premiums and taxes to "cover the costs" of people who, legally or illegally, harm themselves willingly?</s>
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You don't have to let the derailment of 'it always costs money' and 'It's the same<mask> with taxes' pull you off point. [NEWLINE] [NEWLINE] You are right to believe that a right to health should exist in this world, wherein<mask> you are ill and others have the capacity to help you recover or remove your pain they should be obliged to do<mask>. Without that, the core benefit of forming social groups (that through mutual support we can achieve more than we could otherwise) falls away. [NEWLINE] [NEWLINE] Economic realities do not and should not dictate moral beliefs. It might be the situation that executing every prisoner was the only way we could afford to enforce the law; that doesn't make the state executing it's citizens a morally different act (I consider it a bad thing).</s>
Label encoding: <s>You don't have to let the derailment of 'it always costs money' and 'It's the same but with taxes' pull you off point. [NEWLINE] [NEWLINE] You are right to believe that a right to health should exist in this world, wherein if you are ill and others have the capacity to help you recover or remove your pain they should be obliged to do so. Without that, the core benefit of forming social groups (that through mutual support we can achieve more than we could otherwise) falls away. [NEWLINE] [NEWLINE] Economic realities do not and should not dictate moral beliefs. It might be the situation that executing every prisoner was the only way we could afford to enforce the law; that doesn't make the state executing it's citizens a morally different act (I consider it a bad thing).</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> the tl;dr of that point is you can't use the term "greedy" without knowing<mask> the money they're taking in compares to their expenses. [ENDQ] [NEWLINE] <mask> not?  I can see that they are making roughly double on my subscription<mask> of the adds<mask> offering little tangible increase in utility.  Mainly that I can watch it on a hand held computer vs a PC:) [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask> he has suggested hypothetically, it were the case that they're in the red with just subscription revenue and the ads are the only thing keeping them in the black, then that doesn't sound "greedy". [ENDQ] [NEWLINE] I don't think that is the case.  After-all they are enough in the black to produce original series:)</s><pad>
Label encoding: <s> [STARTQ] So the tl;dr of that point is you can't use the term "greedy" without knowing how the money they're taking in compares to their expenses. [ENDQ] [NEWLINE] Why not?  I can see that they are making roughly double on my subscription because of the adds while offering little tangible increase in utility.  Mainly that I can watch it on a hand held computer vs a PC:) [NEWLINE] [NEWLINE] [STARTQ] If, as he has suggested hypothetically, it were the case that they're in the red with just subscription revenue and the ads are the only thing keeping them in the black, then that doesn't sound "greedy". [ENDQ] [NEWLINE] I don't think that is the case.  After-all they are enough in the black to produce original series:)</s><pad>
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right, it's not perfect, and they're friggin finally starting to seriously move towards having a 110 limit on some newer motorways. [NEWLINE] [NEWLINE] My theory on those straight country roads is<mask> it's safe to speed then you'll be able to see with certainty there are no potential hazards by the side of the road, like police... [NEWLINE] [NEWLINE] We generally have pretty  good sign-posting for low-speed corners<mask> well,<mask> for people driving a long way that's a poor consolation for<mask> windy and narrow most roads are. [NEWLINE] [NEWLINE] They're<mask> *usually* pretty good at having visual cues for different speed limits, like narrower roads and<mask> on. [NEWLINE] [NEWLINE] Hope you spent some time travelling the south island and didn't just loaf around Auckland.</s>
Label encoding: <s>You're right, it's not perfect, and they're friggin finally starting to seriously move towards having a 110 limit on some newer motorways. [NEWLINE] [NEWLINE] My theory on those straight country roads is if it's safe to speed then you'll be able to see with certainty there are no potential hazards by the side of the road, like police... [NEWLINE] [NEWLINE] We generally have pretty  good sign-posting for low-speed corners as well, but for people driving a long way that's a poor consolation for how windy and narrow most roads are. [NEWLINE] [NEWLINE] They're also *usually* pretty good at having visual cues for different speed limits, like narrower roads and so on. [NEWLINE] [NEWLINE] Hope you spent some time travelling the south island and didn't just loaf around Auckland.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0716, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We've had a few family get-togethers over the years, and her being gay just never came up and everything was fine.<mask><mask>, I didn't even know my aunt was gay until the last few years<mask> my dad told me. It just never came up during conversations [ENDQ] [NEWLINE] Yeah, this isn't "fine." This is someone who is<mask> deeply ashamed of her daughter's homosexuality that it "just doesn't come up." [NEWLINE] [NEWLINE] SOURCE: Me, a gay man raised in a baptist home. [NEWLINE] [NEWLINE] I'm not saying you should fight about it,<mask> it isn't "fine" that someone has to lead their life in a closet.<mask><mask>, it is a horrible tragedy that gives your Aunt much suffering, no doubt. </s><pad>
Label encoding: <s> [STARTQ] We've had a few family get-togethers over the years, and her being gay just never came up and everything was fine. In fact, I didn't even know my aunt was gay until the last few years when my dad told me. It just never came up during conversations [ENDQ] [NEWLINE] Yeah, this isn't "fine." This is someone who is so deeply ashamed of her daughter's homosexuality that it "just doesn't come up." [NEWLINE] [NEWLINE] SOURCE: Me, a gay man raised in a baptist home. [NEWLINE] [NEWLINE] I'm not saying you should fight about it, but it isn't "fine" that someone has to lead their life in a closet. In fact, it is a horrible tragedy that gives your Aunt much suffering, no doubt. </s><pad>
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry, I don't quite follow<mask> you're trying to say with the last sentence? [NEWLINE] [NEWLINE] [STARTQ] <mask> you get creeped out by them and don't want to invite them to a ceremony just<mask> of who they are,<mask> you think they are just<mask> valid<mask> heterosexual relationships, except<mask> it comes to having a personal ceremony to commemorate their love for each other. They don't have a right to that. [ENDQ] [NEWLINE] <mask> I've said several times, its just their actions that they may do at my wedding that may creep me out and I'm not sure that I want that.<mask> a lot of people here may be convincing me that its time just to expose myself more to situations that scare me. Then maybe I'll reach a point of acceptance.</s>
Label encoding: <s>I'm sorry, I don't quite follow what you're trying to say with the last sentence? [NEWLINE] [NEWLINE] [STARTQ] So you get creeped out by them and don't want to invite them to a ceremony just because of who they are, but you think they are just as valid as heterosexual relationships, except when it comes to having a personal ceremony to commemorate their love for each other. They don't have a right to that. [ENDQ] [NEWLINE] AS I've said several times, its just their actions that they may do at my wedding that may creep me out and I'm not sure that I want that. But a lot of people here may be convincing me that its time just to expose myself more to situations that scare me. Then maybe I'll reach a point of acceptance.</s>
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As an armchair historian, I have a pretty good sense of the issues that led to the US Civil War. Over and over again, I hear from people--who I assume to be Southerners--that slavery was really a secondary issue and that the South went to war due to infringement of their states' rights. I tend to think that these opinions are the product of intellectual gymnastics on the part of Southern culture to maintain the narrative that the were justified in rebelling and not the "bad" side. This narrative is taught to Southerners in public schools and<mask><mask> it should stop. In general,<mask><mask> that the South should look upon their rebellion with a sense of contrition for<mask> their ancestors did instead of hero worshiping the Confederacy. </s><pad>
Label encoding: <s>As an armchair historian, I have a pretty good sense of the issues that led to the US Civil War. Over and over again, I hear from people--who I assume to be Southerners--that slavery was really a secondary issue and that the South went to war due to infringement of their states' rights. I tend to think that these opinions are the product of intellectual gymnastics on the part of Southern culture to maintain the narrative that the were justified in rebelling and not the "bad" side. This narrative is taught to Southerners in public schools and I think it should stop. In general, I think that the South should look upon their rebellion with a sense of contrition for what their ancestors did instead of hero worshiping the Confederacy. </s><pad>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Point of Inquiry / Clarification:** [NEWLINE] [NEWLINE] Your position is rather nebulous in your original statement. You cite reactions that are bridges too far in either direction,<mask> never take a stand<mask> to<mask> is the correct reaction.<mask>,<mask> could anyone properly engage your position<mask> you've only gone<mask><mask><mask> to declare yourself a moderate between two very extreme poles,<mask> no materialized example of<mask> your correct amount of support would be. Could you list some appropriate acts of support in your view,<mask> that we could understand<mask> you stand? [NEWLINE] [NEWLINE] <mask> shooting people dead for an offensive cartoon is unacceptable,<mask><mask> donating money to print the publication in the future is the other extreme.<mask>'s your prescription for engaging the publication given the events that have unfolded?</s>
Label encoding: <s>**Point of Inquiry / Clarification:** [NEWLINE] [NEWLINE] Your position is rather nebulous in your original statement. You cite reactions that are bridges too far in either direction, but never take a stand as to what is the correct reaction. Therefore, how could anyone properly engage your position when you've only gone as far as to declare yourself a moderate between two very extreme poles, but no materialized example of what your correct amount of support would be. Could you list some appropriate acts of support in your view, so that we could understand where you stand? [NEWLINE] [NEWLINE] So shooting people dead for an offensive cartoon is unacceptable, while also donating money to print the publication in the future is the other extreme. What's your prescription for engaging the publication given the events that have unfolded?</s>
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just want to say that that's a very interesting post. [NEWLINE] [NEWLINE] <mask>, one version of morality represents our current understood way of living together to maximize well-being for people. We've concluded that harming others physically is bad for a number of reasons, even<mask> we were small tribes. Do onto others<mask> they would do onto you is one of the best ways to deal with people. For this rule to be overridden by logic, it would take some tremendous like massive overpopulation or some other strange event. [NEWLINE] [NEWLINE] In other words, this scenario seems more like a programming error than a problem with logic based morality.<mask> maybe we do only need 2-3 guys and tons of women to reach Jupiter... there's only one way to find out. BRB</s>
Label encoding: <s>Just want to say that that's a very interesting post. [NEWLINE] [NEWLINE] However, one version of morality represents our current understood way of living together to maximize well-being for people. We've concluded that harming others physically is bad for a number of reasons, even when we were small tribes. Do onto others as they would do onto you is one of the best ways to deal with people. For this rule to be overridden by logic, it would take some tremendous like massive overpopulation or some other strange event. [NEWLINE] [NEWLINE] In other words, this scenario seems more like a programming error than a problem with logic based morality. But maybe we do only need 2-3 guys and tons of women to reach Jupiter... there's only one way to find out. BRB</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a different political ideology that is probably a better fit for a strong, aggressive military than socialism. [NEWLINE] [NEWLINE] A highly nationalist ideology that is associated with surveillance on citizens<mask><mask> the government does not exist to serve workers,<mask> with socialism,<mask> primarily serve business and political leaders (with the assumption that this will be best for the nation): *Fascism*. [NEWLINE] [NEWLINE] In fascism, the objective of "defense" spending isn't to create a defense for the nation at all. It's to glorify the nation and subjugate peoples viewed<mask> inferior. It is not meant to be a public service. [NEWLINE] [NEWLINE] And the US military, having regularly invaded or attacked other nations without provocation for over a century, demonstrates a long history of just such aggression.</s>
Label encoding: <s>There's a different political ideology that is probably a better fit for a strong, aggressive military than socialism. [NEWLINE] [NEWLINE] A highly nationalist ideology that is associated with surveillance on citizens but where the government does not exist to serve workers, as with socialism, but primarily serve business and political leaders (with the assumption that this will be best for the nation): *Fascism*. [NEWLINE] [NEWLINE] In fascism, the objective of "defense" spending isn't to create a defense for the nation at all. It's to glorify the nation and subjugate peoples viewed as inferior. It is not meant to be a public service. [NEWLINE] [NEWLINE] And the US military, having regularly invaded or attacked other nations without provocation for over a century, demonstrates a long history of just such aggression.</s>
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Have you visited the 56 trillion other planets in the universe? No? Then<mask> are you<mask> sure we're the only one with intelligent life? [ENDQ] [NEWLINE] No, haha, I haven't. I admit that my knowledge about outer space is very limited and I didn't know that there are 56 trillion other planets. [NEWLINE] [NEWLINE] My view isn't completely changed<mask> now I'm not 100% sure that I'm right about the non - existence of intelligent life on other planets. Good job. [NEWLINE] [NEWLINE] Edited to include the &amp;#8710; [NEWLINE] [NEWLINE] Edit #2: /u/deltabot seems to have missed the &amp;#8710; that I gave. Hopefully this edit still get you the delta you deserve. </s>
Label encoding: <s> [STARTQ] Have you visited the 56 trillion other planets in the universe? No? Then how are you so sure we're the only one with intelligent life? [ENDQ] [NEWLINE] No, haha, I haven't. I admit that my knowledge about outer space is very limited and I didn't know that there are 56 trillion other planets. [NEWLINE] [NEWLINE] My view isn't completely changed but now I'm not 100% sure that I'm right about the non - existence of intelligent life on other planets. Good job. [NEWLINE] [NEWLINE] Edited to include the &amp;#8710; [NEWLINE] [NEWLINE] Edit #2: /u/deltabot seems to have missed the &amp;#8710; that I gave. Hopefully this edit still get you the delta you deserve. </s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0797, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah,<mask><mask>. Personally,<mask> I'm happy with my relationship it's very easy for me to be monogamous. I won't even think twice. [NEWLINE] [NEWLINE] On other hand<mask> I'm not happy, I'm very likely looking for that next one. Typically, this happens a month or two into dating someone once I get over the relationship's novelty. I don't think this means anything about my ability to be monogamous. I choose to be monogamous<mask> I feel like it and<mask><mask> it's a very healthy attitude. [NEWLINE] [NEWLINE] You can stay stuck in a relationship for years being monogamous for<mask> many different reasons,<mask> I don't think that says much about your potential for being monogamous in the next one. Or vice versa.</s>
Label encoding: <s>Yeah, I agree. Personally, if I'm happy with my relationship it's very easy for me to be monogamous. I won't even think twice. [NEWLINE] [NEWLINE] On other hand if I'm not happy, I'm very likely looking for that next one. Typically, this happens a month or two into dating someone once I get over the relationship's novelty. I don't think this means anything about my ability to be monogamous. I choose to be monogamous when I feel like it and I think it's a very healthy attitude. [NEWLINE] [NEWLINE] You can stay stuck in a relationship for years being monogamous for SO many different reasons, but I don't think that says much about your potential for being monogamous in the next one. Or vice versa.</s>
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In a way, maybe. [NEWLINE] [NEWLINE] <mask><mask><mask> /u/Moronica is saying is that poetry's purpose is fulfilled by its methods and style. Some people in this thread say that it must evoke emotion,<mask> I don't necessarily agree. I've analyzed and read (for fun and other purposes) many poems, and some of the best ones don't evoke deep emotion... they're just thought-provoking or interesting in their own way. [NEWLINE] [NEWLINE] It's subjective<mask> an individual may get a kick out of reading a speech, or listening to the reading of an essay.<mask>,<mask> these things are written to be heard in specific contexts and through certain mediums, the artist's perspective is best portrayed<mask> translated to specific art forms. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>In a way, maybe. [NEWLINE] [NEWLINE] I think what /u/Moronica is saying is that poetry's purpose is fulfilled by its methods and style. Some people in this thread say that it must evoke emotion, but I don't necessarily agree. I've analyzed and read (for fun and other purposes) many poems, and some of the best ones don't evoke deep emotion... they're just thought-provoking or interesting in their own way. [NEWLINE] [NEWLINE] It's subjective because an individual may get a kick out of reading a speech, or listening to the reading of an essay. However, because these things are written to be heard in specific contexts and through certain mediums, the artist's perspective is best portrayed when translated to specific art forms. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Philosophically... even<mask>  you were programmed to enjoy it, isn't it still real? There's a lot of evidence that says basically all of our sexual habits are "programmed". Nobody wakes up and says "I want to be sexually attracted to pantyhose and Asian women!" [NEWLINE] [NEWLINE] More directly - people don't choose to be gay. It's "programmed". [NEWLINE] [NEWLINE] I doubt a gay person would be like "well I don't REALLY like men, I was just programmed that way" -<mask> from a philosophical perspective I can't really see<mask> you could<mask><mask> some of your attractions or fetishes are more real/valid than others just<mask> you can identify a potential source of "programming" in some cases.</s>
Label encoding: <s>Philosophically... even if  you were programmed to enjoy it, isn't it still real? There's a lot of evidence that says basically all of our sexual habits are "programmed". Nobody wakes up and says "I want to be sexually attracted to pantyhose and Asian women!" [NEWLINE] [NEWLINE] More directly - people don't choose to be gay. It's "programmed". [NEWLINE] [NEWLINE] I doubt a gay person would be like "well I don't REALLY like men, I was just programmed that way" - so from a philosophical perspective I can't really see how you could argue that some of your attractions or fetishes are more real/valid than others just because you can identify a potential source of "programming" in some cases.</s>
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0671, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>???<mask> can you make this claim? [NEWLINE] [NEWLINE] [STARTQ] Only Americans ask<mask> they can do, peasants in Brazil or Kurds in Turkey or other oppressed groups thorughout the world do not ask<mask> they can do, they do it and make moves forward. They have less resources than us, less influence, less power, and are more likely to be murdered<mask> they already do far more than you or I do not<mask> they are closer to the problem<mask><mask> they are simply not hypocrites. [ENDQ] [NEWLINE] <mask> is your definition of hypocrite, here? Sorry<mask> I misunderstood your earlier post,<mask> you are going to have explain your definition of hypocrite, and<mask> evidence you have to show that people in the the third world aren't<mask> hypocrites.</s>
Label encoding: <s>??? How can you make this claim? [NEWLINE] [NEWLINE] [STARTQ] Only Americans ask what they can do, peasants in Brazil or Kurds in Turkey or other oppressed groups thorughout the world do not ask what they can do, they do it and make moves forward. They have less resources than us, less influence, less power, and are more likely to be murdered but they already do far more than you or I do not because they are closer to the problem but because they are simply not hypocrites. [ENDQ] [NEWLINE] What is your definition of hypocrite, here? Sorry if I misunderstood your earlier post, but you are going to have explain your definition of hypocrite, and what evidence you have to show that people in the the third world aren't also hypocrites.</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ I'm going to give you a delta. [NEWLINE] [NEWLINE] At first OP's point made a lot of sense, half the people are going to be in the bottom anyway<mask><mask> try and change the racial structure of them. [NEWLINE] [NEWLINE] I interpreted your point, which make more sense to me, to be that it is about the treatment of the individual rather than global statistics. [NEWLINE] [NEWLINE] You could have a world<mask> every race is proportionally represented in every government office and economic class<mask> different races would fight each other on sight. [NEWLINE] [NEWLINE] <mask><mask><mask><mask> every person of a certain race could be in the lowest echelons of society<mask><mask><mask><mask> no one ever judges them on their race then that world has more racial equality than the first </s>
Label encoding: <s>∆ I'm going to give you a delta. [NEWLINE] [NEWLINE] At first OP's point made a lot of sense, half the people are going to be in the bottom anyway so why try and change the racial structure of them. [NEWLINE] [NEWLINE] I interpreted your point, which make more sense to me, to be that it is about the treatment of the individual rather than global statistics. [NEWLINE] [NEWLINE] You could have a world where every race is proportionally represented in every government office and economic class but different races would fight each other on sight. [NEWLINE] [NEWLINE] On the other hand every person of a certain race could be in the lowest echelons of society but as long as no one ever judges them on their race then that world has more racial equality than the first </s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well for starters, I don't necessarily believe any poster should be outright banned.  I just don't respect it<mask> a tactic,<mask><mask><mask> the majority<mask> people who eat meat have already made the choice to ignore<mask> goes on in slaughterhouses. <mask> the majority of meat-eaters genuinely didn't know about the source of their food,<mask><mask> a simple poster that said "Ask us about<mask> your food comes from" with graphic information contained in pamphlets would still make the information available to the honestly ignorant,<mask> would still respect the desires of many to pretend that their meat comes from old macdonald's farm. <mask><mask><mask> these opinions of mine come down to my own ideas about respect, and shouldn't be matters of legality.</s>
Label encoding: <s>Well for starters, I don't necessarily believe any poster should be outright banned.  I just don't respect it as a tactic, because I think the majority if people who eat meat have already made the choice to ignore what goes on in slaughterhouses.  If the majority of meat-eaters genuinely didn't know about the source of their food, I think a simple poster that said "Ask us about where your food comes from" with graphic information contained in pamphlets would still make the information available to the honestly ignorant, but would still respect the desires of many to pretend that their meat comes from old macdonald's farm.  But I think these opinions of mine come down to my own ideas about respect, and shouldn't be matters of legality.</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because that's<mask> people do. Look at any other venue that isn't already sexualized - they do it there too. Both men and women.<mask> we were to de-genderify the bathrooms it would increase severely. Snarky comments<mask> people walk by and<mask> on. I'd say women would mostly not want to share the room with the men. Some men would probably embrace it, I know I wouldn't. [NEWLINE] [NEWLINE] [Here]( [URL] %2FMarkle+Pulse+Poll+Finds%3A+Harassment+of+Women+on+the+Street+Is...-a062870396) you go, an article on<mask> many women suffer from sexual harassment ON THE STREET. </s>
Label encoding: <s>Because that's what people do. Look at any other venue that isn't already sexualized - they do it there too. Both men and women. If we were to de-genderify the bathrooms it would increase severely. Snarky comments as people walk by and so on. I'd say women would mostly not want to share the room with the men. Some men would probably embrace it, I know I wouldn't. [NEWLINE] [NEWLINE] [Here]( [URL] %2FMarkle+Pulse+Poll+Finds%3A+Harassment+of+Women+on+the+Street+Is...-a062870396) you go, an article on how many women suffer from sexual harassment ON THE STREET. </s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think I approached this from the wrong angle. I guess the key thing is, it spreads fecal matter in the area around the toilet. [NEWLINE] [NEWLINE] You're probably safe from disease or whatever from this, I mean lots of people do it,<mask> there is a chance that you'll get sick from doing<mask>,<mask> your exposure to potentially harmful bacteria goes up. [NEWLINE] [NEWLINE] The key thing here isn't whether this is OK or not. I obv think it isn't<mask> that's not the point. The point is, there is a real physical and biological difference between the two options. It is not *only* a matter of who's time is more valuable, whose opinion matters more, etc which was the assertion of the OP.</s>
Label encoding: <s>I think I approached this from the wrong angle. I guess the key thing is, it spreads fecal matter in the area around the toilet. [NEWLINE] [NEWLINE] You're probably safe from disease or whatever from this, I mean lots of people do it, but there is a chance that you'll get sick from doing so, since your exposure to potentially harmful bacteria goes up. [NEWLINE] [NEWLINE] The key thing here isn't whether this is OK or not. I obv think it isn't but that's not the point. The point is, there is a real physical and biological difference between the two options. It is not *only* a matter of who's time is more valuable, whose opinion matters more, etc which was the assertion of the OP.</s>
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In the case of property taxes, they're already progressive in<mask> much<mask> they depend on the fair market value of an asset you own. [ENDQ] [NEWLINE] That's not "progressive." Progressive means charging richer people more and poorer people less than they would pay given a linear scale. [NEWLINE] [NEWLINE] A fictional example of progressive taxes: $100k property might have no tax, $200k property might have 10% tax, and  $300k property might have 40% tax. [NEWLINE] [NEWLINE] That's progressive: you give up a more-than-linear chunk<mask> you go up the scale. [NEWLINE] [NEWLINE] It's a sort of socialism, redistributing proportionally more assets from those who can seemingly afford it, to those who can't.</s>
Label encoding: <s> [STARTQ] In the case of property taxes, they're already progressive in as much as they depend on the fair market value of an asset you own. [ENDQ] [NEWLINE] That's not "progressive." Progressive means charging richer people more and poorer people less than they would pay given a linear scale. [NEWLINE] [NEWLINE] A fictional example of progressive taxes: $100k property might have no tax, $200k property might have 10% tax, and  $300k property might have 40% tax. [NEWLINE] [NEWLINE] That's progressive: you give up a more-than-linear chunk as you go up the scale. [NEWLINE] [NEWLINE] It's a sort of socialism, redistributing proportionally more assets from those who can seemingly afford it, to those who can't.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm actually from Germany<mask> I'd guess we have similar legislations due to the EU. [NEWLINE] [NEWLINE] It kind of sucks<mask>, Germans are notorious misers<mask> it comes to food,<mask><mask> we actually pay the least of all EU countries for our food and that comes at a price. [NEWLINE] [NEWLINE] [STARTQ] Having said that, I would think any effort at only buying free range could only help. [ENDQ] [NEWLINE] It'd probably be an improvement in general, yes. I don't eat meat<mask> I try to buy<mask> much organic butter, milk and eggs<mask> possible<mask> I'm a student and on a rather tight budget. I'm thinking of ditching eggs compeletly, I could never ditch cheese or milk<mask>, sadly. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I'm actually from Germany so I'd guess we have similar legislations due to the EU. [NEWLINE] [NEWLINE] It kind of sucks though, Germans are notorious misers when it comes to food, I think we actually pay the least of all EU countries for our food and that comes at a price. [NEWLINE] [NEWLINE] [STARTQ] Having said that, I would think any effort at only buying free range could only help. [ENDQ] [NEWLINE] It'd probably be an improvement in general, yes. I don't eat meat but I try to buy as much organic butter, milk and eggs as possible although I'm a student and on a rather tight budget. I'm thinking of ditching eggs compeletly, I could never ditch cheese or milk though, sadly. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> we might later discover something that corroborates our story [ENDQ] [NEWLINE] I suppose this is the only way you would find out. [NEWLINE] [NEWLINE] Until you have some other source of verification, you're basing everything on subjective interpretation of scripture, and have no way of checking<mask> you're getting closer or further away from the hypothetical "perfect interpretation". Doesn't that make it a largely pointless exercise until a way to verify your interpretation manifests itself? [NEWLINE] [NEWLINE] Until this happens, everybody is wandering in the dark and arguing over opinions which can't be verified. The best you can hope for is a personal interpretation which is "honest" (not disingenuous),<mask> you can hardly claim with any confidence that it is closer to perfection than other interpretations.</s>
Label encoding: <s> [STARTQ] Also we might later discover something that corroborates our story [ENDQ] [NEWLINE] I suppose this is the only way you would find out. [NEWLINE] [NEWLINE] Until you have some other source of verification, you're basing everything on subjective interpretation of scripture, and have no way of checking if you're getting closer or further away from the hypothetical "perfect interpretation". Doesn't that make it a largely pointless exercise until a way to verify your interpretation manifests itself? [NEWLINE] [NEWLINE] Until this happens, everybody is wandering in the dark and arguing over opinions which can't be verified. The best you can hope for is a personal interpretation which is "honest" (not disingenuous), but you can hardly claim with any confidence that it is closer to perfection than other interpretations.</s>
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Formal studies, probably not. Anecdotally it happens fairly often even today. I've never heard of it being anything other than a really, really bad idea. [There was a thread on the ask transgender sub a month ago]( [URL] /). [NEWLINE] [NEWLINE] <mask> consider that it is VERY common for a trans woman to go through a hyper masculine streak and a trans man to go through a hyper feminine streak, both in an attempt to just "try harder" at being their assigned at birth sex. Trans women are twice<mask> likely<mask> the general population to join the military, for example. Again, I've never heard of this working out nicely for anyone, and it's usually one of the most regretful times in their lives.</s>
Label encoding: <s>Formal studies, probably not. Anecdotally it happens fairly often even today. I've never heard of it being anything other than a really, really bad idea. [There was a thread on the ask transgender sub a month ago]( [URL] /). [NEWLINE] [NEWLINE] Also consider that it is VERY common for a trans woman to go through a hyper masculine streak and a trans man to go through a hyper feminine streak, both in an attempt to just "try harder" at being their assigned at birth sex. Trans women are twice as likely as the general population to join the military, for example. Again, I've never heard of this working out nicely for anyone, and it's usually one of the most regretful times in their lives.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is a bit of double speak in your responce. [NEWLINE] [NEWLINE] You say that you don't treat people differently,<mask> you<mask> say this [NEWLINE] [NEWLINE] Mainly just black men who remind me of the way he looked and spoke. [NEWLINE] [NEWLINE] That's treating people differently based on their race. [NEWLINE] [NEWLINE] Which is somewhat the root of the problem.<mask> we just forget about color than we will still be making decisions such do I cross the street or do I stay were I am waking based on racial reasons. [NEWLINE] [NEWLINE] <mask> we aren't aware that we are making choices based on race than a person can never tackle those issues. [NEWLINE] [NEWLINE] Just thinking in color blind terms doesn't really work. It is band aid on a stab wound. </s>
Label encoding: <s>There is a bit of double speak in your responce. [NEWLINE] [NEWLINE] You say that you don't treat people differently, but you also say this [NEWLINE] [NEWLINE] Mainly just black men who remind me of the way he looked and spoke. [NEWLINE] [NEWLINE] That's treating people differently based on their race. [NEWLINE] [NEWLINE] Which is somewhat the root of the problem. If we just forget about color than we will still be making decisions such do I cross the street or do I stay were I am waking based on racial reasons. [NEWLINE] [NEWLINE] If we aren't aware that we are making choices based on race than a person can never tackle those issues. [NEWLINE] [NEWLINE] Just thinking in color blind terms doesn't really work. It is band aid on a stab wound. </s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What<mask> the family needs the money that they would be paying to keep you alive to have a good life themselves? [NEWLINE] [NEWLINE] <mask> you don't want to talk about things and examples then don't start a CMV thread.  Good lord, man. <mask> did you expect? People to just be like: "oh yeah, man.  you're totally right. good on ya." [NEWLINE] [NEWLINE] The example shows someone sacrificing themselves for the betterment of others.  It's the exact same mindset that goes through the mind of the soldier.  I understand<mask> you don't want to acknowledge that,<mask> you could at least not act like a child. <mask>, I appreciate you dismissing my whole post without reading.  </s>
Label encoding: <s>What if the family needs the money that they would be paying to keep you alive to have a good life themselves? [NEWLINE] [NEWLINE] If you don't want to talk about things and examples then don't start a CMV thread.  Good lord, man.  What did you expect? People to just be like: "oh yeah, man.  you're totally right. good on ya." [NEWLINE] [NEWLINE] The example shows someone sacrificing themselves for the betterment of others.  It's the exact same mindset that goes through the mind of the soldier.  I understand if you don't want to acknowledge that, but you could at least not act like a child.  Though, I appreciate you dismissing my whole post without reading.  </s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It can be a thin line between artist and performer ( even more<mask> a work of art and a replica),<mask> not all covers are equal and some change the meaning of the original song. For example [here's the original version of I will survive by Gloria Gaynor]( [URL] -iI3-I) and [here's a cover by Cake]( [URL] -u4o) [NEWLINE] [NEWLINE] Notice<mask> the original has a brighter tone and has a bottom line hopeful message.<mask> Cake's cover has a darker tone and seems sad or sarcastic with an ambiguous message<mask> of it. [NEWLINE] [NEWLINE] These changes were carefully crafted and<mask> it borrows heavily from source material it stands on its own<mask> a creative piece of work or art. </s>
Label encoding: <s>It can be a thin line between artist and performer ( even more so a work of art and a replica), but not all covers are equal and some change the meaning of the original song. For example [here's the original version of I will survive by Gloria Gaynor]( [URL] -iI3-I) and [here's a cover by Cake]( [URL] -u4o) [NEWLINE] [NEWLINE] Notice how the original has a brighter tone and has a bottom line hopeful message. But Cake's cover has a darker tone and seems sad or sarcastic with an ambiguous message because of it. [NEWLINE] [NEWLINE] These changes were carefully crafted and while it borrows heavily from source material it stands on its own as a creative piece of work or art. </s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly,<mask><mask><mask><mask> men are unfeeling, nor do I posit they DONT suffer greatly from the pain of unplanned pregnancy. [NEWLINE] [NEWLINE] I can't imagine knowing for instance your child is going to be aborted and being powerless. Or that your child was adopted out to strangers without your knowledge. [NEWLINE] [NEWLINE] <mask><mask> a parent myself, I can say you are wrong about vengeance.<mask> my child does something I see<mask> wrong, and I tell them not to, and they do it anyway and get hurt, it's not vengeance. It's an unfortunate natural consequence. [NEWLINE] [NEWLINE] In a perfect world, no one male OR female would be forced to parent.<mask> we live on Earth,<mask> fucking makes babies. </s>
Label encoding: <s>Firstly, I do not think men are unfeeling, nor do I posit they DONT suffer greatly from the pain of unplanned pregnancy. [NEWLINE] [NEWLINE] I can't imagine knowing for instance your child is going to be aborted and being powerless. Or that your child was adopted out to strangers without your knowledge. [NEWLINE] [NEWLINE] But as a parent myself, I can say you are wrong about vengeance. if my child does something I see as wrong, and I tell them not to, and they do it anyway and get hurt, it's not vengeance. It's an unfortunate natural consequence. [NEWLINE] [NEWLINE] In a perfect world, no one male OR female would be forced to parent. But we live on Earth, where fucking makes babies. </s>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just some food for thought before I go to bed. [NEWLINE] [NEWLINE] *<mask> about people who don't have jobs<mask> of discrimination, physical and/or mental disabilities? [NEWLINE] * The federal reserve artificially controls unemployment to prevent the economy from overheating ([source]( [URL] /)),<mask> about the repercussions of being unable to do that? [NEWLINE] * Welfare programs have help reduce poverty by ~12% [NEWLINE] * Welfare for the unemployed is only a small part of the entire welfare program [NEWLINE] * Take a look at the Great Depression to see<mask> happens<mask> there is no welfare, and see<mask> you think it was right (maybe you do maybe you don't). [NEWLINE] [NEWLINE] Well that's it for me, hope I changed your view a bit :) [NEWLINE] </s>
Label encoding: <s>Just some food for thought before I go to bed. [NEWLINE] [NEWLINE] * What about people who don't have jobs because of discrimination, physical and/or mental disabilities? [NEWLINE] * The federal reserve artificially controls unemployment to prevent the economy from overheating ([source]( [URL] /)), what about the repercussions of being unable to do that? [NEWLINE] * Welfare programs have help reduce poverty by ~12% [NEWLINE] * Welfare for the unemployed is only a small part of the entire welfare program [NEWLINE] * Take a look at the Great Depression to see what happens when there is no welfare, and see if you think it was right (maybe you do maybe you don't). [NEWLINE] [NEWLINE] Well that's it for me, hope I changed your view a bit :) [NEWLINE] </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My understanding of<mask> people believes in things goes like this. People believe things for 3 reasons: [NEWLINE] 1)<mask> they're supported by the person's understanding and perception of observable evidence. [NEWLINE] 2)<mask> the person receives some sort of benefit, or avoids a disbenefit, for having that belief (even<mask> the only benefit is that it gives them peace of mind, makes them feel good, or provides justification or validation of their existing view of the world and their place in it) [NEWLINE] 3) "Just<mask> "; i.e. out of intellectual laziness, lack of imagination, or lack of known alternatives. [NEWLINE] [NEWLINE] Option 1 is out the window<mask> it comes to religion. That leaves the other two. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>My understanding of why people believes in things goes like this. People believe things for 3 reasons: [NEWLINE] 1) Because they're supported by the person's understanding and perception of observable evidence. [NEWLINE] 2) Because the person receives some sort of benefit, or avoids a disbenefit, for having that belief (even if the only benefit is that it gives them peace of mind, makes them feel good, or provides justification or validation of their existing view of the world and their place in it) [NEWLINE] 3) "Just because "; i.e. out of intellectual laziness, lack of imagination, or lack of known alternatives. [NEWLINE] [NEWLINE] Option 1 is out the window when it comes to religion. That leaves the other two. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm Canadian,<mask> we are usually in between these "American" v. "British" issues. The spelling is pretty much already interchangeable except for things like "colour, harbour, and favour." Things with the "-our" are ingrained,<mask> things like "offence/offense" or "recognise/recognize" are used interchangeably.<mask> it has been kind of like an experiment here for<mask> you are describing. It doesn't cause any problems that I can identify.<mask> yes, the distinction is pointless,<mask> that is to imply people think it has a point, or ever did have a point. It is just historically<mask> the languages evolved. In short, we should all use Canadian spelling instead.</s>
Label encoding: <s>I'm Canadian, so we are usually in between these "American" v. "British" issues. The spelling is pretty much already interchangeable except for things like "colour, harbour, and favour." Things with the "-our" are ingrained, but things like "offence/offense" or "recognise/recognize" are used interchangeably. So it has been kind of like an experiment here for what you are describing. It doesn't cause any problems that I can identify. So yes, the distinction is pointless, but that is to imply people think it has a point, or ever did have a point. It is just historically how the languages evolved. In short, we should all use Canadian spelling instead.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0878, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You keep bring up the "right to your own body" argument,<mask> that doesn't really apply. You have the right to your body. You are free to decide who you have sex with or not. That is not taken away from anyone. The aspect being regulated is commerce. And it is entirely common and normal for regulations affecting commerce. You need a business license for just about anything. There is no right to a business license, it is a privilege. [NEWLINE] [NEWLINE] Now, I'm not addressing your other points. There is solid logic in some of them even<mask> I don't entirely agree.<mask>, the right to your body logic is flawed<mask> that right is not being infringed, just commerce is being regulated.</s>
Label encoding: <s>You keep bring up the "right to your own body" argument, but that doesn't really apply. You have the right to your body. You are free to decide who you have sex with or not. That is not taken away from anyone. The aspect being regulated is commerce. And it is entirely common and normal for regulations affecting commerce. You need a business license for just about anything. There is no right to a business license, it is a privilege. [NEWLINE] [NEWLINE] Now, I'm not addressing your other points. There is solid logic in some of them even if I don't entirely agree. However, the right to your body logic is flawed because that right is not being infringed, just commerce is being regulated.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you want to prevent bad outcomes, you can't just say "don't be an idiot" to the victim and pretend the problem will correct itself. [ENDQ] [NEWLINE] Right, "don't be an idiot" is not useful information. Being smart doesn't necessarily prevent bad outcomes. You'd say "always lock your door". [NEWLINE] [NEWLINE] The few times something 'bad' happens in my area, there's Facebook groups dedicated to warning people about it. False magazine sellers to scout apartments<mask> the owners are out. And to remove a notice<mask> you get back, to ensure they don't come to your building. [NEWLINE] [NEWLINE] <mask> you ignore that advice, then it's partly your fault. Pay attention, and use caution.</s>
Label encoding: <s> [STARTQ] If you want to prevent bad outcomes, you can't just say "don't be an idiot" to the victim and pretend the problem will correct itself. [ENDQ] [NEWLINE] Right, "don't be an idiot" is not useful information. Being smart doesn't necessarily prevent bad outcomes. You'd say "always lock your door". [NEWLINE] [NEWLINE] The few times something 'bad' happens in my area, there's Facebook groups dedicated to warning people about it. False magazine sellers to scout apartments where the owners are out. And to remove a notice when you get back, to ensure they don't come to your building. [NEWLINE] [NEWLINE] If you ignore that advice, then it's partly your fault. Pay attention, and use caution.</s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand that it is very difficult for victims to come forward and report instances of rape,<mask> the results from someone being charged with rape are **incredibly** severe. [NEWLINE] [NEWLINE] Simply having such a thing on your record will guarantee you to be at the bottom of the job recruitment pool, destroy your reputation both personally and professionally. The extreme of the example is being named a felon, which would go even further by removing your right to vote, carry a weapon, be near schools/children, and many other privileges. [NEWLINE] [NEWLINE] The results of a rape accusation are hefty<mask>, and<mask> it is true that they are severe for a reason, just implying someone may be a rapist is enough to completely destroy their life. </s>
Label encoding: <s>I understand that it is very difficult for victims to come forward and report instances of rape, but the results from someone being charged with rape are **incredibly** severe. [NEWLINE] [NEWLINE] Simply having such a thing on your record will guarantee you to be at the bottom of the job recruitment pool, destroy your reputation both personally and professionally. The extreme of the example is being named a felon, which would go even further by removing your right to vote, carry a weapon, be near schools/children, and many other privileges. [NEWLINE] [NEWLINE] The results of a rape accusation are hefty indeed, and while it is true that they are severe for a reason, just implying someone may be a rapist is enough to completely destroy their life. </s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What, in your opinion, is the actual difference between an "obviously censor-worthy sub-reddit" and a "borderline" sub-reddit like /r/whiterights? [NEWLINE] [NEWLINE] Right now, the difference seems to be that you simply feel more passionate disagreement with the message of the sub-reddits whose themes are rape, racism, holocaust-denial, and misogyny. [NEWLINE] [NEWLINE] The entire point of Reddit is to be a platform<mask> people share content.<mask> you don't like it, there are "downvote" and "unsubscribe" buttons. No one is forcing you to look at the content. You are forgetting that half of free speech is the ability *not to listen*</s>
Label encoding: <s>What, in your opinion, is the actual difference between an "obviously censor-worthy sub-reddit" and a "borderline" sub-reddit like /r/whiterights? [NEWLINE] [NEWLINE] Right now, the difference seems to be that you simply feel more passionate disagreement with the message of the sub-reddits whose themes are rape, racism, holocaust-denial, and misogyny. [NEWLINE] [NEWLINE] The entire point of Reddit is to be a platform where people share content. If you don't like it, there are "downvote" and "unsubscribe" buttons. No one is forcing you to look at the content. You are forgetting that half of free speech is the ability *not to listen*</s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Pipstydoo, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=Pipstydoo+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry Pipstydoo, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=Pipstydoo+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0935, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>* aljezeera.com [NEWLINE] * amazon.com (browsing even<mask> signed in) [NEWLINE] * bbc.co.uk [NEWLINE] * bing.com [NEWLINE] * cnn.com [NEWLINE] * craigslist.org [NEWLINE] * ebay.com [NEWLINE] * linkedin.com [NEWLINE] * motherless.com [NEWLINE] * netflix.com (browsing<mask> logged in) [NEWLINE] * nytimes.com [NEWLINE] * rdio.com [NEWLINE] * redtube.com [NEWLINE] * skype.com [NEWLINE] * speigel.de [NEWLINE] * walmart.com [NEWLINE] * xhamster.com [NEWLINE] [NEWLINE] [A little over half the sites in the Alexa top 1000]( [URL].txt).</s>
Label encoding: <s>* aljezeera.com [NEWLINE] * amazon.com (browsing even when signed in) [NEWLINE] * bbc.co.uk [NEWLINE] * bing.com [NEWLINE] * cnn.com [NEWLINE] * craigslist.org [NEWLINE] * ebay.com [NEWLINE] * linkedin.com [NEWLINE] * motherless.com [NEWLINE] * netflix.com (browsing when logged in) [NEWLINE] * nytimes.com [NEWLINE] * rdio.com [NEWLINE] * redtube.com [NEWLINE] * skype.com [NEWLINE] * speigel.de [NEWLINE] * walmart.com [NEWLINE] * xhamster.com [NEWLINE] [NEWLINE] [A little over half the sites in the Alexa top 1000]( [URL].txt).</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0782, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry ConnivingToad, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=ConnivingToad+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry ConnivingToad, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=ConnivingToad+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you *enjoy* your steak well done, eat it. [NEWLINE] [NEWLINE] <mask>, cooking a steak to that temperature will remove most of the juices *and* cook the muscle fibers<mask> much that it will become tough and chewy in comparison to a medium-rare steak. [NEWLINE] [NEWLINE] In that regard,<mask> you like well done, eat it.<mask>, you are wasting your money by cooking a nice steak well done,<mask> you are effectively destroying the fibers that make a steak<mask> tender and juicy.<mask> you're going to cook it well done, buy a select cut (dont waste your $ on choice or prime) and a cheap cut of steak,<mask> it won't make a difference.</s>
Label encoding: <s>If you *enjoy* your steak well done, eat it. [NEWLINE] [NEWLINE] However, cooking a steak to that temperature will remove most of the juices *and* cook the muscle fibers so much that it will become tough and chewy in comparison to a medium-rare steak. [NEWLINE] [NEWLINE] In that regard, if you like well done, eat it. But, you are wasting your money by cooking a nice steak well done, because you are effectively destroying the fibers that make a steak so tender and juicy. If you're going to cook it well done, buy a select cut (dont waste your $ on choice or prime) and a cheap cut of steak, because it won't make a difference.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A decade is a pretty small timeframe for these things. Re-alignments typically take 20-40 years. Look at the last one - the South was a solid Democratic block for a century after the Civil War. LBJ signed the civil rights act in 1964 which marked the turning point,<mask> the south didn't become a dependable GOP voting block until Reagan, and the process didn't complete until 2004 (<mask> Zel Miller, the last Dixiecrat, flipped parties). [NEWLINE] [NEWLINE] <mask><mask> history will ultimately judge 2008<mask> the end of the "Reagan Coalition" and the start of a new re-alignment in US politics.<mask> it won't really be obvious for another decade or two. </s>
Label encoding: <s>A decade is a pretty small timeframe for these things. Re-alignments typically take 20-40 years. Look at the last one - the South was a solid Democratic block for a century after the Civil War. LBJ signed the civil rights act in 1964 which marked the turning point, but the south didn't become a dependable GOP voting block until Reagan, and the process didn't complete until 2004 ( When Zel Miller, the last Dixiecrat, flipped parties). [NEWLINE] [NEWLINE] I think history will ultimately judge 2008 as the end of the "Reagan Coalition" and the start of a new re-alignment in US politics. But it won't really be obvious for another decade or two. </s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I never said that one shouldn't<mask> have a general education alongside with that - I'm not sure about your university<mask> mine required that every major take some courses in other fields. [NEWLINE] [NEWLINE] Well, who's to say<mask> the society would be like? Plenty of people deemed successful in society have had no higher education and have gone on to do great things - artists and science alike. Beautiful things come out of all aspects of the spectrum.<mask>, it certainly would not be<mask> diverse of a society<mask> we lacked a pursuit of the humanities. That I do not doubt. [NEWLINE] [NEWLINE] <mask>, my problem is with people that treat college purely<mask> job training and not<mask> a way of applying themselves to something. </s>
Label encoding: <s>I never said that one shouldn't also have a general education alongside with that - I'm not sure about your university but mine required that every major take some courses in other fields. [NEWLINE] [NEWLINE] Well, who's to say what the society would be like? Plenty of people deemed successful in society have had no higher education and have gone on to do great things - artists and science alike. Beautiful things come out of all aspects of the spectrum. However, it certainly would not be as diverse of a society if we lacked a pursuit of the humanities. That I do not doubt. [NEWLINE] [NEWLINE] However, my problem is with people that treat college purely as job training and not as a way of applying themselves to something. </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Opinion: A desire for self preservation.<mask> you run around trying to out-muscle people for everything you want, you're going to lose, a lot, and your life is going to be severely harmed. Cooperation is easier and less risky. [NEWLINE] [NEWLINE] <mask>, most people have empathy. Only a small number of people who don't. The real murders, rapists, etc, don't actually feel bad for<mask> they do.<mask> most people understand they do not want themselves to be harmed, and<mask> feel harming others is wrong. [NEWLINE] [NEWLINE] It is not laws that keep most people in check.<mask> tomorrow there was no government, I wouldn't go around murdering people for sport and money.</s>
Label encoding: <s>Opinion: A desire for self preservation. If you run around trying to out-muscle people for everything you want, you're going to lose, a lot, and your life is going to be severely harmed. Cooperation is easier and less risky. [NEWLINE] [NEWLINE] Also, most people have empathy. Only a small number of people who don't. The real murders, rapists, etc, don't actually feel bad for what they do. But most people understand they do not want themselves to be harmed, and so feel harming others is wrong. [NEWLINE] [NEWLINE] It is not laws that keep most people in check. If tomorrow there was no government, I wouldn't go around murdering people for sport and money.</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It comes down to personal preference, your friends and<mask> games you play at the end of the day, and trying to change a fan boys opinion is often difficult. [NEWLINE] [NEWLINE] I myself am a PC gamer and haven't played on console for a<mask>,<mask> I see no need to hate it.<mask> people prefer one over another that's fine for whatever reasons. [NEWLINE] [NEWLINE] Sorry<mask> I'm wrong<mask> this seems like your baiting for an argument<mask> someone claiming to be part of the "master race" wouldn't change their views. [NEWLINE] [NEWLINE] <mask> you're serious then yeah, some people just prefer the plug and play simplicity and the odd exclusive for each system. Who says you can have only one?</s>
Label encoding: <s>It comes down to personal preference, your friends and what games you play at the end of the day, and trying to change a fan boys opinion is often difficult. [NEWLINE] [NEWLINE] I myself am a PC gamer and haven't played on console for a while, but I see no need to hate it. If people prefer one over another that's fine for whatever reasons. [NEWLINE] [NEWLINE] Sorry if I'm wrong but this seems like your baiting for an argument since someone claiming to be part of the "master race" wouldn't change their views. [NEWLINE] [NEWLINE] If you're serious then yeah, some people just prefer the plug and play simplicity and the odd exclusive for each system. Who says you can have only one?</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>IPAs are everywhere and it's a bad thing for only one reason...<mask><mask> you don't like IPA? I don't, I haven't found one that I liked. Dark lager and porter is my jam at the moment, neither of which are particularly bitter (to me at least)<mask> the amount of beer I have to choose from is dwindling unless I want to drive an extended distance or order online. [NEWLINE] [NEWLINE] I have only a few beers to choose from and I'd like to explore<mask> exploration equates to more fucking IPA. I wouldn't mind some more variety. It should be noted I'm not in a medium or large metro area<mask> that might have an effect to it.</s>
Label encoding: <s>IPAs are everywhere and it's a bad thing for only one reason... what if you don't like IPA? I don't, I haven't found one that I liked. Dark lager and porter is my jam at the moment, neither of which are particularly bitter (to me at least) but the amount of beer I have to choose from is dwindling unless I want to drive an extended distance or order online. [NEWLINE] [NEWLINE] I have only a few beers to choose from and I'd like to explore but exploration equates to more fucking IPA. I wouldn't mind some more variety. It should be noted I'm not in a medium or large metro area so that might have an effect to it.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of Mice and Men, We Came<mask> Romans, Blessthefall, I See Stars, Miss May I, Underoath, My Ticket Home, Memphis May Fire, Like Moths To Flames, Woe Is Me, Chiodos, Dance Gavin Dance, Ice Nine Kills, The Word Alive, Adept, Adestria, Alexisonfire, Attack Attack, The Amity Affliction and many more related artists you can find on YouTube or Last.fm. Heavy metal is okay<mask><mask> you're trying to transition into it, I would suggest giving these above bands a try for they have singing + screaming...with both highs and lows that will really get your adrenaline pumping. [NEWLINE] </s>
Label encoding: <s>Of Mice and Men, We Came As Romans, Blessthefall, I See Stars, Miss May I, Underoath, My Ticket Home, Memphis May Fire, Like Moths To Flames, Woe Is Me, Chiodos, Dance Gavin Dance, Ice Nine Kills, The Word Alive, Adept, Adestria, Alexisonfire, Attack Attack, The Amity Affliction and many more related artists you can find on YouTube or Last.fm. Heavy metal is okay but if you're trying to transition into it, I would suggest giving these above bands a try for they have singing + screaming...with both highs and lows that will really get your adrenaline pumping. [NEWLINE] </s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry AnticPosition, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=AnticPosition+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry AnticPosition, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=AnticPosition+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I almost feel like they are too extreme. [ENDQ] [NEWLINE] <mask><mask>. I would consider myself center in many matters,<mask> it's just good knowing that someone in parliament is bringing up these points, that we have a communist to discuss with the libertarian. [NEWLINE] [NEWLINE] [STARTQ] does it make the government more inefficient<mask> of the fact that there are<mask> many ideas being represented. I realize its good to have the people represented,<mask> it almost seems like giving voice to the extreme could cause problems. [ENDQ] [NEWLINE] Yes, more voices make the government more inefficient, with the single voice of the dictator being the most efficient. I see the many-party system<mask> a closer thing to democracy than the two-party one.</s>
Label encoding: <s> [STARTQ] I almost feel like they are too extreme. [ENDQ] [NEWLINE] I agree. I would consider myself center in many matters, but it's just good knowing that someone in parliament is bringing up these points, that we have a communist to discuss with the libertarian. [NEWLINE] [NEWLINE] [STARTQ] does it make the government more inefficient because of the fact that there are so many ideas being represented. I realize its good to have the people represented, but it almost seems like giving voice to the extreme could cause problems. [ENDQ] [NEWLINE] Yes, more voices make the government more inefficient, with the single voice of the dictator being the most efficient. I see the many-party system as a closer thing to democracy than the two-party one.</s>
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The essential difference between companies and the govt is that the govt uses coercion.  Companies have to rely on your free choice. <mask> they used coercion, the would become less and less profitable.  At some point, they would eventually have to rely exclusively on force to support their existing force.  At this point, yes, you could call them a government. [NEWLINE] [NEWLINE] You ask me to deny one of the essential functions of govt, then ask<mask> there's any other difference between govt and companies<mask> that essential function. [NEWLINE] [NEWLINE] You're either a poor debater or a troll.  Either way, I don't think you can teach me anything regarding this subject. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>The essential difference between companies and the govt is that the govt uses coercion.  Companies have to rely on your free choice.  If they used coercion, the would become less and less profitable.  At some point, they would eventually have to rely exclusively on force to support their existing force.  At this point, yes, you could call them a government. [NEWLINE] [NEWLINE] You ask me to deny one of the essential functions of govt, then ask if there's any other difference between govt and companies besides that essential function. [NEWLINE] [NEWLINE] You're either a poor debater or a troll.  Either way, I don't think you can teach me anything regarding this subject. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [NEWLINE] Sorry pcgumby, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=pcgumby+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \)) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [NEWLINE] [NEWLINE] Sorry pcgumby, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=pcgumby+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \)) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I completely agree with you, I have to point out that this is a self-referential fallacy that is commonly used to argue in favor of religion ("The Bible is true<mask> it says in the Bible that the Bible is true.") [NEWLINE] [NEWLINE] That aside,<mask><mask> that scientific reasoning is the best way to understand the natural world and gain empirical knowledge.<mask>, I don't agree that it is the best way to know reality,<mask> by reality you are referring to Everything That Exists. A system cannot be fully understood from within itself. We can simply understand smaller systems. [NEWLINE] [NEWLINE] <mask>, we never "know" anything<mask><mask><mask> of scientific experimentation. We just get highly proven theories.</s>
Label encoding: <s>While I completely agree with you, I have to point out that this is a self-referential fallacy that is commonly used to argue in favor of religion ("The Bible is true because it says in the Bible that the Bible is true.") [NEWLINE] [NEWLINE] That aside, I agree that scientific reasoning is the best way to understand the natural world and gain empirical knowledge. However, I don't agree that it is the best way to know reality, if by reality you are referring to Everything That Exists. A system cannot be fully understood from within itself. We can simply understand smaller systems. [NEWLINE] [NEWLINE] Also, we never "know" anything as a result of scientific experimentation. We just get highly proven theories.</s>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a South East Asian, who are not familiar with Confucius, I don't think<mask><mask> with this. Asian staple food is carbs, rice or noodles or other starches.. And we don't eat steak. Fish or chicken can be dealt with spoons (or chopsticks,<mask> we're not talking about the Chinese here) easily.. And beef is often cooked in small pieces<mask> the spices are absorbed better. [NEWLINE] [NEWLINE] <mask> the need to cut something on your plate is really not common. [NEWLINE] [NEWLINE] Edit: to think of it spoons are probably a heritage from the European colonial era..<mask> they had knives too<mask> they don't get adopted nearly<mask> well<mask> spoons. </s>
Label encoding: <s>As a South East Asian, who are not familiar with Confucius, I don't think I agree with this. Asian staple food is carbs, rice or noodles or other starches.. And we don't eat steak. Fish or chicken can be dealt with spoons (or chopsticks, but we're not talking about the Chinese here) easily.. And beef is often cooked in small pieces so the spices are absorbed better. [NEWLINE] [NEWLINE] So the need to cut something on your plate is really not common. [NEWLINE] [NEWLINE] Edit: to think of it spoons are probably a heritage from the European colonial era.. But they had knives too but they don't get adopted nearly as well as spoons. </s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It sounds like you very much missed the point of the use of sex in 1984. Julia and Winston's illicit sexual relationship is a large part of their rebellion, particularly for Julia, considering she is part of the "Anti-Sex League." Their sexual relationship reflects the fact that the government controls every facet of their lives, even the part that's supposed to be most intimate. This is hugely important to the book, particularly the ending,<mask> it makes his and Julia's betrayals even more heartbreaking. [NEWLINE] [NEWLINE] I find your desire to avoid any books with sex in them a little strange. After all, good literature often comments on the human experience and sex is often a large part of that. </s>
Label encoding: <s>It sounds like you very much missed the point of the use of sex in 1984. Julia and Winston's illicit sexual relationship is a large part of their rebellion, particularly for Julia, considering she is part of the "Anti-Sex League." Their sexual relationship reflects the fact that the government controls every facet of their lives, even the part that's supposed to be most intimate. This is hugely important to the book, particularly the ending, as it makes his and Julia's betrayals even more heartbreaking. [NEWLINE] [NEWLINE] I find your desire to avoid any books with sex in them a little strange. After all, good literature often comments on the human experience and sex is often a large part of that. </s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay, first, your "laws exist" argument falls flat in an epic way<mask> the majority of states (read, nearly ALL) have laws, alright...laws that fly in the face of your argument, exempting breastfeeding from "indecent exposure" laws/arguments, and many have laws preventing people like you from harassing a nursing mother. [NEWLINE] [NEWLINE] Second, anyone who gets uncomfortable around a nursing mother is free to leave, and go somewhere else. The problem is theirs, not the mother's, not the baby's.<mask> you're sexualizing the most natural or processes, that's<mask> something is wrong in YOUR mind. There's nothing sexual or lewd about it. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Okay, first, your "laws exist" argument falls flat in an epic way since the majority of states (read, nearly ALL) have laws, alright...laws that fly in the face of your argument, exempting breastfeeding from "indecent exposure" laws/arguments, and many have laws preventing people like you from harassing a nursing mother. [NEWLINE] [NEWLINE] Second, anyone who gets uncomfortable around a nursing mother is free to leave, and go somewhere else. The problem is theirs, not the mother's, not the baby's. If you're sexualizing the most natural or processes, that's because something is wrong in YOUR mind. There's nothing sexual or lewd about it. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do you get to decide<mask> theft is?  To me, theft is failing to pay your lawful taxes.  By doing this you would increase the tax burden on those of us who aren't stealing from the government. [NEWLINE] [NEWLINE] There's a solution to taxes<mask> you don't like them: convince enough of your fellow citizrns that your point of view is correct that they will elect representitives to eliminate them.  That's going to be a heavy lift<mask> the vast majority of us feel you're wrong. [NEWLINE] [NEWLINE] In the end that's<mask> it comes down to.  You can define theft any way you want to,<mask> it doesn't mean your definition is correct.</s>
Label encoding: <s>Why do you get to decide what theft is?  To me, theft is failing to pay your lawful taxes.  By doing this you would increase the tax burden on those of us who aren't stealing from the government. [NEWLINE] [NEWLINE] There's a solution to taxes if you don't like them: convince enough of your fellow citizrns that your point of view is correct that they will elect representitives to eliminate them.  That's going to be a heavy lift because the vast majority of us feel you're wrong. [NEWLINE] [NEWLINE] In the end that's what it comes down to.  You can define theft any way you want to, but it doesn't mean your definition is correct.</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We have to remember<mask> that some of the older feminists are from a time<mask> females really were *far* from being equal. It is easy for the younger ones amongst us to forget that. [NEWLINE] [NEWLINE] In the 60's (67<mask> my memory doesn't fail me) the first female marathon runner competed in the Boston Runner. She wasn't actually even allowed, it wasn't until 5 years later women were allowed to compete. And this runner had the organiser(/owner?) chasing after her screaming something like "Get the hell out of my race and give me those numbers!!" [NEWLINE] [NEWLINE] <mask> a lot of these vocal minorities are from a vastly different time that was not<mask> long ago.</s>
Label encoding: <s>We have to remember also that some of the older feminists are from a time where females really were *far* from being equal. It is easy for the younger ones amongst us to forget that. [NEWLINE] [NEWLINE] In the 60's (67 if my memory doesn't fail me) the first female marathon runner competed in the Boston Runner. She wasn't actually even allowed, it wasn't until 5 years later women were allowed to compete. And this runner had the organiser(/owner?) chasing after her screaming something like "Get the hell out of my race and give me those numbers!!" [NEWLINE] [NEWLINE] So a lot of these vocal minorities are from a vastly different time that was not so long ago.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0727, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I'm trying to say, is that people should not be forced into believing and supporting gay marriage, gay/lesbian/transgender/pansexual and all that—they should be allowed to express their freedom of religion... [ENDQ] [NEWLINE] People are already free to practice their religion, they don't have to support LGBT people and they're not being forced to either.<mask> they can't do, is impose their religious beliefs on other people, they can't get their religious beliefs enshrined into law. Religious people are in no way being oppressed by two gay people getting married, they just aren't allowed to dictate the rules on who gets to get married based on their religion. [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] What I'm trying to say, is that people should not be forced into believing and supporting gay marriage, gay/lesbian/transgender/pansexual and all that—they should be allowed to express their freedom of religion... [ENDQ] [NEWLINE] People are already free to practice their religion, they don't have to support LGBT people and they're not being forced to either. What they can't do, is impose their religious beliefs on other people, they can't get their religious beliefs enshrined into law. Religious people are in no way being oppressed by two gay people getting married, they just aren't allowed to dictate the rules on who gets to get married based on their religion. [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>bullshit they knew<mask> they were getting into, the only way they knew is<mask> you had a identical twin that was born earlier, [NEWLINE] [NEWLINE] your messing with the wrong "score" [NEWLINE] [NEWLINE] its like [NEWLINE] [NEWLINE] 'human' 5 respect + "parent" 10 respect =  15 respect  - "alcoholic" 3 respect etc : final respect = [NEWLINE] [NEWLINE] now parent variable can be "parent" "great parent" "amazing parent" etc,<mask> the flaws they have are subtracted after,<mask> the flaws they have are human things, not parent things, [NEWLINE] [NEWLINE] even<mask> your parents abandoned you in an orphanage at birth they would still have the "parent" 10 respect</s>
Label encoding: <s>bullshit they knew what they were getting into, the only way they knew is if you had a identical twin that was born earlier, [NEWLINE] [NEWLINE] your messing with the wrong "score" [NEWLINE] [NEWLINE] its like [NEWLINE] [NEWLINE] 'human' 5 respect + "parent" 10 respect =  15 respect  - "alcoholic" 3 respect etc : final respect = [NEWLINE] [NEWLINE] now parent variable can be "parent" "great parent" "amazing parent" etc, but the flaws they have are subtracted after, because the flaws they have are human things, not parent things, [NEWLINE] [NEWLINE] even if your parents abandoned you in an orphanage at birth they would still have the "parent" 10 respect</s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0922, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0819, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0887, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Try thinking of it this way- is it unethical for a website to collect revenue from the advertiser by showing ads to people who have no intention of buying the advertised product? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The advertiser gives money to the website in exchange for presenting its ads to viewers who may buy the product or service in the ad. All users of adblocking software have decided to not purchase *any* products from ads on websites, and are<mask> firm in this decision they wish to not view the ads at all. An effort to bypass (by law or otherwise) adblocking software shortchanges the advertisers by making them pay for useless ad views to people who detest ads on websites. </s>
Label encoding: <s>Try thinking of it this way- is it unethical for a website to collect revenue from the advertiser by showing ads to people who have no intention of buying the advertised product? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The advertiser gives money to the website in exchange for presenting its ads to viewers who may buy the product or service in the ad. All users of adblocking software have decided to not purchase *any* products from ads on websites, and are so firm in this decision they wish to not view the ads at all. An effort to bypass (by law or otherwise) adblocking software shortchanges the advertisers by making them pay for useless ad views to people who detest ads on websites. </s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It most assuredly does, this is not even under question.<mask> women have a better success rate under gender-blind review then the study would be out claiming they have been discriminated against,<mask> is it not the same for men? [ENDQ] [NEWLINE] Whether women do better under gender blind review does not yield any insight into whether men do. There is only evidence of gender bias<mask> one gender improves significantly more than the other under gender blind review. The men did not receive a larger bonus than the women under the gender-blind review,<mask> there is no evidence of gender bias against men. The facts<mask> presented do not support the claim that women are favored over men by reviewers. </s>
Label encoding: <s> [STARTQ] It most assuredly does, this is not even under question. If women have a better success rate under gender-blind review then the study would be out claiming they have been discriminated against, why is it not the same for men? [ENDQ] [NEWLINE] Whether women do better under gender blind review does not yield any insight into whether men do. There is only evidence of gender bias if one gender improves significantly more than the other under gender blind review. The men did not receive a larger bonus than the women under the gender-blind review, so there is no evidence of gender bias against men. The facts as presented do not support the claim that women are favored over men by reviewers. </s>
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think of it like buying a novel hardcover vs waiting for paperback. Paperback releases of popular books often come out almost a full year after the book's initial release, and are significantly cheaper. [NEWLINE] [NEWLINE] Example: There are many anthologies of The Walking Dead that contain the first jillion or<mask> individual releases, and the value is much better than buying each comic.<mask><mask> you are spending $2.99 on an individual, new, comic, you are paying premium to get the story<mask> fast<mask> possible,<mask> opposed to waiting for the anthology (or paperback). [NEWLINE] [NEWLINE] <mask> you're right, it's not a good value,<mask> that's not the point.</s>
Label encoding: <s>I think of it like buying a novel hardcover vs waiting for paperback. Paperback releases of popular books often come out almost a full year after the book's initial release, and are significantly cheaper. [NEWLINE] [NEWLINE] Example: There are many anthologies of The Walking Dead that contain the first jillion or so individual releases, and the value is much better than buying each comic. But when you are spending $2.99 on an individual, new, comic, you are paying premium to get the story as fast as possible, as opposed to waiting for the anthology (or paperback). [NEWLINE] [NEWLINE] So you're right, it's not a good value, but that's not the point.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I am a very strong proponent of the metric system, I take objection with your claim that Imperial is "useless".  Just like with any language, the things you say are only<mask> meaningful<mask> someone else's ability to understand them. <mask> no one I'm talking to understands<mask> a meter is,<mask> they all know<mask> a foot is, then telling them something in feet is going to be infinitely more useful than using meters. [NEWLINE] [NEWLINE] Yes, for calculation purposes, the metric system is obviously far simpler,<mask> communicating measurements to people can only work<mask> they understand<mask> you're saying.  Telling someone that something is 15 miles away works perfectly well.  </s>
Label encoding: <s>While I am a very strong proponent of the metric system, I take objection with your claim that Imperial is "useless".  Just like with any language, the things you say are only as meaningful as someone else's ability to understand them.  If no one I'm talking to understands what a meter is, but they all know what a foot is, then telling them something in feet is going to be infinitely more useful than using meters. [NEWLINE] [NEWLINE] Yes, for calculation purposes, the metric system is obviously far simpler, but communicating measurements to people can only work if they understand what you're saying.  Telling someone that something is 15 miles away works perfectly well.  </s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All races in the country are subject to poverty and oppression (to varying degrees)<mask> result of something.  Be it oppression due to race, socioeconomic class, greedy corporations, etc.  There are certainly white families who have been impoverished to the same level for generations, much like there are certainly black families who are wealthier than white families.  Given affirmative action laws today one could<mask><mask> it's more difficult for a white family to get out of poverty than a black family. [NEWLINE] [NEWLINE] I'm obviously not condoning oppression for anyone,<mask><mask><mask> with arguments that paint the problem<mask> a black vs white issue.  There should be programs that help the impoverished and oppressed equally.</s>
Label encoding: <s>All races in the country are subject to poverty and oppression (to varying degrees) as result of something.  Be it oppression due to race, socioeconomic class, greedy corporations, etc.  There are certainly white families who have been impoverished to the same level for generations, much like there are certainly black families who are wealthier than white families.  Given affirmative action laws today one could argue that it's more difficult for a white family to get out of poverty than a black family. [NEWLINE] [NEWLINE] I'm obviously not condoning oppression for anyone, but I disagree with arguments that paint the problem as a black vs white issue.  There should be programs that help the impoverished and oppressed equally.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "Equal rights, equal lefts" is a view usually espoused by anti-feminist MRA types. [ENDQ] [NEWLINE] Not sure<mask> that would matter. Just<mask> something is believed by bad people doesn't mean it's *wrong*. I'm sure they<mask> agree that Lincoln is the state capital of Nebraska,<mask> the legislative building isn't going to move to Omaha<mask> of that. [NEWLINE] [NEWLINE] I'm no MRA,<mask><mask> you get clobbered in a fist fight *you* started, you aren't going to get much sympathy from me. Especially<mask> you started a fight with somebody bigger than you. That's not changed by the gender configuration involved.</s>
Label encoding: <s> [STARTQ] "Equal rights, equal lefts" is a view usually espoused by anti-feminist MRA types. [ENDQ] [NEWLINE] Not sure why that would matter. Just because something is believed by bad people doesn't mean it's *wrong*. I'm sure they also agree that Lincoln is the state capital of Nebraska, but the legislative building isn't going to move to Omaha because of that. [NEWLINE] [NEWLINE] I'm no MRA, but if you get clobbered in a fist fight *you* started, you aren't going to get much sympathy from me. Especially if you started a fight with somebody bigger than you. That's not changed by the gender configuration involved.</s>
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Zero evidence should force us to re-evaluate the reasonable-ness of an explanation, no?<mask> I encounter an explanation and then observe zero evidence to favor it, I should scale my level of trust in that explanation to that amount of evidence. [NEWLINE] [NEWLINE] The explanations for fairy dust / unicorns / gods / "supernatural" consciousness are not all that trustworthy. You don't have to rule out something completely...we often don't get to operate with epistemological certainty<mask> that doesn't mean we can't make statements about unlikelihood of certain explanations. [NEWLINE] [NEWLINE] Agreed on questioning whether OP should award a delta at all. Seems like it's square 1 all over.</s>
Label encoding: <s>Zero evidence should force us to re-evaluate the reasonable-ness of an explanation, no? If I encounter an explanation and then observe zero evidence to favor it, I should scale my level of trust in that explanation to that amount of evidence. [NEWLINE] [NEWLINE] The explanations for fairy dust / unicorns / gods / "supernatural" consciousness are not all that trustworthy. You don't have to rule out something completely...we often don't get to operate with epistemological certainty but that doesn't mean we can't make statements about unlikelihood of certain explanations. [NEWLINE] [NEWLINE] Agreed on questioning whether OP should award a delta at all. Seems like it's square 1 all over.</s>
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0715, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>when did i say mens problems are fake or lies? just that generally they only apply to a very small number of men who decide to work in female dominated fields, which are female dominated<mask> women are generally forced into them<mask> they aren't considered able to do "mens jobs". this is a much much greater problem than the issue of tiny number of men who are treated badly (generally by other men)<mask> they decide to work in female dominated jobs. the whole paradigm is essentially created by male oppression of women in certain jobs and education fields.<mask> people such<mask> yourself expect these problems to receive the exact same consideration<mask> the much greater problem or else it's sexist.</s>
Label encoding: <s>when did i say mens problems are fake or lies? just that generally they only apply to a very small number of men who decide to work in female dominated fields, which are female dominated because women are generally forced into them as they aren't considered able to do "mens jobs". this is a much much greater problem than the issue of tiny number of men who are treated badly (generally by other men) because they decide to work in female dominated jobs. the whole paradigm is essentially created by male oppression of women in certain jobs and education fields. yet people such as yourself expect these problems to receive the exact same consideration as the much greater problem or else it's sexist.</s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0923, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Only the Sinai peninsula, the city of Quneitra, and Southern Lebanon were "returned" [ENDQ] [NEWLINE] Well, I said returned most of it, not all of it. [NEWLINE] [NEWLINE] [STARTQ] even then Sinai was returned on the condition that Egypt never stations military there without permission from Israel [ENDQ] [NEWLINE] <mask>? The Sinai peninsula is still theirs. [NEWLINE] [NEWLINE] [STARTQ] Every other territory occupied by Israel is still occupied and/or controlled, including the Golan Heights, the Gaza Strip and the West Bank. [ENDQ] [NEWLINE] You're right about the Golan Heights,<mask> Israel offered both Gaza and the West bank to their former owners - Egypt and Jordan respectively,<mask> they refused. </s>
Label encoding: <s> [STARTQ] Only the Sinai peninsula, the city of Quneitra, and Southern Lebanon were "returned" [ENDQ] [NEWLINE] Well, I said returned most of it, not all of it. [NEWLINE] [NEWLINE] [STARTQ] even then Sinai was returned on the condition that Egypt never stations military there without permission from Israel [ENDQ] [NEWLINE] So? The Sinai peninsula is still theirs. [NEWLINE] [NEWLINE] [STARTQ] Every other territory occupied by Israel is still occupied and/or controlled, including the Golan Heights, the Gaza Strip and the West Bank. [ENDQ] [NEWLINE] You're right about the Golan Heights, but Israel offered both Gaza and the West bank to their former owners - Egypt and Jordan respectively, but they refused. </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You know<mask>'s really sad? [NEWLINE] [NEWLINE] 1. the original meaning of coon.  Came from the word "baracoos" - which is a... slavic?  word for "pen" - aka, people kept in pens. [NEWLINE] [NEWLINE] 2. Coontown could be a super cute cuddly cartoon city full of cartoon raccoons<mask> you went on fantastical adventures.  Just like Great Apes could be a sub all about Gorillaz and monkeys! :D  OH MY GOD I'M DYING FROM THE CUTENESS IN MY IMAGINATION!!! [NEWLINE] [NEWLINE]...<mask> no.  It's not :(</s>
Label encoding: <s>You know what's really sad? [NEWLINE] [NEWLINE] 1. the original meaning of coon.  Came from the word "baracoos" - which is a... slavic?  word for "pen" - aka, people kept in pens. [NEWLINE] [NEWLINE] 2. Coontown could be a super cute cuddly cartoon city full of cartoon raccoons where you went on fantastical adventures.  Just like Great Apes could be a sub all about Gorillaz and monkeys! :D  OH MY GOD I'M DYING FROM THE CUTENESS IN MY IMAGINATION!!! [NEWLINE] [NEWLINE]... but no.  It's not :(</s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Oh yes, I believe this is the most important thing ever,<mask> I didn't tell you about it until you were 12<mask> it's controversial." &lt;-- crazy parent. [NEWLINE] [NEWLINE] You cannot choose to present<mask> you believe<mask> fact and VERY IMPORTANT<mask> some blah blah personal decision whoo wah that you need to be mature to make a decision on.<mask> you believe it's fact, plain and simple. It's unethical to present it that way, it's essentially lying by omission. You have to teach your children<mask> you believe, no matter<mask> that's morals, politics, religion, or whatever, even<mask> they're controversial. </s>
Label encoding: <s>"Oh yes, I believe this is the most important thing ever, but I didn't tell you about it until you were 12 because it's controversial." &lt;-- crazy parent. [NEWLINE] [NEWLINE] You cannot choose to present what you believe as fact and VERY IMPORTANT as some blah blah personal decision whoo wah that you need to be mature to make a decision on. Because you believe it's fact, plain and simple. It's unethical to present it that way, it's essentially lying by omission. You have to teach your children what you believe, no matter if that's morals, politics, religion, or whatever, even if they're controversial. </s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] that the harm the surgery causes is more based on the social tension it creates [ENDQ] [NEWLINE] Um... Get over it? [NEWLINE] [NEWLINE] [STARTQ] that transgender people are putting themselves in harms way in a social sense and there is some precedent that we do try to prevent people from putting themselves in harm's way. Do we have a responsibility to prevent people from harming themselves? [ENDQ] [NEWLINE] <mask> society has a responsibility, it's to let other people do<mask> makes their lives better. They should stop feeling affected by things that objectively do not involve them. You're saying trans people should do with your bodies<mask> you say<mask> society gets the heebie jeebies over it? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] that the harm the surgery causes is more based on the social tension it creates [ENDQ] [NEWLINE] Um... Get over it? [NEWLINE] [NEWLINE] [STARTQ] that transgender people are putting themselves in harms way in a social sense and there is some precedent that we do try to prevent people from putting themselves in harm's way. Do we have a responsibility to prevent people from harming themselves? [ENDQ] [NEWLINE] If society has a responsibility, it's to let other people do what makes their lives better. They should stop feeling affected by things that objectively do not involve them. You're saying trans people should do with your bodies what you say because society gets the heebie jeebies over it? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You can fix this problem with the excellent old English custom of requiring the same attorneys to work alternately<mask> prosecutors and public defendants.<mask> this were adopted, a shortfall in.one means the other is equally short; no consultants needed. [ENDQ] [NEWLINE] <mask> do you compel attorneys to work these cases at all then? [NEWLINE] [NEWLINE] [STARTQ] An alternate fix<mask> we are price-fixing is to set the public defender's salary just above the private salary. There would be no shortage of lawyers queuing up to consult. Not that I like pricefixing<mask> there needntneedn't be any slavery. [ENDQ] [NEWLINE] And we spend $1 trillion to accomplish this or...?</s>
Label encoding: <s> [STARTQ] You can fix this problem with the excellent old English custom of requiring the same attorneys to work alternately as prosecutors and public defendants. If this were adopted, a shortfall in.one means the other is equally short; no consultants needed. [ENDQ] [NEWLINE] How do you compel attorneys to work these cases at all then? [NEWLINE] [NEWLINE] [STARTQ] An alternate fix if we are price-fixing is to set the public defender's salary just above the private salary. There would be no shortage of lawyers queuing up to consult. Not that I like pricefixing but there needntneedn't be any slavery. [ENDQ] [NEWLINE] And we spend $1 trillion to accomplish this or...?</s>
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They're not harassment<mask> of their volume. They're harassment<mask> of their quality.<mask> there were 100+ incidences of people stopping me in 10 hours to ask me for directions, that would be annoying. 100+ incidences of people specifically targeting me for sexual interaction (come on, now, be honest - you can hear the difference between "hello" and "<mask> *you* doin'"; calling her "girl"<mask> she's a full grown adult; and you know the difference between specifically referring to her<mask> "beautiful"<mask> actual friendly social interaction should not be based on the attractiveness of the girl, etc etc; right?) is harassment.</s>
Label encoding: <s>They're not harassment because of their volume. They're harassment because of their quality. If there were 100+ incidences of people stopping me in 10 hours to ask me for directions, that would be annoying. 100+ incidences of people specifically targeting me for sexual interaction (come on, now, be honest - you can hear the difference between "hello" and " how *you* doin'"; calling her "girl" when she's a full grown adult; and you know the difference between specifically referring to her as "beautiful" when actual friendly social interaction should not be based on the attractiveness of the girl, etc etc; right?) is harassment.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, are you suggesting in your first point that<mask> my kids want to be Buddhists or Zoroastrians or Jews or whatever that you think that they shouldn't be allowed to? [NEWLINE] [NEWLINE] Second of all, you contradict yourself. You say that I have a right to remain ignorant<mask> that my children don't? Now, I'm not saying that that's an easily defensible right,<mask> in a free society I should be able to believe whatever I want,<mask> you suggest indoctrinating children to believe in one particular mindset and that mindset only. Now in this case,<mask><mask> with that mindset,<mask> that doesn't negate the flimsy reasoning.</s>
Label encoding: <s>First of all, are you suggesting in your first point that if my kids want to be Buddhists or Zoroastrians or Jews or whatever that you think that they shouldn't be allowed to? [NEWLINE] [NEWLINE] Second of all, you contradict yourself. You say that I have a right to remain ignorant but that my children don't? Now, I'm not saying that that's an easily defensible right, but in a free society I should be able to believe whatever I want, but you suggest indoctrinating children to believe in one particular mindset and that mindset only. Now in this case, I agree with that mindset, but that doesn't negate the flimsy reasoning.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am unsure<mask> the adjective "scientific" can be applied to any of<mask> you posted. I didn'tt see anything empirical in your post. You haven't justified that a) people wouldn't eat a dog<mask> they are intelligent (<mask>,<mask> this is a descriptive statement is blatantly wrong; people *do* eat dogs. Some people specifically don't eat cows, and pigs are more intelligent than cats, dogs, or horses.) Or that b) complexity of experience is morally important. I'm especially unsure<mask> one would "scientifically" arrive at b. Note that demonstrating a doesn't prove b,<mask> that would constitute an appeal to majority.</s><pad>
Label encoding: <s>I am unsure how the adjective "scientific" can be applied to any of what you posted. I didn'tt see anything empirical in your post. You haven't justified that a) people wouldn't eat a dog because they are intelligent ( moreover, since this is a descriptive statement is blatantly wrong; people *do* eat dogs. Some people specifically don't eat cows, and pigs are more intelligent than cats, dogs, or horses.) Or that b) complexity of experience is morally important. I'm especially unsure how one would "scientifically" arrive at b. Note that demonstrating a doesn't prove b, as that would constitute an appeal to majority.</s><pad>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Whilst I'd need to be a military attorney to know whether no longer following orders makes one by definition no longer a member of the military, your point is irrelevant,<mask> the facts remain: [NEWLINE] [NEWLINE] 1. There is nothing which makes soldiers incapable of disobeying orders such that they are no longer responsible for their actions. [NEWLINE] [NEWLINE] 2. <mask> there were, then they'd be responsible for having knowingly put themselves in that situation. [NEWLINE] [NEWLINE] I don't get to swear to do whatever a known gang leader tells me to do, and then act like it wasn't my fault I murdered someone<mask> he told me to. [NEWLINE] [NEWLINE] edit:typo</s>
Label encoding: <s>Whilst I'd need to be a military attorney to know whether no longer following orders makes one by definition no longer a member of the military, your point is irrelevant, because the facts remain: [NEWLINE] [NEWLINE] 1. There is nothing which makes soldiers incapable of disobeying orders such that they are no longer responsible for their actions. [NEWLINE] [NEWLINE] 2.  If there were, then they'd be responsible for having knowingly put themselves in that situation. [NEWLINE] [NEWLINE] I don't get to swear to do whatever a known gang leader tells me to do, and then act like it wasn't my fault I murdered someone when he told me to. [NEWLINE] [NEWLINE] edit:typo</s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The entire premise revolves around a villain hacking Batman's computer and stealing his contingency plans for<mask> to neutralize the Justice League. [ENDQ] [NEWLINE] [NEWLINE] Are you saying that Batman came up with plans to take out the rest of the Justice League in case they ever turned on him/each other? That's pretty awesome. [NEWLINE] [NEWLINE] <mask><mask> instances like this are<mask> he proves his worth<mask> a superhero. He takes an outside-the-box approach and is always prepared for everything. Other guys like Superman can run into the action with full force,<mask> often times leave a lot of collateral damage. Batman can save the day without anyone ever knowing he was there. [NEWLINE] </s>
Label encoding: <s> [STARTQ] The entire premise revolves around a villain hacking Batman's computer and stealing his contingency plans for how to neutralize the Justice League. [ENDQ] [NEWLINE] [NEWLINE] Are you saying that Batman came up with plans to take out the rest of the Justice League in case they ever turned on him/each other? That's pretty awesome. [NEWLINE] [NEWLINE] I think instances like this are where he proves his worth as a superhero. He takes an outside-the-box approach and is always prepared for everything. Other guys like Superman can run into the action with full force, but often times leave a lot of collateral damage. Batman can save the day without anyone ever knowing he was there. [NEWLINE] </s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Learning to accept these things was<mask> the book was about [ENDQ] [NEWLINE] Citation needed. [NEWLINE] [NEWLINE] <mask> you want to personally read it<mask> a condolence of torture and oppression then I guess I can't stop you,<mask> that certainly isn't a factual claim. Orwell was vehemently anti-soviet and anti-fascist.  There is simply no way you can claim that "learning to accept" these things was the authorial intent of the novel,<mask><mask> you wish to read it this way then I can't stop you anymore than I can stop you from thinking that Star Wars was really about<mask> Emperor Palpatine wasn't such a bad guy.</s>
Label encoding: <s> [STARTQ] Learning to accept these things was what the book was about [ENDQ] [NEWLINE] Citation needed. [NEWLINE] [NEWLINE] If you want to personally read it as a condolence of torture and oppression then I guess I can't stop you, but that certainly isn't a factual claim. Orwell was vehemently anti-soviet and anti-fascist.  There is simply no way you can claim that "learning to accept" these things was the authorial intent of the novel, however if you wish to read it this way then I can't stop you anymore than I can stop you from thinking that Star Wars was really about how Emperor Palpatine wasn't such a bad guy.</s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You believe humans don't have an innate sense of ethics about us?<mask><mask> I argued nearly every established society has produced generally the same ethics, and then said<mask> of that, those general ethics are ones that should be considered innate? That's<mask> I would immediate think. Sure, people are brought up by their surroundings,<mask> it's not just arbitrary, nothing-really-matters, hitler is cool<mask> his society deemed him ethical, bullshit. Come on now! [NEWLINE] [NEWLINE] Maltreatment is maltreatment, and people generally have a sense of sanctity (about<mask> doesn't matter for now), would you at least agree to that? [NEWLINE] </s>
Label encoding: <s>You believe humans don't have an innate sense of ethics about us? What if I argued nearly every established society has produced generally the same ethics, and then said because of that, those general ethics are ones that should be considered innate? That's what I would immediate think. Sure, people are brought up by their surroundings, but it's not just arbitrary, nothing-really-matters, hitler is cool because his society deemed him ethical, bullshit. Come on now! [NEWLINE] [NEWLINE] Maltreatment is maltreatment, and people generally have a sense of sanctity (about what doesn't matter for now), would you at least agree to that? [NEWLINE] </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am a vegan and I argued the opposite on here recently, that anti-abortion people are hypocritical for not being vegetarian. [NEWLINE] [NEWLINE] To me, it makes little sense<mask> to<mask> one would find aborting a fetus, which is not conscious and does not have feelings and emotions in early stages, abhorrent,<mask> be perfectly fine with the brutal torture and murder that happens to animals that are fully conscious and feel everything. [NEWLINE] [NEWLINE] <mask>, I am pro choice,<mask> I still don't like abortion and don't personally condone it.  Similarly, I am vegan and against aggression towards animals<mask> I don't want to force everyone to be like me.</s>
Label encoding: <s>I am a vegan and I argued the opposite on here recently, that anti-abortion people are hypocritical for not being vegetarian. [NEWLINE] [NEWLINE] To me, it makes little sense as to why one would find aborting a fetus, which is not conscious and does not have feelings and emotions in early stages, abhorrent, yet be perfectly fine with the brutal torture and murder that happens to animals that are fully conscious and feel everything. [NEWLINE] [NEWLINE] Also, I am pro choice, but I still don't like abortion and don't personally condone it.  Similarly, I am vegan and against aggression towards animals but I don't want to force everyone to be like me.</s>
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Akoustyk, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Akoustyk+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry Akoustyk, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Akoustyk+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Jrook, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Jrook+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry Jrook, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=Jrook+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sentencing is in general done by judges,<mask> there are actually eight states that allow for [jury sentencing even in non-capital cases]( [URL].cgi?article=1139&amp;context=urbanlaw). [NEWLINE] [NEWLINE] Jury sentencing tends to be harsher<mask> juries don't know<mask> typical punishments are and usually go for the "middle" of the penalty range, whereas judges will usually go closer to the bottom of the range. Juries just see that 5-40 year range and figure they're being lenient by giving 20 years. The result is in the southern state that allow jury sentencing, there are much fewer jury trials. </s>
Label encoding: <s>Sentencing is in general done by judges, but there are actually eight states that allow for [jury sentencing even in non-capital cases]( [URL].cgi?article=1139&amp;context=urbanlaw). [NEWLINE] [NEWLINE] Jury sentencing tends to be harsher because juries don't know what typical punishments are and usually go for the "middle" of the penalty range, whereas judges will usually go closer to the bottom of the range. Juries just see that 5-40 year range and figure they're being lenient by giving 20 years. The result is in the southern state that allow jury sentencing, there are much fewer jury trials. </s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of individual feminists are decent human beings and<mask> such care about men too.<mask><mask> you look at feminist theory it does tend to sideline male issues per default. Women are seen<mask> *the* oppressed gender and men are *the* privileged gender. And in the cases<mask> men suffer from sexist norms (of which feminists are usually aware of only a very small part) it is seen<mask> an accidental side effect of the more important and widespread women's issues. [NEWLINE] [NEWLINE] <mask><mask> most feminists will admit that men have *some* issues; on the whole it's a very binary way of thinking about gender discrimination<mask> women are put first. </s>
Label encoding: <s>A lot of individual feminists are decent human beings and as such care about men too. But if you look at feminist theory it does tend to sideline male issues per default. Women are seen as *the* oppressed gender and men are *the* privileged gender. And in the cases where men suffer from sexist norms (of which feminists are usually aware of only a very small part) it is seen as an accidental side effect of the more important and widespread women's issues. [NEWLINE] [NEWLINE] Even though most feminists will admit that men have *some* issues; on the whole it's a very binary way of thinking about gender discrimination where women are put first. </s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It will take several mountains of evidence that cannot be refuted before I believe that bad handwriting is something that can't be fixed in someone. [NEWLINE] [NEWLINE] I'll accept that some people are predisposed to better motor skills,<mask> to suggest that some people will have bad handwriting no matter<mask> is insane (don't you dare compare it to depression). It makes more sense to me that they're predisposed to start off worse and before they make any real progress they become demoralised,<mask> they don't put effort in. [NEWLINE] [NEWLINE] You know, the same reason people are better and worse at literally ANYTHING that has something to do with muscle control.</s>
Label encoding: <s>It will take several mountains of evidence that cannot be refuted before I believe that bad handwriting is something that can't be fixed in someone. [NEWLINE] [NEWLINE] I'll accept that some people are predisposed to better motor skills, but to suggest that some people will have bad handwriting no matter what is insane (don't you dare compare it to depression). It makes more sense to me that they're predisposed to start off worse and before they make any real progress they become demoralised, so they don't put effort in. [NEWLINE] [NEWLINE] You know, the same reason people are better and worse at literally ANYTHING that has something to do with muscle control.</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>FallingSnowAngel wasn't making an assumption, they're arguing that the taboo exists to deter people from abusing power they might have over their family members. You're talking about individual cases and they're talking about things at a societal level. [NEWLINE] [NEWLINE] <mask>, the second paragraph directly addresses your point: [NEWLINE] [NEWLINE] [STARTQ] The taboo isn't supposed to judge those in loving, compassionate relationships. It's just meant to make the possibility for abuse more difficult... [ENDQ] [NEWLINE] <mask><mask> it's very similar to students who end up in relationships with their teachers. There's a taboo<mask> of the increased possibility of abuse,<mask><mask> whether or not individual relationships are abusive.</s>
Label encoding: <s>FallingSnowAngel wasn't making an assumption, they're arguing that the taboo exists to deter people from abusing power they might have over their family members. You're talking about individual cases and they're talking about things at a societal level. [NEWLINE] [NEWLINE] Also, the second paragraph directly addresses your point: [NEWLINE] [NEWLINE] [STARTQ] The taboo isn't supposed to judge those in loving, compassionate relationships. It's just meant to make the possibility for abuse more difficult... [ENDQ] [NEWLINE] I think it's very similar to students who end up in relationships with their teachers. There's a taboo because of the increased possibility of abuse, regardless of whether or not individual relationships are abusive.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0777, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like your different approach to the issue. I can see that in my goal of supporting myself and loved ones<mask> we're alive, a charity's work could play a role. [NEWLINE] [NEWLINE] I'm<mask> open to finding a different balance of saving vs. giving now.<mask><mask> it's mostly the uncertainty of health care costs that makes me feel like I need to save<mask> much<mask> possible...especially for end of life care which can rise to unimaginable sums. I wouldn't want whoever survives me to be on the hook for that. [NEWLINE] [NEWLINE] <mask> again, for making my position more flexible, you<mask> get a &amp;#8710;</s>
Label encoding: <s>I like your different approach to the issue. I can see that in my goal of supporting myself and loved ones while we're alive, a charity's work could play a role. [NEWLINE] [NEWLINE] I'm also open to finding a different balance of saving vs. giving now. I think it's mostly the uncertainty of health care costs that makes me feel like I need to save as much as possible...especially for end of life care which can rise to unimaginable sums. I wouldn't want whoever survives me to be on the hook for that. [NEWLINE] [NEWLINE] But again, for making my position more flexible, you also get a &amp;#8710;</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can't make an argument from philosophy and<mask> you the very practical examples of the media that people are experiencing right now. [NEWLINE] [NEWLINE] It is almost like your view is I would eat<mask> much better<mask> I had a private chef cater my meals to my personal wishes. And then<mask> someone counters with, "Can you afford that?" you retreat back into the philosophical realm. [NEWLINE] [NEWLINE] <mask> you are going to use real world examples to support your philosophical beliefs. then we can counter with real world counter arguments.<mask> you just want to wrap yourself up in philosophy then ditch your real world support. You're kinda muddling the middle here. </s><pad>
Label encoding: <s>You can't make an argument from philosophy and also you the very practical examples of the media that people are experiencing right now. [NEWLINE] [NEWLINE] It is almost like your view is I would eat so much better if I had a private chef cater my meals to my personal wishes. And then when someone counters with, "Can you afford that?" you retreat back into the philosophical realm. [NEWLINE] [NEWLINE] If you are going to use real world examples to support your philosophical beliefs. then we can counter with real world counter arguments. If you just want to wrap yourself up in philosophy then ditch your real world support. You're kinda muddling the middle here. </s><pad>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agree.  Donating to charity<mask> OP suggests to "better the world" can<mask> be fairly risky.  Some charities are simply shit, and squander most of the money given to them.  Others have their hearts in the right places<mask> are simply going down a dead end (for example, it's hard to know<mask> kinds of renewable energy research to donate to<mask> it isn't at all clear which has the best chance of taking off and becoming a substantial contributor to our future energy needs).  It's not out of the question that some charities are actively damaging society in either direct or indirect ways with money you give them.</s>
Label encoding: <s>Agree.  Donating to charity as OP suggests to "better the world" can also be fairly risky.  Some charities are simply shit, and squander most of the money given to them.  Others have their hearts in the right places but are simply going down a dead end (for example, it's hard to know what kinds of renewable energy research to donate to when it isn't at all clear which has the best chance of taking off and becoming a substantial contributor to our future energy needs).  It's not out of the question that some charities are actively damaging society in either direct or indirect ways with money you give them.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe it's<mask> gun control, pre-Columbine, wasn't such a hot political topic that these things become bigger news than they did before. My point is<mask>, school shootings per capita haven't really increased<mask> of media coverage. Kids have always gotten bullied, humiliated, jealous, angry and have acted out in an irrational, murderous manner.<mask> anything, general awareness has helped us be able to spot a kid that is this troubled and prevent the tragedy from coming to fruition. Even better, we've gained a lot of understanding<mask> to<mask> kids get this troubled to begin with, possibly preventing the underlying frustration and humiliation altogether. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Maybe it's because gun control, pre-Columbine, wasn't such a hot political topic that these things become bigger news than they did before. My point is though, school shootings per capita haven't really increased because of media coverage. Kids have always gotten bullied, humiliated, jealous, angry and have acted out in an irrational, murderous manner. If anything, general awareness has helped us be able to spot a kid that is this troubled and prevent the tragedy from coming to fruition. Even better, we've gained a lot of understanding as to how kids get this troubled to begin with, possibly preventing the underlying frustration and humiliation altogether. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Given the nature of humans, someone will always rule.  The rulers will always be coercive in some fashion.  The rulers will always act in unjust ways<mask> only<mask> one groups justice is another groups injustice. [NEWLINE] [NEWLINE] Governments have evolved over the centuries in part to try to mitigated injustice to some extent.  This mitigation is by no means perfect and of course favors those in power. <mask> it is better than being governed by warlords and their gangs of thugs.  Under anarchy there is nothing to prevent warlords from forming their gangs and doing their thing. [NEWLINE] [NEWLINE] <mask> bad<mask> government is no government is worse. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Given the nature of humans, someone will always rule.  The rulers will always be coercive in some fashion.  The rulers will always act in unjust ways if only because one groups justice is another groups injustice. [NEWLINE] [NEWLINE] Governments have evolved over the centuries in part to try to mitigated injustice to some extent.  This mitigation is by no means perfect and of course favors those in power.  However it is better than being governed by warlords and their gangs of thugs.  Under anarchy there is nothing to prevent warlords from forming their gangs and doing their thing. [NEWLINE] [NEWLINE] As bad as government is no government is worse. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In the context of this problem, a country without a universal healthcare system would not pay for any of the medical costs of people that injured themselves via failed suicide attempts. <mask> there would be no financial advantage to preventing these people from attempting suicide (from the governments perspective)<mask> they wouldn't be paying for it anyway. [NEWLINE] [NEWLINE] <mask> this argument (and all other arguments about suicide prevention) really only apply to systems<mask> the government is looking after the poor people and has some form of universal healthcare system in place (in the US medicaid and EMTALA, Canada, UK, most European countries and Australia and their universal healthcare etc).</s>
Label encoding: <s>In the context of this problem, a country without a universal healthcare system would not pay for any of the medical costs of people that injured themselves via failed suicide attempts.  So there would be no financial advantage to preventing these people from attempting suicide (from the governments perspective) because they wouldn't be paying for it anyway. [NEWLINE] [NEWLINE] Therefore this argument (and all other arguments about suicide prevention) really only apply to systems where the government is looking after the poor people and has some form of universal healthcare system in place (in the US medicaid and EMTALA, Canada, UK, most European countries and Australia and their universal healthcare etc).</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Couldn't agree more. <mask> I<mask> don't understand<mask> people think they won't get good service<mask> they pay for it. <mask>'s the problem with telling a server, "here's<mask> I want, here's<mask> I'll give you<mask> you do it".  That's<mask> every other job works.  The original comment seems to ignore the fact that<mask> a customer, you have the power to say something<mask> the service style is not to your liking.  It's not a big deal. <mask><mask><mask> you pay, you'll be fine.  Pretty simple concept that people blow out the water.</s>
Label encoding: <s>Couldn't agree more.  But I also don't understand how people think they won't get good service if they pay for it.  What's the problem with telling a server, "here's what I want, here's what I'll give you if you do it".  That's how every other job works.  The original comment seems to ignore the fact that as a customer, you have the power to say something if the service style is not to your liking.  It's not a big deal.  As long as you pay, you'll be fine.  Pretty simple concept that people blow out the water.</s>
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But<mask>? Just<mask> she says she's going to keep it itms suddenly a person? Whether something's a living person or not had nothing to do with intent. My mom may intend to kill me tomorrow, I'm still alive<mask> she does it. She doesn't have the right to kill a person. Whether she thinks I'm a person or not doesn't change the fact that I am. [NEWLINE] [NEWLINE] In the same sense the fetus is either a person or not.<mask> it's not a person<mask> the mother decides to keep it, it's still not a person<mask>. Her decision changed nothing physically about the fetus. </s>
Label encoding: <s>But why? Just because she says she's going to keep it itms suddenly a person? Whether something's a living person or not had nothing to do with intent. My mom may intend to kill me tomorrow, I'm still alive when she does it. She doesn't have the right to kill a person. Whether she thinks I'm a person or not doesn't change the fact that I am. [NEWLINE] [NEWLINE] In the same sense the fetus is either a person or not. If it's not a person but the mother decides to keep it, it's still not a person yet. Her decision changed nothing physically about the fetus. </s>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Let's assume that the cut hours were only 90%<mask> productive<mask> the rest of the hours. That is still a large drop in overall productivity. [ENDQ] [NEWLINE] One of my main points is that I feel you can cut some hours without losing much productivity.  You are just providing a contrary assumption with no evidence. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> I'm making 50k this year, I'd have to expect to be making 50k 10 years from now<mask> well? I'd be unhappy with that scenario. [ENDQ] [NEWLINE] Even<mask> it meant 3 day weekends or 6 hour work days?  Personally, I'd be happy with that scenario.</s>
Label encoding: <s> [STARTQ] Let's assume that the cut hours were only 90% as productive as the rest of the hours. That is still a large drop in overall productivity. [ENDQ] [NEWLINE] One of my main points is that I feel you can cut some hours without losing much productivity.  You are just providing a contrary assumption with no evidence. [NEWLINE] [NEWLINE] [STARTQ] So if I'm making 50k this year, I'd have to expect to be making 50k 10 years from now as well? I'd be unhappy with that scenario. [ENDQ] [NEWLINE] Even if it meant 3 day weekends or 6 hour work days?  Personally, I'd be happy with that scenario.</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] they don't think they about anyone<mask> themselves and go kill themselves without thinking of<mask> it affects people around them and<mask> consequences will come to others by their actions. [ENDQ] [NEWLINE] Some people do think it over and carry it out. It's better of them to ease out of personal bonds or to wait until loved ones die,<mask> there are surely cases in which it's preferable to commit suicide earlier<mask> of the suffering that would be experienced in the interim. [NEWLINE] [NEWLINE] Another problem I have with your account is that you view life<mask> an opportunity rather than a burden. Do we have better reason to think of it<mask> an opportunity?</s>
Label encoding: <s> [STARTQ] they don't think they about anyone but themselves and go kill themselves without thinking of how it affects people around them and what consequences will come to others by their actions. [ENDQ] [NEWLINE] Some people do think it over and carry it out. It's better of them to ease out of personal bonds or to wait until loved ones die, but there are surely cases in which it's preferable to commit suicide earlier because of the suffering that would be experienced in the interim. [NEWLINE] [NEWLINE] Another problem I have with your account is that you view life as an opportunity rather than a burden. Do we have better reason to think of it as an opportunity?</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I am a parent and my child needs a kidney and I am the only match, I am still not obligated by law to be forced to allow my dependent child to use my body to survive. Just being a dependent does not mean that the person who the dependent is dependent on must sacrifice or in any way use their own body and give up their own body rights to ensure the survival of the dependent.<mask> my son were to rape me, that would still be illegal<mask> I do not consent to him using my body that way (<mask> well<mask> due to incest laws,<mask> it would be illegal without those anyway). [ENDQ] [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] If I am a parent and my child needs a kidney and I am the only match, I am still not obligated by law to be forced to allow my dependent child to use my body to survive. Just being a dependent does not mean that the person who the dependent is dependent on must sacrifice or in any way use their own body and give up their own body rights to ensure the survival of the dependent. If my son were to rape me, that would still be illegal because I do not consent to him using my body that way ( as well as due to incest laws, but it would be illegal without those anyway). [ENDQ] [NEWLINE] </s><pad>
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One thing missing from your list is holiday/vacation days. [NEWLINE] [NEWLINE] I'm from the UK, and my brother moved to the US, and we work in similar jobs (professional, computer-based, in an office). [NEWLINE] [NEWLINE] He earns about 50% more than me,<mask> I have three times<mask> much holiday<mask> him. [NEWLINE] [NEWLINE] He has two weeks (10 days) of holiday per *year*. That's pretty appalling by UK standards. I have 30 days. [NEWLINE] [NEWLINE] To me, it doesn't make much sense to have a good income<mask> you don't have the time to use it. </s>
Label encoding: <s>One thing missing from your list is holiday/vacation days. [NEWLINE] [NEWLINE] I'm from the UK, and my brother moved to the US, and we work in similar jobs (professional, computer-based, in an office). [NEWLINE] [NEWLINE] He earns about 50% more than me, but I have three times as much holiday as him. [NEWLINE] [NEWLINE] He has two weeks (10 days) of holiday per *year*. That's pretty appalling by UK standards. I have 30 days. [NEWLINE] [NEWLINE] To me, it doesn't make much sense to have a good income if you don't have the time to use it. </s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are aware that people put their butt on things at the gym, right? And there is a good reason called "urinary tract infection" that makes me not want to put my lady parts on a surface<mask> potentially somebody placed their unclean butt before me. Especially<mask> there is a certain amount of pressure from the task that pushes my body on or into the seat or rubs it against it. No, thank you. [NEWLINE] <mask>, most women have a certain amount of discharge all the time, not just<mask> they are on your period. I really don't need that to come in contact with any sporting equipment. </s>
Label encoding: <s>You are aware that people put their butt on things at the gym, right? And there is a good reason called "urinary tract infection" that makes me not want to put my lady parts on a surface where potentially somebody placed their unclean butt before me. Especially when there is a certain amount of pressure from the task that pushes my body on or into the seat or rubs it against it. No, thank you. [NEWLINE] Also, most women have a certain amount of discharge all the time, not just when they are on your period. I really don't need that to come in contact with any sporting equipment. </s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Every time I've seen a feminist say that, they were quickly revealed to be a troll trying to discredit the movement. [NEWLINE] [NEWLINE] [STARTQ] <mask>, anytime men try to bring up the fact that men are sexually assaulted and victims of domestic violence, feminists will simply say that men are "derailing" feminist discussions. [ENDQ] [NEWLINE] Not all mentions of male rape are used to derail conversations,<mask> some are. It's gotten to the point<mask> some male rape victims have had to explicitly call out MRAs for using the topic of male rape<mask> a means of silencing or hijacking discussions of female rape (or consent in general).</s>
Label encoding: <s>Every time I've seen a feminist say that, they were quickly revealed to be a troll trying to discredit the movement. [NEWLINE] [NEWLINE] [STARTQ] Also, anytime men try to bring up the fact that men are sexually assaulted and victims of domestic violence, feminists will simply say that men are "derailing" feminist discussions. [ENDQ] [NEWLINE] Not all mentions of male rape are used to derail conversations, but some are. It's gotten to the point where some male rape victims have had to explicitly call out MRAs for using the topic of male rape as a means of silencing or hijacking discussions of female rape (or consent in general).</s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We could do that,<mask><mask> should we?<mask> my wife ever needs more money,<mask> her car breaks down or something like that, then of course I will help her.<mask> there is no need for us to have the same amount of money after bills. Not like we would spend that free money every month. [NEWLINE] [NEWLINE] We already help each other out, that's<mask> we don't split bills 50/50, that's<mask> we both cook and clean, that's<mask> she is the one decorating everything,<mask> I take care of the tech stuff.<mask> those are all things that need to be done.</s>
Label encoding: <s>We could do that, but why should we? If my wife ever needs more money, because her car breaks down or something like that, then of course I will help her. But there is no need for us to have the same amount of money after bills. Not like we would spend that free money every month. [NEWLINE] [NEWLINE] We already help each other out, that's why we don't split bills 50/50, that's why we both cook and clean, that's why she is the one decorating everything, while I take care of the tech stuff. But those are all things that need to be done.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How do you decide<mask> a law is working or not? [NEWLINE] [NEWLINE] I laid out<mask> my law could be enforced, and<mask><mask> people in this thread have listed examples of very similar ideas that are already in place today. They have monitors which will inform the court<mask> a person consumes alcohol, and they do weekly pee tests on some people to ensure they do not consume drugs and alcohol. One person said that in Minnesota they even have a drivers license restriction for people with multiple DUIs that forces them to be alcohol and drug free, and makes it<mask> bars cannot serve them, just like<mask> I proposed. [URL] /</s>
Label encoding: <s>How do you decide if a law is working or not? [NEWLINE] [NEWLINE] I laid out how my law could be enforced, and in fact people in this thread have listed examples of very similar ideas that are already in place today. They have monitors which will inform the court if a person consumes alcohol, and they do weekly pee tests on some people to ensure they do not consume drugs and alcohol. One person said that in Minnesota they even have a drivers license restriction for people with multiple DUIs that forces them to be alcohol and drug free, and makes it so bars cannot serve them, just like what I proposed. [URL] /</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> you know, the reason for the bad CGI (especially in the first season) is due to the way the show is funded. Here in America, shows are funded privately. Doctor Who is on BBC, which means it's entirely paid for by the British government through a television tax (or something similar to a tax). Each year they spend X amount on television shows and that's it.<mask> it got more popular, it got more funding. And here's something different too, the merchandise doesn't pay for the show either. That money goes to BBC Worldwide, which is a different division of the BBC.</s>
Label encoding: <s>Just so you know, the reason for the bad CGI (especially in the first season) is due to the way the show is funded. Here in America, shows are funded privately. Doctor Who is on BBC, which means it's entirely paid for by the British government through a television tax (or something similar to a tax). Each year they spend X amount on television shows and that's it. As it got more popular, it got more funding. And here's something different too, the merchandise doesn't pay for the show either. That money goes to BBC Worldwide, which is a different division of the BBC.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with most of<mask> you said.<mask> I have done my research on the topic and read that med schools are in no shortage of corpses. [NEWLINE] [NEWLINE] <mask> we *are* lacking,<mask>, are organ donors. The problem is that you have to die in a hospital, in order for the docs to preserve your innards. Once they've gutted you, they'll sew you back up and stick you in an oven or in the ground. Whatever you prefer. [NEWLINE] [NEWLINE] That way, you might save a life by giving away your heart or something and your family will have a place to grieve.</s>
Label encoding: <s>I agree with most of what you said. So I have done my research on the topic and read that med schools are in no shortage of corpses. [NEWLINE] [NEWLINE] What we *are* lacking, however, are organ donors. The problem is that you have to die in a hospital, in order for the docs to preserve your innards. Once they've gutted you, they'll sew you back up and stick you in an oven or in the ground. Whatever you prefer. [NEWLINE] [NEWLINE] That way, you might save a life by giving away your heart or something and your family will have a place to grieve.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ok, I'll bite. Lets use Super Bowl<mask> an example<mask> you mentioned. The premise of this season is to get first place, no one else is rewarded,<mask> should your kid (even second place) get something? Again, I'm not trying to be cold<mask><mask><mask> children all deserve to be happy I'm just curious<mask> to<mask> someone should earn a trophy for 8th place? in the real world 4th place gets you laid off<mask> you didn't try hard enough. I guess I'm trying to get the mentality of a parent to see<mask> they do this for their kid? </s>
Label encoding: <s>ok, I'll bite. Lets use Super Bowl as an example as you mentioned. The premise of this season is to get first place, no one else is rewarded, why should your kid (even second place) get something? Again, I'm not trying to be cold as I think children all deserve to be happy I'm just curious as to why someone should earn a trophy for 8th place? in the real world 4th place gets you laid off because you didn't try hard enough. I guess I'm trying to get the mentality of a parent to see why they do this for their kid? </s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There have already been two Presidents who were children of a former President, one who was a grandson of a former President, two additional ones who were closely related to a former President, and several more who came from political families. This is nothing new to American politics. And<mask>, someone like Obama could still get elected with no dynasty behind him whatsoever. [NEWLINE] [NEWLINE] One of the things we have that your elected kings did not was term limits. Let's say Clinton or Bush wins the Presidency. Who runs against him in 2020? Who gets the Presidency in 2024? Neither family will have an especially qualified candidate then. </s>
Label encoding: <s>There have already been two Presidents who were children of a former President, one who was a grandson of a former President, two additional ones who were closely related to a former President, and several more who came from political families. This is nothing new to American politics. And yet, someone like Obama could still get elected with no dynasty behind him whatsoever. [NEWLINE] [NEWLINE] One of the things we have that your elected kings did not was term limits. Let's say Clinton or Bush wins the Presidency. Who runs against him in 2020? Who gets the Presidency in 2024? Neither family will have an especially qualified candidate then. </s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> this education is good their notion will match the actual law [ENDQ] [NEWLINE] This will be true<mask> they actually care,<mask> not<mask> they don't.  You can educate a bad cop all day long about<mask> rights someone is supposed to have.  I doubt there are too many police officers out there that genuinely believe (especially these days) that people are not allowed to videotape the police.  Every cop alive knows that that's a protected right, and<mask> every week you see a cop confiscating someone's phone, breaking it, or straight up assaulting the person doing the taping.  </s>
Label encoding: <s> [STARTQ] If this education is good their notion will match the actual law [ENDQ] [NEWLINE] This will be true if they actually care, but not if they don't.  You can educate a bad cop all day long about what rights someone is supposed to have.  I doubt there are too many police officers out there that genuinely believe (especially these days) that people are not allowed to videotape the police.  Every cop alive knows that that's a protected right, and yet every week you see a cop confiscating someone's phone, breaking it, or straight up assaulting the person doing the taping.  </s>
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that some instances require those kinds of actions--for instance,<mask><mask> [this]( [URL] ) was the right response.  The trooper was fired and is being charged. <mask> not every example is quite this easy to pick out.  For example, I'm not convinced that the Ferguson police officer who shot Michael Brown was acting with too much force. [NEWLINE] [NEWLINE] In other words,<mask> I do agree that they should apply the law fairly, I don't think that should mean they hang officers out to dry<mask> the public has a skewed opinion of<mask> was likely a justifiable use of force.</s>
Label encoding: <s>I agree that some instances require those kinds of actions--for instance, I think [this]( [URL] ) was the right response.  The trooper was fired and is being charged.  But not every example is quite this easy to pick out.  For example, I'm not convinced that the Ferguson police officer who shot Michael Brown was acting with too much force. [NEWLINE] [NEWLINE] In other words, while I do agree that they should apply the law fairly, I don't think that should mean they hang officers out to dry when the public has a skewed opinion of what was likely a justifiable use of force.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Again, not super familiar with that particular monopoly,<mask> its my understanding they've worked their asses off making deals with the South African govt. [NEWLINE] [NEWLINE] <mask><mask><mask> this, the Soviets found a huge diamond mine on their territory and Debeers was forced to make a deal with them to keep the supply artificially low. [NEWLINE] [NEWLINE] This illustrates another point about monopolies: its impossible for a company to maintain an abusive position unless they own everything, everywhere, to start with.  No matter<mask> much Debeers hoards diamonds,<mask> they're not really that rare, other diamond mines were found.</s>
Label encoding: <s>Again, not super familiar with that particular monopoly, but its my understanding they've worked their asses off making deals with the South African govt. [NEWLINE] [NEWLINE] In spite of this, the Soviets found a huge diamond mine on their territory and Debeers was forced to make a deal with them to keep the supply artificially low. [NEWLINE] [NEWLINE] This illustrates another point about monopolies: its impossible for a company to maintain an abusive position unless they own everything, everywhere, to start with.  No matter how much Debeers hoards diamonds, since they're not really that rare, other diamond mines were found.</s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So it is illegal, under any circumstances, to terminate a late term (3rd trimester) pregnancy? [NEWLINE] [NEWLINE] [STARTQ] The Supreme Court has held that bans must include exceptions for threats to the woman's life, physical health, and mental health,<mask> four states allow late-term abortions only<mask> the woman's life is at risk; four allow them<mask> the woman's life or physical health is at risk,<mask> use a definition of health that pro-choice organizations believe is impermissibly narrow. [ENDQ] [NEWLINE] Seems like they still like that whole bodily autonomy thing<mask> the mother's life is threatened.</s>
Label encoding: <s>So it is illegal, under any circumstances, to terminate a late term (3rd trimester) pregnancy? [NEWLINE] [NEWLINE] [STARTQ] The Supreme Court has held that bans must include exceptions for threats to the woman's life, physical health, and mental health, but four states allow late-term abortions only when the woman's life is at risk; four allow them when the woman's life or physical health is at risk, but use a definition of health that pro-choice organizations believe is impermissibly narrow. [ENDQ] [NEWLINE] Seems like they still like that whole bodily autonomy thing when the mother's life is threatened.</s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0837, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is quite unfortunate that philosophers and physicists don't communicate<mask> much<mask> they should; I've witnessed attempts at this first hand and they are not pretty (I'm a grad student of physics at a university with a strong philosphy department with philosophy of science experts). There are,<mask>, plenty of physicists interested in the philosophical implications of physical theories. Here's a blog post by one of the foremost theoretical physicists today, who specialises in quantum field theory and cosmology among other topics: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] He details an argument similar to mine, albeit fleshed out and far more convincing.</s><pad>
Label encoding: <s>It is quite unfortunate that philosophers and physicists don't communicate as much as they should; I've witnessed attempts at this first hand and they are not pretty (I'm a grad student of physics at a university with a strong philosphy department with philosophy of science experts). There are, however, plenty of physicists interested in the philosophical implications of physical theories. Here's a blog post by one of the foremost theoretical physicists today, who specialises in quantum field theory and cosmology among other topics: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] He details an argument similar to mine, albeit fleshed out and far more convincing.</s><pad>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> by the price that sellers are willing to accept.<mask> my simplistic answer is "greed" here. [ENDQ] [NEWLINE] That's half of it,<mask><mask> whatever a seller wants,<mask> no one will by a pizza for $600, the seller is not going to sell it. And like OP said,<mask> no one is willing or even able to pay anything, it's going to drastically drive prices down. [NEWLINE] [NEWLINE] Companies need customers to sell things to to make money. You're assuming a company can charge whatever it wants and still make a profit, which is not true at all.</s>
Label encoding: <s> [STARTQ] Also by the price that sellers are willing to accept. So my simplistic answer is "greed" here. [ENDQ] [NEWLINE] That's half of it, but despite whatever a seller wants, if no one will by a pizza for $600, the seller is not going to sell it. And like OP said, if no one is willing or even able to pay anything, it's going to drastically drive prices down. [NEWLINE] [NEWLINE] Companies need customers to sell things to to make money. You're assuming a company can charge whatever it wants and still make a profit, which is not true at all.</s>
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Direct democracy facilitated by or done over the Internet would weaken national security. By performing legally-binding elections over the Internet, we open legal system up to hackers of all sorts: from script kiddies to corporations to other countries' intelligence services. The technology to execute elections online just isn't there<mask>, and it won't come anytime soon. The requirements (both legal and technical) are incredibly strict and often contradictory; for instance,<mask> do we perform voter authentication online<mask> ensuring a secret ballot, all the<mask> allowing for post-election audits and defending against voter coercion and vote-buying?</s>
Label encoding: <s>Direct democracy facilitated by or done over the Internet would weaken national security. By performing legally-binding elections over the Internet, we open legal system up to hackers of all sorts: from script kiddies to corporations to other countries' intelligence services. The technology to execute elections online just isn't there yet, and it won't come anytime soon. The requirements (both legal and technical) are incredibly strict and often contradictory; for instance, how do we perform voter authentication online while ensuring a secret ballot, all the while allowing for post-election audits and defending against voter coercion and vote-buying?</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes and no. Everyone is trying to prove that every (good) socio-economic change was caused by a technological advancement or a scientific discovery.<mask> they did a lot to try to prove this, it stills seem to me that these were factors that helped and not actual causes for these advancements. [NEWLINE] [NEWLINE] The original point<mask>, was that we tend to give too much credit to<mask> science and engineering help humanity and not enough to other branches of knowledge and I still hold that belief (maybe through stubbornness,<mask> to be honest I don't think anyone gave a very convincing argument).  </s>
Label encoding: <s>Yes and no. Everyone is trying to prove that every (good) socio-economic change was caused by a technological advancement or a scientific discovery. Although they did a lot to try to prove this, it stills seem to me that these were factors that helped and not actual causes for these advancements. [NEWLINE] [NEWLINE] The original point though, was that we tend to give too much credit to how science and engineering help humanity and not enough to other branches of knowledge and I still hold that belief (maybe through stubbornness, although to be honest I don't think anyone gave a very convincing argument).  </s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Irrational thinking doesn't make people insane. [NEWLINE] [NEWLINE] Insane is like, something literally went wrong in the brain at some point and now you have people who would be dysfunctional without or without religion in the world. [NEWLINE] [NEWLINE] It's true that religion can be a clinging point for the insane to rally to,<mask> religion itself is not indicitive of mental illness. Irrationality is actually a biologically evolved mechanism. It's the whole reason we can conceptualize things like gravity, black holes, quantum mechanics, particle physics, etc.<mask> irrationality is merely abstract thinking based on an incorrect premise.</s>
Label encoding: <s>Irrational thinking doesn't make people insane. [NEWLINE] [NEWLINE] Insane is like, something literally went wrong in the brain at some point and now you have people who would be dysfunctional without or without religion in the world. [NEWLINE] [NEWLINE] It's true that religion can be a clinging point for the insane to rally to, but religion itself is not indicitive of mental illness. Irrationality is actually a biologically evolved mechanism. It's the whole reason we can conceptualize things like gravity, black holes, quantum mechanics, particle physics, etc. Because irrationality is merely abstract thinking based on an incorrect premise.</s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Glad I could help change your mind :) I tend towards the analytical side of things myself, and it always helps to remind yourself that there are different perspectives, assumptions, and perceptions about the world than our own. Logic and rationality are ways to navigate our own worldviews -<mask> those worldviews are always necessarily limited. Looking at problems differently (for example, reframing the genetics question to one of everyday practicality) can help us see new solutions and positions on issues. Thank you for having the openness and willingness to post here and honestly engage in changing your view. Good luck out there.</s>
Label encoding: <s>Glad I could help change your mind :) I tend towards the analytical side of things myself, and it always helps to remind yourself that there are different perspectives, assumptions, and perceptions about the world than our own. Logic and rationality are ways to navigate our own worldviews - but those worldviews are always necessarily limited. Looking at problems differently (for example, reframing the genetics question to one of everyday practicality) can help us see new solutions and positions on issues. Thank you for having the openness and willingness to post here and honestly engage in changing your view. Good luck out there.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In terms of sexual harassment, maybe. This generally covers women being groped or cat-called.<mask> an attractive woman,<mask>, I will tell you I get just<mask> much of this in jeans<mask> a tshirt<mask> otherwise. [NEWLINE] [NEWLINE] Rape. Absolutely not. Rape is about power, not about sex. This explains<mask> the vast majority of rape is "aquaintance rape."<mask> you realize this, it doesn't really make sense to think that a girl just looked incredibly sexy one day<mask> her acquaintance just couldn't help himself. [NEWLINE] [NEWLINE] Do some research on the subject.</s>
Label encoding: <s>In terms of sexual harassment, maybe. This generally covers women being groped or cat-called. As an attractive woman, though, I will tell you I get just as much of this in jeans as a tshirt as otherwise. [NEWLINE] [NEWLINE] Rape. Absolutely not. Rape is about power, not about sex. This explains why the vast majority of rape is "aquaintance rape." When you realize this, it doesn't really make sense to think that a girl just looked incredibly sexy one day so her acquaintance just couldn't help himself. [NEWLINE] [NEWLINE] Do some research on the subject.</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you had $20 in BTC you can have that amount of US Dollars deposited into the bank of your choosing.  I use Coinbase for that.  I send Bitcoins to my coinbase account.  Click on the "Sell Bitcoins" button, it deposits the cash equivalent in my checking account.  I have linked the checking account to be a deposit address for any coins I choose to sell on Coinbase.  It couldn't be easier.  Buying coins goes the same way. [NEWLINE] [NEWLINE] I just typed this [LINK HERE]( [URL] ) to describe mining in this thread.</s>
Label encoding: <s>If you had $20 in BTC you can have that amount of US Dollars deposited into the bank of your choosing.  I use Coinbase for that.  I send Bitcoins to my coinbase account.  Click on the "Sell Bitcoins" button, it deposits the cash equivalent in my checking account.  I have linked the checking account to be a deposit address for any coins I choose to sell on Coinbase.  It couldn't be easier.  Buying coins goes the same way. [NEWLINE] [NEWLINE] I just typed this [LINK HERE]( [URL] ) to describe mining in this thread.</s>
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe that<mask> people were informed enough about the risks of obesity,<mask> they should be about everything else discussed here, they have the potential of making the right decision. And I don't believe in a genetic disposition towards certain actions.<mask><mask> that it is harder for a child, just<mask> it is harder for a poor person, to eat healthier given the diets within they reach,<mask> self discipline should always be the first and most tried effort here. Allowing idiotic people to believe that their children are fat<mask> they are depressed does absolutely nothing to fight the underlying causes for everything. </s>
Label encoding: <s>I believe that if people were informed enough about the risks of obesity, as they should be about everything else discussed here, they have the potential of making the right decision. And I don't believe in a genetic disposition towards certain actions. I agree that it is harder for a child, just as it is harder for a poor person, to eat healthier given the diets within they reach, but self discipline should always be the first and most tried effort here. Allowing idiotic people to believe that their children are fat because they are depressed does absolutely nothing to fight the underlying causes for everything. </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This doesn't make sense. I understand now that it is possible to create your own meaning from life,<mask> that doesn't change the fact that there is no greater meaning to it all -- no higher power watching and evaluating. [NEWLINE] [NEWLINE] I'm not saying that life is meaningless *unless you believe in god*. I'm saying life has no meaning, whether you believe in god or not. Those who believe,<mask>, find comfort in the illusion that there is meaning. [NEWLINE] [NEWLINE] I<mask> don't understand<mask> other people's beliefs would piss you off. Please don't make baseless assumptions.</s>
Label encoding: <s>This doesn't make sense. I understand now that it is possible to create your own meaning from life, but that doesn't change the fact that there is no greater meaning to it all -- no higher power watching and evaluating. [NEWLINE] [NEWLINE] I'm not saying that life is meaningless *unless you believe in god*. I'm saying life has no meaning, whether you believe in god or not. Those who believe, however, find comfort in the illusion that there is meaning. [NEWLINE] [NEWLINE] I also don't understand why other people's beliefs would piss you off. Please don't make baseless assumptions.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>At this point,<mask>, you're riding on the fact that the loss of utility due to loosening the ethical demand is less than the loss of utility due to the ethical demand being unappealing, and I'm really not sure that it's even possible to substantiate that, nor do<mask><mask> it seems intuitively likely, or unlikely, for that matter, it seems too abstract to really quantify [NEWLINE] [NEWLINE] Of course, that utility is too abstract to quantify would be an entirely separate critique of OP's view,<mask> I'm not sure it's a particularly good one in this case</s>
Label encoding: <s>At this point, though, you're riding on the fact that the loss of utility due to loosening the ethical demand is less than the loss of utility due to the ethical demand being unappealing, and I'm really not sure that it's even possible to substantiate that, nor do I think it seems intuitively likely, or unlikely, for that matter, it seems too abstract to really quantify [NEWLINE] [NEWLINE] Of course, that utility is too abstract to quantify would be an entirely separate critique of OP's view, but I'm not sure it's a particularly good one in this case</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> for<mask> the federal government can do all those things, it's<mask> no one else wants to [ENDQ] [NEWLINE] <mask><mask><mask>, each state still enumerates its own drinking age.  They are just coerced into making it 21<mask> of the [National Minimum Drinking Age Act.]( [URL] )  Similarly, it is feasible for Congress to implement a system whereby a national minimum curriculum is set, and those states that fail to meet that minimum curriculum have their federal education funding cut. [NEWLINE] [NEWLINE] Yay constitutional law<mask>.  Creative interpretation of Congressional power is the name of the game.</s>
Label encoding: <s> [STARTQ] As for why the federal government can do all those things, it's because no one else wants to [ENDQ] [NEWLINE] First of all, each state still enumerates its own drinking age.  They are just coerced into making it 21 because of the [National Minimum Drinking Age Act.]( [URL] )  Similarly, it is feasible for Congress to implement a system whereby a national minimum curriculum is set, and those states that fail to meet that minimum curriculum have their federal education funding cut. [NEWLINE] [NEWLINE] Yay constitutional law indeed.  Creative interpretation of Congressional power is the name of the game.</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point of the game is the game itself, not the details of the game.<mask> you play<mask> pacman you don't have desires to eat your way through mazes,<mask> you play mario, you don't have the desire to jump on everything. You're just playing the game and trying to win. [NEWLINE] [NEWLINE] Penn and Teller did a bullshit on this: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] ____ [NEWLINE] [NEWLINE] Trust me this has been studied over and over and your view is wrong. I'm only being<mask> direct<mask> I've looked into this<mask> much in the past.</s>
Label encoding: <s>The point of the game is the game itself, not the details of the game. When you play as pacman you don't have desires to eat your way through mazes, when you play mario, you don't have the desire to jump on everything. You're just playing the game and trying to win. [NEWLINE] [NEWLINE] Penn and Teller did a bullshit on this: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] ____ [NEWLINE] [NEWLINE] Trust me this has been studied over and over and your view is wrong. I'm only being so direct because I've looked into this so much in the past.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>yeah I know about the New York Times article and it's bull shit. First off, you can drop the whole left side of the chart. Most of those people will get a refund at the end of the year. The ones that don't are allowed to simply not file their taxes<mask><mask> you are filing jointly you don't have to file your taxes<mask> you make less than [$20,000]( [URL].jsp?article_id=66547) [NEWLINE] [NEWLINE] This is<mask> all ignoring the point that it is perfectly in a married couples right to file taxes separately.</s>
Label encoding: <s>yeah I know about the New York Times article and it's bull shit. First off, you can drop the whole left side of the chart. Most of those people will get a refund at the end of the year. The ones that don't are allowed to simply not file their taxes because if you are filing jointly you don't have to file your taxes if you make less than [$20,000]( [URL].jsp?article_id=66547) [NEWLINE] [NEWLINE] This is also all ignoring the point that it is perfectly in a married couples right to file taxes separately.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not a matter of justice or<mask>'s right. The fact is that we value the life of a human over the life of an animal.<mask> an animal attacks, even<mask> provoked, there is a significant chance that it will attack again. The risk of someone else being harmed or killed is too great to favour the life of that animal. [NEWLINE] [NEWLINE] <mask> you disagree with this concept,<mask> is your right, you may<mask> want to reconsider<mask> the meat you buy comes from or<mask> the milk you put in your cereal is collected (assuming you are not already a vegan)</s>
Label encoding: <s>It's not a matter of justice or what's right. The fact is that we value the life of a human over the life of an animal. When an animal attacks, even if provoked, there is a significant chance that it will attack again. The risk of someone else being harmed or killed is too great to favour the life of that animal. [NEWLINE] [NEWLINE] If you disagree with this concept, as is your right, you may also want to reconsider where the meat you buy comes from or how the milk you put in your cereal is collected (assuming you are not already a vegan)</s>
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Again, this is a cultural issue, it's a black sounding name vs a white sounding name. There is no rule that says that a white guy can't have the name Marquis, and no rule that says a black guy can't be named Cody. Names are part of culture, not race.<mask> race was the determined people names then names wouldn't change that much throughout society. There black kids born in Kenya would have the same name<mask> the black kids born in Harlem,<mask> that's not the case. Same skin color, very different cultures, very different names.</s><pad>
Label encoding: <s>Again, this is a cultural issue, it's a black sounding name vs a white sounding name. There is no rule that says that a white guy can't have the name Marquis, and no rule that says a black guy can't be named Cody. Names are part of culture, not race. If race was the determined people names then names wouldn't change that much throughout society. There black kids born in Kenya would have the same name as the black kids born in Harlem, but that's not the case. Same skin color, very different cultures, very different names.</s><pad>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi! I have a few learning disabilities, and i feel like you don't understand the term. Treat it more like a difference. Tests are not equal to a surgery. Tests can be harder for some, just<mask> they can't function on a test doesn't mean they can't perform surgery. They may be geniuses! Just<mask> i can't do math (i can't) or write by hand (i can't) doesn't mean i couldn't, say, talk to you for hours on end about computer science (i can). Disability ≠ idiocy.</s>
Label encoding: <s>Hi! I have a few learning disabilities, and i feel like you don't understand the term. Treat it more like a difference. Tests are not equal to a surgery. Tests can be harder for some, just because they can't function on a test doesn't mean they can't perform surgery. They may be geniuses! Just because i can't do math (i can't) or write by hand (i can't) doesn't mean i couldn't, say, talk to you for hours on end about computer science (i can). Disability ≠ idiocy.</s>
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand your point,<mask> you weaken your argument by resorting to name calling,<mask> well<mask> violating the subreddit rules. <mask> it's not just about one person<mask> you are making it out to be.  There were potentially a hundred people who were inconvenienced.  Some of those people may get to rarely see a movie<mask> it has become something of a luxury these days. [NEWLINE] [NEWLINE] <mask> to play devils advocate<mask> should the family with the child who was the disturbance get to say "Us, us, us!" <mask> about the entire rest of the theater?</s>
Label encoding: <s>I understand your point, but you weaken your argument by resorting to name calling, as well as violating the subreddit rules.  Also it's not just about one person as you are making it out to be.  There were potentially a hundred people who were inconvenienced.  Some of those people may get to rarely see a movie as it has become something of a luxury these days. [NEWLINE] [NEWLINE] So to play devils advocate why should the family with the child who was the disturbance get to say "Us, us, us!"  What about the entire rest of the theater?</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Likewise, saying you should dress conservatively, lock your house, and stay in at night is unreasonable. [ENDQ] [NEWLINE] Answer me this: Does your door have a lock?<mask> yes, you are a fucking hypocrite.<mask><mask><mask> you, a lock is absolutely pointless. That it does *nothing* to prevent break-ins. [NEWLINE] [NEWLINE] <mask> waste money on a lock? Hell, I see more people *lock themselves out* than get broken into. And apparently the lock doesn't prevent that anyway.<mask> might<mask> well go for convenience and not have one.</s>
Label encoding: <s> [STARTQ] Likewise, saying you should dress conservatively, lock your house, and stay in at night is unreasonable. [ENDQ] [NEWLINE] Answer me this: Does your door have a lock? If yes, you are a fucking hypocrite. Because according to you, a lock is absolutely pointless. That it does *nothing* to prevent break-ins. [NEWLINE] [NEWLINE] Why waste money on a lock? Hell, I see more people *lock themselves out* than get broken into. And apparently the lock doesn't prevent that anyway. So might as well go for convenience and not have one.</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It isn't an ideal world I am describing. I am saying a cop's job is to make arrests which are punishable by the legal system.<mask> a judge finds you guilty then an arrest is justified. Even<mask> the person is falsely convicted the cop still made a justified arrest. [NEWLINE] [NEWLINE] Basically<mask> you don't agree with<mask> you are being arrested it doesn't matter. All that matters is<mask> the rest of the legal system decides. That is<mask> allows our civilization to exist. We have to have a legal system and be able to enforce it to have meaningful laws.</s>
Label encoding: <s>It isn't an ideal world I am describing. I am saying a cop's job is to make arrests which are punishable by the legal system. If a judge finds you guilty then an arrest is justified. Even if the person is falsely convicted the cop still made a justified arrest. [NEWLINE] [NEWLINE] Basically if you don't agree with why you are being arrested it doesn't matter. All that matters is what the rest of the legal system decides. That is what allows our civilization to exist. We have to have a legal system and be able to enforce it to have meaningful laws.</s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who will you accept telling you that you're worthwhile? [NEWLINE] There are six kinds of facts, usually: [NEWLINE] Descriptive, contingent, concrete, normative, modal, and abstract. [NEWLINE] Which of these will you accept? [NEWLINE] [NEWLINE] <mask> you say you believe something,<mask> you don't tell us '<mask>'you will trust another view from other than your own view of your worthlessness, then you're saying your ability to gauge your worthlessness is worth more than any other viewpoint anyone else would have. Which is the literal opposite of maintaining that you are worthless.</s>
Label encoding: <s>Who will you accept telling you that you're worthwhile? [NEWLINE] There are six kinds of facts, usually: [NEWLINE] Descriptive, contingent, concrete, normative, modal, and abstract. [NEWLINE] Which of these will you accept? [NEWLINE] [NEWLINE] If you say you believe something, but you don't tell us'where'you will trust another view from other than your own view of your worthlessness, then you're saying your ability to gauge your worthlessness is worth more than any other viewpoint anyone else would have. Which is the literal opposite of maintaining that you are worthless.</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Jobs was a visionary<mask> you need to put him in a box. [NEWLINE] He had a vision of the future of computing, and put it together using pieces that he didn't invent. He saw that product design was a flaw in the computing, cellphone and tablet markets. He thought outside of the box and implanted his visions. I give him respect for that. [NEWLINE] Most of these big tech guys took things already created and put them together on new ways. Of course many might of been more technically skilled and hands on,<mask> that really doesn't make them more important.</s>
Label encoding: <s>Jobs was a visionary if you need to put him in a box. [NEWLINE] He had a vision of the future of computing, and put it together using pieces that he didn't invent. He saw that product design was a flaw in the computing, cellphone and tablet markets. He thought outside of the box and implanted his visions. I give him respect for that. [NEWLINE] Most of these big tech guys took things already created and put them together on new ways. Of course many might of been more technically skilled and hands on, but that really doesn't make them more important.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> the OP believes effeminate mannerisms / camp behaviour are not gay traits [ENDQ] [NEWLINE] AKA the OP wants people that don't act in a way he likes to act in a way he approves of.<mask> he's really saying is "Act<mask> gay<mask> you want,<mask><mask><mask> acting gay doesn't involve acting effeminate." He's defining<mask> good gay is and<mask> bad gay is; not only does he<mask> an individual not have a right, he<mask> a non-gay person really shouldn't be judging another person's experience at all. </s>
Label encoding: <s> [STARTQ] if the OP believes effeminate mannerisms / camp behaviour are not gay traits [ENDQ] [NEWLINE] AKA the OP wants people that don't act in a way he likes to act in a way he approves of. What he's really saying is "Act as gay as you want, as long as acting gay doesn't involve acting effeminate." He's defining what good gay is and what bad gay is; not only does he as an individual not have a right, he as a non-gay person really shouldn't be judging another person's experience at all. </s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So? <mask> you've got that kind of money to burn, more power to you.  For all the people that don't, it would mean that the child is being raised with far less support than it actually needs, which has consequences.  We already know that kids raised in a frequent state of malnutrition have mental deficits<mask><mask><mask> that are very difficult for them to overcome later in life.  I don't imagine that under-supporting a special needs child of<mask> they need would have any less drastic of an impact in their quality of life. </s>
Label encoding: <s>So?  If you've got that kind of money to burn, more power to you.  For all the people that don't, it would mean that the child is being raised with far less support than it actually needs, which has consequences.  We already know that kids raised in a frequent state of malnutrition have mental deficits as a result that are very difficult for them to overcome later in life.  I don't imagine that under-supporting a special needs child of what they need would have any less drastic of an impact in their quality of life. </s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> for the expulsion of the Palestinians [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] Their own leaders encouraged them to leave<mask> they weren't caught in the crossfire<mask> the surrounding Arab nations invaded. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The Hamas charter explicitly states that all Jews everywhere must be killed in the name of Islam. That's just not the kind of belief that has any place in the modern world,<mask> too many people refuse to see this conflict<mask> anything other than "Israel stole the Palestinians' land" (never mind that the UN created Israel, with a partition plan that the Palestinian leaders rejected).</s>
Label encoding: <s> [STARTQ] As for the expulsion of the Palestinians [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] Their own leaders encouraged them to leave so they weren't caught in the crossfire when the surrounding Arab nations invaded. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] The Hamas charter explicitly states that all Jews everywhere must be killed in the name of Islam. That's just not the kind of belief that has any place in the modern world, but too many people refuse to see this conflict as anything other than "Israel stole the Palestinians' land" (never mind that the UN created Israel, with a partition plan that the Palestinian leaders rejected).</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's called a donation for a reason. Just<mask> I give a homeless man 5 dollars<mask> I could afford to give him 10 doesn't make me a terrible person. After all I could have chosen not to give any money. Same thing with rich people. Even a small percentage of a rich persons income will make a difference. And charity is not supposed to me measured.<mask><mask> are you trying to measure it? [NEWLINE] [NEWLINE] I am tired of people hating rich people.<mask> someone makes more money than you. Who cares. Forget about it and move on. </s>
Label encoding: <s>It's called a donation for a reason. Just because I give a homeless man 5 dollars when I could afford to give him 10 doesn't make me a terrible person. After all I could have chosen not to give any money. Same thing with rich people. Even a small percentage of a rich persons income will make a difference. And charity is not supposed to me measured. So why are you trying to measure it? [NEWLINE] [NEWLINE] I am tired of people hating rich people. So someone makes more money than you. Who cares. Forget about it and move on. </s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I prefer awareness. We have separate lanes here and<mask> about 3 cars hit me every year. Last one was the day before yesterday. It's just<mask> the driver doesn't understand<mask> it's like to drive a bike. Here the awareness is good<mask> those few that hit me simply don't seem to respect and understand bikers in general. And before someone starts yanking my chain: I'm a freakin bike saint who follows every last rule to the book. Perhaps you should not be allowed to drive a car<mask> you didn't get a biking license first.</s>
Label encoding: <s>I prefer awareness. We have separate lanes here and yet about 3 cars hit me every year. Last one was the day before yesterday. It's just because the driver doesn't understand what it's like to drive a bike. Here the awareness is good but those few that hit me simply don't seem to respect and understand bikers in general. And before someone starts yanking my chain: I'm a freakin bike saint who follows every last rule to the book. Perhaps you should not be allowed to drive a car if you didn't get a biking license first.</s>
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're not wrong,<mask><mask><mask> you're overestimating<mask> much corruption hurts a business compared to<mask> much a corrupt company would benefit from the ability to influence the government proportionately to its wealth. Or to put it in simpler terms,<mask> corruption is bad for business and we can trust successful companies to do<mask>'s best for their own interests, then we shouldn't be seeing nearly<mask> much corruption in the corporate world<mask> we do.<mask><mask> Wal-Mart would see corruption<mask> a small price to pay to have the voting power of tens of millions of people.</s>
Label encoding: <s>You're not wrong, but I think you're overestimating how much corruption hurts a business compared to how much a corrupt company would benefit from the ability to influence the government proportionately to its wealth. Or to put it in simpler terms, if corruption is bad for business and we can trust successful companies to do what's best for their own interests, then we shouldn't be seeing nearly as much corruption in the corporate world as we do. I think Wal-Mart would see corruption as a small price to pay to have the voting power of tens of millions of people.</s>
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think they are too stupid. I just don't think it's possible for a policy to take every circumstance into consideration.<mask><mask><mask><mask><mask> the people living in that circumstance are much more able to come up with a solution then policy makers. [NEWLINE] [NEWLINE] And<mask> the system prescribes that the money be spent on healthy food, the system needs to allocate much more than it currently does. I would<mask><mask> this goal can be achieved with a lower total amount of money simply by giving those who receive assistance control over<mask> that money is spent. </s>
Label encoding: <s>I don't think they are too stupid. I just don't think it's possible for a policy to take every circumstance into consideration. As a result I think the people living in that circumstance are much more able to come up with a solution then policy makers. [NEWLINE] [NEWLINE] And if the system prescribes that the money be spent on healthy food, the system needs to allocate much more than it currently does. I would argue that this goal can be achieved with a lower total amount of money simply by giving those who receive assistance control over how that money is spent. </s>
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Clarifying question:<mask>'s the end goal of this? Are you just trying to set up a system<mask> either poor people vote for small government or they get taxed into starvation? I understand that you have to create a balance between providing for the marginalized in society and preventing people from literally voting themselves money<mask><mask> you're proposing would destroy all forms of public assistance and welfare by forcing it to do exactly the opposite of<mask> it was intended. Welfare would cease to have any effect, the only people that cold afford it are the people who don't need it. </s>
Label encoding: <s>Clarifying question: what's the end goal of this? Are you just trying to set up a system where either poor people vote for small government or they get taxed into starvation? I understand that you have to create a balance between providing for the marginalized in society and preventing people from literally voting themselves money but what you're proposing would destroy all forms of public assistance and welfare by forcing it to do exactly the opposite of what it was intended. Welfare would cease to have any effect, the only people that cold afford it are the people who don't need it. </s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't want random adults to decide to adopt kids just<mask> there are kids in danger. I mean once people settle into a happy relationship and they think they're ready to have kids, it makes sense to a adopt a kid<mask> it's such a charitable thing to do and you're gonna take care of a child either way. [NEWLINE] [NEWLINE] Furthermore, in my experience, there's not really a diminished love between genetic children and their parents and adopted children and their parents. Once someone's raised a child in a loving home they're going to love each other.</s>
Label encoding: <s>I don't want random adults to decide to adopt kids just because there are kids in danger. I mean once people settle into a happy relationship and they think they're ready to have kids, it makes sense to a adopt a kid since it's such a charitable thing to do and you're gonna take care of a child either way. [NEWLINE] [NEWLINE] Furthermore, in my experience, there's not really a diminished love between genetic children and their parents and adopted children and their parents. Once someone's raised a child in a loving home they're going to love each other.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While it is true that supervillain prisons are basically cardboard boxes, Your argument hinges on the premise that death or handicapping will be MORE EFFECTIVE then these cardboard boxes. Do you know<mask> happens<mask> you kill the joker? He comes back<mask> its "hillarious" or something, you cut of someones arms? He becomes "ARMGUN"- the man with guns for arms! [NEWLINE] The eternal problem that superheroes have is that their universe will not allow them to get rid of their enemies, whether they want to or not. [NEWLINE] </s>
Label encoding: <s>While it is true that supervillain prisons are basically cardboard boxes, Your argument hinges on the premise that death or handicapping will be MORE EFFECTIVE then these cardboard boxes. Do you know what happens when you kill the joker? He comes back because its "hillarious" or something, you cut of someones arms? He becomes "ARMGUN"- the man with guns for arms! [NEWLINE] The eternal problem that superheroes have is that their universe will not allow them to get rid of their enemies, whether they want to or not. [NEWLINE] </s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your argument is based on the idea that religion is an inherently good thing. It's demonstrably bad for a child to do any of the things that you list, the short *and* long-term consequences are evident.<mask>, not trying to force any particular  religion on a child is not demonstrably bad. The only people to whom consequences are apparent are believers of the faith. We're not trying to argue this from their point of view, or anyone's for that matter. We're trying to argue this from an objective point of view.</s>
Label encoding: <s>Your argument is based on the idea that religion is an inherently good thing. It's demonstrably bad for a child to do any of the things that you list, the short *and* long-term consequences are evident. However, not trying to force any particular  religion on a child is not demonstrably bad. The only people to whom consequences are apparent are believers of the faith. We're not trying to argue this from their point of view, or anyone's for that matter. We're trying to argue this from an objective point of view.</s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, they don't get to choose whether to give it for free,<mask> you do? [NEWLINE] [NEWLINE] Who said "huge profit". <mask> about any profit? <mask> you are an author or a studio-type musician you won't be able to do it full time<mask> you don't have a way to make money off of it. [NEWLINE] [NEWLINE] You're right, there will not be "no" incentive,<mask> there's going to be a lot more artists working office jobs to pay the rent<mask> you refuse to pay for<mask> you consume.</s>
Label encoding: <s>So, they don't get to choose whether to give it for free, but you do? [NEWLINE] [NEWLINE] Who said "huge profit".  How about any profit?  If you are an author or a studio-type musician you won't be able to do it full time if you don't have a way to make money off of it. [NEWLINE] [NEWLINE] You're right, there will not be "no" incentive, but there's going to be a lot more artists working office jobs to pay the rent because you refuse to pay for what you consume.</s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0753, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] a recommender familiar with the company does the same job<mask> any interviewer... [ENDQ] [NEWLINE]...placing the applicant far down the normal hiring flowchart, past HR screening, etc. I'm not arguing that an interview isn't a valuable tool for establishing an individual's merit,<mask> networking frequently subverts the situation that all other (meritorious<mask> unconnected) applicants must partake in to get to that point. "Knowing someone on the inside" is the irrelevant aspect that networking makes a prime criteria for hiring, not having someone's recommendation.</s>
Label encoding: <s> [STARTQ] a recommender familiar with the company does the same job as any interviewer... [ENDQ] [NEWLINE]...placing the applicant far down the normal hiring flowchart, past HR screening, etc. I'm not arguing that an interview isn't a valuable tool for establishing an individual's merit, but networking frequently subverts the situation that all other (meritorious but unconnected) applicants must partake in to get to that point. "Knowing someone on the inside" is the irrelevant aspect that networking makes a prime criteria for hiring, not having someone's recommendation.</s>
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Even good actors at tools for a director to leverage to make a movie. A great example of this is John Malkovitch versus Daniel day lewis. With a great director they could probably play the same character<mask> it wouldn't be the same<mask> of not only acting<mask> the intrinsic things about them that shine through<mask> they act. [NEWLINE] [NEWLINE] A great example of this is Mr. Magoriums wonder emporium. Not<mask> it is a fantastic film,<mask><mask> many good actors all play the same role under the same director.</s>
Label encoding: <s>Even good actors at tools for a director to leverage to make a movie. A great example of this is John Malkovitch versus Daniel day lewis. With a great director they could probably play the same character but it wouldn't be the same because of not only acting but the intrinsic things about them that shine through when they act. [NEWLINE] [NEWLINE] A great example of this is Mr. Magoriums wonder emporium. Not because it is a fantastic film, but because many good actors all play the same role under the same director.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0819, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They released a statement saying that their organization isn't involved in fundraisers relating to the incidents in Ferguson: [NEWLINE] [NEWLINE] &gt; [Contrary to recent posts on social media, BackStoppers is not participating in nor has benefited from any fundraising activity involving the Ferguson matter,” BackStoppers stated Monday morning in a release provided by Joyce. “We scrutinize our contributions and<mask> we receive funds involving the Ferguson matter, those funds would be rejected by the Board of Directors.”]( [URL]?mode=jqm)</s>
Label encoding: <s>They released a statement saying that their organization isn't involved in fundraisers relating to the incidents in Ferguson: [NEWLINE] [NEWLINE] &gt; [Contrary to recent posts on social media, BackStoppers is not participating in nor has benefited from any fundraising activity involving the Ferguson matter,” BackStoppers stated Monday morning in a release provided by Joyce. “We scrutinize our contributions and if we receive funds involving the Ferguson matter, those funds would be rejected by the Board of Directors.”]( [URL]?mode=jqm)</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We don't weigh other people's misfortunes and compare them to ours and decide which one is greater and go on the path of least misfortune in our society!<mask> someone is speeding<mask> they are en route to the hospital to deliver a baby, they are still breaking the law,<mask><mask> their reasoning. They should get a ticket, whether or not their getting to the hospital is more or less important than other drivers not having the safest driving experience<mask> of them. [NEWLINE] We are not responsible for other people's misfortunes.</s>
Label encoding: <s>We don't weigh other people's misfortunes and compare them to ours and decide which one is greater and go on the path of least misfortune in our society! If someone is speeding because they are en route to the hospital to deliver a baby, they are still breaking the law, regardless of their reasoning. They should get a ticket, whether or not their getting to the hospital is more or less important than other drivers not having the safest driving experience because of them. [NEWLINE] We are not responsible for other people's misfortunes.</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I came here to say something similar to this. I grew up in a very closed Baptist society, and most of the atheists I know were raised in religious households. A lot of it is either reactionary or simply something that resonates with past experiences. [NEWLINE] [NEWLINE] It's<mask> I get annoyed<mask> people assert I must be angry at god. I'm not angry with god, I'm angry with the humans that forced him into my subconscious mind. I'm not the only one, and /r/atheism is one result of that. </s>
Label encoding: <s>I came here to say something similar to this. I grew up in a very closed Baptist society, and most of the atheists I know were raised in religious households. A lot of it is either reactionary or simply something that resonates with past experiences. [NEWLINE] [NEWLINE] It's why I get annoyed when people assert I must be angry at god. I'm not angry with god, I'm angry with the humans that forced him into my subconscious mind. I'm not the only one, and /r/atheism is one result of that. </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Texas Republicans!= All Republicans. [ENDQ] [NEWLINE] Texas Republicans reflect that actual values of the Republican party.  Can you honestly tell me that the rest of the Republican party (the Creationist, global warming denying GOP) is pro-critical thinking?  Seriously? [NEWLINE] [NEWLINE] [STARTQ] I would say that less than 20% of Republicans would fight the teaching of critical thinking skills. [ENDQ] [NEWLINE] Reword that this way:   "Should kid learn critical thinking skills like those used in colleges?"   And that number jumps to 75%+ opposing.</s>
Label encoding: <s> [STARTQ] Texas Republicans!= All Republicans. [ENDQ] [NEWLINE] Texas Republicans reflect that actual values of the Republican party.  Can you honestly tell me that the rest of the Republican party (the Creationist, global warming denying GOP) is pro-critical thinking?  Seriously? [NEWLINE] [NEWLINE] [STARTQ] I would say that less than 20% of Republicans would fight the teaching of critical thinking skills. [ENDQ] [NEWLINE] Reword that this way:   "Should kid learn critical thinking skills like those used in colleges?"   And that number jumps to 75%+ opposing.</s>
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I question is the utility of educating others at all; fairly few people will argue with you that a black guy in a 1995 Honda on his way to school is more likely to get pulled over than a white guy in a 2015 Audi on his way to sell cocaine. [NEWLINE] [NEWLINE] Do they have to understand that this is the intersection of white privilege and class privilege to understand that this is wrong? Not really, it isn't fair, and that's obvious. The concrete is easier to sell than the theoretical, and more likely to create action.</s>
Label encoding: <s>What I question is the utility of educating others at all; fairly few people will argue with you that a black guy in a 1995 Honda on his way to school is more likely to get pulled over than a white guy in a 2015 Audi on his way to sell cocaine. [NEWLINE] [NEWLINE] Do they have to understand that this is the intersection of white privilege and class privilege to understand that this is wrong? Not really, it isn't fair, and that's obvious. The concrete is easier to sell than the theoretical, and more likely to create action.</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Gender refers to the socially constructed roles, behaviours, activities, and attributes that a given society considers appropriate for men and women. [ENDQ] [NEWLINE] Given this definition, most people are *misidentifying*<mask> gender they are.<mask> there are plenty of males with dominantly female interests (some gay people come to mind) and plenty of females with dominantly male interests (we even have a word for this: tomboy). [NEWLINE] [NEWLINE] <mask> either you are being incredibly sexist. Or a *lot* of people are misidentifying their gender.</s>
Label encoding: <s> [STARTQ] Gender refers to the socially constructed roles, behaviours, activities, and attributes that a given society considers appropriate for men and women. [ENDQ] [NEWLINE] Given this definition, most people are *misidentifying* what gender they are. As there are plenty of males with dominantly female interests (some gay people come to mind) and plenty of females with dominantly male interests (we even have a word for this: tomboy). [NEWLINE] [NEWLINE] So either you are being incredibly sexist. Or a *lot* of people are misidentifying their gender.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While voter ID is a good idea that would decrease voter fraud, statistically minorities like Hispanics, elderly, and black people typically lack a photo ID and statistically a lot of those people happen to vote democrat<mask> both sides don't care for moral reasons, they just want to win elections.  It's especially an issue in places like Wisconsin<mask> we recalled our gov. Scott walker<mask> he won the recall election<mask> it's obvious that people are very split.(<mask> Wisconsin is a swing state and the presidential elections are coming up relatively soon.</s>
Label encoding: <s>While voter ID is a good idea that would decrease voter fraud, statistically minorities like Hispanics, elderly, and black people typically lack a photo ID and statistically a lot of those people happen to vote democrat so both sides don't care for moral reasons, they just want to win elections.  It's especially an issue in places like Wisconsin because we recalled our gov. Scott walker but he won the recall election so it's obvious that people are very split.( also Wisconsin is a swing state and the presidential elections are coming up relatively soon.</s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think<mask>. The pictures are out there now. Whether one individual looks at them, jerks off to them, or ignores them, is totally irrelevant. The issue is with the hacker who made it public,<mask> pretending it is still private is just silly. [NEWLINE] [NEWLINE] <mask><mask> the op used a good analogy with NSA and<mask> the issue wouldn't be with a person reading your public info. I can't find one reason<mask> we are supposed to pretend the photos aren't leaked and not acknowledge them<mask> we want to.</s>
Label encoding: <s>I don't think so. The pictures are out there now. Whether one individual looks at them, jerks off to them, or ignores them, is totally irrelevant. The issue is with the hacker who made it public, but pretending it is still private is just silly. [NEWLINE] [NEWLINE] I think the op used a good analogy with NSA and why the issue wouldn't be with a person reading your public info. I can't find one reason why we are supposed to pretend the photos aren't leaked and not acknowledge them if we want to.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm proud that I live in a system like it and that I continually support many of the things he mentioned. [ENDQ] [NEWLINE] <mask> are you proud? Was living in the system you live in a matter of chance?<mask><mask>, are you proud of any other matters of chance like race or gender? [NEWLINE] [NEWLINE] [STARTQ] <mask> I'm proud that I'm contributing to a community of shared ideals. [ENDQ] [NEWLINE] I see no problem with this. I am proud of the same things,<mask> it has literally nothing to do with my government.</s>
Label encoding: <s> [STARTQ] I'm proud that I live in a system like it and that I continually support many of the things he mentioned. [ENDQ] [NEWLINE] Why are you proud? Was living in the system you live in a matter of chance? If so, are you proud of any other matters of chance like race or gender? [NEWLINE] [NEWLINE] [STARTQ] Moreover I'm proud that I'm contributing to a community of shared ideals. [ENDQ] [NEWLINE] I see no problem with this. I am proud of the same things, but it has literally nothing to do with my government.</s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why would I ever use it<mask>?  The fact that it is deflationary means that there will be a decreasing amount in circulation and its 'value' will<mask> trend upwards. <mask>... there's no reason for me to spend my bitcoins, which I expect to go up in value,<mask> I could spend my more current, practical currencies like GBP or USD, which I expect will be worth vaguely less in the future. [NEWLINE] [NEWLINE] Your desire for it to retain precisely the same value is tremendously undesirable for a functional economy.</s>
Label encoding: <s>Why would I ever use it though?  The fact that it is deflationary means that there will be a decreasing amount in circulation and its 'value' will therefore trend upwards.  So... there's no reason for me to spend my bitcoins, which I expect to go up in value, when I could spend my more current, practical currencies like GBP or USD, which I expect will be worth vaguely less in the future. [NEWLINE] [NEWLINE] Your desire for it to retain precisely the same value is tremendously undesirable for a functional economy.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>~~<mask><mask>,<mask> you think about it, Satan is<mask> called the Deciever, iirc. Do you think he could actually have those cites at that time talking to Jesus, or would he have given them to him later?~~ [NEWLINE] [NEWLINE] Edit: Looked it up, and<mask><mask> the Deciever name is possibly incorrect,<mask> adding on to this, Satan is given authority over Earth,<mask> Satan had the power to give Jesus the kingdoms,<mask> God allowed it. I was wrong<mask> God gave Satan the power.</s>
Label encoding: <s>~~ I disagree, if you think about it, Satan is also called the Deciever, iirc. Do you think he could actually have those cites at that time talking to Jesus, or would he have given them to him later?~~ [NEWLINE] [NEWLINE] Edit: Looked it up, and I think the Deciever name is possibly incorrect, also adding on to this, Satan is given authority over Earth, so Satan had the power to give Jesus the kingdoms, if God allowed it. I was wrong if God gave Satan the power.</s>
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To further comment on the issue of implied consent,<mask><mask> that associating physical pleasure with consent is unreliable.  People (and I'm guessing animals) can be sexually aroused and experience pleasure even<mask> they don't want it, or it is causing them distress or pain at the same time. [NEWLINE] [NEWLINE] A person can experience an orgasm or physical pleasure during a rape.  That does not make the rape acceptable in any way.  This is one flaw I see in your argument that I believe would apply to animals<mask> well. </s>
Label encoding: <s>To further comment on the issue of implied consent, I think that associating physical pleasure with consent is unreliable.  People (and I'm guessing animals) can be sexually aroused and experience pleasure even if they don't want it, or it is causing them distress or pain at the same time. [NEWLINE] [NEWLINE] A person can experience an orgasm or physical pleasure during a rape.  That does not make the rape acceptable in any way.  This is one flaw I see in your argument that I believe would apply to animals as well. </s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you have any evidence, or is this personal speculation?<mask> I would say "bulking" is generally accompanied by self-discipline in the gym.<mask>, bulking is generally accompanied by cutting,<mask> the person loses the fat gained. The self-discipline gained in the gym aids this person in controlling their food intake and reducing their weight. [NEWLINE] [NEWLINE] [STARTQ] There is no scientifically approved diet you can follow. [ENDQ] [NEWLINE] Caloric deficit. Tell me<mask> someone can avoid losing weight<mask> eating below their metabolic rate?</s>
Label encoding: <s>Do you have any evidence, or is this personal speculation? Because I would say "bulking" is generally accompanied by self-discipline in the gym. Additionally, bulking is generally accompanied by cutting, where the person loses the fat gained. The self-discipline gained in the gym aids this person in controlling their food intake and reducing their weight. [NEWLINE] [NEWLINE] [STARTQ] There is no scientifically approved diet you can follow. [ENDQ] [NEWLINE] Caloric deficit. Tell me how someone can avoid losing weight while eating below their metabolic rate?</s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The *rate* of crime is negatively correlated with income, not just the absolute amount. The only case<mask> this is not true is drug offenses. [NEWLINE] [NEWLINE] It has nothing to do with the number of rich people vs. poor people. [NEWLINE] [NEWLINE] This is true even<mask> you discount conviction rates by doing proper statistical surveys. [NEWLINE] [NEWLINE] [URL] ;pg=PA36&amp;source=gbs_toc_r&amp;cad=4#v=onepage&amp;q&amp;f=false</s><pad>
Label encoding: <s>The *rate* of crime is negatively correlated with income, not just the absolute amount. The only case where this is not true is drug offenses. [NEWLINE] [NEWLINE] It has nothing to do with the number of rich people vs. poor people. [NEWLINE] [NEWLINE] This is true even if you discount conviction rates by doing proper statistical surveys. [NEWLINE] [NEWLINE] [URL] ;pg=PA36&amp;source=gbs_toc_r&amp;cad=4#v=onepage&amp;q&amp;f=false</s><pad>
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] before injuries. [ENDQ] [NEWLINE] <mask> you are willing can I ask<mask> injuries and<mask> severe (like do they still impact you today and will continue to in the future.) [NEWLINE] [NEWLINE] I ask this<mask> from my education I know that athletes are frequently overworked and the eventual serious injury that they receive is due to small micro injuries over time. And that is<mask> i support op's view becuase these athletes often receive lifelong complications. [NEWLINE] [NEWLINE] Olympians are just the athletes that made it without have the big serious one.</s>
Label encoding: <s> [STARTQ] before injuries. [ENDQ] [NEWLINE] If you are willing can I ask what injuries and how severe (like do they still impact you today and will continue to in the future.) [NEWLINE] [NEWLINE] I ask this because from my education I know that athletes are frequently overworked and the eventual serious injury that they receive is due to small micro injuries over time. And that is why i support op's view becuase these athletes often receive lifelong complications. [NEWLINE] [NEWLINE] Olympians are just the athletes that made it without have the big serious one.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Imagine two cavemen who live in different areas. They meet up. They begin talking (this would have to be after the brain development that allowed for speech) and one mentions that some fruit in his area is very good. The other goes to check it out. [NEWLINE] [NEWLINE] Obviously, not every social interaction we experience is purely for survival purposes,<mask><mask> our society evolved and intellectual property became just<mask> valuable<mask> food and shelter, our communication did too.<mask>, from this perspective, sociality is a result of evolutionary reasons.</s>
Label encoding: <s>Imagine two cavemen who live in different areas. They meet up. They begin talking (this would have to be after the brain development that allowed for speech) and one mentions that some fruit in his area is very good. The other goes to check it out. [NEWLINE] [NEWLINE] Obviously, not every social interaction we experience is purely for survival purposes, but as our society evolved and intellectual property became just as valuable as food and shelter, our communication did too. Therefore, from this perspective, sociality is a result of evolutionary reasons.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think<mask> you grew up in a society<mask> all changing rooms and toilet blocks were mixed sex, you probably wouldn't think twice about sharing the open space with the opposite sex,<mask> now that society is moving in the direction of mixed sex changing rooms, it will just take a bit of getting used to... I've been in mixed sex changing rooms at swimming pools and it was fine<mask> there were private cubicles... presumably your toilet block at school has private cubicles,<mask> you are not standing naked in front of everyone?</s>
Label encoding: <s>I think if you grew up in a society where all changing rooms and toilet blocks were mixed sex, you probably wouldn't think twice about sharing the open space with the opposite sex, so now that society is moving in the direction of mixed sex changing rooms, it will just take a bit of getting used to... I've been in mixed sex changing rooms at swimming pools and it was fine because there were private cubicles... presumably your toilet block at school has private cubicles, so you are not standing naked in front of everyone?</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0741, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Legalizing drugs actually makes kids less likely to get hold of them,<mask> then they will only be sold to people 21 and up.<mask> I was in high school, it was always much easier to get marijuana than alcohol. College students who grow marijuana are more than happy to sell to high schoolers,<mask> liquor stores would never provide alcohol to a minor. [NEWLINE] [NEWLINE] The Netherlands,<mask> marijuana has been sold in coffee shops for decades to people 21+, has less than half the rate of teen marijuana use<mask> the US.</s><pad>
Label encoding: <s>Legalizing drugs actually makes kids less likely to get hold of them, because then they will only be sold to people 21 and up. When I was in high school, it was always much easier to get marijuana than alcohol. College students who grow marijuana are more than happy to sell to high schoolers, but liquor stores would never provide alcohol to a minor. [NEWLINE] [NEWLINE] The Netherlands, where marijuana has been sold in coffee shops for decades to people 21+, has less than half the rate of teen marijuana use as the US.</s><pad>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I nevet said or implied they shouldn't tell an adult...its a matter of being offended by something on a screen. We create an environment<mask> we tell kids that something someone says online SHOULD offend them instead of teaching the ways of handling it in other ways.  You take away their ability to bully you<mask> you just see them<mask> words.  They only have meaning you give them. [NEWLINE] [NEWLINE] We do need to hold the bulky accountable for their actions<mask> we<mask> give the bullied tools to handle the situation. </s>
Label encoding: <s>I nevet said or implied they shouldn't tell an adult...its a matter of being offended by something on a screen. We create an environment where we tell kids that something someone says online SHOULD offend them instead of teaching the ways of handling it in other ways.  You take away their ability to bully you when you just see them as words.  They only have meaning you give them. [NEWLINE] [NEWLINE] We do need to hold the bulky accountable for their actions while we also give the bullied tools to handle the situation. </s>
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A resold engagement or wedding ring carries all sorts of nasty baggage. Most people wouldn't buy one to give to their beloved, which means that they can't really be sold for more than scrap.<mask> you buy a good-quality diamond to start out with,<mask>, there are a variety of online communities which buy and sell gems, and<mask> you knew your stuff to start out with, you can resell the gem itself (<mask> not the mount) for a substantial portion of<mask> you paid in the first place.</s>
Label encoding: <s>A resold engagement or wedding ring carries all sorts of nasty baggage. Most people wouldn't buy one to give to their beloved, which means that they can't really be sold for more than scrap. If you buy a good-quality diamond to start out with, however, there are a variety of online communities which buy and sell gems, and if you knew your stuff to start out with, you can resell the gem itself ( if not the mount) for a substantial portion of what you paid in the first place.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The best way to make the process faster is to not have carry on bags that go into the overhead bins. Then to speed it up more you just have people line up via seat number from back of plane to front with window seats going before seats in the middle or aisle. [NEWLINE] [NEWLINE] <mask> soon<mask> you're dealing with people trying to play overhead tetris you're going to be boarding quite slowly no matter<mask> people are lined up. And<mask> you want to make sure there's space for your luggage you line up sooner.</s>
Label encoding: <s>The best way to make the process faster is to not have carry on bags that go into the overhead bins. Then to speed it up more you just have people line up via seat number from back of plane to front with window seats going before seats in the middle or aisle. [NEWLINE] [NEWLINE] As soon as you're dealing with people trying to play overhead tetris you're going to be boarding quite slowly no matter how people are lined up. And since you want to make sure there's space for your luggage you line up sooner.</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, sorry for confusion. [NEWLINE] [NEWLINE] I see nothing erroneous with concepts being valid from birth (citizenship, civil rights, being born into slavery, into royal successorship). Concepts can be innate.<mask> rights cannot be<mask><mask><mask> without a benefactor.<mask> a constitution grants them to particular citizens, they are not a universal higher rule, they are granted by the constitution. [NEWLINE] [NEWLINE] I am simply trying to<mask><mask> without a benefactor, a right is only a fantasy an/or aspiration/wish.</s>
Label encoding: <s>Ok, sorry for confusion. [NEWLINE] [NEWLINE] I see nothing erroneous with concepts being valid from birth (citizenship, civil rights, being born into slavery, into royal successorship). Concepts can be innate. But rights cannot be in my opinion without a benefactor. If a constitution grants them to particular citizens, they are not a universal higher rule, they are granted by the constitution. [NEWLINE] [NEWLINE] I am simply trying to argue that without a benefactor, a right is only a fantasy an/or aspiration/wish.</s>
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because most Nice Guys complain about being rejected<mask> being "nice", and criticize a woman's choice of a partner who is "not nice". And then complain about "Friend-zoning". [NEWLINE] [NEWLINE] <mask> is friend-zone? Friendzoning means, "A woman gives me platonic affection in exchange for displaying platonic qualities. Instead, I want ROMANTIC affection in exchange for displaying platonic qualities. I hate women who provide romantic affection to men who display romantic qualities<mask> it is unfair and shallow." [NEWLINE] </s>
Label encoding: <s>Because most Nice Guys complain about being rejected despite being "nice", and criticize a woman's choice of a partner who is "not nice". And then complain about "Friend-zoning". [NEWLINE] [NEWLINE] What is friend-zone? Friendzoning means, "A woman gives me platonic affection in exchange for displaying platonic qualities. Instead, I want ROMANTIC affection in exchange for displaying platonic qualities. I hate women who provide romantic affection to men who display romantic qualities because it is unfair and shallow." [NEWLINE] </s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not OP<mask> my understanding of heaven is kind of different to yours. It is said that heaven is perfect, and subjects in it are eternally happy. I don't think it's a place to improve yourself and learn.<mask> all sins are forgiven, you are essentially perfect, much like in sims<mask> your stats are all maxed out. You're just there to enjoy yourself with Happiness/health etc at max. In other words, you've reached end game, with nothing to strive for.</s>
Label encoding: <s>I'm not OP but my understanding of heaven is kind of different to yours. It is said that heaven is perfect, and subjects in it are eternally happy. I don't think it's a place to improve yourself and learn. Since all sins are forgiven, you are essentially perfect, much like in sims where your stats are all maxed out. You're just there to enjoy yourself with Happiness/health etc at max. In other words, you've reached end game, with nothing to strive for.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How would you feel<mask> the government decided that a view different than yours is correct? This is easy to think about<mask> people are only allowed to hold, or defend your views.<mask><mask><mask> this was implemented such that you had to be Christian? Or you weren't allowed to be liberal? [NEWLINE] [NEWLINE] You have to realize that your beliefs aren't the absolute truth,<mask> those who differ are wrong. This is more than just religion, it goes for politics, philosophy, even to a good amount of science.</s>
Label encoding: <s>How would you feel if the government decided that a view different than yours is correct? This is easy to think about when people are only allowed to hold, or defend your views. But what if this was implemented such that you had to be Christian? Or you weren't allowed to be liberal? [NEWLINE] [NEWLINE] You have to realize that your beliefs aren't the absolute truth, while those who differ are wrong. This is more than just religion, it goes for politics, philosophy, even to a good amount of science.</s>
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And a new thought just occurred. [NEWLINE] [NEWLINE] <mask> is it<mask> the jet did crash they haven't found the fuel slick or even a single bit of debris<mask>? [NEWLINE] [NEWLINE] Something on that jet would be floating on the surface. [NEWLINE] [NEWLINE] <mask> that flight went down over the ocean years back there was flaming debris marking it at night and a fuel slick in the area for miles marking it during the day. [NEWLINE] [NEWLINE] Those were the first things the rescue crews were looking for<mask> telltale signs and<mask>....... nothing.</s>
Label encoding: <s>And a new thought just occurred. [NEWLINE] [NEWLINE] How is it if the jet did crash they haven't found the fuel slick or even a single bit of debris yet? [NEWLINE] [NEWLINE] Something on that jet would be floating on the surface. [NEWLINE] [NEWLINE] When that flight went down over the ocean years back there was flaming debris marking it at night and a fuel slick in the area for miles marking it during the day. [NEWLINE] [NEWLINE] Those were the first things the rescue crews were looking for as telltale signs and yet....... nothing.</s>
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0787, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except it doesn't actually stand for anything bad. [NEWLINE] [NEWLINE] The civil war was not about slavery.   Slavery was simply the justification. [NEWLINE] [NEWLINE] The real reasons for the war were trade agreements or lack thereof between the North and the South. [NEWLINE] [NEWLINE] You wanna know who's racist<mask> fuck?  Abraham Lincoln.  He didn't want to free the slaves for their benefit, they were just pawns to force the South to secede.  Lincoln wanted to ship them all back to Africa.</s>
Label encoding: <s>Except it doesn't actually stand for anything bad. [NEWLINE] [NEWLINE] The civil war was not about slavery.   Slavery was simply the justification. [NEWLINE] [NEWLINE] The real reasons for the war were trade agreements or lack thereof between the North and the South. [NEWLINE] [NEWLINE] You wanna know who's racist as fuck?  Abraham Lincoln.  He didn't want to free the slaves for their benefit, they were just pawns to force the South to secede.  Lincoln wanted to ship them all back to Africa.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0745, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sociopath might not understand the *<mask> * of the law,<mask> sociopathy won't itself preclude them from understanding whether an act is prohibited and that<mask> they do that act, there will be legal consequences. [NEWLINE] [NEWLINE] They might not understand that murder is morally wrong,<mask> they are capable of understanding there will be legal consequences<mask> they do it. In other words, they choose in full knowledge of the consequences to do something that is legally prohibited, even<mask> they don't get<mask> it should be.</s>
Label encoding: <s>Sociopath might not understand the * why * of the law, but sociopathy won't itself preclude them from understanding whether an act is prohibited and that if they do that act, there will be legal consequences. [NEWLINE] [NEWLINE] They might not understand that murder is morally wrong, but they are capable of understanding there will be legal consequences if they do it. In other words, they choose in full knowledge of the consequences to do something that is legally prohibited, even if they don't get why it should be.</s>
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>IIiiinteresting. I have always wondered<mask> my experiences can generalize to Canada vs US, or<mask> it's more Canada vs California. Or even Manitoba vs California. [NEWLINE] [NEWLINE] In any case, I'm from Winnipeg. I've been to most Canadian Capitals,<mask> never for long enough to really get a feel for things. [NEWLINE] [NEWLINE] <mask>, for<mask> it's worth, I've gone prospecting for digital gold in Silicon Valley, and I take it that here is its own special brand of unusual.</s>
Label encoding: <s>IIiiinteresting. I have always wondered if my experiences can generalize to Canada vs US, or if it's more Canada vs California. Or even Manitoba vs California. [NEWLINE] [NEWLINE] In any case, I'm from Winnipeg. I've been to most Canadian Capitals, but never for long enough to really get a feel for things. [NEWLINE] [NEWLINE] Also, for what it's worth, I've gone prospecting for digital gold in Silicon Valley, and I take it that here is its own special brand of unusual.</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Compared to my now 20 year-old self<mask><mask> I was an idiot at 16 sure,<mask> my 25 year-old self will think the me *now* is silly too, and<mask> on until the day I die. [ENDQ] [NEWLINE] Side note: props to you for realizing this by age 20.<mask> you remember this from now on (and *act*<mask><mask> true belief in it) it will serve you well. I know people twice your age who've<mask> to figure this out. </s>
Label encoding: <s> [STARTQ] Compared to my now 20 year-old self I think I was an idiot at 16 sure, but my 25 year-old self will think the me *now* is silly too, and so on until the day I die. [ENDQ] [NEWLINE] Side note: props to you for realizing this by age 20. If you remember this from now on (and *act* according to true belief in it) it will serve you well. I know people twice your age who've yet to figure this out. </s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0769, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This gives sort of a two sided issue. One being yes, one can remain empathetic and simply filter out emotions that would otherwise hinder one's judgement, or one could have no empathy, and then simply teach themselves<mask> to deal with the situations that empathy is beneficial towards. I argue<mask> that it would be easier to conduct the latter,<mask> during the process to achieve both states of maximum benefit, one would have an easier time starting with no empathy instead of having a vast, draining, amount. </s>
Label encoding: <s>This gives sort of a two sided issue. One being yes, one can remain empathetic and simply filter out emotions that would otherwise hinder one's judgement, or one could have no empathy, and then simply teach themselves how to deal with the situations that empathy is beneficial towards. I argue though that it would be easier to conduct the latter, because during the process to achieve both states of maximum benefit, one would have an easier time starting with no empathy instead of having a vast, draining, amount. </s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see<mask> you're coming from<mask><mask><mask> you have a dog who does not learn quickly,<mask> instead continues to try to escape. [NEWLINE] [NEWLINE] <mask> they can tolerate the pain close to the fence-line, they can escape. Once they escape, they are rewarded with no more shocks,<mask> the dog learns that escaping is a good thing. [NEWLINE] [NEWLINE] Or it may invoke an aggressive response and do more harm than good, psychologically for the animal, and physically<mask> it were to panic and lash out.</s>
Label encoding: <s>I see where you're coming from but what if you have a dog who does not learn quickly, but instead continues to try to escape. [NEWLINE] [NEWLINE] If they can tolerate the pain close to the fence-line, they can escape. Once they escape, they are rewarded with no more shocks, so the dog learns that escaping is a good thing. [NEWLINE] [NEWLINE] Or it may invoke an aggressive response and do more harm than good, psychologically for the animal, and physically if it were to panic and lash out.</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The logic of that's just<mask> it was done doesn't make it right. Plenty of bad things have been done in the past<mask> that's just<mask> things were done. Slavery and segregation were common practices were done in the past<mask> that's<mask> things were done,<mask> those aren't right. [NEWLINE] [NEWLINE] And sure, Japan could have attacked a civilian target like Honolulu for example,<mask> it didn't,<mask> there's no reason to assume that they would have<mask> they attacked a military target instead.</s>
Label encoding: <s>The logic of that's just how it was done doesn't make it right. Plenty of bad things have been done in the past because that's just how things were done. Slavery and segregation were common practices were done in the past because that's how things were done, yet those aren't right. [NEWLINE] [NEWLINE] And sure, Japan could have attacked a civilian target like Honolulu for example, but it didn't, so there's no reason to assume that they would have when they attacked a military target instead.</s>
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well in high school and college I remember learning about<mask> the two distinct brains develop differently both before and after birth...<mask> I suppose I can't say I've seen (or looked for) a study linking a person's sex with their chess ability, the fact...yes fact...that men and women's brain are structurally different would explain away<mask> generally aren't that good or seemingly uninterested at chess. I know<mask> to play basketball,<mask> I don't<mask> I'm very bad at it.</s>
Label encoding: <s>Well in high school and college I remember learning about how the two distinct brains develop differently both before and after birth... while I suppose I can't say I've seen (or looked for) a study linking a person's sex with their chess ability, the fact...yes fact...that men and women's brain are structurally different would explain away why generally aren't that good or seemingly uninterested at chess. I know how to play basketball, but I don't because I'm very bad at it.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think there is a massive leap between not believing in god and refusing to consider the possibility.<mask><mask> that most atheists you are likely to meet would describe themselves<mask> agnostic atheists. This means that they do not know for certain, and do not believe in a god. [NEWLINE] [NEWLINE] <mask> for your consciousness question, I'm sure many would be very interested in a reasonable hypothesis about the origins and functionality of consciousness. They just don't see the god hypothesis<mask> viable,<mask> they think it lacks evidence.</s>
Label encoding: <s>I think there is a massive leap between not believing in god and refusing to consider the possibility. I think that most atheists you are likely to meet would describe themselves as agnostic atheists. This means that they do not know for certain, and do not believe in a god. [NEWLINE] [NEWLINE] As for your consciousness question, I'm sure many would be very interested in a reasonable hypothesis about the origins and functionality of consciousness. They just don't see the god hypothesis as viable, because they think it lacks evidence.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The only places you might need to implement such laws would be in large cities like Chicago or New York, or other urban areas that have an extremely large traffic volume.  It's not something you'd need to worry about in, say, Green Bay, Wisconsin.  The laws would be unnecessary for any<mask> the largest of cities. <mask> the computers themselves are flawless, there'd be no way to get into an accident<mask> they'd likely be programmed to take over before the driver could hit something.</s>
Label encoding: <s>The only places you might need to implement such laws would be in large cities like Chicago or New York, or other urban areas that have an extremely large traffic volume.  It's not something you'd need to worry about in, say, Green Bay, Wisconsin.  The laws would be unnecessary for any but the largest of cities.  If the computers themselves are flawless, there'd be no way to get into an accident as they'd likely be programmed to take over before the driver could hit something.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0692, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see<mask> you're coming from,<mask><mask>.<mask>,<mask> all that truly matters is who can bring superior force, then<mask> point is there in even using the terms "right and wrong"? The idea that Quaritch was "right" implies that there is some moral code<mask> his actions were good.<mask> the only thing that matters is<mask> you can get away with, then there is no need to say that anything is right or wrong, only whether the person in question was successful.</s>
Label encoding: <s>I see where you're coming from, I think. However, if all that truly matters is who can bring superior force, then what point is there in even using the terms "right and wrong"? The idea that Quaritch was "right" implies that there is some moral code where his actions were good. If the only thing that matters is what you can get away with, then there is no need to say that anything is right or wrong, only whether the person in question was successful.</s>
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't really want to achieve anything. I don't get a lot of enjoyment out of achievements, and<mask> I do it's very short lived, like a drug.  I do get a lot of enjoyment out of experiences.  Even experiencing the small things in life, like going for a run in a park, gives me more enjoyment than getting a job promotion. [NEWLINE] [NEWLINE] And yes,<mask><mask> I'm conditioned. Hard work is an ethic that was drilled into me<mask> a child.</s>
Label encoding: <s>I don't really want to achieve anything. I don't get a lot of enjoyment out of achievements, and if I do it's very short lived, like a drug.  I do get a lot of enjoyment out of experiences.  Even experiencing the small things in life, like going for a run in a park, gives me more enjoyment than getting a job promotion. [NEWLINE] [NEWLINE] And yes, I think I'm conditioned. Hard work is an ethic that was drilled into me as a child.</s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Cool. Glad you posted that<mask> I really did think it was pink goop. [NEWLINE] [NEWLINE] I noted McNuggets taste pretty good and have typical nutrition compared to most none-low end nuggets<mask> figured whatever they did, they didn't sacrifice the taste or nutrition,<mask> I didn't really care. [NEWLINE] [NEWLINE] (Obviously McNuggets aren't a health food,<mask> the low end nuggets tend to have way less protein, and typically taste crappier and have a less pleasant texture) </s>
Label encoding: <s>Cool. Glad you posted that as I really did think it was pink goop. [NEWLINE] [NEWLINE] I noted McNuggets taste pretty good and have typical nutrition compared to most none-low end nuggets so figured whatever they did, they didn't sacrifice the taste or nutrition, so I didn't really care. [NEWLINE] [NEWLINE] (Obviously McNuggets aren't a health food, but the low end nuggets tend to have way less protein, and typically taste crappier and have a less pleasant texture) </s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I expect people to act in a way that people like them commonly act,<mask> they don't,  there's nothing wrong with that.   I don't see anything selfish about this view. [NEWLINE] [NEWLINE] I make balloon animals. This is atypical. I'm not insulted<mask> you call it atypical. [NEWLINE] [NEWLINE] Prepare for some irony here.  The difference between all words is purely semantics. I knew<mask> you meant<mask>,  which makes this comment purely semantics. </s>
Label encoding: <s>I expect people to act in a way that people like them commonly act, if they don't,  there's nothing wrong with that.   I don't see anything selfish about this view. [NEWLINE] [NEWLINE] I make balloon animals. This is atypical. I'm not insulted if you call it atypical. [NEWLINE] [NEWLINE] Prepare for some irony here.  The difference between all words is purely semantics. I knew what you meant though,  which makes this comment purely semantics. </s>
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The argument you quoted there is<mask><mask> having children does not count<mask> doing something without regard for harming others. I would<mask><mask> having children does intrinsically create harm<mask> that child will suffer during their lives and the only way to avoid that suffering is to not bring life into the world at all. [NEWLINE] [NEWLINE] And even<mask> overpopulation is not that big of an issue<mask> we think, the destruction of the environment certainly is. Having children increases your carbon footprint massively. [NEWLINE] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] [URL] </s><pad>
Label encoding: <s>The argument you quoted there is assuming that having children does not count as doing something without regard for harming others. I would argue that having children does intrinsically create harm because that child will suffer during their lives and the only way to avoid that suffering is to not bring life into the world at all. [NEWLINE] [NEWLINE] And even if overpopulation is not that big of an issue as we think, the destruction of the environment certainly is. Having children increases your carbon footprint massively. [NEWLINE] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] [URL] </s><pad>
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's<mask> life works<mask>. You villify people for doing bad things<mask> they are alive, often<mask> the people speaking against it want the behavior to change. They don't want these people to die<mask>, and the death of a talented individual will always be considered tragedy<mask> whatever issues they had in life. [NEWLINE] [NEWLINE] <mask> soon<mask> you die, your ability to rectify your situation and turn your life around is now eliminated. That's the defining feature of the problem you described. </s>
Label encoding: <s>That's how life works though. You villify people for doing bad things while they are alive, often because the people speaking against it want the behavior to change. They don't want these people to die however, and the death of a talented individual will always be considered tragedy despite whatever issues they had in life. [NEWLINE] [NEWLINE] As soon as you die, your ability to rectify your situation and turn your life around is now eliminated. That's the defining feature of the problem you described. </s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You would need enough of those people to not be able to make a jury.  I bet the poll had a large amount of people who did not respond<mask> they didn't know the details, didn't know who he was, or didn't care to know.  All you need is those people to make a jury. [NEWLINE] [NEWLINE] <mask>, the idea that 55% of people polled thinks he did the right thing is not nearly indicative of<mask> a society thinks (that is hardly a majority).</s>
Label encoding: <s>You would need enough of those people to not be able to make a jury.  I bet the poll had a large amount of people who did not respond because they didn't know the details, didn't know who he was, or didn't care to know.  All you need is those people to make a jury. [NEWLINE] [NEWLINE] Also, the idea that 55% of people polled thinks he did the right thing is not nearly indicative of what a society thinks (that is hardly a majority).</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My point was, independent of my current social relationships,<mask> I had female genitalia tomorrow and everyone had known me<mask> a woman, then I would have no problem identifying<mask> one. It's incredibly different from the suggestion that I decide that I will state that I am woman starting tomorrow. Again, a big part of this post revolves around the fact that, gender and the things tied to it don't seem to be something inherently mental,<mask> rather something<mask><mask><mask> of social expectation. </s>
Label encoding: <s>My point was, independent of my current social relationships, if I had female genitalia tomorrow and everyone had known me as a woman, then I would have no problem identifying as one. It's incredibly different from the suggestion that I decide that I will state that I am woman starting tomorrow. Again, a big part of this post revolves around the fact that, gender and the things tied to it don't seem to be something inherently mental, but rather something as a result of social expectation. </s>
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think a further thing to consider is that mental illness really blocks perception in the first place.<mask><mask> from the outside it's easy to see that the person can either carry on being mentally ill, commit suicide or do A, B or C, the mental illness clouds the perception of the world<mask> they can only see the first two choices.<mask> they still have free will,<mask> from their mentally ill view of the world their choices are much more limited than<mask> is actually reality. </s>
Label encoding: <s>I think a further thing to consider is that mental illness really blocks perception in the first place. So while from the outside it's easy to see that the person can either carry on being mentally ill, commit suicide or do A, B or C, the mental illness clouds the perception of the world so they can only see the first two choices. So they still have free will, but from their mentally ill view of the world their choices are much more limited than what is actually reality. </s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Galileo started with a hypothesis that he had to keep secret from everyone<mask> he feared mockery, imprisonment, and possible death (execution).  Later, with some data, he started talking about his idea and was imprisoned.  He gradually added data to support his hypothesis. [NEWLINE] [NEWLINE] This is beside the point,<mask>.  My concern is with your mocking tone. <mask> mock people whose ideas are different from yours just<mask> their ideas go against<mask> you believe to be correct?</s><pad>
Label encoding: <s>Galileo started with a hypothesis that he had to keep secret from everyone because he feared mockery, imprisonment, and possible death (execution).  Later, with some data, he started talking about his idea and was imprisoned.  He gradually added data to support his hypothesis. [NEWLINE] [NEWLINE] This is beside the point, however.  My concern is with your mocking tone.  Why mock people whose ideas are different from yours just because their ideas go against what you believe to be correct?</s><pad>
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But having less people in prison would be much better for a country than having those in prison well looked after.<mask><mask><mask> well they're being kept, businesses running prisons offer a financial incentive for putting more people behind bars. This can only lead to an increase in man hours in prison,<mask> other commenters have eloquently pointed out, with longer jail times for minor offenses etc. I'd<mask><mask> the quality of prisoner life is a moot point, in this CMV at least. </s>
Label encoding: <s>But having less people in prison would be much better for a country than having those in prison well looked after. Regardless of how well they're being kept, businesses running prisons offer a financial incentive for putting more people behind bars. This can only lead to an increase in man hours in prison, as other commenters have eloquently pointed out, with longer jail times for minor offenses etc. I'd argue that the quality of prisoner life is a moot point, in this CMV at least. </s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you think private hospitals should be able to turn away someone bleeding to death<mask> they are a particular race or sexual orientation?<mask> not,<mask> do we draw the line?<mask> causes the line to be drawn at one point and not another and<mask> is that justified?<mask> you DO think private hospitals should be allowed to discriminate, then you are putting the right-to-discriminate above the right to life, which I have a fundamental problem with. Perhaps you don't?</s>
Label encoding: <s>Do you think private hospitals should be able to turn away someone bleeding to death because they are a particular race or sexual orientation? If not, where do we draw the line? What causes the line to be drawn at one point and not another and how is that justified? If you DO think private hospitals should be allowed to discriminate, then you are putting the right-to-discriminate above the right to life, which I have a fundamental problem with. Perhaps you don't?</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> the best candidate had a single error [ENDQ] [NEWLINE] This all comes down to opinion<mask> theres no point here. [NEWLINE] [NEWLINE] [STARTQ] The most careful, detail-oriented person can have unexpected things happen in creating/printing the resume which can lead to an error on it. [ENDQ] [NEWLINE] <mask> this..just no. I work in Quality Management/Assurance and can garauntee with 100% certainty that any resume I submit will be entirely free from errors. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] But if the best candidate had a single error [ENDQ] [NEWLINE] This all comes down to opinion so theres no point here. [NEWLINE] [NEWLINE] [STARTQ] The most careful, detail-oriented person can have unexpected things happen in creating/printing the resume which can lead to an error on it. [ENDQ] [NEWLINE] But this..just no. I work in Quality Management/Assurance and can garauntee with 100% certainty that any resume I submit will be entirely free from errors. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A quick and dirty answer response would be the fact that dating sites are actually one of the few types of sites that actually do match you with random people to meet. You couldn't just start talking to random people on social media such<mask> Facebook or chat apps. That's usually deemed taboo. Dating sites,<mask><mask><mask><mask>, do introduce you to new people. Sure, there are chat apps and such,<mask> with a dating site, the idea of meeting is still an implication.</s>
Label encoding: <s>A quick and dirty answer response would be the fact that dating sites are actually one of the few types of sites that actually do match you with random people to meet. You couldn't just start talking to random people on social media such as Facebook or chat apps. That's usually deemed taboo. Dating sites, on the other hand, do introduce you to new people. Sure, there are chat apps and such, but with a dating site, the idea of meeting is still an implication.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Same. Grew up in an atheist household and now earn a living calling people out on their bullshit and telling people to check sources. And I basically have had this attitude<mask><mask><mask> I remember. My parents taught it<mask> "question authority" and put it into practice by letting us kids challenge<mask> we thought were unreasonable household rules, with our reasoned arguments.<mask> that expanded into the lesson of not blindly accepting statements/directions from people generally,<mask> using our judgment.</s>
Label encoding: <s>Same. Grew up in an atheist household and now earn a living calling people out on their bullshit and telling people to check sources. And I basically have had this attitude as long as I remember. My parents taught it as "question authority" and put it into practice by letting us kids challenge what we thought were unreasonable household rules, with our reasoned arguments. But that expanded into the lesson of not blindly accepting statements/directions from people generally, but using our judgment.</s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0864, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I am a woman, and I was brought up to work hard and have a career. I have a hard time comprehending the idea of giving that up, and still having the same amount of respect for myself and pride in<mask> I do. [ENDQ] [NEWLINE] I... You're... Are you saying you wouldn't be proud of your children<mask> you didn't have a job? Childcare isn't a napfest, especially<mask> factoring in caring for disabled children. </s>
Label encoding: <s> [STARTQ] I am a woman, and I was brought up to work hard and have a career. I have a hard time comprehending the idea of giving that up, and still having the same amount of respect for myself and pride in what I do. [ENDQ] [NEWLINE] I... You're... Are you saying you wouldn't be proud of your children if you didn't have a job? Childcare isn't a napfest, especially when factoring in caring for disabled children. </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Thinking that innovation will endless solve our problems is comfortable<mask> not practical in most areas of life. [ENDQ] [NEWLINE] Except it still is in many areas<mask> well, such<mask> medicine and safety devices. [NEWLINE] [NEWLINE] <mask><mask><mask> you are mostly referring to is the application of technological advancements and<mask> we proceed using them<mask> a society. I still think there is no doubt we will continue to have technological advancements, and it's not 'foolhardy' to assume<mask>.</s>
Label encoding: <s> [STARTQ] Thinking that innovation will endless solve our problems is comfortable but not practical in most areas of life. [ENDQ] [NEWLINE] Except it still is in many areas as well, such as medicine and safety devices. [NEWLINE] [NEWLINE] I think what you are mostly referring to is the application of technological advancements and how we proceed using them as a society. I still think there is no doubt we will continue to have technological advancements, and it's not 'foolhardy' to assume so.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>!delta I'm awarding this delta<mask> the magnitude of the issue that I perceived is a lot smaller, quantitatively, than I previously thought.<mask>, I'm still uncomfortable with the concept and am realizing that that probably won't be changed.<mask> my own personal feelings, I am generally in favor of a society in which people are free to make their own decisions, including ones that I'd be uncomfortable making. These numbers strengthen my resolve in this particular issue. </s>
Label encoding: <s>!delta I'm awarding this delta because the magnitude of the issue that I perceived is a lot smaller, quantitatively, than I previously thought. However, I'm still uncomfortable with the concept and am realizing that that probably won't be changed. Despite my own personal feelings, I am generally in favor of a society in which people are free to make their own decisions, including ones that I'd be uncomfortable making. These numbers strengthen my resolve in this particular issue. </s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The founders did NOT include the second amendment to allow citizens to rise up against government - the root of the second amendment was Shays rebellion in 1786 and 1787.  The purpose of a well regulated militia<mask> a means to protect the free state meant that the right to bear arms comes with the responsibility of protecting the government<mask> the government is<mask> makes the country a free state. <mask> would the founders tell the people they could take up arms against themselves? </s><pad>
Label encoding: <s>The founders did NOT include the second amendment to allow citizens to rise up against government - the root of the second amendment was Shays rebellion in 1786 and 1787.  The purpose of a well regulated militia as a means to protect the free state meant that the right to bear arms comes with the responsibility of protecting the government as the government is what makes the country a free state.  Why would the founders tell the people they could take up arms against themselves? </s><pad>
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0841, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That Gallup poll doesn't mention the Robin Hood tax.<mask>, with regards to taxing the rich to redistribute wealth, the article says: [NEWLINE] [NEWLINE] [STARTQ] <mask> the plurality response has shifted modestly between the "favor" and "oppose" positions, Americans have been generally divided on the issue. [ENDQ] [NEWLINE] I don't think your statement that a redistributive tax system (like the Robin Hood tax) is popular is borne out by the resource you chose.</s>
Label encoding: <s>That Gallup poll doesn't mention the Robin Hood tax. Additionally, with regards to taxing the rich to redistribute wealth, the article says: [NEWLINE] [NEWLINE] [STARTQ] Although the plurality response has shifted modestly between the "favor" and "oppose" positions, Americans have been generally divided on the issue. [ENDQ] [NEWLINE] I don't think your statement that a redistributive tax system (like the Robin Hood tax) is popular is borne out by the resource you chose.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with the word based system is that people will be more inclined to construct passwords from public knowledge such<mask> their address, high schools or pets. **This** is the main cause of hacking: social engineering. [NEWLINE] [NEWLINE] For example, the security questions for accounts is often mother's maiden name which can be achieved through a Facebook search<mask> people have made their info public. We need to encourage people to be using passwords that are not associated with their publicly available data.</s>
Label encoding: <s>The problem with the word based system is that people will be more inclined to construct passwords from public knowledge such as their address, high schools or pets. **This** is the main cause of hacking: social engineering. [NEWLINE] [NEWLINE] For example, the security questions for accounts is often mother's maiden name which can be achieved through a Facebook search if people have made their info public. We need to encourage people to be using passwords that are not associated with their publicly available data.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right, really it gets into *babies* at that point (any pre-pubescent child is a baby in my book, at least<mask><mask><mask> this awkward discussion is concerned), which<mask> beyond creepy, is still up to debate at to<mask> it should be treated. Some of the "teenage girls" stuff would be presumably *normal* for a teenage boy. Lolicon was perhaps an extreme example for the concept I was trying to use.</s>
Label encoding: <s>Right, really it gets into *babies* at that point (any pre-pubescent child is a baby in my book, at least as far as this awkward discussion is concerned), which while beyond creepy, is still up to debate at to how it should be treated. Some of the "teenage girls" stuff would be presumably *normal* for a teenage boy. Lolicon was perhaps an extreme example for the concept I was trying to use.</s>
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0658, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] &gt; That is<mask> a woman can force a man to be a biological father against his consent (even<mask> financial abortion was legal), **<mask> a man cannot do the same.** [ENDQ] [NEWLINE] [STARTQ] Are you saying that a man cannot force a woman to be a biological mother against her consent?<mask> that's very, very wrong in many ways. [ENDQ] [NEWLINE] Last I checked rape was illegal and abortion legal. <mask> are you getting at? </s>
Label encoding: <s> [STARTQ] &gt; That is why a woman can force a man to be a biological father against his consent (even if financial abortion was legal), ** while a man cannot do the same.** [ENDQ] [NEWLINE] [STARTQ] Are you saying that a man cannot force a woman to be a biological mother against her consent? Because that's very, very wrong in many ways. [ENDQ] [NEWLINE] Last I checked rape was illegal and abortion legal.  What are you getting at? </s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What the OP is describing is known<mask> the "Untied" States of America theory, after the book written by Juan Enriquez in 2004. [NEWLINE] [NEWLINE] Relevant links: [NEWLINE] [NEWLINE] * [Amazon]( [URL] ;ie=UTF8&amp;qid=1398696537&amp;sr=1-4&amp;keywords=juan+enriquez) [NEWLINE] * [Alternet review of the book]( [URL] )</s>
Label encoding: <s>What the OP is describing is known as the "Untied" States of America theory, after the book written by Juan Enriquez in 2004. [NEWLINE] [NEWLINE] Relevant links: [NEWLINE] [NEWLINE] * [Amazon]( [URL] ;ie=UTF8&amp;qid=1398696537&amp;sr=1-4&amp;keywords=juan+enriquez) [NEWLINE] * [Alternet review of the book]( [URL] )</s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The only thing I can think of is that the North wasn't really fighting to abolish slavery, they were fighting to keep the country intact.  The South seceded (in part)<mask> they believed Lincoln would abolish slavery, and they didn't want that to happen...<mask> Lincoln wasn't planning on doing that.  During the war, there were riots instigated by people too poor to avoid the draft,<mask> they were pro-slavery.</s>
Label encoding: <s>The only thing I can think of is that the North wasn't really fighting to abolish slavery, they were fighting to keep the country intact.  The South seceded (in part) because they believed Lincoln would abolish slavery, and they didn't want that to happen... but Lincoln wasn't planning on doing that.  During the war, there were riots instigated by people too poor to avoid the draft, since they were pro-slavery.</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I feel like the term judgement is proving a little difficult here.<mask><mask> you can "judge" someone's character and maintain a high level of respect for someone,<mask> still be pragmatic in<mask> you deal with that person. For example you could maintain respect for a certain criminal<mask> still sentence him to prison<mask> he broke the law. This wouldn't mean that the respect you offer him is meaningless just<mask> it couldn't keep him out of prison.</s>
Label encoding: <s>I feel like the term judgement is proving a little difficult here. I think you can "judge" someone's character and maintain a high level of respect for someone, but still be pragmatic in how you deal with that person. For example you could maintain respect for a certain criminal but still sentence him to prison because he broke the law. This wouldn't mean that the respect you offer him is meaningless just because it couldn't keep him out of prison.</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Minor leagues pay crap salaries. About 2% of college players will make it to the NFL. For the other 98%, getting a degree is far more valuable than making &lt;40K for a couple years. [NEWLINE] [NEWLINE] Even<mask> you still believe the players are better off in the minors, it's not feasible. Other football leagues have tried in the US and consistently failed. Subpar football doesn't draw crowds and money without the school affiliation. </s>
Label encoding: <s>Minor leagues pay crap salaries. About 2% of college players will make it to the NFL. For the other 98%, getting a degree is far more valuable than making &lt;40K for a couple years. [NEWLINE] [NEWLINE] Even if you still believe the players are better off in the minors, it's not feasible. Other football leagues have tried in the US and consistently failed. Subpar football doesn't draw crowds and money without the school affiliation. </s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Gay people wishing to have equal rights<mask> straight people does not devalue gay culture or gay history. Black history and black culture have remained intact after the Civil Rights Act was signed into law, and it has remained intact after Loving v. Virginia, which allowed interracial marriage in the US.<mask> black history and culture can remain intact after black people were given all legal protections,<mask> should gay culture wither and die at the hands of the same action?</s>
Label encoding: <s>Gay people wishing to have equal rights as straight people does not devalue gay culture or gay history. Black history and black culture have remained intact after the Civil Rights Act was signed into law, and it has remained intact after Loving v. Virginia, which allowed interracial marriage in the US. If black history and culture can remain intact after black people were given all legal protections, why should gay culture wither and die at the hands of the same action?</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But even brute force is incredibly difficult.<mask> do you guarantee that<mask> you're doing your search, one doesn't get in? Or that one doesn't hatch? [NEWLINE] [NEWLINE] Can you guarantee that you looked in every hole that exists in the house, that you've prevented new cockroaches from entering the parts you've examined, and that a cockroach couldn't just have moved in such a way<mask> to remain out of your sight all the time?</s>
Label encoding: <s>But even brute force is incredibly difficult. How do you guarantee that while you're doing your search, one doesn't get in? Or that one doesn't hatch? [NEWLINE] [NEWLINE] Can you guarantee that you looked in every hole that exists in the house, that you've prevented new cockroaches from entering the parts you've examined, and that a cockroach couldn't just have moved in such a way as to remain out of your sight all the time?</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0889, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Basic human rights are whatever we say they are, and they change over time.  These are called many things, the term usually applied around here is natural rights,<mask> there is nothing natural about them. [NEWLINE] [NEWLINE] <mask> a society, we decide<mask> rights are included in the package we call basic.  Contraception hasn't been included in the past,<mask> that is changing and it's a change for the better.  Everyone will benefit.</s>
Label encoding: <s>Basic human rights are whatever we say they are, and they change over time.  These are called many things, the term usually applied around here is natural rights, but there is nothing natural about them. [NEWLINE] [NEWLINE] As a society, we decide what rights are included in the package we call basic.  Contraception hasn't been included in the past, but that is changing and it's a change for the better.  Everyone will benefit.</s>
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean in the same sense that *every* country is unique, sure.<mask> then<mask>'s the point in even saying we're different from other countries? [NEWLINE] [NEWLINE] Typically someone only brings up the fact that we're different in a context of *and better than everyone else*. At least that's<mask> I'm seeing it. Otherwise I recognize that Earth is occupied by almost 200 or<mask> nation states. Some similar to ours, some very different.</s>
Label encoding: <s>I mean in the same sense that *every* country is unique, sure. But then what's the point in even saying we're different from other countries? [NEWLINE] [NEWLINE] Typically someone only brings up the fact that we're different in a context of *and better than everyone else*. At least that's how I'm seeing it. Otherwise I recognize that Earth is occupied by almost 200 or so nation states. Some similar to ours, some very different.</s>
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I just think about it like a simple matter of probability. First we assume that the universe is infinite. Then I say that there is the probability that life exists out there somewhere.<mask> I take this event to infinitely many chances in the infinite universe the event is guaranteed to be true. [NEWLINE] [NEWLINE] And<mask>, in an infinite universe there's a guaranteed chance that life exists out there somewhere. [NEWLINE] [NEWLINE] That's<mask><mask><mask> of it, anyway.</s>
Label encoding: <s>I just think about it like a simple matter of probability. First we assume that the universe is infinite. Then I say that there is the probability that life exists out there somewhere. As I take this event to infinitely many chances in the infinite universe the event is guaranteed to be true. [NEWLINE] [NEWLINE] And so, in an infinite universe there's a guaranteed chance that life exists out there somewhere. [NEWLINE] [NEWLINE] That's how I think of it, anyway.</s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I did explain it further in the next paragraph, that you had problems with. [NEWLINE] [NEWLINE] And, yes, of course it's not everyone. <mask> those with money are far more likely to have the conditions I laid out.  This without are unlikely to.  And there is a huge wealth disparity between black and white.  Yes, there are rich blacks and poor whites,<mask> statistically, it's the other way around. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I did explain it further in the next paragraph, that you had problems with. [NEWLINE] [NEWLINE] And, yes, of course it's not everyone.  But those with money are far more likely to have the conditions I laid out.  This without are unlikely to.  And there is a huge wealth disparity between black and white.  Yes, there are rich blacks and poor whites, but statistically, it's the other way around. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course they say they regret it<mask> they don't want to be consigned to our horrific involuntary mental health care system for the rest of their already unsatisfactory lives. [NEWLINE] [NEWLINE] This should be obvious<mask> of<mask> many will go on to make multiple attempts. [NEWLINE] [NEWLINE] Even<mask> making suicide harder does reduce the suicide rate you haven't shown that is a good thing<mask>.<mask> is less suicide on an already overpopulated world desirable? </s>
Label encoding: <s>Of course they say they regret it because they don't want to be consigned to our horrific involuntary mental health care system for the rest of their already unsatisfactory lives. [NEWLINE] [NEWLINE] This should be obvious because of how many will go on to make multiple attempts. [NEWLINE] [NEWLINE] Even if making suicide harder does reduce the suicide rate you haven't shown that is a good thing yet. Why is less suicide on an already overpopulated world desirable? </s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No you don't. You can go into international waters and renounce your citizenship. [ENDQ] [NEWLINE] Or, you know, I could just kill myself. [NEWLINE] [NEWLINE] The point is that the natural state of human existence has been obviated by the state. And the [few pockets of people living without state influence are being brought under heel]( [URL] ). Your "solution" isn't one,<mask> nobody can survive on a boat indefinitely.</s>
Label encoding: <s> [STARTQ] No you don't. You can go into international waters and renounce your citizenship. [ENDQ] [NEWLINE] Or, you know, I could just kill myself. [NEWLINE] [NEWLINE] The point is that the natural state of human existence has been obviated by the state. And the [few pockets of people living without state influence are being brought under heel]( [URL] ). Your "solution" isn't one, because nobody can survive on a boat indefinitely.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right. They might have depended on being light sleepers and keeping their weapons close,<mask> that wouldn't help against anyone with a bow, or throwing axes or spears. Maybe they could have mitigated the risk somewhat by sleeping far enough apart that no one could get off a clear shot with a ranged weapon, except they quite clearly didn't - they were all close enough together for Katniss to drop a tracker jacker nest on them.</s>
Label encoding: <s>Right. They might have depended on being light sleepers and keeping their weapons close, but that wouldn't help against anyone with a bow, or throwing axes or spears. Maybe they could have mitigated the risk somewhat by sleeping far enough apart that no one could get off a clear shot with a ranged weapon, except they quite clearly didn't - they were all close enough together for Katniss to drop a tracker jacker nest on them.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Isn't the difference between legitimately suicidal, premeditated folks and crisis-related suicidal folks just in the longevity of the crisis they are failing to manage?<mask> difference is it whether it is a single short term crisis or some long term misery/pain? [NEWLINE] [NEWLINE] Long or short term crisis, suicidal people have something in common; They have all reasoned about their situation<mask> failed to see an alternative, and there is always an alternative.  </s>
Label encoding: <s>Isn't the difference between legitimately suicidal, premeditated folks and crisis-related suicidal folks just in the longevity of the crisis they are failing to manage? What difference is it whether it is a single short term crisis or some long term misery/pain? [NEWLINE] [NEWLINE] Long or short term crisis, suicidal people have something in common; They have all reasoned about their situation but failed to see an alternative, and there is always an alternative.  </s>
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0957, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe I was not clear enough in my explanation,<mask> this is one of the things I was alluding to.<mask> terrorists were a (mostly) unified group with a specific goal,<mask> they are often portrayed<mask>, debating the level of reason of such a goal would be easier,<mask> they just cause destruction wherever they go, indiscriminately. The example I used was stabbing the old lady down the street<mask> you hate the President. </s>
Label encoding: <s>Maybe I was not clear enough in my explanation, but this is one of the things I was alluding to. If terrorists were a (mostly) unified group with a specific goal, as they are often portrayed as, debating the level of reason of such a goal would be easier, but they just cause destruction wherever they go, indiscriminately. The example I used was stabbing the old lady down the street because you hate the President. </s>
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're forgetting<mask> sports exist and that's to bring in revenue, which require people to watch them and be interested.  Those sports<mask> there are subjective judges are even more popular<mask> people themselves will disagree and cheer for who they personally thought should win. [NEWLINE] [NEWLINE] Think American Idol and all it's international counterparts. [NEWLINE] [NEWLINE] All forms of highly publicized, competitive entertainment are about profit for the sponsors first and actual competition a very distant second.</s>
Label encoding: <s>You're forgetting why sports exist and that's to bring in revenue, which require people to watch them and be interested.  Those sports where there are subjective judges are even more popular because people themselves will disagree and cheer for who they personally thought should win. [NEWLINE] [NEWLINE] Think American Idol and all it's international counterparts. [NEWLINE] [NEWLINE] All forms of highly publicized, competitive entertainment are about profit for the sponsors first and actual competition a very distant second.</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0771, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And no Johnny's ever grow up to be actors? With all the people who've played roles<mask> killers, thugs, psychos, wouldn't you expect at least some spike in likelihood of violent behavior form these? Care to show us any examples? [NEWLINE] [NEWLINE] Hell, I'll give you a counter example. Danny Trejo used to be a criminal...and now gets that rush from acting and<mask> doesn't bother doing it irl.</s>
Label encoding: <s>And no Johnny's ever grow up to be actors? With all the people who've played roles as killers, thugs, psychos, wouldn't you expect at least some spike in likelihood of violent behavior form these? Care to show us any examples? [NEWLINE] [NEWLINE] Hell, I'll give you a counter example. Danny Trejo used to be a criminal...and now gets that rush from acting and thus doesn't bother doing it irl.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The the "buzz" you get from nicotine is actually acts<mask> a treatment for the symptoms of ADHD.  This is<mask> you see more than 40% of the adult ADHD population smoking compared with 26% in the general population.  Nicotine is no substitute for methylphenidate,<mask> many un-diagnosed adults with ADHD turn to it<mask> it provides a temporary relief. [NEWLINE] [NEWLINE] &gt; [source]( [URL].pdf)</s>
Label encoding: <s>The the "buzz" you get from nicotine is actually acts as a treatment for the symptoms of ADHD.  This is why you see more than 40% of the adult ADHD population smoking compared with 26% in the general population.  Nicotine is no substitute for methylphenidate, but many un-diagnosed adults with ADHD turn to it because it provides a temporary relief. [NEWLINE] [NEWLINE] &gt; [source]( [URL].pdf)</s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> would it take for someone to be a racist in your eyes,<mask> holding negative stereotypes about members of a particular race and acting on those ideas doesn't count? [ENDQ] [NEWLINE] In my eyes, someone is "a racist"<mask> they hold strong irrational feelings of hate and/or contempt and/or general dislike for one or more races for whatever reasons. Whether said person keeps the beliefs in private, or openly expresses them to others.</s>
Label encoding: <s> [STARTQ] What would it take for someone to be a racist in your eyes, if holding negative stereotypes about members of a particular race and acting on those ideas doesn't count? [ENDQ] [NEWLINE] In my eyes, someone is "a racist" when they hold strong irrational feelings of hate and/or contempt and/or general dislike for one or more races for whatever reasons. Whether said person keeps the beliefs in private, or openly expresses them to others.</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're the one making it about gender.  A man or woman can choose to do housework.  The thing they chose to do is a feminine activity.  The point is the activity itself is feminine.  Just like hunting is masculine.  It doesn't mean only men do it or can be good at it, it's just a testosterone driven activity.  I hardly see<mask>'s offensive or sexist about that. </s>
Label encoding: <s>You're the one making it about gender.  A man or woman can choose to do housework.  The thing they chose to do is a feminine activity.  The point is the activity itself is feminine.  Just like hunting is masculine.  It doesn't mean only men do it or can be good at it, it's just a testosterone driven activity.  I hardly see what's offensive or sexist about that. </s>
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How is that even possible? Bush's famous golf press conference occurred in 2002. Reddit wasn't founded until 2005. [NEWLINE] [NEWLINE] Nonetheless, I don't recall Bush being widely criticized on reddit or otherwise for taking vacations quite like Obama has. Maybe it's just selective memory,<mask> I'd welcome examples that disprove me,<mask> regardless, I don't think either should be criticized for taking a day or weekend off once in a<mask>.</s>
Label encoding: <s>How is that even possible? Bush's famous golf press conference occurred in 2002. Reddit wasn't founded until 2005. [NEWLINE] [NEWLINE] Nonetheless, I don't recall Bush being widely criticized on reddit or otherwise for taking vacations quite like Obama has. Maybe it's just selective memory, so I'd welcome examples that disprove me, but regardless, I don't think either should be criticized for taking a day or weekend off once in a while.</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is no law (at least in USA) that restricts someone from playing publicly available games due to their age. A 12 year old can play COD all he wants without any legal repercussions. The ESRB rating is something that game publishers and retailers have agreed to follow, it has no legal basis.<mask>,<mask> OP is a minor, his parents can restrict him (within reason) from certain activities<mask> they see fit.</s>
Label encoding: <s>There is no law (at least in USA) that restricts someone from playing publicly available games due to their age. A 12 year old can play COD all he wants without any legal repercussions. The ESRB rating is something that game publishers and retailers have agreed to follow, it has no legal basis. However, since OP is a minor, his parents can restrict him (within reason) from certain activities as they see fit.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One reason these drugs are expensive (and<mask> you are thinking of something that bodybuilders regularly use, then it's not even expensive) is that they are illegal. [NEWLINE] [NEWLINE] Making them legal would bring down the price a lot. [NEWLINE] [NEWLINE] Sponsorship and recruiting<mask> begins very early on. Teams/Universities look out for talents in High-School and offer scholarships, etc.<mask> makes you think that would not include drugs?</s>
Label encoding: <s>One reason these drugs are expensive (and if you are thinking of something that bodybuilders regularly use, then it's not even expensive) is that they are illegal. [NEWLINE] [NEWLINE] Making them legal would bring down the price a lot. [NEWLINE] [NEWLINE] Sponsorship and recruiting also begins very early on. Teams/Universities look out for talents in High-School and offer scholarships, etc. What makes you think that would not include drugs?</s>
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You have a family now,<mask> it's not about you anymore. Depends,<mask> your wife makes good money that's enough for you all, otherwise take a job that makes money, even<mask> it's against your principles. Shit sucks,<mask> such is life<mask> people depend on you. I'm pretty sure your kid won't say thanks<mask> you're gonna live in poverty just<mask> his dad wanted to play a good guy.</s>
Label encoding: <s>You have a family now, so it's not about you anymore. Depends, if your wife makes good money that's enough for you all, otherwise take a job that makes money, even if it's against your principles. Shit sucks, but such is life when people depend on you. I'm pretty sure your kid won't say thanks if you're gonna live in poverty just because his dad wanted to play a good guy.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe shitty things cause shitty problems.  Maybe your sacred cows are just cows. [NEWLINE] [NEWLINE] God, don't ever entertain that intent doesn't equal result, you uncritical partisan.  You're talking about directly impacting the monetary policy, and it's balanced on the backs of the poor.  That inequality rises the more meddling we've done.  Correlation doesn't imply causation,<mask> it does suggestively wiggle its eyebrows.</s>
Label encoding: <s>Maybe shitty things cause shitty problems.  Maybe your sacred cows are just cows. [NEWLINE] [NEWLINE] God, don't ever entertain that intent doesn't equal result, you uncritical partisan.  You're talking about directly impacting the monetary policy, and it's balanced on the backs of the poor.  That inequality rises the more meddling we've done.  Correlation doesn't imply causation, but it does suggestively wiggle its eyebrows.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have read that in the Netherlands,<mask> prostitution is legal, there is quite a problem with human trafficking from Eastern Europe, Asia, and Africa. <mask> read that there are major problems with pimping,<mask> theoretically, it is illegal. [NEWLINE] [NEWLINE] <mask> you want to check into this quickly, read the "Foreign Prostitutes" and lower-down  "Human Trafficking" sections of this wiki page: [NEWLINE] [URL] </s>
Label encoding: <s>I have read that in the Netherlands, where prostitution is legal, there is quite a problem with human trafficking from Eastern Europe, Asia, and Africa.  Also read that there are major problems with pimping, although theoretically, it is illegal. [NEWLINE] [NEWLINE] If you want to check into this quickly, read the "Foreign Prostitutes" and lower-down  "Human Trafficking" sections of this wiki page: [NEWLINE] [URL] </s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right. I was forgetting that bears seem to only attack<mask> they think that humans are providing their food, not necessarily that humans are their food. Obviously to a bear, food is food.<mask> at the same time I still feel like that's the wild, and a risk that you are taking<mask> you are hiking or camping in an area with a large bear population. I'm not sold,<mask> I'm torn now.</s>
Label encoding: <s>You're right. I was forgetting that bears seem to only attack because they think that humans are providing their food, not necessarily that humans are their food. Obviously to a bear, food is food. But at the same time I still feel like that's the wild, and a risk that you are taking if you are hiking or camping in an area with a large bear population. I'm not sold, but I'm torn now.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Surely politics... rely quite heavily on experience? [ENDQ] [NEWLINE] Perhaps<mask> said experience was professional in nature.<mask> someone has foreign policy or research experience, obviously their opinions hold more weight than a 16 year old debater.<mask><mask> it pertains to the general public, experience in life doesn't really matter in most cases.<mask> matters is the aggregate evidence each party collects from experts and then presents to the other. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Surely politics... rely quite heavily on experience? [ENDQ] [NEWLINE] Perhaps if said experience was professional in nature. If someone has foreign policy or research experience, obviously their opinions hold more weight than a 16 year old debater. But as it pertains to the general public, experience in life doesn't really matter in most cases. What matters is the aggregate evidence each party collects from experts and then presents to the other. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This delta is currently disallowed<mask> your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for<mask> /u/LeonardPeikoff changed your view.<mask> you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>This delta is currently disallowed as your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for how /u/LeonardPeikoff changed your view. If you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, it's a region thing. In my country I have never ever heard someone say something in imperial/US or UK units. Everything here is in metrics. [NEWLINE] [NEWLINE] I know, I was rough on that title (I tried to change it after posted), I meant in practicality it *might not be useful. [NEWLINE] [NEWLINE] <mask> in conclusion it's a region (only American) thing that they still use. =)</s><pad>
Label encoding: <s>Well, it's a region thing. In my country I have never ever heard someone say something in imperial/US or UK units. Everything here is in metrics. [NEWLINE] [NEWLINE] I know, I was rough on that title (I tried to change it after posted), I meant in practicality it *might not be useful. [NEWLINE] [NEWLINE] So in conclusion it's a region (only American) thing that they still use. =)</s><pad>
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Class has always been a thing<mask> the first civilization. Skin based racism is a relatively new thing. Cultural racism was a thing for a long time<mask>. I could see a post apocalyptic state getting over racism<mask> still having a ton of class struggle.<mask> don't forget the people of the capitol have had decades to develop a separate culture that could see the people of the districts<mask> barbarians, much like ancient romans.</s>
Label encoding: <s>Class has always been a thing since the first civilization. Skin based racism is a relatively new thing. Cultural racism was a thing for a long time though. I could see a post apocalyptic state getting over racism but still having a ton of class struggle. Also don't forget the people of the capitol have had decades to develop a separate culture that could see the people of the districts as barbarians, much like ancient romans.</s>
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm saying that an unexpected, negative change which affects both partners in a relationship is something that should be up for compromise. [NEWLINE] [NEWLINE] Let it be known that I made no comparison between alcoholism and libido. [NEWLINE] [NEWLINE] I never said that low libido is<mask> grave<mask> alcoholism. I used my example to illustrate that not all change is healthy, and simply that decisions that affect both partners should be up for discussion and compromise.</s><pad>
Label encoding: <s>I'm saying that an unexpected, negative change which affects both partners in a relationship is something that should be up for compromise. [NEWLINE] [NEWLINE] Let it be known that I made no comparison between alcoholism and libido. [NEWLINE] [NEWLINE] I never said that low libido is as grave as alcoholism. I used my example to illustrate that not all change is healthy, and simply that decisions that affect both partners should be up for discussion and compromise.</s><pad>
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0774, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not of the mission<mask> of his job, and in general in obligation to reputation of<mask> he works. That's generally true for most jobs: you can get work done, bit<mask> you're a net detriment to the company, they have no reason to hire you in the first place. [NEWLINE] [NEWLINE] In not saying this guy is a net detriment,<mask> the statement that "his only responsibility is the mission" is baseless.</s>
Label encoding: <s>Not of the mission but of his job, and in general in obligation to reputation of where he works. That's generally true for most jobs: you can get work done, bit if you're a net detriment to the company, they have no reason to hire you in the first place. [NEWLINE] [NEWLINE] In not saying this guy is a net detriment, but the statement that "his only responsibility is the mission" is baseless.</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is probably a shitty argument,<mask> seriously. Meet someone who has had sexual abuse/rape during childhood. Flashbacks and trust issues are permanent without a lot of therapy. Even with therapy, permanent traumatization can still be there. [NEWLINE] [NEWLINE] I can't explain<mask> goes through the mind of someone after they have been raped. I don't think anyone who hasn't been raped can truly understand<mask> it's like. </s>
Label encoding: <s>This is probably a shitty argument, but seriously. Meet someone who has had sexual abuse/rape during childhood. Flashbacks and trust issues are permanent without a lot of therapy. Even with therapy, permanent traumatization can still be there. [NEWLINE] [NEWLINE] I can't explain what goes through the mind of someone after they have been raped. I don't think anyone who hasn't been raped can truly understand what it's like. </s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Russia/China have relatively high standard of living compared to countires in political dissaray.  I believe that Canada is<mask> much more socialist than the U.S.<mask> I could be wrong.  Switzerland was already mentioned in /u/deten's post.  I don't know much about the rest.  I wasn't disagreeing with you before, just wanted you to make you argument more concrete.</s>
Label encoding: <s>Russia/China have relatively high standard of living compared to countires in political dissaray.  I believe that Canada is also much more socialist than the U.S. but I could be wrong.  Switzerland was already mentioned in /u/deten's post.  I don't know much about the rest.  I wasn't disagreeing with you before, just wanted you to make you argument more concrete.</s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So you're saying that deliberately killing someone is okay simply<mask> we are not sure<mask> we are killing someone.<mask> someone set up a lever that had a 50% chance of murdering an innocent person,<mask> the right to move your arm and press down levers is protected by our freedom of choice, would you consider it in their right to press that lever simply<mask> they can never know<mask> it will actually kill the innocent person?</s>
Label encoding: <s>So you're saying that deliberately killing someone is okay simply because we are not sure if we are killing someone. So someone set up a lever that had a 50% chance of murdering an innocent person, while the right to move your arm and press down levers is protected by our freedom of choice, would you consider it in their right to press that lever simply because they can never know if it will actually kill the innocent person?</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because people are not intelligent. They don't realize that sometimes the constitution must be stretched in order for our country to be successful. They believe that the forefathers were infallible and made a perfect document which never has to be amended.<mask>, any law that does not follow this document is immoral. Their belief in the constitution is<mask> embedded in them, and I have no idea<mask>. The constitution is plainly flawed.</s>
Label encoding: <s>Because people are not intelligent. They don't realize that sometimes the constitution must be stretched in order for our country to be successful. They believe that the forefathers were infallible and made a perfect document which never has to be amended. Therefore, any law that does not follow this document is immoral. Their belief in the constitution is so embedded in them, and I have no idea why. The constitution is plainly flawed.</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not talking about the average suicide bomber prodded by Hamas into blowing up an Israeli bus.  The leadership of Al Qaeda deliberately sought to goad the USA into a devastating invasion of Afghanistan,<mask> they were quite comfortable triggering massive retribution and destroying innocent Muslims. [NEWLINE] [NEWLINE] Are you<mask> desperate to advance a contrarian, counter-'Murican position that you would rather trust Al Qaeda than Obama with a nuke?</s>
Label encoding: <s>I'm not talking about the average suicide bomber prodded by Hamas into blowing up an Israeli bus.  The leadership of Al Qaeda deliberately sought to goad the USA into a devastating invasion of Afghanistan, so they were quite comfortable triggering massive retribution and destroying innocent Muslims. [NEWLINE] [NEWLINE] Are you so desperate to advance a contrarian, counter-'Murican position that you would rather trust Al Qaeda than Obama with a nuke?</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was born into a Romanian orphanage. I don't know who my birth father is. [NEWLINE] [NEWLINE] I'm probably not going to have sex until vasectomies can be reliably reversed - I can't raise a child<mask>, and I'm **not** going to fuck up like my birth parents - my child will grow up in goddamn luxury or I won't have one. [NEWLINE] [NEWLINE] I hate the human body.</s>
Label encoding: <s>I was born into a Romanian orphanage. I don't know who my birth father is. [NEWLINE] [NEWLINE] I'm probably not going to have sex until vasectomies can be reliably reversed - I can't raise a child yet, and I'm **not** going to fuck up like my birth parents - my child will grow up in goddamn luxury or I won't have one. [NEWLINE] [NEWLINE] I hate the human body.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> Home Owner Associations (HOAs) are good and *everyone* should live<mask> there is one. [ENDQ] [NEWLINE] <mask>,<mask> we could pull just one example of a HOA horror story, that would be enough to CYV, right?<mask> just one person who reasonably and justifiably hates their HOA with the boiling fire of a thousand suns falls into the group of everyone, right?</s>
Label encoding: <s> [STARTQ] I think Home Owner Associations (HOAs) are good and *everyone* should live where there is one. [ENDQ] [NEWLINE] So, if we could pull just one example of a HOA horror story, that would be enough to CYV, right? Because just one person who reasonably and justifiably hates their HOA with the boiling fire of a thousand suns falls into the group of everyone, right?</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right,<mask><mask> the server does not end up making the equivalent of minimum wage, the employer must compensate them sufficiently to make it<mask> they do.  Basically,<mask> nobody ever tipped their servers,<mask> strictly speaking their wage might be around $3/hr, their employers would be required to make up the extra $4.25/hr to make it<mask> they are paid at *least* minimum wage.</s>
Label encoding: <s>Right, but if the server does not end up making the equivalent of minimum wage, the employer must compensate them sufficiently to make it so they do.  Basically, if nobody ever tipped their servers, although strictly speaking their wage might be around $3/hr, their employers would be required to make up the extra $4.25/hr to make it so they are paid at *least* minimum wage.</s>
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Decision trees like CHAID work pretty well for things like direct marketing, medicine, and psychiatry<mask> classification and predictability are important.<mask> you're trying to make politically unified areas that could work,<mask> gerrymandering of this fashion produces extreme politicians. Do we really want to create a situation<mask> we're actively creating situations<mask> the Michelle Bachmann's of the world have an easy road to political power?</s>
Label encoding: <s>Decision trees like CHAID work pretty well for things like direct marketing, medicine, and psychiatry where classification and predictability are important. If you're trying to make politically unified areas that could work, but gerrymandering of this fashion produces extreme politicians. Do we really want to create a situation where we're actively creating situations where the Michelle Bachmann's of the world have an easy road to political power?</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's pretty widely accepted<mask> common knowledge that "intelligence" (not actually a quantifiable genetic trait,<mask> receptiveness to information is,) is passed down genetically. [NEWLINE] [NEWLINE] <mask>, being good or bad with money, being a hard worker, being able to lift yourself out of bad situations are not "intelligence." They are habits, and they are not predominately passed down through genetics, they are learned behaviours. </s>
Label encoding: <s>It's pretty widely accepted as common knowledge that "intelligence" (not actually a quantifiable genetic trait, but receptiveness to information is,) is passed down genetically. [NEWLINE] [NEWLINE] However, being good or bad with money, being a hard worker, being able to lift yourself out of bad situations are not "intelligence." They are habits, and they are not predominately passed down through genetics, they are learned behaviours. </s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can see<mask> you would say that and can understand<mask> you could see it that way.<mask> in my mind<mask> the heart starts beating they are a living person and<mask> there should be a good reason for ending the life of that person. [NEWLINE] [NEWLINE] I feel like all the responses are missing the question. Everyone wants to argue on abortion yes or no not on should we be focusing on abortion or contracepton. [NEWLINE] </s>
Label encoding: <s>I can see how you would say that and can understand how you could see it that way. But in my mind when the heart starts beating they are a living person and therefore there should be a good reason for ending the life of that person. [NEWLINE] [NEWLINE] I feel like all the responses are missing the question. Everyone wants to argue on abortion yes or no not on should we be focusing on abortion or contracepton. [NEWLINE] </s>
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This delta is currently disallowed<mask> your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for<mask> /u/huadpe changed your view.<mask> you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>This delta is currently disallowed as your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for how /u/huadpe changed your view. If you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Something to keep in mind is to realize that there were schools like this in the US for minority students in the early/mid 1900s that,<mask> they didn't solve the issue in it of themselves, they gathered enough data to<mask> "regular" schools could tackle the issue of harassment and segregation of the populace in a meaningful way. This isn't a permanent solution; just a way for us to get to one.</s>
Label encoding: <s>Something to keep in mind is to realize that there were schools like this in the US for minority students in the early/mid 1900s that, while they didn't solve the issue in it of themselves, they gathered enough data to where "regular" schools could tackle the issue of harassment and segregation of the populace in a meaningful way. This isn't a permanent solution; just a way for us to get to one.</s>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you mean in general (i.e. on average?), or do you mean that those majors<mask> a whole require less critical thinking? [NEWLINE] [NEWLINE] I did philosophy and literature<mask> majors and it was a high level of critical thinking. A lot of STEM work isn't critical thinking at all, it's mastery of a body of techniques and creative thinking about<mask> to solve problems, this isn't necessarily critical thinking.</s>
Label encoding: <s>Do you mean in general (i.e. on average?), or do you mean that those majors as a whole require less critical thinking? [NEWLINE] [NEWLINE] I did philosophy and literature as majors and it was a high level of critical thinking. A lot of STEM work isn't critical thinking at all, it's mastery of a body of techniques and creative thinking about how to solve problems, this isn't necessarily critical thinking.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, I read a lot of articles, including your link, about omniscience and free will and, boy, is that a difficult combination! [NEWLINE] [NEWLINE] And<mask> we're gonna agree that free will is an illusion, God doesn't even have to be outside of time.<mask><mask> He is *outside* of time or simply knows everything in the future, to me is different words for the same outcome.</s><pad>
Label encoding: <s>Yeah, I read a lot of articles, including your link, about omniscience and free will and, boy, is that a difficult combination! [NEWLINE] [NEWLINE] And if we're gonna agree that free will is an illusion, God doesn't even have to be outside of time. Although if He is *outside* of time or simply knows everything in the future, to me is different words for the same outcome.</s><pad>
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you have a shitty job,<mask> exactly is that my problem? [ENDQ] [NEWLINE] That right there is the very essence of being an asshole<mask>. [NEWLINE] [NEWLINE] Of course its not your problem. That's the point. It's their problem. And you are causing them problems.<mask> you are a nice person you consider other peoples problems and try to avoid causing them, not<mask><mask> you are an asshole.</s>
Label encoding: <s> [STARTQ] So you have a shitty job, how exactly is that my problem? [ENDQ] [NEWLINE] That right there is the very essence of being an asshole though. [NEWLINE] [NEWLINE] Of course its not your problem. That's the point. It's their problem. And you are causing them problems. If you are a nice person you consider other peoples problems and try to avoid causing them, not so if you are an asshole.</s>
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I went to a four year university in WI and was charged with possession my freshman year and my financial aid was not taken away. I know about 10 people that got possession tickets<mask> university students at various schools in WI and not one got their aid taken away.  Even a friend of mine that got arrested for selling didn't lose his aid.  I've never personally heard of someone losing it over weed.</s>
Label encoding: <s>I went to a four year university in WI and was charged with possession my freshman year and my financial aid was not taken away. I know about 10 people that got possession tickets while university students at various schools in WI and not one got their aid taken away.  Even a friend of mine that got arrested for selling didn't lose his aid.  I've never personally heard of someone losing it over weed.</s>
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agreed, I just wanted to strictly speak to that one statement. The comment in itself<mask><mask> with<mask> it all comes with being the one everyone blames for everything during his presidency. That's the only reason I'd say I don't get<mask> people do it for the purpose of prestige, power or leaving a legacy<mask> he's being used<mask> a shield by the people who make some of the mistakes.</s>
Label encoding: <s>Agreed, I just wanted to strictly speak to that one statement. The comment in itself I agree with but it all comes with being the one everyone blames for everything during his presidency. That's the only reason I'd say I don't get why people do it for the purpose of prestige, power or leaving a legacy since he's being used as a shield by the people who make some of the mistakes.</s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] which is the metabolic variation [ENDQ] [NEWLINE] No.  The variation in men's and women's metabolism is expressed the other example,<mask> the difference is 228, not 600 (<mask> you said) or 562 (<mask> the "average man/woman" example shows). [NEWLINE] [NEWLINE] A woman does not have to eat "much less" than a man of the same size to acheive the same results.</s><pad>
Label encoding: <s> [STARTQ] which is the metabolic variation [ENDQ] [NEWLINE] No.  The variation in men's and women's metabolism is expressed the other example, where the difference is 228, not 600 ( as you said) or 562 ( as the "average man/woman" example shows). [NEWLINE] [NEWLINE] A woman does not have to eat "much less" than a man of the same size to acheive the same results.</s><pad>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] unethical [ENDQ] [NEWLINE] [STARTQ] contract [ENDQ] [NEWLINE] Pick one.<mask> you find a loophole in your country's Law that lets you abuse others for personal gain, you will probably still find it unethical. [NEWLINE] [NEWLINE] My point is that there being no contract that legally binds you to watch the ads on a website, doesn't imply it isn't unethical to bypass them. Not on its own. [NEWLINE] [NEWLINE] Edit: spelling</s>
Label encoding: <s> [STARTQ] unethical [ENDQ] [NEWLINE] [STARTQ] contract [ENDQ] [NEWLINE] Pick one. If you find a loophole in your country's Law that lets you abuse others for personal gain, you will probably still find it unethical. [NEWLINE] [NEWLINE] My point is that there being no contract that legally binds you to watch the ads on a website, doesn't imply it isn't unethical to bypass them. Not on its own. [NEWLINE] [NEWLINE] Edit: spelling</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Marriage is a contract, in order to get a marriage license you must affirm that both you and your partner are eligible to marry (both of age, consenting adults, not already married, not directly related, etc.)  In claiming to be eligible<mask> not meeting all the criteria you have committed a felony in lying under oath, possibly up to three times depending on the form of the marriage. </s>
Label encoding: <s>Marriage is a contract, in order to get a marriage license you must affirm that both you and your partner are eligible to marry (both of age, consenting adults, not already married, not directly related, etc.)  In claiming to be eligible despite not meeting all the criteria you have committed a felony in lying under oath, possibly up to three times depending on the form of the marriage. </s>
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; Thanks for the evidence! [NEWLINE] [NEWLINE] Edit: /u/anriana posted the only response with actual evidence. Scientific evidence is more persuasive to me than anecdotal evidence (<mask> anecdotal evidence is useful too!). The article cited research which showed that kids are not damaged by being in daycare, and<mask> really matters is whether or not the daycare is high-quality.</s>
Label encoding: <s>&amp;#8710; Thanks for the evidence! [NEWLINE] [NEWLINE] Edit: /u/anriana posted the only response with actual evidence. Scientific evidence is more persuasive to me than anecdotal evidence ( although anecdotal evidence is useful too!). The article cited research which showed that kids are not damaged by being in daycare, and what really matters is whether or not the daycare is high-quality.</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No,<mask><mask> he's talking about the [MythBuster's episode on this.]( [URL] )<mask> the test resulted that all toothbrushes had fecal matter on them, albeit tiny tiny particles, even the control ones. [It might<mask> be that 93% of our shoes contain fecal matter]( [URL] /)<mask><mask>,<mask>...try not to think about that too hard, ok?</s>
Label encoding: <s>No, I think he's talking about the [MythBuster's episode on this.]( [URL] ) But the test resulted that all toothbrushes had fecal matter on them, albeit tiny tiny particles, even the control ones. [It might also be that 93% of our shoes contain fecal matter]( [URL] /) also though, so...try not to think about that too hard, ok?</s>
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I haven't thought about the case of monuments to soldiers in the American revolution. [NEWLINE] [NEWLINE] [STARTQ] <mask> about the Jefferson Memorial? He owned slaves. [ENDQ] [NEWLINE] I need to think more on this<mask> well,<mask> right now it seems the best course of action would be to acknowledge that he owned slaves and told all of the truth with a plaque or something in the most prominent and central part of Jefferson Memorial.</s>
Label encoding: <s>I haven't thought about the case of monuments to soldiers in the American revolution. [NEWLINE] [NEWLINE] [STARTQ] What about the Jefferson Memorial? He owned slaves. [ENDQ] [NEWLINE] I need to think more on this as well, but right now it seems the best course of action would be to acknowledge that he owned slaves and told all of the truth with a plaque or something in the most prominent and central part of Jefferson Memorial.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While it is certainly true that China and Russia have the bigger military neither,<mask> especially China, really have the force projection capability to launch an actual attack on the UK. A nuclear strike on the UK would be devastating with or without trident,<mask> having it more or less means that any strike would aim for total annihilation, hoping that the orders to the submarines<mask> written aren't to retaliate. </s>
Label encoding: <s>While it is certainly true that China and Russia have the bigger military neither, though especially China, really have the force projection capability to launch an actual attack on the UK. A nuclear strike on the UK would be devastating with or without trident, but having it more or less means that any strike would aim for total annihilation, hoping that the orders to the submarines as written aren't to retaliate. </s>
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay.<mask> abortion doesn't happen before people exist, or to non life. That was a reactionary postmodernist idea made<mask> people realized that people were not going to continue supporting the real logic that was used for it. That is, the logic about bodily usage, which ultimately was just an extension of overall, certain people wanting abortion in general at any cost, and willing to ignore the ethics.</s>
Label encoding: <s>Okay. But abortion doesn't happen before people exist, or to non life. That was a reactionary postmodernist idea made since people realized that people were not going to continue supporting the real logic that was used for it. That is, the logic about bodily usage, which ultimately was just an extension of overall, certain people wanting abortion in general at any cost, and willing to ignore the ethics.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it depends on which tv station you watch and which songs you listen to. It's not like pop-culture revolves around one single aspect. Having actually gone to high school, I can tell you that girls are much less focused on hooking up for the night and MUCH more focused on all that "i wanna spend my life wiff yhu babyyyy" kinda bs. </s>
Label encoding: <s>I think it depends on which tv station you watch and which songs you listen to. It's not like pop-culture revolves around one single aspect. Having actually gone to high school, I can tell you that girls are much less focused on hooking up for the night and MUCH more focused on all that "i wanna spend my life wiff yhu babyyyy" kinda bs. </s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You seem to be misinformed about the wide availability of expensive journals. University libraries are struggling with the exploding cost of the - very profitable - publishers. Look up the long list of scientists who are boycotting Elsevier for this exact reason. [NEWLINE] [NEWLINE] And congratulations for deciding that students and scientists in most of the world should be blocked from scientific research results<mask> you personally don't see a problem.</s>
Label encoding: <s>You seem to be misinformed about the wide availability of expensive journals. University libraries are struggling with the exploding cost of the - very profitable - publishers. Look up the long list of scientists who are boycotting Elsevier for this exact reason. [NEWLINE] [NEWLINE] And congratulations for deciding that students and scientists in most of the world should be blocked from scientific research results because you personally don't see a problem.</s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] compared to other illegal activities, like human trafficking, it seems a lot less severe in terms of immorality. [ENDQ] [NEWLINE] I don't agree with this reasoning. <mask> something is bad, then it is bad, full stop.  I won't absolve a robber of their blame simply<mask> they didn't murder anyone, and I won't do the same for poachers.</s>
Label encoding: <s> [STARTQ] compared to other illegal activities, like human trafficking, it seems a lot less severe in terms of immorality. [ENDQ] [NEWLINE] I don't agree with this reasoning.  If something is bad, then it is bad, full stop.  I won't absolve a robber of their blame simply because they didn't murder anyone, and I won't do the same for poachers.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I recognize slow deaths are horrific, and I sympathize with people who choose to die. Hell,<mask> I was diagnosed with terminal cancer I might myself choose death.<mask> my point is there's no dignity in that decision. [NEWLINE] [NEWLINE] [STARTQ] <mask> their mind still works, their body still functions [ENDQ] [NEWLINE] This is misleading. It implies that dying with these abilities somehow preserves them. Death preserves nothing.</s>
Label encoding: <s>I recognize slow deaths are horrific, and I sympathize with people who choose to die. Hell, if I was diagnosed with terminal cancer I might myself choose death. But my point is there's no dignity in that decision. [NEWLINE] [NEWLINE] [STARTQ] while their mind still works, their body still functions [ENDQ] [NEWLINE] This is misleading. It implies that dying with these abilities somehow preserves them. Death preserves nothing.</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0783, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If a woman goes into a guys apartment, and then says yes to the sex, or consents in other ways it's her fault.<mask><mask> she doesn't consent, she is raped. I'm not trying to blame the victim, I'm saying the victim can't use their vulnerability<mask> an excuse to null the consent that they did give (<mask><mask> they put themselves in their position).</s>
Label encoding: <s>If a woman goes into a guys apartment, and then says yes to the sex, or consents in other ways it's her fault. But if she doesn't consent, she is raped. I'm not trying to blame the victim, I'm saying the victim can't use their vulnerability as an excuse to null the consent that they did give ( given that they put themselves in their position).</s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually I've got some friends who really appreciate trigger warnings<mask> it gives them a "heads up"<mask> they can take a couple of minutes to mentally prepare themselves before reading or watching the content. Or they can choose to avoid it or put it off until they're in a better state of mind or whatever. Just helps you make a more informed decision and not get taken by surprise, basically.</s>
Label encoding: <s>Actually I've got some friends who really appreciate trigger warnings because it gives them a "heads up" so they can take a couple of minutes to mentally prepare themselves before reading or watching the content. Or they can choose to avoid it or put it off until they're in a better state of mind or whatever. Just helps you make a more informed decision and not get taken by surprise, basically.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Any decision made by a person is going to affect the rest of society in some way, whether they want to or not. And on average, making risky choices like smoking will affect the rest of society negatively. It doesn't matter that the person making the decision didn't want to make his or her choices affect society. [ENDQ] [NEWLINE] <mask> does an individuals choice to do drug harm others?</s>
Label encoding: <s> [STARTQ] Any decision made by a person is going to affect the rest of society in some way, whether they want to or not. And on average, making risky choices like smoking will affect the rest of society negatively. It doesn't matter that the person making the decision didn't want to make his or her choices affect society. [ENDQ] [NEWLINE] How does an individuals choice to do drug harm others?</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If rules and morals were simple, we could have given control of the justice system to computers. Of course views are changeable<mask> they are held by humans... I don't quite understand<mask> you want to have changed. And<mask><mask>, wouldn't having this view changed paradoxically reinforce this view? [NEWLINE] [NEWLINE]...Did you make this CMV just to short circuit our internal logic processors?</s>
Label encoding: <s>If rules and morals were simple, we could have given control of the justice system to computers. Of course views are changeable if they are held by humans... I don't quite understand what you want to have changed. And in fact, wouldn't having this view changed paradoxically reinforce this view? [NEWLINE] [NEWLINE]...Did you make this CMV just to short circuit our internal logic processors?</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I acknowledge there are some problems with such conflicts of interest [ENDQ] [NEWLINE] Well, first think that the purpose of the institution is that people obey the law. [NEWLINE] And this institution gets funded by people not obeying the law. [NEWLINE] [NEWLINE] See the contradiction?  It's like firemen getting paid only<mask> there are fires, and the longer the fire the more they get paid.  </s>
Label encoding: <s> [STARTQ] I acknowledge there are some problems with such conflicts of interest [ENDQ] [NEWLINE] Well, first think that the purpose of the institution is that people obey the law. [NEWLINE] And this institution gets funded by people not obeying the law. [NEWLINE] [NEWLINE] See the contradiction?  It's like firemen getting paid only when there are fires, and the longer the fire the more they get paid.  </s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you know musical theory?<mask> you do, you might be aware that a 'normal, musical' ending is not hard at all to create (<mask> with cadences and whatnot).<mask> it's not really that much about effort and,<mask><mask>, in some cases it fits better than other endings. I hope this is convincing, this is my first post in this subreddit :)</s>
Label encoding: <s>Do you know musical theory? If you do, you might be aware that a 'normal, musical' ending is not hard at all to create ( what with cadences and whatnot). So it's not really that much about effort and, IMO, in some cases it fits better than other endings. I hope this is convincing, this is my first post in this subreddit :)</s>
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The 'rape is rape' isn't to say both types are the same<mask> more like 'theft is theft'<mask> dealing with theft crimes that tend to be washed away or dismissed. Theft is wrong. It doesn't matter<mask> you are in a business suit and behind a company face, theft is theft, and<mask> a society we deem theft to be a bad thing. </s>
Label encoding: <s>The 'rape is rape' isn't to say both types are the same but more like 'theft is theft' when dealing with theft crimes that tend to be washed away or dismissed. Theft is wrong. It doesn't matter if you are in a business suit and behind a company face, theft is theft, and as a society we deem theft to be a bad thing. </s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What do you think you are doing,<mask> you let a HEAVILY corporately influenced government write regulations, other than giving control to those corporations? [NEWLINE] [NEWLINE] "A chance" to affect change through government?  A chance?  You opposition does not rely on chance, they directly affect change in government, to their favor. <mask> does a chance compare to the sure thing?</s>
Label encoding: <s>What do you think you are doing, when you let a HEAVILY corporately influenced government write regulations, other than giving control to those corporations? [NEWLINE] [NEWLINE] "A chance" to affect change through government?  A chance?  You opposition does not rely on chance, they directly affect change in government, to their favor.  How does a chance compare to the sure thing?</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If he said that,<mask> made good policy decisions right up to saying that on ethical principles, than no matter<mask> he said, getting elected would work. [NEWLINE] [NEWLINE] And a draft of everyone, maybe not.<mask> of everyone who is a government employee? That might be a better option, like<mask> military don't necessarily have a choice in deployments,<mask> don't have to serve.</s>
Label encoding: <s>If he said that, but made good policy decisions right up to saying that on ethical principles, than no matter what he said, getting elected would work. [NEWLINE] [NEWLINE] And a draft of everyone, maybe not. But of everyone who is a government employee? That might be a better option, like how military don't necessarily have a choice in deployments, but don't have to serve.</s>
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know r/aww isn't just for baby pictures,<mask> they do come up sometimes… I only picked that sub<mask> it's one of the more popular ones and I see a lot more baby-related hate there.<mask> your comparison to me going into /r/religion is a lot better than the one I made about /r/christianity haha.</s>
Label encoding: <s>I know r/aww isn't just for baby pictures, but they do come up sometimes… I only picked that sub because it's one of the more popular ones and I see a lot more baby-related hate there. But your comparison to me going into /r/religion is a lot better than the one I made about /r/christianity haha.</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your hatred should NEVER be aimed at a member of the armed forces. They aren't doing anything wrong.<mask> you want to complain about<mask> the army actually does, do it with your vote and your dollar.<mask> being angry at military people is absurd. You should try talking to one of them and ask<mask> it is they do. You'll be surprised at<mask> you hear.</s>
Label encoding: <s>Your hatred should NEVER be aimed at a member of the armed forces. They aren't doing anything wrong. If you want to complain about what the army actually does, do it with your vote and your dollar. But being angry at military people is absurd. You should try talking to one of them and ask what it is they do. You'll be surprised at what you hear.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] i know a lot of people that are great at science who would be *terrible* leaders. [ENDQ] [NEWLINE] I don't think leadership is a necessary component of government officials.  It's certainly something we select for at the moment,<mask><mask><mask> the purpose of politicians is to design and maintain a platform upon which a country can run.  Engineers, not celebrities. [NEWLINE] </s>
Label encoding: <s> [STARTQ] i know a lot of people that are great at science who would be *terrible* leaders. [ENDQ] [NEWLINE] I don't think leadership is a necessary component of government officials.  It's certainly something we select for at the moment, but I think the purpose of politicians is to design and maintain a platform upon which a country can run.  Engineers, not celebrities. [NEWLINE] </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0804, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean, I'm pretty sure that not causing harm to others always supersedes bodily autonomy.<mask> you could get an implant that made you exhale a deadly gas every time you breathed, it should be outlawed,<mask> you would kill people.<mask> we determine that a person starts existing at conception, getting an abortion *is* murder, and should supersede bodily autonomy.</s>
Label encoding: <s>I mean, I'm pretty sure that not causing harm to others always supersedes bodily autonomy. If you could get an implant that made you exhale a deadly gas every time you breathed, it should be outlawed, because you would kill people. If we determine that a person starts existing at conception, getting an abortion *is* murder, and should supersede bodily autonomy.</s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Eeek, that's... That's not<mask> it was intended to be consumed. [NEWLINE] [NEWLINE] I mean, to a certain extent, art (food included) is subjective,<mask><mask> you go look at the mona lisa through inverted Neon-Yellow tinted glasses, Leonardo Da Vinci's gonna be like "that's not<mask> you're meant to experience it!"</s>
Label encoding: <s>Eeek, that's... That's not how it was intended to be consumed. [NEWLINE] [NEWLINE] I mean, to a certain extent, art (food included) is subjective, but if you go look at the mona lisa through inverted Neon-Yellow tinted glasses, Leonardo Da Vinci's gonna be like "that's not how you're meant to experience it!"</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are other factors<mask> intelligence that could be selected. Physical strength, immune robustness, empathy and sensitivity, or simply to avoid the traits we identify<mask> bad. [NEWLINE] [NEWLINE] We can ponder this all we want,<mask> applying it universally will be meet with resistance, and<mask><mask><mask> there are any philosophical questions, any form of eugenics will be controversial at best.</s>
Label encoding: <s>There are other factors besides intelligence that could be selected. Physical strength, immune robustness, empathy and sensitivity, or simply to avoid the traits we identify as bad. [NEWLINE] [NEWLINE] We can ponder this all we want, but applying it universally will be meet with resistance, and as long as there are any philosophical questions, any form of eugenics will be controversial at best.</s>
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The tyranny of the majority.  <mask> young people were to become the majority at some point, a very strong majority, it is possible that majority could go and elect a person that is<mask> very young<mask> not for good reason<mask> for silly reasons.   Youth has a tendency to do this and it is something that could be very detrimental to the country.  </s>
Label encoding: <s>The tyranny of the majority.   If young people were to become the majority at some point, a very strong majority, it is possible that majority could go and elect a person that is also very young but not for good reason but for silly reasons.   Youth has a tendency to do this and it is something that could be very detrimental to the country.  </s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry LinguaManiac, your post has been removed: [NEWLINE] [NEWLINE] &gt; Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5)</s>
Label encoding: <s>Sorry LinguaManiac, your post has been removed: [NEWLINE] [NEWLINE] &gt; Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5)</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your analogy is actually very incorrect,<mask> the conclusion it would lead to is not that the science of medicine is impossible,<mask> that it isn't a scientific necessity to value medicine, which is true. The equivalent to "the science of medicine is possible" is "it is possible to be moral"; both are true,<mask> that doesn't explain<mask> we should value them.</s>
Label encoding: <s>Your analogy is actually very incorrect, because the conclusion it would lead to is not that the science of medicine is impossible, but that it isn't a scientific necessity to value medicine, which is true. The equivalent to "the science of medicine is possible" is "it is possible to be moral"; both are true, but that doesn't explain why we should value them.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/TurtleBeansforAll. [NEWLINE] [NEWLINE] [^TurtleBeansforAll's ^delta ^history](/r/ChangeMyView/wiki/user/turtlebeansforall) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/TurtleBeansforAll. [NEWLINE] [NEWLINE] [^TurtleBeansforAll's ^delta ^history](/r/ChangeMyView/wiki/user/turtlebeansforall) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In this in between phase<mask><mask> 24 fps might be best;<mask><mask> about the future? It's conceivable that graphics will be<mask> absolutely amazing in 50 years that there will be no way to distinguish a fake city from a real one (visually), and you will need 60 fps to get a full real life virtual reality experience. [NEWLINE] [NEWLINE] <mask> about then? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>In this in between phase I agree 24 fps might be best; but what about the future? It's conceivable that graphics will be so absolutely amazing in 50 years that there will be no way to distinguish a fake city from a real one (visually), and you will need 60 fps to get a full real life virtual reality experience. [NEWLINE] [NEWLINE] What about then? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You cannot decide to date a person and then tell them<mask> to change. [ENDQ] [NEWLINE] You can<mask> tell him that you dont like it, and ask him to stop. [NEWLINE] [NEWLINE] [STARTQ] <mask> he enjoys a thing that you hate then you should not have started this relationship. [ENDQ] [NEWLINE] Looking to the past is pointless, it doesnt help deal with current issues. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] You cannot decide to date a person and then tell them how to change. [ENDQ] [NEWLINE] You can however tell him that you dont like it, and ask him to stop. [NEWLINE] [NEWLINE] [STARTQ] If he enjoys a thing that you hate then you should not have started this relationship. [ENDQ] [NEWLINE] Looking to the past is pointless, it doesnt help deal with current issues. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It might,<mask> sometimes I've only got a quick moment and don't want to go re-research everything I've already learned.  Prior to calling someone else out, I always verify my information. <mask> I have trouble either finding or refuting someone's position, then I'd ask for help finding the source, typically without all caps bold face.  </s>
Label encoding: <s>It might, but sometimes I've only got a quick moment and don't want to go re-research everything I've already learned.  Prior to calling someone else out, I always verify my information.  If I have trouble either finding or refuting someone's position, then I'd ask for help finding the source, typically without all caps bold face.  </s>
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆. [NEWLINE] [NEWLINE] You win, man. [NEWLINE] [NEWLINE] This is an awesome metaphor. Strange,<mask> awesome.<mask><mask> it were me I'd go for a solid tungsten ring. Tungsten is pretty cool. I would pick ytterbium<mask> it tends to corrode easily,<mask><mask> its color is really unusual and nice. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>∆. [NEWLINE] [NEWLINE] You win, man. [NEWLINE] [NEWLINE] This is an awesome metaphor. Strange, but awesome. Though if it were me I'd go for a solid tungsten ring. Tungsten is pretty cool. I would pick ytterbium but it tends to corrode easily, even though its color is really unusual and nice. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] **and** (c) The order was not manifestly unlawful. [ENDQ] [NEWLINE] This absolutely does *not* make "just following orders" an "unofficial exception". It certainly raises the bar for prosecutions against "footsoldiers",<mask> it absolutely does not absolve them of their guilt<mask> they choose to participate in acts that are "manifestly unlawful".</s>
Label encoding: <s> [STARTQ] **and** (c) The order was not manifestly unlawful. [ENDQ] [NEWLINE] This absolutely does *not* make "just following orders" an "unofficial exception". It certainly raises the bar for prosecutions against "footsoldiers", but it absolutely does not absolve them of their guilt when they choose to participate in acts that are "manifestly unlawful".</s>
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Donating to charity might prevent a child from dying from cancer,<mask> I am not a child and I don't have cancer,<mask> that extra dollar from my pocket is better spent being used<mask> tax on my latest board game purchase. [ENDQ] [NEWLINE] <mask> you say "better", do you mean better for you, better for the world<mask> a whole,<mask>? </s>
Label encoding: <s> [STARTQ] Donating to charity might prevent a child from dying from cancer, but I am not a child and I don't have cancer, so that extra dollar from my pocket is better spent being used as tax on my latest board game purchase. [ENDQ] [NEWLINE] When you say "better", do you mean better for you, better for the world as a whole, what? </s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I, for one, would never buy a car that I couldn't drive myself. Don't get me wrong, I'd love to have one for rush hour<mask> all I do is sit on my ass and let the clutch in and out to shuffle forward once every five minutes.<mask> I'm on the back roads,<mask>, I want to drive myself. </s>
Label encoding: <s>I, for one, would never buy a car that I couldn't drive myself. Don't get me wrong, I'd love to have one for rush hour when all I do is sit on my ass and let the clutch in and out to shuffle forward once every five minutes. If I'm on the back roads, though, I want to drive myself. </s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Consider it this way. Immoral and illegal are different things and should remain different. [NEWLINE] [NEWLINE] People consider pre-marital sex immoral, should it be illegal? Some people consider extra-marital affair immoral, should it be illegal? [NEWLINE] [NEWLINE] You consider abortion immoral? Start it with yourself. Legislate yourself. Don't force it upon other's body.</s>
Label encoding: <s>Consider it this way. Immoral and illegal are different things and should remain different. [NEWLINE] [NEWLINE] People consider pre-marital sex immoral, should it be illegal? Some people consider extra-marital affair immoral, should it be illegal? [NEWLINE] [NEWLINE] You consider abortion immoral? Start it with yourself. Legislate yourself. Don't force it upon other's body.</s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you seriously using godandscience.com, a known propaganda site,<mask> a serious source? Come on now. [NEWLINE] [NEWLINE] Anyways,<mask> you asked: [NEWLINE] [NEWLINE] * [URL].11/marijuana.html [NEWLINE] * [URL] [NEWLINE] [NEWLINE] I'm not going to<mask><mask> marijuana is risk-free and 100% safe,<mask> the risks are very low.</s>
Label encoding: <s>Are you seriously using godandscience.com, a known propaganda site, as a serious source? Come on now. [NEWLINE] [NEWLINE] Anyways, since you asked: [NEWLINE] [NEWLINE] * [URL].11/marijuana.html [NEWLINE] * [URL] [NEWLINE] [NEWLINE] I'm not going to argue that marijuana is risk-free and 100% safe, but the risks are very low.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that some form of grammar Nazism is undesirable, like cussing over contracting conversation into convo. [NEWLINE] [NEWLINE] <mask> trying to remind people of the difference between then than they're or its/it's is important<mask> confusion on these words cause ambiguity. [NEWLINE] [NEWLINE] Evolution of language is inevitable,<mask> we should help it evolve in a pragmatic way.</s>
Label encoding: <s>I agree that some form of grammar Nazism is undesirable, like cussing over contracting conversation into convo. [NEWLINE] [NEWLINE] But trying to remind people of the difference between then than they're or its/it's is important because confusion on these words cause ambiguity. [NEWLINE] [NEWLINE] Evolution of language is inevitable, but we should help it evolve in a pragmatic way.</s>
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right.<mask><mask> we agree. Deficit spending on war does stimulate the economy,<mask><mask> increases the debt. Spending on infrastructure or social services probably stimulates the economy more than making things and blowing them up. Of course, part of war spending is wages for soldiers.<mask><mask> the wages get sent home, or stay in the country, this stimulates the economy.</s><pad>
Label encoding: <s>Right. I think we agree. Deficit spending on war does stimulate the economy, but also increases the debt. Spending on infrastructure or social services probably stimulates the economy more than making things and blowing them up. Of course, part of war spending is wages for soldiers. Assuming that the wages get sent home, or stay in the country, this stimulates the economy.</s><pad>
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/OSkorzeny. [NEWLINE] [NEWLINE] [^OSkorzeny's ^delta ^history](/r/ChangeMyView/wiki/user/oskorzeny) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/OSkorzeny. [NEWLINE] [NEWLINE] [^OSkorzeny's ^delta ^history](/r/ChangeMyView/wiki/user/oskorzeny) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0765, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well obviously the interesting part of this question, is<mask> the kid is NOT over 18. [NEWLINE] [NEWLINE] Whats you'r thoughts on that? [NEWLINE] [NEWLINE] EDIT: To clarify, parenthood can be a pretty fucking strong feeling.<mask> her telling him to fuck off might not be enough. He still might try to make a relationship with his child, regardless.</s>
Label encoding: <s>Well obviously the interesting part of this question, is when the kid is NOT over 18. [NEWLINE] [NEWLINE] Whats you'r thoughts on that? [NEWLINE] [NEWLINE] EDIT: To clarify, parenthood can be a pretty fucking strong feeling. So her telling him to fuck off might not be enough. He still might try to make a relationship with his child, regardless.</s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I just really don't think that secularists are being oppressed and I don't think it's doing anybody any harm [ENDQ] [NEWLINE] I wouldn't be<mask> sure of that. [Packages marked<mask> atheistic seem to do much worse in the US postal system than those without such markings.]( [URL] )  This, by a branch of the US government...</s><pad>
Label encoding: <s> [STARTQ] I just really don't think that secularists are being oppressed and I don't think it's doing anybody any harm [ENDQ] [NEWLINE] I wouldn't be so sure of that. [Packages marked as atheistic seem to do much worse in the US postal system than those without such markings.]( [URL] )  This, by a branch of the US government...</s><pad>
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Omg jumping on a grenade again.<mask> is it with you people and your grenades? CTRL+F "grenade" and see the numerous answers I've already given to this example. [NEWLINE] [NEWLINE] <mask><mask> : It's not suicide, it's dying to save other people, which I specifically stated in the OP has nothing to do with this thread.</s>
Label encoding: <s>Omg jumping on a grenade again. What is it with you people and your grenades? CTRL+F "grenade" and see the numerous answers I've already given to this example. [NEWLINE] [NEWLINE] TLDR : It's not suicide, it's dying to save other people, which I specifically stated in the OP has nothing to do with this thread.</s>
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Non-mobile: [URL] [NEWLINE] [NEWLINE] ^That's ^<mask> ^I'm ^here, ^I ^don't ^judge ^you. ^PM ^/u/xl0 ^<mask> ^I'm ^causing ^any ^trouble. [^WUT?]( [URL]'s-this-all-about%3F)</s>
Label encoding: <s>Non-mobile: [URL] [NEWLINE] [NEWLINE] ^That's ^ why ^I'm ^here, ^I ^don't ^judge ^you. ^PM ^/u/xl0 ^ if ^I'm ^causing ^any ^trouble. [^WUT?]( [URL]'s-this-all-about%3F)</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not blaming the immigrants,<mask><mask> I ought to clarify. I blame the Blair government which deliberately encouraged this<mask> Government policy. I don't hate immigrants, I just think its disappointing that they were eager to come to a country famous for<mask> was once a brilliant, proud culture and arrived to find that the country they had loved was dead. </s>
Label encoding: <s>I'm not blaming the immigrants, I think I ought to clarify. I blame the Blair government which deliberately encouraged this as Government policy. I don't hate immigrants, I just think its disappointing that they were eager to come to a country famous for what was once a brilliant, proud culture and arrived to find that the country they had loved was dead. </s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why should they accept that their tax money is used for things they disagree with or that the *right* to education should only be a right to an education<mask> inferior to be inadequate to the point of sometimes being even harmful? Especially<mask> they are mandatory for anyone not privileged with a lot of money and on top of that seemingly indoctrinate disagreeable beliefs.</s>
Label encoding: <s>Why should they accept that their tax money is used for things they disagree with or that the *right* to education should only be a right to an education so inferior to be inadequate to the point of sometimes being even harmful? Especially when they are mandatory for anyone not privileged with a lot of money and on top of that seemingly indoctrinate disagreeable beliefs.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Basic human rights to bodily autonomy and choice of<mask> to do with their own lives. We know they exist<mask> human rights are pretty much across the board accepted<mask> something that exists. [NEWLINE] [NEWLINE] <mask> you don't believe those rights exist, then you don't need to be having this conversation at all<mask> abortion would be the least of your worries. </s>
Label encoding: <s>Basic human rights to bodily autonomy and choice of what to do with their own lives. We know they exist because human rights are pretty much across the board accepted as something that exists. [NEWLINE] [NEWLINE] If you don't believe those rights exist, then you don't need to be having this conversation at all because abortion would be the least of your worries. </s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0598, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Has this literally happened more than once?<mask> I ever don't have a store card, it's<mask> I'm too lazy to fill out the paperwork, and no one has ever forced one on me. And you clearly tout putting up false information for misdirection in other threads here -<mask> not just put fake information into the store card form?</s>
Label encoding: <s>Has this literally happened more than once? If I ever don't have a store card, it's because I'm too lazy to fill out the paperwork, and no one has ever forced one on me. And you clearly tout putting up false information for misdirection in other threads here - why not just put fake information into the store card form?</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Someone transgender is going to be transgender<mask><mask> the environs they are raised in. They are born that way. They may grow up not knowing it's an option or that they can do it,<mask> they will still feel that way. [NEWLINE] [NEWLINE] Would the 'trans racial' girl feel black of she grew up in a white suburb? </s>
Label encoding: <s>Someone transgender is going to be transgender regardless of the environs they are raised in. They are born that way. They may grow up not knowing it's an option or that they can do it, but they will still feel that way. [NEWLINE] [NEWLINE] Would the 'trans racial' girl feel black of she grew up in a white suburb? </s>
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/thisdude415. [NEWLINE] [NEWLINE] [^thisdude415's ^delta ^history](/r/ChangeMyView/wiki/user/thisdude415) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/thisdude415. [NEWLINE] [NEWLINE] [^thisdude415's ^delta ^history](/r/ChangeMyView/wiki/user/thisdude415) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To create an accurate analogy between men's and women's nipples and sexuality, you need to consider<mask> many women consider men's nipples sexy, not<mask> many men are aroused by nipple stimulation. For all you know, all women objectify all men's nipples in a sexual manner,<mask> men can still take their shirts off whenever they want to.</s>
Label encoding: <s>To create an accurate analogy between men's and women's nipples and sexuality, you need to consider how many women consider men's nipples sexy, not how many men are aroused by nipple stimulation. For all you know, all women objectify all men's nipples in a sexual manner, but men can still take their shirts off whenever they want to.</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry for<mask> happened with your uncle. The laws are<mask> biased. The records of the students who lied must be tainted for the rest of their lives. They must pay for<mask> they did to this family. [NEWLINE] [NEWLINE] Juveniles? Yea,<mask> mark it on their records. False accusations should be a criminal offense. </s>
Label encoding: <s>I'm sorry for what happened with your uncle. The laws are so biased. The records of the students who lied must be tainted for the rest of their lives. They must pay for what they did to this family. [NEWLINE] [NEWLINE] Juveniles? Yea, but mark it on their records. False accusations should be a criminal offense. </s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know enough about the issue to comment on several of your points,<mask> for at least the bit about whether it encourages rape in real life, [statistics]( [URL] ) would make that seem unlikely.  I'd recommend reading that article to see<mask> I'm confident that the relationship between rape and internet access is not just correlation.</s><pad>
Label encoding: <s>I don't know enough about the issue to comment on several of your points, but for at least the bit about whether it encourages rape in real life, [statistics]( [URL] ) would make that seem unlikely.  I'd recommend reading that article to see why I'm confident that the relationship between rape and internet access is not just correlation.</s><pad>
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it's<mask>inine to try to make comparisons between racial groups to determine who has it worse. I don't think suffering and discrimination needs to be a dick measuring contest. I was just using some historical examples of large scale slights against the Asian community to demonstrate that they may not be<mask> favored as you argue in your OP.</s>
Label encoding: <s>I think it's asinine to try to make comparisons between racial groups to determine who has it worse. I don't think suffering and discrimination needs to be a dick measuring contest. I was just using some historical examples of large scale slights against the Asian community to demonstrate that they may not be as favored as you argue in your OP.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except that OP sounds like the kind of person who ignores all reasonable requests. [NEWLINE] [NEWLINE] "Please stop letting your cat into my yard, he tried to bite my kids and he leaves dead birds on my porch." [NEWLINE] [NEWLINE] "Aiiiieeee!! Easements! Right of way! MUST CAT-PROOF YOUR OWN YARD!"</s>
Label encoding: <s>Except that OP sounds like the kind of person who ignores all reasonable requests. [NEWLINE] [NEWLINE] "Please stop letting your cat into my yard, he tried to bite my kids and he leaves dead birds on my porch." [NEWLINE] [NEWLINE] "Aiiiieeee!! Easements! Right of way! MUST CAT-PROOF YOUR OWN YARD!"</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't<mask> I can't keep my balance (I fall backwards due to not being able to bend my ankles far enough),<mask> it's a thing Slavs (particularly Russians) are [known to do]( [URL].com/entries/icons/original/000/014/591/GS5mY6x.jpg).</s>
Label encoding: <s>I don't because I can't keep my balance (I fall backwards due to not being able to bend my ankles far enough), but it's a thing Slavs (particularly Russians) are [known to do]( [URL].com/entries/icons/original/000/014/591/GS5mY6x.jpg).</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] that being a policy that is enforced by most departments? [ENDQ] [NEWLINE] <mask> he's saying, is that people will be able to make FOIA requests and prove that nice looking white girls get let off with a warning 4 billion times<mask> often<mask> unkept black males. [NEWLINE] [NEWLINE] <mask><mask> putting a spotlight on this would be a positive thing.</s><pad>
Label encoding: <s> [STARTQ] that being a policy that is enforced by most departments? [ENDQ] [NEWLINE] What he's saying, is that people will be able to make FOIA requests and prove that nice looking white girls get let off with a warning 4 billion times as often as unkept black males. [NEWLINE] [NEWLINE] I think putting a spotlight on this would be a positive thing.</s><pad>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They would work the same<mask> they do now. People with a red light have to yield to the traffic and pedestrians going in the direction with a green light or "hand" [NEWLINE] [NEWLINE] I don't see<mask> this would change that. People are already allowed to make a right on red, and have to yield to pedestrians before doing<mask>.</s>
Label encoding: <s>They would work the same as they do now. People with a red light have to yield to the traffic and pedestrians going in the direction with a green light or "hand" [NEWLINE] [NEWLINE] I don't see how this would change that. People are already allowed to make a right on red, and have to yield to pedestrians before doing so.</s>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except application of learned material is the best way to learn.<mask> we teach in depth knowing that student a will forget a lot of the material, in the hopes that they will remember the concepts.<mask> we taught only concepts<mask> many things would students forget? You're kidding yourself<mask> you think high school students would purse knowledge on their own.</s><pad>
Label encoding: <s>Except application of learned material is the best way to learn. Also we teach in depth knowing that student a will forget a lot of the material, in the hopes that they will remember the concepts. If we taught only concepts how many things would students forget? You're kidding yourself if you think high school students would purse knowledge on their own.</s><pad>
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wow...<mask> do you live<mask> it only costs $100-300 dollars to upgrade to first class? I usually fly Delta in the US, and they almost never offer an upgrade for less than $500. I do get upgraded fairly often<mask> of my frequent flier status,<mask> I'd never pay for a first class upgrade. </s>
Label encoding: <s>Wow... where do you live where it only costs $100-300 dollars to upgrade to first class? I usually fly Delta in the US, and they almost never offer an upgrade for less than $500. I do get upgraded fairly often because of my frequent flier status, but I'd never pay for a first class upgrade. </s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First we went to London for two days, then Portsmouth, then Bath. [NEWLINE] [NEWLINE] Becides from just hangning out in the sparetime, we went to Stonehenge, London Eye, a Rivercrouse, the theater, and a few other things. We<mask> lived in home hospitality. [NEWLINE] [NEWLINE] Great trip overall. [NEWLINE] </s>
Label encoding: <s>First we went to London for two days, then Portsmouth, then Bath. [NEWLINE] [NEWLINE] Becides from just hangning out in the sparetime, we went to Stonehenge, London Eye, a Rivercrouse, the theater, and a few other things. We also lived in home hospitality. [NEWLINE] [NEWLINE] Great trip overall. [NEWLINE] </s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While using a pirated game might make you with to buy a sequel, you could just have well<mask> pirated the sequel. Liking it doesn't necessarily implicate you wish to pay the developers. You could just decide not to buy the sequel either. This attitude would simply lead to people delaying or possibly even denying recompense.</s>
Label encoding: <s>While using a pirated game might make you with to buy a sequel, you could just have well as pirated the sequel. Liking it doesn't necessarily implicate you wish to pay the developers. You could just decide not to buy the sequel either. This attitude would simply lead to people delaying or possibly even denying recompense.</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You may want to make the statement without fear of retribution. [ENDQ] [NEWLINE] The whole point of posting provocative pictures of Mohammed is to make the statement that the fear of retribution isn't going to take away your freedom of speech.<mask> you make that statement from a hiding place, it looks like the fear of retribution is working pretty well.</s>
Label encoding: <s> [STARTQ] You may want to make the statement without fear of retribution. [ENDQ] [NEWLINE] The whole point of posting provocative pictures of Mohammed is to make the statement that the fear of retribution isn't going to take away your freedom of speech. If you make that statement from a hiding place, it looks like the fear of retribution is working pretty well.</s>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In all honesty i used to make rap songs in high school with my friends. I<mask> produced and sold beats for several years. I don't know<mask> i can link to my personal songs,<mask> i can i will.<mask> not everybody is going to make it<mask> a rapper.<mask> it's<mask> saturated with lack of skill.</s>
Label encoding: <s>In all honesty i used to make rap songs in high school with my friends. I also produced and sold beats for several years. I don't know if i can link to my personal songs, if i can i will. But not everybody is going to make it as a rapper. Because it's so saturated with lack of skill.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a difference between selfish and individualistic. You<mask> can't judge a voting coalition by its ideological extreme.<mask> you're going to accuse the right of wanting to completely abandon those who can't look after themselves, you have to accuse the left of complete denial of property rights. You can't ignore nuance in only one direction.</s>
Label encoding: <s>There's a difference between selfish and individualistic. You also can't judge a voting coalition by its ideological extreme. If you're going to accuse the right of wanting to completely abandon those who can't look after themselves, you have to accuse the left of complete denial of property rights. You can't ignore nuance in only one direction.</s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>sure, human sacrifice is bad,<mask><mask> is the killing of 8 million natives by the belgians in the congo and the disfigurement and mutilation of countless more. [NEWLINE] [NEWLINE] go to /r/askhistorians or /r/badhistory, they've got plenty of resources on places other than europe.</s>
Label encoding: <s>sure, human sacrifice is bad, but so is the killing of 8 million natives by the belgians in the congo and the disfigurement and mutilation of countless more. [NEWLINE] [NEWLINE] go to /r/askhistorians or /r/badhistory, they've got plenty of resources on places other than europe.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] We could set different policies for birth control pills being sought for medical purposes and birth control pills being sought for solely contraceptive purposes. [ENDQ] [NEWLINE] On a practical level, I'm really not sure<mask> one would make that distinction. Once the birth control pills are paid for, the seller has no guarantee on<mask> they are used for.</s>
Label encoding: <s> [STARTQ] We could set different policies for birth control pills being sought for medical purposes and birth control pills being sought for solely contraceptive purposes. [ENDQ] [NEWLINE] On a practical level, I'm really not sure how one would make that distinction. Once the birth control pills are paid for, the seller has no guarantee on how they are used for.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I should have said that you should be prepared to kill<mask> you shoot. [ENDQ] [NEWLINE] This<mask><mask> with.  For example, most people(I don't know about everyone's training) aren't trained to "shoot to wound". [NEWLINE] [NEWLINE] <mask> the gun comes out, it's<mask> you intend to use lethal force.</s>
Label encoding: <s> [STARTQ] I should have said that you should be prepared to kill if you shoot. [ENDQ] [NEWLINE] This I agree with.  For example, most people(I don't know about everyone's training) aren't trained to "shoot to wound". [NEWLINE] [NEWLINE] If the gun comes out, it's because you intend to use lethal force.</s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok,<mask> would you go about conclusively proving that there isn't a cockroach anywhere in your house? [NEWLINE] [NEWLINE] I don't mean just making a plausible argument that the odds are better that there isn't one,<mask> being able to assert with 100% certainty that there can't be one hiding in some obscure corner. [NEWLINE] </s>
Label encoding: <s>Ok, how would you go about conclusively proving that there isn't a cockroach anywhere in your house? [NEWLINE] [NEWLINE] I don't mean just making a plausible argument that the odds are better that there isn't one, but being able to assert with 100% certainty that there can't be one hiding in some obscure corner. [NEWLINE] </s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] All factors being the same it boils down to who applies and<mask> skills and experience are needed for the job. [ENDQ] [NEWLINE] Yes. <mask> all factors were the same, that's<mask> it would come down to. <mask> women didn't carry the added cost of maternity, that's<mask> it would come down to.</s>
Label encoding: <s> [STARTQ] All factors being the same it boils down to who applies and what skills and experience are needed for the job. [ENDQ] [NEWLINE] Yes.  If all factors were the same, that's what it would come down to.  If women didn't carry the added cost of maternity, that's what it would come down to.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Similar to informal beliefs, every formal belief system ultimately rests on faith in un-provable assumptions/axioms. There is a famous proof of this called [ [URL] %C3%B6del's_incompleteness_theorems]Gödel's Incompleteness Theorem. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Similar to informal beliefs, every formal belief system ultimately rests on faith in un-provable assumptions/axioms. There is a famous proof of this called [ [URL] %C3%B6del's_incompleteness_theorems]Gödel's Incompleteness Theorem. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And<mask> the woman exercises either of her options on the child.  End scenario. [NEWLINE] [NEWLINE] Right now, you have a woman saying pay me or I will make you pay for 18 years, in the scenario<mask> a man has the ability to sign off rights, the woman can sign off too and it's a wash.</s><pad>
Label encoding: <s>And so the woman exercises either of her options on the child.  End scenario. [NEWLINE] [NEWLINE] Right now, you have a woman saying pay me or I will make you pay for 18 years, in the scenario where a man has the ability to sign off rights, the woman can sign off too and it's a wash.</s><pad>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Person 2, failed: I'm pro refugee immigration<mask> the refugees need help. [ENDQ] [NEWLINE] [STARTQ] Person 4, failed: I'm against refugee immigration<mask> the immigrants are bad for society. [ENDQ] [NEWLINE] Both of these people express noncontradictory and valid reasons for their viewpoints. <mask> would you fail them?</s>
Label encoding: <s> [STARTQ] Person 2, failed: I'm pro refugee immigration because the refugees need help. [ENDQ] [NEWLINE] [STARTQ] Person 4, failed: I'm against refugee immigration because the immigrants are bad for society. [ENDQ] [NEWLINE] Both of these people express noncontradictory and valid reasons for their viewpoints.  Why would you fail them?</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In an ideal world I would be,<mask> I understand it creates a loophole for arseholes to exploit ("it's not a payment it's a gift!" and such)<mask> I'm not against taxing it. I believe inheritance is different<mask><mask> a dead person probably doesn't give many shits about avoiding taxes.</s>
Label encoding: <s>In an ideal world I would be, but I understand it creates a loophole for arseholes to exploit ("it's not a payment it's a gift!" and such) so I'm not against taxing it. I believe inheritance is different though since a dead person probably doesn't give many shits about avoiding taxes.</s>
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>yup, and even<mask> its a violent intruder,<mask> he isn't carrying a gun you can just keep him at bay until the cops arrive rather then put one between his eyes,<mask> lets face it even burglars are people,<mask> you don't have to shoot to kill you should at least attempt it.</s>
Label encoding: <s>yup, and even if its a violent intruder, if he isn't carrying a gun you can just keep him at bay until the cops arrive rather then put one between his eyes, because lets face it even burglars are people, if you don't have to shoot to kill you should at least attempt it.</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've already conceded that in the current political climate a voter ID law would cause more harm than good. I still believe that steps should be taken to change this, such<mask> a "right to identity"<mask> citizens are given unrestricted access to documents proving they are citizens and that these documents will suffice for voter ID.</s>
Label encoding: <s>I've already conceded that in the current political climate a voter ID law would cause more harm than good. I still believe that steps should be taken to change this, such as a "right to identity" where citizens are given unrestricted access to documents proving they are citizens and that these documents will suffice for voter ID.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They aren't necessarily lying. Psychology effects<mask> you perceive a lot of things.<mask> you think that wine itself should have classy natures then<mask> you think of it will effect your perception of it. They might even like the novelty of it more than the direct taste obviously.<mask> that is not a "lie."</s>
Label encoding: <s>They aren't necessarily lying. Psychology effects how you perceive a lot of things. If you think that wine itself should have classy natures then what you think of it will effect your perception of it. They might even like the novelty of it more than the direct taste obviously. But that is not a "lie."</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Glory2Hypnotoad. ^[[History](/r/changemyview/wiki/user/Glory2Hypnotoad)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Glory2Hypnotoad. ^[[History](/r/changemyview/wiki/user/Glory2Hypnotoad)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, it's a bit unfortunate for everyone involved. [NEWLINE] [NEWLINE] Society will not allow them to satisfy their urges, and for good reason! [NEWLINE] [NEWLINE] <mask>, they're still human beings,<mask> we can't just leave them behind. [NEWLINE] [NEWLINE] I truly wonder<mask> future medical science will tackle this issue.</s><pad>
Label encoding: <s>Yeah, it's a bit unfortunate for everyone involved. [NEWLINE] [NEWLINE] Society will not allow them to satisfy their urges, and for good reason! [NEWLINE] [NEWLINE] However, they're still human beings, so we can't just leave them behind. [NEWLINE] [NEWLINE] I truly wonder how future medical science will tackle this issue.</s><pad>
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/whattodo-whattodo. ^[[History](/r/changemyview/wiki/user/whattodo-whattodo)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/whattodo-whattodo. ^[[History](/r/changemyview/wiki/user/whattodo-whattodo)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Has no stranger ever opened a door for you? [NEWLINE] Has no-one ever let you cross the road<mask> they weren't legally required to? [NEWLINE] Has no-one ever pointed out that you'd dropped something? [NEWLINE] Has no-one ever stepped out of your way<mask> you were walking down the street?</s>
Label encoding: <s>Has no stranger ever opened a door for you? [NEWLINE] Has no-one ever let you cross the road when they weren't legally required to? [NEWLINE] Has no-one ever pointed out that you'd dropped something? [NEWLINE] Has no-one ever stepped out of your way while you were walking down the street?</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with that definition is that,<mask> you say, it places The Lord of the Rings<mask> middle fantasy,<mask> most people would agree it is most definitely high fantasy. [NEWLINE] [NEWLINE] Personally I thought the existence of elves and dwarves was one major component of high fantasy, along with magic of some sort.</s>
Label encoding: <s>The problem with that definition is that, as you say, it places The Lord of the Rings as middle fantasy, while most people would agree it is most definitely high fantasy. [NEWLINE] [NEWLINE] Personally I thought the existence of elves and dwarves was one major component of high fantasy, along with magic of some sort.</s>
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But you are choosing to play a sport every time you go play. It's within your control every day. Heroin is completely out of your control, addicts will hurt family and friends to maintain said addiction.<mask> sports usually do the opposite, by bringing family and friends together for a fun bonding experience.</s>
Label encoding: <s>But you are choosing to play a sport every time you go play. It's within your control every day. Heroin is completely out of your control, addicts will hurt family and friends to maintain said addiction. While sports usually do the opposite, by bringing family and friends together for a fun bonding experience.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was saying that no owner should allow their dog to get close enough TO lick you,<mask><mask> we want to go down this path: [NEWLINE] [NEWLINE] It would normally be bad behavior for me to put my hands on you.<mask><mask> you're backing into me, am I not justified in doing<mask>?</s>
Label encoding: <s>I was saying that no owner should allow their dog to get close enough TO lick you, but if we want to go down this path: [NEWLINE] [NEWLINE] It would normally be bad behavior for me to put my hands on you. But if you're backing into me, am I not justified in doing so?</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>just like raw milk is contributory, not causative for typhoid... am I wrong here?  Raw milk does not *cause* disease, it is simply more likely to carry it.  Cigarettes do not *cause* respiratory infections,<mask> smokers are more likely to carry them.  </s>
Label encoding: <s>just like raw milk is contributory, not causative for typhoid... am I wrong here?  Raw milk does not *cause* disease, it is simply more likely to carry it.  Cigarettes do not *cause* respiratory infections, but smokers are more likely to carry them.  </s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point you are missing is that machine intelligence at human levels is Moores law scalable in a thought seconds per hour sense,  and<mask> it will be running on a well understood platform,  *recursively self improving*.  This will give benefits beyond current  Moores law scales.  </s>
Label encoding: <s>The point you are missing is that machine intelligence at human levels is Moores law scalable in a thought seconds per hour sense,  and because it will be running on a well understood platform,  *recursively self improving*.  This will give benefits beyond current  Moores law scales.  </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> somehow you think that a supernatural, spiritual entity like God is still possible.<mask><mask> this is an unexamined flaw in your thinking. [ENDQ] [NEWLINE] [NEWLINE] I'll point you to my edits in the OP: I don't believe god is supernatural, merely beyond our understanding of nature. [NEWLINE] </s>
Label encoding: <s> [STARTQ] yet somehow you think that a supernatural, spiritual entity like God is still possible. I think this is an unexamined flaw in your thinking. [ENDQ] [NEWLINE] [NEWLINE] I'll point you to my edits in the OP: I don't believe god is supernatural, merely beyond our understanding of nature. [NEWLINE] </s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm saying that you're expecting the benefits of society (a sense of justice) without wanting to be part of society. [ENDQ] [NEWLINE] No I'm not.  I'm saying that in order for the social contract to be considered just (within the system) it must offer an alternative. </s>
Label encoding: <s> [STARTQ] I'm saying that you're expecting the benefits of society (a sense of justice) without wanting to be part of society. [ENDQ] [NEWLINE] No I'm not.  I'm saying that in order for the social contract to be considered just (within the system) it must offer an alternative. </s>
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are aspects of family relationships that make these unions far worse. It is well documented in psychology that child victims of incestual bonds often fare worse that child victims of underage rape or molestation by non family members. Just basing this on research and science, not assumptions and wishful thinking. </s>
Label encoding: <s>There are aspects of family relationships that make these unions far worse. It is well documented in psychology that child victims of incestual bonds often fare worse that child victims of underage rape or molestation by non family members. Just basing this on research and science, not assumptions and wishful thinking. </s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This award is currently disallowed<mask> your comment doesn't include enough text ([comment rule 4]( [URL] #wiki_rule_4)). Please add an explanation for<mask> /u/cr0kus changed your view. Responding to this comment will cause me to recheck your delta comment.</s>
Label encoding: <s>This award is currently disallowed as your comment doesn't include enough text ([comment rule 4]( [URL] #wiki_rule_4)). Please add an explanation for how /u/cr0kus changed your view. Responding to this comment will cause me to recheck your delta comment.</s>
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Stanislawiii. ^[[History](/r/changemyview/wiki/user/Stanislawiii)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Stanislawiii. ^[[History](/r/changemyview/wiki/user/Stanislawiii)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/maxpenny42. ^[[History](/r/changemyview/wiki/user/maxpenny42)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/maxpenny42. ^[[History](/r/changemyview/wiki/user/maxpenny42)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] MRAs do not advocate hating, stalking or beating women at all. [ENDQ] [NEWLINE] MRAs advocate for equal treatment of men and women in regards to violence.  That some read that and take it for advocating an increase in violence against women is more telling about the reader than anything else.</s>
Label encoding: <s> [STARTQ] MRAs do not advocate hating, stalking or beating women at all. [ENDQ] [NEWLINE] MRAs advocate for equal treatment of men and women in regards to violence.  That some read that and take it for advocating an increase in violence against women is more telling about the reader than anything else.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/PopeJohnPaulII. ^[[History](/r/changemyview/wiki/user/PopeJohnPaulII)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/PopeJohnPaulII. ^[[History](/r/changemyview/wiki/user/PopeJohnPaulII)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Water is an economic good, with scarcity - it acts like all other economic goods. It is all well and good to say that it should be a universal right,<mask><mask><mask> it costs money to have enough water for everyone (purifying, etc) someone has to pay for that.</s>
Label encoding: <s>Water is an economic good, with scarcity - it acts like all other economic goods. It is all well and good to say that it should be a universal right, but given that it costs money to have enough water for everyone (purifying, etc) someone has to pay for that.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You get buzzed, I would describe the effects somewhere in between downing a bunch of coffee<mask> simultaneously having a drunkeness aspect to it. I remember the first few time smoking a cigarette I was driving and I thought I was gonna run off the road I was<mask> dizzy. </s>
Label encoding: <s>You get buzzed, I would describe the effects somewhere in between downing a bunch of coffee while simultaneously having a drunkeness aspect to it. I remember the first few time smoking a cigarette I was driving and I thought I was gonna run off the road I was so dizzy. </s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/AnalogKid2112. ^[[History](/r/changemyview/wiki/AnalogKid2112)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/AnalogKid2112. ^[[History](/r/changemyview/wiki/AnalogKid2112)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I asked OP to clarify his viewpoint for me. You might be entirely correct about his interpretation,<mask> I'm waiting for him to respond. OP had not stated that eating disorders were exempt until after I had commented and shared my perspective,<mask> I would like him to clarify his views. </s>
Label encoding: <s>I asked OP to clarify his viewpoint for me. You might be entirely correct about his interpretation, but I'm waiting for him to respond. OP had not stated that eating disorders were exempt until after I had commented and shared my perspective, therefore I would like him to clarify his views. </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/ThereOnceWasAMan. ^[[History](/r/changemyview/wiki/user/ThereOnceWasAMan)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/ThereOnceWasAMan. ^[[History](/r/changemyview/wiki/user/ThereOnceWasAMan)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why are you making a biological explanation for something that is really a sociological one. [NEWLINE] [NEWLINE] It isn't just a one person getting praise and one person not. It is one person getting praise and the other one get public shame. [NEWLINE] [NEWLINE] time, rime, thyme. </s>
Label encoding: <s>Why are you making a biological explanation for something that is really a sociological one. [NEWLINE] [NEWLINE] It isn't just a one person getting praise and one person not. It is one person getting praise and the other one get public shame. [NEWLINE] [NEWLINE] time, rime, thyme. </s>
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Personally, my objection to the death penalty and cruel executions in particular is pretty simple:  We're supposed to be better than they are. <mask><mask> it's exceptionally hypocritical to condemn somebody for violent crimes<mask> you operate a facility built for the purpose of torturing people to death.</s>
Label encoding: <s>Personally, my objection to the death penalty and cruel executions in particular is pretty simple:  We're supposed to be better than they are.  I think it's exceptionally hypocritical to condemn somebody for violent crimes while you operate a facility built for the purpose of torturing people to death.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I never said you shouldbe held accountable for the actions of your ancestors, that's a backward way of thinking. The problem with "privilege" is that many people think of it that way, that's<mask><mask><mask> using terms like luck of the draw would be more effective.</s>
Label encoding: <s>I never said you shouldbe held accountable for the actions of your ancestors, that's a backward way of thinking. The problem with "privilege" is that many people think of it that way, that's why I think using terms like luck of the draw would be more effective.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am asserting that it was a war primarily for profit.  wealthy individuals saw an opportunity, took advantage of our hurt and fervor, and made bank.  to us, the american people, it is about terrorism,<mask> to the people making decisions it is about money.</s>
Label encoding: <s>I am asserting that it was a war primarily for profit.  wealthy individuals saw an opportunity, took advantage of our hurt and fervor, and made bank.  to us, the american people, it is about terrorism, but to the people making decisions it is about money.</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> you're arguing for here,<mask> SWATTING is a form of proxy violence. Even<mask> the perpetrator doesn't intend for their victims to be killed, it's still a strong possibility. It's waving a bunch of guns in someone's face remotely. </s>
Label encoding: <s>I'm not sure what you're arguing for here, but SWATTING is a form of proxy violence. Even if the perpetrator doesn't intend for their victims to be killed, it's still a strong possibility. It's waving a bunch of guns in someone's face remotely. </s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Nepene. ^[[History](/r/changemyview/wiki/user/Nepene)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Nepene. ^[[History](/r/changemyview/wiki/user/Nepene)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Pretty sure it was Newton that wrote Principia Mathematica, no? [NEWLINE] [NEWLINE] **EDIT:** I was kind of right. Newton wrote a different book with a similar title - [URL] %C3%A6_Naturalis_Principia_Mathematica</s>
Label encoding: <s>Pretty sure it was Newton that wrote Principia Mathematica, no? [NEWLINE] [NEWLINE] **EDIT:** I was kind of right. Newton wrote a different book with a similar title - [URL] %C3%A6_Naturalis_Principia_Mathematica</s>
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0691, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/aguafiestas. ^[[History](/r/changemyview/wiki/user/aguafiestas)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/aguafiestas. ^[[History](/r/changemyview/wiki/user/aguafiestas)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you don't see the harm in a world<mask> everyone is almost entirely isolated<mask> it is easier and faster to interact with machines than other people? [ENDQ] [NEWLINE] You'd have to go back 100 years to find the last technological age requiring human interaction to get a weather report. </s>
Label encoding: <s> [STARTQ] you don't see the harm in a world where everyone is almost entirely isolated because it is easier and faster to interact with machines than other people? [ENDQ] [NEWLINE] You'd have to go back 100 years to find the last technological age requiring human interaction to get a weather report. </s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your comment makes me really angry.  Every time these issues are brought up, there is always someone ready to chime in and blame the complainers. [NEWLINE] [NEWLINE] "Just go speak to him in his office hours and learn his accent!" [NEWLINE] [NEWLINE] "Just study harder!"</s>
Label encoding: <s>Your comment makes me really angry.  Every time these issues are brought up, there is always someone ready to chime in and blame the complainers. [NEWLINE] [NEWLINE] "Just go speak to him in his office hours and learn his accent!" [NEWLINE] [NEWLINE] "Just study harder!"</s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to cut my food into small pieces normally. [NEWLINE] [NEWLINE] [NEWLINE] I can rip my food before putting it my mouth [NEWLINE] [NEWLINE] [NEWLINE] I already do this. [NEWLINE] [NEWLINE] [NEWLINE] Debatable. [NEWLINE] [NEWLINE] [NEWLINE] This is cleaner than leaving them in your bacteria ridden mouth. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I have to cut my food into small pieces normally. [NEWLINE] [NEWLINE] [NEWLINE] I can rip my food before putting it my mouth [NEWLINE] [NEWLINE] [NEWLINE] I already do this. [NEWLINE] [NEWLINE] [NEWLINE] Debatable. [NEWLINE] [NEWLINE] [NEWLINE] This is cleaner than leaving them in your bacteria ridden mouth. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Addressing the tone of someone's argument rather than the argument itself is a quick way to admit that you don't have anything more substantial to say... [ENDQ] [NEWLINE] This is rather childish to write in response to a single point in a long reply filled with other substantial points.</s>
Label encoding: <s> [STARTQ] Addressing the tone of someone's argument rather than the argument itself is a quick way to admit that you don't have anything more substantial to say... [ENDQ] [NEWLINE] This is rather childish to write in response to a single point in a long reply filled with other substantial points.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the problem is that there are people that are asshole to women engineers, and they are the ones that the women remember.  Are the majority of male engineers nice and have no problem with female engineers?  Probably.  Are there asshole that do?  Yes.</s>
Label encoding: <s>I think the problem is that there are people that are asshole to women engineers, and they are the ones that the women remember.  Are the majority of male engineers nice and have no problem with female engineers?  Probably.  Are there asshole that do?  Yes.</s>
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not the intended use of the product,<mask> that is<mask> happens<mask> people do use the product. [NEWLINE] [NEWLINE] you could say that it wasn't the intended intent of Jarts to piece the skull<mask> thrown,<mask> that is<mask> happened<mask> people used them. </s>
Label encoding: <s>That's not the intended use of the product, but that is what happens when people do use the product. [NEWLINE] [NEWLINE] you could say that it wasn't the intended intent of Jarts to piece the skull when thrown, but that is what happened when people used them. </s>
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But the Supreme Court can rule against their actions and the president is liable to be impeached. The Court's decisions essentially become law, something that even the president can't do. The closest they can do is issue a doctrine or bypass congressional approval with the United Nations. </s><pad>
Label encoding: <s>But the Supreme Court can rule against their actions and the president is liable to be impeached. The Court's decisions essentially become law, something that even the president can't do. The closest they can do is issue a doctrine or bypass congressional approval with the United Nations. </s><pad>
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The ass towel is near the bidet. The face/hands towel is near the sink. It's kind of hard to mess it up. And the bidet towels (at least the ones I've seen) are usually much smaller<mask> you can recognize them right away.</s>
Label encoding: <s>The ass towel is near the bidet. The face/hands towel is near the sink. It's kind of hard to mess it up. And the bidet towels (at least the ones I've seen) are usually much smaller so you can recognize them right away.</s>
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0676, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/klw. ^[[History](/r/changemyview/wiki/user/klw)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/klw. ^[[History](/r/changemyview/wiki/user/klw)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of suicidal people feel that way, and it's<mask> of<mask> easy it is to idealize. No more work, no more needs, no more worrying, no more sadness, no more pain. No more anything. To them it seems like true peace.</s>
Label encoding: <s>A lot of suicidal people feel that way, and it's because of how easy it is to idealize. No more work, no more needs, no more worrying, no more sadness, no more pain. No more anything. To them it seems like true peace.</s>
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But<mask> you recognize that the fetus has rights, no court would let a women kill a dependent. That's<mask> it is a fetus rights issue instead of a women's rights issue. The whole question is whether or not a fetus had the right not to be killed.</s>
Label encoding: <s>But if you recognize that the fetus has rights, no court would let a women kill a dependent. That's why it is a fetus rights issue instead of a women's rights issue. The whole question is whether or not a fetus had the right not to be killed.</s>
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Honestly, I really liked the 9th doctor<mask> well, it was more that i didn't like his companion, and he only ever got one. And by comparison I really liked Martha and Donna,<mask> the 10th was generally a better experience for me. </s>
Label encoding: <s>Honestly, I really liked the 9th doctor as well, it was more that i didn't like his companion, and he only ever got one. And by comparison I really liked Martha and Donna, so the 10th was generally a better experience for me. </s>
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It works both ways. Mental illness can cause low income - for example,<mask> people are unable to function in a job. The stresses of being poor can cause mental illness - for example, [poor mothers  become more likely to develop an anxiety disorder]( [URL] ).</s>
Label encoding: <s>It works both ways. Mental illness can cause low income - for example, when people are unable to function in a job. The stresses of being poor can cause mental illness - for example, [poor mothers  become more likely to develop an anxiety disorder]( [URL] ).</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So change the rule. Until then, you're just making up your own way of doing it. The whole point of language is for it to be a common form of communication--not for every person to have their own personal spelling/grammar/syntax.</s><pad>
Label encoding: <s>So change the rule. Until then, you're just making up your own way of doing it. The whole point of language is for it to be a common form of communication--not for every person to have their own personal spelling/grammar/syntax.</s><pad>
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People really enjoy watching sports,<mask> much<mask> that they are willing to pay money for it. [NEWLINE] [NEWLINE] <mask> I pay $100 towards watching my home team, and a million other people<mask> pay the same amount,<mask> should that $100 million of revenue go?</s>
Label encoding: <s>People really enjoy watching sports, so much so that they are willing to pay money for it. [NEWLINE] [NEWLINE] If I pay $100 towards watching my home team, and a million other people also pay the same amount, where should that $100 million of revenue go?</s>
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While<mask><mask> that no one has used it in that context before (that I've seen). I don't think it falls under straw man only<mask> you used the word "impossible". They are simply demonstrating that it IS possible by giving an example. </s>
Label encoding: <s>While I agree that no one has used it in that context before (that I've seen). I don't think it falls under straw man only because you used the word "impossible". They are simply demonstrating that it IS possible by giving an example. </s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are a lot less likely to be alive than I am.<mask> it causes disease and bad health along the way. A cigarette being smoked can cause a large area to stink. Smokers don't realize it themselves put the smell is really terrible and gross.</s>
Label encoding: <s>You are a lot less likely to be alive than I am. Also it causes disease and bad health along the way. A cigarette being smoked can cause a large area to stink. Smokers don't realize it themselves put the smell is really terrible and gross.</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The vast majority. Enough that the persecution complex has cemented itself in Jewish culture. Of course arguing whether or not something is part of a culture is difficult,<mask> my argument isn't just that it's there - it's that it's irrational<mask> well.</s>
Label encoding: <s>The vast majority. Enough that the persecution complex has cemented itself in Jewish culture. Of course arguing whether or not something is part of a culture is difficult, although my argument isn't just that it's there - it's that it's irrational as well.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0717, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's hard to change your mind about this<mask> stated premise includes the word "generally."  Are you claiming an absolute or not?  Sometimes fade-outs sound great, everybody knows that.  Your premise is<mask> un-falsifiable.</s>
Label encoding: <s>It's hard to change your mind about this if stated premise includes the word "generally."  Are you claiming an absolute or not?  Sometimes fade-outs sound great, everybody knows that.  Your premise is therefore un-falsifiable.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The link in my post, [URL] [NEWLINE] [NEWLINE] Welfare consists of the following: health, education, housing, energy, agriculture, transportation, labor, social security, and small business administration. [NEWLINE] [NEWLINE] All of these represent welfare payments given to individuals. [NEWLINE] </s>
Label encoding: <s>The link in my post, [URL] [NEWLINE] [NEWLINE] Welfare consists of the following: health, education, housing, energy, agriculture, transportation, labor, social security, and small business administration. [NEWLINE] [NEWLINE] All of these represent welfare payments given to individuals. [NEWLINE] </s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>At the very least, coverage tends to begin at the beginning of the moth and require application at least a week or<mask> before that,<mask> you'd need to sit around for minimum one week with a broken arm, or you know, car crash injuries.</s>
Label encoding: <s>At the very least, coverage tends to begin at the beginning of the moth and require application at least a week or so before that, so you'd need to sit around for minimum one week with a broken arm, or you know, car crash injuries.</s>
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0751, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Both of you have plausible assertions. Maybe women's chess leagues draw more girls to the game; maybe they hold them back. At this point the speculation has done its job and can cease. We need an actual study to find out<mask> happens. </s>
Label encoding: <s>Both of you have plausible assertions. Maybe women's chess leagues draw more girls to the game; maybe they hold them back. At this point the speculation has done its job and can cease. We need an actual study to find out what happens. </s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would just rather less people have to go<mask> the stress of having an illness and needing to go to the hospital. Yes people will be able to survive in this environment<mask> it means more resources used, more people in the hospital, etc.. [NEWLINE] </s>
Label encoding: <s>I would just rather less people have to go though the stress of having an illness and needing to go to the hospital. Yes people will be able to survive in this environment but it means more resources used, more people in the hospital, etc.. [NEWLINE] </s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The post didn't just say CMV: I don't believe in God anymore, it<mask> said that he/she wants to. <mask><mask> it's completely reasonable to change that part of his/her view<mask> it is the first. </s><pad>
Label encoding: <s>The post didn't just say CMV: I don't believe in God anymore, it also said that he/she wants to.  I think it's completely reasonable to change that part of his/her view as it is the first. </s><pad>
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is basic existentialism.<mask> the best way to have a solid view on this is to read up on that. In anycase,<mask> life is pointless, then it could mean whatever you want it to mean. Isn't that great?</s>
Label encoding: <s>This is basic existentialism. So the best way to have a solid view on this is to read up on that. In anycase, if life is pointless, then it could mean whatever you want it to mean. Isn't that great?</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Balance is the underlying principle.<mask> the Jedi interpeted balance was that they were the ones who balanced good and evil, and that the Sith were chaos.<mask><mask> it turned out was that they were each one side of the coin. </s>
Label encoding: <s>Balance is the underlying principle. How the Jedi interpeted balance was that they were the ones who balanced good and evil, and that the Sith were chaos. However how it turned out was that they were each one side of the coin. </s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Alright, convince me. My argument is that<mask> our society has evolved to such a degree<mask> we can rationalize not passing our genes on<mask> we really want to.<mask> someone holds this opinion, they can<mask> justify not being in relationships.</s>
Label encoding: <s>Alright, convince me. My argument is that since our society has evolved to such a degree where we can rationalize not passing our genes on if we really want to. If someone holds this opinion, they can also justify not being in relationships.</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hence the "almost all" part. [NEWLINE] [NEWLINE] And 70k in savings is enough to live just fine for a few months, anyway. He should be feeling the pain of<mask> he is doing to our country, just like everyone else.</s>
Label encoding: <s>Hence the "almost all" part. [NEWLINE] [NEWLINE] And 70k in savings is enough to live just fine for a few months, anyway. He should be feeling the pain of what he is doing to our country, just like everyone else.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0737, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most women don't think of strippers and dominatrixes<mask> a "sexy and powerful" stereotype,<mask> they represent situations<mask> a woman's sexuality exists primarily for the purpose of male fantasy and male sexual gratification.   </s>
Label encoding: <s>Most women don't think of strippers and dominatrixes as a "sexy and powerful" stereotype, because they represent situations where a woman's sexuality exists primarily for the purpose of male fantasy and male sexual gratification.   </s>
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Is losing customers worth the fraction of the time it takes to ring in 17 vs 15 items in terms of monetary loss? [NEWLINE] [NEWLINE] [NEWLINE] Absolutely not. This is a terrible idea an angry cashier thinks up, not a viable business tactic.</s>
Label encoding: <s>Is losing customers worth the fraction of the time it takes to ring in 17 vs 15 items in terms of monetary loss? [NEWLINE] [NEWLINE] [NEWLINE] Absolutely not. This is a terrible idea an angry cashier thinks up, not a viable business tactic.</s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This might be worthy of a whole other CMV<mask> I don't think increased access to scientific materials necessarily causes increased production, especially in science<mask> there is considerable cost (in resources, time, training and expertise) required to publish. </s>
Label encoding: <s>This might be worthy of a whole other CMV but I don't think increased access to scientific materials necessarily causes increased production, especially in science where there is considerable cost (in resources, time, training and expertise) required to publish. </s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for the links. I'll take a softer tone with that one next time. [NEWLINE] [NEWLINE] The scan/upload concept is neat,<mask> it could only ever be a copy of the subject,<mask> opposed to their actual consciousness. </s>
Label encoding: <s>Thank you for the links. I'll take a softer tone with that one next time. [NEWLINE] [NEWLINE] The scan/upload concept is neat, but it could only ever be a copy of the subject, as opposed to their actual consciousness. </s>
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh, I like this answer, the impracticality and aims it could be exploited towards were fresh,<mask> I still think its vastly better than the current system. You raise great points. ∆ First delta use, hope its well placed.</s>
Label encoding: <s>Oh, I like this answer, the impracticality and aims it could be exploited towards were fresh, but I still think its vastly better than the current system. You raise great points. ∆ First delta use, hope its well placed.</s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But<mask> about the EXTREME cases? [NEWLINE] [NEWLINE] <mask><mask> someone found a an exploit that takes you straight from the beginning screen to the game-over you won screen. [NEWLINE] [NEWLINE] Would that still be a speed run? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>But what about the EXTREME cases? [NEWLINE] [NEWLINE] What if someone found a an exploit that takes you straight from the beginning screen to the game-over you won screen. [NEWLINE] [NEWLINE] Would that still be a speed run? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For me, personally? Yes. I most certainly would not have bought it, unless I would have bought it a number of years from now,<mask> I don't currently possess the financial stability to even consider buying high-level software.</s>
Label encoding: <s>For me, personally? Yes. I most certainly would not have bought it, unless I would have bought it a number of years from now, because I don't currently possess the financial stability to even consider buying high-level software.</s>
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We both agree that we do need an age for<mask> we stop punishing people for statutory rape.<mask> let me ask then,<mask> do you think is an appropriate age?<mask> we both agree, I just believe the answer is eighteen.</s>
Label encoding: <s>We both agree that we do need an age for when we stop punishing people for statutory rape. So let me ask then, what do you think is an appropriate age? Because we both agree, I just believe the answer is eighteen.</s>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ Very well put.  I admit I had not considered that people would not necessarily differentiate between extremists and peaceful Muslims.  I find it disappointing to think that there are a great deal of people who cannot distinguish between the two.</s>
Label encoding: <s>∆ Very well put.  I admit I had not considered that people would not necessarily differentiate between extremists and peaceful Muslims.  I find it disappointing to think that there are a great deal of people who cannot distinguish between the two.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I am completely unswayed and unmoved by TV ads [ENDQ] [NEWLINE] <mask> do you know? You'll come up with some explanation that doesn't include ads<mask> you don't want to admit you are controlled by your subconscious.</s>
Label encoding: <s> [STARTQ] I am completely unswayed and unmoved by TV ads [ENDQ] [NEWLINE] How do you know? You'll come up with some explanation that doesn't include ads because you don't want to admit you are controlled by your subconscious.</s>
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that person should file a lawsuit on the basis of discrimination<mask><mask><mask> there was no specific stipulation about speaking on the manner in their contract. [NEWLINE] [NEWLINE] <mask> once again, I fail to see the relevance here. </s>
Label encoding: <s>I think that person should file a lawsuit on the basis of discrimination so long as there was no specific stipulation about speaking on the manner in their contract. [NEWLINE] [NEWLINE] But once again, I fail to see the relevance here. </s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, and<mask><mask> with that. I just<mask> think it's important to remember Gandhi's core message didn't place prime importance on nonviolence. Gandhi cared much more strongly about resistance to tyranny than the principle of nonviolence.</s>
Label encoding: <s>Yeah, and I agree with that. I just also think it's important to remember Gandhi's core message didn't place prime importance on nonviolence. Gandhi cared much more strongly about resistance to tyranny than the principle of nonviolence.</s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many. <mask> the difference between a $25 bottle and a $100 bottle can be difficult for many to realize, the difference between a $10 bottle and a $25 bottle is amazingly obvious, even in a blind tasting.</s>
Label encoding: <s>Many.  While the difference between a $25 bottle and a $100 bottle can be difficult for many to realize, the difference between a $10 bottle and a $25 bottle is amazingly obvious, even in a blind tasting.</s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OK, this differentiation seems to be necessary to make your argument work, thanks for spelling it out. [NEWLINE] [NEWLINE] Edit: OK, apparently I'm not allowed to thank you for changing my mind. Sorry about that. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>OK, this differentiation seems to be necessary to make your argument work, thanks for spelling it out. [NEWLINE] [NEWLINE] Edit: OK, apparently I'm not allowed to thank you for changing my mind. Sorry about that. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't want to be presumptuous,<mask> we were having a pretty good dialogue, then you went silent.<mask> I made you think differently about the subject, can you please award a delta? Thank you. </s>
Label encoding: <s>I don't want to be presumptuous, but we were having a pretty good dialogue, then you went silent. If I made you think differently about the subject, can you please award a delta? Thank you. </s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I guess we can conclude that it's futile clapping with the expectation of anyone responsible for the creation of said movie acknowledging us.<mask> reasons for clapping are highly subjective and situational<mask> on the surface it looks generic. </s>
Label encoding: <s>I guess we can conclude that it's futile clapping with the expectation of anyone responsible for the creation of said movie acknowledging us. Also reasons for clapping are highly subjective and situational although on the surface it looks generic. </s>
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It seemed<mask><mask> he really meant it, wouldn't it be a little easy<mask> you could just say 'I love BB' to get out of the system.<mask><mask> you have to really mean it to be killed.</s><pad>
Label encoding: <s>It seemed as if he really meant it, wouldn't it be a little easy if you could just say 'I love BB' to get out of the system. I think you have to really mean it to be killed.</s><pad>
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The government interferes with the free market all the time to protect it's citizens. We have other laws about family medical leave, disability leave and health standards in the work place.<mask> should maternity care be any different?</s>
Label encoding: <s>The government interferes with the free market all the time to protect it's citizens. We have other laws about family medical leave, disability leave and health standards in the work place. Why should maternity care be any different?</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I try to do that<mask> I can,<mask> the situations in which I have a crisis of conscience are those<mask> I honestly don't have time to go in, wait in line, and buy<mask> ever they need. </s>
Label encoding: <s>I try to do that when I can, but the situations in which I have a crisis of conscience are those when I honestly don't have time to go in, wait in line, and buy what ever they need. </s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually it worked out better than<mask> they had before. You can check out the pdf [Better off stateless: Somalia before and after government collapse]( [URL].pdf) or [this video]( [URL] ) for more details.</s><pad>
Label encoding: <s>Actually it worked out better than what they had before. You can check out the pdf [Better off stateless: Somalia before and after government collapse]( [URL].pdf) or [this video]( [URL] ) for more details.</s><pad>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Having sex is not a contract, there are [requirements]( [URL] ) that sex does not fill. A woman is not legally obligated to carry a child to term<mask> she consented to sex, that makes no sense.</s>
Label encoding: <s>Having sex is not a contract, there are [requirements]( [URL] ) that sex does not fill. A woman is not legally obligated to carry a child to term because she consented to sex, that makes no sense.</s>
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A man and a woman creates a child. [ENDQ] [NEWLINE] No, a woman and a man create a pregnancy.  The way the law works, the woman has the sole authority to create a child from a pregnancy. </s>
Label encoding: <s> [STARTQ] A man and a woman creates a child. [ENDQ] [NEWLINE] No, a woman and a man create a pregnancy.  The way the law works, the woman has the sole authority to create a child from a pregnancy. </s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0907, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah. [NEWLINE] [NEWLINE] <mask> I tend to consider the probabilities, and the probability of the US Government turning against its citizens is astronomically low. [NEWLINE] [NEWLINE] I infer that you somehow think it is an inevitable eventuality.</s>
Label encoding: <s>Yeah. [NEWLINE] [NEWLINE] But I tend to consider the probabilities, and the probability of the US Government turning against its citizens is astronomically low. [NEWLINE] [NEWLINE] I infer that you somehow think it is an inevitable eventuality.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ours was like that too, somewhat,<mask> aside from electives most classes had primarily a single grade. For example my freshman year I took geometry<mask> 90% of the students in the class were sophomores.</s>
Label encoding: <s>Ours was like that too, somewhat, although aside from electives most classes had primarily a single grade. For example my freshman year I took geometry but 90% of the students in the class were sophomores.</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So they already teach similar topics in highschool,<mask> clearly highschool kids can handle it. At the risk of slight redundancy,<mask> not have a health class that includes that info specifically in the context of safe sex?</s>
Label encoding: <s>So they already teach similar topics in highschool, so clearly highschool kids can handle it. At the risk of slight redundancy, why not have a health class that includes that info specifically in the context of safe sex?</s>
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> some dude came along painting black squares [ENDQ] [NEWLINE] [Color Field Painting]( [URL] ) is definately a thing. Mark Rothko comes to mind. He painted [this]( [URL].jpg) in 1953.</s>
Label encoding: <s> [STARTQ] if some dude came along painting black squares [ENDQ] [NEWLINE] [Color Field Painting]( [URL] ) is definately a thing. Mark Rothko comes to mind. He painted [this]( [URL].jpg) in 1953.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0951, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why should Person A get citizenship by birth?<mask> is voting tied to citizenship? I know *<mask> * the system works, I want to know<mask> there's a reason *<mask> * it doesn't work better.</s>
Label encoding: <s>Why should Person A get citizenship by birth? Why is voting tied to citizenship? I know * how * the system works, I want to know if there's a reason * why * it doesn't work better.</s>
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not a massive point,<mask><mask><mask> it contributes and I didn't see it mentioned<mask>.<mask> maybe there's another reason for sex-segregation in that men *hate* losing to women?</s>
Label encoding: <s>It's not a massive point, but I think it contributes and I didn't see it mentioned yet. But maybe there's another reason for sex-segregation in that men *hate* losing to women?</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They certainly did with zip guns. [ENDQ] [NEWLINE] Never heard of a crime being committed with a zip gun.  Now, go Google me an example.  Now go Google crimes committed with a real weapon.</s>
Label encoding: <s> [STARTQ] They certainly did with zip guns. [ENDQ] [NEWLINE] Never heard of a crime being committed with a zip gun.  Now, go Google me an example.  Now go Google crimes committed with a real weapon.</s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except gender/trans* people aren't always non-standard oriented sexually (you have gay trans,<mask> you have hetero trans too. And pan, bi, etc. That still doesn't tell anything)</s>
Label encoding: <s>Except gender/trans* people aren't always non-standard oriented sexually (you have gay trans, but you have hetero trans too. And pan, bi, etc. That still doesn't tell anything)</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This weird thing with computer science is that it's rated<mask> one of the highest paying majors,<mask> that's without considering that programming is<mask> one of the highest paying jobs you can do without a degree.</s>
Label encoding: <s>This weird thing with computer science is that it's rated as one of the highest paying majors, but that's without considering that programming is also one of the highest paying jobs you can do without a degree.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I just used the term<mask> it is used in porn. I thought that there was a difference between them and transwomen. In the future I will just call any woman with a penis a transwoman.</s>
Label encoding: <s>I just used the term because it is used in porn. I thought that there was a difference between them and transwomen. In the future I will just call any woman with a penis a transwoman.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] sick days [ENDQ] [NEWLINE] Yeah don't use that<mask> evidence, unless you're some kind of Dickensian factory owner. People here work<mask> sick, and it's both gross and costs us money.</s>
Label encoding: <s> [STARTQ] sick days [ENDQ] [NEWLINE] Yeah don't use that as evidence, unless you're some kind of Dickensian factory owner. People here work while sick, and it's both gross and costs us money.</s>
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First, I know<mask> an analogy is. <mask>, maybe we should be focusing on changing the hurtful definitions, rather than complaining that all shifts in definition should be considered one way or another.</s>
Label encoding: <s>First, I know what an analogy is.  Secondly, maybe we should be focusing on changing the hurtful definitions, rather than complaining that all shifts in definition should be considered one way or another.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Then your CMV is really CMV: People shouldn't be dicks? Don't think anyone can really argue against that [ENDQ] [NEWLINE] Careful now, this is reddit we're talking about...</s>
Label encoding: <s> [STARTQ] Then your CMV is really CMV: People shouldn't be dicks? Don't think anyone can really argue against that [ENDQ] [NEWLINE] Careful now, this is reddit we're talking about...</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But we *have* moved forward. Homosexuality is much more accepted today than it was even 10 years ago.<mask> makes you think this progress is going to stop all of the sudden?</s>
Label encoding: <s>But we *have* moved forward. Homosexuality is much more accepted today than it was even 10 years ago. What makes you think this progress is going to stop all of the sudden?</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah,<mask> OP explicitly said that Consititutional law would only have to be changed slightly, and that is not true.<mask>, my point is directly challenging OP's view in some way.</s>
Label encoding: <s>Yeah, but OP explicitly said that Consititutional law would only have to be changed slightly, and that is not true. Thus, my point is directly challenging OP's view in some way.</s>
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That money could buy a lot of anything. The drones are not capable in air to air combat. You might<mask> well spend the money on bouncy balls<mask> it will have the same effect.</s>
Label encoding: <s>That money could buy a lot of anything. The drones are not capable in air to air combat. You might as well spend the money on bouncy balls because it will have the same effect.</s>
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Exactly. This attitude is the reason you can only be responsible for your safety. Don't drive unsafe just<mask> entitled individuals are willing to put others at risk<mask> "they have shit to do".</s>
Label encoding: <s>Exactly. This attitude is the reason you can only be responsible for your safety. Don't drive unsafe just because entitled individuals are willing to put others at risk because "they have shit to do".</s>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fair enough. I still think that there is an underlying issue at play<mask> someone is not able to understand<mask> the drug is affecting them and/or<mask> it should be used safely. </s>
Label encoding: <s>Fair enough. I still think that there is an underlying issue at play if someone is not able to understand how the drug is affecting them and/or how it should be used safely. </s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't understand. Organic producers aren't attempting to get "normal" agriculture *banned*. They're just choosing not to engage in it.<mask> will this lead to anyone starving?</s>
Label encoding: <s>I don't understand. Organic producers aren't attempting to get "normal" agriculture *banned*. They're just choosing not to engage in it. How will this lead to anyone starving?</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry, your post has been removed. [NEWLINE] [NEWLINE] &gt; Rule 1. Direct responses to a CMV post must challenge at least one aspect of OP’s stated view.</s>
Label encoding: <s>Sorry, your post has been removed. [NEWLINE] [NEWLINE] &gt; Rule 1. Direct responses to a CMV post must challenge at least one aspect of OP’s stated view.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The ultimate end of your train of logic seems to be that every Catholic should just sit down and stop doing anything,<mask><mask> it's God's will, than it will happen anyways.</s>
Label encoding: <s>The ultimate end of your train of logic seems to be that every Catholic should just sit down and stop doing anything, because if it's God's will, than it will happen anyways.</s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Medicine is founded on invasive advertising in america. It is not<mask> and more accessible in lots of other parts of the globe. [ENDQ] [NEWLINE] Is entertainment not the same way?</s>
Label encoding: <s> [STARTQ] Medicine is founded on invasive advertising in america. It is not so and more accessible in lots of other parts of the globe. [ENDQ] [NEWLINE] Is entertainment not the same way?</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some of them wish that.<mask> many don't. It's wrong either way. [NEWLINE] [NEWLINE] One of the main points here are that the goal doesn't always justify the means  </s>
Label encoding: <s>Some of them wish that. While many don't. It's wrong either way. [NEWLINE] [NEWLINE] One of the main points here are that the goal doesn't always justify the means  </s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm not sure<mask> they're anti-woman or<mask> they're just stupid. [ENDQ] [NEWLINE] Or they just don't believe in using evil means to a good end. </s>
Label encoding: <s> [STARTQ] I'm not sure if they're anti-woman or if they're just stupid. [ENDQ] [NEWLINE] Or they just don't believe in using evil means to a good end. </s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I can't look at someones nipples to see<mask> they're happy or sad. [ENDQ] [NEWLINE] <mask> you can use them to determine<mask> a person is a comfortable temperature! </s>
Label encoding: <s> [STARTQ] I can't look at someones nipples to see if they're happy or sad. [ENDQ] [NEWLINE] But you can use them to determine if a person is a comfortable temperature! </s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Exis007. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Exis007. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think he was following the logic<mask> he called them terrorists, yes.<mask> the statement about 50% of progressives and liberals hating white dudes? That was his own ridiculous statement.</s>
Label encoding: <s>I think he was following the logic when he called them terrorists, yes. But the statement about 50% of progressives and liberals hating white dudes? That was his own ridiculous statement.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's easy to assume you don't need to shoot<mask> the assumption is that the suspect doesn't have a gun. [NEWLINE] [NEWLINE] That's clearly not the case in the US.</s>
Label encoding: <s>It's easy to assume you don't need to shoot when the assumption is that the suspect doesn't have a gun. [NEWLINE] [NEWLINE] That's clearly not the case in the US.</s>
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Islam has virtually never been violent,<mask> it has almost never even been political [ENDQ] [NEWLINE] I literally just need to read a book about Constantinople to throw that away<mask> folly. </s>
Label encoding: <s> [STARTQ] Islam has virtually never been violent, indeed it has almost never even been political [ENDQ] [NEWLINE] I literally just need to read a book about Constantinople to throw that away as folly. </s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not much,<mask> the land in Western Canada is surveyed in a 1 mile by 2 mile grid.  Miles and acres aren't going anywhere for a long time.</s>
Label encoding: <s>It's not much, but the land in Western Canada is surveyed in a 1 mile by 2 mile grid.  Miles and acres aren't going anywhere for a long time.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0710, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[Ever read about Rotherham, UK?]( [URL]?_r=0) Now that sounds like a third world country. Sort your shit out, England.</s>
Label encoding: <s>[Ever read about Rotherham, UK?]( [URL]?_r=0) Now that sounds like a third world country. Sort your shit out, England.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0627, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The state level is exactly<mask> women's and LGBT rights are being attacked. Something like 200 bills restricting abortion were proposed in state congresses over the past couple of years.</s>
Label encoding: <s>The state level is exactly where women's and LGBT rights are being attacked. Something like 200 bills restricting abortion were proposed in state congresses over the past couple of years.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agreed, I was simply contesting the idea that something<mask> simple<mask> "do unto others<mask> you would have others do unto you" could encompass one's morality.</s>
Label encoding: <s>Agreed, I was simply contesting the idea that something as simple as "do unto others as you would have others do unto you" could encompass one's morality.</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OK,<mask><mask><mask> this isn't one of those? I'd put money on there being more stories with villain-driven plots than hero-driven ones.</s>
Label encoding: <s>OK, but what if this isn't one of those? I'd put money on there being more stories with villain-driven plots than hero-driven ones.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think<mask> tbh, I've never seen anyone walk and eat in Hong Kong or Mexico City and they are huge cities, bigger than NYC.</s>
Label encoding: <s>I don't think so tbh, I've never seen anyone walk and eat in Hong Kong or Mexico City and they are huge cities, bigger than NYC.</s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm pretty sure i will never be the employer or employee involved in a multimillion dollar deal<mask> the best I can do is judge from the outside. </s>
Label encoding: <s>I'm pretty sure i will never be the employer or employee involved in a multimillion dollar deal so the best I can do is judge from the outside. </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well,<mask> they talk about it a lot it would certainly indicate that they are fixated on it,<mask> yeah that's a reasonable assumption to make.</s>
Label encoding: <s>Well, if they talk about it a lot it would certainly indicate that they are fixated on it, so yeah that's a reasonable assumption to make.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710;  The sheer "amount" of regulation isn't necessarily proportional to the resultant benefits/assurances to the patient.</s>
Label encoding: <s>&amp;#8710;  The sheer "amount" of regulation isn't necessarily proportional to the resultant benefits/assurances to the patient.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How about yourself? <mask> are you deemed responsible to receive real dollars instead of food stamps, mortgage stamps, gas stamps, and alcohol stamps?</s>
Label encoding: <s>How about yourself?  Why are you deemed responsible to receive real dollars instead of food stamps, mortgage stamps, gas stamps, and alcohol stamps?</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It still wouldn't work.  Machiavelli understood very well the problems with mercenaries - they don't need customers<mask> they have victims.</s><pad>
Label encoding: <s>It still wouldn't work.  Machiavelli understood very well the problems with mercenaries - they don't need customers when they have victims.</s><pad>
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0989, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maleficent's minions are ugly goblin things. [NEWLINE] [NEWLINE] Gaston's minions are a set of swooning blonde triplets. </s>
Label encoding: <s>Maleficent's minions are ugly goblin things. [NEWLINE] [NEWLINE] Gaston's minions are a set of swooning blonde triplets. </s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yahweh is used to refer to God<mask> a whole, not Jesus specifically. The living person was not named Yahweh.</s>
Label encoding: <s>Yahweh is used to refer to God as a whole, not Jesus specifically. The living person was not named Yahweh.</s>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The show did not cause the suicides.  The show did not make someone kill themselves; that is a choice they made for themselves.</s><pad>
Label encoding: <s>The show did not cause the suicides.  The show did not make someone kill themselves; that is a choice they made for themselves.</s><pad>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No it didn't.<mask> I recall correctly, 35 is typically good enough. I can dig out the ol stats text book<mask>.</s><pad>
Label encoding: <s>No it didn't. If I recall correctly, 35 is typically good enough. I can dig out the ol stats text book though.</s><pad>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One denomination, which makes up less then 1 percent of all Christians, is a really bad way to generalize an entire faith.</s><pad>
Label encoding: <s>One denomination, which makes up less then 1 percent of all Christians, is a really bad way to generalize an entire faith.</s><pad>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why does this need to be ensured? Shouldn't an educated electorate be able to decide the most capable candidate for themselves?</s><pad>
Label encoding: <s>Why does this need to be ensured? Shouldn't an educated electorate be able to decide the most capable candidate for themselves?</s><pad>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why don't we just sit in corners and never get offended. We should never talk to each about anything<mask> superficial shit.</s>
Label encoding: <s>Why don't we just sit in corners and never get offended. We should never talk to each about anything but superficial shit.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you familiar with the Blue Falcon award?<mask> not, look it up. You're a contender for sure.  </s>
Label encoding: <s>Are you familiar with the Blue Falcon award? If not, look it up. You're a contender for sure.  </s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you kidding me. [NEWLINE] [NEWLINE] Source that women don't have to pay child support for children raised by their father?</s>
Label encoding: <s>Are you kidding me. [NEWLINE] [NEWLINE] Source that women don't have to pay child support for children raised by their father?</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not really. Just<mask> something is worth discussing on the internet doesn't mean it's worth including in the education system.</s>
Label encoding: <s>Not really. Just because something is worth discussing on the internet doesn't mean it's worth including in the education system.</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Direct responses to a CMV post must challenge at least one aspect of OP’s stated view. [ENDQ] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Direct responses to a CMV post must challenge at least one aspect of OP’s stated view. [ENDQ] [NEWLINE] </s>
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0816, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agreed. Maybe that would convince them to work harder in order to not have to have a "bullshit major".</s>
Label encoding: <s>Agreed. Maybe that would convince them to work harder in order to not have to have a "bullshit major".</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You wanted something more complicated? Sometimes all that's required is a simple answer, even<mask> the question is complex.</s>
Label encoding: <s>You wanted something more complicated? Sometimes all that's required is a simple answer, even if the question is complex.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oprah was bodyshamed her entire fucking career. I don't think you know<mask> a class is.</s>
Label encoding: <s>Oprah was bodyshamed her entire fucking career. I don't think you know what a class is.</s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> it's racism from an Asian race against another Asian race, doesn't make it not racism against Asians</s>
Label encoding: <s>Just because it's racism from an Asian race against another Asian race, doesn't make it not racism against Asians</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course. :) I just meant, they both supplied their own books *and* accepted donated bibles.</s>
Label encoding: <s>Of course. :) I just meant, they both supplied their own books *and* accepted donated bibles.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I still do not want to remove the downvote function and I still want people to use it more often.</s>
Label encoding: <s>I still do not want to remove the downvote function and I still want people to use it more often.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I never put that together. I am the worst scientist ever. In the world. Of all time. </s>
Label encoding: <s>I never put that together. I am the worst scientist ever. In the world. Of all time. </s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] ~$144,000 in my pocket right now. [ENDQ] [NEWLINE] And then you'd be homeless...</s><pad>
Label encoding: <s> [STARTQ] ~$144,000 in my pocket right now. [ENDQ] [NEWLINE] And then you'd be homeless...</s><pad>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But<mask> would you judge<mask> is worse? Maybe a loss for you is a gain for someone else.</s>
Label encoding: <s>But how would you judge what is worse? Maybe a loss for you is a gain for someone else.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I do that with Firefox<mask>. It's not a Google chrome thing. It's a Google thing.</s>
Label encoding: <s>I do that with Firefox though. It's not a Google chrome thing. It's a Google thing.</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What makes humans<mask> important to you that their personal enjoyment should come above the life of another being?</s>
Label encoding: <s>What makes humans so important to you that their personal enjoyment should come above the life of another being?</s>
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Go be a man and buy a motorcycle and you'll learn real quick that loud pipes save lives.</s>
Label encoding: <s>Go be a man and buy a motorcycle and you'll learn real quick that loud pipes save lives.</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>what I'm saying is, you're downplaying<mask> significant that biological challenge is. </s>
Label encoding: <s>what I'm saying is, you're downplaying how significant that biological challenge is. </s>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Then I expect my delta<mask> you are now aware that I did<mask> comment on it.</s>
Label encoding: <s>Then I expect my delta since you are now aware that I did indeed comment on it.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Much like all people hopelessly devoted to religion, your belief doesnt measure up to reality </s><pad>
Label encoding: <s>Much like all people hopelessly devoted to religion, your belief doesnt measure up to reality </s><pad>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Would you engage in a life-changing activity that had a 50% failure rate?</s>
Label encoding: <s>Would you engage in a life-changing activity that had a 50% failure rate?</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This question may seem out of place<mask> do you believe people have free will?</s>
Label encoding: <s>This question may seem out of place but do you believe people have free will?</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Presumably it was<mask> family bloodlines wouldn't be ended by the policy</s>
Label encoding: <s>Presumably it was so family bloodlines wouldn't be ended by the policy</s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Inhuman =\= object. There's a pretty big difference. </s>
Label encoding: <s>Inhuman =\= object. There's a pretty big difference. </s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure this addresses<mask> the OP was talking about.</s>
Label encoding: <s>I'm not sure this addresses what the OP was talking about.</s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You probably meant "we live in a democratic republic." </s>
Label encoding: <s>You probably meant "we live in a democratic republic." </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ah the rule of three. Everything is great in three.</s><pad>
Label encoding: <s>Ah the rule of three. Everything is great in three.</s><pad>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you have any studys on healthy rested people?</s>
Label encoding: <s>Do you have any studys on healthy rested people?</s>
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, you can,<mask> you won't win.</s>
Label encoding: <s>Well, you can, but you won't win.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1314, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Shinto is a larger practice in Japan than Buddhism...</s>
Label encoding: <s>Shinto is a larger practice in Japan than Buddhism...</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, C++ doesn't look simple...</s>
Label encoding: <s>Well, C++ doesn't look simple...</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0781, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.1467e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.6812e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.8109e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.5357e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.1619e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.9429e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.0755e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.7637e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.3674e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.0302e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.0066e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.0867e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.2938e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.5316e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.0539e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.1025e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.5552e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.2372e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.9426e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.1694e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.8757e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.3746e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.6405e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.9877e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.7245e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.7989e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.8846e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.8902e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.5007e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.1833e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.9842e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.7308e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.0671e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.5323e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.2559e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.8488e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.0432e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.1206e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.1532e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.6950e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.5660e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.1344e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.5747e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.3022e-05, device='cuda:0', grad_fn=<NllLossBackward>)
--------------EPOCH 3-------------
Test Accuracy: tensor(0.7838, device='cuda:0')
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Saving model at iteration: 0
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Going up 20% for $100k more does seem like quite a bit more<mask> I'm not sure<mask> it would fit my definition of a slightly higher tax rate. It is my fault for making the title<mask> ambiguous,<mask> I apologize for that. [ENDQ] [NEWLINE] Not really you,<mask> I have seen plenty of people advocating that we tax millionaires at effective rates upwards of 60%.<mask> I wanted to start my own business in no small part to become a millionaire (or, more diplomatically put, to ensure my and my children's financial security)<mask> my effective tax rate at, say, $150,000 remained<mask> it is now (around 30%)<mask> jumped to 60% starting at $1 million, I'm not sure I'd have the motivation to start my own business entailing all the risks and long hours and crap that comes with it<mask> I only ended up with $400,000 of every million I made. That's not to say I wouldn't take that $400,000 in a heartbeat<mask> it was a guarantee,<mask> starting a risky venture is hardly a guarantee and we want to encourage and reward people for adding value to society (or even attempting to), not punish them for succeeding, which is<mask> a 60%+ tax rate would probably do. [NEWLINE] [NEWLINE] [STARTQ] Do you think the amount of people a higher tax rate deters from taking a more challenging job could have such a large effect on the economy<mask> there is<mask> much competition for jobs? [ENDQ] [NEWLINE] Are you asking that<mask> there is<mask> much competition for jobs, lowering taxes wouldn't matter<mask> much<mask> those who might then seek a more challenging job still wouldn't be able to find one? That's a really tough question to answer, and I'm not an economist or even particularly smart,<mask> I'll try to explain<mask> I see it. We pretty much always want more money in the economy, right?<mask> not _too_ much and not too quickly. Some inflation due to economic growth is good,<mask> not _too_ much.<mask> with that<mask> a given,<mask> I'm pretty sure of is that that more education = more jobs and more money = more jobs.<mask> competition for jobs (ie, lack of demand for labor) is probably due to one or both of those factors. [NEWLINE] [NEWLINE] <mask> there is a large supply of workers without the necessary education,<mask> that doesn't really say _anything_ about<mask> many jobs are _actually_ available to qualified candidates,<mask> we'll put it aside and assume we're qualified.
Label encoding: <s> [STARTQ] Going up 20% for $100k more does seem like quite a bit more so I'm not sure if it would fit my definition of a slightly higher tax rate. It is my fault for making the title so ambiguous, so I apologize for that. [ENDQ] [NEWLINE] Not really you, but I have seen plenty of people advocating that we tax millionaires at effective rates upwards of 60%. If I wanted to start my own business in no small part to become a millionaire (or, more diplomatically put, to ensure my and my children's financial security) but my effective tax rate at, say, $150,000 remained where it is now (around 30%) but jumped to 60% starting at $1 million, I'm not sure I'd have the motivation to start my own business entailing all the risks and long hours and crap that comes with it if I only ended up with $400,000 of every million I made. That's not to say I wouldn't take that $400,000 in a heartbeat if it was a guarantee, but starting a risky venture is hardly a guarantee and we want to encourage and reward people for adding value to society (or even attempting to), not punish them for succeeding, which is what a 60%+ tax rate would probably do. [NEWLINE] [NEWLINE] [STARTQ] Do you think the amount of people a higher tax rate deters from taking a more challenging job could have such a large effect on the economy when there is so much competition for jobs? [ENDQ] [NEWLINE] Are you asking that because there is so much competition for jobs, lowering taxes wouldn't matter so much since those who might then seek a more challenging job still wouldn't be able to find one? That's a really tough question to answer, and I'm not an economist or even particularly smart, but I'll try to explain how I see it. We pretty much always want more money in the economy, right? But not _too_ much and not too quickly. Some inflation due to economic growth is good, but not _too_ much. So with that as a given, what I'm pretty sure of is that that more education = more jobs and more money = more jobs. So competition for jobs (ie, lack of demand for labor) is probably due to one or both of those factors. [NEWLINE] [NEWLINE] So there is a large supply of workers without the necessary education, but that doesn't really say _anything_ about how many jobs are _actually_ available to qualified candidates, so we'll put it aside and assume we're qualified.
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"<mask> do you KNOW that you are fighting for the right?" [NEWLINE] In Iraq fighting against groups that are responsible for ethnic cleansing and encouraging political reconciliation between dissident groups<mask> they can disagree peacefully earns the moral high ground. [NEWLINE] [NEWLINE] In Afghanistan the Taliban are not in the business of winning "hearts and minds", or giving the Afghans better quality of life, they are in the business of reconquering the country, and in the process have been absolutely brutal. I recall story<mask> in Nuristan province a U.S. commander was having a shura with local tribal elders who were fence sitting. He simply asked them "Who do you think has a better vision for Afghanistan, the Taliban or us?" The answer is obvious. You can disagree with the political motivations for going to war in the first place,<mask> in terms of the wars in Iraq and Afghanistan you will have an impossible time trying to prove that the intentions and efforts of the occupying forces doesn't hold the moral high ground<mask> opposed to the opposition they face. [NEWLINE] [NEWLINE] "I've seen alot of soldiers state that they are there for peaceful purposes, and<mask> that is the case,<mask> don't we have a peacekeeping organization that can be signed up for opposed to the military.<mask><mask><mask>, 2 organizations seperated like this would have completely different view on<mask> to handle situations.<mask> we do have an organization like this,<mask> not join that?" [NEWLINE] [NEWLINE] Modern counter insurgency conflicts are not just military operations any more, its a massive spectrum of tasks and goals that go into rehabilitating a failed state<mask><mask> to whatever challenges the insurgency faces. In Iraq and Afghanistan,<mask> of the nature and goal of the wars, there were dozens of organizations both American and non-governmental international organizations working on the ground.<mask> the fundamental problem is that it is a warzone, the security situation is the first concern to be addressed, and humanitarian work cannot commence<mask> there is no semblance of security. In Afghanistan and Iraq humanitarian workers for non military organizations have been killed and kidnapped by insurgents, construction contractors<mask> well. A famous incident was Al Qaeda bombing  the headquarters of the U.N. Assistance mission in Iraq in 2003, which killed the U.N. high commissioner for human rights Sergio de Mello. The role of the soldier in these recent wars hasn't been solely to drop firepower, it has been to provide population security, facilitate good governance, political mediation, and improving the local economy through reconstruction funds, to name a few of the hats that soldiers have been made
Label encoding: <s>" how do you KNOW that you are fighting for the right?" [NEWLINE] In Iraq fighting against groups that are responsible for ethnic cleansing and encouraging political reconciliation between dissident groups so they can disagree peacefully earns the moral high ground. [NEWLINE] [NEWLINE] In Afghanistan the Taliban are not in the business of winning "hearts and minds", or giving the Afghans better quality of life, they are in the business of reconquering the country, and in the process have been absolutely brutal. I recall story when in Nuristan province a U.S. commander was having a shura with local tribal elders who were fence sitting. He simply asked them "Who do you think has a better vision for Afghanistan, the Taliban or us?" The answer is obvious. You can disagree with the political motivations for going to war in the first place, but in terms of the wars in Iraq and Afghanistan you will have an impossible time trying to prove that the intentions and efforts of the occupying forces doesn't hold the moral high ground as opposed to the opposition they face. [NEWLINE] [NEWLINE] "I've seen alot of soldiers state that they are there for peaceful purposes, and if that is the case, why don't we have a peacekeeping organization that can be signed up for opposed to the military. In my opinion, 2 organizations seperated like this would have completely different view on how to handle situations. If we do have an organization like this, why not join that?" [NEWLINE] [NEWLINE] Modern counter insurgency conflicts are not just military operations any more, its a massive spectrum of tasks and goals that go into rehabilitating a failed state in addition to whatever challenges the insurgency faces. In Iraq and Afghanistan, because of the nature and goal of the wars, there were dozens of organizations both American and non-governmental international organizations working on the ground. But the fundamental problem is that it is a warzone, the security situation is the first concern to be addressed, and humanitarian work cannot commence if there is no semblance of security. In Afghanistan and Iraq humanitarian workers for non military organizations have been killed and kidnapped by insurgents, construction contractors as well. A famous incident was Al Qaeda bombing  the headquarters of the U.N. Assistance mission in Iraq in 2003, which killed the U.N. high commissioner for human rights Sergio de Mello. The role of the soldier in these recent wars hasn't been solely to drop firepower, it has been to provide population security, facilitate good governance, political mediation, and improving the local economy through reconstruction funds, to name a few of the hats that soldiers have been made
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> not? You choose<mask> to believe yourself, your view of the world is<mask> much a choice<mask> the view of someone who has faith [ENDQ] [NEWLINE] I'm sorry<mask> I just can't buy this. Some things are facts.<mask> I kick a ball forward it will move forward. Every time. It will never go backward<mask> I apply a force forward. It has nothing to do with<mask> I believe. I *happen* to believe that<mask> I kick it forward it will move forward<mask> I know that's<mask> happens. And I would be *100% wrong*<mask> I chose to believe that it would go backward<mask> I kick it forward. [NEWLINE] [NEWLINE] [STARTQ] Remember, we're not discussing<mask> you should believe in God, we're discussing whether the existence of belief in a God can be defended. [ENDQ] [NEWLINE] Saying that God exists is a statement of fact. Not a belief.<mask> I were to say that a blue haired mongoose exists, I would have to prove it to you. A belief that a blue haired mongoose exists would have to be defended with evidence of a blue haired mongoose. [NEWLINE] [NEWLINE] I believe that a belief that God exists is impossible to defend logically<mask> there is zero objective proof of its existence.<mask> else would you go about saying that a belief in him is defensible? [NEWLINE] [NEWLINE] Believing that things that do not exist actually exist in the real world is not supported for anything other than God. I happen to be someone who doesn't understand<mask> we say "well of course (<mask><mask> the idea of Santa Claus exists) Santa Claus doesn't *actually* exist in the real world", and I dont understand<mask> we don't say, "well of course God (<mask><mask> the idea of an infinite, omniscient, omnipotent being exists) doesn't exists in the *real world*".<mask> people still say that<mask> the possibility of him exists, they believe he **actually does* exist. That's<mask> my issue arises. [NEWLINE] [NEWLINE] **It is not logically defensible to say that God actually exists in the real world.** This is<mask> I defend. And I don't understand<mask> one could go about defending the opposite claim. [NEWLINE] [NEWLINE] [STARTQ] it may be possible to derive a perfect set of morals based on logic alone, the proof has<mask> to be made. [ENDQ] [NEWLINE] I would<mask><mask> the collective set of all literature, art, culture, sociology, etc. etc. knowledge would provide more than enough basis for a set
Label encoding: <s> [STARTQ] Why not? You choose what to believe yourself, your view of the world is as much a choice as the view of someone who has faith [ENDQ] [NEWLINE] I'm sorry but I just can't buy this. Some things are facts. If I kick a ball forward it will move forward. Every time. It will never go backward if I apply a force forward. It has nothing to do with what I believe. I *happen* to believe that if I kick it forward it will move forward because I know that's what happens. And I would be *100% wrong* if I chose to believe that it would go backward if I kick it forward. [NEWLINE] [NEWLINE] [STARTQ] Remember, we're not discussing why you should believe in God, we're discussing whether the existence of belief in a God can be defended. [ENDQ] [NEWLINE] Saying that God exists is a statement of fact. Not a belief. If I were to say that a blue haired mongoose exists, I would have to prove it to you. A belief that a blue haired mongoose exists would have to be defended with evidence of a blue haired mongoose. [NEWLINE] [NEWLINE] I believe that a belief that God exists is impossible to defend logically because there is zero objective proof of its existence. How else would you go about saying that a belief in him is defensible? [NEWLINE] [NEWLINE] Believing that things that do not exist actually exist in the real world is not supported for anything other than God. I happen to be someone who doesn't understand why we say "well of course ( even though the idea of Santa Claus exists) Santa Claus doesn't *actually* exist in the real world", and I dont understand why we don't say, "well of course God ( even though the idea of an infinite, omniscient, omnipotent being exists) doesn't exists in the *real world*". But people still say that because the possibility of him exists, they believe he **actually does* exist. That's where my issue arises. [NEWLINE] [NEWLINE] **It is not logically defensible to say that God actually exists in the real world.** This is what I defend. And I don't understand how one could go about defending the opposite claim. [NEWLINE] [NEWLINE] [STARTQ] it may be possible to derive a perfect set of morals based on logic alone, the proof has yet to be made. [ENDQ] [NEWLINE] I would argue that the collective set of all literature, art, culture, sociology, etc. etc. knowledge would provide more than enough basis for a set
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's [grab heroin]( [URL].socialsciences) for a moment,<mask> I don't have much information on similar types of studies for crack. [NEWLINE] [NEWLINE] [STARTQ] In Liverpool, during the early 1990s, Dr John Marks used a special Home Office licence to prescribe heroin to addicts. Police reported a 96% reduction in acquisitive crime among a group of addict patients. Deaths from locally acquired HIV infection and drug-related overdoses fell to zero.<mask>, under intense pressure from the government, the project was closed down. In its 10 years' work, not one of its patients had died. In the first two years after it was closed, 41 died. [ENDQ] [NEWLINE] <mask><mask> it doesn't say in the article is that part of the deal for giving them this drug was they had to stay on the clean side of the law.<mask> they got arrested, no more supply. That, in large part, accounts for the dramatic drop in crime. [NEWLINE] [NEWLINE] The reason the program was shut down was that [60 Minutes]( [URL] ) did a lengthy story on it and the US gov't pressured the UK to stop him: [NEWLINE] [NEWLINE] [STARTQ] Ed: To get drugs from the clinic rather than the Mafia, addicts have to take a urine test to prove they are taking the drug [ENDQ] they say they are. And unlike most other addiction clinics<mask> you have to say you want to kick the habit before they'll take you in, addicts here have to convince Dr. Marks, a nurse and a social worker they intend to stay on drugs come<mask> may.<mask> does Dr. Marks try to cure people? [NEWLINE] [NEWLINE] [STARTQ] Dr. Marks: Cure people? Nobody can.<mask><mask> whether you stick them in prison, put them in mental hospitals and give them shock treatment, we have done all these things, put them in a nice rehab center away in the country, give them a nice social worker and pat them on the head, give them drugs, give them no drugs, does not matter<mask> you do. 5% per annum, 1 in 20 per year, get off spontaneously. Compound interested up that reaches about 50% (50/50) after ten years are off drugs. They seem to mature out of addiction<mask><mask> any intervention in the interim<mask> you can keep them alive and healthy and legal during that 10 years,<mask> you<mask> wish to. [ENDQ] [NEWLINE] [STARTQ] Ed: By giving them drugs? [ENDQ] [NEWLINE] [STARTQ] Dr. Marks: It doesn't get them off drugs, it doesn't prolong their addiction, either.
Label encoding: <s>Let's [grab heroin]( [URL].socialsciences) for a moment, because I don't have much information on similar types of studies for crack. [NEWLINE] [NEWLINE] [STARTQ] In Liverpool, during the early 1990s, Dr John Marks used a special Home Office licence to prescribe heroin to addicts. Police reported a 96% reduction in acquisitive crime among a group of addict patients. Deaths from locally acquired HIV infection and drug-related overdoses fell to zero. But, under intense pressure from the government, the project was closed down. In its 10 years' work, not one of its patients had died. In the first two years after it was closed, 41 died. [ENDQ] [NEWLINE] So what it doesn't say in the article is that part of the deal for giving them this drug was they had to stay on the clean side of the law. If they got arrested, no more supply. That, in large part, accounts for the dramatic drop in crime. [NEWLINE] [NEWLINE] The reason the program was shut down was that [60 Minutes]( [URL] ) did a lengthy story on it and the US gov't pressured the UK to stop him: [NEWLINE] [NEWLINE] [STARTQ] Ed: To get drugs from the clinic rather than the Mafia, addicts have to take a urine test to prove they are taking the drug [ENDQ] they say they are. And unlike most other addiction clinics where you have to say you want to kick the habit before they'll take you in, addicts here have to convince Dr. Marks, a nurse and a social worker they intend to stay on drugs come what may. But does Dr. Marks try to cure people? [NEWLINE] [NEWLINE] [STARTQ] Dr. Marks: Cure people? Nobody can. Regardless of whether you stick them in prison, put them in mental hospitals and give them shock treatment, we have done all these things, put them in a nice rehab center away in the country, give them a nice social worker and pat them on the head, give them drugs, give them no drugs, does not matter what you do. 5% per annum, 1 in 20 per year, get off spontaneously. Compound interested up that reaches about 50% (50/50) after ten years are off drugs. They seem to mature out of addiction regardless of any intervention in the interim but you can keep them alive and healthy and legal during that 10 years, if you so wish to. [ENDQ] [NEWLINE] [STARTQ] Ed: By giving them drugs? [ENDQ] [NEWLINE] [STARTQ] Dr. Marks: It doesn't get them off drugs, it doesn't prolong their addiction, either.
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They eat poo. [ENDQ] [NEWLINE] This is a sign of severe mental problems in a dog. <mask> a dog is eating poop, something is wrong with them and they some sort of major change in their life.  It very much doesn't apply to the majority of dogs. [NEWLINE] [NEWLINE] [STARTQ] They are annoying. Many people have really strong opinions against smoking in public,<mask> it is upsetting them and<mask> laws have been past in many countries that limit/prohibit smoking in public transport, public buildings, etc.<mask>, it is not uncommon to be leg-humped by a dog in the bus, or step into dog-poo in the park, which I find really annoying. This seems to be regarded<mask> unproblematic, some pepole even get annoyed<mask> you do not want to touch their animal companion with questionable hygiene standards.* You may say that (contrary to smoking) these are merely inconveniences,<mask><mask><mask>... [ENDQ] [NEWLINE] <mask><mask> I live away from the city and would rarely even bring the dog onto property that I don't own? <mask> does any of this apply? <mask>, even in a city, this stuff is not an issue with a well trained and watched after dog. [NEWLINE] [NEWLINE] [STARTQ] Some dogs are plain dangerous. Have a look at the "Fatal dog attacks"[1] wiki entry. Granted, there are MUCH more people dying from cars, cigaretts, cancer (and that's only deadly stuff with a C), etc.<mask><mask><mask><mask> even one person would be too much. There are a lot of young children on the list<mask> well. [ENDQ] [NEWLINE] 100 times more people a year are killed by drowning in a pool (mostly targeting young children).  Should we outlaw [swimming pools]( [URL] ) too? [NEWLINE] [NEWLINE] [STARTQ] They eat. In a world<mask> people are starving this is<mask><mask><mask> morally not justifiable. [ENDQ] [NEWLINE] The world produces enough food to feed the entire world no problem, and feeding pets has little impact on that.  The issue comes not from food shortages,<mask> the difficulties in large scale distribution.  Us not feeding dogs and cats in the US will not suddenly put food on the plate of people in the Sudan. [NEWLINE] [NEWLINE] [STARTQ] Most importantly: They produce greenhouse gases - and quite a lot of them! The co2-"paw print" of a big dog that gets fed mostly meat may even be bigger than the emissions caused by an SUV. [ENDQ] [NEWLINE] Neither add up to being [major sources of greenhouse gasses on
Label encoding: <s> [STARTQ] They eat poo. [ENDQ] [NEWLINE] This is a sign of severe mental problems in a dog.  If a dog is eating poop, something is wrong with them and they some sort of major change in their life.  It very much doesn't apply to the majority of dogs. [NEWLINE] [NEWLINE] [STARTQ] They are annoying. Many people have really strong opinions against smoking in public, because it is upsetting them and accordingly laws have been past in many countries that limit/prohibit smoking in public transport, public buildings, etc. However, it is not uncommon to be leg-humped by a dog in the bus, or step into dog-poo in the park, which I find really annoying. This seems to be regarded as unproblematic, some pepole even get annoyed if you do not want to touch their animal companion with questionable hygiene standards.* You may say that (contrary to smoking) these are merely inconveniences, but in fact... [ENDQ] [NEWLINE] What If I live away from the city and would rarely even bring the dog onto property that I don't own?  How does any of this apply?  Also, even in a city, this stuff is not an issue with a well trained and watched after dog. [NEWLINE] [NEWLINE] [STARTQ] Some dogs are plain dangerous. Have a look at the "Fatal dog attacks"[1] wiki entry. Granted, there are MUCH more people dying from cars, cigaretts, cancer (and that's only deadly stuff with a C), etc. but in my opinion even one person would be too much. There are a lot of young children on the list as well. [ENDQ] [NEWLINE] 100 times more people a year are killed by drowning in a pool (mostly targeting young children).  Should we outlaw [swimming pools]( [URL] ) too? [NEWLINE] [NEWLINE] [STARTQ] They eat. In a world where people are starving this is in my opinion morally not justifiable. [ENDQ] [NEWLINE] The world produces enough food to feed the entire world no problem, and feeding pets has little impact on that.  The issue comes not from food shortages, but the difficulties in large scale distribution.  Us not feeding dogs and cats in the US will not suddenly put food on the plate of people in the Sudan. [NEWLINE] [NEWLINE] [STARTQ] Most importantly: They produce greenhouse gases - and quite a lot of them! The co2-"paw print" of a big dog that gets fed mostly meat may even be bigger than the emissions caused by an SUV. [ENDQ] [NEWLINE] Neither add up to being [major sources of greenhouse gasses on
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that people may be poor judges of their own happiness, purpose, etc. I<mask> agree that<mask> we feel at one time doesn't necessarily mean we'll continue to feel that way. We're irrational and imperfect.<mask> by this argument,<mask> should we ever be allowed to make a decision?<mask> can a judge sentence a criminal to imprisonment? Couldn't he change his mind later? Couldn't he be wrong? Yes,<mask> nonetheless we can only act<mask><mask> our *best* interpretation of things. And<mask> your *best* interpretation of the quality of your life is that it is not worth living, that is<mask> valid a determination<mask> can be made and that is the feeling on which you should act. (Another example:<mask> I decide I want to have a child and procreate, couldn't I change my mind in a few days and not want a child anymore? Couldn't I merely be misestimating the challenges parenthood will entail? Of course,<mask> it doesn't seem like you'd be willing to<mask><mask> I'm wrong or unjustified in choosing to be a parent<mask> my determination is that I want to be and am suited to be a parent). [NEWLINE] [NEWLINE] [NEWLINE] Suffering is only "useless"<mask> a concept in the sense that it is an imperfect concept. We make imperfect decisions, and they're based on imperfect analyses. "Good" and "evil" or "right" and "wrong" are<mask> imperfect ideas, impossible to fully define, and impossible to have actions fully assigned to one or the other.<mask> we still need to choose our actions based on some metric, even<mask> the metric is a bit fuzzy. Pleasure vs. suffering such a metric, and I see no reason<mask> it's not valid to consider. I can't *measure*<mask> much suffering I will cause someone by kicking their teeth in,<mask> I still know I shouldn't do it. [NEWLINE] [NEWLINE] [NEWLINE] I don't want to get too far off track here,<mask> regarding the necessity of simplifying philosophy: every academic discipline simplifies its subject matter to make modeling, discussing, and thinking about it possible. Is it a "made up game" to discuss DNA replication without naming every single bonding interaction between every molecule in DNA and transcriptase? Is it wrong to discuss classical or operant conditioning<mask> we don't know every neurotransmitter released in every synapse of the brain being conditioned? With philosophy, we cannot discuss every possible thing that makes life seem good or bad to every possible
Label encoding: <s>I agree that people may be poor judges of their own happiness, purpose, etc. I also agree that how we feel at one time doesn't necessarily mean we'll continue to feel that way. We're irrational and imperfect. But by this argument, why should we ever be allowed to make a decision? Why can a judge sentence a criminal to imprisonment? Couldn't he change his mind later? Couldn't he be wrong? Yes, but nonetheless we can only act according to our *best* interpretation of things. And if your *best* interpretation of the quality of your life is that it is not worth living, that is as valid a determination as can be made and that is the feeling on which you should act. (Another example: if I decide I want to have a child and procreate, couldn't I change my mind in a few days and not want a child anymore? Couldn't I merely be misestimating the challenges parenthood will entail? Of course, but it doesn't seem like you'd be willing to argue that I'm wrong or unjustified in choosing to be a parent if my determination is that I want to be and am suited to be a parent). [NEWLINE] [NEWLINE] [NEWLINE] Suffering is only "useless" as a concept in the sense that it is an imperfect concept. We make imperfect decisions, and they're based on imperfect analyses. "Good" and "evil" or "right" and "wrong" are also imperfect ideas, impossible to fully define, and impossible to have actions fully assigned to one or the other. But we still need to choose our actions based on some metric, even if the metric is a bit fuzzy. Pleasure vs. suffering such a metric, and I see no reason why it's not valid to consider. I can't *measure* how much suffering I will cause someone by kicking their teeth in, but I still know I shouldn't do it. [NEWLINE] [NEWLINE] [NEWLINE] I don't want to get too far off track here, but regarding the necessity of simplifying philosophy: every academic discipline simplifies its subject matter to make modeling, discussing, and thinking about it possible. Is it a "made up game" to discuss DNA replication without naming every single bonding interaction between every molecule in DNA and transcriptase? Is it wrong to discuss classical or operant conditioning because we don't know every neurotransmitter released in every synapse of the brain being conditioned? With philosophy, we cannot discuss every possible thing that makes life seem good or bad to every possible
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I didnt like the way you come off<mask> crazy and rude.<mask> I decided to reply in the vein. [NEWLINE] [NEWLINE] [STARTQ] No, the cause of an imbalanced ratio of men vs women in prison is largely<mask> of socializing men to use violence to solve their problems. Men commit the vast majority of violent crimes, and this is something that is true in nearly every culture globally,<mask> with less violent societies, there is less social pressure for men to be violent. [ENDQ] [NEWLINE] Only one time in my life did I feel "socialized" or a pressured to be violent. A man at a bar said something rude to my then GF, she then wanted me to fight for her honor.<mask> I guess maybe women are just trying to have the men be violent for them? [NEWLINE] [NEWLINE] [STARTQ] Is this another thing that you only feel, or do you have any actual evidence for this claim? [ENDQ] [NEWLINE] (personal bias incoming) [NEWLINE] [NEWLINE] This, this bit right here I disliked the most. Im a divorced, single father. My ex is on government assistance, by her being on government assistance she is legally entitled to not pay one dime in child support and I still get to pay her alimony(that is seriously the law, I get nothing and MUST pay alimony)! YAY fair and balanced womens lib groups!!! She has a job, she makes 5k/year less than I do (and has 0 child time)<mask><mask> she applied and was granted government assistance (a womans only health policy offered by my state to any mother with a child under 8). She gets a pass on paying for her kid and I get the stick! (NO idea<mask> some men would be offended by this! Us penis having asshats!) [NEWLINE] [NEWLINE] [STARTQ] I find it clear that you have no idea<mask> feminism is. Feminism is not about "human rights" or "equality". Feminism is about defining<mask> "human rights" or "equality" even mean. [ENDQ] [NEWLINE] Is this just a thing that only you feel?<mask> you went to every crack pot on the world wide net your going to get a different definition for<mask> it means. [NEWLINE] [NEWLINE] <mask> the Feminist movement is reactionary anti-male movement that is primarily about burning bras and letting out **emotions** rather than any sort of respect for intellectualism, academic rigor, or actual social change! [NEWLINE] [NEWLINE] [STARTQ] I don't know<mask> you know about first wave feminism [ENDQ] [NEWLINE] Honestly zero, next to nothing. Can you surf on it? [NEWLINE] [NEWLINE] [STARTQ] By this
Label encoding: <s>I didnt like the way you come off as crazy and rude. So I decided to reply in the vein. [NEWLINE] [NEWLINE] [STARTQ] No, the cause of an imbalanced ratio of men vs women in prison is largely because of socializing men to use violence to solve their problems. Men commit the vast majority of violent crimes, and this is something that is true in nearly every culture globally, but with less violent societies, there is less social pressure for men to be violent. [ENDQ] [NEWLINE] Only one time in my life did I feel "socialized" or a pressured to be violent. A man at a bar said something rude to my then GF, she then wanted me to fight for her honor. So I guess maybe women are just trying to have the men be violent for them? [NEWLINE] [NEWLINE] [STARTQ] Is this another thing that you only feel, or do you have any actual evidence for this claim? [ENDQ] [NEWLINE] (personal bias incoming) [NEWLINE] [NEWLINE] This, this bit right here I disliked the most. Im a divorced, single father. My ex is on government assistance, by her being on government assistance she is legally entitled to not pay one dime in child support and I still get to pay her alimony(that is seriously the law, I get nothing and MUST pay alimony)! YAY fair and balanced womens lib groups!!! She has a job, she makes 5k/year less than I do (and has 0 child time) but because she applied and was granted government assistance (a womans only health policy offered by my state to any mother with a child under 8). She gets a pass on paying for her kid and I get the stick! (NO idea why some men would be offended by this! Us penis having asshats!) [NEWLINE] [NEWLINE] [STARTQ] I find it clear that you have no idea what feminism is. Feminism is not about "human rights" or "equality". Feminism is about defining what "human rights" or "equality" even mean. [ENDQ] [NEWLINE] Is this just a thing that only you feel? If you went to every crack pot on the world wide net your going to get a different definition for what it means. [NEWLINE] [NEWLINE] Meanwhile the Feminist movement is reactionary anti-male movement that is primarily about burning bras and letting out **emotions** rather than any sort of respect for intellectualism, academic rigor, or actual social change! [NEWLINE] [NEWLINE] [STARTQ] I don't know what you know about first wave feminism [ENDQ] [NEWLINE] Honestly zero, next to nothing. Can you surf on it? [NEWLINE] [NEWLINE] [STARTQ] By this
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you that the description of radical feminism above was somewhat watered-down.  Most of the description sounded like general feminism to me. <mask>, I have to take issue with some of the logic of your post. [NEWLINE] [NEWLINE] [STARTQ] "Men can choose not to take on that role and can choose not to propagate it." This implies beforehand that men are predisposed to oppressive and misogynist actions - and<mask> extremely sexist and offensive. [ENDQ] [NEWLINE] Yes.  Yes we are.  We *are* predisposed towards oppressive and misogynist actions.  Not through any fault of our own, not<mask> it is some inherent trait of being male,<mask><mask> we have been raised in a society<mask> we are taught sexism from day one -- not explicitly,<mask> through a collective attitude that expresses itself in a myriad of little ways. [NEWLINE] [NEWLINE] "Well, gee," you might think, "I'm not a sexist. <mask><mask> women are and should be equal to men." And that's probably true, on a rational and intellectual level.  It is and always has been for me. <mask><mask> sexism shows up most is on a more primitive level - responses and judgments we make on an automatic, instinctual, or emotional level.  I know I'm guilty of this. Even<mask> I notice it (which is definitely not a given) and am cognizent of the rational argument against whatever automatic reaction or judgment I'm making, it is still hard to overcome the conditioning I've acquired through years of constant exposure to the same. [NEWLINE] [NEWLINE] <mask> often have you judged women for sexual promiscuity that you would congratulate, or at least not judge, a man for? Have you ever thought of a woman<mask> a bitch instead of an asshole,<mask><mask> her being a jerk is somehow tied into her gender identity? Do you unquestioningly accept the overtly sexual nature of female characters in video games<mask> expecting diversity and depth in male characters? Have you ever unconsciously judged a woman's competence<mask> she doesn't project a traditional manly aura of strength that you wouldn't even need to observe in a man to avoid judging him the same way? There are<mask> many examples like this: most pretty minor, almost all unthinking,<mask><mask><mask><mask> they make up 90% of the problem.  Maybe a minority of men are wife-beaters and rapists,<mask> most of us exhibit this kind of unconscious behavior that reflect attitudes we don't even realize we really have. <mask> we aren't *thinking* about
Label encoding: <s>I agree with you that the description of radical feminism above was somewhat watered-down.  Most of the description sounded like general feminism to me.  However, I have to take issue with some of the logic of your post. [NEWLINE] [NEWLINE] [STARTQ] "Men can choose not to take on that role and can choose not to propagate it." This implies beforehand that men are predisposed to oppressive and misogynist actions - and therefore extremely sexist and offensive. [ENDQ] [NEWLINE] Yes.  Yes we are.  We *are* predisposed towards oppressive and misogynist actions.  Not through any fault of our own, not because it is some inherent trait of being male, but because we have been raised in a society where we are taught sexism from day one -- not explicitly, but through a collective attitude that expresses itself in a myriad of little ways. [NEWLINE] [NEWLINE] "Well, gee," you might think, "I'm not a sexist.  I think women are and should be equal to men." And that's probably true, on a rational and intellectual level.  It is and always has been for me.  But where sexism shows up most is on a more primitive level - responses and judgments we make on an automatic, instinctual, or emotional level.  I know I'm guilty of this. Even when I notice it (which is definitely not a given) and am cognizent of the rational argument against whatever automatic reaction or judgment I'm making, it is still hard to overcome the conditioning I've acquired through years of constant exposure to the same. [NEWLINE] [NEWLINE] How often have you judged women for sexual promiscuity that you would congratulate, or at least not judge, a man for? Have you ever thought of a woman as a bitch instead of an asshole, as if her being a jerk is somehow tied into her gender identity? Do you unquestioningly accept the overtly sexual nature of female characters in video games while expecting diversity and depth in male characters? Have you ever unconsciously judged a woman's competence because she doesn't project a traditional manly aura of strength that you wouldn't even need to observe in a man to avoid judging him the same way? There are so many examples like this: most pretty minor, almost all unthinking, but in my opinion they make up 90% of the problem.  Maybe a minority of men are wife-beaters and rapists, but most of us exhibit this kind of unconscious behavior that reflect attitudes we don't even realize we really have.  Because we aren't *thinking* about
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's say there are only two people on an island. They arrive there at the same time. Person A starts developing a little plot of land near a coconut tree. We don't know<mask> Person B is doing at this time. A few days after arriving at the island, Person B shows up at Person A's little plot of land. Person A built a small shelter, a trap to catch wild animals, and padding down near the base of the tree for the coconuts to fall on. Person B likes it, wants it, and declares himself the owner of this little plot of land that Person A spent days developing. Naturally, Person A disputes this and says that he is the exclusive owner. [NEWLINE] [NEWLINE] In this situation, there are only two members of society that have conflicting claims.<mask> a social contract is the thing that endows rights, who has the right to this property? I don't think social contract has an answer to this question. [NEWLINE] [NEWLINE] In the case of taxation in the US, I am certainly a minority in saying taxes, no matter<mask> slight, are a violation of my inherent rights<mask> a human being. It is reasonable to say that taxes being legitimate are part of the social contract,<mask><mask> I don't like it, tough. Leave the social contract. Presently, you have more or less a historical case for such a social contract (that is, most people think there is a social contract). Note:<mask><mask> it is *reasonable*<mask> not correct. [NEWLINE] [NEWLINE] The situation that I provided isn't<mask> clear. Based on all of the comments in this thread, a social contract "exists"<mask> there is a majority consensus in a society, and the terms of the social contract are whatever the consensus is. There is not a social contract in my situation,<mask> under your view, no rights have been granted to anyone.<mask> it is, both parties want only themselves to be the owner,<mask> there is nothing<mask> an impasse. [NEWLINE] [NEWLINE] In my scenario, who is correct in saying it is his? Under my worldview, it is simple: Person A owns the property<mask><mask><mask> many times Person B says it is<mask> Person A found the prescriptive use for it. [NEWLINE] [NEWLINE] Under your worldview, who owns it? Neither party?<mask> it is not owned by anyone, can someone claim it<mask> his? Can Person B legitimately attack Person A for it? [NEWLINE] [NEWLINE] [STARTQ] <mask> does him "using" something make it his? By that definition<mask> I take it from
Label encoding: <s>Let's say there are only two people on an island. They arrive there at the same time. Person A starts developing a little plot of land near a coconut tree. We don't know what Person B is doing at this time. A few days after arriving at the island, Person B shows up at Person A's little plot of land. Person A built a small shelter, a trap to catch wild animals, and padding down near the base of the tree for the coconuts to fall on. Person B likes it, wants it, and declares himself the owner of this little plot of land that Person A spent days developing. Naturally, Person A disputes this and says that he is the exclusive owner. [NEWLINE] [NEWLINE] In this situation, there are only two members of society that have conflicting claims. If a social contract is the thing that endows rights, who has the right to this property? I don't think social contract has an answer to this question. [NEWLINE] [NEWLINE] In the case of taxation in the US, I am certainly a minority in saying taxes, no matter how slight, are a violation of my inherent rights as a human being. It is reasonable to say that taxes being legitimate are part of the social contract, therefore if I don't like it, tough. Leave the social contract. Presently, you have more or less a historical case for such a social contract (that is, most people think there is a social contract). Note: I think it is *reasonable* but not correct. [NEWLINE] [NEWLINE] The situation that I provided isn't as clear. Based on all of the comments in this thread, a social contract "exists" if there is a majority consensus in a society, and the terms of the social contract are whatever the consensus is. There is not a social contract in my situation, therefore under your view, no rights have been granted to anyone. As it is, both parties want only themselves to be the owner, so there is nothing but an impasse. [NEWLINE] [NEWLINE] In my scenario, who is correct in saying it is his? Under my worldview, it is simple: Person A owns the property regardless of how many times Person B says it is because Person A found the prescriptive use for it. [NEWLINE] [NEWLINE] Under your worldview, who owns it? Neither party? If it is not owned by anyone, can someone claim it as his? Can Person B legitimately attack Person A for it? [NEWLINE] [NEWLINE] [STARTQ] Why does him "using" something make it his? By that definition if I take it from
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That's literally like<mask> you and I were having a fight, I vaporise your parents<mask> we don't have to get any more cuts and bruises. [ENDQ] [NEWLINE] Cuts and bruises? [Not exactly.]( [URL] #Estimated_casualties) [NEWLINE] [NEWLINE] [STARTQ] <mask> the U.S. military planners assumed "that operations in this area will be opposed not only by the available organized military forces of the Empire,<mask><mask> by a fanatically hostile population",[11] high casualties were thought to be inevitable,<mask> nobody knew with certainty<mask> high. Several people made estimates,<mask> they varied widely in numbers, assumptions, and purposes, which included advocating for and against the invasion. Afterwards, they were reused in the debate over the atomic bombings of Hiroshima and Nagasaki. [ENDQ] [NEWLINE] [STARTQ] Casualty estimates were based on the experience of the preceding campaigns, drawing different lessons: [ENDQ] [NEWLINE] [STARTQ] *    In a letter sent to Gen. Curtis LeMay from Gen. Lauris Norstad,<mask> LeMay assumed command of the B-29 force on Guam, Norstad told LeMay that<mask> an invasion took place, it would cost the US "half a million" dead.[52] [ENDQ] [NEWLINE] [STARTQ] *    In a study done by the Joint Chiefs of Staff in April, the figures of 7.45 casualties/1,000 man-days and 1.78 fatalities/1,000 man-days were developed. This implied that a 90-day Olympic campaign would cost 456,000 casualties, including 109,000 dead or missing.<mask> Coronet took another 90 days, the combined cost would be 1,200,000 casualties, with 267,000 fatalities.[53] [ENDQ] [NEWLINE] [STARTQ] *    A study done by Adm. Nimitz's staff in May estimated 49,000 U.S casualties in the first 30 days, including 5,000 at sea.[54] A study done by General MacArthur's staff in June estimated 23,000 US casualties in the first 30 days and 125,000 after 120 days.[55]<mask> these figures were questioned by General Marshall, MacArthur submitted a revised estimate of 105,000, in part by deducting wounded men able to return to duty.[56] [ENDQ] [NEWLINE] [STARTQ] *    In a conference with President Truman on June 18, Marshall, taking the Battle of Luzon<mask> the best model for Olympic, thought the Americans would suffer 31,000 casualties in the first 30 days (and ultimately 20% of Japanese casualties, which implied
Label encoding: <s> [STARTQ] That's literally like if you and I were having a fight, I vaporise your parents so we don't have to get any more cuts and bruises. [ENDQ] [NEWLINE] Cuts and bruises? [Not exactly.]( [URL] #Estimated_casualties) [NEWLINE] [NEWLINE] [STARTQ] Because the U.S. military planners assumed "that operations in this area will be opposed not only by the available organized military forces of the Empire, but also by a fanatically hostile population",[11] high casualties were thought to be inevitable, but nobody knew with certainty how high. Several people made estimates, but they varied widely in numbers, assumptions, and purposes, which included advocating for and against the invasion. Afterwards, they were reused in the debate over the atomic bombings of Hiroshima and Nagasaki. [ENDQ] [NEWLINE] [STARTQ] Casualty estimates were based on the experience of the preceding campaigns, drawing different lessons: [ENDQ] [NEWLINE] [STARTQ] *    In a letter sent to Gen. Curtis LeMay from Gen. Lauris Norstad, when LeMay assumed command of the B-29 force on Guam, Norstad told LeMay that if an invasion took place, it would cost the US "half a million" dead.[52] [ENDQ] [NEWLINE] [STARTQ] *    In a study done by the Joint Chiefs of Staff in April, the figures of 7.45 casualties/1,000 man-days and 1.78 fatalities/1,000 man-days were developed. This implied that a 90-day Olympic campaign would cost 456,000 casualties, including 109,000 dead or missing. If Coronet took another 90 days, the combined cost would be 1,200,000 casualties, with 267,000 fatalities.[53] [ENDQ] [NEWLINE] [STARTQ] *    A study done by Adm. Nimitz's staff in May estimated 49,000 U.S casualties in the first 30 days, including 5,000 at sea.[54] A study done by General MacArthur's staff in June estimated 23,000 US casualties in the first 30 days and 125,000 after 120 days.[55] When these figures were questioned by General Marshall, MacArthur submitted a revised estimate of 105,000, in part by deducting wounded men able to return to duty.[56] [ENDQ] [NEWLINE] [STARTQ] *    In a conference with President Truman on June 18, Marshall, taking the Battle of Luzon as the best model for Olympic, thought the Americans would suffer 31,000 casualties in the first 30 days (and ultimately 20% of Japanese casualties, which implied
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A farmer instead has incentive to make tomatoes that last longer, and can be grown year round. Making more beneficial tomatoes for society is asking farmers to sacrifice profits. [ENDQ] [NEWLINE] Error #1, assumption of benefit. You assume that a "good tasting, nutritional" tomato is<mask> is of most social value. Tomatoes that last longer can reach more people, poison them less, and make more efficient use of food (rather than throw out a larger percentage). Having them grown year round means there are tomatoes available year round. Even *<mask> * some reduced nutrition, it is better nutrition than not eating them at all<mask> they aren't available. Mass production of tomatoes likewise cuts costs and makes tomatoes more affordable<mask> that the poor can eat them (and improve nutrition), and everybody else can afford to eat more of them. [NEWLINE] [NEWLINE] The rest of your points have similar problems,<mask> I'll refrain from addressing them one by one. It appears to me that you've just never thought through the tradeoffs and net social value. You see the item itself in a bubble and don't consider its cost, its value, access to it, or the industry on the whole. [NEWLINE] [NEWLINE] One variation that I will address is the "ethically reasonable" meat production comment. Yes, you are correct that flat out profit aims for pure efficiency and may lead to unethical behaviour.<mask>, that is not a problem with profit motivation or capitalism in general; it is a problem with the costs and benefits tied to the activity. The solution tends to take the form of either collective regulation (via our collective government), or otherwise tying costs to the problems themselves (known<mask> externalities in economics). For example, carbon tax adds the cost of cleaning the environment to each transaction that generates the carbon that results in the need to clean it up. [NEWLINE] [NEWLINE] Based on your title too, you imply that there are other systems that would do better to help things "reach full potential". To be clear, the trend towards maximizing efficiency in lieu of other things is an inherent mathematical property. Natural selection works on it even. Issues like the Prisoners Dilemma and Tragedy of the Commons are not products of a given economic system,<mask> are fundamental properties of systems with social transactions between multiple stakeholders. It matters not<mask> economic system you use, they will always exist. [NEWLINE] [NEWLINE] For example, imagine a system that encourages you to spend all day making one really good tomato,<mask> tasty and nutritious<mask> you can imagine. OK,<mask><mask> do you survive? You could eat it
Label encoding: <s> [STARTQ] A farmer instead has incentive to make tomatoes that last longer, and can be grown year round. Making more beneficial tomatoes for society is asking farmers to sacrifice profits. [ENDQ] [NEWLINE] Error #1, assumption of benefit. You assume that a "good tasting, nutritional" tomato is what is of most social value. Tomatoes that last longer can reach more people, poison them less, and make more efficient use of food (rather than throw out a larger percentage). Having them grown year round means there are tomatoes available year round. Even * if * some reduced nutrition, it is better nutrition than not eating them at all because they aren't available. Mass production of tomatoes likewise cuts costs and makes tomatoes more affordable so that the poor can eat them (and improve nutrition), and everybody else can afford to eat more of them. [NEWLINE] [NEWLINE] The rest of your points have similar problems, so I'll refrain from addressing them one by one. It appears to me that you've just never thought through the tradeoffs and net social value. You see the item itself in a bubble and don't consider its cost, its value, access to it, or the industry on the whole. [NEWLINE] [NEWLINE] One variation that I will address is the "ethically reasonable" meat production comment. Yes, you are correct that flat out profit aims for pure efficiency and may lead to unethical behaviour. However, that is not a problem with profit motivation or capitalism in general; it is a problem with the costs and benefits tied to the activity. The solution tends to take the form of either collective regulation (via our collective government), or otherwise tying costs to the problems themselves (known as externalities in economics). For example, carbon tax adds the cost of cleaning the environment to each transaction that generates the carbon that results in the need to clean it up. [NEWLINE] [NEWLINE] Based on your title too, you imply that there are other systems that would do better to help things "reach full potential". To be clear, the trend towards maximizing efficiency in lieu of other things is an inherent mathematical property. Natural selection works on it even. Issues like the Prisoners Dilemma and Tragedy of the Commons are not products of a given economic system, but are fundamental properties of systems with social transactions between multiple stakeholders. It matters not what economic system you use, they will always exist. [NEWLINE] [NEWLINE] For example, imagine a system that encourages you to spend all day making one really good tomato, as tasty and nutritious as you can imagine. OK, so how do you survive? You could eat it
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't know exactly<mask> smooth the process was for you to establish a sense of trust. [ENDQ] [NEWLINE] Well, this part is kind of weird. I came in<mask> a cynical person with trust issues,<mask> I had no problems with establishing a connection of trust with him.<mask> I said, I came in there on my own free will. I always entered the room with this idea: "You can read me like a book,<mask> I want to get over with it<mask>ap." Makes the job easier for both of you. Depends on<mask> deep you have fallen, I guess...<mask> you really have the need to change a situation, it goes a lot easier. [NEWLINE] [NEWLINE] [STARTQ] you can never be sure<mask> your playing that chess game just with yourself [ENDQ] [NEWLINE] I had that problem a lot and the psy asked me to say everything I was thinking out loud. It makes it for him clearer<mask> the points of interest are and in the process, I found answers a lot easier for some reason. He would write them down and share them with his team later on to decide<mask> would be the next step. [NEWLINE] [NEWLINE] [STARTQ] Now this is just one way this can play out it depends on the therapist, and he's only human,<mask> you see<mask> things can get out of hand, and the worst part is<mask> he is more intelligent than you, you may not even realize it. [ENDQ] [NEWLINE] Hold on there, that's one problem I had<mask> well: making predictions. You are making predictions, possible outcomes of<mask> could happen in a therapy session. [NEWLINE] Everything is possible<mask> you can not predict<mask> will happen. No one can. [NEWLINE] [NEWLINE] Problem is that you are making these to self fulfilling prophecy. Every sign you get, you will try to link to one of your predictions and before you know it, you end up running away<mask> you are convinced that your psy is a murderer (A very extreme example,<mask> you get the point). Now that's a result of the trust issues... Short answer: you can't know<mask> will happen. It will only be the source of a lot of frustration, missed opportunities, guilt,... [NEWLINE] [NEWLINE] [STARTQ] I'm fairly certain you know<mask> this is, it's called hypnosis. [ENDQ] [NEWLINE] I'm pretty sure you are aware<mask> your mind starts numbing down. And hypnosis works only on people who are receptive to hypnosis. [NEWLINE] I<mask> made that scenario in my head and made some rules: [NEWLINE] Hypnosis should be announced in advance and the session should be recorded on
Label encoding: <s> [STARTQ] I don't know exactly how smooth the process was for you to establish a sense of trust. [ENDQ] [NEWLINE] Well, this part is kind of weird. I came in as a cynical person with trust issues, though I had no problems with establishing a connection of trust with him. As I said, I came in there on my own free will. I always entered the room with this idea: "You can read me like a book, because I want to get over with it asap." Makes the job easier for both of you. Depends on how deep you have fallen, I guess... If you really have the need to change a situation, it goes a lot easier. [NEWLINE] [NEWLINE] [STARTQ] you can never be sure when your playing that chess game just with yourself [ENDQ] [NEWLINE] I had that problem a lot and the psy asked me to say everything I was thinking out loud. It makes it for him clearer where the points of interest are and in the process, I found answers a lot easier for some reason. He would write them down and share them with his team later on to decide what would be the next step. [NEWLINE] [NEWLINE] [STARTQ] Now this is just one way this can play out it depends on the therapist, and he's only human, so you see how things can get out of hand, and the worst part is if he is more intelligent than you, you may not even realize it. [ENDQ] [NEWLINE] Hold on there, that's one problem I had as well: making predictions. You are making predictions, possible outcomes of what could happen in a therapy session. [NEWLINE] Everything is possible as you can not predict what will happen. No one can. [NEWLINE] [NEWLINE] Problem is that you are making these to self fulfilling prophecy. Every sign you get, you will try to link to one of your predictions and before you know it, you end up running away because you are convinced that your psy is a murderer (A very extreme example, but you get the point). Now that's a result of the trust issues... Short answer: you can't know what will happen. It will only be the source of a lot of frustration, missed opportunities, guilt,... [NEWLINE] [NEWLINE] [STARTQ] I'm fairly certain you know what this is, it's called hypnosis. [ENDQ] [NEWLINE] I'm pretty sure you are aware when your mind starts numbing down. And hypnosis works only on people who are receptive to hypnosis. [NEWLINE] I also made that scenario in my head and made some rules: [NEWLINE] Hypnosis should be announced in advance and the session should be recorded on
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You're<mask><mask> a large market of people partaking in something means they're enjoying it. [ENDQ] [NEWLINE] Yes. Barring *physical* addiction people regularly consuming something indicates that it brings them utility - which (in the case of entertainment) could be quantified<mask> *fun*.<mask><mask> we assume that there isn't a heroine-esque biological response associated with going to a loud-music-bar then we should accept that people do it<mask> they like it. [NEWLINE] [NEWLINE] Your example of a used car is not valid<mask> the once in 10 years overestimation of the "value" of a new vs. used car is not the same<mask> going to a loud-music-bar 30 times a year for multiple years in your 20s and maybe early 30s. (<mask> an aside just<mask> the resale value of a care diminishes rapidly doesn't mean that there aren't actual benefits to buying new vs used - not everyone who buys new does it<mask> they think "<mask> fun!"). [NEWLINE] [NEWLINE] [STARTQ] <mask> to me it seems that people get dragged along to the establishments that exist<mask> it's something to do,<mask> I don't perceive that anyone is enjoying themselves<mask> I've never heard anyone say that they were. [ENDQ] [NEWLINE] <mask> no one likes these places who is doing the dragging.<mask> would someone agree to be repeatedly taken to a place they actively dislike? [NEWLINE] [NEWLINE] [STARTQ] I've never been to a club or heard of dress codes existing for them [ENDQ] [NEWLINE] Clubs, like all types of drinking establishments, come in many forms. They are often stigmatized<mask> being more "high maintenance" than a bar<mask> people expect more expensive clothing, bottle/table service, lines at the door, and possible drug culture. This isn't true for all (or maybe even most) clubs<mask> it is the impression that they have and a reason many people prefer to go to loud-music-bars<mask>,<mask><mask>, a night club could provide a more functional venue for that type of enjoyment. [NEWLINE] [NEWLINE] [STARTQ] a bar is not inexpensive versus drinking at home. [ENDQ] [NEWLINE] $6 beers and mixed drinks &lt; $15 "cocktails" and $500 table service. [NEWLINE] [NEWLINE] [STARTQ] <mask> could one dance in a bar? [ENDQ] [NEWLINE] Extremely easily. [NEWLINE] [NEWLINE] [STARTQ] There's no room for it and I've never seen it done. [ENDQ] [NEWLINE] You appear to be arguing from a place of ignorance. Dancing is pretty fundamental to loud-music-bars. You've indicated that you do not frequent sports bars or clubs and I
Label encoding: <s> [STARTQ] You're assuming that a large market of people partaking in something means they're enjoying it. [ENDQ] [NEWLINE] Yes. Barring *physical* addiction people regularly consuming something indicates that it brings them utility - which (in the case of entertainment) could be quantified as *fun*. So if we assume that there isn't a heroine-esque biological response associated with going to a loud-music-bar then we should accept that people do it because they like it. [NEWLINE] [NEWLINE] Your example of a used car is not valid because the once in 10 years overestimation of the "value" of a new vs. used car is not the same as going to a loud-music-bar 30 times a year for multiple years in your 20s and maybe early 30s. ( As an aside just because the resale value of a care diminishes rapidly doesn't mean that there aren't actual benefits to buying new vs used - not everyone who buys new does it because they think " what fun!"). [NEWLINE] [NEWLINE] [STARTQ] So to me it seems that people get dragged along to the establishments that exist because it's something to do, but I don't perceive that anyone is enjoying themselves because I've never heard anyone say that they were. [ENDQ] [NEWLINE] If no one likes these places who is doing the dragging. Why would someone agree to be repeatedly taken to a place they actively dislike? [NEWLINE] [NEWLINE] [STARTQ] I've never been to a club or heard of dress codes existing for them [ENDQ] [NEWLINE] Clubs, like all types of drinking establishments, come in many forms. They are often stigmatized as being more "high maintenance" than a bar as people expect more expensive clothing, bottle/table service, lines at the door, and possible drug culture. This isn't true for all (or maybe even most) clubs but it is the impression that they have and a reason many people prefer to go to loud-music-bars when, in fact, a night club could provide a more functional venue for that type of enjoyment. [NEWLINE] [NEWLINE] [STARTQ] a bar is not inexpensive versus drinking at home. [ENDQ] [NEWLINE] $6 beers and mixed drinks &lt; $15 "cocktails" and $500 table service. [NEWLINE] [NEWLINE] [STARTQ] How could one dance in a bar? [ENDQ] [NEWLINE] Extremely easily. [NEWLINE] [NEWLINE] [STARTQ] There's no room for it and I've never seen it done. [ENDQ] [NEWLINE] You appear to be arguing from a place of ignorance. Dancing is pretty fundamental to loud-music-bars. You've indicated that you do not frequent sports bars or clubs and I
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, I'm sorry for<mask> happened to you and I would hope that that never happens to anybody. [NEWLINE] [NEWLINE] [STARTQ] I've always thought that people who use violence and fight to settle arguments are really just too stupid to win an argument with words.<mask><mask><mask> upbringing, in all of my 19 years I have never punched or kicked a person, even in self defense. [ENDQ] [NEWLINE] On the first part you are mostly right. I would say that<mask> I walked into a bar and there was somebody fighting, I would make the same judgement of the both of them.<mask>, self defense is very important and<mask><mask> that to not defend yourself is ridiculous, especially<mask> it is a life or death situation. Sure it can be better to run at times,<mask><mask> you don't have a choice, you need to go for it, you can't let your past control whether or not you live in a life or death situation. [NEWLINE] [NEWLINE] [STARTQ] I completely negate that there is intelligence needed for these fighting sports [ENDQ] [NEWLINE] This is wrong.<mask> there are plenty of'stupid' fighters, they are not stupid people<mask> it comes to fighting. Fights, especially<mask> you progress up and up involve a lot of strategy. It is basically a game of wills, and the person who can keep going<mask> the other person quits will win. Technique takes forever to learn and even longer to master. They might be dumb in terms of book smarts,<mask> they are definitely not dumb. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask><mask> that being the best fighter is a talent, or at least a talent that should be endorsed by society and money. [ENDQ] [NEWLINE] You can think that all you want,<mask><mask> it comes down to is whether or not society thinks it should be.<mask> the rest of us agreed with you, they wouldn't exist. It exists solely<mask> people want to support it. [NEWLINE] [NEWLINE] [STARTQ] I realize that I am not in the majority, and everytime I bring up any of these views people basically call me a wimp and can never get a serious answer out of anyone<mask> to<mask> these sports are beneficial to society. [ENDQ] [NEWLINE] You are not a wimp for saying you don't like them at all and I perfectly understand<mask> you are coming from. I don't like boxing<mask><mask><mask> it's pointless,<mask> I like MMA<mask> it has more dimensions to it and<mask><mask> they are more skilled. These sports are beneficial to society<mask> they provide entertainment to a lot of people, plain and simple. I don
Label encoding: <s>First of all, I'm sorry for what happened to you and I would hope that that never happens to anybody. [NEWLINE] [NEWLINE] [STARTQ] I've always thought that people who use violence and fight to settle arguments are really just too stupid to win an argument with words. Because of this upbringing, in all of my 19 years I have never punched or kicked a person, even in self defense. [ENDQ] [NEWLINE] On the first part you are mostly right. I would say that if I walked into a bar and there was somebody fighting, I would make the same judgement of the both of them. However, self defense is very important and I think that to not defend yourself is ridiculous, especially when it is a life or death situation. Sure it can be better to run at times, but if you don't have a choice, you need to go for it, you can't let your past control whether or not you live in a life or death situation. [NEWLINE] [NEWLINE] [STARTQ] I completely negate that there is intelligence needed for these fighting sports [ENDQ] [NEWLINE] This is wrong. Although there are plenty of'stupid' fighters, they are not stupid people when it comes to fighting. Fights, especially as you progress up and up involve a lot of strategy. It is basically a game of wills, and the person who can keep going when the other person quits will win. Technique takes forever to learn and even longer to master. They might be dumb in terms of book smarts, but they are definitely not dumb. [NEWLINE] [NEWLINE] [STARTQ] I do not think that being the best fighter is a talent, or at least a talent that should be endorsed by society and money. [ENDQ] [NEWLINE] You can think that all you want, but what it comes down to is whether or not society thinks it should be. If the rest of us agreed with you, they wouldn't exist. It exists solely because people want to support it. [NEWLINE] [NEWLINE] [STARTQ] I realize that I am not in the majority, and everytime I bring up any of these views people basically call me a wimp and can never get a serious answer out of anyone as to why these sports are beneficial to society. [ENDQ] [NEWLINE] You are not a wimp for saying you don't like them at all and I perfectly understand where you are coming from. I don't like boxing because I think it's pointless, but I like MMA because it has more dimensions to it and I think they are more skilled. These sports are beneficial to society because they provide entertainment to a lot of people, plain and simple. I don
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This CMV is not about whether you or I believe aliens exist. It is about the widely held point of view:<mask> you don’t think alien life exists elsewhere in the universe, you’re crazy. I understand<mask> this view comes from; the [Universe is<mask> unimaginably huge]( [URL] ) and<mask> filled with [galaxies]( [URL] ), and<mask> stars, and<mask> [planets]( [URL] ), there simply *must* be other life out there. My view is that this approach is fundamentally unscientific and should not be held<mask> it is based on probability and not actual scientific evidence. [NEWLINE] [NEWLINE] My background: I am a physics and astronomy teacher, and I teach about this subject in depth every year.<mask><mask> about it and research it often. I’m well acquainted with the [Drake Equation]( [URL] /), the [Fermi Paradox]( [URL] ) and [its many possible solutions]( [URL] -uI), the [Great Filter]( [URL] ), and the debate over alien life in general. [NEWLINE] [NEWLINE] My own belief in the matter used to be strongly on the ‘of course aliens exist out there’ side,<mask> I thought the distances were just too vast for us to ever observe them.<mask>,<mask><mask> was strongly shifted by the book *Alone in the Universe* by [John Gribbin]( [URL] ), and I now think that the great profusion of life here on Earth is<mask> rare that we are alone in the Universe. [NEWLINE] [NEWLINE] Don’t get me wrong, I’m completely open to the idea of aliens existing. I kind of hope they do,<mask><mask><mask> they don’t destroy us all!<mask> until we get any kind of scientific evidence that they exist — an organized signal, clear alien-made trace elements on a planet’s spectroscopy, anything measurable —<mask><mask> the correct scientific approach is that they don’t exist. CMV. [NEWLINE] [NEWLINE] Edit 1: Some good thoughts in here, thanks.<mask>, some incorrect assumptions about<mask> I'm saying. Probability is of course a useful scientific tool, the key to our understanding of quantum mechanics.<mask> guesses about the Drake Equation boil down to probability based on no data, quite different from the data-based probability of QM. [NEWLINE] [NEWLINE] The most compelling argument that I've read below is that<mask> we know life happened once in the Universe with us, it could happen again. Physical laws of symmetry point to the idea
Label encoding: <s>This CMV is not about whether you or I believe aliens exist. It is about the widely held point of view: if you don’t think alien life exists elsewhere in the universe, you’re crazy. I understand where this view comes from; the [Universe is so unimaginably huge]( [URL] ) and so filled with [galaxies]( [URL] ), and therefore stars, and therefore [planets]( [URL] ), there simply *must* be other life out there. My view is that this approach is fundamentally unscientific and should not be held because it is based on probability and not actual scientific evidence. [NEWLINE] [NEWLINE] My background: I am a physics and astronomy teacher, and I teach about this subject in depth every year. I think about it and research it often. I’m well acquainted with the [Drake Equation]( [URL] /), the [Fermi Paradox]( [URL] ) and [its many possible solutions]( [URL] -uI), the [Great Filter]( [URL] ), and the debate over alien life in general. [NEWLINE] [NEWLINE] My own belief in the matter used to be strongly on the ‘of course aliens exist out there’ side, but I thought the distances were just too vast for us to ever observe them. However, my opinion was strongly shifted by the book *Alone in the Universe* by [John Gribbin]( [URL] ), and I now think that the great profusion of life here on Earth is so rare that we are alone in the Universe. [NEWLINE] [NEWLINE] Don’t get me wrong, I’m completely open to the idea of aliens existing. I kind of hope they do, as long as they don’t destroy us all! But until we get any kind of scientific evidence that they exist — an organized signal, clear alien-made trace elements on a planet’s spectroscopy, anything measurable — I think the correct scientific approach is that they don’t exist. CMV. [NEWLINE] [NEWLINE] Edit 1: Some good thoughts in here, thanks. Also, some incorrect assumptions about what I'm saying. Probability is of course a useful scientific tool, the key to our understanding of quantum mechanics. But guesses about the Drake Equation boil down to probability based on no data, quite different from the data-based probability of QM. [NEWLINE] [NEWLINE] The most compelling argument that I've read below is that because we know life happened once in the Universe with us, it could happen again. Physical laws of symmetry point to the idea
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a lot of fun at the expense of people with two last names. This is a tough one,<mask> I do believe a woman shouldn't be socially pressured into taking her's husband's last name, and gay spouses<mask> have a conundrum of their own. [NEWLINE] [NEWLINE] <mask>,<mask><mask> that a family should have a family name,<mask> otherwise you are choosing to make everyone else's life difficult. One small example is<mask> you work in an industry<mask> you deal with a lot of names and families and meeting people blind (like I do). It's helpful to know in advance who is related to who by just looking at a piece of paper. [NEWLINE] [NEWLINE] <mask>,<mask> the wife and the husband choose not to share a last name, please God settle on one name for your kid. Just flip a coin for which last name your kid takes<mask> that's<mask> you need to do. Don't make everyone around him have to deal with saying and writing out Reginald Henry Lieberman-Montgomery. [NEWLINE] [NEWLINE] A metaphor I would make for this would be gender pronouns. You have the right to tell me whether you identify<mask> a man or a woman,<mask> I swear to God<mask> you insist that everyone around you refer to you strictly with the gender neutral pronouns Ni, Nem and Nir, I will be having none of it. [NEWLINE] [NEWLINE] <mask> a counterpoint, I will say that I can't ever imagine myself taking my wife's last name. And<mask> she chose not to take mine,<mask><mask> I would feel a little disappointed. Further, I can't imagine feeling comfortable with my kid not taking my last name.<mask> I already recognize a bit of the hypocrisy. In that way, I see that I am posing a problem without an apparent solution. On top of your rebuttals, I would<mask> be curious to hear alternative ideas/social structures that could work to level the playing field, AND prevent me from having to say a mouthful. [NEWLINE] [NEWLINE] **EDIT:** [NEWLINE] [NEWLINE] Apologies for the day long delay-- car emergency took all night and all morning. [NEWLINE] [NEWLINE] Thanks for all your responses! Here are some of my views you've changed. [NEWLINE] [NEWLINE] Spain [STARTQ] 'Murica. There's a clear system in place in Spanish countries that actually speaks to my "one family name" point. People consistently take the fathers, then mother's surname. Then the kid's kids will do the same. Gives you family history and<mask> something to put on your mailbox. [ENDQ] [NEWLINE] <mask> I personally like about this is
Label encoding: <s>I have a lot of fun at the expense of people with two last names. This is a tough one, because I do believe a woman shouldn't be socially pressured into taking her's husband's last name, and gay spouses also have a conundrum of their own. [NEWLINE] [NEWLINE] However, I think that a family should have a family name, because otherwise you are choosing to make everyone else's life difficult. One small example is if you work in an industry where you deal with a lot of names and families and meeting people blind (like I do). It's helpful to know in advance who is related to who by just looking at a piece of paper. [NEWLINE] [NEWLINE] BUT, if the wife and the husband choose not to share a last name, please God settle on one name for your kid. Just flip a coin for which last name your kid takes if that's what you need to do. Don't make everyone around him have to deal with saying and writing out Reginald Henry Lieberman-Montgomery. [NEWLINE] [NEWLINE] A metaphor I would make for this would be gender pronouns. You have the right to tell me whether you identify as a man or a woman, but I swear to God if you insist that everyone around you refer to you strictly with the gender neutral pronouns Ni, Nem and Nir, I will be having none of it. [NEWLINE] [NEWLINE] As a counterpoint, I will say that I can't ever imagine myself taking my wife's last name. And if she chose not to take mine, I think I would feel a little disappointed. Further, I can't imagine feeling comfortable with my kid not taking my last name. So I already recognize a bit of the hypocrisy. In that way, I see that I am posing a problem without an apparent solution. On top of your rebuttals, I would also be curious to hear alternative ideas/social structures that could work to level the playing field, AND prevent me from having to say a mouthful. [NEWLINE] [NEWLINE] **EDIT:** [NEWLINE] [NEWLINE] Apologies for the day long delay-- car emergency took all night and all morning. [NEWLINE] [NEWLINE] Thanks for all your responses! Here are some of my views you've changed. [NEWLINE] [NEWLINE] Spain [STARTQ] 'Murica. There's a clear system in place in Spanish countries that actually speaks to my "one family name" point. People consistently take the fathers, then mother's surname. Then the kid's kids will do the same. Gives you family history and also something to put on your mailbox. [ENDQ] [NEWLINE] What I personally like about this is
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here in Israel, we are taught "tanach" (Hebrew for Bible, which is obviously the first testament) and literature (short stories, songs, poems, novels, etc.) in highschool. [NEWLINE] [NEWLINE] They are part of a few required subjects along with English, math, history and Hebrew.   I don't think literature and Bible should be in the same group of required subjects and<mask><mask> they should be completely optional altogether. [NEWLINE] [NEWLINE] It's obvious<mask> to<mask> English, Hebrew (the countries main language) history and math are objectively important. [NEWLINE] Math is a subject used daily.  Kids will always use the "I don't need this in life" argument<mask><mask><mask> with that<mask> it comes to these 4 subjects.  Math is insanely useful and even the super long equations are,<mask> they develop your thinking and your thought process and all that stuff.  It's got obvious benefits, aside from the fact that with 5 "units" of math, the highest you can do here in high-school, you can get pretty much anywhere. [NEWLINE] [NEWLINE] English is another obvious one, it's the language of the world. You can't leave the country without English. Everyone on earth should have some basic English speaking ability. [NEWLINE] [NEWLINE] Hebrew is even more obvious, I won't even go into that one. [NEWLINE] [NEWLINE] Finally, history. History is deserving of being a require subject,<mask><mask>,<mask> it's literally just a bunch of facts that lead to<mask> we are today.  It's simply the shit that happened before I was here.  It's important. It comes up in every day conversations all the time.  It's important that we all never stop leaving history<mask> that it doesn't repeat. History, put simply is the things that happened.  You can't get any more important than that, in terms of<mask> should be taught<mask><mask>. [NEWLINE] [NEWLINE] [NEWLINE] And then there's literature and Bible.  Two subjects made up of pure fictional nonsense.  I mean, I guess it depends on your beliefs<mask> it comes to Bible<mask><mask><mask> 95% of people, even in Israel, can agree that seas splitting and bushes burning never actually happened. and even<mask> most people do think it happened, its<mask> of bible being taught in public highschool. we know that seas cannot be split, this is fact,<mask><mask> is it being taught<mask> fact that they *can* be split? thats<mask> bad<mask> telling kids that climate change isn't happening. its a flat out lie
Label encoding: <s>Here in Israel, we are taught "tanach" (Hebrew for Bible, which is obviously the first testament) and literature (short stories, songs, poems, novels, etc.) in highschool. [NEWLINE] [NEWLINE] They are part of a few required subjects along with English, math, history and Hebrew.   I don't think literature and Bible should be in the same group of required subjects and I think they should be completely optional altogether. [NEWLINE] [NEWLINE] It's obvious as to why English, Hebrew (the countries main language) history and math are objectively important. [NEWLINE] Math is a subject used daily.  Kids will always use the "I don't need this in life" argument but I disagree with that when it comes to these 4 subjects.  Math is insanely useful and even the super long equations are, because they develop your thinking and your thought process and all that stuff.  It's got obvious benefits, aside from the fact that with 5 "units" of math, the highest you can do here in high-school, you can get pretty much anywhere. [NEWLINE] [NEWLINE] English is another obvious one, it's the language of the world. You can't leave the country without English. Everyone on earth should have some basic English speaking ability. [NEWLINE] [NEWLINE] Hebrew is even more obvious, I won't even go into that one. [NEWLINE] [NEWLINE] Finally, history. History is deserving of being a require subject, IMO, because it's literally just a bunch of facts that lead to where we are today.  It's simply the shit that happened before I was here.  It's important. It comes up in every day conversations all the time.  It's important that we all never stop leaving history so that it doesn't repeat. History, put simply is the things that happened.  You can't get any more important than that, in terms of what should be taught IMO. [NEWLINE] [NEWLINE] [NEWLINE] And then there's literature and Bible.  Two subjects made up of pure fictional nonsense.  I mean, I guess it depends on your beliefs when it comes to Bible but I think 95% of people, even in Israel, can agree that seas splitting and bushes burning never actually happened. and even if most people do think it happened, its because of bible being taught in public highschool. we know that seas cannot be split, this is fact, so why is it being taught as fact that they *can* be split? thats as bad as telling kids that climate change isn't happening. its a flat out lie
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That is<mask> the majority of bicyclists seem to be acting<mask> it's just not true. [ENDQ] [NEWLINE] Do you have any statistics or studies to back this up? Or are you just acting off [emotion]( [URL] ) and<mask> appears to be a bias and prejudice towards a certain user of the road? [NEWLINE] [NEWLINE] [STARTQ] You and your broken spoke endanger everyone around you. Suppose your bike breaks and throws you into the street. The car behind you swerves to avoid running you over and you crashes into another car. YOU and your faulty bike just caused an accident.<mask> does the car driver's insurance have to pay for that? It was your fault. [ENDQ] [NEWLINE] Suppose your car has a loose hubcap just waiting to fall off, or a bumper duct taped on about to fall off. [NEWLINE] Suppose your truck with a load in the bed that's not properly secured and risks falling off? [NEWLINE] Suppose your car has a blowout. [NEWLINE] Suppose your car hits a pothole. [NEWLINE] Suppose a child (or wildlife) hops out in the street in front of you. [NEWLINE] [NEWLINE] <mask> any of this happens and you're in your car, big fucking deal. You're surrounded by an entire car body, a seat belt (hopefully), probably an air bag.<mask> happens to the cyclist in these situations? They're probably seriously injured or dead whether it's their fault or not. I'm under the impression that you're valuing your car's condition and your ability to travel a road bike/frustration free<mask> more important than the life and well being of another human being. You should CMV on that. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Or you are riding down a street full of parked cars and you get too close and knock off a parked cars mirror assembly, or scratch all down the side of the car, or dent it. Thousands of dollars in body work that YOU caused.<mask> should the car owner's insurance have to pay that? Or the owner the deductible? Or his rates go up<mask> of an accident YOU caused? Or have to hunt you down to go through the headache and expense to sue you in small claims court? [ENDQ] [NEWLINE] Do you have a single instance<mask> a cyclist's actions caused **thousands** of dollars in body work?? [NEWLINE] Driving/riding/walking, etc... involves risk of some sort. [NEWLINE] We could play this "<mask> -<mask> " game all day and it wouldn't get us anywhere. [NEWLINE] <mask> happens<mask> you park your car crooked in a parking a
Label encoding: <s> [STARTQ] That is how the majority of bicyclists seem to be acting but it's just not true. [ENDQ] [NEWLINE] Do you have any statistics or studies to back this up? Or are you just acting off [emotion]( [URL] ) and what appears to be a bias and prejudice towards a certain user of the road? [NEWLINE] [NEWLINE] [STARTQ] You and your broken spoke endanger everyone around you. Suppose your bike breaks and throws you into the street. The car behind you swerves to avoid running you over and you crashes into another car. YOU and your faulty bike just caused an accident. Why does the car driver's insurance have to pay for that? It was your fault. [ENDQ] [NEWLINE] Suppose your car has a loose hubcap just waiting to fall off, or a bumper duct taped on about to fall off. [NEWLINE] Suppose your truck with a load in the bed that's not properly secured and risks falling off? [NEWLINE] Suppose your car has a blowout. [NEWLINE] Suppose your car hits a pothole. [NEWLINE] Suppose a child (or wildlife) hops out in the street in front of you. [NEWLINE] [NEWLINE] If any of this happens and you're in your car, big fucking deal. You're surrounded by an entire car body, a seat belt (hopefully), probably an air bag. What happens to the cyclist in these situations? They're probably seriously injured or dead whether it's their fault or not. I'm under the impression that you're valuing your car's condition and your ability to travel a road bike/frustration free as more important than the life and well being of another human being. You should CMV on that. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Or you are riding down a street full of parked cars and you get too close and knock off a parked cars mirror assembly, or scratch all down the side of the car, or dent it. Thousands of dollars in body work that YOU caused. Why should the car owner's insurance have to pay that? Or the owner the deductible? Or his rates go up because of an accident YOU caused? Or have to hunt you down to go through the headache and expense to sue you in small claims court? [ENDQ] [NEWLINE] Do you have a single instance where a cyclist's actions caused **thousands** of dollars in body work?? [NEWLINE] Driving/riding/walking, etc... involves risk of some sort. [NEWLINE] We could play this " what - if " game all day and it wouldn't get us anywhere. [NEWLINE] What happens when you park your car crooked in a parking a
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for this response (∆). It was extremely informative, and<mask><mask> it raises an interesting issue. [NEWLINE] For me, the problem with this definition of feminism seems to be that it captures enough viewpoints that the label becomes somewhat meaningless.<mask><mask> that most people would consider themselves some sort of feminist in this sense – I certainly would.<mask>, the political/social agenda pushed by many feminist groups is not one<mask><mask> with. [NEWLINE] [NEWLINE] The clip that jumps out at me is somewhat old,<mask> still relevant: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Basically the NOW New York City President is debating Patrice O'neal on whether O&amp;A should be fired from the radio for a comment that was offensive to women. This particular example has nothing to do with laws or policies and<mask> they impact women. Personally,<mask><mask> that O&amp;A are hilarious.<mask> someone is horribly offended by them, I don’t see the problem with them just changing the channel.<mask><mask> that many feminists would agree on this point. In this particular case, the woman debating is not an O&amp;A listener. She certainly heard about this offense second hand and quite possibly never even listened to the actual clip. To me, this seems like an invasion of a space that she has no stake in.<mask><mask>, NOW has influence that can lead to a show’s cancellation. This power derives from a large base of people whose consider themselves feminists, many of whom only hold feminism’s most general paradigm.<mask> much of this base would be indifferent towards NOW’s actions in the particular situation, and many would have problems with it. The same thing happens with various radical factions within religion. Would creationism in schools be a concern<mask> not for the fact that many weakly religious or “spiritual” people put down Christian on the census? [NEWLINE] [NEWLINE] That is one of the reasons<mask> I find it hard to consider myself a feminist<mask> my acceptance of its general tenants. Many of my friends (male and female)<mask> find this problematic. [NEWLINE] [NEWLINE] The other problem that I have is that feminism (or at least movements motivated by feminism) tends to limit open debate. There is a lot of intolerance in this country, and that is a huge problem. On a legal level, bigotry, misogyny, homophobia (and the like) need to be ended immediately. On a social level, it takes time. At this level,<mask><mask> that a lot of the fight against intolerance is off base
Label encoding: <s>Thank you for this response (∆). It was extremely informative, and I think it raises an interesting issue. [NEWLINE] For me, the problem with this definition of feminism seems to be that it captures enough viewpoints that the label becomes somewhat meaningless. I think that most people would consider themselves some sort of feminist in this sense – I certainly would. However, the political/social agenda pushed by many feminist groups is not one I agree with. [NEWLINE] [NEWLINE] The clip that jumps out at me is somewhat old, but still relevant: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Basically the NOW New York City President is debating Patrice O'neal on whether O&amp;A should be fired from the radio for a comment that was offensive to women. This particular example has nothing to do with laws or policies and how they impact women. Personally, I think that O&amp;A are hilarious. If someone is horribly offended by them, I don’t see the problem with them just changing the channel. I think that many feminists would agree on this point. In this particular case, the woman debating is not an O&amp;A listener. She certainly heard about this offense second hand and quite possibly never even listened to the actual clip. To me, this seems like an invasion of a space that she has no stake in. In fact, NOW has influence that can lead to a show’s cancellation. This power derives from a large base of people whose consider themselves feminists, many of whom only hold feminism’s most general paradigm. But much of this base would be indifferent towards NOW’s actions in the particular situation, and many would have problems with it. The same thing happens with various radical factions within religion. Would creationism in schools be a concern if not for the fact that many weakly religious or “spiritual” people put down Christian on the census? [NEWLINE] [NEWLINE] That is one of the reasons why I find it hard to consider myself a feminist despite my acceptance of its general tenants. Many of my friends (male and female) also find this problematic. [NEWLINE] [NEWLINE] The other problem that I have is that feminism (or at least movements motivated by feminism) tends to limit open debate. There is a lot of intolerance in this country, and that is a huge problem. On a legal level, bigotry, misogyny, homophobia (and the like) need to be ended immediately. On a social level, it takes time. At this level, I think that a lot of the fight against intolerance is off base
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>u/TMLFAN11 has given a pretty thorough response to this,<mask> I'll only add my two cents<mask> I differ from him. His answers about asymmetric information are excellent (and<mask> they're coming from an econ major that's no surprise!) [NEWLINE] [NEWLINE] [STARTQ] For... the non emergency stuff...<mask> cant someone pick a hospital based on reviews on yelp? [ENDQ] [NEWLINE] Actually, I don't see any reason, in principle,<mask> you shouldn't be allowed to travel for non-emergency treatment.<mask><mask>, the NHS already allows this,<mask> GPs are not employees of the state<mask> private contractors who bill the NHS for their time. My dad lives in London<mask> travels two hours to see the GP he went to<mask> he lived further North. [NEWLINE] [NEWLINE] One reason we might not want people doing that for routine surgery,<mask>, is that it directly impacts a hospital's competence at the harder stuff.<mask> you read the *Daily Mail* you will be aware the NHS is closing a lot of child heart surgery units<mask> of... immigrants or something (it's the *Mail*,<mask> are you going to do?) The actual reason is that infant cardiology is<mask> complicated that<mask> your team isn't doing about 400 operations a year you are a liability to patients.<mask><mask> your routine (inpatient) pacemaker insertion means that the team at your local hospital 'keep their hand in' and that means the difference between life and death for someone with an enormous heart valve haematoma following a gunshot or something it might be worth it that you have slightly worse hospital food (or whatever). [NEWLINE] [NEWLINE] That's a personal judgement<mask> ; some people think individual rights always trump obligations to society, some people think the exact opposite and most people lie somewhere in the middle. [NEWLINE] [NEWLINE] Just on the specific point of using Yelp reviews<mask>, the actual system would have to be enormously more complicated in order that it wasn't distortionary. A simple example would be someone rating a hospital highly based on a 'halo effect' from the fact they had pretty receptionists and clean-looking wards<mask> a much better hospital was rated poorly<mask> their food was crap tasting<mask> nutritious. (<mask> an aside, the exact weight to give'soft' aspects of hospital management like tasty food and clean waiting rooms in our judgement of<mask> well a hospital performs is<mask> I am working on at the moment). More crucially, this can affect matters of life and death. An example (using made up numbers); would you rather send
Label encoding: <s>u/TMLFAN11 has given a pretty thorough response to this, so I'll only add my two cents where I differ from him. His answers about asymmetric information are excellent (and since they're coming from an econ major that's no surprise!) [NEWLINE] [NEWLINE] [STARTQ] For... the non emergency stuff... why cant someone pick a hospital based on reviews on yelp? [ENDQ] [NEWLINE] Actually, I don't see any reason, in principle, why you shouldn't be allowed to travel for non-emergency treatment. In fact, the NHS already allows this, since GPs are not employees of the state but private contractors who bill the NHS for their time. My dad lives in London but travels two hours to see the GP he went to when he lived further North. [NEWLINE] [NEWLINE] One reason we might not want people doing that for routine surgery, though, is that it directly impacts a hospital's competence at the harder stuff. If you read the *Daily Mail* you will be aware the NHS is closing a lot of child heart surgery units because of... immigrants or something (it's the *Mail*, what are you going to do?) The actual reason is that infant cardiology is so complicated that if your team isn't doing about 400 operations a year you are a liability to patients. So if your routine (inpatient) pacemaker insertion means that the team at your local hospital 'keep their hand in' and that means the difference between life and death for someone with an enormous heart valve haematoma following a gunshot or something it might be worth it that you have slightly worse hospital food (or whatever). [NEWLINE] [NEWLINE] That's a personal judgement though ; some people think individual rights always trump obligations to society, some people think the exact opposite and most people lie somewhere in the middle. [NEWLINE] [NEWLINE] Just on the specific point of using Yelp reviews though, the actual system would have to be enormously more complicated in order that it wasn't distortionary. A simple example would be someone rating a hospital highly based on a 'halo effect' from the fact they had pretty receptionists and clean-looking wards while a much better hospital was rated poorly because their food was crap tasting but nutritious. ( As an aside, the exact weight to give'soft' aspects of hospital management like tasty food and clean waiting rooms in our judgement of how well a hospital performs is what I am working on at the moment). More crucially, this can affect matters of life and death. An example (using made up numbers); would you rather send
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Taking a quick moment to address your addendum to the OP. [NEWLINE] [NEWLINE] I feel you are leaving a crucial component of the argument out.  The reason that it is<mask> difficult for you to find anything wrong with the cannibalism in principal is<mask> you are framing it devoid of context.  Cannibalism in "the right" context already has widespread acceptance, albeit begrudgingly. [NEWLINE] [NEWLINE] <mask> people hear about cannibalism used<mask> a last ditch means to survive most see an occurrence that is sad, disgusting and regrettable. <mask> is it wrong? <mask> the standard in society against cannibalism most can be convinced of the validity of the action considering the extreme circumstances.  Some even see it<mask> being worse<mask> they HADN'T eaten the other person, viewing it<mask> better that somebody survived rather than nobody.  It may even be that it is the last great thing someone accomplished for their comrade in suffering, providing them with the needed nutrition to survive. [NEWLINE] [NEWLINE] Their exist today numerous tribes that still practice ritual cannibalism within their community<mask> a part of death rites and such.  We see it<mask> disgusting of course<mask> only evangelical bigots invade their community and attempt to make them stop.  The practices are in a way just<mask> crazy<mask> our habits of preserving them and burying them in a box<mask> likely stem from the same goal.  Dispose of the body in a way that causes the least disease and attracts the least predators and scavengers. [NEWLINE] ***** [NEWLINE] At this point you may be wondering "Aren't you supposed to be changing my view?  This is pretty strong for my case."  At this point I am challenging not the issues the question brings up<mask> the question itself. [NEWLINE] [NEWLINE] You cannot possibly have a discussion about the morality of a given event<mask> that event is taking place in a vacuum.  Once someone has reached a certain point in logical thinking statements like "(Insert context free event here) is not immoral<mask> done in the right way." become essentially no-brainers.  Without context anything that can be imagined is possible and the arguments ceases to be about the event and instead becomes a tennis match wherein the participants throw various hypothetical contexts at each other attempting to reach favorable conclusions.  Just<mask> their is no human without society, there are no morals without context. [NEWLINE] [NEWLINE] By presupposing the context<mask> you have done in point 4 (consent), the Edit (new society starting from scratch) and your latest addendum (cleanliness, safety and access are now non-
Label encoding: <s>Taking a quick moment to address your addendum to the OP. [NEWLINE] [NEWLINE] I feel you are leaving a crucial component of the argument out.  The reason that it is so difficult for you to find anything wrong with the cannibalism in principal is because you are framing it devoid of context.  Cannibalism in "the right" context already has widespread acceptance, albeit begrudgingly. [NEWLINE] [NEWLINE] When people hear about cannibalism used as a last ditch means to survive most see an occurrence that is sad, disgusting and regrettable.  But is it wrong?  Despite the standard in society against cannibalism most can be convinced of the validity of the action considering the extreme circumstances.  Some even see it as being worse if they HADN'T eaten the other person, viewing it as better that somebody survived rather than nobody.  It may even be that it is the last great thing someone accomplished for their comrade in suffering, providing them with the needed nutrition to survive. [NEWLINE] [NEWLINE] Their exist today numerous tribes that still practice ritual cannibalism within their community as a part of death rites and such.  We see it as disgusting of course but only evangelical bigots invade their community and attempt to make them stop.  The practices are in a way just as crazy as our habits of preserving them and burying them in a box but likely stem from the same goal.  Dispose of the body in a way that causes the least disease and attracts the least predators and scavengers. [NEWLINE] ***** [NEWLINE] At this point you may be wondering "Aren't you supposed to be changing my view?  This is pretty strong for my case."  At this point I am challenging not the issues the question brings up but the question itself. [NEWLINE] [NEWLINE] You cannot possibly have a discussion about the morality of a given event when that event is taking place in a vacuum.  Once someone has reached a certain point in logical thinking statements like "(Insert context free event here) is not immoral if done in the right way." become essentially no-brainers.  Without context anything that can be imagined is possible and the arguments ceases to be about the event and instead becomes a tennis match wherein the participants throw various hypothetical contexts at each other attempting to reach favorable conclusions.  Just as their is no human without society, there are no morals without context. [NEWLINE] [NEWLINE] By presupposing the context as you have done in point 4 (consent), the Edit (new society starting from scratch) and your latest addendum (cleanliness, safety and access are now non-
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I worked for 5 years supporting people with physical, cognitive, and sensory disabilities. Everything from an otherwise regular kid with a missing leg to a blind kid in a wheelchair with autism. [NEWLINE] [NEWLINE] I'm not these kids' parents. I get to go home at the end of the day or the end of the week and take a break. I never had to pay their medical bills, save for their future<mask> I die, struggle to come to terms with their disability, blame myself for their suffering, or see my marriage or relationships destroyed by the stress of raising them. [NEWLINE] [NEWLINE] At the same time, I do have experience with the ups and downs, the personal care, the insensitive doctors, the public stares, the sleepless nights, the violence, and the difficulty of day-to-day life<mask> they're in my care (try buying groceries with a 6'3 blind man who compulsively masturbates<mask> he has anxiety). And I can tell you, it's not for everyone.<mask><mask> you can deal with hardship and take pride in small successes, it's absolutely worth it. [NEWLINE] [NEWLINE] All kids are proud of their accomplishments. And all parents are proud of their kids.<mask><mask> both luck and society are stacked up against them in every way, every win is that much more meaningful. Think parents are proud<mask> their able-bodied upper-middle-class white kid graduates from high school? Think<mask> they feel<mask> their non-verbal 8-year old whose generic anomalies meant he had "no chance of living past age 4" learns to tie his shoes, serve his own dinner, and dance the Macarena. [NEWLINE] [NEWLINE] Think about the sense of relief<mask> a 30-year old man with cerebral palsy and the intellectual capacity of a 5-year old stops his 3-hour tantrum, let's you put a bandaid on the cut on his head that tore open<mask> he was smashing it against the wall, apologizes, gives you a hug, and thanks you for always being there for him. [NEWLINE] [NEWLINE] Think of the bonding moments you have<mask> the 15-year old whose diaper you're changing<mask> he has no feeling or strength below the waist can't stop giggling at his own erection, which you won't go near and he has no control over,<mask> you're both literally covered in shit. Think you're close with your kids? This is intimacy. This is trust. [NEWLINE] [NEWLINE] Think of that moment<mask> your 9-year old, who never speaks, doesn't walk, and keeps her
Label encoding: <s>I worked for 5 years supporting people with physical, cognitive, and sensory disabilities. Everything from an otherwise regular kid with a missing leg to a blind kid in a wheelchair with autism. [NEWLINE] [NEWLINE] I'm not these kids' parents. I get to go home at the end of the day or the end of the week and take a break. I never had to pay their medical bills, save for their future when I die, struggle to come to terms with their disability, blame myself for their suffering, or see my marriage or relationships destroyed by the stress of raising them. [NEWLINE] [NEWLINE] At the same time, I do have experience with the ups and downs, the personal care, the insensitive doctors, the public stares, the sleepless nights, the violence, and the difficulty of day-to-day life when they're in my care (try buying groceries with a 6'3 blind man who compulsively masturbates when he has anxiety). And I can tell you, it's not for everyone. But if you can deal with hardship and take pride in small successes, it's absolutely worth it. [NEWLINE] [NEWLINE] All kids are proud of their accomplishments. And all parents are proud of their kids. But when both luck and society are stacked up against them in every way, every win is that much more meaningful. Think parents are proud when their able-bodied upper-middle-class white kid graduates from high school? Think how they feel when their non-verbal 8-year old whose generic anomalies meant he had "no chance of living past age 4" learns to tie his shoes, serve his own dinner, and dance the Macarena. [NEWLINE] [NEWLINE] Think about the sense of relief when a 30-year old man with cerebral palsy and the intellectual capacity of a 5-year old stops his 3-hour tantrum, let's you put a bandaid on the cut on his head that tore open when he was smashing it against the wall, apologizes, gives you a hug, and thanks you for always being there for him. [NEWLINE] [NEWLINE] Think of the bonding moments you have when the 15-year old whose diaper you're changing because he has no feeling or strength below the waist can't stop giggling at his own erection, which you won't go near and he has no control over, while you're both literally covered in shit. Think you're close with your kids? This is intimacy. This is trust. [NEWLINE] [NEWLINE] Think of that moment when your 9-year old, who never speaks, doesn't walk, and keeps her
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not familiar with Fitzgerald<mask> I do know Carrier's work. His argument basically boils down to this: Paul was really talking about a spirit named Jesus who spiritually died and rose again. Everyone following Paul came up with the human Jesus after-the-fact. [NEWLINE] [NEWLINE] The primary criticism Carrier receives in academia is that<mask> much of Paul's writings don't really make all that much sense<mask> this is true.<mask> Carrier comes up with a bunch of *ad hoc* explanations to each of the problems. I just don't see any reason to think Carrier's explanations are correct other than to simply defend his thesis. They certainly don't seem like the most plausible interpretation on their face. [NEWLINE] [NEWLINE] Carrier originally took up this cross in order to see<mask> mythicism really should be put to rest. It was one last hurrah in an era<mask> he thought mythicism wasn't given a fair scholarly chance. I still think Carrier's arguments all fail. [NEWLINE] [NEWLINE] [STARTQ] There are no extrabiblical attestations of any significant event from the life of Jesus.<mask><mask><mask>, events such<mask> the slaughter of the innocents, the census of all the Empire are clearly fictitious, and multiple miracles, the triumphal entry to Jerusalem, and multiple events surrounding the crucifixion are absent from all historical records<mask> there's a reasonable chance that some account of them would have survived, had they occurred at all. [ENDQ] [NEWLINE] Well, we do have accounts that survived. They are biblical. First, we should ask the question whether or not these works being later canonized is relevant.<mask><mask> not. Next, we should simply analyze<mask> we do have and come up with the best explanation possible. Regardless, there are some extrabiblical sources<mask> some of the comments have already mentioned. [NEWLINE] [NEWLINE] [STARTQ] The synoptic problem indicates that we are working from at most one source that even approaches being primary, and even that most likely written much later, anonymously, and<mask> hagiography rather than history. [ENDQ] [NEWLINE] I'm not sure most scholars would agree with there being at most one synoptic source. Most scholars think that the sources the synoptics have are Mark, M (Matthew), L (Luke), and sometimes Q. One would need to argue against the reasons for these theories to conclude that we are left with only one source. Stylistically, this conclusion just doesn't seem all that likely. [NEWLINE] [NEWLINE] [STARTQ] Well into the third century, pagan sources mostly recount the existence of Christians and document the claims of Christians. This is at
Label encoding: <s>I'm not familiar with Fitzgerald but I do know Carrier's work. His argument basically boils down to this: Paul was really talking about a spirit named Jesus who spiritually died and rose again. Everyone following Paul came up with the human Jesus after-the-fact. [NEWLINE] [NEWLINE] The primary criticism Carrier receives in academia is that so much of Paul's writings don't really make all that much sense if this is true. So Carrier comes up with a bunch of *ad hoc* explanations to each of the problems. I just don't see any reason to think Carrier's explanations are correct other than to simply defend his thesis. They certainly don't seem like the most plausible interpretation on their face. [NEWLINE] [NEWLINE] Carrier originally took up this cross in order to see if mythicism really should be put to rest. It was one last hurrah in an era where he thought mythicism wasn't given a fair scholarly chance. I still think Carrier's arguments all fail. [NEWLINE] [NEWLINE] [STARTQ] There are no extrabiblical attestations of any significant event from the life of Jesus. On the contrary, events such as the slaughter of the innocents, the census of all the Empire are clearly fictitious, and multiple miracles, the triumphal entry to Jerusalem, and multiple events surrounding the crucifixion are absent from all historical records when there's a reasonable chance that some account of them would have survived, had they occurred at all. [ENDQ] [NEWLINE] Well, we do have accounts that survived. They are biblical. First, we should ask the question whether or not these works being later canonized is relevant. I think not. Next, we should simply analyze what we do have and come up with the best explanation possible. Regardless, there are some extrabiblical sources as some of the comments have already mentioned. [NEWLINE] [NEWLINE] [STARTQ] The synoptic problem indicates that we are working from at most one source that even approaches being primary, and even that most likely written much later, anonymously, and as hagiography rather than history. [ENDQ] [NEWLINE] I'm not sure most scholars would agree with there being at most one synoptic source. Most scholars think that the sources the synoptics have are Mark, M (Matthew), L (Luke), and sometimes Q. One would need to argue against the reasons for these theories to conclude that we are left with only one source. Stylistically, this conclusion just doesn't seem all that likely. [NEWLINE] [NEWLINE] [STARTQ] Well into the third century, pagan sources mostly recount the existence of Christians and document the claims of Christians. This is at
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think I'm idealizing anything -- there is some contention<mask> to<mask> quickly the butterfly effect operates in practice,<mask> it's hard to<mask><mask> it doesn't, especially<mask> you give it millions of years. [NEWLINE] [NEWLINE] The butterfly effect isn't about<mask> the slightest perturbation can plunge the world into random chaos.<mask> there are forces at hand, these forces will keep applying and will keep steering the system in a particular direction, for instance towards lower energy states. The butterfly effect won't make the climate change.<mask> it sure<mask> hell will change the *details* of the phenomena which,<mask> averaged, form climate.<mask> it won't change the climate,<mask> it will change the winds. It won't change the overall march of evolution,<mask> it will change the species on its path. It won't change human nature,<mask> it will change the particular humans that exist. You get<mask> I'm saying? The bigger the system, the more robust it is to perturbation,<mask> it's only robust<mask> it's an average.<mask><mask> our ego may say, the existence of an individual human in the context of humanity is a detail, not an average. [NEWLINE] [NEWLINE] [STARTQ] the effect of any one thing is trivial<mask> it never has a chance to gain any momentum,<mask> any force it produces is immediately interrupted by the force of another [ENDQ] [NEWLINE] Sure. "Interrupted".<mask><mask>? Well, perhaps it is done by balancing things around.<mask> there's a hurricane here, there won't be one there.<mask> I don't take this job, somebody else will take the place.<mask> we don't conceive a child now,<mask>, we'll conceive it tomorrow. You try to stop a mugger from mugging you? Take that knife in the ribs. Humanity fails to evolve sapience? Whatever. That's just an opportunity for crows. [NEWLINE] [NEWLINE] Let me put it this way: over a certain time span, an invariant is only going to be preserved<mask> it is an "end" of the system, something that the system is actively working to preserve. For instance,<mask> your goal is to have children, your behavior will constantly adapt in ways that put you in contact with the opposite gender. The fact that you have children will be robust to all sorts of changes during your life<mask> you work towards it and you adapt yourself with the goal in mind. Which particular children you have,<mask>, is something that's a lot more sensitive to perturbations<mask> it matters much less.
Label encoding: <s>I don't think I'm idealizing anything -- there is some contention as to how quickly the butterfly effect operates in practice, but it's hard to argue that it doesn't, especially if you give it millions of years. [NEWLINE] [NEWLINE] The butterfly effect isn't about how the slightest perturbation can plunge the world into random chaos. When there are forces at hand, these forces will keep applying and will keep steering the system in a particular direction, for instance towards lower energy states. The butterfly effect won't make the climate change. But it sure as hell will change the *details* of the phenomena which, when averaged, form climate. So it won't change the climate, but it will change the winds. It won't change the overall march of evolution, but it will change the species on its path. It won't change human nature, but it will change the particular humans that exist. You get what I'm saying? The bigger the system, the more robust it is to perturbation, but it's only robust because it's an average. Despite what our ego may say, the existence of an individual human in the context of humanity is a detail, not an average. [NEWLINE] [NEWLINE] [STARTQ] the effect of any one thing is trivial because it never has a chance to gain any momentum, because any force it produces is immediately interrupted by the force of another [ENDQ] [NEWLINE] Sure. "Interrupted". But how? Well, perhaps it is done by balancing things around. If there's a hurricane here, there won't be one there. If I don't take this job, somebody else will take the place. If we don't conceive a child now, why, we'll conceive it tomorrow. You try to stop a mugger from mugging you? Take that knife in the ribs. Humanity fails to evolve sapience? Whatever. That's just an opportunity for crows. [NEWLINE] [NEWLINE] Let me put it this way: over a certain time span, an invariant is only going to be preserved if it is an "end" of the system, something that the system is actively working to preserve. For instance, if your goal is to have children, your behavior will constantly adapt in ways that put you in contact with the opposite gender. The fact that you have children will be robust to all sorts of changes during your life because you work towards it and you adapt yourself with the goal in mind. Which particular children you have, however, is something that's a lot more sensitive to perturbations because it matters much less.
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0854, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> yes, people collectively changing their behaviour to do something about the impact is still a natural process. You wouldn’t say it’s unnatural for humans to want to preserve the earth, and by extension themselves, by recycling? [ENDQ] [NEWLINE] This is kind of an asside<mask> i feel its a distractor<mask> :  No other species has the ability to foresee their impact on the environment and their natural world, let alone<mask> it will impact them.  Animals go through boom and bust population cycles all the time, without ever realizing that their behaviors perpetuate them (for example, wolf populations grow with wide consumption of deer,<mask> the population of deer decreases, the food supply for wolves dries up, and most wolves die the next year.  Now, with much fewer preditors culling the herd, the deer population grows rapidly.  Sometimes, these busts can be very dramatic and lead to population extinctions.  Logically, individual wolves would be better served by keeping their population small to ensure a stable, consistent food supply,<mask> they don't.) [NEWLINE] [NEWLINE] [STARTQ] I still can’t see<mask> there is this distinction,<mask> we are<mask> natural<mask> anything else. [ENDQ] [NEWLINE] Recycling was a bad example. <mask> about farming?  Only in the last 5000 years have humans raising crops for food.  They existed a good 100,000 years before that.  Or burning fossil fuel for energy?  humans have only started using coal and oil in the last 150-200 years. <mask> we look at the established pattern of behavior over the long term, can you really<mask><mask> burning fossil fuel<mask> a "natural," inherent behavior by humans? [NEWLINE] [NEWLINE] Sometimes, these things have detrimental impacts on the environment.  "Natural" in this context distinguishing between human-caused consecuences and non human-caused impacts. [NEWLINE] Saying that humans are part of nature,<mask> all behavior we partake in  is natural, is a very comfortable and dangerous lie to fall into.  Now, we don't have any incentive to recycle, or reduce our consumption,<mask> its all a part of our "natural" behavior. [NEWLINE] [NEWLINE] [STARTQ] And yeah, those star trek people disrupting the natural order would again be a part of nature. The prime directive may<mask> well be to not disrupt the ‘order’ of other worlds. [ENDQ] [NEWLINE] Sorry, I didn't elaborate on this point enough.  With the Kingon civil war, the federation couldn't intervene with the natural order of
Label encoding: <s> [STARTQ] But yes, people collectively changing their behaviour to do something about the impact is still a natural process. You wouldn’t say it’s unnatural for humans to want to preserve the earth, and by extension themselves, by recycling? [ENDQ] [NEWLINE] This is kind of an asside because i feel its a distractor but :  No other species has the ability to foresee their impact on the environment and their natural world, let alone how it will impact them.  Animals go through boom and bust population cycles all the time, without ever realizing that their behaviors perpetuate them (for example, wolf populations grow with wide consumption of deer, as the population of deer decreases, the food supply for wolves dries up, and most wolves die the next year.  Now, with much fewer preditors culling the herd, the deer population grows rapidly.  Sometimes, these busts can be very dramatic and lead to population extinctions.  Logically, individual wolves would be better served by keeping their population small to ensure a stable, consistent food supply, but they don't.) [NEWLINE] [NEWLINE] [STARTQ] I still can’t see why there is this distinction, since we are as natural as anything else. [ENDQ] [NEWLINE] Recycling was a bad example.  What about farming?  Only in the last 5000 years have humans raising crops for food.  They existed a good 100,000 years before that.  Or burning fossil fuel for energy?  humans have only started using coal and oil in the last 150-200 years.  If we look at the established pattern of behavior over the long term, can you really argue that burning fossil fuel as a "natural," inherent behavior by humans? [NEWLINE] [NEWLINE] Sometimes, these things have detrimental impacts on the environment.  "Natural" in this context distinguishing between human-caused consecuences and non human-caused impacts. [NEWLINE] Saying that humans are part of nature, so all behavior we partake in  is natural, is a very comfortable and dangerous lie to fall into.  Now, we don't have any incentive to recycle, or reduce our consumption, since its all a part of our "natural" behavior. [NEWLINE] [NEWLINE] [STARTQ] And yeah, those star trek people disrupting the natural order would again be a part of nature. The prime directive may as well be to not disrupt the ‘order’ of other worlds. [ENDQ] [NEWLINE] Sorry, I didn't elaborate on this point enough.  With the Kingon civil war, the federation couldn't intervene with the natural order of
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Most of the events in the video don't qualify<mask> harassment individually,<mask> put them together every 5 minutes all day evey day and it all amounts to harrassment. In the same way one piece of paper doesn't contaminate a city, littering is illegal in most large cities<mask> of the cummulative damage<mask><mask><mask>. [ENDQ] [NEWLINE] &amp;#8710; [NEWLINE] [NEWLINE] I’ve decided to award delta to the above,<mask> that post pointed out that cumulative effect can be greater than the sum of its parts. I’ve known that before, I just haven’t thought about it in relation to this particular issue. [This changed my view in a sense that I now see that a campaign against approaching random strangers on the street may<mask> be worthwhile in some cases] [NEWLINE] [NEWLINE] <mask>, I haven’t changed my mind on my main thesis that calling compliments and greetings “harassment” serves nobody. I will explain<mask> : [NEWLINE] [NEWLINE] 1. This cumulative effect may be limited to big cities like New York, and may not apply to other areas, especially rural areas. [NEWLINE] [NEWLINE] 2. It is not useful to have different definitions for the word harassment in different areas. Internet is global,<mask> any website or video that claims that saying “hi” to a stranger is harassment will spread that idea to every viewer,<mask><mask> his location. [NEWLINE] [NEWLINE] 3.<mask> I said before, it seems just wrong to me to label a guy that says “hi” to a stranger on a street the same<mask> a guy that grabs girl's ass. I still think it is not respectful to victims of real harassment such<mask> stalking or something even worse.<mask> you disagree, then ask yourself<mask> should we not call every harassment a rape? (i.e. “He complimented me” becomes “He raped me”).<mask> words mean something, that’s<mask>. [NEWLINE] [NEWLINE] 4. It may not even be beneficial to the proponents of that view,<mask> of the backlash. The authors of “10 Hours of Walking in NYC<mask> a Woman” had to disable Youtube ratings<mask> the video received<mask> many dislikes. [NEWLINE] [NEWLINE] 5. Some women have different views on this topic<mask> they made video responses to that video, disagreeing with it.<mask>, clearly, not all women feel harassed by that kind of behavior. [NEWLINE] [NEWLINE] I don’t know<mask> a solution to this issue should be.<mask><mask>
Label encoding: <s> [STARTQ] Most of the events in the video don't qualify as harassment individually, but put them together every 5 minutes all day evey day and it all amounts to harrassment. In the same way one piece of paper doesn't contaminate a city, littering is illegal in most large cities because of the cummulative damage as a result. [ENDQ] [NEWLINE] &amp;#8710; [NEWLINE] [NEWLINE] I’ve decided to award delta to the above, because that post pointed out that cumulative effect can be greater than the sum of its parts. I’ve known that before, I just haven’t thought about it in relation to this particular issue. [This changed my view in a sense that I now see that a campaign against approaching random strangers on the street may indeed be worthwhile in some cases] [NEWLINE] [NEWLINE] However, I haven’t changed my mind on my main thesis that calling compliments and greetings “harassment” serves nobody. I will explain why : [NEWLINE] [NEWLINE] 1. This cumulative effect may be limited to big cities like New York, and may not apply to other areas, especially rural areas. [NEWLINE] [NEWLINE] 2. It is not useful to have different definitions for the word harassment in different areas. Internet is global, so any website or video that claims that saying “hi” to a stranger is harassment will spread that idea to every viewer, regardless of his location. [NEWLINE] [NEWLINE] 3. As I said before, it seems just wrong to me to label a guy that says “hi” to a stranger on a street the same as a guy that grabs girl's ass. I still think it is not respectful to victims of real harassment such as stalking or something even worse. If you disagree, then ask yourself why should we not call every harassment a rape? (i.e. “He complimented me” becomes “He raped me”). Because words mean something, that’s why. [NEWLINE] [NEWLINE] 4. It may not even be beneficial to the proponents of that view, because of the backlash. The authors of “10 Hours of Walking in NYC as a Woman” had to disable Youtube ratings because the video received so many dislikes. [NEWLINE] [NEWLINE] 5. Some women have different views on this topic as they made video responses to that video, disagreeing with it. So, clearly, not all women feel harassed by that kind of behavior. [NEWLINE] [NEWLINE] I don’t know what a solution to this issue should be. I think
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that it goes without saying that I believe people should get<mask> they want out of life.  That being saying, in the &lt;<mask> I'm guessing is [STARTQ] ultra-rare scenario that someone WANTS their partner to lie and cheat on them, they should get<mask> they want,<mask> it seems all like a big game at best.  Almost<mask><mask><mask> they're both in on the "sleeping around and lying about it" than are they really lying at all? <mask> it's within the scope of the conditions they came together on regarding the relationship, is it "cheating"?  Is it "lying"<mask> you expect it?  Regardless, there's a lot of grey area... At the end of the day, I'm hopeful that people find someone (or a group of peoples) that make them happy. [ENDQ] [NEWLINE] I find myself totally on the flip-side of that.  I'll admit that even the idea of my partner not being trustworthy gives me a big ol' case of the bad-feels.  It disgusts me on some visceral level.  Even for instance,<mask> infidelity happens in a comedy movie, I find it appalling.  That song “Scotty Doesn’t Know” is probably my most hated song of all time.  Like; it makes me physically sick.  I know that this isn’t a “normal” way of viewing the world. <mask> in, in the year 2014, people don’t feel a similar negative reaction to infidelity<mask> they do to murder,<mask> that is just<mask> my brain reacts to it.  It’s like it’s not even a big deal in our society to lie to a person that’s supposed to be your closest friend and ally in life. [NEWLINE] [NEWLINE] Anyway,<mask> that’s part of the reason I support open relationships<mask> much. <mask> we didn’t live in a time and place<mask> it was shameful to establish things upfront like “I love you,<mask><mask> we’re going to continue to have sex and live together, there are going to be other people here having sex with us…  or just me in other places.  Is that something that sounds like a good time to you?” then we could cut out<mask> much of the pain in people’s lives.  Will it still hurt that other person?  ABSOLUTELY!  Will it
Label encoding: <s>I think that it goes without saying that I believe people should get what they want out of life.  That being saying, in the &lt; what I'm guessing is [STARTQ] ultra-rare scenario that someone WANTS their partner to lie and cheat on them, they should get what they want, but it seems all like a big game at best.  Almost as if since they're both in on the "sleeping around and lying about it" than are they really lying at all?  If it's within the scope of the conditions they came together on regarding the relationship, is it "cheating"?  Is it "lying" if you expect it?  Regardless, there's a lot of grey area... At the end of the day, I'm hopeful that people find someone (or a group of peoples) that make them happy. [ENDQ] [NEWLINE] I find myself totally on the flip-side of that.  I'll admit that even the idea of my partner not being trustworthy gives me a big ol' case of the bad-feels.  It disgusts me on some visceral level.  Even for instance, when infidelity happens in a comedy movie, I find it appalling.  That song “Scotty Doesn’t Know” is probably my most hated song of all time.  Like; it makes me physically sick.  I know that this isn’t a “normal” way of viewing the world.  As in, in the year 2014, people don’t feel a similar negative reaction to infidelity as they do to murder, but that is just how my brain reacts to it.  It’s like it’s not even a big deal in our society to lie to a person that’s supposed to be your closest friend and ally in life. [NEWLINE] [NEWLINE] Anyway, so that’s part of the reason I support open relationships so much.  If we didn’t live in a time and place where it was shameful to establish things upfront like “I love you, but if we’re going to continue to have sex and live together, there are going to be other people here having sex with us…  or just me in other places.  Is that something that sounds like a good time to you?” then we could cut out so much of the pain in people’s lives.  Will it still hurt that other person?  ABSOLUTELY!  Will it
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry,<mask> gender roles (strict or otherwise) and a patriarchy are not at all the same thing. It is entirely possible to have either one without the other,<mask> your point is moot from the start. [NEWLINE] [NEWLINE] <mask> per your definition, a patriarchy is a system in which men hold the majority of direct political or social power,<mask> whether or not that actually creates any gender roles at all is not an inherent consequence of the definition, and<mask> such it's completely disingenuous to claim the two phenomena are related. The whole things absolutely *reeks* of trying to blame the entire phenomenon of gender roles on men by way of tenuously connecting their positions in power to the status quo regarding gender roles, which is entirely false<mask> we all know,<mask> men are most definitely not the only ones perpetuating any gender stereotypes, and they are most certainly not handed down from above. [NEWLINE] [NEWLINE] It's a complete non sequitur, the two unrelated phenomena of gender roles *et al* being connected to which gender has direct political power. Either you define patriarchy<mask> mere gender roles which cast men<mask> being leaders, in which case it's an alarmist misnomer that has nothing to do with men specifically, or you say it's a *de facto* system of political and governmental organization (which is<mask> the term means, by the way, and which you - more or less - have used), in which case gender roles are an unrelated phenomenon. [NEWLINE] [NEWLINE] In short,<mask> you're using patriarchy<mask> the term is and ought to be understood,<mask> a system<mask> men hold the majority of direct political power, then it is in and of itself a non-issue,<mask> made clear previously.<mask>,<mask><mask><mask><mask>, you are using it<mask> a by-word for gender roles which cast men<mask> aggressive, ambitious leaders, you are using it in a despicable attempt to shift blame onto a gender by insinuating that their political positions somehow grant them power to determine gender roles unilaterally. [NEWLINE] [NEWLINE] [STARTQ] <mask> legally the system no longer discriminates [ENDQ] [NEWLINE] Women receive shorter sentences for the same crimes... [NEWLINE] [NEWLINE] [STARTQ] And men can represent women just fine, it's just more likely that a woman will understand women's issues better<mask> she is a woman. A man has certain privileges that he might not acknowledge, and<mask> may not even know<mask> kind of work needs to be done to help the situation whereas a woman has likely experienced these problems. [ENDQ] [NEWLINE] There are two problems with this logic. First, gender is at
Label encoding: <s>I'm sorry, but gender roles (strict or otherwise) and a patriarchy are not at all the same thing. It is entirely possible to have either one without the other, so your point is moot from the start. [NEWLINE] [NEWLINE] As per your definition, a patriarchy is a system in which men hold the majority of direct political or social power, but whether or not that actually creates any gender roles at all is not an inherent consequence of the definition, and as such it's completely disingenuous to claim the two phenomena are related. The whole things absolutely *reeks* of trying to blame the entire phenomenon of gender roles on men by way of tenuously connecting their positions in power to the status quo regarding gender roles, which is entirely false as we all know, since men are most definitely not the only ones perpetuating any gender stereotypes, and they are most certainly not handed down from above. [NEWLINE] [NEWLINE] It's a complete non sequitur, the two unrelated phenomena of gender roles *et al* being connected to which gender has direct political power. Either you define patriarchy as mere gender roles which cast men as being leaders, in which case it's an alarmist misnomer that has nothing to do with men specifically, or you say it's a *de facto* system of political and governmental organization (which is what the term means, by the way, and which you - more or less - have used), in which case gender roles are an unrelated phenomenon. [NEWLINE] [NEWLINE] In short, if you're using patriarchy as the term is and ought to be understood, as a system where men hold the majority of direct political power, then it is in and of itself a non-issue, as made clear previously. If, on the other hand, you are using it as a by-word for gender roles which cast men as aggressive, ambitious leaders, you are using it in a despicable attempt to shift blame onto a gender by insinuating that their political positions somehow grant them power to determine gender roles unilaterally. [NEWLINE] [NEWLINE] [STARTQ] While legally the system no longer discriminates [ENDQ] [NEWLINE] Women receive shorter sentences for the same crimes... [NEWLINE] [NEWLINE] [STARTQ] And men can represent women just fine, it's just more likely that a woman will understand women's issues better because she is a woman. A man has certain privileges that he might not acknowledge, and so may not even know what kind of work needs to be done to help the situation whereas a woman has likely experienced these problems. [ENDQ] [NEWLINE] There are two problems with this logic. First, gender is at
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I would never expect or follow through with the destruction of the human race, I believe it to be the simplest way of dealing with our problems. [NEWLINE] [NEWLINE] I believe that we will never be satisfied with<mask> we have, there will never be equality for all, we'll continue to destroy the environment and in the end, we will all die a much worse death trying to live happily than one that would come form purposefully erasing our existence. Looking at it from another perspective, it's like eradicating the parts of DNA needed to make tonsils<mask> that humans never again deal with tonsillitis. In that analogy, the human body would be the Earth and tonsils would be the people. Tonsillitis would be the possibility of people doing bad things and removing the DNA would be<mask> we do to get rid of humans. I just think that, not taking feelings into account and going with the quickest and simplest option, remove the problems from the root is ideal. Once again, this does not take into account feelings of people<mask> obviously most of us don't want to die, regardless<mask> it would be better for the planet. [NEWLINE] [NEWLINE] Of course, erasing humans from the planet is much simpler than you'd expect. It<mask> doesn't have to be prolonged or painful<mask> would be our deaths from a polluted world and from wars. I am sure that within 50 years, a group could easily genetically engineer a strain of virus to get the job done. Mass production of asbestos and coordinated release all over the world could work too.<mask> I'm saying is that it's pretty easy and could be done quickly. Nothing more, nothing less. [NEWLINE] [NEWLINE] Now, this is all assuming a few things: [NEWLINE] [NEWLINE] - That<mask> humanity continues, we will eventually cease to exist against our will. [NEWLINE] - That humanity will never have true equality. [NEWLINE] - That war will always continue in some form. [NEWLINE] - That we will continue polluting/degrading the Earth<mask><mask><mask> we exist. [NEWLINE] [NEWLINE] <mask> each human somehow became the "God" of their own universe (<mask> I have no idea<mask> that would come about) then I believe there would be no problem with each individual existence,<mask> controlling your own universe means that no matter<mask> you do you are in the right. Of course, that is all BS/fiction and we live just like all the animals on the Earth, aside from being special in our own little ways. I don't think we<mask> a whole can ever fully agree on anything, and that our disagreements
Label encoding: <s>While I would never expect or follow through with the destruction of the human race, I believe it to be the simplest way of dealing with our problems. [NEWLINE] [NEWLINE] I believe that we will never be satisfied with what we have, there will never be equality for all, we'll continue to destroy the environment and in the end, we will all die a much worse death trying to live happily than one that would come form purposefully erasing our existence. Looking at it from another perspective, it's like eradicating the parts of DNA needed to make tonsils so that humans never again deal with tonsillitis. In that analogy, the human body would be the Earth and tonsils would be the people. Tonsillitis would be the possibility of people doing bad things and removing the DNA would be what we do to get rid of humans. I just think that, not taking feelings into account and going with the quickest and simplest option, remove the problems from the root is ideal. Once again, this does not take into account feelings of people as obviously most of us don't want to die, regardless if it would be better for the planet. [NEWLINE] [NEWLINE] Of course, erasing humans from the planet is much simpler than you'd expect. It also doesn't have to be prolonged or painful as would be our deaths from a polluted world and from wars. I am sure that within 50 years, a group could easily genetically engineer a strain of virus to get the job done. Mass production of asbestos and coordinated release all over the world could work too. What I'm saying is that it's pretty easy and could be done quickly. Nothing more, nothing less. [NEWLINE] [NEWLINE] Now, this is all assuming a few things: [NEWLINE] [NEWLINE] - That if humanity continues, we will eventually cease to exist against our will. [NEWLINE] - That humanity will never have true equality. [NEWLINE] - That war will always continue in some form. [NEWLINE] - That we will continue polluting/degrading the Earth as long as we exist. [NEWLINE] [NEWLINE] If each human somehow became the "God" of their own universe ( though I have no idea how that would come about) then I believe there would be no problem with each individual existence, as controlling your own universe means that no matter what you do you are in the right. Of course, that is all BS/fiction and we live just like all the animals on the Earth, aside from being special in our own little ways. I don't think we as a whole can ever fully agree on anything, and that our disagreements
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As the title says, I feel that that character Boba Fett isn't the awesome bounty hunter many people make him out to be, and his popularity is mostly a result of people jumping on the bandwagon. Now before I go on to explain<mask>, I need to clarify something first: [NEWLINE] [NEWLINE] I'm aware of the Star Wars expanded universe<mask> I haven't read any of it. I'm reluctant to accept any actions of Boba Fett from the expanded universe for 2 reasons: [NEWLINE] [NEWLINE] 1. Like I said, I haven't read anything from the expanded universe, *and I would go<mask> far<mask> to<mask><mask> most people who praise Boba Fett haven't either*. [NEWLINE] [NEWLINE] 2. I only view the movies<mask> canon<mask> the expanded universe,<mask><mask><mask>, was just a money grab by Lucas who was trying to increase his profits from merchandising. Just<mask> Lucas gave the "okay" doesn't mean it should be canon.<mask> that were the case, then we have to deal with the fact that the Star Wars Christmas Special is valid lore. [NEWLINE] [NEWLINE] With that out of the way, I can discuss the "<mask> " part. Boba Fett appears (<mask> an actual bounty hunter) in two of the movies, "The Empire Strikes Back" and "Return of the Jedi". He has minimal screen time compared to other characters and is rarely doing anything at all, let alone something awesome. The vast majority of his screen time is either him standing around trying to look cool, or him walking around with his blaster in hand, still trying to look cool. Aside from those instances, there's a brief moment of him in his ship Slave One getting ready to track the Millennium Falcon (more on this in a bit). Next is him shooting at Luke Skywalker in Cloud City (and missing, I might add). Finally we get to the fun part in Return of the Jedi<mask> he actually gets involved in some action during the sarlac pit scene.<mask> does Boba Fett do<mask> the time for him to finally shine happens? He not only fails to put a stop to a measly 4 prisoners trying to escape with *all* of Jabba's henchmen there helping him,<mask> then he gets his ass handed to him, accidentally, by a **blind** Han Solo.<mask> anything, Boba Fett has charisma and a strong presence,<mask> not talent. [NEWLINE] [NEWLINE] I've brought this up in casual conversations before, and the only counter argument I've heard is such:
Label encoding: <s>As the title says, I feel that that character Boba Fett isn't the awesome bounty hunter many people make him out to be, and his popularity is mostly a result of people jumping on the bandwagon. Now before I go on to explain why, I need to clarify something first: [NEWLINE] [NEWLINE] I'm aware of the Star Wars expanded universe though I haven't read any of it. I'm reluctant to accept any actions of Boba Fett from the expanded universe for 2 reasons: [NEWLINE] [NEWLINE] 1. Like I said, I haven't read anything from the expanded universe, *and I would go so far as to argue that most people who praise Boba Fett haven't either*. [NEWLINE] [NEWLINE] 2. I only view the movies as canon because the expanded universe, in my opinion, was just a money grab by Lucas who was trying to increase his profits from merchandising. Just because Lucas gave the "okay" doesn't mean it should be canon. If that were the case, then we have to deal with the fact that the Star Wars Christmas Special is valid lore. [NEWLINE] [NEWLINE] With that out of the way, I can discuss the " why " part. Boba Fett appears ( as an actual bounty hunter) in two of the movies, "The Empire Strikes Back" and "Return of the Jedi". He has minimal screen time compared to other characters and is rarely doing anything at all, let alone something awesome. The vast majority of his screen time is either him standing around trying to look cool, or him walking around with his blaster in hand, still trying to look cool. Aside from those instances, there's a brief moment of him in his ship Slave One getting ready to track the Millennium Falcon (more on this in a bit). Next is him shooting at Luke Skywalker in Cloud City (and missing, I might add). Finally we get to the fun part in Return of the Jedi where he actually gets involved in some action during the sarlac pit scene. What does Boba Fett do when the time for him to finally shine happens? He not only fails to put a stop to a measly 4 prisoners trying to escape with *all* of Jabba's henchmen there helping him, but then he gets his ass handed to him, accidentally, by a **blind** Han Solo. If anything, Boba Fett has charisma and a strong presence, but not talent. [NEWLINE] [NEWLINE] I've brought this up in casual conversations before, and the only counter argument I've heard is such:
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't think the mental and physical torture of children by their peers in a mandated educational situation is a problem worth fighting. CMV [ENDQ] [NEWLINE] That's<mask> your saying. Really. [NEWLINE] [NEWLINE] [STARTQ] I just feel that bullying is one of those things that is self-correcting. [ENDQ] [NEWLINE] It is. The people being bullied end up feeling like they deserved it and accept their fate. Or they lash out violently. Or they kill themselves.(Hint: Neither of these is a desirable outcome.) [NEWLINE] [NEWLINE] [STARTQ] Over-coddling children and coming to their rescue never teaches people to actually solve their problems [ENDQ] [NEWLINE] The problem of bullying is **not** over-coddling. A major problem with bullying is parents and educators seeing it<mask> "boys will be boys," or "part of growing up." [NEWLINE] [NEWLINE] [STARTQ] I just think that adult intervention in bullying causes the victims to not have to remedy problems themselves, and it never gives the bullies a true, natural realization that<mask> they are doing is wrong. [ENDQ] [NEWLINE] Adult intervention is by and large the only way to bring bullying to a peaceful end. Victims don't peacefully stand up for themselves. Or they maybe try that once. They get depressed or they lash out. [NEWLINE] [NEWLINE] <mask>,<mask><mask> is disapproval from adults an ineffective way for dealing with children who behave badly? [NEWLINE] [NEWLINE] [STARTQ] It just seems that of course it isn't fair,<mask> neither is anything in life. [ENDQ] [NEWLINE] And that means we shouldn't strive to make it fair? Or at least more fair? [NEWLINE] [NEWLINE] [STARTQ] There are plenty of terrible situations in life<mask> you will find yourself being the victim in one way or another, and in those situations you typically need to find a way to solve the problem, avoid the problem, or cope with something. [ENDQ] [NEWLINE] <mask><mask><mask>, children get assistance with a lot of stuff. Like half<mask> not more of<mask> they accomplish is<mask> adults help them. Kids need to learn,<mask> they need to do<mask> in a (relatively) environment. You don't just hand a kid money and expect them to feed themselves. You don't just throw them a book and expect them to learn<mask> to read. [NEWLINE] [NEWLINE] And let me just say this again.<mask> a kid has to find a problem on its own, there are the solution: [NEWLINE] [NEWLINE] * Be unhappy until you change schools (and maybe get bullied again<mask> your self-esteem has become non-existent.) [NEWLINE] [NEWLINE] * Lash out, violently. This can be against the bullies,<mask><mask>
Label encoding: <s> [STARTQ] I don't think the mental and physical torture of children by their peers in a mandated educational situation is a problem worth fighting. CMV [ENDQ] [NEWLINE] That's what your saying. Really. [NEWLINE] [NEWLINE] [STARTQ] I just feel that bullying is one of those things that is self-correcting. [ENDQ] [NEWLINE] It is. The people being bullied end up feeling like they deserved it and accept their fate. Or they lash out violently. Or they kill themselves.(Hint: Neither of these is a desirable outcome.) [NEWLINE] [NEWLINE] [STARTQ] Over-coddling children and coming to their rescue never teaches people to actually solve their problems [ENDQ] [NEWLINE] The problem of bullying is **not** over-coddling. A major problem with bullying is parents and educators seeing it as "boys will be boys," or "part of growing up." [NEWLINE] [NEWLINE] [STARTQ] I just think that adult intervention in bullying causes the victims to not have to remedy problems themselves, and it never gives the bullies a true, natural realization that what they are doing is wrong. [ENDQ] [NEWLINE] Adult intervention is by and large the only way to bring bullying to a peaceful end. Victims don't peacefully stand up for themselves. Or they maybe try that once. They get depressed or they lash out. [NEWLINE] [NEWLINE] Also, since when is disapproval from adults an ineffective way for dealing with children who behave badly? [NEWLINE] [NEWLINE] [STARTQ] It just seems that of course it isn't fair, but neither is anything in life. [ENDQ] [NEWLINE] And that means we shouldn't strive to make it fair? Or at least more fair? [NEWLINE] [NEWLINE] [STARTQ] There are plenty of terrible situations in life where you will find yourself being the victim in one way or another, and in those situations you typically need to find a way to solve the problem, avoid the problem, or cope with something. [ENDQ] [NEWLINE] First of all, children get assistance with a lot of stuff. Like half if not more of what they accomplish is because adults help them. Kids need to learn, but they need to do so in a (relatively) environment. You don't just hand a kid money and expect them to feed themselves. You don't just throw them a book and expect them to learn how to read. [NEWLINE] [NEWLINE] And let me just say this again. If a kid has to find a problem on its own, there are the solution: [NEWLINE] [NEWLINE] * Be unhappy until you change schools (and maybe get bullied again because your self-esteem has become non-existent.) [NEWLINE] [NEWLINE] * Lash out, violently. This can be against the bullies, but also
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm a young, college educated, bisexual woman who needs contraceptives for medical reasons and who doesn't believe in God.  The GOP's stances on social issues are downright appalling to me. I know that people tend to surround themselves with like minded people, be it consciously or not,<mask> the thought of the GOP being unable to ever win another presidential election has crawled into my head. ***(TL;DR in bold.)*** [NEWLINE] [NEWLINE] ***The social issues I'm talking about are:*** [NEWLINE] [NEWLINE] * ***LGBT issues*** such<mask> same sex marriage, religious freedom laws, "bathroom bills" involving trans people, etc. [NEWLINE] [NEWLINE] * Their black and white stance on ***abortion***- some of them want no exceptions for cases of terrible fetal deformities, threats to the life of the mother (Scott Walker in the last debate), rape, incest, etc. [NEWLINE] [NEWLINE] * Fighting against ***contraceptives*** and the morning after pill that could prevent unwanted pregnancies and prevent abortions [NEWLINE] [NEWLINE] * Pushing ***abstinence only sex ed.*** [37 states allow for medically inaccurate information to be taught to students<mask> "fact",<mask><mask><mask> it scares them out of having sex. My school was one of the ones that taught blatantly incorrect information.]( [URL] ) [NEWLINE] [NEWLINE] * Claiming that anthroprogenic ***climate change*** isn't real, and pushing the teaching of ***intelligent design*** in public schools [NEWLINE] [NEWLINE] * ***Marijuana legalization*** [NEWLINE] [NEWLINE] * Being<mask> ***bible/religion*** based<mask> the population seems to be shifting away from religion [NEWLINE] [NEWLINE] ***Please please please change my view! I live in Ohio,<mask> I really can't let myself slide into apathy thinking that the GOP has no hope of winning anything! You don't need to change my views on the social issues at hand, just convince me that another Republican president isn't out of the question!*** [NEWLINE] [NEWLINE] Edit 1: Wow! Lots of responses- awesome! It's going to take me a few minutes to sift through all this. :D [NEWLINE] [NEWLINE] Edit 2: Delta awarded-<mask> a candidate like Hillary gets the democratic nomination and young people who are frustrated with government express that frustration by just staying home on election day, it will definitely give Republicans an advantage, no matter their stance on the issues. [NEWLINE] [NEWLINE] Edit 3: Please cite sources<mask> possible. Feelings and anecdotes don't hold the same weight<mask> things like polling data from reputable institutions, academic studies, etc. [NEWLINE] [NEWLINE] Edit 4: Another delta awarded
Label encoding: <s>I'm a young, college educated, bisexual woman who needs contraceptives for medical reasons and who doesn't believe in God.  The GOP's stances on social issues are downright appalling to me. I know that people tend to surround themselves with like minded people, be it consciously or not, so the thought of the GOP being unable to ever win another presidential election has crawled into my head. ***(TL;DR in bold.)*** [NEWLINE] [NEWLINE] ***The social issues I'm talking about are:*** [NEWLINE] [NEWLINE] * ***LGBT issues*** such as same sex marriage, religious freedom laws, "bathroom bills" involving trans people, etc. [NEWLINE] [NEWLINE] * Their black and white stance on ***abortion***- some of them want no exceptions for cases of terrible fetal deformities, threats to the life of the mother (Scott Walker in the last debate), rape, incest, etc. [NEWLINE] [NEWLINE] * Fighting against ***contraceptives*** and the morning after pill that could prevent unwanted pregnancies and prevent abortions [NEWLINE] [NEWLINE] * Pushing ***abstinence only sex ed.*** [37 states allow for medically inaccurate information to be taught to students as "fact", so long as it scares them out of having sex. My school was one of the ones that taught blatantly incorrect information.]( [URL] ) [NEWLINE] [NEWLINE] * Claiming that anthroprogenic ***climate change*** isn't real, and pushing the teaching of ***intelligent design*** in public schools [NEWLINE] [NEWLINE] * ***Marijuana legalization*** [NEWLINE] [NEWLINE] * Being so ***bible/religion*** based when the population seems to be shifting away from religion [NEWLINE] [NEWLINE] ***Please please please change my view! I live in Ohio, so I really can't let myself slide into apathy thinking that the GOP has no hope of winning anything! You don't need to change my views on the social issues at hand, just convince me that another Republican president isn't out of the question!*** [NEWLINE] [NEWLINE] Edit 1: Wow! Lots of responses- awesome! It's going to take me a few minutes to sift through all this. :D [NEWLINE] [NEWLINE] Edit 2: Delta awarded- if a candidate like Hillary gets the democratic nomination and young people who are frustrated with government express that frustration by just staying home on election day, it will definitely give Republicans an advantage, no matter their stance on the issues. [NEWLINE] [NEWLINE] Edit 3: Please cite sources if possible. Feelings and anecdotes don't hold the same weight as things like polling data from reputable institutions, academic studies, etc. [NEWLINE] [NEWLINE] Edit 4: Another delta awarded
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I take issue with your third premise, OP, on two counts. [NEWLINE] [NEWLINE] <mask>, I would<mask><mask> the majority of drunken drivers did not make an informed choice (re: "even<mask> they understand the risks") to commit their particular crime. [NEWLINE] <mask>, well, this is a larger issue with your post. Societies restrict the behavior of individuals for a lot of reasons, not all of them good,<mask> usually<mask> those behaviors in some way infringe on the rights of others. Someone may want to drive drunk or enjoy drunken driving<mask> a recreational activity. That doesn't validate the act of drunken driving, which is more likely to kill other drivers, pedestrians, bikers, etc. than sober driving. Similarly, to choose an extreme case, we restrict the ability of citizens to murder other citizens<mask> we believe people have a right to live without being murdered. The desires of the murderer are irrelevant. A freedom to harm others has been removed,<mask> that people are free to live with a greater likelihood of not being harmed. That's the balance a good law creates. [NEWLINE] [NEWLINE] You posit that "You are equally free to choose to drive without drinking and driving laws. You may choose not to<mask> it is less safe,<mask> you are not coerced..." [NEWLINE] [NEWLINE] I mean, technically, yes,<mask> this is a little strange to see in a post whose main complaint is against intellectual dishonesty. No outside entity would be restricting me from driving in a drunken-driver-filled world,<mask> saying 'You're free to make this choice that is more likely to kill you' does not strike me<mask> a truly free choice.<mask> I have a knife at your throat, OP, and am telling you to sign a contract or be killed, by your argument this would still be a free choice. I wouldn't be restricting your actions. You could still choose not to sign and subsequently die. I could still choose to go on the road and subsequently get hit by a drunken driver. It's all about<mask> you define<mask> a restriction, and for me, knowing that a given behavior carries a higher risk will usually restrict my actions, whether or not an external entity is imposing that restriction. [NEWLINE] [NEWLINE] I might go even further and say that in both cases, laws and no laws, that there is an external restriction. One restriction is a legal penalty, the other is an increased risk of death or injury. In one case, a formal entity is imposing it; in others, the state of the world is imposing it. In both cases
Label encoding: <s>I take issue with your third premise, OP, on two counts. [NEWLINE] [NEWLINE] Firstly, I would argue that the majority of drunken drivers did not make an informed choice (re: "even if they understand the risks") to commit their particular crime. [NEWLINE] Secondly, well, this is a larger issue with your post. Societies restrict the behavior of individuals for a lot of reasons, not all of them good, but usually because those behaviors in some way infringe on the rights of others. Someone may want to drive drunk or enjoy drunken driving as a recreational activity. That doesn't validate the act of drunken driving, which is more likely to kill other drivers, pedestrians, bikers, etc. than sober driving. Similarly, to choose an extreme case, we restrict the ability of citizens to murder other citizens because we believe people have a right to live without being murdered. The desires of the murderer are irrelevant. A freedom to harm others has been removed, so that people are free to live with a greater likelihood of not being harmed. That's the balance a good law creates. [NEWLINE] [NEWLINE] You posit that "You are equally free to choose to drive without drinking and driving laws. You may choose not to because it is less safe, but you are not coerced..." [NEWLINE] [NEWLINE] I mean, technically, yes, but this is a little strange to see in a post whose main complaint is against intellectual dishonesty. No outside entity would be restricting me from driving in a drunken-driver-filled world, but saying 'You're free to make this choice that is more likely to kill you' does not strike me as a truly free choice. If I have a knife at your throat, OP, and am telling you to sign a contract or be killed, by your argument this would still be a free choice. I wouldn't be restricting your actions. You could still choose not to sign and subsequently die. I could still choose to go on the road and subsequently get hit by a drunken driver. It's all about what you define as a restriction, and for me, knowing that a given behavior carries a higher risk will usually restrict my actions, whether or not an external entity is imposing that restriction. [NEWLINE] [NEWLINE] I might go even further and say that in both cases, laws and no laws, that there is an external restriction. One restriction is a legal penalty, the other is an increased risk of death or injury. In one case, a formal entity is imposing it; in others, the state of the world is imposing it. In both cases
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I want to say two things: [NEWLINE] [NEWLINE] First, let me tell you a story about the first time I ever did shrooms. [NEWLINE] [NEWLINE] I was raised in a really, really abusive household. My father raped me and beat me and locked me in closets and mentally battered me. My mom was an alcoholic. My sister hated me. I hated myself.<mask> I did shrooms, all of that was different. It was the first time in my life I actually felt<mask> 'being happy' was. It was completely new. Things were brighter, I could laugh, I *felt* things. It was<mask> the first time that I thought I could actually have sex with someone and not feel ashamed. [NEWLINE] [NEWLINE] I don't advocate for drug use,<mask> I don't think that temporary happiness is bad. I couldn't just 'fix it',<mask> it was my life. I couldn't 'fix' my mother or 'fix' my father. I was too afraid to call the police, and my dad manipulated me into thinking that it was *normal* anyway. I thought everyone's life was like mine, and everyone wanted to kill themselves.<mask> this gave me an experience to hold on to. Did I abuse the drugs in the future looking for this feeling? Yeah. Did it keep me alive? I like to think<mask>. I honestly think that<mask> I had never felt this ephemeral happiness,<mask> I had honestly thought that life was just the misery I lived in, I would have killed myself. [NEWLINE] [NEWLINE] <mask>, I'd like to talk about your view on unhappiness. Depression and anxiety are actual disorders. They are the result of a lot of things,<mask> mainly there's a lack of certain chemicals in your brain. This is<mask> SSRIs and other antidepressants and anti anxiety medication exists. With therapy, drugs, and other things, it's totally possible to recover from depression (just<mask> you know: they say you're in remission from depression. Not cured). [NEWLINE] [NEWLINE] <mask>, you can't just tell someone to 'fix' their lives. People are complicated, and their reasons for doing things are complicated. A person in an abusive relationship (usually) can't just get up and leave, and often don't realize they're in it to begin with. A college student who's depressed<mask> they don't have enough serotonin in their brain can't just 'fix it' and make it better. Sure, yes, you can exercise and you can eat certain things and you can try real hard and
Label encoding: <s>I want to say two things: [NEWLINE] [NEWLINE] First, let me tell you a story about the first time I ever did shrooms. [NEWLINE] [NEWLINE] I was raised in a really, really abusive household. My father raped me and beat me and locked me in closets and mentally battered me. My mom was an alcoholic. My sister hated me. I hated myself. When I did shrooms, all of that was different. It was the first time in my life I actually felt what 'being happy' was. It was completely new. Things were brighter, I could laugh, I *felt* things. It was also the first time that I thought I could actually have sex with someone and not feel ashamed. [NEWLINE] [NEWLINE] I don't advocate for drug use, but I don't think that temporary happiness is bad. I couldn't just 'fix it', because it was my life. I couldn't 'fix' my mother or 'fix' my father. I was too afraid to call the police, and my dad manipulated me into thinking that it was *normal* anyway. I thought everyone's life was like mine, and everyone wanted to kill themselves. But this gave me an experience to hold on to. Did I abuse the drugs in the future looking for this feeling? Yeah. Did it keep me alive? I like to think so. I honestly think that if I had never felt this ephemeral happiness, if I had honestly thought that life was just the misery I lived in, I would have killed myself. [NEWLINE] [NEWLINE] Secondly, I'd like to talk about your view on unhappiness. Depression and anxiety are actual disorders. They are the result of a lot of things, but mainly there's a lack of certain chemicals in your brain. This is why SSRIs and other antidepressants and anti anxiety medication exists. With therapy, drugs, and other things, it's totally possible to recover from depression (just so you know: they say you're in remission from depression. Not cured). [NEWLINE] [NEWLINE] Therefore, you can't just tell someone to 'fix' their lives. People are complicated, and their reasons for doing things are complicated. A person in an abusive relationship (usually) can't just get up and leave, and often don't realize they're in it to begin with. A college student who's depressed because they don't have enough serotonin in their brain can't just 'fix it' and make it better. Sure, yes, you can exercise and you can eat certain things and you can try real hard and
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Before pointing out the reasons for their existence, I do understand that the deficit is<mask> of the century 0-100 AD, which is referred to<mask> the "1st Century", making the 100s the "2nd." [NEWLINE] [NEWLINE] <mask>, it is better to stick with one term or another, especially<mask> little reference is made at all to Ancient times. It stands to reason that<mask> it comes to referencing centuries versus specific years, that the numerical system rather than the ordinal system is obviously superior; we can both designate the "1900s" and "1914",<mask> we cannot designate an individual year with "20th century." [NEWLINE] [NEWLINE] This system<mask> makes it more difficult for non-English speakers to understand,<mask> they have enough on their plate figuring out our mish-mash of a language, let alone<mask> many dates they could get wrong during translation. Hell, I sometimes get it wrong<mask> I'm not paying enough attention, especially<mask> it gets back into less recent history - English speakers know automatically that "19th century" and "20th century" refer to 1800s and 1900s,<mask><mask> many of us automatically make the shift<mask> we hear "7th century"? [NEWLINE] [NEWLINE] The only issue I can see is that<mask> discussing Roman-era history, around the time 100 BC-100 AD, it might make referring to the "1st Century", 1-100 AD a little more difficult.<mask>, the problem persists even in BC with the confusion of the ordinal century lagging behind the actual year date. [NEWLINE] [NEWLINE] My CMV is that this is a confusing system and bad for international students, casual readers, and even for historians and archaeologists. The below proposal is just a suggestion for discussion, and shredding it,<mask> I'm sure this subreddit will, will **not necessarily** change my view on the whole subject. [NEWLINE] [NEWLINE] <mask>,<mask><mask> that,<mask> referencing these two centuries, it might be more instructive to refer to them in a similar manner that some refer to the most recent full decade, the "oughts".<mask><mask> 2000 is technically part of the 1990s, this means that 2001-2010 are referred to on occasion<mask> the "oughts"<mask> referring to decades, in lieu of the '20s, '30s, and<mask> forth.<mask> we should refer to the 1st century AD<mask> "the ought century" AD, and then continue with 100s AD, 200s AD, and same for BC. [NEWLINE] [NEWLINE] That was just a suggestion. On
Label encoding: <s>Before pointing out the reasons for their existence, I do understand that the deficit is because of the century 0-100 AD, which is referred to as the "1st Century", making the 100s the "2nd." [NEWLINE] [NEWLINE] However, it is better to stick with one term or another, especially when little reference is made at all to Ancient times. It stands to reason that when it comes to referencing centuries versus specific years, that the numerical system rather than the ordinal system is obviously superior; we can both designate the "1900s" and "1914", but we cannot designate an individual year with "20th century." [NEWLINE] [NEWLINE] This system also makes it more difficult for non-English speakers to understand, as they have enough on their plate figuring out our mish-mash of a language, let alone how many dates they could get wrong during translation. Hell, I sometimes get it wrong when I'm not paying enough attention, especially when it gets back into less recent history - English speakers know automatically that "19th century" and "20th century" refer to 1800s and 1900s, but how many of us automatically make the shift when we hear "7th century"? [NEWLINE] [NEWLINE] The only issue I can see is that when discussing Roman-era history, around the time 100 BC-100 AD, it might make referring to the "1st Century", 1-100 AD a little more difficult. However, the problem persists even in BC with the confusion of the ordinal century lagging behind the actual year date. [NEWLINE] [NEWLINE] My CMV is that this is a confusing system and bad for international students, casual readers, and even for historians and archaeologists. The below proposal is just a suggestion for discussion, and shredding it, as I'm sure this subreddit will, will **not necessarily** change my view on the whole subject. [NEWLINE] [NEWLINE] Therefore, I think that, when referencing these two centuries, it might be more instructive to refer to them in a similar manner that some refer to the most recent full decade, the "oughts". Given that 2000 is technically part of the 1990s, this means that 2001-2010 are referred to on occasion as the "oughts" when referring to decades, in lieu of the '20s, '30s, and so forth. So we should refer to the 1st century AD as "the ought century" AD, and then continue with 100s AD, 200s AD, and same for BC. [NEWLINE] [NEWLINE] That was just a suggestion. On
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Even<mask> you don't trust the police,<mask> is disobeying an officer's command a good idea for you personally? In<mask> scenario will disobeying their orders make things better for you? [ENDQ] [NEWLINE] A police officer could reasonably knock on your door and say "Open this door." Now, you have the right to privacy, and you don't have to open your door to police<mask> they lack a warrant. There may be any reason they are asking you to open the door: To speak with you about a crime in the area (gathering information), to investigate a possible crime in your home (suspicion), or even to invite you to the local police bake sale (community outreach).<mask> you are under no legal obligation in this context to comply. You can tell the police officer to go away (refusal), you can ignore him, or you could open the door (comply). [NEWLINE] [NEWLINE] <mask> you comply, and you have an illegal substance, weapon, or even something that looks like one in your house visible from the door, the police officer has reasonable cause to search your home-- immediately. Even<mask> the police officer *thinks* he smells an illicit substance, he can search your home.<mask> compliance in this case absolutely has negative consequences that are avoided by refusal. [NEWLINE] [NEWLINE] <mask><mask> you have done nothing wrong/illegal? You should still not comply, and here's<mask> : Are you the only person who has ever been in your home?<mask> not, it is possible that someone other than you has hidden an illegal substance or item in your home.<mask> the police search your home on other grounds, they may find it, landing you in trouble. Yes, your name may be cleared later,<mask> for now you are likely facing a possession charge. [NEWLINE] [NEWLINE] I believe this gives an example of<mask> disobeying a police officer is to your benefit. Now,<mask> the police officer has reasonable suspicion before you open the door, and you refuse, they will get a warrant and search your home anyway.<mask> you always have a right to refuse certain orders from the police, and this is one of them. [NEWLINE] [NEWLINE] Other examples include: [NEWLINE] [NEWLINE] 1. [<mask> you aren't driving and police do not suspect you of a crime, you are not required to produce identification upon their request/order.]( [URL] /)<mask> they are investigating a crime nearby and simply want a record of you having been in the area, they don't have the right to require it of you-- and you shouldn't
Label encoding: <s> [STARTQ] Even if you don't trust the police, how is disobeying an officer's command a good idea for you personally? In what scenario will disobeying their orders make things better for you? [ENDQ] [NEWLINE] A police officer could reasonably knock on your door and say "Open this door." Now, you have the right to privacy, and you don't have to open your door to police if they lack a warrant. There may be any reason they are asking you to open the door: To speak with you about a crime in the area (gathering information), to investigate a possible crime in your home (suspicion), or even to invite you to the local police bake sale (community outreach). But you are under no legal obligation in this context to comply. You can tell the police officer to go away (refusal), you can ignore him, or you could open the door (comply). [NEWLINE] [NEWLINE] If you comply, and you have an illegal substance, weapon, or even something that looks like one in your house visible from the door, the police officer has reasonable cause to search your home-- immediately. Even if the police officer *thinks* he smells an illicit substance, he can search your home. So compliance in this case absolutely has negative consequences that are avoided by refusal. [NEWLINE] [NEWLINE] What if you have done nothing wrong/illegal? You should still not comply, and here's why : Are you the only person who has ever been in your home? If not, it is possible that someone other than you has hidden an illegal substance or item in your home. If the police search your home on other grounds, they may find it, landing you in trouble. Yes, your name may be cleared later, but for now you are likely facing a possession charge. [NEWLINE] [NEWLINE] I believe this gives an example of where disobeying a police officer is to your benefit. Now, if the police officer has reasonable suspicion before you open the door, and you refuse, they will get a warrant and search your home anyway. But you always have a right to refuse certain orders from the police, and this is one of them. [NEWLINE] [NEWLINE] Other examples include: [NEWLINE] [NEWLINE] 1. [ If you aren't driving and police do not suspect you of a crime, you are not required to produce identification upon their request/order.]( [URL] /) If they are investigating a crime nearby and simply want a record of you having been in the area, they don't have the right to require it of you-- and you shouldn't
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just about every religion is in its purest form, incompatible with modern Western society.<mask>, a lot of religious followers adapt<mask> their culture shifts to the more modern culture. Islam is no exception. [NEWLINE] [NEWLINE] I live on Long Island in New York. There's a huge Muslim population. There's<mask> a huge Jewish population, huge Chrisian population, huge Buddhist population, huge Hindu population...basically, very diverse in religion. And bar none, all of these religions have believers who are on a spectrum from very orthodox/conservative to very liberal/progressive in<mask> they reconcile old faiths with moderns ideas and society. [NEWLINE] [NEWLINE] It's a microcosm for the world,<mask> at the end of the day, every religion has differences in<mask> they are observed.<mask> an atheist who was raised Jewish, my family was very modern, and view the Torah<mask> a collection of books that were written to help guide Jews back in the days of the old Kingdom of Judea. Some of the stories still give good lessons,<mask> the essence of Judaism wasn't following the laws and rules set forth in the Tanakh,<mask> to live life in a way God would approve. Other Jews who lived in the same neighborhood lived their lives by the letter of Judaic law, keeping kosher, observing the Sabbath, etc. [NEWLINE] [NEWLINE] I've seen that with my Muslim friends. One of them is very strict in following the Quran and the Haddith, only eating halal meat, praying 5 times a day, etc. Most of them are less strict and follow a much looser code of conduct that was basically the same way I was raised. And the same could be said for my Christian friends. [NEWLINE] [NEWLINE] The problem is that for a lot of people, the only image of Islam they are exposed to isn't the people on Long Island who are harmless,<mask> the people who live in the MENA regions. Their religions might<mask> well be completely different,<mask> the ones who live on Long Island and are raised with Western culture have assimilated their faith with their culture. Those in MENA have done the same,<mask> they aren't assimilating with Western culture<mask> with a harsher culture that is born out of lower socioeconomic standards, a history of cultural oppression, etc. [NEWLINE] [NEWLINE] A better comparison for the MENA Muslims who make headlines is the Christians in sub-Saharan Africa. Culturally they are very similar in many ways (much more orthodox/conservative, fewer women's rights, prevalence of FGM, lack of literacy, etc.).
Label encoding: <s>Just about every religion is in its purest form, incompatible with modern Western society. However, a lot of religious followers adapt as their culture shifts to the more modern culture. Islam is no exception. [NEWLINE] [NEWLINE] I live on Long Island in New York. There's a huge Muslim population. There's also a huge Jewish population, huge Chrisian population, huge Buddhist population, huge Hindu population...basically, very diverse in religion. And bar none, all of these religions have believers who are on a spectrum from very orthodox/conservative to very liberal/progressive in how they reconcile old faiths with moderns ideas and society. [NEWLINE] [NEWLINE] It's a microcosm for the world, because at the end of the day, every religion has differences in how they are observed. As an atheist who was raised Jewish, my family was very modern, and view the Torah as a collection of books that were written to help guide Jews back in the days of the old Kingdom of Judea. Some of the stories still give good lessons, but the essence of Judaism wasn't following the laws and rules set forth in the Tanakh, but to live life in a way God would approve. Other Jews who lived in the same neighborhood lived their lives by the letter of Judaic law, keeping kosher, observing the Sabbath, etc. [NEWLINE] [NEWLINE] I've seen that with my Muslim friends. One of them is very strict in following the Quran and the Haddith, only eating halal meat, praying 5 times a day, etc. Most of them are less strict and follow a much looser code of conduct that was basically the same way I was raised. And the same could be said for my Christian friends. [NEWLINE] [NEWLINE] The problem is that for a lot of people, the only image of Islam they are exposed to isn't the people on Long Island who are harmless, but the people who live in the MENA regions. Their religions might as well be completely different, because the ones who live on Long Island and are raised with Western culture have assimilated their faith with their culture. Those in MENA have done the same, but they aren't assimilating with Western culture but with a harsher culture that is born out of lower socioeconomic standards, a history of cultural oppression, etc. [NEWLINE] [NEWLINE] A better comparison for the MENA Muslims who make headlines is the Christians in sub-Saharan Africa. Culturally they are very similar in many ways (much more orthodox/conservative, fewer women's rights, prevalence of FGM, lack of literacy, etc.).
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] CDC numbers say that the flu shot is about 40 to 60 % effective,<mask><mask> are they measuring that versus people who don't get the shot at all? It seems to me your chances of getting the flu are higher WITH the shot<mask> you are absolutely exposing yourself to the flu versus maybe not being exposed at all. Do they actually have scientific studies to back their claims up? [ENDQ] [NEWLINE] [All of this is answered on this site.]( [URL] )  To summerize, the rate of effectiveness is determined using a double blind test (the standard for medical science).  The percentage given is the reduction in the likelihood of a person getting the flu with the vaccine<mask> compared to without. <mask><mask> they say it is 40% to 60% effective, that means<mask> you get the vaccine you are about half<mask> likely to get the flu<mask><mask> you didn't. <mask> for vaccines making you less likely to get the desies dispite exposing you to it, [this principle has been known<mask> the 1700's]( [URL] #History) and is kind of old news in the realm of science. [NEWLINE] [NEWLINE] [STARTQ] Not only that<mask> the vaccines have mercury in them with is a known poison that doesn't leave your system easily (<mask> at all). [ENDQ] [NEWLINE] The mercury in vaccines is contained in a compound known<mask> thimerosal.  [Thimerosal has an LD50 of 75 mg/kg,]( [URL].php?msdsId=9925236) meaning that a 150 pound person will need about 5103 mg for a 50% chance of a lethal dose.  The amount of thimerosal in the vaccines that have it (not all do and you can easily request one that does not have it) have only a trace amount [(0.01%)]( [URL] ),<mask> you will need to get a massive volume of the vaccine in your system for it to even have an affect on you, let alone be dangerous. [NEWLINE] [NEWLINE] [STARTQ] People say "herd immunity" which makes sense for things like malaria and polio,<mask> the flu evolves every year and they're just guessing on the variety they put in the flu shot. Most importantly the side effects are severe and fairly common including some side effects that are far worse than almost any that you can get from the actual flu. [ENDQ] [NEWLINE] The method to determine exactly<mask> strain to create a vaccine for has some pretty complicated math invovled (which I admit I don't fully understand), [<mask> the criteria they look at is explained here.]( [URL] #virus
Label encoding: <s> [STARTQ] CDC numbers say that the flu shot is about 40 to 60 % effective, but how are they measuring that versus people who don't get the shot at all? It seems to me your chances of getting the flu are higher WITH the shot since you are absolutely exposing yourself to the flu versus maybe not being exposed at all. Do they actually have scientific studies to back their claims up? [ENDQ] [NEWLINE] [All of this is answered on this site.]( [URL] )  To summerize, the rate of effectiveness is determined using a double blind test (the standard for medical science).  The percentage given is the reduction in the likelihood of a person getting the flu with the vaccine as compared to without.  So if they say it is 40% to 60% effective, that means if you get the vaccine you are about half as likely to get the flu as if you didn't.  As for vaccines making you less likely to get the desies dispite exposing you to it, [this principle has been known since the 1700's]( [URL] #History) and is kind of old news in the realm of science. [NEWLINE] [NEWLINE] [STARTQ] Not only that but the vaccines have mercury in them with is a known poison that doesn't leave your system easily ( if at all). [ENDQ] [NEWLINE] The mercury in vaccines is contained in a compound known as thimerosal.  [Thimerosal has an LD50 of 75 mg/kg,]( [URL].php?msdsId=9925236) meaning that a 150 pound person will need about 5103 mg for a 50% chance of a lethal dose.  The amount of thimerosal in the vaccines that have it (not all do and you can easily request one that does not have it) have only a trace amount [(0.01%)]( [URL] ), so you will need to get a massive volume of the vaccine in your system for it to even have an affect on you, let alone be dangerous. [NEWLINE] [NEWLINE] [STARTQ] People say "herd immunity" which makes sense for things like malaria and polio, but the flu evolves every year and they're just guessing on the variety they put in the flu shot. Most importantly the side effects are severe and fairly common including some side effects that are far worse than almost any that you can get from the actual flu. [ENDQ] [NEWLINE] The method to determine exactly what strain to create a vaccine for has some pretty complicated math invovled (which I admit I don't fully understand), [ but the criteria they look at is explained here.]( [URL] #virus
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, analysis of published preliminary casualty lists demonstrates very clearly that Israel is targeting combatants.  [This blog]( [URL] ) did an analysis of the casualty list published by *Al Jazeera*.  It found males make up ~82% of the casualties (~51% of population).  Of the men killed, more than 66% were between the ages of 18-38.  <mask><mask> children under the age of 14 make up ~44% of Gaza's population, those under the age of 18 make up just 18% of Palestinian casualties in this conflict<mask> far.  The point is not that every man aged 18-38 that's killed has been a combatant (just<mask> not every woman or 17 year old male is a non-combatant). <mask> unfortunately,<mask> there is very little accurate reporting of combat vs. civilian deaths in these conflicts, it's a good metric for looking at targeting. <mask> Israel was indiscriminately or intentionally targeting civilians, you would expect to see an astronomically higher proportion of children killed.  This is exacerbated by the fact that children are less able to withstand severe trauma,<mask> their death rate from injury should inflate their numbers further (this<mask> true of the elderly, and is visible in the analysis: ~4.7% killed over 65 vs. 2.6% of population).  You would<mask> expect much more balance of men vs women, i.e. much closer to the 51-49% respective proportion. <mask> you actual see is that the casualty statistics do not correspond to the population data.  It skews towards the common combatant sub-groups quite heavily. [NEWLINE] [NEWLINE] <mask>, you are entirely misunderstanding<mask> "proportionate response/force" means.  It has nothing to do with relative counts of casualties.  From the wikipedia article on [Proportionality]( [URL] %28law%29#International_humanitarian_law)(my emphasis added): [NEWLINE] [NEWLINE] [STARTQ] Luis Moreno-Ocampo was the Chief Prosecutor at the International Criminal Court who investigated allegations of war crimes during the 2003 invasion of Iraq. He published an open letter containing his findings; in a section titled "Allegations concerning War Crimes", he elucidates this use of proportionality: [ENDQ] Under international humanitarian law and the Rome Statute, the death of civilians during an armed conflict, no matter<mask> grave and regrettable, does not in itself constitute a war crime. International humanitarian law and the Rome Statute permit belligerents to carry out proportionate attacks against military
Label encoding: <s>Firstly, analysis of published preliminary casualty lists demonstrates very clearly that Israel is targeting combatants.  [This blog]( [URL] ) did an analysis of the casualty list published by *Al Jazeera*.  It found males make up ~82% of the casualties (~51% of population).  Of the men killed, more than 66% were between the ages of 18-38.   Even though children under the age of 14 make up ~44% of Gaza's population, those under the age of 18 make up just 18% of Palestinian casualties in this conflict so far.  The point is not that every man aged 18-38 that's killed has been a combatant (just as not every woman or 17 year old male is a non-combatant).  But unfortunately, as there is very little accurate reporting of combat vs. civilian deaths in these conflicts, it's a good metric for looking at targeting.  If Israel was indiscriminately or intentionally targeting civilians, you would expect to see an astronomically higher proportion of children killed.  This is exacerbated by the fact that children are less able to withstand severe trauma, so their death rate from injury should inflate their numbers further (this also true of the elderly, and is visible in the analysis: ~4.7% killed over 65 vs. 2.6% of population).  You would also expect much more balance of men vs women, i.e. much closer to the 51-49% respective proportion.  What you actual see is that the casualty statistics do not correspond to the population data.  It skews towards the common combatant sub-groups quite heavily. [NEWLINE] [NEWLINE] Secondly, you are entirely misunderstanding what "proportionate response/force" means.  It has nothing to do with relative counts of casualties.  From the wikipedia article on [Proportionality]( [URL] %28law%29#International_humanitarian_law)(my emphasis added): [NEWLINE] [NEWLINE] [STARTQ] Luis Moreno-Ocampo was the Chief Prosecutor at the International Criminal Court who investigated allegations of war crimes during the 2003 invasion of Iraq. He published an open letter containing his findings; in a section titled "Allegations concerning War Crimes", he elucidates this use of proportionality: [ENDQ] Under international humanitarian law and the Rome Statute, the death of civilians during an armed conflict, no matter how grave and regrettable, does not in itself constitute a war crime. International humanitarian law and the Rome Statute permit belligerents to carry out proportionate attacks against military
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think your criticism of Singer's argument is intellectually honest. Singer's argument is based on preferences. With the person in the coma the preference to live is assumed, unless the act that caused them to go into the coma was some type of attempted suicide. [NEWLINE] [NEWLINE] In the case of the fetus, or newborn baby, the ability to have preferences isn't there.<mask> the mothers preference wins by default. [NEWLINE] [NEWLINE] The only rebuttal you could really make,<mask><mask><mask> I can see, would be to ask<mask> we don't assume the preferences of a newborn. For the reason<mask> we don't I would<mask><mask> the only reason<mask> we assume the preferences of the coma patient is<mask> they were previously'sentient' and obviously had a desire to continue living (<mask> they didn't attempt to end their life). [NEWLINE] [NEWLINE] With the fetus/newborn it doesn't matter<mask> it never had the ability to develop preferences in the first place. [NEWLINE] [NEWLINE] I<mask> tend to feel that these abortion debates get too philosophical and people tend to forget that the underlying idea of modern ethics is to create a society we want to live in. There is no ultimate 'right answer'. [NEWLINE] [NEWLINE] In short, the evidence shows us that the killing an unborn baby, or even one that has just been born, doesn't harm the baby in the same way you would kill<mask> you kill your neighbor in cold blood. Even<mask> the baby/fetus has the circuity necessary to experience pain the science shows us that it's experience of pain is far diminished compared to that of something like an older human. The pain is of course only one facet. Other things enter into the equation in the case of you killing your neighbor including,<mask> not limited to, his clear desire to live, his fear of death, etc. [NEWLINE] [NEWLINE] A fetus or newborn simply doesn't have the cognitive capacity to experience these things. The coma counter example, even<mask> you want to allow for it to work, isn't necessary and just isn't applicable. We're not trying to kill potentially viable coma patients.<mask> we are trying to kill is unborn babies.<mask>?<mask> they're not wanted and killing them doesn't significantly effect their experience<mask> they don't have any type of significant experience to begin with. [NEWLINE] [NEWLINE] <mask> for<mask> we draw the line?<mask> we wanted to be EXACT we would need to do a case by case basis. This isn't feasible.<mask><mask> not simply draw the line at before the baby is out? Or 6 months into the
Label encoding: <s>I don't think your criticism of Singer's argument is intellectually honest. Singer's argument is based on preferences. With the person in the coma the preference to live is assumed, unless the act that caused them to go into the coma was some type of attempted suicide. [NEWLINE] [NEWLINE] In the case of the fetus, or newborn baby, the ability to have preferences isn't there. Therefore the mothers preference wins by default. [NEWLINE] [NEWLINE] The only rebuttal you could really make, as far as I can see, would be to ask why we don't assume the preferences of a newborn. For the reason why we don't I would argue that the only reason why we assume the preferences of the coma patient is because they were previously'sentient' and obviously had a desire to continue living ( if they didn't attempt to end their life). [NEWLINE] [NEWLINE] With the fetus/newborn it doesn't matter because it never had the ability to develop preferences in the first place. [NEWLINE] [NEWLINE] I also tend to feel that these abortion debates get too philosophical and people tend to forget that the underlying idea of modern ethics is to create a society we want to live in. There is no ultimate 'right answer'. [NEWLINE] [NEWLINE] In short, the evidence shows us that the killing an unborn baby, or even one that has just been born, doesn't harm the baby in the same way you would kill when you kill your neighbor in cold blood. Even if the baby/fetus has the circuity necessary to experience pain the science shows us that it's experience of pain is far diminished compared to that of something like an older human. The pain is of course only one facet. Other things enter into the equation in the case of you killing your neighbor including, but not limited to, his clear desire to live, his fear of death, etc. [NEWLINE] [NEWLINE] A fetus or newborn simply doesn't have the cognitive capacity to experience these things. The coma counter example, even if you want to allow for it to work, isn't necessary and just isn't applicable. We're not trying to kill potentially viable coma patients. What we are trying to kill is unborn babies. Why? Because they're not wanted and killing them doesn't significantly effect their experience because they don't have any type of significant experience to begin with. [NEWLINE] [NEWLINE] As for where we draw the line? If we wanted to be EXACT we would need to do a case by case basis. This isn't feasible. So why not simply draw the line at before the baby is out? Or 6 months into the
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The brain is a pattern maker and pattern matcher. And the patterns it makes, it finds in the outside world with ease. And the more deeply held and treasured a pattern (belief), the more it must find them out there in the world,<mask> that they can be matched are validated to be true. And the more invested it becomes, the more it sees<mask> it believes, even<mask> "out there" is the very poorest of matches. [NEWLINE] [NEWLINE] To challenge your own beliefs is to threaten your very identity! For it's our patterns/beliefs/values that make us who we are, and to change them for something else is an act of self-destruction without the guarantee of catharsis at the other end!<mask> kudos to you! [NEWLINE] [NEWLINE] My Dad was a mathematician by training and a hobby astrologer/astronomer, and he made thousands of charts. He sort a statistical proof of astrology for about a decade, collecting thousands of birth dates/times/locations for people in different professions. I remember there was a famous sportsmen set, and I remember he had a huge amount of data on volcanic eruptions and earthquakes. He eventually discounted the popular field<mask> completely rubbish, and was left with two statistical correlations (mars in some house for sportsmen and for earthquakes!) That was 20 years ago, and<mask> remains from my perspective, and conversations with him, is his spiritual sense that "existential causation" is "top down", not "bottom-up"... [NEWLINE] [NEWLINE] Ahh, Dad. [NEWLINE] [NEWLINE] Anyhow, you discount confirmation bias,<mask><mask> sure are you with the validity of your self-diagnosis? It's<mask> a small part of known [cognitive biases]( [URL] ) -<mask> many ways for the mind to trick itself to retain it's values/beliefs! [NEWLINE] [NEWLINE] You love science and thinking critically,<mask> no doubt you have a hunch that there is a contradiction you need to resolve. [NEWLINE] [NEWLINE] Science doesn't do<mask> well proving the non-existence of something,<mask> that something doesn't exist or is untestable  -<mask> there is no proof to find! There's nothing for science to point to and say "Oh, there it isn't!"<mask><mask> you make a very specific claim,<mask> your words have set meanings and definitions, then Science or just well applied logic can find proof of absence (or proof of impossibility) much easier. For example, for Science, you have to make a specific claim like
Label encoding: <s>The brain is a pattern maker and pattern matcher. And the patterns it makes, it finds in the outside world with ease. And the more deeply held and treasured a pattern (belief), the more it must find them out there in the world, so that they can be matched are validated to be true. And the more invested it becomes, the more it sees what it believes, even if "out there" is the very poorest of matches. [NEWLINE] [NEWLINE] To challenge your own beliefs is to threaten your very identity! For it's our patterns/beliefs/values that make us who we are, and to change them for something else is an act of self-destruction without the guarantee of catharsis at the other end! So kudos to you! [NEWLINE] [NEWLINE] My Dad was a mathematician by training and a hobby astrologer/astronomer, and he made thousands of charts. He sort a statistical proof of astrology for about a decade, collecting thousands of birth dates/times/locations for people in different professions. I remember there was a famous sportsmen set, and I remember he had a huge amount of data on volcanic eruptions and earthquakes. He eventually discounted the popular field as completely rubbish, and was left with two statistical correlations (mars in some house for sportsmen and for earthquakes!) That was 20 years ago, and what remains from my perspective, and conversations with him, is his spiritual sense that "existential causation" is "top down", not "bottom-up"... [NEWLINE] [NEWLINE] Ahh, Dad. [NEWLINE] [NEWLINE] Anyhow, you discount confirmation bias, but how sure are you with the validity of your self-diagnosis? It's but a small part of known [cognitive biases]( [URL] ) - so many ways for the mind to trick itself to retain it's values/beliefs! [NEWLINE] [NEWLINE] You love science and thinking critically, so no doubt you have a hunch that there is a contradiction you need to resolve. [NEWLINE] [NEWLINE] Science doesn't do so well proving the non-existence of something, if that something doesn't exist or is untestable  - because there is no proof to find! There's nothing for science to point to and say "Oh, there it isn't!" But if you make a very specific claim, where your words have set meanings and definitions, then Science or just well applied logic can find proof of absence (or proof of impossibility) much easier. For example, for Science, you have to make a specific claim like
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> the mother's life is not in immediate danger the baby should not be aborted? [NEWLINE] [NEWLINE] Ok<mask> to the thousands of women who get raped every year and become pregnant, Tough? You are now forced against your will to take care of a baby that for the rest of your life every single day will be a reminder of<mask> happened to you one day/night? I mean the mother's life is not at risk and neither is the baby's<mask> both will be healthy for the birth<mask> the mother did not even have any fault here, she did not have a choice and now you are FORCING her to have a baby and take care of it. [NEWLINE] [NEWLINE] The counter argument is that she can put it up for adoption, she still has to go through birth... something which I am assuming you are a male by looking at your post history and you have no idea<mask> painful it is (don't worry, I am a guy too)<mask> they still have to go through that process<mask> you,<mask> someone who never even has to go through this process at all thinks it is wrong. [NEWLINE] [NEWLINE] I don't know<mask> profession you are, let's say you are a construction worker for the sake of example. You work hard and you do good work and all is well in your life. Suddenly there is an energy crisis going on in the city. There is an outcry and city officials need to make sure there is enough energy for emergency things like hospitals and police dept ect. ect.<mask><mask> do they do? they put an energy limit on all construction worksites. You now can't use cranes, jacks, lifts, nail guns electric saws. [NEWLINE] [NEWLINE] No sorry, now everything for you is done manually. You know<mask>?<mask> people who don't even do<mask> you do are telling you<mask> you can and can't do your job. Seems like bullshit right? Same exact concept applies to abortion. [NEWLINE] [NEWLINE] Now another reason<mask> abortion should not be illegal is accidental pregnancies. There are<mask> many teenagers and kids who get pregnant<mask> of stupid mistakes and their lives are not considerably shaken and altered<mask> they can get an abortion. It does not even have to be young people even older people in their 20s or maybe in their 30s who are not ready for kids get abortions all the time<mask> their life simply can not support taking care of a new child. [NEWLINE] [NEWLINE] An adoption argument is not going to be realistic,<mask> we replace every single abortion in the US per year [This shows<mask> many
Label encoding: <s>So if the mother's life is not in immediate danger the baby should not be aborted? [NEWLINE] [NEWLINE] Ok so to the thousands of women who get raped every year and become pregnant, Tough? You are now forced against your will to take care of a baby that for the rest of your life every single day will be a reminder of what happened to you one day/night? I mean the mother's life is not at risk and neither is the baby's because both will be healthy for the birth but the mother did not even have any fault here, she did not have a choice and now you are FORCING her to have a baby and take care of it. [NEWLINE] [NEWLINE] The counter argument is that she can put it up for adoption, she still has to go through birth... something which I am assuming you are a male by looking at your post history and you have no idea how painful it is (don't worry, I am a guy too) but they still have to go through that process because you, as someone who never even has to go through this process at all thinks it is wrong. [NEWLINE] [NEWLINE] I don't know what profession you are, let's say you are a construction worker for the sake of example. You work hard and you do good work and all is well in your life. Suddenly there is an energy crisis going on in the city. There is an outcry and city officials need to make sure there is enough energy for emergency things like hospitals and police dept ect. ect. so what do they do? they put an energy limit on all construction worksites. You now can't use cranes, jacks, lifts, nail guns electric saws. [NEWLINE] [NEWLINE] No sorry, now everything for you is done manually. You know why? because people who don't even do what you do are telling you how you can and can't do your job. Seems like bullshit right? Same exact concept applies to abortion. [NEWLINE] [NEWLINE] Now another reason why abortion should not be illegal is accidental pregnancies. There are so many teenagers and kids who get pregnant because of stupid mistakes and their lives are not considerably shaken and altered because they can get an abortion. It does not even have to be young people even older people in their 20s or maybe in their 30s who are not ready for kids get abortions all the time because their life simply can not support taking care of a new child. [NEWLINE] [NEWLINE] An adoption argument is not going to be realistic, if we replace every single abortion in the US per year [This shows how many
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You don't seem to understand my argument. One of my premises is that laws do not,<mask><mask>, force you to choose. Laws provide an incentive for you to choose one type of behavior over another,<mask><mask> I said, they do not govern your actual ability to make a choice. You can choose to perform any action you can otherwise perform. You say that I am "directly coerced into not making that choice", which is simply untrue.<mask> we're talking dictionary definitions, then [coercion]( [URL].com/dictionary/coerce) is to compel someone to do something by force or threat...<mask> a law can fail to coerce me. The law can *try* to coerce me,<mask> it cannot *make* me do anything. It can only,<mask> I wrote in the above post, administer consequences for my actions. You are misusing the word ‘coercion’. In one of your above posts, you wrote “<mask> it was illegal to drive on St Patrick's day, I would be coerced into not driving through no choice of my own”. Well, no, that’s false.  You would still be making a choice to obey or disobey the law. Coercion depends on the subject’s willingness to be coerced—this is<mask> drunk driving exists today (or, by extension,<mask> anyone breaks any law). Drunk people aren’t making a rational choice to drive drunk,<mask> the decision-making processes that govern rational behavior have been disabled! The coercion that you seem to think is<mask> powerful does nothing to prevent them from drunken driving. [NEWLINE] [NEWLINE] You wrote "<mask> driving was illegal, would you say that you are equally<mask> free to make the choice to drive<mask> not?" and the answer is,<mask> you have a car, yes! You're free, by your definition of freedom--and I'm going by this explicitly stated line in your above post, "It is not<mask> YOU CHOOSE<mask> YOUR ABILITY to choose which defines freedom"--to choose to drive in that scenario, and get pulled over for it. Your entire premise about risk is that knowing the potential consequences of an action doesn’t prevent you from taking said action. I don’t see<mask> you can reasonably hold the belief that the law is different. [NEWLINE] [NEWLINE] /u/DerekReinbold and /u/gpunotpsu seem to be handling the issue of the definition of freedom quite well
Label encoding: <s>You don't seem to understand my argument. One of my premises is that laws do not, in fact, force you to choose. Laws provide an incentive for you to choose one type of behavior over another, but as I said, they do not govern your actual ability to make a choice. You can choose to perform any action you can otherwise perform. You say that I am "directly coerced into not making that choice", which is simply untrue. If we're talking dictionary definitions, then [coercion]( [URL].com/dictionary/coerce) is to compel someone to do something by force or threat... but a law can fail to coerce me. The law can *try* to coerce me, but it cannot *make* me do anything. It can only, as I wrote in the above post, administer consequences for my actions. You are misusing the word ‘coercion’. In one of your above posts, you wrote “ If it was illegal to drive on St Patrick's day, I would be coerced into not driving through no choice of my own”. Well, no, that’s false.  You would still be making a choice to obey or disobey the law. Coercion depends on the subject’s willingness to be coerced—this is why drunk driving exists today (or, by extension, why anyone breaks any law). Drunk people aren’t making a rational choice to drive drunk, because the decision-making processes that govern rational behavior have been disabled! The coercion that you seem to think is so powerful does nothing to prevent them from drunken driving. [NEWLINE] [NEWLINE] You wrote " If driving was illegal, would you say that you are equally as free to make the choice to drive as not?" and the answer is, if you have a car, yes! You're free, by your definition of freedom--and I'm going by this explicitly stated line in your above post, "It is not WHAT YOU CHOOSE but YOUR ABILITY to choose which defines freedom"--to choose to drive in that scenario, and get pulled over for it. Your entire premise about risk is that knowing the potential consequences of an action doesn’t prevent you from taking said action. I don’t see how you can reasonably hold the belief that the law is different. [NEWLINE] [NEWLINE] /u/DerekReinbold and /u/gpunotpsu seem to be handling the issue of the definition of freedom quite well
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The phrase was deliberately inserted into the pledge<mask> a way to differentiate between the United States and state atheism ( Merriman, Scott A. (2007). Religion and the Law in America: An Encyclopedia of Personal Belief) [ENDQ] [NEWLINE] That sounds like a perfectly reasonable thing to do.  It's<mask> not an endorsement of religion,<mask> a distinction between a secular and atheistic state.  A truly secular state would have to have no official position on religion to include the conspicuous absence of any evidence that religion existed. [NEWLINE] [NEWLINE] [STARTQ] This endorses religion over the lack thereof, which is discriminatory against atheists is nothing else. [ENDQ] [NEWLINE] No it doesn't.  I've already explained<mask> it doesn't.  Saying a word is not an endorsement...it just doesn't logically follow. [NEWLINE] [NEWLINE] [STARTQ] <mask>, I would<mask><mask> it goes further than that<mask> it doesn't pander to a specific denomination, it does<mask> reflect christian beliefs specifically. [ENDQ] [NEWLINE] Okay, feel free to do that.  You certainly haven't<mask>, and I'm not sure you read past the first paragraph of the article you linked to.  It says literally nothing else about the 9th Circuit ruling other than the exact words you quoted "a proud recitation of the ideals on which our Republic was founded."  You have to perform several leaps of logic to assume that<mask> that means is that the court was endorsing Christianity and you don't have anything approaching sufficient evidence to make that argument. [NEWLINE] [NEWLINE] Beyond that, the rest of the article *is about the ambiguity of the phrase*: [NEWLINE] [NEWLINE] [STARTQ] Does it affirm our faith in God or assert that we have his special protection? Is it a ceremonial deist formula with no especial religious character? Or is it merely a historical nod to the beliefs of the founders,<mask> the 9th Circuit majority said? You can take this wherever you like,<mask> "under God" is another hapax legomenon that doesn't occur anywhere else in modern English. People don't say things like "Western Europe isn't under God anymore," or "She only goes out with men who are under God." [ENDQ] [NEWLINE] [STARTQ] That ambiguity has certain advantages.<mask> it actually came about<mask> of a linguistic misunderstanding. The words were taken from the Gettysburg Address,<mask> Lincoln asked his listeners to resolve that "this nation, under God, shall have a new birth of freedom." Except that in the Gettysburg Address, "under God" didn't modify "this nation"<mask> the following phrase, "have a
Label encoding: <s> [STARTQ] The phrase was deliberately inserted into the pledge as a way to differentiate between the United States and state atheism ( Merriman, Scott A. (2007). Religion and the Law in America: An Encyclopedia of Personal Belief) [ENDQ] [NEWLINE] That sounds like a perfectly reasonable thing to do.  It's also not an endorsement of religion, but a distinction between a secular and atheistic state.  A truly secular state would have to have no official position on religion to include the conspicuous absence of any evidence that religion existed. [NEWLINE] [NEWLINE] [STARTQ] This endorses religion over the lack thereof, which is discriminatory against atheists is nothing else. [ENDQ] [NEWLINE] No it doesn't.  I've already explained how it doesn't.  Saying a word is not an endorsement...it just doesn't logically follow. [NEWLINE] [NEWLINE] [STARTQ] But, I would argue that it goes further than that although it doesn't pander to a specific denomination, it does indeed reflect christian beliefs specifically. [ENDQ] [NEWLINE] Okay, feel free to do that.  You certainly haven't yet, and I'm not sure you read past the first paragraph of the article you linked to.  It says literally nothing else about the 9th Circuit ruling other than the exact words you quoted "a proud recitation of the ideals on which our Republic was founded."  You have to perform several leaps of logic to assume that what that means is that the court was endorsing Christianity and you don't have anything approaching sufficient evidence to make that argument. [NEWLINE] [NEWLINE] Beyond that, the rest of the article *is about the ambiguity of the phrase*: [NEWLINE] [NEWLINE] [STARTQ] Does it affirm our faith in God or assert that we have his special protection? Is it a ceremonial deist formula with no especial religious character? Or is it merely a historical nod to the beliefs of the founders, as the 9th Circuit majority said? You can take this wherever you like, because "under God" is another hapax legomenon that doesn't occur anywhere else in modern English. People don't say things like "Western Europe isn't under God anymore," or "She only goes out with men who are under God." [ENDQ] [NEWLINE] [STARTQ] That ambiguity has certain advantages. But it actually came about because of a linguistic misunderstanding. The words were taken from the Gettysburg Address, where Lincoln asked his listeners to resolve that "this nation, under God, shall have a new birth of freedom." Except that in the Gettysburg Address, "under God" didn't modify "this nation" but the following phrase, "have a
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A few months back there was a tragic case<mask> a poor young girl accidentally shot a Superviser at a gunrange. The girl was nine year old and firing a Uzi submachine gun on full automatic. [NEWLINE] [NEWLINE] From some comments various people have left on this and the reporting it seems that it is not unusual for kids even younger then that to shoot real firearms. [NEWLINE] [NEWLINE] This is ridiculous. Children (and I define that in this case<mask> everyone under 18) are not nearly responsible enough to handle them. There is a reason we do not let little children drive cars or drink alcohol etc. [NEWLINE] [NEWLINE] Childrens Brains (and even young adult's,<mask> you have to draw the line somewhere) are still in the process of developing. We all have done stupid shit<mask> children. Hell, just here on reddit I read about some guy who shot his brother *twice*<mask> a kid. You may choose not be believe that story,<mask> for me it seems plausible. [NEWLINE] [NEWLINE] I know btw that the Instructor in this case made some critical mistakes<mask> well, and that children shooting firearms have to be supervised.<mask> still,<mask> the Girl had the experiance, wisdom and patience that comes with being an adult, this would have never happened. [NEWLINE] [NEWLINE] An adult acts<mask> another layer of protection to prevent an accident. In my view, having a child handle a firearm is a little like not exercising proper trigger discipline.<mask> you do everything else right, nothing bad can really happen.<mask> you should never rely on doing everything else right. [NEWLINE] [NEWLINE] You might know some very mature and patient 14 year old who is really interested in hunting or sport shooting. In that case, congratulations.<mask> I still stand by my view that that kid should wait and grow four more years. Sure, they might not get to shoot for four your years,<mask> nobody will die. [NEWLINE] [NEWLINE] EDIT: [NEWLINE] [NEWLINE] /u/caw81 has slightly changed my view. Now<mask><mask> only people under the age of 16 should be barred. [NEWLINE] [NEWLINE] EDIT 2: [NEWLINE] [NEWLINE] /u/incruente and /u/Zorthianator have completely changed my view. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it
Label encoding: <s>A few months back there was a tragic case where a poor young girl accidentally shot a Superviser at a gunrange. The girl was nine year old and firing a Uzi submachine gun on full automatic. [NEWLINE] [NEWLINE] From some comments various people have left on this and the reporting it seems that it is not unusual for kids even younger then that to shoot real firearms. [NEWLINE] [NEWLINE] This is ridiculous. Children (and I define that in this case as everyone under 18) are not nearly responsible enough to handle them. There is a reason we do not let little children drive cars or drink alcohol etc. [NEWLINE] [NEWLINE] Childrens Brains (and even young adult's, but you have to draw the line somewhere) are still in the process of developing. We all have done stupid shit as children. Hell, just here on reddit I read about some guy who shot his brother *twice* as a kid. You may choose not be believe that story, but for me it seems plausible. [NEWLINE] [NEWLINE] I know btw that the Instructor in this case made some critical mistakes as well, and that children shooting firearms have to be supervised. But still, if the Girl had the experiance, wisdom and patience that comes with being an adult, this would have never happened. [NEWLINE] [NEWLINE] An adult acts as another layer of protection to prevent an accident. In my view, having a child handle a firearm is a little like not exercising proper trigger discipline. If you do everything else right, nothing bad can really happen. But you should never rely on doing everything else right. [NEWLINE] [NEWLINE] You might know some very mature and patient 14 year old who is really interested in hunting or sport shooting. In that case, congratulations. But I still stand by my view that that kid should wait and grow four more years. Sure, they might not get to shoot for four your years, but nobody will die. [NEWLINE] [NEWLINE] EDIT: [NEWLINE] [NEWLINE] /u/caw81 has slightly changed my view. Now I think only people under the age of 16 should be barred. [NEWLINE] [NEWLINE] EDIT 2: [NEWLINE] [NEWLINE] /u/incruente and /u/Zorthianator have completely changed my view. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, well, lets start with the abuses, just to prove that it is a problem that needs to be solved.  Without going into it comprehensively,<mask> it is very difficult to get complete numbers, here is a reasonable article about the number of people who are wrongfully killed by police in a year: [URL] / [NEWLINE] [NEWLINE] <mask>,<mask> there isn't great evidence, unless you have better evidence,<mask><mask> it is safe to assume that this is accurate info.  And that about 15% of deaths by police are abuses of some kind. <mask> you ask me, that is a significant amount of abuse that needs to be addressed, and it probably results in a significant cost to society.  Actually, economists try very hard to estimate the value of human life, and the Office of Management and Budget in the USA values the average human life at between $7 and $9 Million dollars ( [URL] / ). Im not going to do the calculation here,<mask> that is just the cost by loss of life. [NEWLINE] [NEWLINE] Now, other than deaths (which is more unjust,<mask> probably less costly),<mask><mask> the biggest police abuse in the USA is the over-policing of poor and minority areas, which results in thousands of excess arrests, which costs billions of dollars a year in direct costs (the cost of the judicial system, and the costs of incarcerating people), and additional billions in indirect costs (the costs of lost productivity<mask><mask><mask> of having to go through the judicial process, and the prison system). [URL] [NEWLINE] [NEWLINE] "In 2007, around $74 billion was spent on corrections.[136] The total number of inmates in 2007 in federal, state, and local lockups was 2,419,241.[22] That comes to around $30,600 per inmate. Church Publishing (CP) estimates the 50 states plus federal government expenditure amounts to $56.9 billion spent on U.S. corrections. The CPs' additional $7.5 billion estimate excludes 'double-counting' state or federal subsidies for local lock-ups which vary to reach $64.4 billion spent on U.S. corrections annually by 2014.[137]" [URL] [NEWLINE] [NEWLINE] And then, let's just talk about body cams, and assume that they reduce the cost of payouts<mask><mask><mask> of lawsuits against police (we are not even deciding whether they are justified, just that they happen). and you are talking billions of dollars in costs a year. [URL] [NEWLINE] [NEWLINE] <mask>, i haven't added up the costs here
Label encoding: <s>Ok, well, lets start with the abuses, just to prove that it is a problem that needs to be solved.  Without going into it comprehensively, although it is very difficult to get complete numbers, here is a reasonable article about the number of people who are wrongfully killed by police in a year: [URL] / [NEWLINE] [NEWLINE] So, because there isn't great evidence, unless you have better evidence, I think it is safe to assume that this is accurate info.  And that about 15% of deaths by police are abuses of some kind.  If you ask me, that is a significant amount of abuse that needs to be addressed, and it probably results in a significant cost to society.  Actually, economists try very hard to estimate the value of human life, and the Office of Management and Budget in the USA values the average human life at between $7 and $9 Million dollars ( [URL] / ). Im not going to do the calculation here, but that is just the cost by loss of life. [NEWLINE] [NEWLINE] Now, other than deaths (which is more unjust, but probably less costly), I think the biggest police abuse in the USA is the over-policing of poor and minority areas, which results in thousands of excess arrests, which costs billions of dollars a year in direct costs (the cost of the judicial system, and the costs of incarcerating people), and additional billions in indirect costs (the costs of lost productivity as a result of having to go through the judicial process, and the prison system). [URL] [NEWLINE] [NEWLINE] "In 2007, around $74 billion was spent on corrections.[136] The total number of inmates in 2007 in federal, state, and local lockups was 2,419,241.[22] That comes to around $30,600 per inmate. Church Publishing (CP) estimates the 50 states plus federal government expenditure amounts to $56.9 billion spent on U.S. corrections. The CPs' additional $7.5 billion estimate excludes 'double-counting' state or federal subsidies for local lock-ups which vary to reach $64.4 billion spent on U.S. corrections annually by 2014.[137]" [URL] [NEWLINE] [NEWLINE] And then, let's just talk about body cams, and assume that they reduce the cost of payouts as a result of lawsuits against police (we are not even deciding whether they are justified, just that they happen). and you are talking billions of dollars in costs a year. [URL] [NEWLINE] [NEWLINE] SO, i haven't added up the costs here
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the purpose of a test is to approximate your ability to understand the material that test is over.<mask><mask> it is designed properly, the grade a child receives on a test should approximately reflect their understanding of the material, and the cumulative score of their tests is<mask> produces the final grade.<mask> that is the case, then<mask> does it matter<mask> they figured out, say, trig identities<mask> the test was first given rather than a week later? To give them a grade that is not reflective of<mask> they actually understand at the end of the course is to defeat the purpose of the evaluation itself. [NEWLINE] [NEWLINE] I'm using "test" broadly.<mask><mask> things that are typically called "projects" are, generally speaking, a type of test. I don't think that repetitious "busywork" should ever be factored in to a grade.<mask>, in case it matters, I'm speaking from the experience of someone who went through the American public education system. [NEWLINE] [NEWLINE] I understand that grades<mask> can be used to measure certain personality factors (conscientiousness, ability to follow instruction, adherence to procedural norms) which all (debatable) might be<mask> important<mask> understanding the material.<mask>, you could simply give a separate evaluation of these factors instead rolling them all in to one. It's been observed that performance at the university level corresponds better to standardized test scores than to high school GPA,<mask> it is helpful to have a grade that isn't a vague amalgamation of different factors. [NEWLINE] [NEWLINE] Note, I set specific parameters in the title for a reason. I understand that a teacher might not have time to give endless make-up tests,<mask>, surely, they can manage a handful. I excluded college<mask><mask><mask> that raises many questions about different obligations and the manner in which people separate based on ability.<mask><mask><mask> it comes to k-12 (especially at public schools) there is more of a strict obligation to be fair and accurate. [NEWLINE] [NEWLINE] P.S. - This my first CMV, or Reddit post for that matter,<mask> I apologize in advance<mask> I've made any missteps in my presentation. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of
Label encoding: <s>I think the purpose of a test is to approximate your ability to understand the material that test is over. Given that it is designed properly, the grade a child receives on a test should approximately reflect their understanding of the material, and the cumulative score of their tests is what produces the final grade. If that is the case, then why does it matter if they figured out, say, trig identities when the test was first given rather than a week later? To give them a grade that is not reflective of what they actually understand at the end of the course is to defeat the purpose of the evaluation itself. [NEWLINE] [NEWLINE] I'm using "test" broadly. I think things that are typically called "projects" are, generally speaking, a type of test. I don't think that repetitious "busywork" should ever be factored in to a grade. Also, in case it matters, I'm speaking from the experience of someone who went through the American public education system. [NEWLINE] [NEWLINE] I understand that grades also can be used to measure certain personality factors (conscientiousness, ability to follow instruction, adherence to procedural norms) which all (debatable) might be as important as understanding the material. However, you could simply give a separate evaluation of these factors instead rolling them all in to one. It's been observed that performance at the university level corresponds better to standardized test scores than to high school GPA, so it is helpful to have a grade that isn't a vague amalgamation of different factors. [NEWLINE] [NEWLINE] Note, I set specific parameters in the title for a reason. I understand that a teacher might not have time to give endless make-up tests, but, surely, they can manage a handful. I excluded college because I think that raises many questions about different obligations and the manner in which people separate based on ability. I think when it comes to k-12 (especially at public schools) there is more of a strict obligation to be fair and accurate. [NEWLINE] [NEWLINE] P.S. - This my first CMV, or Reddit post for that matter, so I apologize in advance if I've made any missteps in my presentation. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I speak English and Spanish. Both fluently.<mask> I speak English I don't have a Spanish accent,<mask> would I intentionally fake an English accent<mask> saying Spanish words. [NEWLINE] [NEWLINE] The reason non-Spanish speaking English speakers say quesadilla the way they do is<mask> they can't easily pronounce it correctly.<mask> would native speakers mispronounce words? [NEWLINE] [NEWLINE] Think about your iPod experience. Did you pronounce it "with an English accent" on purpose to be "exclusionary"? No, you just spoke the way you speak<mask> coming to a word you know<mask> to pronounce. [NEWLINE] [NEWLINE] For me to pronounce quesadilla they way you suggest requires me to fake something, to use an accent I don't have - this I would consider to be more pretentious then just saying it the way you know to say it. [NEWLINE] [NEWLINE] Your position is that I should pronounce words from my native language incorrectly -<mask> just<mask> incorrectly should I pronounce them? You say you don't want people to use hard L's.<mask> not, I've heard non-spanish speakers say it that way.<mask> they used your logic and heard you pronounce it with the Y sound, they would call you pretentious and exclusionary, and tell you to use three proper English accent. [NEWLINE] [NEWLINE] Now look<mask> someone doesn't understand me, it'll take me a second to realise it and say then fake it to sound a way they might understand - I have no problem with that. I'm not going to default to wrong pronunciations<mask>. [NEWLINE] [NEWLINE] <mask> there are cases<mask> I do default to the English version of the word - a consequence of being born in the US and only having received education in English. For example Colorado Montana and Florida are all Spanish words,<mask><mask> referring to the States I always default to the English version - often even<mask> speaking in Spanish, which feels strange each time<mask> I become aware of<mask> in saying it<mask> it sounds out of place to me. I natively default to both versions in different contexts. It's all confusing enough already without faking an accent preemptively, and guessing<mask> a given individual might say the word themselves. [NEWLINE] [NEWLINE] There is another aspect to this that makes<mask> you ask difficult. I've grown up with a  predominantly bilingual family -<mask> that happens you get a lot of Spanglish. It happens for the same reason English and Spanish pick up loan words from each other -<mask> that group realizes that the other group has a good word for something.<mask> two people
Label encoding: <s>I speak English and Spanish. Both fluently. When I speak English I don't have a Spanish accent, why would I intentionally fake an English accent when saying Spanish words. [NEWLINE] [NEWLINE] The reason non-Spanish speaking English speakers say quesadilla the way they do is because they can't easily pronounce it correctly. Why would native speakers mispronounce words? [NEWLINE] [NEWLINE] Think about your iPod experience. Did you pronounce it "with an English accent" on purpose to be "exclusionary"? No, you just spoke the way you speak when coming to a word you know how to pronounce. [NEWLINE] [NEWLINE] For me to pronounce quesadilla they way you suggest requires me to fake something, to use an accent I don't have - this I would consider to be more pretentious then just saying it the way you know to say it. [NEWLINE] [NEWLINE] Your position is that I should pronounce words from my native language incorrectly - so just how incorrectly should I pronounce them? You say you don't want people to use hard L's. Why not, I've heard non-spanish speakers say it that way. If they used your logic and heard you pronounce it with the Y sound, they would call you pretentious and exclusionary, and tell you to use three proper English accent. [NEWLINE] [NEWLINE] Now look if someone doesn't understand me, it'll take me a second to realise it and say then fake it to sound a way they might understand - I have no problem with that. I'm not going to default to wrong pronunciations though. [NEWLINE] [NEWLINE] Also there are cases where I do default to the English version of the word - a consequence of being born in the US and only having received education in English. For example Colorado Montana and Florida are all Spanish words, but when referring to the States I always default to the English version - often even when speaking in Spanish, which feels strange each time because I become aware of how in saying it since it sounds out of place to me. I natively default to both versions in different contexts. It's all confusing enough already without faking an accent preemptively, and guessing how a given individual might say the word themselves. [NEWLINE] [NEWLINE] There is another aspect to this that makes what you ask difficult. I've grown up with a  predominantly bilingual family - when that happens you get a lot of Spanglish. It happens for the same reason English and Spanish pick up loan words from each other - because that group realizes that the other group has a good word for something. When two people
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How do you propose to separate high quality posts that have controversial opinions from the garbage posts that present that same idea?<mask> looking at huge communities (like /r/movies) it is nearly impossible to prevent it from devolving into a pandering to the lowest common denominator, that's exactly<mask> happens in a purely democratic system. The general public is stupid, really, really stupid. This is<mask> pure democracies never have, and never will work in very large groups. Leaders(voters) need to be educated in the matters they have power over, and the general public is not educated in these matters<mask> they will make poor decisions<mask> left to their own devices. [NEWLINE] [NEWLINE] The solution that I believe is best for this is the reputation system that is implemented on many other public forums. People get voting power based on the quality of their own posts. Of course this doesn't really matter in subreddits that deal with trivial jokes and shitposting (see /r/funny, pics, videos, advice animals, askreddit to some degree too)<mask> in that case you shouldn't really expect to see any quality posts in these places<mask> with their group sizes you will only see the exact phenomenon that you posted originally about. [NEWLINE] [NEWLINE] The reputation system would be helpful in giving power to people who consistently make high quality posts,<mask>, from<mask> I've observed in many well moderated subreddits, this hive-mind mentality isn't an issue. Look at /r/askscience, they have a fantastic community, that is absolutely huge,<mask> still manages to make the democratic voting system work<mask> of their strict quality standards and posting guidelines. Posting a controversial opinion there doesn't matter,<mask> does matter is supporting arguments and evidence. Even the most hive-mindy of posts there get shot down immediately<mask> they don't have substantial evidence to support their claims. [NEWLINE] [NEWLINE] Another very niche and small subreddit I'd like to point out is /r/SSBM. It is a subreddit dedicated to educational discussion of a popular competitive fighting game. This subreddit<mask> consistently shoots down hive-mindy posts<mask> the OP isn't able to present some reasoning behind their opinion. This subreddit<mask> benefits from being relatively small<mask> with allows this system to succeed without super strict posting rules. [NEWLINE] [NEWLINE] I personally have observed many posts in popular subreddits get upvoted<mask> having controversial opinions<mask> they present their case well. I have a feeling from your post history that you tend to make very low-effort posts, which is likely<mask> they would get
Label encoding: <s>How do you propose to separate high quality posts that have controversial opinions from the garbage posts that present that same idea? When looking at huge communities (like /r/movies) it is nearly impossible to prevent it from devolving into a pandering to the lowest common denominator, that's exactly what happens in a purely democratic system. The general public is stupid, really, really stupid. This is why pure democracies never have, and never will work in very large groups. Leaders(voters) need to be educated in the matters they have power over, and the general public is not educated in these matters so they will make poor decisions if left to their own devices. [NEWLINE] [NEWLINE] The solution that I believe is best for this is the reputation system that is implemented on many other public forums. People get voting power based on the quality of their own posts. Of course this doesn't really matter in subreddits that deal with trivial jokes and shitposting (see /r/funny, pics, videos, advice animals, askreddit to some degree too) so in that case you shouldn't really expect to see any quality posts in these places as with their group sizes you will only see the exact phenomenon that you posted originally about. [NEWLINE] [NEWLINE] The reputation system would be helpful in giving power to people who consistently make high quality posts, however, from what I've observed in many well moderated subreddits, this hive-mind mentality isn't an issue. Look at /r/askscience, they have a fantastic community, that is absolutely huge, but still manages to make the democratic voting system work because of their strict quality standards and posting guidelines. Posting a controversial opinion there doesn't matter, what does matter is supporting arguments and evidence. Even the most hive-mindy of posts there get shot down immediately if they don't have substantial evidence to support their claims. [NEWLINE] [NEWLINE] Another very niche and small subreddit I'd like to point out is /r/SSBM. It is a subreddit dedicated to educational discussion of a popular competitive fighting game. This subreddit also consistently shoots down hive-mindy posts if the OP isn't able to present some reasoning behind their opinion. This subreddit however benefits from being relatively small though with allows this system to succeed without super strict posting rules. [NEWLINE] [NEWLINE] I personally have observed many posts in popular subreddits get upvoted despite having controversial opinions because they present their case well. I have a feeling from your post history that you tend to make very low-effort posts, which is likely why they would get
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If<mask> I mean is "irrelevant" then it doesn't put the demand on me to be clear about<mask> I mean. It's irrelevant<mask> I mean.<mask> I say "Wow, that got bigger." and someone thinks I said the n-word, by your definition - [NEWLINE] [NEWLINE] [STARTQ] <mask> you mean it is irrelevant. [ENDQ] <mask> it is heard is literally all that matters. [NEWLINE] [NEWLINE] then I said the n-word. It just happened. It doesn't matter<mask> sounds I used or<mask> I meant by them or anything else -<mask> the person I'm talking to heard something and took offense, then it's my fault.<mask> I say "Hey dawg!" and their dog just died and I didn't know and they break out into tears and pull out a gun and shoot themselves, I've just murdered them<mask> my word "caused" them to take that action no matter<mask> my intent was. [NEWLINE] [NEWLINE] That's absurd. [NEWLINE] [NEWLINE] Intent matters. It's not irrelevant. It's a huge portion of communication<mask> it's all I have control over. It's my responsibility to do<mask> I can to make sure that the outcome that results from my actions is the one I wanted to happen and<mask> I do something that results in an unexpected outcome, I need to learn from that, adapt, and often times clarify.<mask> it's not just the end of the road, slap the cuffs on me, and throw away the key<mask> someone took offense to the word "Fork" and I had no way of knowing that it would be a trigger for them. [NEWLINE] [NEWLINE] And no, it's not "...between friends who are both black."<mask> by that definition, some black people are using it incorrectly. You've just defined the word such that of the demographic of people we've agreed are "allowed" to use the word (<mask> has been the case throughout this post, "allowed" doesn't mean having the freedom to do<mask>,<mask> having an utter lack of controversy<mask> doing<mask> ), a portion of them are using the word incorrect. [NEWLINE] [NEWLINE] And on top of that, it presumes the conclusion - that it's only a word that is allowed to be said by black people and is offensive or racist at all other times<mask> historical context etc. You're begging the question - not proving it. This thread is about<mask> white people aren't "allowed" to use the n-word in the same context<mask> black people<mask>,<mask> they use it, it's not offensive. You
Label encoding: <s>If what I mean is "irrelevant" then it doesn't put the demand on me to be clear about what I mean. It's irrelevant what I mean. If I say "Wow, that got bigger." and someone thinks I said the n-word, by your definition - [NEWLINE] [NEWLINE] [STARTQ] How you mean it is irrelevant. [ENDQ] How it is heard is literally all that matters. [NEWLINE] [NEWLINE] then I said the n-word. It just happened. It doesn't matter what sounds I used or what I meant by them or anything else - if the person I'm talking to heard something and took offense, then it's my fault. If I say "Hey dawg!" and their dog just died and I didn't know and they break out into tears and pull out a gun and shoot themselves, I've just murdered them because my word "caused" them to take that action no matter what my intent was. [NEWLINE] [NEWLINE] That's absurd. [NEWLINE] [NEWLINE] Intent matters. It's not irrelevant. It's a huge portion of communication because it's all I have control over. It's my responsibility to do what I can to make sure that the outcome that results from my actions is the one I wanted to happen and if I do something that results in an unexpected outcome, I need to learn from that, adapt, and often times clarify. But it's not just the end of the road, slap the cuffs on me, and throw away the key because someone took offense to the word "Fork" and I had no way of knowing that it would be a trigger for them. [NEWLINE] [NEWLINE] And no, it's not "...between friends who are both black." because by that definition, some black people are using it incorrectly. You've just defined the word such that of the demographic of people we've agreed are "allowed" to use the word ( as has been the case throughout this post, "allowed" doesn't mean having the freedom to do so, but having an utter lack of controversy when doing so ), a portion of them are using the word incorrect. [NEWLINE] [NEWLINE] And on top of that, it presumes the conclusion - that it's only a word that is allowed to be said by black people and is offensive or racist at all other times besides historical context etc. You're begging the question - not proving it. This thread is about why white people aren't "allowed" to use the n-word in the same context as black people if, when they use it, it's not offensive. You
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>(1a) First, you seem to believe that<mask> in the past something was good for the propagation of your genes, it is<mask> important to do that thing today.  For example, hunting mammoths was once very important for your genes. Gathering roots was once very important for your genes. Staying within earshot of other members of your clan and threatening to kill people outside the clan was important for your genes. Have you done any of these things lately? No? [NEWLINE] [NEWLINE] Okay. Maybe you don't think behaviors that lead to the propagation of your genes in the past are that important. [NEWLINE] [NEWLINE] (1b) You<mask> seem to imply that it works the other way around -<mask> something wasn't important to the propagation of genes, it can't be very important, or not more important than something that was.<mask><mask> about going to a great job interview? Following traffic laws? Undergoing fertility treatments? Inventing penicillin? On your account, these things can't be important to human beings,<mask> they weren't important to the propagation of genes in the distant past. [NEWLINE] [NEWLINE] Note: (1a) and (1b) both address the fallacy that the status of some activity<mask> humans were evolving automatically has significance for its modern status. [NEWLINE] [NEWLINE] (2) Even<mask> you assume that important activities are activities that have an evolutionary significance to your species, that doesn't imply that having babies is the only role with evolutionary significance.<mask> you look at most species of social animals, only a tiny number of individuals are actually sexually active - the remainder are contributing to the community and the propagation of its genes in some other way. Humans have menopause precisely<mask> elders have functions that are far more important than pumping out more babies. It's perfectly possible to contribute far more to the propagation of your genes through non-reproductive efforts than through bumping uglies. [NEWLINE] [NEWLINE] Take Ian Fleming, for example - by discovering penicillin, he may well have doubled the population of the world. From a genetic point of view, he has done far more to propagate his genes than anyone else in recorded history. [NEWLINE] [NEWLINE] (3) Perhaps you want to go in for the far stronger assumption that it is not your *genes* that you want to propagate,<mask> *attributes that are specific to you and rare in the human population generally*.<mask> even<mask> you accept this assumption (which, like -2-, is dubious),<mask> would you want to pass on traits that are most-prom
Label encoding: <s>(1a) First, you seem to believe that if in the past something was good for the propagation of your genes, it is therefore important to do that thing today.  For example, hunting mammoths was once very important for your genes. Gathering roots was once very important for your genes. Staying within earshot of other members of your clan and threatening to kill people outside the clan was important for your genes. Have you done any of these things lately? No? [NEWLINE] [NEWLINE] Okay. Maybe you don't think behaviors that lead to the propagation of your genes in the past are that important. [NEWLINE] [NEWLINE] (1b) You also seem to imply that it works the other way around - if something wasn't important to the propagation of genes, it can't be very important, or not more important than something that was. So what about going to a great job interview? Following traffic laws? Undergoing fertility treatments? Inventing penicillin? On your account, these things can't be important to human beings, because they weren't important to the propagation of genes in the distant past. [NEWLINE] [NEWLINE] Note: (1a) and (1b) both address the fallacy that the status of some activity while humans were evolving automatically has significance for its modern status. [NEWLINE] [NEWLINE] (2) Even if you assume that important activities are activities that have an evolutionary significance to your species, that doesn't imply that having babies is the only role with evolutionary significance. If you look at most species of social animals, only a tiny number of individuals are actually sexually active - the remainder are contributing to the community and the propagation of its genes in some other way. Humans have menopause precisely because elders have functions that are far more important than pumping out more babies. It's perfectly possible to contribute far more to the propagation of your genes through non-reproductive efforts than through bumping uglies. [NEWLINE] [NEWLINE] Take Ian Fleming, for example - by discovering penicillin, he may well have doubled the population of the world. From a genetic point of view, he has done far more to propagate his genes than anyone else in recorded history. [NEWLINE] [NEWLINE] (3) Perhaps you want to go in for the far stronger assumption that it is not your *genes* that you want to propagate, but *attributes that are specific to you and rare in the human population generally*. But even if you accept this assumption (which, like -2-, is dubious), why would you want to pass on traits that are most-prom
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't disagree with everything you've said,<mask> I do disagree with your conclusion. [NEWLINE] [NEWLINE] [STARTQ] We don't waste our time protesting something political, like the liberal arts majors at my school (UC Berkeley) do. We are smart to realize that yelling loudly and occupying buildings is illogical: the opportunity cost is huge, and the tie would be better spent on doing well in school, gaining marketable skills, and pursuing activities that won't land us in jail. [ENDQ] [NEWLINE] No? I find it rather hard to believe that no STEM major is politically active. Non-STEM students may have a greater propensity to engage in such activity than their counterparts,<mask> it is certainly not unique to them. I<mask> find it slightly amusing that you say this in<mask> amounts to a sociological essay about something that that I very much doubt affects your own life in any material way. [NEWLINE] [NEWLINE] [STARTQ] We are more intelligent about social issues, economics, and politics than many humanities majors [ENDQ] [NEWLINE] Not in my experience. [NEWLINE] [NEWLINE] [STARTQ] And the funny thing is, b/c we're more intelligent, we STEM majors have a more logical and nuanced perspective of politics than many liberal arts majors. [ENDQ] [NEWLINE] Maybe.<mask> it is usually just<mask> wrong<mask> anyone elses. The major difference (pun intended) I've seen is that liberal-arts or humanities majors are less likely to exhibit a pretense of knowledge - they admit and understand that *they do not understand*. Most engineers and information technologies personel I know have no such self awareness, and claim definitive knowledge on subjects they have *at best* a cursory understanding of. [NEWLINE] [NEWLINE] [STARTQ] The only liberal arts majors I respect are philosophy and economics. Economics is very rigorous on a mathematical level, and many philosophers were<mask> mathematicians. [ENDQ] [NEWLINE] I'm glad my field has been able to live up to your standards. [NEWLINE] [NEWLINE] [STARTQ] <mask> engineering majors tend to<mask> kick ass on various graduate school admissions tests, like the GRE, GMAT, and LSAT. [ENDQ] [NEWLINE] That is true,<mask> concluding that STEM majors are better at law than people who studied law in undergrad is not an accurate assesment. I suspect the reason this occurs is that only STEM majors who feel especially confident in their ability to pass the test take it, whereas a far greater number of non-STEM majors (who are obviously underqualified) take it even<mask> they feel unconfident. The grades of the STEM students are, in a sense, inflated. [NEWLINE] [NEWLINE] [STARTQ] And people who are competent in math (whether
Label encoding: <s>I don't disagree with everything you've said, but I do disagree with your conclusion. [NEWLINE] [NEWLINE] [STARTQ] We don't waste our time protesting something political, like the liberal arts majors at my school (UC Berkeley) do. We are smart to realize that yelling loudly and occupying buildings is illogical: the opportunity cost is huge, and the tie would be better spent on doing well in school, gaining marketable skills, and pursuing activities that won't land us in jail. [ENDQ] [NEWLINE] No? I find it rather hard to believe that no STEM major is politically active. Non-STEM students may have a greater propensity to engage in such activity than their counterparts, but it is certainly not unique to them. I also find it slightly amusing that you say this in what amounts to a sociological essay about something that that I very much doubt affects your own life in any material way. [NEWLINE] [NEWLINE] [STARTQ] We are more intelligent about social issues, economics, and politics than many humanities majors [ENDQ] [NEWLINE] Not in my experience. [NEWLINE] [NEWLINE] [STARTQ] And the funny thing is, b/c we're more intelligent, we STEM majors have a more logical and nuanced perspective of politics than many liberal arts majors. [ENDQ] [NEWLINE] Maybe. But it is usually just as wrong as anyone elses. The major difference (pun intended) I've seen is that liberal-arts or humanities majors are less likely to exhibit a pretense of knowledge - they admit and understand that *they do not understand*. Most engineers and information technologies personel I know have no such self awareness, and claim definitive knowledge on subjects they have *at best* a cursory understanding of. [NEWLINE] [NEWLINE] [STARTQ] The only liberal arts majors I respect are philosophy and economics. Economics is very rigorous on a mathematical level, and many philosophers were also mathematicians. [ENDQ] [NEWLINE] I'm glad my field has been able to live up to your standards. [NEWLINE] [NEWLINE] [STARTQ] Also engineering majors tend to also kick ass on various graduate school admissions tests, like the GRE, GMAT, and LSAT. [ENDQ] [NEWLINE] That is true, but concluding that STEM majors are better at law than people who studied law in undergrad is not an accurate assesment. I suspect the reason this occurs is that only STEM majors who feel especially confident in their ability to pass the test take it, whereas a far greater number of non-STEM majors (who are obviously underqualified) take it even if they feel unconfident. The grades of the STEM students are, in a sense, inflated. [NEWLINE] [NEWLINE] [STARTQ] And people who are competent in math (whether
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> your cost-benefit is pretty accurate,<mask> the study you've linked points out that that is only in the short term. The study points out that for the next 20-25 years, [NEWLINE] [NEWLINE] [STARTQ] High construction costs for nuclear plants, especially relative to natural gas-fired plants, make additional options for new nuclear capacity uneconomical **until the later years of the projection**. [ENDQ] [NEWLINE] Emphasis mine. In less than half a decade, shifting to nukes *will* be economically incentivized.<mask> it become more difficult to extract the fossil fuels our civilization needs for<mask> much more than energy, that economic pressure to switch to nuclear plants will increase even more. [NEWLINE] [NEWLINE] On a more speculative note, I suspect the social pressures that keep nuclear unfairly despised in the public eye will fade<mask> the Cold War becomes more and more a thing of the past. Nuclear proliferation faced public support before people got<mask> scared of MAD, after all, and<mask><mask> that that distance, combined with economic pressure, will make nuclearization of our energy grid that much more likely. [NEWLINE] [NEWLINE] <mask><mask> that the statement [NEWLINE] [NEWLINE] [STARTQ] You couldn't possibly hope for the market to do anything about this<mask> the market will always be biased towards fossil fuel<mask> its<mask> convenient and the logistical apparatus is already in place for processing and distribution. [ENDQ] [NEWLINE] is pretty backwards, to be totally honest. Yes, it holds for the short term,<mask><mask> I mentioned before, that same economic system will be pressuring a shift *away* from fossil fuels<mask> soon<mask> they start to become unprofitable to simply burn. We *need* oil for things like plastics and rocket fuel, not to keep our houses warm. The market will eventually ensure that fossil fuels are squeezed out of the energy market. *I* only hope that that doesn't happen too late. [NEWLINE] [NEWLINE] <mask> for the [NEWLINE] [NEWLINE] [STARTQ] need to drastically curtail individual travel and impose strict regulations on diet, reproduction, and recreation [ENDQ] [NEWLINE] I think you've gotten a little far ahead of yourself. There is plenty of arable land, and GM crops only increase<mask> much we can grow, especially in the developing world.<mask><mask><mask> we take care of the arable land and stop trying to grow fuel on it, sustainable mass-farming is perfectly possible. The OP we're responding to discussed the UN's views on the peak population,<mask> reproductive sanctions are probably a counterproductive idea to even *mention*,<mask><mask> there are severe ethical implications to that sort of thing.<mask> for travel and recreation
Label encoding: <s>I think your cost-benefit is pretty accurate, but the study you've linked points out that that is only in the short term. The study points out that for the next 20-25 years, [NEWLINE] [NEWLINE] [STARTQ] High construction costs for nuclear plants, especially relative to natural gas-fired plants, make additional options for new nuclear capacity uneconomical **until the later years of the projection**. [ENDQ] [NEWLINE] Emphasis mine. In less than half a decade, shifting to nukes *will* be economically incentivized. As it become more difficult to extract the fossil fuels our civilization needs for so much more than energy, that economic pressure to switch to nuclear plants will increase even more. [NEWLINE] [NEWLINE] On a more speculative note, I suspect the social pressures that keep nuclear unfairly despised in the public eye will fade as the Cold War becomes more and more a thing of the past. Nuclear proliferation faced public support before people got so scared of MAD, after all, and I think that that distance, combined with economic pressure, will make nuclearization of our energy grid that much more likely. [NEWLINE] [NEWLINE] I think that the statement [NEWLINE] [NEWLINE] [STARTQ] You couldn't possibly hope for the market to do anything about this because the market will always be biased towards fossil fuel because its so convenient and the logistical apparatus is already in place for processing and distribution. [ENDQ] [NEWLINE] is pretty backwards, to be totally honest. Yes, it holds for the short term, but as I mentioned before, that same economic system will be pressuring a shift *away* from fossil fuels as soon as they start to become unprofitable to simply burn. We *need* oil for things like plastics and rocket fuel, not to keep our houses warm. The market will eventually ensure that fossil fuels are squeezed out of the energy market. *I* only hope that that doesn't happen too late. [NEWLINE] [NEWLINE] As for the [NEWLINE] [NEWLINE] [STARTQ] need to drastically curtail individual travel and impose strict regulations on diet, reproduction, and recreation [ENDQ] [NEWLINE] I think you've gotten a little far ahead of yourself. There is plenty of arable land, and GM crops only increase how much we can grow, especially in the developing world. So long as we take care of the arable land and stop trying to grow fuel on it, sustainable mass-farming is perfectly possible. The OP we're responding to discussed the UN's views on the peak population, so reproductive sanctions are probably a counterproductive idea to even *mention*, given that there are severe ethical implications to that sort of thing. As for travel and recreation
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [NEWLINE] [STARTQ] The Rose of Versailles, Evangelion, Princess Tutu, Dennou Coil, FMA, maybe<mask> I am not the target audience. I<mask> don't like material that is artificially stretched over too many episodes such<mask> DBZ or One Piece.<mask> I found odd is that you put Cowboy Bebop, the Ghibli stuff, Ghost in the Shell, Samurai Champloo and Akira in the same list and to me these are astounding and completely different from the others works listed.<mask><mask> Akira is specifically on another level. [ENDQ] [NEWLINE] <mask> I said, I only like DB<mask> of nostalgia, I know it's not that great and I can't even be bothered watching the show again. I just enjoyed the recent movies<mask> of nostalgia. [NEWLINE] Personally I would put Evangelion (both show and movie) on their same level: it was a completely new take on the mecha genre (well, actually there was Gunbuster before that<mask> it's by the same author anyway and I consider it more of a prototype), it had some of the best character development in any anime I've watched and I love<mask> it pisses off people who watch anime<mask> a form of escapism and expect the main character to make them feel badass (nothing wrong with that,<mask><mask> you consider a character whiny<mask> you'd probably fare<mask> bad or worse in his situation you're delusional), the movie has some of the best imagery and mind blowing scenes I've ever seen, not even Enter the Void did that to me. [NEWLINE] [NEWLINE] FMA Brotherhood is one of the few shonen shows I would consider close to perfection (in terms of shonen standards): great characters, great plot, fights that aren't just about powering up and a memorable OST. [NEWLINE] [NEWLINE] Roses of Versailles has one of the best female characters I've seen in any anime (seriously<mask> cringey<mask> I find waifu choices to be, I'd definitely take Oscar<mask> a waifu), an interesting setting and again a memorable soundtrack. I'm not even into shojo stuff<mask> I still enjoyed it. [NEWLINE] [NEWLINE] Denno Coil was a nice SciFi story with an interesting turn of events. Not a masterpiece<mask> still good. [NEWLINE] [NEWLINE] Princess Tutu did<mask> Madoka attempted to in a much better way: it took the magic girl genre and turned it on its head with a dark twist. Madoka,<mask><mask><mask><mask>, did<mask> I didn't like about Attack on Titan
Label encoding: <s> [NEWLINE] [NEWLINE] [STARTQ] The Rose of Versailles, Evangelion, Princess Tutu, Dennou Coil, FMA, maybe because I am not the target audience. I also don't like material that is artificially stretched over too many episodes such as DBZ or One Piece. What I found odd is that you put Cowboy Bebop, the Ghibli stuff, Ghost in the Shell, Samurai Champloo and Akira in the same list and to me these are astounding and completely different from the others works listed. I think Akira is specifically on another level. [ENDQ] [NEWLINE] As I said, I only like DB because of nostalgia, I know it's not that great and I can't even be bothered watching the show again. I just enjoyed the recent movies because of nostalgia. [NEWLINE] Personally I would put Evangelion (both show and movie) on their same level: it was a completely new take on the mecha genre (well, actually there was Gunbuster before that but it's by the same author anyway and I consider it more of a prototype), it had some of the best character development in any anime I've watched and I love how it pisses off people who watch anime as a form of escapism and expect the main character to make them feel badass (nothing wrong with that, but when you consider a character whiny when you'd probably fare as bad or worse in his situation you're delusional), the movie has some of the best imagery and mind blowing scenes I've ever seen, not even Enter the Void did that to me. [NEWLINE] [NEWLINE] FMA Brotherhood is one of the few shonen shows I would consider close to perfection (in terms of shonen standards): great characters, great plot, fights that aren't just about powering up and a memorable OST. [NEWLINE] [NEWLINE] Roses of Versailles has one of the best female characters I've seen in any anime (seriously as cringey as I find waifu choices to be, I'd definitely take Oscar as a waifu), an interesting setting and again a memorable soundtrack. I'm not even into shojo stuff but I still enjoyed it. [NEWLINE] [NEWLINE] Denno Coil was a nice SciFi story with an interesting turn of events. Not a masterpiece but still good. [NEWLINE] [NEWLINE] Princess Tutu did what Madoka attempted to in a much better way: it took the magic girl genre and turned it on its head with a dark twist. Madoka, on the other hand, did what I didn't like about Attack on Titan
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your entire premise is apparently based on the idea that people who aren't reporting are too stupid to weigh for themselves the pros (chance of the rapist being convicted) and the cons (chance that accusing the person will ruin their life). [ENDQ] [NEWLINE] No, that's not<mask> I'm saying at all. It's not that people are too stupid, it's that rape is an emotionally loaded subject which makes it incredibly difficult for *anyone* to do<mask> they ought to do. No matter<mask> the statistics are of conviction, reporting it is the only way to make it go up.<mask> every victim did<mask> they ought do do and report rape, then conviction rates would go up. *Your* entire premise fails to take into account that successful conviction rates are hampered by the sheer amount of people who refuse to report, and<mask><mask><mask> can't even enter the study. [NEWLINE] [NEWLINE] [STARTQ] That's not only incredibly disrespectful of rape victims,<mask> it reveals that you have<mask> I can only charitably term an under-informed understanding of<mask> rape cases are prosecuted,<mask> frequently rapists are convicted, and<mask> often happens to rape victims after they make accusations. [ENDQ] [NEWLINE] First, keep it civil.<mask> you believe I'm mistaken, you could at the very least go through the trouble of 1) saying it politely, or 2) actually provide some studies to back up your point. I try to make sure<mask> I say is backed up by information and fact, and I don't really see<mask> I've missed here.<mask> you have an example, I would be glad,<mask> I honestly want to know the best answer. The only thing you could say about my misinformation would be my example of a serial rapist, which I would agree is not the typical example.<mask>, it was a hypothetical example to prove a flawed point. [NEWLINE] [NEWLINE] [Out of every 100 rapes, 40 are reported, and of those 40 reports, 8 will be prosecuted.]( [URL] ) Of those 8 that will be prosecuted, 3 will spend a day in prison. A 20% prosecution rate and a 37.5% success rate in trial. [NEWLINE] [NEWLINE] <mask> the numbers show, clearly, is that you don't need to go to trial to report rape. Often times,<mask> I've stated previously, prosecutors will not continue to trial without consent of the victim.<mask> I'm saying, is that people have the obligation to, at the very least, contribute by reporting their rape to make the prosecutor's job easier<mask> the next victim does want to go to trial.
Label encoding: <s> [STARTQ] Your entire premise is apparently based on the idea that people who aren't reporting are too stupid to weigh for themselves the pros (chance of the rapist being convicted) and the cons (chance that accusing the person will ruin their life). [ENDQ] [NEWLINE] No, that's not what I'm saying at all. It's not that people are too stupid, it's that rape is an emotionally loaded subject which makes it incredibly difficult for *anyone* to do what they ought to do. No matter what the statistics are of conviction, reporting it is the only way to make it go up. If every victim did what they ought do do and report rape, then conviction rates would go up. *Your* entire premise fails to take into account that successful conviction rates are hampered by the sheer amount of people who refuse to report, and as a result can't even enter the study. [NEWLINE] [NEWLINE] [STARTQ] That's not only incredibly disrespectful of rape victims, but it reveals that you have what I can only charitably term an under-informed understanding of how rape cases are prosecuted, how frequently rapists are convicted, and what often happens to rape victims after they make accusations. [ENDQ] [NEWLINE] First, keep it civil. If you believe I'm mistaken, you could at the very least go through the trouble of 1) saying it politely, or 2) actually provide some studies to back up your point. I try to make sure what I say is backed up by information and fact, and I don't really see what I've missed here. If you have an example, I would be glad, because I honestly want to know the best answer. The only thing you could say about my misinformation would be my example of a serial rapist, which I would agree is not the typical example. However, it was a hypothetical example to prove a flawed point. [NEWLINE] [NEWLINE] [Out of every 100 rapes, 40 are reported, and of those 40 reports, 8 will be prosecuted.]( [URL] ) Of those 8 that will be prosecuted, 3 will spend a day in prison. A 20% prosecution rate and a 37.5% success rate in trial. [NEWLINE] [NEWLINE] What the numbers show, clearly, is that you don't need to go to trial to report rape. Often times, as I've stated previously, prosecutors will not continue to trial without consent of the victim. What I'm saying, is that people have the obligation to, at the very least, contribute by reporting their rape to make the prosecutor's job easier when the next victim does want to go to trial.
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My view is that in some situations a nazi comparison can be a useful and valid tool in an argument, and that<mask> such useful comparisons are made, "Godwin's Law" and other similar cultural memes serve only to make the discussion less fruitful. [An example]( [URL] ) showed up earlier today in an /r/Foodforthought thread. [NEWLINE] [NEWLINE] /u/Westlondonwannabe said: [NEWLINE] [STARTQ] Don't totally disagree with you.<mask> something someone said to me once has always stuck : The death penalty is society saying, we collectively, choose to remove you from our group due to your heinous actions. The government is simply acting on our wishes<mask> a society. [ENDQ] [NEWLINE] To which /u/tvrr replied: [NEWLINE] [STARTQ] This one didn't fly to well in Nazi Germany. [ENDQ] [NEWLINE] /u/tvrr was downvoted and people replied with comments about Godwin's law, and<mask> you "just can't compare" the two. [NEWLINE] [NEWLINE] Now, I'm not saying that the comparison is exactly correct,<mask> rather that elucidating exactly<mask> the death penalty is different from the holocaust (i.e. actually responding to the content of the comparison, rather than just dismissing it) would be an important contribution to the conversation, allowing us to more clearly understand the parameters under which /u/Westlondonwannabe's statement is true, which to me is at the heart of<mask> people are trying to get from this discussion. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, I'm not saying that there are no circumstances under which Nazi comparisons can be wrongly used. Two examples that<mask><mask> of immediately are: [NEWLINE] [NEWLINE] * Person A makes the comparison, to which person B responds with 'This situation is different<mask>..." and then person A comes back with 'Oh,<mask> you're on the Nazi's side?' [NEWLINE] * The comparison is 'The Nazis did X too',<mask> X has nothing to do with<mask> people hate the Nazis. [NEWLINE] [NEWLINE] <mask>, go ahead and CMV! [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdown
Label encoding: <s>My view is that in some situations a nazi comparison can be a useful and valid tool in an argument, and that when such useful comparisons are made, "Godwin's Law" and other similar cultural memes serve only to make the discussion less fruitful. [An example]( [URL] ) showed up earlier today in an /r/Foodforthought thread. [NEWLINE] [NEWLINE] /u/Westlondonwannabe said: [NEWLINE] [STARTQ] Don't totally disagree with you. But something someone said to me once has always stuck : The death penalty is society saying, we collectively, choose to remove you from our group due to your heinous actions. The government is simply acting on our wishes as a society. [ENDQ] [NEWLINE] To which /u/tvrr replied: [NEWLINE] [STARTQ] This one didn't fly to well in Nazi Germany. [ENDQ] [NEWLINE] /u/tvrr was downvoted and people replied with comments about Godwin's law, and how you "just can't compare" the two. [NEWLINE] [NEWLINE] Now, I'm not saying that the comparison is exactly correct, but rather that elucidating exactly how the death penalty is different from the holocaust (i.e. actually responding to the content of the comparison, rather than just dismissing it) would be an important contribution to the conversation, allowing us to more clearly understand the parameters under which /u/Westlondonwannabe's statement is true, which to me is at the heart of what people are trying to get from this discussion. [NEWLINE] [NEWLINE] On the other hand, I'm not saying that there are no circumstances under which Nazi comparisons can be wrongly used. Two examples that I think of immediately are: [NEWLINE] [NEWLINE] * Person A makes the comparison, to which person B responds with 'This situation is different because..." and then person A comes back with 'Oh, so you're on the Nazi's side?' [NEWLINE] * The comparison is 'The Nazis did X too', when X has nothing to do with why people hate the Nazis. [NEWLINE] [NEWLINE] So, go ahead and CMV! [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdown
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] First Amendment<mask> it relates to corporate personhood and lobbying [ENDQ] [NEWLINE] Citizens United could not have been ruled any other way. <mask> you, I, and 300 of our closest friends want to spend money to buy a commercial to express an idea, our options are to give money to one person, whom we then have to trust won't just walk off with it, *or* to create a legal entity to further that expression *for us.*  That's the only way that groups of average people can get their message heard<mask> we're in competition with Koch, Soros, Cocacola, UPS, etc. [NEWLINE] [NEWLINE] Or do you think there should be a cap on money spent on all first amendment rights?  "You're allowed to worship<mask> you want,<mask> you're not allowed to build your Synagogue,<mask> it's more than $200k.  You can express yourself,<mask><mask> you spend more than $10k on printing flyers to hand out, that runs afoul of campaign finance laws." [NEWLINE] [NEWLINE] TL;DR for *Citizens United*: [NEWLINE] [NEWLINE] * Corporations are *made up of* People, and to limit their expression is to limit the expression of those people. [NEWLINE] * Money isn't speech,<mask> to limit money limits speech. [NEWLINE] [NEWLINE] from the second link: [NEWLINE] [NEWLINE] [STARTQ] the frequency of *gun* murders, and the shockingly high number of annual *gun* deaths [emphasis added] [ENDQ] [NEWLINE] Oh, I was completely unaware that [victims of knife crime]( [URL] ) get better after they've been murdered... [NEWLINE] [NEWLINE] <mask> I can cite things, too.  In my case, rather than some op-ed by someone who is afraid of guns rather than murderers, [I'm going to cite Harvard (PDF)]( [URL].pdf) who found, to their surprise, that there is *no* evidence that gun control has any benefit<mask><mask> ever,<mask> that there *is* some (inconclusive) evidence that it *hurts* things. [NEWLINE] [NEWLINE] <mask><mask> actual *scientists* have found no problem with the implications of the 2nd amendment...<mask> would you update about it? [NEWLINE] [NEWLINE] [STARTQ] Fourth Amendment<mask> it relates to privacy, specifically digital privacy [ENDQ] [NEWLINE] <mask>,<mask>'s the problem here?  That they aren't treating "persons, papers, houses, and effects" rationally for the modern, digital world? That could be fixed by a simple law stating that files, phones, etc, are effectively "papers and
Label encoding: <s> [STARTQ] First Amendment as it relates to corporate personhood and lobbying [ENDQ] [NEWLINE] Citizens United could not have been ruled any other way.  If you, I, and 300 of our closest friends want to spend money to buy a commercial to express an idea, our options are to give money to one person, whom we then have to trust won't just walk off with it, *or* to create a legal entity to further that expression *for us.*  That's the only way that groups of average people can get their message heard when we're in competition with Koch, Soros, Cocacola, UPS, etc. [NEWLINE] [NEWLINE] Or do you think there should be a cap on money spent on all first amendment rights?  "You're allowed to worship however you want, but you're not allowed to build your Synagogue, because it's more than $200k.  You can express yourself, but if you spend more than $10k on printing flyers to hand out, that runs afoul of campaign finance laws." [NEWLINE] [NEWLINE] TL;DR for *Citizens United*: [NEWLINE] [NEWLINE] * Corporations are *made up of* People, and to limit their expression is to limit the expression of those people. [NEWLINE] * Money isn't speech, but to limit money limits speech. [NEWLINE] [NEWLINE] from the second link: [NEWLINE] [NEWLINE] [STARTQ] the frequency of *gun* murders, and the shockingly high number of annual *gun* deaths [emphasis added] [ENDQ] [NEWLINE] Oh, I was completely unaware that [victims of knife crime]( [URL] ) get better after they've been murdered... [NEWLINE] [NEWLINE] But I can cite things, too.  In my case, rather than some op-ed by someone who is afraid of guns rather than murderers, [I'm going to cite Harvard (PDF)]( [URL].pdf) who found, to their surprise, that there is *no* evidence that gun control has any benefit what so ever, but that there *is* some (inconclusive) evidence that it *hurts* things. [NEWLINE] [NEWLINE] Given that actual *scientists* have found no problem with the implications of the 2nd amendment... what would you update about it? [NEWLINE] [NEWLINE] [STARTQ] Fourth Amendment as it relates to privacy, specifically digital privacy [ENDQ] [NEWLINE] So, what's the problem here?  That they aren't treating "persons, papers, houses, and effects" rationally for the modern, digital world? That could be fixed by a simple law stating that files, phones, etc, are effectively "papers and
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Quick answer: Language don't work that way. No language that I'm aware of has ever had a "regulator" and even<mask> one did it wouldn't have made a difference. Language is constructed by need, not by design, and those constructed languages I'm aware of have all failed. The closest you might come to a successful constructed language is Various sign languages created for the deaf and mute,<mask> even those have branched out and changed based on usage. [ENDQ] [NEWLINE] You are mistaken.  [Many,<mask> not most, major world languages have regulatory bodies.]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] In my quick reading regarding anglish this seem to come up a lot. I'm not really sure it's a valid criticism. Basically the argument is that you don't know<mask> words mean until you've learned<mask> those words mean. I'm not sure<mask> the problem is with that. I was equally unaware of<mask> "goodnessfrod" or "cracklore of bones" meant before I read your link. [ENDQ] [NEWLINE] You know<mask> "crack", "lore", and "bones" mean. <mask> you saw the term in context, I'm sure you would have been able to work out the meaning.  (Clearly, it refers to drug-induced fan theories about the tv series starring Emily Deschanel :P ) <mask> would be even easier<mask> you had grown up speaking Anglish, and were accustomed to hearing "crack" used in a medical context, and "lore" used to mean a field of academic study. [NEWLINE] [NEWLINE] [STARTQ] Find me a living spoken language that has these. [ENDQ] [NEWLINE] A lot of languages have one or more of those traits, especially the regulated ones.  (At least officially.  There may be some rule-breaking in slang, of course,<mask> at least the official version provides an ideal for the educated to strive towards.) [NEWLINE] [NEWLINE] [STARTQ] <mask>'s<mask> great about Englishes roots?<mask> stop there? Wouldn't Germanic be even better? More pure? [ENDQ] [NEWLINE] Actually, Anglish does use roots from other Germanic languages<mask> an appropriate Anglo-Saxon word can't be found. [NEWLINE] [NEWLINE] It's not about any language being better than any other.  It would be just<mask> well to speak pure Greek or pure Latin.  It's about consistency.  Keeping the number of phonemes from becoming overwhelming.  Using words that are conjugated the same way,<mask><mask> the same rules.  Using compounds of common words<mask>
Label encoding: <s> [STARTQ] Quick answer: Language don't work that way. No language that I'm aware of has ever had a "regulator" and even if one did it wouldn't have made a difference. Language is constructed by need, not by design, and those constructed languages I'm aware of have all failed. The closest you might come to a successful constructed language is Various sign languages created for the deaf and mute, but even those have branched out and changed based on usage. [ENDQ] [NEWLINE] You are mistaken.  [Many, if not most, major world languages have regulatory bodies.]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] In my quick reading regarding anglish this seem to come up a lot. I'm not really sure it's a valid criticism. Basically the argument is that you don't know what words mean until you've learned what those words mean. I'm not sure where the problem is with that. I was equally unaware of what "goodnessfrod" or "cracklore of bones" meant before I read your link. [ENDQ] [NEWLINE] You know what "crack", "lore", and "bones" mean.  If you saw the term in context, I'm sure you would have been able to work out the meaning.  (Clearly, it refers to drug-induced fan theories about the tv series starring Emily Deschanel :P )  If would be even easier if you had grown up speaking Anglish, and were accustomed to hearing "crack" used in a medical context, and "lore" used to mean a field of academic study. [NEWLINE] [NEWLINE] [STARTQ] Find me a living spoken language that has these. [ENDQ] [NEWLINE] A lot of languages have one or more of those traits, especially the regulated ones.  (At least officially.  There may be some rule-breaking in slang, of course, but at least the official version provides an ideal for the educated to strive towards.) [NEWLINE] [NEWLINE] [STARTQ] What's so great about Englishes roots? Why stop there? Wouldn't Germanic be even better? More pure? [ENDQ] [NEWLINE] Actually, Anglish does use roots from other Germanic languages when an appropriate Anglo-Saxon word can't be found. [NEWLINE] [NEWLINE] It's not about any language being better than any other.  It would be just as well to speak pure Greek or pure Latin.  It's about consistency.  Keeping the number of phonemes from becoming overwhelming.  Using words that are conjugated the same way, according to the same rules.  Using compounds of common words when
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You know<mask>, I mostly agree with you. <mask> can one person do to change this vicious cycle?  My life feels like an individual ant being forced to go with the flow of the colony.  The major question for me is<mask> the direction of the colony (human nature) is truly beyond hope or<mask> it could maybe change for the better.  Greed is ingrained in human nature it's hard to ever imagine that changing.  I too fear humans will someday go extinct<mask><mask><mask> of our own action,<mask> maybe not... [NEWLINE] [NEWLINE] In the past 5-10 years it's become increasingly popular to "go green".  I've actually been surprised to see<mask> much mainstream it has become to care about the environment.  Nonetheless, that greed still exists and corporations are doing everything they can to sustain their business and ensure things like oil don't become irrelevant.  The ability for a few to control the masses is extremely depressing to me.  Perhaps the problem is beyond a few individuals and it is our will to resist change that is the root cause. [NEWLINE] [NEWLINE] I, too, eat<mask> much meat<mask> I want and don't think twice about taking road trips.  My personal view is that people are going to do<mask> they want to do.  That said, instead of hoping to change human nature, perhaps we need to work on better alternatives. <mask> humans are in a pinch, we are blessed with ingenuity.  It's sad to think we may have to wait until the oil reserves dry up before changing,<mask> I have a strong belief we are capable of creating better tools. <mask><mask><mask>, I don't think humans will easily go extinct. <mask><mask> we're extremely crafty and will find a way to survive.  It would be great to see the negative aspects of our nature disappear,<mask> they probably won't.  That doesn't mean all is lost<mask><mask> don't give up on us! [NEWLINE] [NEWLINE] I don't think you should hold any guilt for doing<mask> most everyone else does too.  The important part is you actually recognize the issue and think twice about it.  I can relate a ton to this attitude,<mask> that doesn't mean we should give up.  It feels hopeless now,<mask> that's only<mask> it's all or nothing.  Our choices at the time being are don't eat meat or don't drive.  Those don't seem very practical now do they?  I'm confident that<mask> enough people are able to *recognize* the
Label encoding: <s>You know what, I mostly agree with you.  What can one person do to change this vicious cycle?  My life feels like an individual ant being forced to go with the flow of the colony.  The major question for me is if the direction of the colony (human nature) is truly beyond hope or if it could maybe change for the better.  Greed is ingrained in human nature it's hard to ever imagine that changing.  I too fear humans will someday go extinct as a result of our own action, but maybe not... [NEWLINE] [NEWLINE] In the past 5-10 years it's become increasingly popular to "go green".  I've actually been surprised to see how much mainstream it has become to care about the environment.  Nonetheless, that greed still exists and corporations are doing everything they can to sustain their business and ensure things like oil don't become irrelevant.  The ability for a few to control the masses is extremely depressing to me.  Perhaps the problem is beyond a few individuals and it is our will to resist change that is the root cause. [NEWLINE] [NEWLINE] I, too, eat as much meat as I want and don't think twice about taking road trips.  My personal view is that people are going to do what they want to do.  That said, instead of hoping to change human nature, perhaps we need to work on better alternatives.  When humans are in a pinch, we are blessed with ingenuity.  It's sad to think we may have to wait until the oil reserves dry up before changing, but I have a strong belief we are capable of creating better tools.  Because of this, I don't think humans will easily go extinct.  I think we're extremely crafty and will find a way to survive.  It would be great to see the negative aspects of our nature disappear, but they probably won't.  That doesn't mean all is lost though so don't give up on us! [NEWLINE] [NEWLINE] I don't think you should hold any guilt for doing what most everyone else does too.  The important part is you actually recognize the issue and think twice about it.  I can relate a ton to this attitude, but that doesn't mean we should give up.  It feels hopeless now, but that's only because it's all or nothing.  Our choices at the time being are don't eat meat or don't drive.  Those don't seem very practical now do they?  I'm confident that if enough people are able to *recognize* the
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, I mean--I know<mask> sort of privileges you must have. <mask><mask> you want us to convince you to do something differently, we (or "I" at least) need a better idea of<mask> you're doing now. [NEWLINE] [NEWLINE] I'm not trying to make you feel bad.  Feeling bad isn't going to help anyone.  That's<mask> not<mask> "check your privilege" is supposed to be telling you to do, either. [NEWLINE] [NEWLINE] For example, say you notice your female coworker getting sexually harassed by your boss.  You *could* just ignore it.  That's<mask> most people would do.  Or you could stick your neck out and try any of a variety of techniques to fix the situation.  One of these options is risky for you,<mask> it's the kind, caring, compassionate thing to do.  The other one sounds kind-of cowardly now that I've phrased the dilemma like this,<mask> it's still very likely to be the option that gets you ahead at your job. [NEWLINE] [NEWLINE] <mask><mask> should you get involved in something like that?  Well, I don't know.  Maybe<mask> you don't like seeing injustice in the world?  Or maybe<mask> your coworker bought you a coffee that one time and you want to pay her back.  Or maybe<mask> part of the way you think of yourself involves you being willing to fight for people who are disadvantaged, sort-of like a badass superhero.  Or maybe<mask> you'll feel good doing it.  Or maybe<mask> the whole scenario makes you angry.  Or maybe<mask> you're actively trying to be a better person by doing nice things for people.  Or maybe<mask> you have a mindset in which you should "be the change that you wish to see in the world" or some such. [NEWLINE] [NEWLINE] Incidentally,<mask> change *would* you like to see in the world?  Do you pretty much want to keep the status quo<mask> it is?  Or would you like to make the world more fair and equitable for people from a variety of different backgrounds?  Maybe you don't really care.  In that case: <mask> do you want from your life?  Are you religious?  Do you want to do anything important, or contribute anything lasting to society or to the world?  Do you want to have kids?  Or, maybe you haven't thought that far ahead.  Maybe you're happy just getting by and living life<mask> it comes, for the time
Label encoding: <s>No, I mean--I know what sort of privileges you must have.  But if you want us to convince you to do something differently, we (or "I" at least) need a better idea of what you're doing now. [NEWLINE] [NEWLINE] I'm not trying to make you feel bad.  Feeling bad isn't going to help anyone.  That's also not what "check your privilege" is supposed to be telling you to do, either. [NEWLINE] [NEWLINE] For example, say you notice your female coworker getting sexually harassed by your boss.  You *could* just ignore it.  That's what most people would do.  Or you could stick your neck out and try any of a variety of techniques to fix the situation.  One of these options is risky for you, but it's the kind, caring, compassionate thing to do.  The other one sounds kind-of cowardly now that I've phrased the dilemma like this, but it's still very likely to be the option that gets you ahead at your job. [NEWLINE] [NEWLINE] So why should you get involved in something like that?  Well, I don't know.  Maybe because you don't like seeing injustice in the world?  Or maybe because your coworker bought you a coffee that one time and you want to pay her back.  Or maybe because part of the way you think of yourself involves you being willing to fight for people who are disadvantaged, sort-of like a badass superhero.  Or maybe because you'll feel good doing it.  Or maybe because the whole scenario makes you angry.  Or maybe because you're actively trying to be a better person by doing nice things for people.  Or maybe because you have a mindset in which you should "be the change that you wish to see in the world" or some such. [NEWLINE] [NEWLINE] Incidentally, what change *would* you like to see in the world?  Do you pretty much want to keep the status quo as it is?  Or would you like to make the world more fair and equitable for people from a variety of different backgrounds?  Maybe you don't really care.  In that case:  what do you want from your life?  Are you religious?  Do you want to do anything important, or contribute anything lasting to society or to the world?  Do you want to have kids?  Or, maybe you haven't thought that far ahead.  Maybe you're happy just getting by and living life as it comes, for the time
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Sometimes you don't throw away your beam after you cut part of it off. Like<mask> you cut 123 cm off a 500 cm beam or something. [ENDQ] [NEWLINE] <mask> often do you have a piece that is a size you need,<mask> the Imperial system prevented you from knowing that? [NEWLINE] [NEWLINE] My original point, was that the world does work in nice clean increments,<mask> having a system that works in the fashion doesn't gain a huge advantage over another system. [NEWLINE] [NEWLINE] [STARTQ] You are basically arguing against the metric system based on transition costs. Transition costs don't enter into a discussion of which system is best.<mask> you want to admit that metric is superior,<mask> then say that transition costs are too high<mask> we should keep using an inferior system, that's a discussion I'd be willing to have<mask> right now I'm just talking about which system is better. [ENDQ] [NEWLINE] I never said the Imperial system was better. [NEWLINE] [NEWLINE] The original topic was the Imperial system being useless, which it is not (it has obviously served us well enough<mask> far). [NEWLINE] [NEWLINE] My argument with that point was that the transition cost would outweigh the benefit of moving to the new system. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Furthermore, there is a satellite industry in Europe too and they use software<mask> well. The way the world is now, either every software developer has to design their program for two different systems of units, or two totally different softwares that do the essentially same thing have to be developed. That is not a good thing. [ENDQ] [NEWLINE] That's great for their satellites,<mask><mask> about ours? These constants are specific for our satellite, which are currently on orbit. The software used is proprietary and only used on these satellites,<mask> it isn't designed around the Metric system. [NEWLINE] [NEWLINE] <mask> we needed to rederive the constants and rewrite the software, we wouldn't have the satellite on-hand to verify it worked correctly. It would all be simulated until we actually made the change, in which case we cross our fingers that the simulation was correct. [NEWLINE] [NEWLINE] [STARTQ] <mask> an engineer I can multiply 343 by 228 without a calculator<mask> I don't want to<mask> it's a waste of time. Just like, yes, I can convert units<mask> it is a waste of time. No one is arguing that converting units is impossible, we are just saying that it is wasteful. [ENDQ] [NEWLINE] It is a valid point that there can be an extra layer of work to convert units. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, having to manage units builds a certain "
Label encoding: <s> [STARTQ] Sometimes you don't throw away your beam after you cut part of it off. Like if you cut 123 cm off a 500 cm beam or something. [ENDQ] [NEWLINE] How often do you have a piece that is a size you need, but the Imperial system prevented you from knowing that? [NEWLINE] [NEWLINE] My original point, was that the world does work in nice clean increments, so having a system that works in the fashion doesn't gain a huge advantage over another system. [NEWLINE] [NEWLINE] [STARTQ] You are basically arguing against the metric system based on transition costs. Transition costs don't enter into a discussion of which system is best. If you want to admit that metric is superior, but then say that transition costs are too high so we should keep using an inferior system, that's a discussion I'd be willing to have but right now I'm just talking about which system is better. [ENDQ] [NEWLINE] I never said the Imperial system was better. [NEWLINE] [NEWLINE] The original topic was the Imperial system being useless, which it is not (it has obviously served us well enough thus far). [NEWLINE] [NEWLINE] My argument with that point was that the transition cost would outweigh the benefit of moving to the new system. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Furthermore, there is a satellite industry in Europe too and they use software as well. The way the world is now, either every software developer has to design their program for two different systems of units, or two totally different softwares that do the essentially same thing have to be developed. That is not a good thing. [ENDQ] [NEWLINE] That's great for their satellites, but what about ours? These constants are specific for our satellite, which are currently on orbit. The software used is proprietary and only used on these satellites, so it isn't designed around the Metric system. [NEWLINE] [NEWLINE] If we needed to rederive the constants and rewrite the software, we wouldn't have the satellite on-hand to verify it worked correctly. It would all be simulated until we actually made the change, in which case we cross our fingers that the simulation was correct. [NEWLINE] [NEWLINE] [STARTQ] As an engineer I can multiply 343 by 228 without a calculator but I don't want to because it's a waste of time. Just like, yes, I can convert units but it is a waste of time. No one is arguing that converting units is impossible, we are just saying that it is wasteful. [ENDQ] [NEWLINE] It is a valid point that there can be an extra layer of work to convert units. [NEWLINE] [NEWLINE] On the other hand, having to manage units builds a certain "
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Honestly you sound like someone who I wouldn't be totally against carrying a weapon. From<mask> you have written, we seem to have a similar understanding of the purpose of guns and<mask> they may be necessary in extreme situations and<mask> they should not be produced (going back to the different convenience store scenarios). [NEWLINE] [NEWLINE] Unfortunately<mask>, not everyone who is able to obtain a gun and be properly licenced for it has the same, or even a similar, understanding. In the robbery scenario<mask> no one would get hurt, they may be the person who would produce their gun and make the situation worse.<mask>, whilst I wasn't aware of the necesary training course to obtain a concealed permit, it does kind of sound like one of those one-day, easy to pass kind of things, like a Responsible Service of Alcohol certification (Australian thing which lets you work at bars etc). From a quick look on Google, it did look like it was something that heavily varied from state to state<mask> well, being non-existent in some. [NEWLINE] [NEWLINE] Please correct me<mask> I am completely wrong on that assumption<mask>. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask> the "compulsory police training". I'm on mobile<mask> I can't really link<mask> there are studies out there that prove police aren't trained enough and don't always have the best aim. [ENDQ] [NEWLINE] Oh well I am not saying exactly like Police training,<mask> like something extensive about<mask> and<mask> not it is ok to produce a gun, etc. Not sure exactly<mask> would be included in such a course<mask> you know<mask> I am getting at. [NEWLINE] [NEWLINE] [STARTQ] it's been posted before they have no legal obligation to help. [ENDQ] [NEWLINE] Wow, I did notice that elsewhere in the thread<mask> it's hard to believe (not disputing it, I am just truly amazed). Can't believe that the Police in America don't have to perform the most fundamental part of their job.<mask><mask><mask> in most cases the Police are going to get involved<mask> they see something wrong (atleast I would hope<mask> ). [NEWLINE] [NEWLINE] [STARTQ] Something you most certainly can't do<mask> your entire population is unarmed. [ENDQ] [NEWLINE] <mask><mask>. I am not arguing for the complete dearmment (is that a word?) of America, I just don't think many,<mask> any, people should be allowed to carry these guns around in everyday life.<mask> even<mask> there are people like yourself who don't take such a responsibility lightly, there are people out there with the same privilege who are completely unapp
Label encoding: <s>Honestly you sound like someone who I wouldn't be totally against carrying a weapon. From what you have written, we seem to have a similar understanding of the purpose of guns and why they may be necessary in extreme situations and when they should not be produced (going back to the different convenience store scenarios). [NEWLINE] [NEWLINE] Unfortunately though, not everyone who is able to obtain a gun and be properly licenced for it has the same, or even a similar, understanding. In the robbery scenario where no one would get hurt, they may be the person who would produce their gun and make the situation worse. Additionally, whilst I wasn't aware of the necesary training course to obtain a concealed permit, it does kind of sound like one of those one-day, easy to pass kind of things, like a Responsible Service of Alcohol certification (Australian thing which lets you work at bars etc). From a quick look on Google, it did look like it was something that heavily varied from state to state as well, being non-existent in some. [NEWLINE] [NEWLINE] Please correct me if I am completely wrong on that assumption though. [NEWLINE] [NEWLINE] [STARTQ] As far as the "compulsory police training". I'm on mobile so I can't really link but there are studies out there that prove police aren't trained enough and don't always have the best aim. [ENDQ] [NEWLINE] Oh well I am not saying exactly like Police training, but like something extensive about when and when not it is ok to produce a gun, etc. Not sure exactly what would be included in such a course but you know what I am getting at. [NEWLINE] [NEWLINE] [STARTQ] it's been posted before they have no legal obligation to help. [ENDQ] [NEWLINE] Wow, I did notice that elsewhere in the thread but it's hard to believe (not disputing it, I am just truly amazed). Can't believe that the Police in America don't have to perform the most fundamental part of their job. Though I think in most cases the Police are going to get involved when they see something wrong (atleast I would hope so ). [NEWLINE] [NEWLINE] [STARTQ] Something you most certainly can't do if your entire population is unarmed. [ENDQ] [NEWLINE] I agree. I am not arguing for the complete dearmment (is that a word?) of America, I just don't think many, if any, people should be allowed to carry these guns around in everyday life. Because even if there are people like yourself who don't take such a responsibility lightly, there are people out there with the same privilege who are completely unapp
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The Anti-PC movement's mistake is acting like "political correctness" is some scary new idea. It's not. [ENDQ] [NEWLINE] You are correct.  We're not the first country/government/people to practice fascism. [NEWLINE] [NEWLINE] [STARTQ] It used to go by names like "common decency," "manners," or "not being a dick." [ENDQ] [NEWLINE] This is equating "you're welcome" to a cultural phenomenon of self-censorship that led to a movement which routinely spawns internet lynch mobs<mask> someone said something to another person who, *in their own subjective opinion*, didn't like it. [NEWLINE] [NEWLINE] [STARTQ] <mask> do you think the world was like before PC culture? [ENDQ] [NEWLINE] Honest.  [Here]( [URL] ;feature=player_detailpage#t=1202) is Carlin on the subject. [NEWLINE] [NEWLINE] [STARTQ] Could you just go around saying whatever you wanted to people and not face any consequences? Of course not. [ENDQ] [NEWLINE] You are stating the obvious. [NEWLINE] [NEWLINE] [STARTQ] Sure people like Trump can run around acting like victims<mask> they feel entitled to be<mask> inconsiderate<mask> they want without having people dislike them [ENDQ] [NEWLINE] People can play the victim for a variety of reasons. [NEWLINE] [NEWLINE] [STARTQ] <mask> in reality there isn't some new censorship conspiracy. [ENDQ] [NEWLINE] <mask> are you trying to say here? [NEWLINE] [NEWLINE] [STARTQ] You can't say "Mexico sends their rapists and maybe some good people," [ENDQ] [NEWLINE] Yes you can. [NEWLINE] [NEWLINE] [STARTQ] you can't walk down the street and call everyone you see ugly. [ENDQ] [NEWLINE] Yes you can. [NEWLINE] [NEWLINE] [STARTQ] Not<mask> people think you shouldn't be allowed [ENDQ] [NEWLINE] This is exactly<mask> supporters of political correctness believe you can't say those things. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> no one wants to be around someone who refuses to make the effort to consider other people<mask> expressing themselves. [ENDQ] [NEWLINE] Which is exactly<mask> I, personally, would do one of the following and in this order.  Ignore it.  Remove myself from the situation. [NEWLINE] [NEWLINE] At it's most basic level an offense is a disagreement and I would *not* attempt - through law or otherwise - to limit<mask> is within the acceptable range of human expression just<mask> I have a different opinion, no matter<mask> abhorrent that opinion. [NEWLINE] [NEWLINE] [STARTQ] <mask> it has been greatly exaggerated by people who want to be assholes with impunity. [ENDQ] [NEWLINE] No it hasn't.  We've come to a point<mask> [a Nobel laureate literally cannot get a job in academia for one innocuous joke]( [URL] /). [NEWLINE]
Label encoding: <s> [STARTQ] The Anti-PC movement's mistake is acting like "political correctness" is some scary new idea. It's not. [ENDQ] [NEWLINE] You are correct.  We're not the first country/government/people to practice fascism. [NEWLINE] [NEWLINE] [STARTQ] It used to go by names like "common decency," "manners," or "not being a dick." [ENDQ] [NEWLINE] This is equating "you're welcome" to a cultural phenomenon of self-censorship that led to a movement which routinely spawns internet lynch mobs because someone said something to another person who, *in their own subjective opinion*, didn't like it. [NEWLINE] [NEWLINE] [STARTQ] What do you think the world was like before PC culture? [ENDQ] [NEWLINE] Honest.  [Here]( [URL] ;feature=player_detailpage#t=1202) is Carlin on the subject. [NEWLINE] [NEWLINE] [STARTQ] Could you just go around saying whatever you wanted to people and not face any consequences? Of course not. [ENDQ] [NEWLINE] You are stating the obvious. [NEWLINE] [NEWLINE] [STARTQ] Sure people like Trump can run around acting like victims because they feel entitled to be as inconsiderate as they want without having people dislike them [ENDQ] [NEWLINE] People can play the victim for a variety of reasons. [NEWLINE] [NEWLINE] [STARTQ] but in reality there isn't some new censorship conspiracy. [ENDQ] [NEWLINE] What are you trying to say here? [NEWLINE] [NEWLINE] [STARTQ] You can't say "Mexico sends their rapists and maybe some good people," [ENDQ] [NEWLINE] Yes you can. [NEWLINE] [NEWLINE] [STARTQ] you can't walk down the street and call everyone you see ugly. [ENDQ] [NEWLINE] Yes you can. [NEWLINE] [NEWLINE] [STARTQ] Not because people think you shouldn't be allowed [ENDQ] [NEWLINE] This is exactly why supporters of political correctness believe you can't say those things. [NEWLINE] [NEWLINE] [STARTQ] but because no one wants to be around someone who refuses to make the effort to consider other people when expressing themselves. [ENDQ] [NEWLINE] Which is exactly why I, personally, would do one of the following and in this order.  Ignore it.  Remove myself from the situation. [NEWLINE] [NEWLINE] At it's most basic level an offense is a disagreement and I would *not* attempt - through law or otherwise - to limit what is within the acceptable range of human expression just because I have a different opinion, no matter how abhorrent that opinion. [NEWLINE] [NEWLINE] [STARTQ] but it has been greatly exaggerated by people who want to be assholes with impunity. [ENDQ] [NEWLINE] No it hasn't.  We've come to a point where [a Nobel laureate literally cannot get a job in academia for one innocuous joke]( [URL] /). [NEWLINE]
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll approach this paragraph-by-paragraph,<mask> I'm just going to quote the header you put on it for simplicity. [NEWLINE] [NEWLINE] [STARTQ] Women are physically inferior to men on average. [ENDQ] [NEWLINE] Fitness standards can be changed. <mask><mask> that it's counterproductive to have fitness standards be less for women than men,<mask> we just need to have an entirely non-discriminatory approach to this rather than keeping women from even having the chance to be discriminated against by excluding them altogether.  The second article that you linked with reference to injuries<mask> mentioned that the difference in injury rate dropped off significantly by the end of basic training, which<mask><mask> suggests that women statistically start off in worse shape<mask> improve more during the session.  Looking at the traditional societal relationship between athletics and sex, that makes sense to me. [NEWLINE] [NEWLINE] [STARTQ] "Standards" are consistently changed in a manner that makes military life "easier" for women than men. [ENDQ] [NEWLINE] Let's just not have different standards, then.  Yes, fewer women will qualify than men,<mask> obviously these standards are in place to ensure that new recruits won't be a danger to themselves or others. [NEWLINE] [NEWLINE] [STARTQ] Pregnancies pull women away from duty. [ENDQ] [NEWLINE] <mask><mask> that it would be reasonable to enact a form of mandatory birth control.  Enforcing it is a more complex issue,<mask> nowhere near impossible.  There are long-term birth control methods that involve the [implantation of a tiny slow-release hormone capsule under the skin]( [URL] ) that prevents ovulation. <mask> I recall correctly, most hormonal birth control methods will<mask> show up in urine tests (and certainly in blood tests),<mask> that's a relatively simple enforcement tool.  In the case of the subdural implant,<mask>, you likely wouldn't need that<mask> they aren't self-administered. [NEWLINE] [NEWLINE] Yes, there is a small chance of failure,<mask> I don't think that it would be enough to effect unit cohesion any more significantly than the various ailments that men can suffer from. [NEWLINE] [NEWLINE] [STARTQ] Women can join the civilian sector equivalents. [ENDQ] [NEWLINE] Many women don't want to serve in support or civilian roles, and I'm sure you know plenty that feel this way.  Ultimately, it's about choice, and the availability of alternatives that are not the same doesn't mean that we should exclude an entire gender. [NEWLINE] [NEWLINE] <mask> an ending note, I basically agree with you that the last three points you mentioned aren't sufficient. <mask><mask><mask> that
Label encoding: <s>I'll approach this paragraph-by-paragraph, but I'm just going to quote the header you put on it for simplicity. [NEWLINE] [NEWLINE] [STARTQ] Women are physically inferior to men on average. [ENDQ] [NEWLINE] Fitness standards can be changed.  I agree that it's counterproductive to have fitness standards be less for women than men, but we just need to have an entirely non-discriminatory approach to this rather than keeping women from even having the chance to be discriminated against by excluding them altogether.  The second article that you linked with reference to injuries also mentioned that the difference in injury rate dropped off significantly by the end of basic training, which I agree suggests that women statistically start off in worse shape but improve more during the session.  Looking at the traditional societal relationship between athletics and sex, that makes sense to me. [NEWLINE] [NEWLINE] [STARTQ] "Standards" are consistently changed in a manner that makes military life "easier" for women than men. [ENDQ] [NEWLINE] Let's just not have different standards, then.  Yes, fewer women will qualify than men, but obviously these standards are in place to ensure that new recruits won't be a danger to themselves or others. [NEWLINE] [NEWLINE] [STARTQ] Pregnancies pull women away from duty. [ENDQ] [NEWLINE] I think that it would be reasonable to enact a form of mandatory birth control.  Enforcing it is a more complex issue, but nowhere near impossible.  There are long-term birth control methods that involve the [implantation of a tiny slow-release hormone capsule under the skin]( [URL] ) that prevents ovulation.  If I recall correctly, most hormonal birth control methods will also show up in urine tests (and certainly in blood tests), so that's a relatively simple enforcement tool.  In the case of the subdural implant, however, you likely wouldn't need that since they aren't self-administered. [NEWLINE] [NEWLINE] Yes, there is a small chance of failure, but I don't think that it would be enough to effect unit cohesion any more significantly than the various ailments that men can suffer from. [NEWLINE] [NEWLINE] [STARTQ] Women can join the civilian sector equivalents. [ENDQ] [NEWLINE] Many women don't want to serve in support or civilian roles, and I'm sure you know plenty that feel this way.  Ultimately, it's about choice, and the availability of alternatives that are not the same doesn't mean that we should exclude an entire gender. [NEWLINE] [NEWLINE] As an ending note, I basically agree with you that the last three points you mentioned aren't sufficient.  While I think that
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0702, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you're going to say the cause of the [horrible event] was ________, you can't stop at the rapist/burglar/attacker/murderer. You have to look at<mask> they acted,<mask>.<mask> 'd the burglar rob the house? Maybe<mask> they're hard into the poverty trap and they couldn't figure out any other way to survive without taking a hit in another major area of their life.<mask> did the rapist rape that girl? Maybe<mask> he has a tumor surpressing that part of his brain responsible for impulse control and emotion regulation.<mask> did that husband murder his wife? Maybe she abused him for 10 years or cheated on him. [NEWLINE] [NEWLINE] And from there you have to keep going.<mask> are there people in poverty,<mask> caused that person to be a sociopath,<mask> caused the person to become<mask> selfish and violent and not see any other way to handle the problem? Etc.... [NEWLINE] [NEWLINE] Human behavior is very complex,<mask> in the end we all act to ease our suffering and gain reward.<mask> people are generally in control of their behavior and the choice to attack is in the hands of the assailant, something in their brain is telling them this is the right move and something the victim did put them in the equation [NEWLINE] [NEWLINE] [NEWLINE] Sometimes you're right, the event was absolutely caused by the selfishness of an assailant who enjoys other peoples pain or having power over someone else to hurt them, or who cares more about getting<mask> they want than the effect it has on the other person,<mask> that's a very small percentage of crimes. And the victims aren't random. The reason someone becomes a victim could be based on their choices or circumstances out of their control (like who else was around, or they were wearing a color the attacker hates)<mask> its important to look at<mask> put them in the position to be hurt in the first place<mask> well<mask><mask> put the other person in the position of attacker. That's basic survival, learning to avoid danger through mistakes, both your own mistakes and other peoples mistakes. Understanding the mistakes of the victim is about preventing becoming a victim yourself, understanding the assailants motives is about preventing others from becoming victims<mask> well. And<mask> we dont try to understand then society is drastically less likely to prevent it from happening again to anyone anywhere and an individual is less likely to avoid danger themselves. [NEWLINE] [NEWLINE] There is no one single reason anything happens, even<mask> it looks like someone hurt someone else to get something they wanted and didn't need to survive
Label encoding: <s>If you're going to say the cause of the [horrible event] was ________, you can't stop at the rapist/burglar/attacker/murderer. You have to look at why they acted, also. Why 'd the burglar rob the house? Maybe because they're hard into the poverty trap and they couldn't figure out any other way to survive without taking a hit in another major area of their life. Why did the rapist rape that girl? Maybe because he has a tumor surpressing that part of his brain responsible for impulse control and emotion regulation. Why did that husband murder his wife? Maybe she abused him for 10 years or cheated on him. [NEWLINE] [NEWLINE] And from there you have to keep going. Why are there people in poverty, what caused that person to be a sociopath, what caused the person to become so selfish and violent and not see any other way to handle the problem? Etc.... [NEWLINE] [NEWLINE] Human behavior is very complex, but in the end we all act to ease our suffering and gain reward. Although people are generally in control of their behavior and the choice to attack is in the hands of the assailant, something in their brain is telling them this is the right move and something the victim did put them in the equation [NEWLINE] [NEWLINE] [NEWLINE] Sometimes you're right, the event was absolutely caused by the selfishness of an assailant who enjoys other peoples pain or having power over someone else to hurt them, or who cares more about getting what they want than the effect it has on the other person, but that's a very small percentage of crimes. And the victims aren't random. The reason someone becomes a victim could be based on their choices or circumstances out of their control (like who else was around, or they were wearing a color the attacker hates) but its important to look at what put them in the position to be hurt in the first place as well as what put the other person in the position of attacker. That's basic survival, learning to avoid danger through mistakes, both your own mistakes and other peoples mistakes. Understanding the mistakes of the victim is about preventing becoming a victim yourself, understanding the assailants motives is about preventing others from becoming victims as well. And if we dont try to understand then society is drastically less likely to prevent it from happening again to anyone anywhere and an individual is less likely to avoid danger themselves. [NEWLINE] [NEWLINE] There is no one single reason anything happens, even if it looks like someone hurt someone else to get something they wanted and didn't need to survive
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0711, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're talking about different things here. [NEWLINE] [NEWLINE] First, road rules need to be followed not<mask> they're right or wrong<mask><mask> of safety. <mask> you're going 60 in a 30 zone, people *expect* you to be traveling at 30<mask> you're not,<mask> you're putting people at risk. <mask> you cross<mask> there's no crosswalk, cars don't know that they'll have to stop for you,<mask> you're creating an unsafe condition.  Violating these road rules puts people in danger specifically<mask> people need to be able to predict<mask> you'll do in order to move about safely. <mask> you're driving at 30 in a 60 zone, that's<mask> dangerous!  In many places there are posted *minimum* speed limits<mask> well.  It's all about the flow of traffic. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, marijuana isn't that. <mask> you break the law by possessing or even selling marijuana, you're not putting anyone at risk by virtue of breaking the law.  See, driving at 60 miles per hour or crossing the street are not crimes;<mask> is a violation is going much faster than the flow of traffic or crossing<mask> cars don't expect it. <mask> for marijuana, the crime is just owning or using some.  There are no victims. <mask> is the point of holding a person prisoner in such a case<mask> the original violation is no longer a crime?  It's a waste of resources, and there's no real deterrent<mask> the person's original crime is now legal. [NEWLINE] [NEWLINE] You shouldn't confuse criminal laws with regulations.  Regulations are simply the rules of the game;<mask> you break these regulations, you're cheating, which is bad in itself: doing X is unfair<mask> build your business around not doing X.  Criminal laws refer to crimes that are bad in themselves: doing Y is bad<mask> Y is a crime. <mask> society repeals both the regulation against X and the law against Y, then doing X is no longer unfair<mask> you can change your business practices, and Y is no longer bad. <mask> you did X before the repeal, you were cheating then and gaining an unfair advantage. <mask> you did Y before the repeal, you were committing an act that was considered bad,<mask> that same act is no longer considered bad.  Your punishment for doing X was for cheating; your punishment for doing Y was<mask> Y was bad. <mask> X is no longer cheating, well, that's OK<mask><mask> you did
Label encoding: <s>You're talking about different things here. [NEWLINE] [NEWLINE] First, road rules need to be followed not because they're right or wrong but because of safety.  If you're going 60 in a 30 zone, people *expect* you to be traveling at 30 but you're not, so you're putting people at risk.  If you cross where there's no crosswalk, cars don't know that they'll have to stop for you, so you're creating an unsafe condition.  Violating these road rules puts people in danger specifically because people need to be able to predict what you'll do in order to move about safely.  If you're driving at 30 in a 60 zone, that's also dangerous!  In many places there are posted *minimum* speed limits as well.  It's all about the flow of traffic. [NEWLINE] [NEWLINE] On the other hand, marijuana isn't that.  If you break the law by possessing or even selling marijuana, you're not putting anyone at risk by virtue of breaking the law.  See, driving at 60 miles per hour or crossing the street are not crimes; what is a violation is going much faster than the flow of traffic or crossing where cars don't expect it.  But for marijuana, the crime is just owning or using some.  There are no victims.  What is the point of holding a person prisoner in such a case if the original violation is no longer a crime?  It's a waste of resources, and there's no real deterrent because the person's original crime is now legal. [NEWLINE] [NEWLINE] You shouldn't confuse criminal laws with regulations.  Regulations are simply the rules of the game; if you break these regulations, you're cheating, which is bad in itself: doing X is unfair so build your business around not doing X.  Criminal laws refer to crimes that are bad in themselves: doing Y is bad therefore Y is a crime.  If society repeals both the regulation against X and the law against Y, then doing X is no longer unfair so you can change your business practices, and Y is no longer bad.  If you did X before the repeal, you were cheating then and gaining an unfair advantage.  If you did Y before the repeal, you were committing an act that was considered bad, but that same act is no longer considered bad.  Your punishment for doing X was for cheating; your punishment for doing Y was because Y was bad.  If X is no longer cheating, well, that's OK because when you did
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are two points that you are leaning on that need to be address: [NEWLINE] [NEWLINE] 1. You are equating popularity to quality.<mask> you state "the numbers don't lie" you are saying that this is a popular product and<mask> a quality product. In today's world most mass produced products are never good in a qualitative sense, unless the goal of the product is to be qualitatively good on a mass produced scale. Mass production typically doesn't lend itself to "good." The franchises that are mass produced (Call of Duty, Assassins Creed, sports franchises, etc) are made cheaply and don't show much innovations between iterations. You need to look throughout a franchise to see major changes, and even then typically you don't see much differentiation. In the broad sense of quality videogames this is usually a bad sign, which is<mask> OP contending the popular choice<mask> he see's the flaws of the mass produced product<mask> out weighing the benefits. [NEWLINE] [NEWLINE] 2. You, and typically a capitalist society<mask> a whole, think that a good product is a product that makes money.<mask> you look at Call of Duty<mask> a product to make money than it is an amazing product.<mask>, through the history of art--and lets be honest here, video games can be art--most products of creativity that push the boundaries of their medium and try to capture an essence that we consider art typically don't make money. This is the problem of AAA video games:<mask> can you justify spending millions of dollars on an artistic/"good" product that may not make back its money? I guess it comes down to not expecting AAA games to produce something other than<mask> is going to hit the biggest target audience, and<mask> they do hit the right balance between originality/artistic free expression and money making--like The Last of Us did-- than all the better. [NEWLINE] [NEWLINE] <mask>, I have to disagree with your point about OP's view on '"overrated and overhyped" being nothing more than an idiosyncratic reflection of his own personal threshold for formulaic videogames'<mask><mask> we start restricting personal experiences for a medium that lives in personal experiences then we might<mask> well not be talking about videogames. By looking at CoD<mask> an experience and then grading that experience by objectively looking at the delivery method then comparing it back onto the franchise you will get a experience that hasn't changed in a decade, hell McDonald's  has changed more than CoD in that time. To those people who are tired of that experience than yes
Label encoding: <s>There are two points that you are leaning on that need to be address: [NEWLINE] [NEWLINE] 1. You are equating popularity to quality. When you state "the numbers don't lie" you are saying that this is a popular product and therefore a quality product. In today's world most mass produced products are never good in a qualitative sense, unless the goal of the product is to be qualitatively good on a mass produced scale. Mass production typically doesn't lend itself to "good." The franchises that are mass produced (Call of Duty, Assassins Creed, sports franchises, etc) are made cheaply and don't show much innovations between iterations. You need to look throughout a franchise to see major changes, and even then typically you don't see much differentiation. In the broad sense of quality videogames this is usually a bad sign, which is why OP contending the popular choice because he see's the flaws of the mass produced product as out weighing the benefits. [NEWLINE] [NEWLINE] 2. You, and typically a capitalist society as a whole, think that a good product is a product that makes money. If you look at Call of Duty as a product to make money than it is an amazing product. However, through the history of art--and lets be honest here, video games can be art--most products of creativity that push the boundaries of their medium and try to capture an essence that we consider art typically don't make money. This is the problem of AAA video games: How can you justify spending millions of dollars on an artistic/"good" product that may not make back its money? I guess it comes down to not expecting AAA games to produce something other than what is going to hit the biggest target audience, and when they do hit the right balance between originality/artistic free expression and money making--like The Last of Us did-- than all the better. [NEWLINE] [NEWLINE] Also, I have to disagree with your point about OP's view on '"overrated and overhyped" being nothing more than an idiosyncratic reflection of his own personal threshold for formulaic videogames' because if we start restricting personal experiences for a medium that lives in personal experiences then we might as well not be talking about videogames. By looking at CoD as an experience and then grading that experience by objectively looking at the delivery method then comparing it back onto the franchise you will get a experience that hasn't changed in a decade, hell McDonald's  has changed more than CoD in that time. To those people who are tired of that experience than yes
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] By the same token, does Mathematics not produce any new knowledge? [ENDQ] [NEWLINE] This is tricky. I am not sure about this one. By my argument, no,<mask> there are no experiments in mathematics.<mask>, I am on the fence with this one,<mask> historically there do seem to be discoveries in mathematics. I can't really say the same for philosophy. [NEWLINE] [NEWLINE] [STARTQ] Wat? Philosophers are confined to defining things intuitively? [ENDQ] [NEWLINE] Let me explain myself,<mask> I wasn't clear. Let's say we have a set of objects A,B, and C. Philosophers will attempt to construct a theory that is consistent with these objects.<mask> a new object D is introduced, then they will change their theory to incorporate D.<mask>, they never are the ones to produce D. Physicists will do the same,<mask> they<mask> attempt to introduce new items into the set. They will be the ones to introduce D. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask>? Is there some scale of "<mask> much has this done to revolutionize our understanding of truth" that I'm not aware of? [ENDQ] [NEWLINE] Well determinism, for an example, is completely dead<mask> a theory thanks to quantum mechanics. Absolute space and time is gone due to relativity. We know that everything is probabilistic rather than defined, and systems always go towards a disordered state. We have some idea of the shape, size, and origins of the universe. We know that we cannot separate observation from reality. We know that math is particularly good at describing the universe. We know the insignificance of our position<mask> a planet. Physics has largely changed our ideas about causation and action at a distance. Spirituality<mask> an explanatory tool is becoming smaller and smaller<mask> physics fills in the gaps of our knowledge. [NEWLINE] [NEWLINE] This is not huge to you? [NEWLINE] [NEWLINE] [STARTQ] At best, this is a criticism of certain branches of Philosophy,<mask> not decision theory, logic etc. [ENDQ] Further, it's genuinely difficult to mathematize certain concepts, it sucks,<mask> that doesn't make those concepts any less true or false, being able to reduce them to mathematics is an epistemic, not ontological concern. Does the fact that your statement in the CMV title is not easily formalizable make it any less true? [NEWLINE] [NEWLINE] I appreciate the irony that I am using philosophy to try and prove a statement that undermines the usefulness of philosophy.<mask><mask> that physics would not be equipped to answer questions such<mask> these. My view is somewhat changed in that I
Label encoding: <s> [STARTQ] By the same token, does Mathematics not produce any new knowledge? [ENDQ] [NEWLINE] This is tricky. I am not sure about this one. By my argument, no, since there are no experiments in mathematics. However, I am on the fence with this one, since historically there do seem to be discoveries in mathematics. I can't really say the same for philosophy. [NEWLINE] [NEWLINE] [STARTQ] Wat? Philosophers are confined to defining things intuitively? [ENDQ] [NEWLINE] Let me explain myself, because I wasn't clear. Let's say we have a set of objects A,B, and C. Philosophers will attempt to construct a theory that is consistent with these objects. If a new object D is introduced, then they will change their theory to incorporate D. However, they never are the ones to produce D. Physicists will do the same, but they also attempt to introduce new items into the set. They will be the ones to introduce D. [NEWLINE] [NEWLINE] [STARTQ] How so? Is there some scale of " how much has this done to revolutionize our understanding of truth" that I'm not aware of? [ENDQ] [NEWLINE] Well determinism, for an example, is completely dead as a theory thanks to quantum mechanics. Absolute space and time is gone due to relativity. We know that everything is probabilistic rather than defined, and systems always go towards a disordered state. We have some idea of the shape, size, and origins of the universe. We know that we cannot separate observation from reality. We know that math is particularly good at describing the universe. We know the insignificance of our position as a planet. Physics has largely changed our ideas about causation and action at a distance. Spirituality as an explanatory tool is becoming smaller and smaller as physics fills in the gaps of our knowledge. [NEWLINE] [NEWLINE] This is not huge to you? [NEWLINE] [NEWLINE] [STARTQ] At best, this is a criticism of certain branches of Philosophy, but not decision theory, logic etc. [ENDQ] Further, it's genuinely difficult to mathematize certain concepts, it sucks, but that doesn't make those concepts any less true or false, being able to reduce them to mathematics is an epistemic, not ontological concern. Does the fact that your statement in the CMV title is not easily formalizable make it any less true? [NEWLINE] [NEWLINE] I appreciate the irony that I am using philosophy to try and prove a statement that undermines the usefulness of philosophy. I agree that physics would not be equipped to answer questions such as these. My view is somewhat changed in that I
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As your many questions indicate, depression is not a simple phenomenon. Even just by virtue of being a function of your brain, makes it a byproduct of an extremely complicated and little-understood system. [NEWLINE] [NEWLINE] **Depression is a disease.** It doesn't have that much to do with<mask> good your life is or<mask> wealthy you are. It is literally a disease of the human brain.<mask> I were born with, or developed asthma (a disease of the lungs) it would not make sense to say "these people I know with it [asthma] have very fortunate lives, and<mask> that, continue to have trouble breathing."<mask> fortunate of a life you have does not directly translate into<mask> your brain, or lungs, or heart, naturally function. [NEWLINE] [NEWLINE] [STARTQ] Depressive illnesses are disorders of the brain. Brain-imaging technologies, such<mask> magnetic resonance imaging (MRI), have shown that the brains of people who have depression look different than those of people without depression. The parts of the brain involved in mood, thinking, sleep, appetite, and behavior appear different.[1] [ENDQ] [NEWLINE] Just<mask> having a fortunate life does not make you immune to asthma, or heart disease, or cancer; having a fortunate life does not make you immune to depression or other diseases of the brain. [NEWLINE] [NEWLINE] It's okay to not understand it very well,<mask> even medical experts do not understand the brain and the disease very well.<mask> we *do* know that it is a disease, and not a fleeting feeling or dramatics. [NEWLINE] [NEWLINE] Going back to asthma, just<mask> you can breathe *some* of the time, doesn't mean that you can breathe healthily *all* of the time. Depression doesn't mean you can never experience happy thoughts or genuine joy,<mask> it is a disease that can cause the brain and body to exhibit some of the symptoms you mentioned (wasting time, struggling to smile, etc). [NEWLINE] [NEWLINE] I highly recommend you read Allie Brosh's depiction of depression; it is very real and illustrates the disease in a compelling and accessible way. Hopefully it will help you understand<mask> depression is like: [NEWLINE] [NEWLINE] [Part One]( [URL]?updated-max=2013-10-02T14:53:00-06:00&amp;max-results=10) [NEWLINE] [NEWLINE] [Part Two]( [URL]?updated-max=2013-10-02T14:53:00-06:00&amp;max-results=10) [NEWLINE] [NEWLINE] [1]
Label encoding: <s>As your many questions indicate, depression is not a simple phenomenon. Even just by virtue of being a function of your brain, makes it a byproduct of an extremely complicated and little-understood system. [NEWLINE] [NEWLINE] **Depression is a disease.** It doesn't have that much to do with how good your life is or how wealthy you are. It is literally a disease of the human brain. If I were born with, or developed asthma (a disease of the lungs) it would not make sense to say "these people I know with it [asthma] have very fortunate lives, and despite that, continue to have trouble breathing." How fortunate of a life you have does not directly translate into how your brain, or lungs, or heart, naturally function. [NEWLINE] [NEWLINE] [STARTQ] Depressive illnesses are disorders of the brain. Brain-imaging technologies, such as magnetic resonance imaging (MRI), have shown that the brains of people who have depression look different than those of people without depression. The parts of the brain involved in mood, thinking, sleep, appetite, and behavior appear different.[1] [ENDQ] [NEWLINE] Just as having a fortunate life does not make you immune to asthma, or heart disease, or cancer; having a fortunate life does not make you immune to depression or other diseases of the brain. [NEWLINE] [NEWLINE] It's okay to not understand it very well, since even medical experts do not understand the brain and the disease very well. But we *do* know that it is a disease, and not a fleeting feeling or dramatics. [NEWLINE] [NEWLINE] Going back to asthma, just because you can breathe *some* of the time, doesn't mean that you can breathe healthily *all* of the time. Depression doesn't mean you can never experience happy thoughts or genuine joy, but it is a disease that can cause the brain and body to exhibit some of the symptoms you mentioned (wasting time, struggling to smile, etc). [NEWLINE] [NEWLINE] I highly recommend you read Allie Brosh's depiction of depression; it is very real and illustrates the disease in a compelling and accessible way. Hopefully it will help you understand what depression is like: [NEWLINE] [NEWLINE] [Part One]( [URL]?updated-max=2013-10-02T14:53:00-06:00&amp;max-results=10) [NEWLINE] [NEWLINE] [Part Two]( [URL]?updated-max=2013-10-02T14:53:00-06:00&amp;max-results=10) [NEWLINE] [NEWLINE] [1]
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ Thank you for your carefully-written and thorough response! [NEWLINE] [NEWLINE] I have to concede about rape culture. My past objections to it were mainly that it seemed to put violence into a hierarchy<mask> only certain victims were taken seriously and that the ad campaigns created<mask> many problems<mask> they did fix them. [NEWLINE] [NEWLINE] To address your second counter-point, my example of expressions about suicide were meant to be applied to the families and friends of people who committed suicide. They still are around to hear insensitive phrases about "blowing your brains out," which are used much more freely, I'd say, albeit in my limited experience, than rape jokes. [NEWLINE] [NEWLINE] Generally, there's an inconsistency in the way we decide which terms and expressions are perpetuating violence and which ones are just unfortunate or casual jokes. [NEWLINE] [NEWLINE] I'd go further and say that these expressions aren't unique to rape or suicide<mask> that we use light-hearted language to talk about all sorts of violence. We talk about "kicking his ass," or "punching her in the face," without really making the same considerations we do<mask> we talk about rape culture or trivializing rape. [NEWLINE] [NEWLINE] This inconsistency leads me to wonder whether we're perpetuating violence by using such expressions. I doubt it only<mask> I don't think there's any correlation between the rate of violence in a city and whether these expressions are used. That's<mask> I had the same doubt about rape jokes and sexually violent expressions in video games. I do realize that language has an impact on us,<mask>. I'm just not sure that putting these expressions out of use will help end such violence.<mask><mask> the fact that they're used is a symptom that,<mask> it may exacerbate the problem, isn't necessarily used with that intention. And<mask> they are used, they serve<mask> an important theme for dialogue about violence and the place it holds in society. Let me know<mask> this premise seems flawed. [NEWLINE] [NEWLINE] I do greatly appreciate the links you've attached. The last one, especially, was an amazing thread and lucidly written. Thank you! [NEWLINE] [NEWLINE] I'd<mask> like to say I'm grateful for forums like these and for people like yourself who respond in a cogent, non-aggresive manner. I do realize that these problems don't exist in a vacuum and people who are personally touched by them feel violated and hostile<mask> they have to explain something they feel to be true. They have the right to feel the way they do<mask> it's always nice to be debated with respect
Label encoding: <s>∆ Thank you for your carefully-written and thorough response! [NEWLINE] [NEWLINE] I have to concede about rape culture. My past objections to it were mainly that it seemed to put violence into a hierarchy where only certain victims were taken seriously and that the ad campaigns created as many problems as they did fix them. [NEWLINE] [NEWLINE] To address your second counter-point, my example of expressions about suicide were meant to be applied to the families and friends of people who committed suicide. They still are around to hear insensitive phrases about "blowing your brains out," which are used much more freely, I'd say, albeit in my limited experience, than rape jokes. [NEWLINE] [NEWLINE] Generally, there's an inconsistency in the way we decide which terms and expressions are perpetuating violence and which ones are just unfortunate or casual jokes. [NEWLINE] [NEWLINE] I'd go further and say that these expressions aren't unique to rape or suicide but that we use light-hearted language to talk about all sorts of violence. We talk about "kicking his ass," or "punching her in the face," without really making the same considerations we do when we talk about rape culture or trivializing rape. [NEWLINE] [NEWLINE] This inconsistency leads me to wonder whether we're perpetuating violence by using such expressions. I doubt it only because I don't think there's any correlation between the rate of violence in a city and whether these expressions are used. That's why I had the same doubt about rape jokes and sexually violent expressions in video games. I do realize that language has an impact on us, however. I'm just not sure that putting these expressions out of use will help end such violence. I think the fact that they're used is a symptom that, though it may exacerbate the problem, isn't necessarily used with that intention. And while they are used, they serve as an important theme for dialogue about violence and the place it holds in society. Let me know if this premise seems flawed. [NEWLINE] [NEWLINE] I do greatly appreciate the links you've attached. The last one, especially, was an amazing thread and lucidly written. Thank you! [NEWLINE] [NEWLINE] I'd also like to say I'm grateful for forums like these and for people like yourself who respond in a cogent, non-aggresive manner. I do realize that these problems don't exist in a vacuum and people who are personally touched by them feel violated and hostile when they have to explain something they feel to be true. They have the right to feel the way they do but it's always nice to be debated with respect
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me address this by pointing one thing out: numbers don't lie. With absolutely anything, there will always be things that people that aren't a fan of. Make major changes and instead of Person A hating the series and Person B loving it, now Person A loves it and Person B hates it. You're entirely in your right to feel that CoD is overrated, and personally it seems futile to change this view. You like<mask> you like. I can only provide objective points to maybe calibrate your perspective. [NEWLINE] [NEWLINE] <mask> back to numbers don't lie. Consider the goal the creators of CoD have. To make money. And they consistently perform this goal well year after year. Now, you're an executive of Activision-Blizzard.<mask> would you tell the developers to do a complete 180 on the series, "for change's sake."? You only run risk and very little reward.<mask> you stick to the tried and true model they've had, you're very nearly guaranteed to make a certain amount of sales. I'm sure they can graph the sales of each subsequent game and make predictions to the gradual decline with each passing year<mask> more and more individuals grow bored of the CoD formula. [NEWLINE] [NEWLINE] Really, your view of "overrated and overhyped" is nothing more than an idiosyncratic reflection of your own *personal* threshold for formulaic video games.<mask> this view is separate from the objective success of the series.<mask> they keep selling games, it means people still enjoy the series for<mask> it is. The idiom, "Don't fix<mask> isn't broken" comes to mind. There are more people who still like the series and will buy new iterations without much change compared to the minority who do find it cliche, hackneyed, etc. I mean, be realistic here,<mask> they promised the most radically new CoD next year, would you buy it? Probably not. You have a bias against the series.<mask> would their target audience be the people who *don't* like CoD? It makes no sense. [NEWLINE] [NEWLINE] The other major point is that video game development is extremely expensive. Making an entirely new engine is costly, takes years to develop, and you never have a guarantee of financial success upon release. It makes more sense to stick with the old engine and old code, which drastically reduces the time (and<mask> cost) it takes to make a new game. Brand new engine CoDs likely wouldn't make sufficiently more money to offset their cost to make it
Label encoding: <s>Let me address this by pointing one thing out: numbers don't lie. With absolutely anything, there will always be things that people that aren't a fan of. Make major changes and instead of Person A hating the series and Person B loving it, now Person A loves it and Person B hates it. You're entirely in your right to feel that CoD is overrated, and personally it seems futile to change this view. You like what you like. I can only provide objective points to maybe calibrate your perspective. [NEWLINE] [NEWLINE] So back to numbers don't lie. Consider the goal the creators of CoD have. To make money. And they consistently perform this goal well year after year. Now, you're an executive of Activision-Blizzard. Why would you tell the developers to do a complete 180 on the series, "for change's sake."? You only run risk and very little reward. If you stick to the tried and true model they've had, you're very nearly guaranteed to make a certain amount of sales. I'm sure they can graph the sales of each subsequent game and make predictions to the gradual decline with each passing year as more and more individuals grow bored of the CoD formula. [NEWLINE] [NEWLINE] Really, your view of "overrated and overhyped" is nothing more than an idiosyncratic reflection of your own *personal* threshold for formulaic video games. But this view is separate from the objective success of the series. If they keep selling games, it means people still enjoy the series for what it is. The idiom, "Don't fix what isn't broken" comes to mind. There are more people who still like the series and will buy new iterations without much change compared to the minority who do find it cliche, hackneyed, etc. I mean, be realistic here, if they promised the most radically new CoD next year, would you buy it? Probably not. You have a bias against the series. Why would their target audience be the people who *don't* like CoD? It makes no sense. [NEWLINE] [NEWLINE] The other major point is that video game development is extremely expensive. Making an entirely new engine is costly, takes years to develop, and you never have a guarantee of financial success upon release. It makes more sense to stick with the old engine and old code, which drastically reduces the time (and therefore cost) it takes to make a new game. Brand new engine CoDs likely wouldn't make sufficiently more money to offset their cost to make it
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OP all your main points, to me, do not summarize the negative impact of porn on society,<mask> rather are all the reasons<mask><mask> much porn is bad or at least only possessing a very limited amount of entertainment value;<mask> these points you laid out are not inherent qualities of porn<mask> rather<mask> the market appears to bear. [NEWLINE] [NEWLINE] The question that I have is '<mask> does the market bear it?' and '<mask> isn't there at least more variety?' and by that I mean '<mask> is it<mask> hard to find erotic videos that suit my adult tastes<mask> I can look at nips and peen and fucking?' [NEWLINE] [NEWLINE] The fact is that most mainstream porn is about<mask> sophisticated<mask> a fourteen year old boy's wank fantasy and<mask><mask> that this is the intentional design.<mask><mask> the porn producers who are successful today are very good at perceiving that most of their market is still made up of people who are a bit sexually prudish by any global standard. That means porn is not designed to be a high value product that a person would collect and cherish over time,<mask> something that someone would pick up on an impulse and discard. [NEWLINE] [NEWLINE] Impulse products usually have a few distinguishing features 1) relatively cheap to produce, 2) rather high markup in comparison to the cost of production, 3) over promise 4)<mask> do not fully satisfy.<mask> you've watched enough porn you can see that it does share all these characteristics. The funny thing about these impulse products is that consumers will buy them over an over looking for that satisfaction that is never fully realized. They are habit forming. In such a sexually repressed nation, it's the only way those poor girls can make a living. [NEWLINE] [NEWLINE] And that's porn in a nutshell. [NEWLINE] [NEWLINE] The fact is that there's a whole bunch of people in America, collectively, who don't want erotic entertainment for anyone and they're part of the reason porn is *<mask> bad* and has to be in this impulse buy format. On one side you have the anti-porn warriors who believe that porn is the cause of all of America's sexual problems rather than a reflection of it. On the other side you have the censoring assholes who like to go after artists who even hint at any form of nudity, eroticism, or sexuality in their work. And then you have the general population of moralists whose sole purpose for existing seems to be to make people feel bad about having normal sexual impulses. [NEWLINE] [NEWLINE] <mask>, porn has to be in a format that
Label encoding: <s>OP all your main points, to me, do not summarize the negative impact of porn on society, but rather are all the reasons why so much porn is bad or at least only possessing a very limited amount of entertainment value; since these points you laid out are not inherent qualities of porn but rather what the market appears to bear. [NEWLINE] [NEWLINE] The question that I have is'why does the market bear it?' and'why isn't there at least more variety?' and by that I mean'why is it so hard to find erotic videos that suit my adult tastes where I can look at nips and peen and fucking?' [NEWLINE] [NEWLINE] The fact is that most mainstream porn is about as sophisticated as a fourteen year old boy's wank fantasy and I think that this is the intentional design. I think the porn producers who are successful today are very good at perceiving that most of their market is still made up of people who are a bit sexually prudish by any global standard. That means porn is not designed to be a high value product that a person would collect and cherish over time, but something that someone would pick up on an impulse and discard. [NEWLINE] [NEWLINE] Impulse products usually have a few distinguishing features 1) relatively cheap to produce, 2) rather high markup in comparison to the cost of production, 3) over promise 4) but do not fully satisfy. If you've watched enough porn you can see that it does share all these characteristics. The funny thing about these impulse products is that consumers will buy them over an over looking for that satisfaction that is never fully realized. They are habit forming. In such a sexually repressed nation, it's the only way those poor girls can make a living. [NEWLINE] [NEWLINE] And that's porn in a nutshell. [NEWLINE] [NEWLINE] The fact is that there's a whole bunch of people in America, collectively, who don't want erotic entertainment for anyone and they're part of the reason porn is * so bad* and has to be in this impulse buy format. On one side you have the anti-porn warriors who believe that porn is the cause of all of America's sexual problems rather than a reflection of it. On the other side you have the censoring assholes who like to go after artists who even hint at any form of nudity, eroticism, or sexuality in their work. And then you have the general population of moralists whose sole purpose for existing seems to be to make people feel bad about having normal sexual impulses. [NEWLINE] [NEWLINE] Therefore, porn has to be in a format that
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe people pursue these degrees partly for the reason you listed: they have an unwarranted fear of mathematics and "hard" sciences. I firmly believe that most people are capable of understanding a topic and pursuing a career in virtually in field<mask><mask><mask> they apply themselves appropriately. The amount of time devoted to studying and learning the topic will vary between individuals,<mask> it is possible nonetheless. [ENDQ] [NEWLINE] I want to specifically concentrate on the "virtually any (I'm assuming you meant any, not in) field<mask><mask><mask> they apply themselves" part. [NEWLINE] [NEWLINE] I'm a psychology major<mask> I LOVE psychology. Love it<mask> much that I read textbooks, journal articles and write research papers in my own time<mask> it *fascinates* me. [NEWLINE] [NEWLINE] I've<mask> had the wonderful opportunity to learn from some truly gifted professors, who happen to be clinical psychologists.<mask> I have learned, and refutes your point, is that it takes a certain sort of person to succeed in that field. Frankly, not everyone is truly capable of deep empathy, nor can most people easily alter their perspective to view the world from someone else point of view. [NEWLINE] [NEWLINE] Psychology<mask> a major, and even more<mask> applicable once you get into graduate degrees, requires a person to be very open-minded and mentally flexible. Math and science are based on hard facts. 2 + 2 WILL = 4. To succeed, I must continue to beat the numbers in until they stick.<mask> in psychology.... you can't 'teach' the ability to shift perspective. You can teach theory all day long, and knowing it will not make you any more effective in your field unless you can *understand* people. [NEWLINE] [NEWLINE] For instance: (insert whatever truly hated person/group you can think of). For my example I will use the man who kidnapped those women in Cleveland. I want to understand him. I want to see the world through his eyes until I can see<mask> this made sense. I want to think like he does until I can imagine "yes I could beat that one woman's face in. It makes sense." I want to empathize with his way of thinking<mask> I can understand<mask> makes him tick. To do<mask>, I must shift my perspective to<mask> I see his deeds<mask> natural or right (<mask> that is<mask> he sees them). I must remove my own compulsion to see him<mask> evil, and shift my world to<mask> he is *normal*. Only then can I see<mask> he did<mask> he did. [NEWLINE]
Label encoding: <s> [STARTQ] I believe people pursue these degrees partly for the reason you listed: they have an unwarranted fear of mathematics and "hard" sciences. I firmly believe that most people are capable of understanding a topic and pursuing a career in virtually in field as long as they apply themselves appropriately. The amount of time devoted to studying and learning the topic will vary between individuals, but it is possible nonetheless. [ENDQ] [NEWLINE] I want to specifically concentrate on the "virtually any (I'm assuming you meant any, not in) field as long as they apply themselves" part. [NEWLINE] [NEWLINE] I'm a psychology major because I LOVE psychology. Love it so much that I read textbooks, journal articles and write research papers in my own time because it *fascinates* me. [NEWLINE] [NEWLINE] I've also had the wonderful opportunity to learn from some truly gifted professors, who happen to be clinical psychologists. What I have learned, and refutes your point, is that it takes a certain sort of person to succeed in that field. Frankly, not everyone is truly capable of deep empathy, nor can most people easily alter their perspective to view the world from someone else point of view. [NEWLINE] [NEWLINE] Psychology as a major, and even more so applicable once you get into graduate degrees, requires a person to be very open-minded and mentally flexible. Math and science are based on hard facts. 2 + 2 WILL = 4. To succeed, I must continue to beat the numbers in until they stick. But in psychology.... you can't 'teach' the ability to shift perspective. You can teach theory all day long, and knowing it will not make you any more effective in your field unless you can *understand* people. [NEWLINE] [NEWLINE] For instance: (insert whatever truly hated person/group you can think of). For my example I will use the man who kidnapped those women in Cleveland. I want to understand him. I want to see the world through his eyes until I can see why this made sense. I want to think like he does until I can imagine "yes I could beat that one woman's face in. It makes sense." I want to empathize with his way of thinking so I can understand what makes him tick. To do so, I must shift my perspective to where I see his deeds as natural or right ( because that is how he sees them). I must remove my own compulsion to see him as evil, and shift my world to where he is *normal*. Only then can I see why he did what he did. [NEWLINE]
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> it comes down to one's ability to experience joy/happiness/contentment/fulfillment. It is these feelings that give life meaning--<mask> they are innately pleasurable and enjoyable in the broadest sense of the terms.<mask> you are able to feel these things with consistency it is painfully clear<mask> killing yourself is an illogical act. I consider,<mask> feeling joy, for it to be wholly self-evident that such an experience has value, and<mask> to deprive the universe of such value would be illogical. [NEWLINE] [NEWLINE] I think you have probably noted already that your situation is just the opposite-- you are experiencing much more pain and suffering than you are joy, and<mask> it would be almost *ill*ogical for you to bestow such an (anti)value on the universe. [NEWLINE] [NEWLINE] <mask>,<mask><mask> you (<mask> you do,<mask><mask>, agree with the above paragraph) would be wrong.<mask> *providing* the universe with value is *better* than simply beginning to have no effect (i.e. through suicide, or having never existed in the first place).<mask><mask> you would accept my logic,<mask> of course the line of argument all hinges upon whether or not you believe that such experiences like joy, happiness, and fulfillment actually provide value. And,<mask> I said from the top I believe it is self-evident, it is an axiom that I am willing to accept based on my experience with these emotions-- In that moment of complete satisfaction or contentment, I just would not accept any argument that tried to convince me that these feelings don't have inherent value. [NEWLINE] [NEWLINE] To me (and I am by no means an expert),<mask> depression is is an inability, or at least a barrier to, feeling these positive emotions.<mask><mask> once you get out of this (and I truly believe it is possible,<mask> I would be lying<mask> I said I knew<mask> ), it will become self-evident to you<mask> well that emotions have inherent value. And<mask> you can see this, you can see<mask> it is illogical to think that life is not worth living-- you would be depriving the universe of this value that could potentially be contributed by you.<mask><mask>,<mask><mask> you need to spend<mask> much mental energy<mask> you can to try and experience things that may bring about these kinds of emotions (save up--travel the world?). This is<mask> you will be able to accept<mask> those who enjoy being alive accept-- that being happy is good
Label encoding: <s>I think it comes down to one's ability to experience joy/happiness/contentment/fulfillment. It is these feelings that give life meaning-- because they are innately pleasurable and enjoyable in the broadest sense of the terms. If you are able to feel these things with consistency it is painfully clear why killing yourself is an illogical act. I consider, when feeling joy, for it to be wholly self-evident that such an experience has value, and so to deprive the universe of such value would be illogical. [NEWLINE] [NEWLINE] I think you have probably noted already that your situation is just the opposite-- you are experiencing much more pain and suffering than you are joy, and so it would be almost *ill*ogical for you to bestow such an (anti)value on the universe. [NEWLINE] [NEWLINE] BUT, I think you ( if you do, in fact, agree with the above paragraph) would be wrong. Because *providing* the universe with value is *better* than simply beginning to have no effect (i.e. through suicide, or having never existed in the first place). I think you would accept my logic, but of course the line of argument all hinges upon whether or not you believe that such experiences like joy, happiness, and fulfillment actually provide value. And, as I said from the top I believe it is self-evident, it is an axiom that I am willing to accept based on my experience with these emotions-- In that moment of complete satisfaction or contentment, I just would not accept any argument that tried to convince me that these feelings don't have inherent value. [NEWLINE] [NEWLINE] To me (and I am by no means an expert), what depression is is an inability, or at least a barrier to, feeling these positive emotions. I think once you get out of this (and I truly believe it is possible, although I would be lying if I said I knew how ), it will become self-evident to you as well that emotions have inherent value. And if you can see this, you can see why it is illogical to think that life is not worth living-- you would be depriving the universe of this value that could potentially be contributed by you. IMHO, I think you need to spend as much mental energy as you can to try and experience things that may bring about these kinds of emotions (save up--travel the world?). This is how you will be able to accept what those who enjoy being alive accept-- that being happy is good
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You seem to be minimizing the situation. [ENDQ] [NEWLINE] I apologize for not making this more apparent,<mask> the olive garden example applies specifically to implicit contracts. [NEWLINE] [NEWLINE] The government owns all of the land, at best you can borrow it. The government is in the business of selling residency/citizenship. Everything else is a fringe benefit, those are the real products.<mask> you don't have one or both of those you can't really do much on government land. Which, remember, is all of the land. The whole country. [NEWLINE] [NEWLINE] <mask> you disagree with that, you have a couple of options. You can call for a redistribution of property based on<mask> you think is more fair, or you can make the case that<mask> someone has property you need to survive you have de facto ownership of that property. There isn't really a third option here. You either accept current property distribution or you want some variety of redistribution based on<mask> you find fair. [NEWLINE] [NEWLINE] [STARTQ] require!= consent. [ENDQ] [NEWLINE] Again I didn't make my point clear,<mask> gimme a sec and I will try again. Conditional requirement can totally be consent. You are required to pay<mask> part of an exchange of goods and services.<mask> you want something someone is selling,<mask> you aren't paying for it you are stealing it. [NEWLINE] [NEWLINE] [STARTQ] You don't pay to live in your country, you pay to use the benefits provided by the government. [ENDQ] [NEWLINE] This is completely incorrect, and<mask>'s worse it's a red herring. You do pay for residency. Even<mask> you don't use anything else you are using residency. And it's a red herring<mask> it doesn't actually matter. Yes, explicit consent is always better than implicit consent,<mask> that probably isn't your core issue with the government.<mask> selling residency were blatant and explicit, you would likely still have problems with our societies' use of force, with government corruption and overeach, with crony capitalism and all of the other legitimate problems that libertarians are almost unique in pointing out in mainstream politics. I know I'd still be pissed about those.<mask> instead of those actual issues internet libertarians get continually diverted by the fact that "they didn't sign any social contract."<mask> the fact that<mask> they were presented with an actual contract they would either sign it or engage in hefty mental gymnastics to explain<mask>  *this time* property rights are open to negotiation. It's a useless red herring that only serves to make your legitimate points easier to dismiss and it
Label encoding: <s> [STARTQ] You seem to be minimizing the situation. [ENDQ] [NEWLINE] I apologize for not making this more apparent, but the olive garden example applies specifically to implicit contracts. [NEWLINE] [NEWLINE] The government owns all of the land, at best you can borrow it. The government is in the business of selling residency/citizenship. Everything else is a fringe benefit, those are the real products. If you don't have one or both of those you can't really do much on government land. Which, remember, is all of the land. The whole country. [NEWLINE] [NEWLINE] If you disagree with that, you have a couple of options. You can call for a redistribution of property based on what you think is more fair, or you can make the case that if someone has property you need to survive you have de facto ownership of that property. There isn't really a third option here. You either accept current property distribution or you want some variety of redistribution based on what you find fair. [NEWLINE] [NEWLINE] [STARTQ] require!= consent. [ENDQ] [NEWLINE] Again I didn't make my point clear, but gimme a sec and I will try again. Conditional requirement can totally be consent. You are required to pay as part of an exchange of goods and services. If you want something someone is selling, if you aren't paying for it you are stealing it. [NEWLINE] [NEWLINE] [STARTQ] You don't pay to live in your country, you pay to use the benefits provided by the government. [ENDQ] [NEWLINE] This is completely incorrect, and what's worse it's a red herring. You do pay for residency. Even if you don't use anything else you are using residency. And it's a red herring because it doesn't actually matter. Yes, explicit consent is always better than implicit consent, but that probably isn't your core issue with the government. If selling residency were blatant and explicit, you would likely still have problems with our societies' use of force, with government corruption and overeach, with crony capitalism and all of the other legitimate problems that libertarians are almost unique in pointing out in mainstream politics. I know I'd still be pissed about those. But instead of those actual issues internet libertarians get continually diverted by the fact that "they didn't sign any social contract." Despite the fact that if they were presented with an actual contract they would either sign it or engage in hefty mental gymnastics to explain how  *this time* property rights are open to negotiation. It's a useless red herring that only serves to make your legitimate points easier to dismiss and it
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Before I continue, I would like to start by saying that I am not a pedophile and that I am strongly against the sexual abuse of children. [NEWLINE] [NEWLINE] My view is based on scientific evidence which strongly suggests that allowing pedophiles to view child porn will result in fewer instances of sexual abuse. [NEWLINE] [NEWLINE] Now I understand and accept that some people will say that children who feature in child porn are re-victimised<mask> they learn that people are viewing images of their abuse,<mask> by legalising viewing, this would mean that the victims will never learn that images of their abuse are being viewed and<mask> they will not suffer re-victimisation. [NEWLINE] [NEWLINE] There is no evidence to suggest that pedophiles who view child porn are more likely than not to go onto abuse children.<mask><mask>, the research suggests the exact opposite. It suggests that pedophiles who have access to child porn are more likely than not to stick with child porn to relieve their desires rather than abuse a child. [NEWLINE] [NEWLINE] My view is based on a study titled ["Legalizing child pornography is linked to lower rates of child sex abuse"]( [URL] +springer/media/springer+select?SGWID=0-11001-6-1042321-0). [NEWLINE] [NEWLINE] Child abuse is abhorrant,<mask><mask><mask> to ignore this research is irresponsible. [NEWLINE] [NEWLINE] I will not be convinced by anecdotal evidence on this by the way. Please make sure you provide reliable sources for your opinions, like I have. [NEWLINE] [NEWLINE] CMV. [NEWLINE] [NEWLINE] Edit: Someone has suggested that I make it clear that I am only talking about the viewing of child porn to be legal and not the production or distribution of images. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Before I continue, I would like to start by saying that I am not a pedophile and that I am strongly against the sexual abuse of children. [NEWLINE] [NEWLINE] My view is based on scientific evidence which strongly suggests that allowing pedophiles to view child porn will result in fewer instances of sexual abuse. [NEWLINE] [NEWLINE] Now I understand and accept that some people will say that children who feature in child porn are re-victimised when they learn that people are viewing images of their abuse, but by legalising viewing, this would mean that the victims will never learn that images of their abuse are being viewed and so they will not suffer re-victimisation. [NEWLINE] [NEWLINE] There is no evidence to suggest that pedophiles who view child porn are more likely than not to go onto abuse children. In fact, the research suggests the exact opposite. It suggests that pedophiles who have access to child porn are more likely than not to stick with child porn to relieve their desires rather than abuse a child. [NEWLINE] [NEWLINE] My view is based on a study titled ["Legalizing child pornography is linked to lower rates of child sex abuse"]( [URL] +springer/media/springer+select?SGWID=0-11001-6-1042321-0). [NEWLINE] [NEWLINE] Child abuse is abhorrant, but I think to ignore this research is irresponsible. [NEWLINE] [NEWLINE] I will not be convinced by anecdotal evidence on this by the way. Please make sure you provide reliable sources for your opinions, like I have. [NEWLINE] [NEWLINE] CMV. [NEWLINE] [NEWLINE] Edit: Someone has suggested that I make it clear that I am only talking about the viewing of child porn to be legal and not the production or distribution of images. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Jealousy" is another name for "selfishness".<mask> you think about binary relationships<mask> standard, it's<mask> you are following the notion that society tells you that they are. Open relationships have evolved from people thinking outside of the box of the "standard relationship". [NEWLINE] [NEWLINE] In pretty much every relationship, there are standards that hold true<mask><mask> the terms of the relationship. [NEWLINE] [NEWLINE] - You want to be happy [NEWLINE] - You want them to be happy [NEWLINE] - They want you to be happy [NEWLINE] - They are happy with you [NEWLINE] [NEWLINE] Now, think about a couple that is sitting together in a park. One of the two sees someone walking by and gets distracted by their attractiveness. After seeing this, the second of the pair gets upset. "<mask> are you looking at that person? that's not me." Now<mask> you think about<mask> it means to truly fulfil the points above; a person shouldn't have to feel guilty for being themselves,<mask> they would ideally be perfect the way they are to the other person. This creates a disharmony of happiness vs. honesty to oneself. It's a small thing,<mask> it is the foundation of the mindset for open relationships that other people have pointed out. [NEWLINE] [NEWLINE] A couple that I knew saw their partner's sexuality<mask> their own possession. The idea in monogamous relationships is that your sexuality belongs to your partner exclusively. This couple I knew would have other people come in and be intimate with them  and the dynamic was fascinating for me. Person A would see Person B with somebody else and actually get *giddy* about it. "Ahhh look at<mask> much fun they're having!"<mask> you stop holding your partner's sexuality<mask> your own possession, it opens them up to share the things that make them happy with you. In the relationship of the people I mentioned, they were always passionate towards each other and their eyes lit up<mask> they would tell stories of the people they were with. [NEWLINE] [NEWLINE] Now you made a comment elsewhere about someone getting pregnant. More often than not, that would be an issue in an open relationship. You typically want to take measures to make sure things like that won't happen with someone that you aren't planning to commit to. Open relationships require a very strong foundation of trust, respect, and responsibility in<mask> you deal with each other<mask> well<mask> other people. It may not be for everyone,<mask> for some people it relieves a lot of guilt in not being able to truly express themselves and be themselves with their partner.</s>
Label encoding: <s>"Jealousy" is another name for "selfishness". If you think about binary relationships as standard, it's because you are following the notion that society tells you that they are. Open relationships have evolved from people thinking outside of the box of the "standard relationship". [NEWLINE] [NEWLINE] In pretty much every relationship, there are standards that hold true regardless of the terms of the relationship. [NEWLINE] [NEWLINE] - You want to be happy [NEWLINE] - You want them to be happy [NEWLINE] - They want you to be happy [NEWLINE] - They are happy with you [NEWLINE] [NEWLINE] Now, think about a couple that is sitting together in a park. One of the two sees someone walking by and gets distracted by their attractiveness. After seeing this, the second of the pair gets upset. " why are you looking at that person? that's not me." Now if you think about what it means to truly fulfil the points above; a person shouldn't have to feel guilty for being themselves, because they would ideally be perfect the way they are to the other person. This creates a disharmony of happiness vs. honesty to oneself. It's a small thing, but it is the foundation of the mindset for open relationships that other people have pointed out. [NEWLINE] [NEWLINE] A couple that I knew saw their partner's sexuality as their own possession. The idea in monogamous relationships is that your sexuality belongs to your partner exclusively. This couple I knew would have other people come in and be intimate with them  and the dynamic was fascinating for me. Person A would see Person B with somebody else and actually get *giddy* about it. "Ahhh look at how much fun they're having!" When you stop holding your partner's sexuality as your own possession, it opens them up to share the things that make them happy with you. In the relationship of the people I mentioned, they were always passionate towards each other and their eyes lit up as they would tell stories of the people they were with. [NEWLINE] [NEWLINE] Now you made a comment elsewhere about someone getting pregnant. More often than not, that would be an issue in an open relationship. You typically want to take measures to make sure things like that won't happen with someone that you aren't planning to commit to. Open relationships require a very strong foundation of trust, respect, and responsibility in how you deal with each other as well as other people. It may not be for everyone, but for some people it relieves a lot of guilt in not being able to truly express themselves and be themselves with their partner.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Yes, he is vengeful, you can see that easily,<mask> surely he would not let people die of diseases like Ebola. Did he give us a choice there? No, not really. [ENDQ] [NEWLINE] Well he did. He has given us the knowledge and capacity to study and come up with treatments and cures for it. We just haven't gotten there<mask>. Its not a choice of<mask> we want it or not,<mask> a choice to actively combat it. You have to remember that the world runs<mask> a system. Ebola is generated through mutations that randomly happen and is already part of our world. We humans have a defined natural immune system that has already been set and determined. God<mask> given us the capabilities to add to the immune system to fight it off.<mask> does there need to be divine intervention to save a few people(<mask> people naturally die anyways)?? [NEWLINE] [NEWLINE] [STARTQ] God could very easily stop a war<mask> he chooses not to. He may have given the 2 leaders in the war a choice<mask> did he give the civilians in the street? No, the civilians in the street can do nothing to stop the war,<mask> the leaders mind is set on war, it will happen. 2 civilians in the street can't change a dictators mind that is set on war. [ENDQ] [NEWLINE] He has given every individual a choice. Again the world is a system. 2 civilians cant do anything.<mask> a majority of the population can<mask> they demand it. Leadership only exist given<mask> the citizens allow. obviously 2 people isnt enough to overthrow the government,<mask> there is a point<mask> there is enough people to. [NEWLINE] [NEWLINE] [STARTQ] We try to combat these diseases,<mask> we need better tech etc etc<mask> people still die from them.<mask> God is all loving<mask> he says he is, he would not allow the diseases to kill<mask> many. He would make them happen<mask> wouldn't let the disease kill<mask> many people. [ENDQ] [NEWLINE] <mask> not? He did promise us an afterlife. Maybe its just all part of the process. heaven is suppose to be better<mask><mask> not let a lot of people go there? [NEWLINE] [NEWLINE] Basically im trying to say that god has already set up a world or system that runs a certain way. miracles that prevents wars and cures people instantly is an immediate action of divine intervention.<mask> he did<mask>, he would show himself regularly to the world. God may work in mysterious ways,<mask> he possibly has intervened in the most discrete way to cause an end to wars and cure for disease. </s>
Label encoding: <s> [STARTQ] Yes, he is vengeful, you can see that easily, but surely he would not let people die of diseases like Ebola. Did he give us a choice there? No, not really. [ENDQ] [NEWLINE] Well he did. He has given us the knowledge and capacity to study and come up with treatments and cures for it. We just haven't gotten there yet. Its not a choice of if we want it or not, but a choice to actively combat it. You have to remember that the world runs as a system. Ebola is generated through mutations that randomly happen and is already part of our world. We humans have a defined natural immune system that has already been set and determined. God as given us the capabilities to add to the immune system to fight it off. Why does there need to be divine intervention to save a few people( because people naturally die anyways)?? [NEWLINE] [NEWLINE] [STARTQ] God could very easily stop a war but he chooses not to. He may have given the 2 leaders in the war a choice but did he give the civilians in the street? No, the civilians in the street can do nothing to stop the war, if the leaders mind is set on war, it will happen. 2 civilians in the street can't change a dictators mind that is set on war. [ENDQ] [NEWLINE] He has given every individual a choice. Again the world is a system. 2 civilians cant do anything. But a majority of the population can if they demand it. Leadership only exist given what the citizens allow. obviously 2 people isnt enough to overthrow the government, but there is a point where there is enough people to. [NEWLINE] [NEWLINE] [STARTQ] We try to combat these diseases, therefore we need better tech etc etc but people still die from them. If God is all loving as he says he is, he would not allow the diseases to kill as many. He would make them happen but wouldn't let the disease kill so many people. [ENDQ] [NEWLINE] why not? He did promise us an afterlife. Maybe its just all part of the process. heaven is suppose to be better so why not let a lot of people go there? [NEWLINE] [NEWLINE] Basically im trying to say that god has already set up a world or system that runs a certain way. miracles that prevents wars and cures people instantly is an immediate action of divine intervention. If he did so, he would show himself regularly to the world. God may work in mysterious ways, but he possibly has intervened in the most discrete way to cause an end to wars and cure for disease. </s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Those people are going against your supposition to begin with.<mask> they abandon their families it's<mask> they think they're going to find more fulfillment outside of them,<mask> they shouldn't be used in your argument against me. [ENDQ] [NEWLINE] Yes which is<mask> I'm looking for people that have obtained extrinsic goals who have done that (<mask> controlling for other issues). [NEWLINE] [NEWLINE] [STARTQ] Your entire premise says that. You propose that those who have children are more fulfilled and<mask> such it is saying that those who choose not to have children have made poorer choices than those who made the other decision. [ENDQ] [NEWLINE] That is not the same<mask> saying that people that don't have children are lesser people.  More fulfilled, does not mean "a better person."  It just mean's more fulfilled. [NEWLINE] [NEWLINE] [STARTQ] And<mask> they still don't know<mask> it would be like to live out the rest of their lives not having children. You're still excluding an entire half of the people you encompass in your original view by saying that their views can't apply. You can't say someone feels less fulfillment than you<mask> they don't have children. Just<mask> you can't say that they have more. It's completely different circumstances. [ENDQ] [NEWLINE] It's only about 15-20% of the population actually (over their lifetime). <mask> I'm excluding 15-20% of the population.  This just happens to be a sample group that is different<mask> of the average age on this site. [NEWLINE] [NEWLINE] <mask><mask> the sex analogy is pretty good<mask> you want to play the analogy game.  Someone who hasn't had sex can't understand<mask> it fully entails.  And<mask> they can't compare sex to going to a great movie. [NEWLINE] [NEWLINE] On paper they can debate it and postulate,<mask> their opinion is going to be fairly ridiculous. <mask> all of the small details are<mask> make sex great, and someone can't understand that until they've had at least a couple good romps. [NEWLINE] [NEWLINE] <mask> I guess<mask> I'm saying is I don't want a thread full of people who have never had sex telling me that "look at the facts, you go to a movie, it's $5, it's an hour long of a totally positive emotional and visual experience.  I've checked rotten tomatoes, it's 95%. <mask> I had sex, it could go wrong and really people have sex thousands of times<mask> only see maybe a hundred movies!" [NEWLINE] [NEWLINE] That's<mask> these arguments would be.</s>
Label encoding: <s> [STARTQ] Those people are going against your supposition to begin with. If they abandon their families it's because they think they're going to find more fulfillment outside of them, therefore they shouldn't be used in your argument against me. [ENDQ] [NEWLINE] Yes which is why I'm looking for people that have obtained extrinsic goals who have done that ( while controlling for other issues). [NEWLINE] [NEWLINE] [STARTQ] Your entire premise says that. You propose that those who have children are more fulfilled and as such it is saying that those who choose not to have children have made poorer choices than those who made the other decision. [ENDQ] [NEWLINE] That is not the same as saying that people that don't have children are lesser people.  More fulfilled, does not mean "a better person."  It just mean's more fulfilled. [NEWLINE] [NEWLINE] [STARTQ] And yet they still don't know how it would be like to live out the rest of their lives not having children. You're still excluding an entire half of the people you encompass in your original view by saying that their views can't apply. You can't say someone feels less fulfillment than you because they don't have children. Just as you can't say that they have more. It's completely different circumstances. [ENDQ] [NEWLINE] It's only about 15-20% of the population actually (over their lifetime).  So I'm excluding 15-20% of the population.  This just happens to be a sample group that is different because of the average age on this site. [NEWLINE] [NEWLINE] I think the sex analogy is pretty good if you want to play the analogy game.  Someone who hasn't had sex can't understand what it fully entails.  And thus they can't compare sex to going to a great movie. [NEWLINE] [NEWLINE] On paper they can debate it and postulate, but their opinion is going to be fairly ridiculous.  Because all of the small details are what make sex great, and someone can't understand that until they've had at least a couple good romps. [NEWLINE] [NEWLINE] So I guess what I'm saying is I don't want a thread full of people who have never had sex telling me that "look at the facts, you go to a movie, it's $5, it's an hour long of a totally positive emotional and visual experience.  I've checked rotten tomatoes, it's 95%.  If I had sex, it could go wrong and really people have sex thousands of times but only see maybe a hundred movies!" [NEWLINE] [NEWLINE] That's what these arguments would be.</s>
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Gonna start by saying im a former volunteer first responder. I highly disagree with<mask> youve said here. Theres a reason that not all places have a paid full time fire department and its mostly<mask> its unfeasible to do<mask>. It would cost a lot of money and serve little to no purpose.<mask> i was on my local department we would get maybe 10 calls a month total for anything. Thats not only fires, thats cats stuck in trees kind of crap, sick people wanting a ambulance ride or just needing something we have (diabetics and hypos were pretty common) and once in a<mask> a death that basically needed to be verified. For all of that kind of thing its once in 20+ calls that we actually dealt with a fire of any kind. [NEWLINE] [NEWLINE] My department did pay,<mask> we were considered volunteer<mask> we only took home about 3$ a hour after firemen association dues and various funds we paid into<mask> we could have community events and crap like that. [NEWLINE] [NEWLINE] <mask> we did<mask> your proposing<mask> and basically having all fire fighters be paid full time workers it wouldnt be long before you run into the same problems police forces run into with budgets and then the fire department would have to start charging for everything which would likely turn into tickets of some sort due to lack of payment for services and eventually people wont call the fire department during emergencies<mask> they cant afford it and dont want to get screwed with either jail time or financially just<mask> they had a low blood sugar or fell down and hurt themselves. [NEWLINE] [NEWLINE] It would be bad for society<mask> the fire department didnt respond to things and charged you to do<mask>. I myself have used the services they provide at least 50 times in my life and<mask> i had to pay for all of those or got ticketed for it or whatever id be<mask> royally screwed at a time<mask> im likely already not doing to well such<mask><mask> i had a heart attack two years ago, or<mask> the roof burned off my house<mask> a teenager due to a downed power line, or all the low blood sugars<mask> i wasnt able to help myself and my wife called them and they gave me some shit to bring my blood sugar back up and get me responsive again, id likely risk my life more by not calling<mask> fast<mask> i should<mask> there was a financial cost to doing<mask> and your proposal would lead to that similar to<mask> police camp all over ticketing people for such stupid minor crimes, cause they have to be funded somehow.</s>
Label encoding: <s>Gonna start by saying im a former volunteer first responder. I highly disagree with what youve said here. Theres a reason that not all places have a paid full time fire department and its mostly because its unfeasible to do so. It would cost a lot of money and serve little to no purpose. When i was on my local department we would get maybe 10 calls a month total for anything. Thats not only fires, thats cats stuck in trees kind of crap, sick people wanting a ambulance ride or just needing something we have (diabetics and hypos were pretty common) and once in a while a death that basically needed to be verified. For all of that kind of thing its once in 20+ calls that we actually dealt with a fire of any kind. [NEWLINE] [NEWLINE] My department did pay, but we were considered volunteer because we only took home about 3$ a hour after firemen association dues and various funds we paid into so we could have community events and crap like that. [NEWLINE] [NEWLINE] If we did what your proposing though and basically having all fire fighters be paid full time workers it wouldnt be long before you run into the same problems police forces run into with budgets and then the fire department would have to start charging for everything which would likely turn into tickets of some sort due to lack of payment for services and eventually people wont call the fire department during emergencies because they cant afford it and dont want to get screwed with either jail time or financially just because they had a low blood sugar or fell down and hurt themselves. [NEWLINE] [NEWLINE] It would be bad for society if the fire department didnt respond to things and charged you to do so. I myself have used the services they provide at least 50 times in my life and if i had to pay for all of those or got ticketed for it or whatever id be so royally screwed at a time when im likely already not doing to well such as when i had a heart attack two years ago, or when the roof burned off my house as a teenager due to a downed power line, or all the low blood sugars where i wasnt able to help myself and my wife called them and they gave me some shit to bring my blood sugar back up and get me responsive again, id likely risk my life more by not calling as fast as i should if there was a financial cost to doing so and your proposal would lead to that similar to how police camp all over ticketing people for such stupid minor crimes, cause they have to be funded somehow.</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&gt; That is not the purpose of academia, it is a place to learn things or to discover new things, not to learn<mask> to think (whatever it means to think correctly.) My view is that it is not worth it for the public to be paying to make progress in these fields. [NEWLINE] [NEWLINE] Aha... and here is the heart of the matter. [NEWLINE] [NEWLINE] I suppose this opinion is the inevitable eventuality for someone who views education primarily<mask> a means of consuming and retaining facts.  The problem, put simply, is that knowing everything in any given field only allows for<mask> much room for progress - and<mask><mask> can often be detrimental<mask> up against a problem that defies known parameters.  The easiest example is probably the theory of special relativity; plenty of learned physicists were nearly stopped in their tracks by the inherent contradictions brought about by the aether and the constancy of time. To finally resolve those inconsistencies, it took someone not only knowledgeable,<mask> curious enough to treat them<mask><mask> they were - assumptions, not facts - and throw them out the window. [NEWLINE] [NEWLINE] It is precisely this kind of thought which /u/yamsx1 refers to<mask> "thinking correctly" - the ability to approach, examine, and solve arbitrary problems not just logically,<mask><mask> creatively. <mask> to think, not just<mask> to know.  This is the skill of raw thought, and should be treated<mask> mathematics is - not only<mask> a worthy pursuit unto itself,<mask><mask> a fundamental building block of most other fields.  Given this, we can properly understand and appreciate fields like theology and philosophy for<mask> they are: logical and creative thought<mask> directly applied to God and society, respectively. [NEWLINE] [NEWLINE] Even<mask> "progress" in these applied fields in modern times is mostly restricted to the ivory towers, the fundamentals are and remain valuable skills for all to learn,<mask><mask> whether they are pursued for their own ends. <mask>?  For the same reason we all learn some level of mathematics; even<mask> you do not pursue a career which uses it, directly or indirectly, the fundamentals are used<mask> broadly in everyday life that it would be flatly irresponsible to *not* teach them.  It is no accident that the very first material taught to any student of philosophy is a crash course on logical relationships and fallacies. <mask> improved would, say, our political debates be<mask> everyone had learned<mask> to reason their way to the end of a simple syllogism right alongside their multiplication tables?</s>
Label encoding: <s>&gt; That is not the purpose of academia, it is a place to learn things or to discover new things, not to learn how to think (whatever it means to think correctly.) My view is that it is not worth it for the public to be paying to make progress in these fields. [NEWLINE] [NEWLINE] Aha... and here is the heart of the matter. [NEWLINE] [NEWLINE] I suppose this opinion is the inevitable eventuality for someone who views education primarily as a means of consuming and retaining facts.  The problem, put simply, is that knowing everything in any given field only allows for so much room for progress - and in fact can often be detrimental when up against a problem that defies known parameters.  The easiest example is probably the theory of special relativity; plenty of learned physicists were nearly stopped in their tracks by the inherent contradictions brought about by the aether and the constancy of time. To finally resolve those inconsistencies, it took someone not only knowledgeable, but curious enough to treat them as what they were - assumptions, not facts - and throw them out the window. [NEWLINE] [NEWLINE] It is precisely this kind of thought which /u/yamsx1 refers to as "thinking correctly" - the ability to approach, examine, and solve arbitrary problems not just logically, but also creatively.  How to think, not just what to know.  This is the skill of raw thought, and should be treated as mathematics is - not only as a worthy pursuit unto itself, but as a fundamental building block of most other fields.  Given this, we can properly understand and appreciate fields like theology and philosophy for what they are: logical and creative thought as directly applied to God and society, respectively. [NEWLINE] [NEWLINE] Even if "progress" in these applied fields in modern times is mostly restricted to the ivory towers, the fundamentals are and remain valuable skills for all to learn, regardless of whether they are pursued for their own ends.  Why?  For the same reason we all learn some level of mathematics; even if you do not pursue a career which uses it, directly or indirectly, the fundamentals are used so broadly in everyday life that it would be flatly irresponsible to *not* teach them.  It is no accident that the very first material taught to any student of philosophy is a crash course on logical relationships and fallacies.  How improved would, say, our political debates be if everyone had learned how to reason their way to the end of a simple syllogism right alongside their multiplication tables?</s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The men's rights movement has everything to do with *literally* bashing women. [NEWLINE] [NEWLINE] [STARTQ] [Equal rights and equal lefts.]( [URL] /) [ENDQ] [NEWLINE] [STARTQ] ["Equal rights, equal fights".]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [Equal rights, equal lefts.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [Equal rights equal lefts]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [Equal lefts.]( [URL] ) (This one's an r/mensrights moderator.) [ENDQ] [NEWLINE] [STARTQ] [I hope that guy knocked her out right in front of her son.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [The guy should have smacked her upside the head with his skateboard.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [I would have hit her across the head with the skateboard, flat end of course.]( [URL] ) [ENDQ] [NEWLINE] Oh, and they kinda just plain hate women too. [NEWLINE] [NEWLINE] [STARTQ] [Name and shame this fat bitch. She needs to learn to keep a leash of her little mistake and keep him off the skateboard park<mask> he doesn't fucking belong. It's her fault, not her child's.<mask> at a stupid hag. Half-wit. He should've knocked her loser, ghetto ass out. I bet she's a real peach to be around.]( [URL] ) [ENDQ] [NEWLINE] Sadly, I didn't even have to *scroll my screen* for 1/2 of these, they were all<mask> close together in just one thread. [NEWLINE] [NEWLINE] [STARTQ] [Equal rights means equal fights.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [equal rights means equal lefts... and uppercuts.]( [URL] ) [ENDQ] [NEWLINE] Okay, I'm tired of this. It's like shooting fish in a scummy violent barrel. [NEWLINE] [NEWLINE] Edit: Eh, here's more. [NEWLINE] [NEWLINE] [STARTQ] [man i never get tired of seeing that. equal rights means equal lefts.]( [URL] ) Regarding a video<mask> a burly male bus driver uppercuts a teenage girl and throws her physically off the bus. The bus driver was later fired. [ENDQ] [NEWLINE] [STARTQ] [He didn't hit hard enough.]( [URL] ) [ENDQ] [NEWLINE] Here's<mask> their penchant for hitting women ties in with their hatred of feminism: [NEWLINE] [NEWLINE] [STARTQ] [Firing the driver was] [ridiculous,<mask> expected<mask> living in a feminist society.]( [URL] ) [ENDQ] [NEWLINE] They relish disproportionate retaliation. [NEWLINE] [NEWLINE] [STARTQ] [Shoulda turned around and decked the bitch in the fucking teeth anyway.]( [URL] ) [ENDQ] </s>
Label encoding: <s>The men's rights movement has everything to do with *literally* bashing women. [NEWLINE] [NEWLINE] [STARTQ] [Equal rights and equal lefts.]( [URL] /) [ENDQ] [NEWLINE] [STARTQ] ["Equal rights, equal fights".]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [Equal rights, equal lefts.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [Equal rights equal lefts]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [Equal lefts.]( [URL] ) (This one's an r/mensrights moderator.) [ENDQ] [NEWLINE] [STARTQ] [I hope that guy knocked her out right in front of her son.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [The guy should have smacked her upside the head with his skateboard.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [I would have hit her across the head with the skateboard, flat end of course.]( [URL] ) [ENDQ] [NEWLINE] Oh, and they kinda just plain hate women too. [NEWLINE] [NEWLINE] [STARTQ] [Name and shame this fat bitch. She needs to learn to keep a leash of her little mistake and keep him off the skateboard park where he doesn't fucking belong. It's her fault, not her child's. What at a stupid hag. Half-wit. He should've knocked her loser, ghetto ass out. I bet she's a real peach to be around.]( [URL] ) [ENDQ] [NEWLINE] Sadly, I didn't even have to *scroll my screen* for 1/2 of these, they were all so close together in just one thread. [NEWLINE] [NEWLINE] [STARTQ] [Equal rights means equal fights.]( [URL] ) [ENDQ] [NEWLINE] [STARTQ] [equal rights means equal lefts... and uppercuts.]( [URL] ) [ENDQ] [NEWLINE] Okay, I'm tired of this. It's like shooting fish in a scummy violent barrel. [NEWLINE] [NEWLINE] Edit: Eh, here's more. [NEWLINE] [NEWLINE] [STARTQ] [man i never get tired of seeing that. equal rights means equal lefts.]( [URL] ) Regarding a video where a burly male bus driver uppercuts a teenage girl and throws her physically off the bus. The bus driver was later fired. [ENDQ] [NEWLINE] [STARTQ] [He didn't hit hard enough.]( [URL] ) [ENDQ] [NEWLINE] Here's how their penchant for hitting women ties in with their hatred of feminism: [NEWLINE] [NEWLINE] [STARTQ] [Firing the driver was] [ridiculous, but expected when living in a feminist society.]( [URL] ) [ENDQ] [NEWLINE] They relish disproportionate retaliation. [NEWLINE] [NEWLINE] [STARTQ] [Shoulda turned around and decked the bitch in the fucking teeth anyway.]( [URL] ) [ENDQ] </s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You know, this used to be my position on it, which made my relationship with my best friend - a trans man - tense at times, which sucked. [NEWLINE] [NEWLINE] <mask><mask><mask> ultimately changed my mind is that<mask> I thought about it in terms of mental health,<mask> I kept coming back to is the idea that you should treat the patient in the way that causes the least harm and does the most good. [NEWLINE] [NEWLINE] Well,<mask> someone's mental model of their body doesn't agree with the physical reality of their body, it can cause serious harm to their wellbeing over time. And<mask> you<mask> a psychiatrist sit down with someone like that and try to figure out the best way forward, you basically have two ways of attacking the issue:<mask> mind and body don't agree, you can try to change the mind, or you can try to change the body. [NEWLINE] [NEWLINE] Right now, medical science is in a place<mask> we can do a lot with surgery that gets more precise and less invasive every day. (I swear, hysterectomies can be done<mask> outpatient procedures now! Wow!)<mask> turn around and take a look at<mask> much we know about<mask> in the brain determines a person's mental model of their sex, and it's<mask> much worse it's staggering. There is no brain surgery we can do to flip someone's "brain-sex" from one to the other, and the only drugs that will start to do that are hormones that will change the body, anyway!<mask><mask> does that leave? Years and years of analysis and "talking it out" on a therapist's couch? [NEWLINE] [NEWLINE] Good luck. With that. The success rate is terrible, and even<mask> it works, it takes years of time spent suffering with something that's deeply disturbing for the patient. (And, seriously, I saw a friend go through it and it's no joke. I can't *relate* to<mask> it feels to be transgender,<mask> I believe that it's real, and serious.) [NEWLINE] [NEWLINE] And that's<mask>,<mask><mask><mask><mask> with you that being transgender is a mental health problem, (and I expect that in the future, we'll see more subtle treatments for it) I strongly believe that the protocol we have of hormone therapy and gender reassignment surgery and the associated treatments is currently the best thing for them. [NEWLINE] [NEWLINE] And, hey, it's<mask> they want to live their lives,<mask> that works out, too. Cheers!</s>
Label encoding: <s>You know, this used to be my position on it, which made my relationship with my best friend - a trans man - tense at times, which sucked. [NEWLINE] [NEWLINE] I think what ultimately changed my mind is that when I thought about it in terms of mental health, what I kept coming back to is the idea that you should treat the patient in the way that causes the least harm and does the most good. [NEWLINE] [NEWLINE] Well, when someone's mental model of their body doesn't agree with the physical reality of their body, it can cause serious harm to their wellbeing over time. And when you as a psychiatrist sit down with someone like that and try to figure out the best way forward, you basically have two ways of attacking the issue: if mind and body don't agree, you can try to change the mind, or you can try to change the body. [NEWLINE] [NEWLINE] Right now, medical science is in a place where we can do a lot with surgery that gets more precise and less invasive every day. (I swear, hysterectomies can be done as outpatient procedures now! Wow!) But turn around and take a look at how much we know about what in the brain determines a person's mental model of their sex, and it's so much worse it's staggering. There is no brain surgery we can do to flip someone's "brain-sex" from one to the other, and the only drugs that will start to do that are hormones that will change the body, anyway! So what does that leave? Years and years of analysis and "talking it out" on a therapist's couch? [NEWLINE] [NEWLINE] Good luck. With that. The success rate is terrible, and even if it works, it takes years of time spent suffering with something that's deeply disturbing for the patient. (And, seriously, I saw a friend go through it and it's no joke. I can't *relate* to how it feels to be transgender, but I believe that it's real, and serious.) [NEWLINE] [NEWLINE] And that's why, even though I agree with you that being transgender is a mental health problem, (and I expect that in the future, we'll see more subtle treatments for it) I strongly believe that the protocol we have of hormone therapy and gender reassignment surgery and the associated treatments is currently the best thing for them. [NEWLINE] [NEWLINE] And, hey, it's how they want to live their lives, so that works out, too. Cheers!</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OK,<mask> we are in agreement that the social contract is not a contract<mask> my consent isn't necessary for it to be binding.  Again, I'm not saying that the social contract is not a valid principle in and of itself,<mask> that its status<mask> not-a-contract renders any attempt to remove<mask> is considered my property<mask> "theft". [NEWLINE] [NEWLINE] <mask> now we come to the question of property ownership. <mask> property is, fundamentally, is an agreement<mask> to the status of a thing. <mask> you and<mask><mask> that this spoon is your property and that fork is my property,<mask> we are in effect saying is that we agree that you have a monopoly of rights over the use of the spoon, and I have a monopoly of rights over the use of the fork - or more colloquially the spoon is assigned the status of "yours" and the fork the status of "mine".  Property need not be legally defined, it is simply a voluntary agreement between two or more parties. [NEWLINE] [NEWLINE] <mask> I can prove I own something<mask> soon<mask> it is recognized<mask> mine communally, by society at large. <mask> my employer or my client transfers a fork to me in exchange for my property or services, the fork is then recognized<mask> mine.  It is not loaned to me, it is not in escrow, it is not in a trust, it is not anything except legitimately my property. [NEWLINE] [NEWLINE] <mask>, at the end of the tax year, I keep my fork and do not hand it over to the government.  Eventually, I get a letter from the government that says that the fork which was my property now belongs to society at large.  My fork went from the status of being "mine"<mask><mask> society at large, to the status of "not mine", and will be removed from me coercively, outside of any valid contract. [NEWLINE] [NEWLINE] <mask> this transfer process is theft<mask><mask> my definition of the term. <mask> theft is something other than the coercive transfer of property outside of a contract, then the process may be something else. [NEWLINE] [NEWLINE] It's the same thing<mask> the government claims immanent domain.  The only way that act can be considered something other than theft, is<mask> the definition of "theft" is based on legality.  Of course of the definition of theft is based on legality, then theft can be anything we want it be, provided the law changes to conform with it.</s>
Label encoding: <s>OK, so we are in agreement that the social contract is not a contract since my consent isn't necessary for it to be binding.  Again, I'm not saying that the social contract is not a valid principle in and of itself, but that its status as not-a-contract renders any attempt to remove what is considered my property as "theft". [NEWLINE] [NEWLINE] But now we come to the question of property ownership.  What property is, fundamentally, is an agreement as to the status of a thing.  When you and I agree that this spoon is your property and that fork is my property, what we are in effect saying is that we agree that you have a monopoly of rights over the use of the spoon, and I have a monopoly of rights over the use of the fork - or more colloquially the spoon is assigned the status of "yours" and the fork the status of "mine".  Property need not be legally defined, it is simply a voluntary agreement between two or more parties. [NEWLINE] [NEWLINE] So I can prove I own something as soon as it is recognized as mine communally, by society at large.  When my employer or my client transfers a fork to me in exchange for my property or services, the fork is then recognized as mine.  It is not loaned to me, it is not in escrow, it is not in a trust, it is not anything except legitimately my property. [NEWLINE] [NEWLINE] So, at the end of the tax year, I keep my fork and do not hand it over to the government.  Eventually, I get a letter from the government that says that the fork which was my property now belongs to society at large.  My fork went from the status of being "mine" according to society at large, to the status of "not mine", and will be removed from me coercively, outside of any valid contract. [NEWLINE] [NEWLINE] So this transfer process is theft according to my definition of the term.  If theft is something other than the coercive transfer of property outside of a contract, then the process may be something else. [NEWLINE] [NEWLINE] It's the same thing when the government claims immanent domain.  The only way that act can be considered something other than theft, is if the definition of "theft" is based on legality.  Of course of the definition of theft is based on legality, then theft can be anything we want it be, provided the law changes to conform with it.</s>
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>These actions are not a way of taking over factory farms. They **are** destroying businesses,<mask> that makes no difference to the general public. It only affects the people who are responsible for animal cruelty. [NEWLINE] [NEWLINE] [STARTQ] You might think you're acting<mask> a liberator,<mask> to others you will be just vandals. [ENDQ] [NEWLINE] I understand this.<mask> imagine the situation from the point of view of an animal. You've spent your whole life from birth inside a cramped cage for months on end, sleeping on your own shit and you've even seen your friends die from self-mutilation after going mentally unstable. Finally, one night, your cage is removed and you are released to live in an animal sanctuary/shelter for the rest of your life.<mask> do the factory farms get authority over a cow, or pig, chicken's life? Whys should anyone get to choose<mask> another being gets to live or die? [NEWLINE] [NEWLINE] [STARTQ] <mask> for the Chinese, the issue isn't<mask> much cost it is more of meat's tie to certain dishes that have been made for thousands of years. Meat is<mask> important that China even has a strategic pork reserve. [ENDQ] [NEWLINE] This is a common fallacy of appeal to tradition. An 'ad antiquitatem'. This is the it's "been always done this way", it's traditional, it's cultural and<mask> it must be correct, attitude. This is a flawed argument. [NEWLINE] [NEWLINE] [STARTQ] You stop the influx of meat and there will be rioting in the streets. The Chinese starved generations ago. The restriction of food,<mask><mask> a cause you see<mask> noble, won't go over well. [ENDQ] [NEWLINE] [STARTQ] <mask> there were full scale meat restrictions it would be like America in the 1920's. Underground restaurants would pop up overnight. Hundreds of them. It is unrealistic to expect a culture that had had thousands of years of cultural ties with meat dishes to abandon them over night. [ENDQ] [NEWLINE] You are correct,<mask>, the ALF and other similar movements are simply for helping animals in the short term. They are not trying to, and will not stop factory farming altogether.<mask> you are referring to veganism, the idea that everyone will **one day** become vegans accepts that it will **not happen overnight** and<mask> it ever does, will happen over hundreds of years.<mask>, in the 18th century, you banned slavery, there would have been riots,<mask> the reduction of slavery happened over hundreds of years.</s>
Label encoding: <s>These actions are not a way of taking over factory farms. They **are** destroying businesses, but that makes no difference to the general public. It only affects the people who are responsible for animal cruelty. [NEWLINE] [NEWLINE] [STARTQ] You might think you're acting as a liberator, but to others you will be just vandals. [ENDQ] [NEWLINE] I understand this. But imagine the situation from the point of view of an animal. You've spent your whole life from birth inside a cramped cage for months on end, sleeping on your own shit and you've even seen your friends die from self-mutilation after going mentally unstable. Finally, one night, your cage is removed and you are released to live in an animal sanctuary/shelter for the rest of your life. Why do the factory farms get authority over a cow, or pig, chicken's life? Whys should anyone get to choose if another being gets to live or die? [NEWLINE] [NEWLINE] [STARTQ] As for the Chinese, the issue isn't so much cost it is more of meat's tie to certain dishes that have been made for thousands of years. Meat is so important that China even has a strategic pork reserve. [ENDQ] [NEWLINE] This is a common fallacy of appeal to tradition. An 'ad antiquitatem'. This is the it's "been always done this way", it's traditional, it's cultural and so it must be correct, attitude. This is a flawed argument. [NEWLINE] [NEWLINE] [STARTQ] You stop the influx of meat and there will be rioting in the streets. The Chinese starved generations ago. The restriction of food, regardless of a cause you see as noble, won't go over well. [ENDQ] [NEWLINE] [STARTQ] If there were full scale meat restrictions it would be like America in the 1920's. Underground restaurants would pop up overnight. Hundreds of them. It is unrealistic to expect a culture that had had thousands of years of cultural ties with meat dishes to abandon them over night. [ENDQ] [NEWLINE] You are correct, however, the ALF and other similar movements are simply for helping animals in the short term. They are not trying to, and will not stop factory farming altogether. If you are referring to veganism, the idea that everyone will **one day** become vegans accepts that it will **not happen overnight** and if it ever does, will happen over hundreds of years. If, in the 18th century, you banned slavery, there would have been riots, however the reduction of slavery happened over hundreds of years.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I were in the building, I wouldnt know it was true. Just like no Christian knows its true. We hear it over the PA and we believe it. [ENDQ] [NEWLINE] My point is that there is not only one source of information that points to one threat and the decision is not binary do I believe the PA or not. There are multiple ideas about the threat out there and either many people can use the PA to spread theirs or none (<mask><mask> a PA implies some sort of authority and seriousnes which doesn't work with the analogy). [NEWLINE] [NEWLINE] [STARTQ] Irrelevant. God not appearing to you doesnt make me a jerk for not being preachy. [ENDQ] [NEWLINE] That's<mask> I said I'm getting off topic. The point there was that god would be a jerk. [NEWLINE] [NEWLINE] [STARTQ] &gt; You are sure that the bomb threat is the real one... [ENDQ] [NEWLINE] [STARTQ] No Im not... [ENDQ] [NEWLINE] <mask> you are no Christian or one that just doesn't believe in the concept of heaven and hell? [NEWLINE] [NEWLINE] [STARTQ] &gt;...and all the others are made up. [ENDQ] [NEWLINE] [STARTQ] No,<mask> even<mask><mask>... [ENDQ] [NEWLINE] <mask> are you telling me that completely different teachings that exclude believers of other teachings from salvation can be true at the same time? [NEWLINE] [NEWLINE] [STARTQ]...I never said wouldnt try. I just said there are more effective ways than being obnoxious.<mask> you were in a building and everybody gets up and goes for the exit, wouldnt you ask<mask> was up? [ENDQ] [NEWLINE] You still seem to imply there is only one decision to make: follow the Christians or don't. Take the bomb threat serious or don't. That's<mask> you just assume everyone knows<mask>'s going on and just doesn't get out of the building<mask> they don't believe it at all<mask> they may believe in a fire burning in the lower floors and safety on the roof or can't decide which threat to take seriously. [NEWLINE] [NEWLINE] <mask> I were in a building and everyone runs in different directions<mask> there seem to be majorities in certain departments running in the same direction I would ask different people from different departments and check some facts<mask> I have the time before deciding one way or the other.<mask> I found out that one threat after another is just hearsay I would probably assume all are until someone shows to me one is real.<mask> yours is real and you base your belief on more than just<mask> your parents told you, you should at least try to convince others.</s>
Label encoding: <s> [STARTQ] If I were in the building, I wouldnt know it was true. Just like no Christian knows its true. We hear it over the PA and we believe it. [ENDQ] [NEWLINE] My point is that there is not only one source of information that points to one threat and the decision is not binary do I believe the PA or not. There are multiple ideas about the threat out there and either many people can use the PA to spread theirs or none ( imo a PA implies some sort of authority and seriousnes which doesn't work with the analogy). [NEWLINE] [NEWLINE] [STARTQ] Irrelevant. God not appearing to you doesnt make me a jerk for not being preachy. [ENDQ] [NEWLINE] That's why I said I'm getting off topic. The point there was that god would be a jerk. [NEWLINE] [NEWLINE] [STARTQ] &gt; You are sure that the bomb threat is the real one... [ENDQ] [NEWLINE] [STARTQ] No Im not... [ENDQ] [NEWLINE] So you are no Christian or one that just doesn't believe in the concept of heaven and hell? [NEWLINE] [NEWLINE] [STARTQ] &gt;...and all the others are made up. [ENDQ] [NEWLINE] [STARTQ] No, but even if so... [ENDQ] [NEWLINE] So are you telling me that completely different teachings that exclude believers of other teachings from salvation can be true at the same time? [NEWLINE] [NEWLINE] [STARTQ]...I never said wouldnt try. I just said there are more effective ways than being obnoxious. If you were in a building and everybody gets up and goes for the exit, wouldnt you ask what was up? [ENDQ] [NEWLINE] You still seem to imply there is only one decision to make: follow the Christians or don't. Take the bomb threat serious or don't. That's why you just assume everyone knows what's going on and just doesn't get out of the building because they don't believe it at all while they may believe in a fire burning in the lower floors and safety on the roof or can't decide which threat to take seriously. [NEWLINE] [NEWLINE] If I were in a building and everyone runs in different directions while there seem to be majorities in certain departments running in the same direction I would ask different people from different departments and check some facts if I have the time before deciding one way or the other. If I found out that one threat after another is just hearsay I would probably assume all are until someone shows to me one is real. If yours is real and you base your belief on more than just what your parents told you, you should at least try to convince others.</s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**Ignosticism**]( [URL] ): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] [STARTQ] __Ignosticism__ or __igtheism__ is the idea that every theological position assumes too much about the [concept of God]( [URL] ) and other theological concepts; including (<mask> not limited to) concepts of faith, spirituality, heaven, hell, afterlife, damnation, salvation, sin and the soul. [ENDQ] [NEWLINE] [STARTQ] Ignosticism is the view that any religious term or theological concept presented must be accompanied by a coherent definition. Without a clear definition such terms cannot be meaningfully discussed. Such terms or concepts must<mask> be [falsifiable]( [URL] ). Lacking this an ignostic takes the [theological noncognitivist]( [URL] ) position that the existence or nature of the terms presented (and all matters of debate) is meaningless. For example,<mask> the term "God" does not refer to anything reasonably defined then there is no conceivable method to test against the existence of god.<mask> the term "God" has no literal significance and need not be debated or discussed. [ENDQ] [NEWLINE] [STARTQ] Some philosophers have seen ignosticism<mask> a variation of [agnosticism]( [URL] ) or [atheism]( [URL] ), <mask> others have considered it to be distinct. [ENDQ] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Theological ^noncognitivism]( [URL] ) ^| [^Irreligion ^by ^country]( [URL] ) ^| [^Agnosticism]( [URL] ) ^| [^Atheism]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cjcqi5p) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cjcqi5p)^. ^Will ^<mask> ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Label encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**Ignosticism**]( [URL] ): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] [STARTQ] __Ignosticism__ or __igtheism__ is the idea that every theological position assumes too much about the [concept of God]( [URL] ) and other theological concepts; including ( but not limited to) concepts of faith, spirituality, heaven, hell, afterlife, damnation, salvation, sin and the soul. [ENDQ] [NEWLINE] [STARTQ] Ignosticism is the view that any religious term or theological concept presented must be accompanied by a coherent definition. Without a clear definition such terms cannot be meaningfully discussed. Such terms or concepts must also be [falsifiable]( [URL] ). Lacking this an ignostic takes the [theological noncognitivist]( [URL] ) position that the existence or nature of the terms presented (and all matters of debate) is meaningless. For example, if the term "God" does not refer to anything reasonably defined then there is no conceivable method to test against the existence of god. Therefore the term "God" has no literal significance and need not be debated or discussed. [ENDQ] [NEWLINE] [STARTQ] Some philosophers have seen ignosticism as a variation of [agnosticism]( [URL] ) or [atheism]( [URL] ),  while others have considered it to be distinct. [ENDQ] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Theological ^noncognitivism]( [URL] ) ^| [^Irreligion ^by ^country]( [URL] ) ^| [^Agnosticism]( [URL] ) ^| [^Atheism]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cjcqi5p) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cjcqi5p)^. ^Will ^ also ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That bit is the problem I feel.<mask> it comes to religiously motivated moral arguments, from the perspective of the non-religious person to say that something is 'wrong' is just your opinion, and not something I am required to share. [ENDQ] [NEWLINE] <mask><mask><mask><mask> that religion has a monopoly on this<mask>. Take something like slavery or murder. Many religious texts outline<mask> these things are ok,<mask> society has largely decided they are not. Are those just commonly held opinions?<mask> makes them right? [NEWLINE] [NEWLINE] More interestingly<mask> we are not making those decisions based on religion,<mask> are we basing those decisions on? Aren't all of our morals merely opinions? [NEWLINE] [NEWLINE] [STARTQ] Of course religious people probably think they are just doing<mask> is right and that they are trying to save people etc. [ENDQ] [NEWLINE] Much the same way that secular society creates laws and rules.<mask> makes those fundamentally right? We have a social understanding that speed limits are essentially just guidelines now,<mask><mask> did we decide that particular law can be broken?<mask> I exploit a legal loophole and benefit by not breaking any laws I am legally right,<mask> am I morally right? [NEWLINE] [NEWLINE] [STARTQ] I can ask them to stop. [ENDQ] [NEWLINE] We can ask someone who is doing something that we do not agree with or feel is harming us to stop. Who decides<mask> they should?<mask> both sides feel that they are right, is it not the opinion that has the moral majority and<mask> the law in a secular society behind it that wins?<mask> happens<mask> a paradigm shift has not reached critical mass?<mask> society has a negative collective opinion on it, does that mean it is wrong?<mask> that collective opinion has changed,<mask> the legal framework hasn't, does that make it right? [NEWLINE] [NEWLINE] [STARTQ] I'm sure they would not be equally understanding<mask> I were to go all richard dawkins and try and'save them from religion' or something<mask>. [ENDQ] [NEWLINE] Conversion is not just a religious issue. In politics, people at different ends of the political spectrum frequently disagree with each other, and often enact laws that the entire jurisdiction has to follow, even<mask> the people within that jurisdiction do not agree with them. Those differences of opinions have very real consequences on very real lives. [NEWLINE] [NEWLINE] Something being morally right or wrong is entirely based on an opinion, and the popularity of that opinion regardless to its source is<mask> dictates who agrees with you. </s>
Label encoding: <s> [STARTQ] That bit is the problem I feel. When it comes to religiously motivated moral arguments, from the perspective of the non-religious person to say that something is 'wrong' is just your opinion, and not something I am required to share. [ENDQ] [NEWLINE] I do not think that religion has a monopoly on this though. Take something like slavery or murder. Many religious texts outline when these things are ok, but society has largely decided they are not. Are those just commonly held opinions? What makes them right? [NEWLINE] [NEWLINE] More interestingly if we are not making those decisions based on religion, what are we basing those decisions on? Aren't all of our morals merely opinions? [NEWLINE] [NEWLINE] [STARTQ] Of course religious people probably think they are just doing what is right and that they are trying to save people etc. [ENDQ] [NEWLINE] Much the same way that secular society creates laws and rules. What makes those fundamentally right? We have a social understanding that speed limits are essentially just guidelines now, but why did we decide that particular law can be broken? If I exploit a legal loophole and benefit by not breaking any laws I am legally right, but am I morally right? [NEWLINE] [NEWLINE] [STARTQ] I can ask them to stop. [ENDQ] [NEWLINE] We can ask someone who is doing something that we do not agree with or feel is harming us to stop. Who decides if they should? If both sides feel that they are right, is it not the opinion that has the moral majority and therefore the law in a secular society behind it that wins? What happens when a paradigm shift has not reached critical mass? If society has a negative collective opinion on it, does that mean it is wrong? If that collective opinion has changed, but the legal framework hasn't, does that make it right? [NEWLINE] [NEWLINE] [STARTQ] I'm sure they would not be equally understanding if I were to go all richard dawkins and try and'save them from religion' or something though. [ENDQ] [NEWLINE] Conversion is not just a religious issue. In politics, people at different ends of the political spectrum frequently disagree with each other, and often enact laws that the entire jurisdiction has to follow, even if the people within that jurisdiction do not agree with them. Those differences of opinions have very real consequences on very real lives. [NEWLINE] [NEWLINE] Something being morally right or wrong is entirely based on an opinion, and the popularity of that opinion regardless to its source is what dictates who agrees with you. </s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've been using Linux<mask> 1997, and first used FreeBSD in 1995.  In that time, I've gradually been teaching myself the POSIX shell utilities for data manipulation, and I've<mask> done some database driven Web programming with PHP and My/PostgreSQL. [NEWLINE] [NEWLINE] Maybe I just haven't been working on sufficiently complex projects,<mask> I've honestly never seen a legitimate reason for XML to exist.  There honestly doesn't seem to be anything that can be done with it, that can't be done much more simply and easily with something else. <mask> I want non-relational data, I can use read or cut with a shell script, and<mask> I need relational data, I can use Postgres with PHP, or possibly even a spreadsheet. [NEWLINE] [NEWLINE] [URL].aspx - Articles like this,<mask> admittedly snarky and sarcastic, only seem to offer reinforcement for<mask><mask>. [NEWLINE] [NEWLINE] <mask><mask> there are two real reasons<mask> this bothers me<mask> much.  It's<mask> whenever I've read any attempt at XML advocacy, it has always had the following two problems. [NEWLINE] [NEWLINE] a}  There will either be an appeal made to corporate buzzwords, or some other appeal to unnecessary complexity, which can always (literally; ***always***) be proven false by practical demonstration. [NEWLINE] [NEWLINE] b}  There will be an appeal to arrogance and elitism.  "XML is awesome,<mask> the reason<mask> I can appreciate its' awesomeness and you can't, is<mask> I'm more intelligent than you, and can<mask> understand said complexity." [NEWLINE] [NEWLINE] I do not,<mask>, want to be a bigot;<mask> it was not for the above two points, XML would not bother me.  I wouldn't even think about it. <mask> someone here can provide me with a genuine demonstration, of a situation in which XML can solve a problem, which cannot be solved more easily with any other method, then I will change my view.  I am,<mask>, fairly confident that nobody will be able to do<mask>. [NEWLINE] [NEWLINE] <mask> this thread does not receive sufficient responses here, I will repost it in /r/learnprogramming,<mask> that would be more appropriate. [NEWLINE] [NEWLINE] EDIT:  I'm wishing I could edit the headline now,<mask> I realise that it was a little harsh.  Still, it's probably good for controversy, at least!</s>
Label encoding: <s>I've been using Linux since 1997, and first used FreeBSD in 1995.  In that time, I've gradually been teaching myself the POSIX shell utilities for data manipulation, and I've also done some database driven Web programming with PHP and My/PostgreSQL. [NEWLINE] [NEWLINE] Maybe I just haven't been working on sufficiently complex projects, but I've honestly never seen a legitimate reason for XML to exist.  There honestly doesn't seem to be anything that can be done with it, that can't be done much more simply and easily with something else.  If I want non-relational data, I can use read or cut with a shell script, and if I need relational data, I can use Postgres with PHP, or possibly even a spreadsheet. [NEWLINE] [NEWLINE] [URL].aspx - Articles like this, while admittedly snarky and sarcastic, only seem to offer reinforcement for my opinion. [NEWLINE] [NEWLINE] I think there are two real reasons why this bothers me so much.  It's because whenever I've read any attempt at XML advocacy, it has always had the following two problems. [NEWLINE] [NEWLINE] a}  There will either be an appeal made to corporate buzzwords, or some other appeal to unnecessary complexity, which can always (literally; ***always***) be proven false by practical demonstration. [NEWLINE] [NEWLINE] b}  There will be an appeal to arrogance and elitism.  "XML is awesome, but the reason why I can appreciate its' awesomeness and you can't, is because I'm more intelligent than you, and can therefore understand said complexity." [NEWLINE] [NEWLINE] I do not, however, want to be a bigot; if it was not for the above two points, XML would not bother me.  I wouldn't even think about it.  If someone here can provide me with a genuine demonstration, of a situation in which XML can solve a problem, which cannot be solved more easily with any other method, then I will change my view.  I am, however, fairly confident that nobody will be able to do so. [NEWLINE] [NEWLINE] If this thread does not receive sufficient responses here, I will repost it in /r/learnprogramming, if that would be more appropriate. [NEWLINE] [NEWLINE] EDIT:  I'm wishing I could edit the headline now, as I realise that it was a little harsh.  Still, it's probably good for controversy, at least!</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>? The researcher would experience the greatest possible happiness possible. In his virtual world he'd cure AIDS and every other disease and fulfill every dream he has. Maybe he'd regain the use of his leg, get married to the love of his life. Denying this happiness to him seems unjustified to me. [ENDQ] [NEWLINE] This makes the situation implausible. You think that the experience machine could be programmed<mask> accurately to the external world<mask> to facilitate scientific *discovery* inside it? [NEWLINE] [NEWLINE] No, the programming of the machine would reflect the current level of scientific understanding. You wouldn't be discovering anything about quarks in the experience machine. [NEWLINE] [NEWLINE] [STARTQ] The machine is not a 24/7 sex drugs and rocknroll party. It gives you the most valuable mental state you have.<mask> you get happiness from a life of monk-like ascetic living, you will get that in the machine. [ENDQ] [NEWLINE] Unless the machine actually shapes the person's reaction to situations<mask> well, the increase in happiness will not be<mask> large<mask> you claim. Suffering is unavoidable in any situation whether from boredom, weariness etc. <mask>,<mask> you allow the machine to dictate response to the imagined scenario, then it is essentially just a dopamine dispenser. [NEWLINE] [NEWLINE] <mask>, people value<mask> they take to be *real* connections to others/the outside world. A person who is a doctor values *actually helping people*. Giving him the experiences of helping imaginary computer programs is, in some sense, deeply wrong. His happiness is illusory in a way. [NEWLINE] [NEWLINE] [STARTQ] I'm not sure I understand this part. Plugging everyone into the machine would seem like the most moral act imaginable to me. Everyone would be experiencing the greatest possible happiness. The end of suffering.<mask> can you argue against that? [ENDQ] [NEWLINE] Plugging everyone in is different than plugging person X into the machine. Plug person X into the machine and the happiness/utility he produced for society goes away<mask> well. [NEWLINE] [NEWLINE] <mask>, remember that not everyone believes that happiness is the end goal. Rights such<mask> self-ownership trump happiness of the individual.<mask>, these people will deny your premise at the very start. No amount of post-plugging happiness can swamp the rights violation that continues<mask> the person is hooked up or being hooked up.  People have a right to choose to suffer, in effect.</s>
Label encoding: <s> [STARTQ] Why? The researcher would experience the greatest possible happiness possible. In his virtual world he'd cure AIDS and every other disease and fulfill every dream he has. Maybe he'd regain the use of his leg, get married to the love of his life. Denying this happiness to him seems unjustified to me. [ENDQ] [NEWLINE] This makes the situation implausible. You think that the experience machine could be programmed so accurately to the external world as to facilitate scientific *discovery* inside it? [NEWLINE] [NEWLINE] No, the programming of the machine would reflect the current level of scientific understanding. You wouldn't be discovering anything about quarks in the experience machine. [NEWLINE] [NEWLINE] [STARTQ] The machine is not a 24/7 sex drugs and rocknroll party. It gives you the most valuable mental state you have. If you get happiness from a life of monk-like ascetic living, you will get that in the machine. [ENDQ] [NEWLINE] Unless the machine actually shapes the person's reaction to situations as well, the increase in happiness will not be as large as you claim. Suffering is unavoidable in any situation whether from boredom, weariness etc.  However, if you allow the machine to dictate response to the imagined scenario, then it is essentially just a dopamine dispenser. [NEWLINE] [NEWLINE] Also, people value what they take to be *real* connections to others/the outside world. A person who is a doctor values *actually helping people*. Giving him the experiences of helping imaginary computer programs is, in some sense, deeply wrong. His happiness is illusory in a way. [NEWLINE] [NEWLINE] [STARTQ] I'm not sure I understand this part. Plugging everyone into the machine would seem like the most moral act imaginable to me. Everyone would be experiencing the greatest possible happiness. The end of suffering. How can you argue against that? [ENDQ] [NEWLINE] Plugging everyone in is different than plugging person X into the machine. Plug person X into the machine and the happiness/utility he produced for society goes away as well. [NEWLINE] [NEWLINE] Also, remember that not everyone believes that happiness is the end goal. Rights such as self-ownership trump happiness of the individual. Thus, these people will deny your premise at the very start. No amount of post-plugging happiness can swamp the rights violation that continues while the person is hooked up or being hooked up.  People have a right to choose to suffer, in effect.</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Nazi Germany was a powerful majority, ghetto culture is a minority. They're entirely different. [ENDQ] [NEWLINE] The difference is irrelevant. The point is that rejecting a culture<mask><mask> its essence is doesn't make it wrong. Nazi Germany's culture largely contained antisemitism, and we reject it for one of those reasons. Likewise,<mask> by "ghetto culture" the OP means is being loud and obnoxious in large groups, instigating violence, actively intimidating people, and rejecting proper authority, then that seems like pretty good reasons to reject that culture. He would not be doing it just<mask> the culture's members are predominately black—much like we wouldn't be reject Nazi Germany's culture just<mask> they're predominately German. [NEWLINE] [NEWLINE] [STARTQ] Furthermore, OP admitted that he's doing it<mask> they "reject society",<mask> by society he means upper-class white society. [ENDQ] [NEWLINE] Really?<mask> he explicitly said<mask> he didn't [here]( [URL] ): [NEWLINE] [NEWLINE] [STARTQ] I would<mask><mask><mask><mask><mask> there is no such thing<mask> "white culture". [ENDQ] [NEWLINE] You argued that they're one of the same,<mask> that honestly doesn't hold water<mask> his point in his paragraph: [NEWLINE] [NEWLINE] [STARTQ] <mask> you are calling white culture is really just mainstream society. I don't really care<mask> races are a part of it. [ENDQ] [NEWLINE] You attempt to argue against it by saying: [NEWLINE] [NEWLINE] [STARTQ] I just explained<mask> it is based on race. "Mainstream society",<mask> you call it, is culture that is typically associated with whiteness. You are rejecting ghetto culture<mask> it is not. That's racial discrimination against a minority: racism. [ENDQ] [NEWLINE] The problem with this<mask> is that: [NEWLINE] [NEWLINE] 1. You haven't really explained<mask> it's the same other than<mask> people of the majority think. It doesn't matter<mask> the majority think; people of all races are still members of the "mainstream" culture—much like people of all races can be members of the "ghetto culture". The OP already pointed this out, and you're not addressing it properly. [NEWLINE] [NEWLINE] 2.<mask> already explained, he is not rejecting the culture<mask> it is merely different. He has already stated reasons, such glorifying crime, travel in loud and obnoxious groups, and exhibit disdain for **society** (not the culture of it<mask> the community of individuals) outside of their own.</s><pad>
Label encoding: <s> [STARTQ] Nazi Germany was a powerful majority, ghetto culture is a minority. They're entirely different. [ENDQ] [NEWLINE] The difference is irrelevant. The point is that rejecting a culture because what its essence is doesn't make it wrong. Nazi Germany's culture largely contained antisemitism, and we reject it for one of those reasons. Likewise, if by "ghetto culture" the OP means is being loud and obnoxious in large groups, instigating violence, actively intimidating people, and rejecting proper authority, then that seems like pretty good reasons to reject that culture. He would not be doing it just because the culture's members are predominately black—much like we wouldn't be reject Nazi Germany's culture just because they're predominately German. [NEWLINE] [NEWLINE] [STARTQ] Furthermore, OP admitted that he's doing it because they "reject society", where by society he means upper-class white society. [ENDQ] [NEWLINE] Really? Because he explicitly said so he didn't [here]( [URL] ): [NEWLINE] [NEWLINE] [STARTQ] I would first of all argue that there is no such thing as "white culture". [ENDQ] [NEWLINE] You argued that they're one of the same, but that honestly doesn't hold water since his point in his paragraph: [NEWLINE] [NEWLINE] [STARTQ] What you are calling white culture is really just mainstream society. I don't really care what races are a part of it. [ENDQ] [NEWLINE] You attempt to argue against it by saying: [NEWLINE] [NEWLINE] [STARTQ] I just explained why it is based on race. "Mainstream society", as you call it, is culture that is typically associated with whiteness. You are rejecting ghetto culture because it is not. That's racial discrimination against a minority: racism. [ENDQ] [NEWLINE] The problem with this though is that: [NEWLINE] [NEWLINE] 1. You haven't really explained why it's the same other than what people of the majority think. It doesn't matter what the majority think; people of all races are still members of the "mainstream" culture—much like people of all races can be members of the "ghetto culture". The OP already pointed this out, and you're not addressing it properly. [NEWLINE] [NEWLINE] 2. As already explained, he is not rejecting the culture because it is merely different. He has already stated reasons, such glorifying crime, travel in loud and obnoxious groups, and exhibit disdain for **society** (not the culture of it but the community of individuals) outside of their own.</s><pad>
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Being born has two meanings:<mask> a child is expelled from the womb and<mask> a new legal entity is recognized by the state. [NEWLINE] [NEWLINE] From [URL].pdf [NEWLINE] [NEWLINE] [STARTQ] <mask> far our inquiry has established that in modern legal jargon "birth" can mean the delivery of a human child, OR the act of bringing into full and complete existence an artificial entity. [ENDQ] [NEWLINE] Changing the definition of birth and<mask> legal protections apply offers many benefits to parents and the state. [NEWLINE] [NEWLINE] <mask> birth is not recognized until the birth certficate has been filed then parents can decide to discard damaged or unwanted babies which would be a legal abortion instead of an illegal murder. [NEWLINE] [NEWLINE] Fathers could opt out of child support by not signing the birth certicate. The number of single parent households would drop<mask> the baby would not have been born until both parents sign the birth certificate. Until the baby has been born, the birth certificate filed, it would not be a citizen. [NEWLINE] [NEWLINE] Cases of misattributed paternity could be eliminated by requiring a paternity test before the father could sign the birth certificate. Prenatal genetic testing can identify paternity<mask> an incidental finding. Birth certificates would identify the parents of the child and not the people intending to raise it<mask> anonymous sperm and egg donation would be illegal. [NEWLINE] [NEWLINE] Prenatal paternity testing would allow for provisional filing of the legal birth certificate before the baby has been physically born. [NEWLINE] [NEWLINE] Babies that have been physically born<mask> not legally born would be in a legal state similar to cohabiting couples who are not married. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Label encoding: <s>Being born has two meanings: when a child is expelled from the womb and when a new legal entity is recognized by the state. [NEWLINE] [NEWLINE] From [URL].pdf [NEWLINE] [NEWLINE] [STARTQ] So far our inquiry has established that in modern legal jargon "birth" can mean the delivery of a human child, OR the act of bringing into full and complete existence an artificial entity. [ENDQ] [NEWLINE] Changing the definition of birth and when legal protections apply offers many benefits to parents and the state. [NEWLINE] [NEWLINE] If birth is not recognized until the birth certficate has been filed then parents can decide to discard damaged or unwanted babies which would be a legal abortion instead of an illegal murder. [NEWLINE] [NEWLINE] Fathers could opt out of child support by not signing the birth certicate. The number of single parent households would drop because the baby would not have been born until both parents sign the birth certificate. Until the baby has been born, the birth certificate filed, it would not be a citizen. [NEWLINE] [NEWLINE] Cases of misattributed paternity could be eliminated by requiring a paternity test before the father could sign the birth certificate. Prenatal genetic testing can identify paternity as an incidental finding. Birth certificates would identify the parents of the child and not the people intending to raise it so anonymous sperm and egg donation would be illegal. [NEWLINE] [NEWLINE] Prenatal paternity testing would allow for provisional filing of the legal birth certificate before the baby has been physically born. [NEWLINE] [NEWLINE] Babies that have been physically born but not legally born would be in a legal state similar to cohabiting couples who are not married. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Well,<mask> you've conceded that your idea has no basis in reality, I've got nothing to argue with you about. [ENDQ] [NEWLINE] [NEWLINE] <mask> you look at the front page of CMV, I guarantee you that a substantial number of proposed policy changes aren't realistic in the current political environment (i.e. abolishing the 2nd amendment, ending U.S. involvement in foreign wars, etc.).  That doesn't mean that they aren't worth debating, partly<mask> it's important to evaluate policy for its own sake, and partly<mask> there may come a time<mask> the political environment IS more amenable to these ideas (or the political environment can be MADE more amenable through advocacy).  The question "<mask> would politicians do?" is a question designed to end debate, rather than to promote the discussion of fresh ideas.  Let's first discuss the substantive issues and establish whether this is a good or bad idea on its merits, and then we can worry about the political advocacy needed to get it implemented. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ]...by that logic, near-native fluency in at least three foreign languages (each from a different language family) should be mandatory for high school graduation... [ENDQ] [NEWLINE] [NEWLINE] No, this is not the case.  You're<mask><mask> the only reason not to teach literally everything is a lack of resources.  Teaching three languages to near-native fluency is neither pedagogically appropriate for high school, nor particularly beneficial in terms of educational outcomes.  Elsewhere, I have made the case for<mask> statistics uniquely meets both of these criteria, and<mask> it could be integrated into a standard school curriculum rather seamlessly. <mask> you're actually interested in debating pedagogy, which is the substantive question here, please refer to the more productive discussions I've had with others on this thread. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] You adhere to additional requirements with zero additional resources regularly? You spend hours teaching one thing without taking time from teaching a mutually exclusive thing? [ENDQ] [NEWLINE] [NEWLINE] Please stop straw-manning.  Schools regularly make minor changes to their programs to accommodate changes in state policy.  Sometimes, these changes require additional expenditures of resources, which ideally come from the state. <mask> they don't, it can be burdensome on schools,<mask> generally not burdensome enough to outweigh the potential benefits, particularly<mask> alternative resource streams are fully accessed.  </s>
Label encoding: <s> [STARTQ] Well, if you've conceded that your idea has no basis in reality, I've got nothing to argue with you about. [ENDQ] [NEWLINE] [NEWLINE] If you look at the front page of CMV, I guarantee you that a substantial number of proposed policy changes aren't realistic in the current political environment (i.e. abolishing the 2nd amendment, ending U.S. involvement in foreign wars, etc.).  That doesn't mean that they aren't worth debating, partly because it's important to evaluate policy for its own sake, and partly because there may come a time when the political environment IS more amenable to these ideas (or the political environment can be MADE more amenable through advocacy).  The question " What would politicians do?" is a question designed to end debate, rather than to promote the discussion of fresh ideas.  Let's first discuss the substantive issues and establish whether this is a good or bad idea on its merits, and then we can worry about the political advocacy needed to get it implemented. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ]...by that logic, near-native fluency in at least three foreign languages (each from a different language family) should be mandatory for high school graduation... [ENDQ] [NEWLINE] [NEWLINE] No, this is not the case.  You're assuming that the only reason not to teach literally everything is a lack of resources.  Teaching three languages to near-native fluency is neither pedagogically appropriate for high school, nor particularly beneficial in terms of educational outcomes.  Elsewhere, I have made the case for why statistics uniquely meets both of these criteria, and how it could be integrated into a standard school curriculum rather seamlessly.  If you're actually interested in debating pedagogy, which is the substantive question here, please refer to the more productive discussions I've had with others on this thread. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] You adhere to additional requirements with zero additional resources regularly? You spend hours teaching one thing without taking time from teaching a mutually exclusive thing? [ENDQ] [NEWLINE] [NEWLINE] Please stop straw-manning.  Schools regularly make minor changes to their programs to accommodate changes in state policy.  Sometimes, these changes require additional expenditures of resources, which ideally come from the state.  If they don't, it can be burdensome on schools, but generally not burdensome enough to outweigh the potential benefits, particularly when alternative resource streams are fully accessed.  </s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> it's meaningless to cast a blank vote,<mask> it doesn't really affect the election at all, and the only thing it really does is help some spin doctors spread some more meaningless propaganda. [NEWLINE] [NEWLINE] And I don't think it's sending a message about you being more engaged and for democracy, than<mask> you didn't vote at all (a common argument). [NEWLINE] [NEWLINE] I live in Denmark,<mask> expect it's the same in other parts of the world; you constantly get told that you HAVE to vote, and<mask> you don't, you shouldn't complain. And<mask> you say that you either don't think it matters in the end anyway, they're all the same, or you just don't like the choices, you get told to vote blank...<mask><mask> that's unaugmented and<mask><mask> there's ways you can change politics, society, and the world much more effectively than voting (demonstrate, change your own behaviour before asking others to do<mask>, donating to causes you believe in, signing up for a interest group, helping others, etc.) [NEWLINE] [NEWLINE] I think the people who vote blank do it mostly for themselves,<mask> people expect them to have an opinion on everything and want to feel like they have somehow contributed to society. [NEWLINE] [NEWLINE] <mask> let me be clear: I don't have anything against people who vote, and<mask><mask> it should be their right; I just don't think it should be expected to vote,<mask> you don't know<mask> you are voting for. [NEWLINE] [NEWLINE] Change my view :) [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I think it's meaningless to cast a blank vote, since it doesn't really affect the election at all, and the only thing it really does is help some spin doctors spread some more meaningless propaganda. [NEWLINE] [NEWLINE] And I don't think it's sending a message about you being more engaged and for democracy, than if you didn't vote at all (a common argument). [NEWLINE] [NEWLINE] I live in Denmark, but expect it's the same in other parts of the world; you constantly get told that you HAVE to vote, and if you don't, you shouldn't complain. And if you say that you either don't think it matters in the end anyway, they're all the same, or you just don't like the choices, you get told to vote blank... I think that's unaugmented and I think there's ways you can change politics, society, and the world much more effectively than voting (demonstrate, change your own behaviour before asking others to do so, donating to causes you believe in, signing up for a interest group, helping others, etc.) [NEWLINE] [NEWLINE] I think the people who vote blank do it mostly for themselves, as people expect them to have an opinion on everything and want to feel like they have somehow contributed to society. [NEWLINE] [NEWLINE] But let me be clear: I don't have anything against people who vote, and I think it should be their right; I just don't think it should be expected to vote, if you don't know what you are voting for. [NEWLINE] [NEWLINE] Change my view :) [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've been thinking about personal finance recently, and this is a conclusion that seemed reasonable.<mask>, I'm sure I haven't thought of all sides of the issue and would be interested in other points of view. [NEWLINE] [NEWLINE] My reasoning: [NEWLINE] [NEWLINE] I believe that<mask> you have the means, you should be responsible for yourself and minimize the amount you take from others. This includes not only current living costs<mask> future needs such<mask> retirement and health care. I want to minimize the amount I take from public and private assistance, freeing that money up for people who really need it. [NEWLINE] [NEWLINE] A disaster can happen at any time and can be incredibly expensive. In the US this is most likely to be a health issue. [NEWLINE] [NEWLINE] I<mask> want to be able to help my family and friends<mask> they need it.<mask> one could<mask><mask> charity recipients are no less deserving than my family, and<mask><mask>,<mask><mask> we all have a unique responsibility to those closest to us.<mask> I gave to everyone who deserved it, I would go broke. I don't have kids currently,<mask> I might some day, and it's better to start saving<mask> early<mask> possible. [NEWLINE] [NEWLINE] <mask><mask><mask>,<mask><mask> the best option is to save<mask> much<mask> I can, letting it grow throughout my life, and then give whatever is left to charity in my will. [NEWLINE] [NEWLINE] For<mask> it's worth, I'm neither rich nor poor. I make above the median individual income,<mask> less than the median household income. [NEWLINE] [NEWLINE] Thanks [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Label encoding: <s>I've been thinking about personal finance recently, and this is a conclusion that seemed reasonable. However, I'm sure I haven't thought of all sides of the issue and would be interested in other points of view. [NEWLINE] [NEWLINE] My reasoning: [NEWLINE] [NEWLINE] I believe that if you have the means, you should be responsible for yourself and minimize the amount you take from others. This includes not only current living costs but future needs such as retirement and health care. I want to minimize the amount I take from public and private assistance, freeing that money up for people who really need it. [NEWLINE] [NEWLINE] A disaster can happen at any time and can be incredibly expensive. In the US this is most likely to be a health issue. [NEWLINE] [NEWLINE] I also want to be able to help my family and friends if they need it. While one could argue that charity recipients are no less deserving than my family, and I agree, I think we all have a unique responsibility to those closest to us. If I gave to everyone who deserved it, I would go broke. I don't have kids currently, but I might some day, and it's better to start saving as early as possible. [NEWLINE] [NEWLINE] Because of this, I think the best option is to save as much as I can, letting it grow throughout my life, and then give whatever is left to charity in my will. [NEWLINE] [NEWLINE] For what it's worth, I'm neither rich nor poor. I make above the median individual income, but less than the median household income. [NEWLINE] [NEWLINE] Thanks [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>He's only really running<mask> a democrat<mask> he can secure funding from the DNC<mask> it ever reaches that point. I'm a huge supporter of him, and hope he can turn around the rest of the dems from turning into the GOP lite, which is<mask> Hillary is. Check her campaign financing and you'll see she's in bed with the corporations and big business<mask> much<mask> republicans. Sanders represents a weird combination of old school liberalism (ie. JFK/FDR) and a new, grassroots, youth movement with his positions regarding student loans and education. [NEWLINE] [NEWLINE] Bill Maher did a pretty good analysis of<mask> Hillary is an awful choice. She panders to certain sensibilities on the far left (one would call them the SJW sensibilities, focusing on women, feminism, and using racial buzzwords),<mask> in reality<mask> her husband's administration is any indicator, she'll be no better than Obama and be a centrist/moderate at best. I don't trust<mask> she says,<mask> she is pandering to that certain niche on the left- her campaign financing is much more telling and she's received massive amounts from Lehman Brothers, Citigroup, Goldman-Sachs, etc. That tells you all you need to know. Hillary represents status quo from the Obama administration (especially<mask> she gets elected with a GOP dominated Congress), and the GOP represents regression. [NEWLINE] [NEWLINE] Back to Sanders, Sanders represents the progressives from before the Red Scare. He's pushing social democratic ideas from Europe (education, reigning in the filthy rich, campaign finance reform, decreasing economic inequality) that were previously branded<mask> "socialist" or "communist" by McCarthy and his ilk in the 1950s. These ideas are now appealing<mask> one can see that European societies such<mask> Denmark, Germany, Benelux, etc. all have socialistic policies that work out quite well and those countries societies haven't collapsed<mask> of<mask>, like the Red Scare folks predicted. These appeal to the younger generation that has been more liberal than previous generations, and are bearing the full brunt of this student loan educational shithole we're currently dealing with. Hillary has never made any comment to the effect of educational reform (at least not nearly to the degree Sanders has) and this is old school liberalism, which the democratic party used to be known for until Citizens United.</s>
Label encoding: <s>He's only really running as a democrat so he can secure funding from the DNC if it ever reaches that point. I'm a huge supporter of him, and hope he can turn around the rest of the dems from turning into the GOP lite, which is what Hillary is. Check her campaign financing and you'll see she's in bed with the corporations and big business as much as republicans. Sanders represents a weird combination of old school liberalism (ie. JFK/FDR) and a new, grassroots, youth movement with his positions regarding student loans and education. [NEWLINE] [NEWLINE] Bill Maher did a pretty good analysis of why Hillary is an awful choice. She panders to certain sensibilities on the far left (one would call them the SJW sensibilities, focusing on women, feminism, and using racial buzzwords), but in reality if her husband's administration is any indicator, she'll be no better than Obama and be a centrist/moderate at best. I don't trust what she says, as she is pandering to that certain niche on the left- her campaign financing is much more telling and she's received massive amounts from Lehman Brothers, Citigroup, Goldman-Sachs, etc. That tells you all you need to know. Hillary represents status quo from the Obama administration (especially if she gets elected with a GOP dominated Congress), and the GOP represents regression. [NEWLINE] [NEWLINE] Back to Sanders, Sanders represents the progressives from before the Red Scare. He's pushing social democratic ideas from Europe (education, reigning in the filthy rich, campaign finance reform, decreasing economic inequality) that were previously branded as "socialist" or "communist" by McCarthy and his ilk in the 1950s. These ideas are now appealing as one can see that European societies such as Denmark, Germany, Benelux, etc. all have socialistic policies that work out quite well and those countries societies haven't collapsed as of yet, like the Red Scare folks predicted. These appeal to the younger generation that has been more liberal than previous generations, and are bearing the full brunt of this student loan educational shithole we're currently dealing with. Hillary has never made any comment to the effect of educational reform (at least not nearly to the degree Sanders has) and this is old school liberalism, which the democratic party used to be known for until Citizens United.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Before I state my view on this I would like to say that it is apparent (whether true or not I just mean visible) that women are more aggressively propositioned than men are. That being said I have been propositioned by both men and women. I can only think of one occasion<mask> a man has propositioned me and he was definitely persistent;<mask> that may have been<mask> I was not clear enough or he was too drunk to realize. [NEWLINE] [NEWLINE] The problem isn't a matter of empathy,<mask> a matter of communication and entitlement. A pleasant conversation between two people at a club can be interpreted<mask> friendly or flirting. Alcohol sure<mask> hell doesn't help make intentions clearer. I believe that drunk people are like children. They feel everything to excess. Anger, happiness, sadness, arousal, and really any other emotion seems to be amplified. This is<mask> people do stupid things<mask> they are hammered. Alcohol does<mask> impair decision making,<mask> the rational thought "Hey I probably should leave this girl alone." does not occur. [NEWLINE] [NEWLINE] I don't think harassing men would do very much to help the situation. This is<mask> the people who make the bar/club unsafe for women are not the people who would see the error of their ways. Rape is an act of power and<mask> placing a person willing to rape in a situation in which they would feel helpless could actually do the opposite of the desired effect. [NEWLINE] [NEWLINE] [NEWLINE] I would<mask> like to state that most guys do not ignore that type of shit.<mask> you see that shit happen at a club/bar try telling a random group of guys<mask> is happening. Regardless<mask> you know them or<mask> is happening they will come to help you and the victim. [NEWLINE] [NEWLINE] I understand<mask> you mean<mask>, most people don't notice it or are willfully ignorant. Putting someone in the situation<mask> they are forced to witness<mask> is happening will force them to act. Especially<mask> they are in greater number. Nobody wants to act alone,<mask> everyone wants to be apart of the crowd doing the right thing. [NEWLINE] [NEWLINE] To sum it up: Placing gay propositioners will do nothing (<mask><mask><mask> ) to make bars/clubs safer. Showing bar/club goers that that sort of behavior will.<mask> we should ban all alcohol. Just kidding alcohol fucking rules. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Before I state my view on this I would like to say that it is apparent (whether true or not I just mean visible) that women are more aggressively propositioned than men are. That being said I have been propositioned by both men and women. I can only think of one occasion where a man has propositioned me and he was definitely persistent; though that may have been because I was not clear enough or he was too drunk to realize. [NEWLINE] [NEWLINE] The problem isn't a matter of empathy, but a matter of communication and entitlement. A pleasant conversation between two people at a club can be interpreted as friendly or flirting. Alcohol sure as hell doesn't help make intentions clearer. I believe that drunk people are like children. They feel everything to excess. Anger, happiness, sadness, arousal, and really any other emotion seems to be amplified. This is why people do stupid things when they are hammered. Alcohol does also impair decision making, so the rational thought "Hey I probably should leave this girl alone." does not occur. [NEWLINE] [NEWLINE] I don't think harassing men would do very much to help the situation. This is because the people who make the bar/club unsafe for women are not the people who would see the error of their ways. Rape is an act of power and therefore placing a person willing to rape in a situation in which they would feel helpless could actually do the opposite of the desired effect. [NEWLINE] [NEWLINE] [NEWLINE] I would also like to state that most guys do not ignore that type of shit. If you see that shit happen at a club/bar try telling a random group of guys what is happening. Regardless if you know them or what is happening they will come to help you and the victim. [NEWLINE] [NEWLINE] I understand what you mean though, most people don't notice it or are willfully ignorant. Putting someone in the situation where they are forced to witness what is happening will force them to act. Especially if they are in greater number. Nobody wants to act alone, but everyone wants to be apart of the crowd doing the right thing. [NEWLINE] [NEWLINE] To sum it up: Placing gay propositioners will do nothing ( in my opinion ) to make bars/clubs safer. Showing bar/club goers that that sort of behavior will. So we should ban all alcohol. Just kidding alcohol fucking rules. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Translation software already exists and is pretty good. Image recognition lets you translate signs, and with voice recognition, it may be possible to have a conversation between two people who don't share a language (<mask> it would be slow and occasionally confusing). Google translate may have its hiccups,<mask> I can understand the majority of its translations and it's only going to get better over the next few years. [NEWLINE] [NEWLINE] I understand that most universities (at least in the US) require some time spent in a foreign language class,<mask> this is really part of<mask> I am arguing against. Most people I have ever taken a foreign language class with resented being there and putting in years of effort to learn a skill they might only use a couple times in their whole life, and will likely forget. [NEWLINE] [NEWLINE] Obviously people in some careers, like diplomats or aid workers, need to learn another language,<mask> I just don't see<mask> the average person will spend enough time with people who don't speak their language, in a place without an interpreter or internet access, to justify the amount of time spent learning a language. [NEWLINE] [NEWLINE] *I am excepting English from this<mask><mask> much of the Internet is available only in English. [NEWLINE] [NEWLINE] **Edit**: /u/RustyRook changed my view by linking a study which showed that learning a language could delay onset of dementia by 4.5 years. To me, this would justify the time requirement<mask> it often doesn't take that long to learn a language. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Translation software already exists and is pretty good. Image recognition lets you translate signs, and with voice recognition, it may be possible to have a conversation between two people who don't share a language ( although it would be slow and occasionally confusing). Google translate may have its hiccups, but I can understand the majority of its translations and it's only going to get better over the next few years. [NEWLINE] [NEWLINE] I understand that most universities (at least in the US) require some time spent in a foreign language class, but this is really part of what I am arguing against. Most people I have ever taken a foreign language class with resented being there and putting in years of effort to learn a skill they might only use a couple times in their whole life, and will likely forget. [NEWLINE] [NEWLINE] Obviously people in some careers, like diplomats or aid workers, need to learn another language, but I just don't see how the average person will spend enough time with people who don't speak their language, in a place without an interpreter or internet access, to justify the amount of time spent learning a language. [NEWLINE] [NEWLINE] *I am excepting English from this because so much of the Internet is available only in English. [NEWLINE] [NEWLINE] **Edit**: /u/RustyRook changed my view by linking a study which showed that learning a language could delay onset of dementia by 4.5 years. To me, this would justify the time requirement as it often doesn't take that long to learn a language. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's<mask><mask><mask> with your points: [NEWLINE] [NEWLINE] [STARTQ] Look above at the list of other novel cinematic technologies I mentioned. People initially had their minds completely blown by the talkies before coming expect nothing else. The same is true of colour and surround sound and even CGI that doesn't look like CGI. [ENDQ] [NEWLINE] The main difference between these technologies and 3D is that 3D already existed before regaining popularity recently. <mask> there is obviously an immense jump in picture quality between 3D movies now and back in the 50s, the technology has existed in some form for nearly 100 years.  This recent resurgence has only made me believe that 3D is just a fad. [NEWLINE] [NEWLINE] [STARTQ] I mentioned Toy Story above; I'll mention it again. The children's market is a vast and important one, and<mask> you really want to<mask><mask> Disney and Dreamworks have made a mistake in choosing to produce works that are aimed squarely in that direction, go right ahead. I won't get in your way,<mask> I<mask> won't agree. [ENDQ] [NEWLINE] This is certainly true,<mask> I would say that Disney and Dreamworks have such great films mostly<mask> of the storytelling, and usage of themes that all kinds of people can relate to.  I wasn't saying it's a mistake to appeal to children,<mask> it certainly is a mistake to think that 3D can make a film<mask> it is. [NEWLINE] [NEWLINE] [STARTQ] Certainly,<mask> it doesn't do this any more or less than other regular features of mainstream, big-budget film-making. [ENDQ] [NEWLINE] True,<mask> this doesn't exactly counter my point.  I said 3D is a factor in the stratification, not the only reason.  I'm just very wary of studio control over the final product<mask> they're pretty much just concerned with making money. [NEWLINE] [NEWLINE] [STARTQ] Nevertheless, the fact that this massively-funded and -popular director is able to make these inroads into the technology is<mask> will help it eventually become more accessible to everyone else.<mask> do you say about this? [ENDQ] [NEWLINE] This may be true,<mask> I personally feel that this current 3D resurgence will end well before the technology becomes accessible. [NEWLINE] [NEWLINE] I'm not entirely sure<mask> I can give you more than one delta,<mask> I'd still like to hear<mask> you think about these points.</s>
Label encoding: <s>Here's where I disagree with your points: [NEWLINE] [NEWLINE] [STARTQ] Look above at the list of other novel cinematic technologies I mentioned. People initially had their minds completely blown by the talkies before coming expect nothing else. The same is true of colour and surround sound and even CGI that doesn't look like CGI. [ENDQ] [NEWLINE] The main difference between these technologies and 3D is that 3D already existed before regaining popularity recently.  While there is obviously an immense jump in picture quality between 3D movies now and back in the 50s, the technology has existed in some form for nearly 100 years.  This recent resurgence has only made me believe that 3D is just a fad. [NEWLINE] [NEWLINE] [STARTQ] I mentioned Toy Story above; I'll mention it again. The children's market is a vast and important one, and if you really want to argue that Disney and Dreamworks have made a mistake in choosing to produce works that are aimed squarely in that direction, go right ahead. I won't get in your way, though I also won't agree. [ENDQ] [NEWLINE] This is certainly true, but I would say that Disney and Dreamworks have such great films mostly because of the storytelling, and usage of themes that all kinds of people can relate to.  I wasn't saying it's a mistake to appeal to children, but it certainly is a mistake to think that 3D can make a film what it is. [NEWLINE] [NEWLINE] [STARTQ] Certainly, but it doesn't do this any more or less than other regular features of mainstream, big-budget film-making. [ENDQ] [NEWLINE] True, but this doesn't exactly counter my point.  I said 3D is a factor in the stratification, not the only reason.  I'm just very wary of studio control over the final product since they're pretty much just concerned with making money. [NEWLINE] [NEWLINE] [STARTQ] Nevertheless, the fact that this massively-funded and -popular director is able to make these inroads into the technology is what will help it eventually become more accessible to everyone else. What do you say about this? [ENDQ] [NEWLINE] This may be true, but I personally feel that this current 3D resurgence will end well before the technology becomes accessible. [NEWLINE] [NEWLINE] I'm not entirely sure if I can give you more than one delta, but I'd still like to hear what you think about these points.</s>
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is not meant<mask> a disrespect to most organized religions, merely an observation i have come to. [NEWLINE] [NEWLINE] The widespread belief or following of most organized religions (Christianity, Islam, Judaism) is a sign of human weakness in a couple of ways. [NEWLINE] [NEWLINE] The need for an afterlife in most religions satisfies our general fears of death, and the impermanence and futility of our lives. The theory isn't backed by much scientific evidence<mask><mask><mask> i know,<mask> the reason to believe in an afterlife isn't that it makes more sense<mask> that it makes life easier. A fear of death and impermanence<mask> strong that one must believe in something that i would categorize<mask> fairytale. This is<mask> i would call a weakness. [NEWLINE] [NEWLINE] The need for moral guidance in life - to need guidance from religion to know the difference between right and wrong is<mask> a sign of weakness in that it shows a lack of judgement and wisdom for one to decide for themselves<mask> is right or wrong. Furthermore, the need of a consequence by eternal damnation<mask> persuasion not to do "bad things", and the need of an incentive by eternal salvation to do good. Is a sign of weakness in that it shows that human-nature is bad, or barbaric in a sense. [NEWLINE] [NEWLINE] [NEWLINE] EDIT: Keep the comments coming guys - "i'll secede on that point - "moral guidance" is not evidence for human weakness.<mask> a motivation to believe in a creator." [NEWLINE] [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>This is not meant as a disrespect to most organized religions, merely an observation i have come to. [NEWLINE] [NEWLINE] The widespread belief or following of most organized religions (Christianity, Islam, Judaism) is a sign of human weakness in a couple of ways. [NEWLINE] [NEWLINE] The need for an afterlife in most religions satisfies our general fears of death, and the impermanence and futility of our lives. The theory isn't backed by much scientific evidence as far as i know, so the reason to believe in an afterlife isn't that it makes more sense but that it makes life easier. A fear of death and impermanence so strong that one must believe in something that i would categorize as fairytale. This is what i would call a weakness. [NEWLINE] [NEWLINE] The need for moral guidance in life - to need guidance from religion to know the difference between right and wrong is also a sign of weakness in that it shows a lack of judgement and wisdom for one to decide for themselves what is right or wrong. Furthermore, the need of a consequence by eternal damnation as persuasion not to do "bad things", and the need of an incentive by eternal salvation to do good. Is a sign of weakness in that it shows that human-nature is bad, or barbaric in a sense. [NEWLINE] [NEWLINE] [NEWLINE] EDIT: Keep the comments coming guys - "i'll secede on that point - "moral guidance" is not evidence for human weakness. But a motivation to believe in a creator." [NEWLINE] [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for the suggestion for the line break, I'll be sure to do that from now on.  Even<mask> your view was already changed, I just going to respond to<mask> you said for the heck of it. [NEWLINE] [NEWLINE] Your age-old argument against immortality is correct....<mask> it all goes back to<mask> I said about the choosers personality and willpower, especially<mask> you're discussing immortality. <mask> the chooser is old enough to understand the pros and cons of the power, then<mask><mask> more good can be done for the world.  And<mask> it is in the persons interest is to better the world, then this can be a great power to do it.  The chooser may realize everything you are saying about immortality,<mask> they may view it<mask> a kind of...self sacrifice.  Their one life can make many more much better. [NEWLINE] [NEWLINE] For super intelligence, I would agree with you there.  And all of those thoughts that Einstein, had are mainly due to the fact that he is not a computer, he's human. <mask> you have a very good point there. <mask>, Mind control for me is evil...very evil.  I don't need to explain<mask> it can be<mask> the user is a super-villain or something silly like that. <mask> using mind control to say...have world peace is an evil act in my mind. <mask> you are suppressing who that person is, and you are putting in their head<mask> you think is correct.  Or stopping an oppressive state that controls it people (ex. 1984), then in the act of stopping them...you become the oppressor of their freedoms. <mask><mask> you are trying them to stop them from oppressing freedoms. [NEWLINE] [NEWLINE] This all comes back to the person<mask>...and I guess the self-sacrifice argument that I used can<mask> be used here,<mask> you mentioned that you wouldn't want to be tired of life.  Well<mask> would you want to be this all controlling being may become<mask> it tries to stop? <mask> this statement assumes that you will<mask> some major things with your powers, like world peace or stopping 1984-like governments.    Either way I understand<mask> you are coming from and it depends<mask> you use your powers for.</s>
Label encoding: <s>Thank you for the suggestion for the line break, I'll be sure to do that from now on.  Even if your view was already changed, I just going to respond to what you said for the heck of it. [NEWLINE] [NEWLINE] Your age-old argument against immortality is correct.... but it all goes back to what I said about the choosers personality and willpower, especially when you're discussing immortality.  If the chooser is old enough to understand the pros and cons of the power, then I think more good can be done for the world.  And if it is in the persons interest is to better the world, then this can be a great power to do it.  The chooser may realize everything you are saying about immortality, but they may view it as a kind of...self sacrifice.  Their one life can make many more much better. [NEWLINE] [NEWLINE] For super intelligence, I would agree with you there.  And all of those thoughts that Einstein, had are mainly due to the fact that he is not a computer, he's human.  So you have a very good point there.  Also, Mind control for me is evil...very evil.  I don't need to explain why it can be if the user is a super-villain or something silly like that.  But using mind control to say...have world peace is an evil act in my mind.  As you are suppressing who that person is, and you are putting in their head what you think is correct.  Or stopping an oppressive state that controls it people (ex. 1984), then in the act of stopping them...you become the oppressor of their freedoms.  Even though you are trying them to stop them from oppressing freedoms. [NEWLINE] [NEWLINE] This all comes back to the person though...and I guess the self-sacrifice argument that I used can also be used here, but you mentioned that you wouldn't want to be tired of life.  Well why would you want to be this all controlling being may become what it tries to stop?  Though this statement assumes that you will so some major things with your powers, like world peace or stopping 1984-like governments.    Either way I understand where you are coming from and it depends what you use your powers for.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>China is not truly on route to becoming the next superpower. In the long term they have to deal with a bloated country with an enormous demand for resources (a demand that cannot be met in a sustainable manner). In order for them to become the next superpower, they must transition into the status of a fully developed nation. [NEWLINE] [NEWLINE] Believe it or not, their enormous population is more of a hindrance than a blessing<mask> it comes to long term development. Aging problems (Far too many old people with not enough young people), pollution, environmental issues, it's truly a burden for any developed nation. They acknowledge this and are taking great strides in reducing birth rates,<mask> with this comes a host of other problems. Social issues (male vs female ratio) and broader economic ones<mask> well (4-2-1 problem). [NEWLINE] [NEWLINE] The problem is that China's success relies on the wealth of the West. Their huge industrial capacity is great for making loads of cheap products and selling them to wealthier countries.<mask> they lag behind technologically and socially. Their entire system of functioning<mask> of now does not permit them to usurp the West in any significant way. Their GDP (nominal) is still half that of the US. Militarily, they aren't even close to the US. Superpowers rely on [force projection]( [URL] ) in modern warfare. In this sense, Air power and Naval power is supreme. China can't even touch the US on this front, and they aren't even close to catching up. One aircraft carrier to the US's ten. [NEWLINE] [NEWLINE] They are growing, yes,<mask> soon this growth will level off<mask> they begin to catch up. They are making use of already established technologies in conjunction with their industrial capacity in order to develop.<mask> true development requires innovation, something that is still firmly in the hands of the West. And once the US and Europe truly departs from this recession, growth there will speed up. [NEWLINE] [NEWLINE] China could potentially already be considered a'superpower'.<mask>, for them to truly usurp the United States? Decades. Thirty, Forty years at least. And I believe that at that point, with influence and development will come more and more democracy. This will lead to less corruption, less nationalism and more human rights for everyone involved.</s>
Label encoding: <s>China is not truly on route to becoming the next superpower. In the long term they have to deal with a bloated country with an enormous demand for resources (a demand that cannot be met in a sustainable manner). In order for them to become the next superpower, they must transition into the status of a fully developed nation. [NEWLINE] [NEWLINE] Believe it or not, their enormous population is more of a hindrance than a blessing when it comes to long term development. Aging problems (Far too many old people with not enough young people), pollution, environmental issues, it's truly a burden for any developed nation. They acknowledge this and are taking great strides in reducing birth rates, but with this comes a host of other problems. Social issues (male vs female ratio) and broader economic ones as well (4-2-1 problem). [NEWLINE] [NEWLINE] The problem is that China's success relies on the wealth of the West. Their huge industrial capacity is great for making loads of cheap products and selling them to wealthier countries. But they lag behind technologically and socially. Their entire system of functioning as of now does not permit them to usurp the West in any significant way. Their GDP (nominal) is still half that of the US. Militarily, they aren't even close to the US. Superpowers rely on [force projection]( [URL] ) in modern warfare. In this sense, Air power and Naval power is supreme. China can't even touch the US on this front, and they aren't even close to catching up. One aircraft carrier to the US's ten. [NEWLINE] [NEWLINE] They are growing, yes, but soon this growth will level off as they begin to catch up. They are making use of already established technologies in conjunction with their industrial capacity in order to develop. But true development requires innovation, something that is still firmly in the hands of the West. And once the US and Europe truly departs from this recession, growth there will speed up. [NEWLINE] [NEWLINE] China could potentially already be considered a'superpower'. However, for them to truly usurp the United States? Decades. Thirty, Forty years at least. And I believe that at that point, with influence and development will come more and more democracy. This will lead to less corruption, less nationalism and more human rights for everyone involved.</s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0890, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I came from the opposite perspective, atheism, to belief. I was raised without belief, my father was openly an Atheist and my mother was vaguely spiritual. I started with the same ideas<mask> you about mainstream interpretations of God.<mask> could one God be right,<mask> there are<mask> many equally valid versions?<mask> could evolution be wrong and religion right? Etc etc. [NEWLINE] [NEWLINE] After thinking about it A LOT (I was a philosophy major), and studying Kant in particular, I began to think of atheism in a different light. Kant described in the difference between the Noumenal World and the Phenomenal World. The Noumenal world is things in themselves, e.g. the table<mask> it actually is. The Phenomenal world is our perceptions of those things, e.g. light bounces off the table and we perceive it<mask> brown and shaped in a certain way. Even our touch just our brains' interpretation of electrical signals between ourselves and things in themselves. For all we know, the table could, in itself, be an Octopus-shaped demon. Even space and time are not things in themselves, there are only the way our brains organize information. [NEWLINE] [NEWLINE] Kant says that we can never, ever know anything about the Noumenal world. It is hidden to us,<mask> we can only interpret it through our subjective perception. [NEWLINE] [NEWLINE] To me, this is absolutely true, and says a lot about God. God,<mask> it exists, exists in the Noumenal world. It is by definition beyond our understanding. Flawed interpretations like organized religion seek to describe God,<mask> it's ultimately a frivolous task. [NEWLINE] [NEWLINE] <mask>, being certain that something beyond us doesn't exist is equally frivolous. It's not that we know that there is no table, it's that we don't know anything about the table in itself. We can't know, until we die (and even then we might not know). [NEWLINE] [NEWLINE] <mask> believing and non-believing are equally unsupported, I now choose to believe. I make that choice<mask> I feel that there is something greater, and it makes me happy and comforted to believe,<mask> choosing to non-believe makes me sad and depressed.<mask> I don't know for certain either way. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I came from the opposite perspective, atheism, to belief. I was raised without belief, my father was openly an Atheist and my mother was vaguely spiritual. I started with the same ideas as you about mainstream interpretations of God. How could one God be right, when there are so many equally valid versions? How could evolution be wrong and religion right? Etc etc. [NEWLINE] [NEWLINE] After thinking about it A LOT (I was a philosophy major), and studying Kant in particular, I began to think of atheism in a different light. Kant described in the difference between the Noumenal World and the Phenomenal World. The Noumenal world is things in themselves, e.g. the table as it actually is. The Phenomenal world is our perceptions of those things, e.g. light bounces off the table and we perceive it as brown and shaped in a certain way. Even our touch just our brains' interpretation of electrical signals between ourselves and things in themselves. For all we know, the table could, in itself, be an Octopus-shaped demon. Even space and time are not things in themselves, there are only the way our brains organize information. [NEWLINE] [NEWLINE] Kant says that we can never, ever know anything about the Noumenal world. It is hidden to us, because we can only interpret it through our subjective perception. [NEWLINE] [NEWLINE] To me, this is absolutely true, and says a lot about God. God, if it exists, exists in the Noumenal world. It is by definition beyond our understanding. Flawed interpretations like organized religion seek to describe God, but it's ultimately a frivolous task. [NEWLINE] [NEWLINE] However, being certain that something beyond us doesn't exist is equally frivolous. It's not that we know that there is no table, it's that we don't know anything about the table in itself. We can't know, until we die (and even then we might not know). [NEWLINE] [NEWLINE] Since believing and non-believing are equally unsupported, I now choose to believe. I make that choice because I feel that there is something greater, and it makes me happy and comforted to believe, where choosing to non-believe makes me sad and depressed. But I don't know for certain either way. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I dropped out of high school<mask> I was a sophomore<mask> they didn't have any sort of gifted programs to get me ahead [ENDQ] [NEWLINE] I know this is a bit off-topic,<mask> I want to clue into this for a moment. [NEWLINE] [NEWLINE] It was not the school's fault that you dropped out.  You dropped out of high school,<mask> *you* didn't want to go.  That is all. [NEWLINE] [NEWLINE] It is not the school's job to entertain you or "get you ahead" (whatever that means).  The school's job is to bring the majority of children to a basic competency level.  That is all. [NEWLINE] [NEWLINE] At the same time, they do not prevent you for exploring stuff on your own.  This is an important point.  You don't need the school's (or anyone else's) permission to learn.  You don't even need their help. [NEWLINE] [NEWLINE] <mask> you really are smart enough that school was boring, then dropping out was not the right course of action.  Instead, you could have attended the school and pursued your interests on the side.  For a truly gifted person, very little time/effort need to be devoted to this.  Sure, it is boring,<mask> that brings me to my next point. [NEWLINE] [NEWLINE] There is going to be a *bunch* of boring bullshit in your life.  That is a fact.  And you are going to be expected to do it.  Even the most exciting jobs have this<mask> an aspect to them.  The sooner you learn to just buckle down, deal with the boring shit, and get it done, the better off you are going to be. [NEWLINE] [NEWLINE] This is<mask> it is sad to have dropped out of high school -- you missed a golden opportunity to learn a valuable life lesson:<mask> to get boring shit done, and still make time for interesting things. [NEWLINE] [NEWLINE] This is my two cents, of course, and you can take it or leave it. <mask> an ability to do mindless stuff, or tedious stuff, or stuff you already know<mask> to do, is going to be a skill that is not only necessary,<mask> valued by any employer (including yourself,<mask> you are self-employed).  Learn it. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I dropped out of high school when I was a sophomore because they didn't have any sort of gifted programs to get me ahead [ENDQ] [NEWLINE] I know this is a bit off-topic, but I want to clue into this for a moment. [NEWLINE] [NEWLINE] It was not the school's fault that you dropped out.  You dropped out of high school, because *you* didn't want to go.  That is all. [NEWLINE] [NEWLINE] It is not the school's job to entertain you or "get you ahead" (whatever that means).  The school's job is to bring the majority of children to a basic competency level.  That is all. [NEWLINE] [NEWLINE] At the same time, they do not prevent you for exploring stuff on your own.  This is an important point.  You don't need the school's (or anyone else's) permission to learn.  You don't even need their help. [NEWLINE] [NEWLINE] If you really are smart enough that school was boring, then dropping out was not the right course of action.  Instead, you could have attended the school and pursued your interests on the side.  For a truly gifted person, very little time/effort need to be devoted to this.  Sure, it is boring, but that brings me to my next point. [NEWLINE] [NEWLINE] There is going to be a *bunch* of boring bullshit in your life.  That is a fact.  And you are going to be expected to do it.  Even the most exciting jobs have this as an aspect to them.  The sooner you learn to just buckle down, deal with the boring shit, and get it done, the better off you are going to be. [NEWLINE] [NEWLINE] This is why it is sad to have dropped out of high school -- you missed a golden opportunity to learn a valuable life lesson: how to get boring shit done, and still make time for interesting things. [NEWLINE] [NEWLINE] This is my two cents, of course, and you can take it or leave it.  But an ability to do mindless stuff, or tedious stuff, or stuff you already know how to do, is going to be a skill that is not only necessary, but valued by any employer (including yourself, if you are self-employed).  Learn it. [NEWLINE] </s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My wife and I had our first child 6 weeks ago. It's been a challenging<mask> fun 6 weeks. We could not be more thrilled to have a healthy peeing, pooping, and crying machine! [NEWLINE] [NEWLINE] With that said, I have no desire to have more than one child. [NEWLINE] [NEWLINE] <mask> we only have one child we'll all be able to live a very comfortable life. We won't need to buy a larger house.  We've already started a college fund for the little guy and he'll be able to go to any college in the country (granted he need to earn his way in!).  We'll be able to buy a nice lake house in 10 years or<mask>. In short, we'll be able ensure he gets all the resources he'll ever need. [NEWLINE] [NEWLINE] <mask> we have more than one those previously mentioned resources will be stretched more thinly. Don't get me wrong, we won't go hungry<mask> we have two children,<mask> we won't be able to fully fund two educations and still live the life we would have been able to with one child. [NEWLINE] [NEWLINE] Money and lifestyle aside - we're very thankful that he's the picture of health.<mask><mask> we have a second and s/he isn't<mask> healthy?<mask> ruin<mask> you already know is a good thing? [NEWLINE] [NEWLINE] I guess I just do not see the value in having a second child.  In my eyes, less is more. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>My wife and I had our first child 6 weeks ago. It's been a challenging but fun 6 weeks. We could not be more thrilled to have a healthy peeing, pooping, and crying machine! [NEWLINE] [NEWLINE] With that said, I have no desire to have more than one child. [NEWLINE] [NEWLINE] If we only have one child we'll all be able to live a very comfortable life. We won't need to buy a larger house.  We've already started a college fund for the little guy and he'll be able to go to any college in the country (granted he need to earn his way in!).  We'll be able to buy a nice lake house in 10 years or so. In short, we'll be able ensure he gets all the resources he'll ever need. [NEWLINE] [NEWLINE] If we have more than one those previously mentioned resources will be stretched more thinly. Don't get me wrong, we won't go hungry if we have two children, but we won't be able to fully fund two educations and still live the life we would have been able to with one child. [NEWLINE] [NEWLINE] Money and lifestyle aside - we're very thankful that he's the picture of health. What if we have a second and s/he isn't as healthy? Why ruin what you already know is a good thing? [NEWLINE] [NEWLINE] I guess I just do not see the value in having a second child.  In my eyes, less is more. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can look for additional sources.  Unfortunately, just describing it is like trying to describe the dead puck era in the NHL; it really had to be witnessed to be understood. [NEWLINE] [NEWLINE] The other factor with the 1990 tournament is that European women's hockey was bigger than in North America,<mask> there were actual female-only teams (which allowed checking) that the European national teams were drawing from.  The Canadian and American women's teams were stocked with players who'd grown up playing hockey against boys the entire way up, which<mask> included checking.  The European teams insisted that checking be allowed (figuring it would be an advantage for them), not realizing that the North American teams could mentally flip a switch and *really* start roughing it up<mask> conditions allowed.  Let's just say that the conditions in the 1990 tournament allowed it, and the European teams (save for always-gritty Finland) were largely beaten into submission. [NEWLINE] [NEWLINE] EDIT: The following comes from [a MacLean's article]( [URL].txt) on April 2, 1990, written by D'Arcy Jenish: [NEWLINE] [STARTQ] <mask>, in last week's tournament, the Canadian women showed that they can play a tough brand of hockey.  Their match against Sweden included several thunderous collisions and 21 minor penalties for such infractions<mask> boarding, roughing and high-sticking. [ENDQ] [NEWLINE] This one [From Sports Illustrated]( [URL].txt): [NEWLINE] [STARTQ] And<mask> with the men, it's hitting.  Just ask U.S. team captain Tina Cardinale, whose right forearm and elbow were a mass of purple-and-blue welts, courtesy of a slash early in the tournament.  Canada's France St.-Louis spent three days in a hospital after taking a stick across the throat, and Finland's Kirsi Hirvonen was carried away with a neck injury after being cross-checked. [ENDQ] [NEWLINE] [STARTQ] Bodychecking in women's hockey is illegal in the U.S.,<mask> tournament [ENDQ] rules allowed for full-contact checking with certain limitations along [NEWLINE] the boards.  That did not present much of a problem for a U.S. team... [NEWLINE] [NEWLINE] &gt; "They're tougher creatures than we ever gave them credit for," said (American coach Don) MacLeod.</s>
Label encoding: <s>I can look for additional sources.  Unfortunately, just describing it is like trying to describe the dead puck era in the NHL; it really had to be witnessed to be understood. [NEWLINE] [NEWLINE] The other factor with the 1990 tournament is that European women's hockey was bigger than in North America, so there were actual female-only teams (which allowed checking) that the European national teams were drawing from.  The Canadian and American women's teams were stocked with players who'd grown up playing hockey against boys the entire way up, which also included checking.  The European teams insisted that checking be allowed (figuring it would be an advantage for them), not realizing that the North American teams could mentally flip a switch and *really* start roughing it up if conditions allowed.  Let's just say that the conditions in the 1990 tournament allowed it, and the European teams (save for always-gritty Finland) were largely beaten into submission. [NEWLINE] [NEWLINE] EDIT: The following comes from [a MacLean's article]( [URL].txt) on April 2, 1990, written by D'Arcy Jenish: [NEWLINE] [STARTQ] But, in last week's tournament, the Canadian women showed that they can play a tough brand of hockey.  Their match against Sweden included several thunderous collisions and 21 minor penalties for such infractions as boarding, roughing and high-sticking. [ENDQ] [NEWLINE] This one [From Sports Illustrated]( [URL].txt): [NEWLINE] [STARTQ] And as with the men, it's hitting.  Just ask U.S. team captain Tina Cardinale, whose right forearm and elbow were a mass of purple-and-blue welts, courtesy of a slash early in the tournament.  Canada's France St.-Louis spent three days in a hospital after taking a stick across the throat, and Finland's Kirsi Hirvonen was carried away with a neck injury after being cross-checked. [ENDQ] [NEWLINE] [STARTQ] Bodychecking in women's hockey is illegal in the U.S., but tournament [ENDQ] rules allowed for full-contact checking with certain limitations along [NEWLINE] the boards.  That did not present much of a problem for a U.S. team... [NEWLINE] [NEWLINE] &gt; "They're tougher creatures than we ever gave them credit for," said (American coach Don) MacLeod.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When I saw the title of your CMV, I was expecting a critique of things like framing a toilet seat, stuff<mask> the artist really is seeking to make people "question art" rather than aim for any aesthetic appeal. [NEWLINE] [NEWLINE] <mask> instead you brought up the "random splats" side of the modern art argument, and here<mask><mask> it's actually not the best one for your CMV angle. <mask> some modern art, like my example of just framing a toilet seat, really doesn't do much else<mask> make you question<mask> is art or not. (<mask> someone who went to art school, I might have then argued that there's still fun to be had in that,<mask> I will easily understand<mask> no one else cares) [NEWLINE] [NEWLINE] <mask> the "random splat" style of modern art, that I would argue has a much better chance of having meaning beyond the questioning itself.  The colors and strokes and the composition itself, even<mask> seemingly random, can give an emotional effect.  Even a completely blank wall of one color can give an emotional effect, or else<mask> would people care about choosing between paint colors<mask> painting their house? [NEWLINE] [NEWLINE] In your link to the test of toddler artwork vs modern art, some of those toddler's works did look very pleasing!  #6 was my favorite, with its bright variety of colors.  #10 and #14 looked nice and soothing, too.  At least, in the areas<mask> they cropped in. <mask> a toddler got lucky enough to keep up that same style and color across a whole sheet large enough to hang on the wall, then I would gladly hang it on my wall,<mask> it would look good.  This is probably going to be a matter of luck,<mask>. [NEWLINE] [NEWLINE] One thing people often don't realize is<mask> *large* some of these works are.  Jackson Pollack's works are usually [about the size of a whole wall]( [URL].jpg).  A bunch of toddlers smearing paint randomly certainly can make stuff that sometimes looks<mask> good. <mask> could they keep that up consistently across the whole wall?  And<mask> many tries would it take, and<mask> much money would you have wasted on paint and canvas by then?</s>
Label encoding: <s>When I saw the title of your CMV, I was expecting a critique of things like framing a toilet seat, stuff where the artist really is seeking to make people "question art" rather than aim for any aesthetic appeal. [NEWLINE] [NEWLINE] But instead you brought up the "random splats" side of the modern art argument, and here I think it's actually not the best one for your CMV angle.  Because some modern art, like my example of just framing a toilet seat, really doesn't do much else besides make you question what is art or not. ( as someone who went to art school, I might have then argued that there's still fun to be had in that, but I will easily understand if no one else cares) [NEWLINE] [NEWLINE] But the "random splat" style of modern art, that I would argue has a much better chance of having meaning beyond the questioning itself.  The colors and strokes and the composition itself, even when seemingly random, can give an emotional effect.  Even a completely blank wall of one color can give an emotional effect, or else why would people care about choosing between paint colors when painting their house? [NEWLINE] [NEWLINE] In your link to the test of toddler artwork vs modern art, some of those toddler's works did look very pleasing!  #6 was my favorite, with its bright variety of colors.  #10 and #14 looked nice and soothing, too.  At least, in the areas where they cropped in.  If a toddler got lucky enough to keep up that same style and color across a whole sheet large enough to hang on the wall, then I would gladly hang it on my wall, because it would look good.  This is probably going to be a matter of luck, though. [NEWLINE] [NEWLINE] One thing people often don't realize is how *large* some of these works are.  Jackson Pollack's works are usually [about the size of a whole wall]( [URL].jpg).  A bunch of toddlers smearing paint randomly certainly can make stuff that sometimes looks as good.  But could they keep that up consistently across the whole wall?  And how many tries would it take, and how much money would you have wasted on paint and canvas by then?</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the expectation that group X should disregard whatever they call themselves<mask> of a vocal minority that is incompatible with the majority's beliefs. [ENDQ] [NEWLINE] In certain situations this may be in the interest of the group, for their self image or gaining public traction on divisive issues. Moderate feminists should get to decide whether to do this for themselves,<mask> OP may just be arguing that it would be in their interest to do<mask>. [NEWLINE] [NEWLINE] Regarding MRA's, the few I've seen come to main subs have all been very reasonable and good advocates of their cause. They cite information, speak calmly, and almost always speak dismissively of extremists. It isn't a cause I would involve myself in,<mask><mask> anything *were* to draw me in it would be my initial impression that the community *may* regulate the extremists and potential woman haters such a group would draw. I don't see this<mask> much with feminism. [NEWLINE] [NEWLINE] <mask> I was a member of the Gawker network I would occasionally comment on Jezebel. The commenters ranged form civil and moderate to extreme and annoyingly snarky (aka SRS stuff). The thing is, people wouldn't disregard the extremists<mask><mask><mask> they were espousing feminist ideas.<mask> someone disagreed with a tenant of feminist ideology,<mask>, everyone would come down on you. The most dismissive I've seen self-proclaimed feminists of the extremists is to say they don't represent the whole; I don't see moderates actively critiquing or dismissing there ideological tenants. [NEWLINE] [NEWLINE] <mask> the above is personal anecdote,<mask><mask> it is<mask> important to remember that feminism is an ideology with specific movements containing specific ideas spanning almost half a century at this point. Usually<mask> one says they 'aren't a feminist' they get the fallacy that '<mask> you believe women should be equal, you should call yourself a feminist'. Which would be fine, except feminism is an ideology beyond just that core idea, and some of its ideas are divisive for a reason. Having an alternative name for moderates who disagree with a large part of the ideology is fine and may help avoid focus on extremists. This is<mask> different form Muslims issue,<mask> moderate feminists don't have a central text in common with extreme ones, for example.</s>
Label encoding: <s> [STARTQ] the expectation that group X should disregard whatever they call themselves because of a vocal minority that is incompatible with the majority's beliefs. [ENDQ] [NEWLINE] In certain situations this may be in the interest of the group, for their self image or gaining public traction on divisive issues. Moderate feminists should get to decide whether to do this for themselves, but OP may just be arguing that it would be in their interest to do so. [NEWLINE] [NEWLINE] Regarding MRA's, the few I've seen come to main subs have all been very reasonable and good advocates of their cause. They cite information, speak calmly, and almost always speak dismissively of extremists. It isn't a cause I would involve myself in, but if anything *were* to draw me in it would be my initial impression that the community *may* regulate the extremists and potential woman haters such a group would draw. I don't see this as much with feminism. [NEWLINE] [NEWLINE] When I was a member of the Gawker network I would occasionally comment on Jezebel. The commenters ranged form civil and moderate to extreme and annoyingly snarky (aka SRS stuff). The thing is, people wouldn't disregard the extremists as long as they were espousing feminist ideas. If someone disagreed with a tenant of feminist ideology, however, everyone would come down on you. The most dismissive I've seen self-proclaimed feminists of the extremists is to say they don't represent the whole; I don't see moderates actively critiquing or dismissing there ideological tenants. [NEWLINE] [NEWLINE] While the above is personal anecdote, I think it is also important to remember that feminism is an ideology with specific movements containing specific ideas spanning almost half a century at this point. Usually when one says they 'aren't a feminist' they get the fallacy that'if you believe women should be equal, you should call yourself a feminist'. Which would be fine, except feminism is an ideology beyond just that core idea, and some of its ideas are divisive for a reason. Having an alternative name for moderates who disagree with a large part of the ideology is fine and may help avoid focus on extremists. This is also different form Muslims issue, since moderate feminists don't have a central text in common with extreme ones, for example.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] don't have any reason to be afraid of terrorism and neither does anybody in the US or the UK. [ENDQ] [STARTQ] Im Australian. It really shocks me the way people will just accept<mask> they're told without thinking about it. [ENDQ] [NEWLINE] I'm a dual Australian and British citizen. <mask> I lived in London the IRA would put bombs in cars, trucks, and on train lines.  Just a few terrorism events that I remember: [NEWLINE] [NEWLINE] - They bombed [Harrod's]( [URL] -uQIFBkQ) which is a major department store. [NEWLINE] [NEWLINE] - They put a bomb underneath a flat bed truck and totally [destroyed]( [URL].jpg) a big office building in a neighbourhood I was living in. [NEWLINE] [NEWLINE] - They bombed a [shopping centre]( [URL] ) in Britain's 2nd largest city. [NEWLINE] [NEWLINE] - They killed lots of people travelling on trains. [NEWLINE] [NEWLINE] You don't live in fear.  There was a sense of normality to it.<mask> I used to go to school, it was quite common to have to detour<mask> there was a bomb scare at my local train station (Kings Cross).   That station<mask> had a [massive fire]( [URL] ) which was non-terrorism related. [NEWLINE] [NEWLINE] It became such a problem that the entire City of London (1 square mile) had cameras installed all over it, and the police got to co-ordinate with all the business owners who had CCTV. [NEWLINE] [NEWLINE] This was at a time<mask> terrorism was not an overused word. [NEWLINE] [NEWLINE] Post 9-11, and after I moved to Australia, London still received terrorist attacks: [NEWLINE] [NEWLINE] - July 2005 [Bus &amp; Tube]( [URL] -OKK0ctVqk) bombings. [NEWLINE] [NEWLINE] [STARTQ] Statistically I have more of a chance of being struck by lightning than killed by a terrorist [ENDQ] [NEWLINE] You<mask> get to choose to stand out in a rain storm.  You personally cannot forecast a terrorist event. Many terrorist events in the UK are aimed at areas<mask> thousands of people go through every day of the week. [NEWLINE] [NEWLINE] Just<mask> there hasn't been a bomb in Sydney or Melbourne or somewhere else, doesn't mean nobody has tried. [NEWLINE] </s>
Label encoding: <s> [STARTQ] don't have any reason to be afraid of terrorism and neither does anybody in the US or the UK. [ENDQ] [STARTQ] Im Australian. It really shocks me the way people will just accept what they're told without thinking about it. [ENDQ] [NEWLINE] I'm a dual Australian and British citizen.  When I lived in London the IRA would put bombs in cars, trucks, and on train lines.  Just a few terrorism events that I remember: [NEWLINE] [NEWLINE] - They bombed [Harrod's]( [URL] -uQIFBkQ) which is a major department store. [NEWLINE] [NEWLINE] - They put a bomb underneath a flat bed truck and totally [destroyed]( [URL].jpg) a big office building in a neighbourhood I was living in. [NEWLINE] [NEWLINE] - They bombed a [shopping centre]( [URL] ) in Britain's 2nd largest city. [NEWLINE] [NEWLINE] - They killed lots of people travelling on trains. [NEWLINE] [NEWLINE] You don't live in fear.  There was a sense of normality to it. When I used to go to school, it was quite common to have to detour because there was a bomb scare at my local train station (Kings Cross).   That station also had a [massive fire]( [URL] ) which was non-terrorism related. [NEWLINE] [NEWLINE] It became such a problem that the entire City of London (1 square mile) had cameras installed all over it, and the police got to co-ordinate with all the business owners who had CCTV. [NEWLINE] [NEWLINE] This was at a time when terrorism was not an overused word. [NEWLINE] [NEWLINE] Post 9-11, and after I moved to Australia, London still received terrorist attacks: [NEWLINE] [NEWLINE] - July 2005 [Bus &amp; Tube]( [URL] -OKK0ctVqk) bombings. [NEWLINE] [NEWLINE] [STARTQ] Statistically I have more of a chance of being struck by lightning than killed by a terrorist [ENDQ] [NEWLINE] You also get to choose to stand out in a rain storm.  You personally cannot forecast a terrorist event. Many terrorist events in the UK are aimed at areas where thousands of people go through every day of the week. [NEWLINE] [NEWLINE] Just because there hasn't been a bomb in Sydney or Melbourne or somewhere else, doesn't mean nobody has tried. [NEWLINE] </s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe free speech is important, and a generally good thing.<mask>,<mask> you state with an idea, let's say "Forks are a blight upon society", and someone disagrees with you,<mask> you immediately jump to free speech<mask> defense of your idea, you are basically admitting that you have no worthwhile arguments. Yes, you are allowed to say that,<mask><mask> you actually had something to back it up, say "sporks are far more versatile, cutting the required utensil manufacturing by a third, reducing global warming 93%", you would say that. Free speech does nothing to back up the validity of your claim,<mask> you are free to be wrong. [NEWLINE] [NEWLINE] EDIT: I have been convinced that it can be a way of simply saying you do not wish to argue the point further, that you want to let the argument stand<mask> it does. [NEWLINE] [NEWLINE] And, to clarify, I am not talking about free speech<mask> a legal concept. Having to prove your speech legal says nothing<mask> to its validity. I mean having to fall back on free speech<mask> an argument does nothing to prove you right. [NEWLINE] [NEWLINE] EDIT 2: I now feel that an argument for, or against someone's right to a viewpoint has no bearing on the validity of the viewpoint.<mask>, the person who first feels the need to argue this right,<mask>, likely has no argument. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I believe free speech is important, and a generally good thing. However, if you state with an idea, let's say "Forks are a blight upon society", and someone disagrees with you, if you immediately jump to free speech as defense of your idea, you are basically admitting that you have no worthwhile arguments. Yes, you are allowed to say that, but if you actually had something to back it up, say "sporks are far more versatile, cutting the required utensil manufacturing by a third, reducing global warming 93%", you would say that. Free speech does nothing to back up the validity of your claim, as you are free to be wrong. [NEWLINE] [NEWLINE] EDIT: I have been convinced that it can be a way of simply saying you do not wish to argue the point further, that you want to let the argument stand where it does. [NEWLINE] [NEWLINE] And, to clarify, I am not talking about free speech as a legal concept. Having to prove your speech legal says nothing as to its validity. I mean having to fall back on free speech as an argument does nothing to prove you right. [NEWLINE] [NEWLINE] EDIT 2: I now feel that an argument for, or against someone's right to a viewpoint has no bearing on the validity of the viewpoint. But, the person who first feels the need to argue this right, however, likely has no argument. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Okay,<mask>, the best argument I know for it being immoral to have sex with animals is that they are incapable of consent and that<mask> sexual behaviour with animals is rape. I cannot imagine that any animal would give its consent to dying just in order to satisfy a person's appetite. Furthermore, I don't see a moral difference between acts of bestiality and some methods of animal husbandry. [ENDQ] [NEWLINE] I don't see the inconsistency. Perhaps sex acts *simply do* require consent, whereas eating non-sapient animals *simply doesn't* require consent.<mask> is the contradiction? [NEWLINE] [NEWLINE] <mask> here are a few moral frameworks, and arguments from each. Note that it isn't<mask> important whether these frameworks are *true*, only that they are consistent frameworks which justify both beliefs in question. [NEWLINE] [NEWLINE] * Natural Law theories seek to ground norms in their natural *purpose*. The *purpose* for sex might be procreation.<mask>, sex with animals is wrong,<mask> procreation is impossible. (An alternate purpose might be solidifying bonds with one's partner.<mask>,<mask> this purpose may exist between humans, it is dubious it could work between humans and non-human animals. Bestiality is plausibly psychologically harmful.) Carnivorism,<mask><mask><mask><mask>, doesn't seem impermissible on natural law theory. [NEWLINE] * Consequentialism holds that consequences are the primary aspect of moral consideration. Perhaps eating animals leads to greater net happiness. (Maybe not the current methods of factory farming,<mask><mask> about responsible free range farming,<mask> animals have good lives before being taken for food?) Perhaps bestiality contributes net negative consequences. Under such conditions (which aren't necessarily that implausible), consequentialism would entail that bestiality is wrong,<mask> eating meat is, in some cases, permissible. [NEWLINE] * Virtue Ethics entails that a person's character attributes are the primary aspect of moral evaluation. Perhaps a virtuous character would eat meat in many circumstances (like<mask> he knew it wasn't prepared irresponsibly with extravagant suffering, for example)<mask> wouldn't have sex with animals. (Perhaps<mask> they believed it destroyed the dignity of both the human and the non-human animal?)</s>
Label encoding: <s> [STARTQ] Okay, so, the best argument I know for it being immoral to have sex with animals is that they are incapable of consent and that therefore sexual behaviour with animals is rape. I cannot imagine that any animal would give its consent to dying just in order to satisfy a person's appetite. Furthermore, I don't see a moral difference between acts of bestiality and some methods of animal husbandry. [ENDQ] [NEWLINE] I don't see the inconsistency. Perhaps sex acts *simply do* require consent, whereas eating non-sapient animals *simply doesn't* require consent. Where is the contradiction? [NEWLINE] [NEWLINE] But here are a few moral frameworks, and arguments from each. Note that it isn't so important whether these frameworks are *true*, only that they are consistent frameworks which justify both beliefs in question. [NEWLINE] [NEWLINE] * Natural Law theories seek to ground norms in their natural *purpose*. The *purpose* for sex might be procreation. Therefore, sex with animals is wrong, since procreation is impossible. (An alternate purpose might be solidifying bonds with one's partner. But, while this purpose may exist between humans, it is dubious it could work between humans and non-human animals. Bestiality is plausibly psychologically harmful.) Carnivorism, on the other hand, doesn't seem impermissible on natural law theory. [NEWLINE] * Consequentialism holds that consequences are the primary aspect of moral consideration. Perhaps eating animals leads to greater net happiness. (Maybe not the current methods of factory farming, but what about responsible free range farming, where animals have good lives before being taken for food?) Perhaps bestiality contributes net negative consequences. Under such conditions (which aren't necessarily that implausible), consequentialism would entail that bestiality is wrong, but eating meat is, in some cases, permissible. [NEWLINE] * Virtue Ethics entails that a person's character attributes are the primary aspect of moral evaluation. Perhaps a virtuous character would eat meat in many circumstances (like if he knew it wasn't prepared irresponsibly with extravagant suffering, for example) but wouldn't have sex with animals. (Perhaps because they believed it destroyed the dignity of both the human and the non-human animal?)</s>
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Define extraterrestrial life? Do you mean life full stop, or intelligent life/more developed life? There's ~~plenty of good evidence that there's microbial life in space~~ evidence that life can survive the harshest conditions of space and theories about spacial origins,<mask> it seems you're more talking about the Fermi Paradox, that somehow there's been no contact with any intelligent life, and<mask> the universe is<mask> vast. [NEWLINE] [NEWLINE] <mask> you're talking about the concept of the universe at large, it is infinite and<mask> there's an infinite probability that there will be life.<mask> you're talking about the observable universe.<mask> there are hundreds of billions of stars and planets in this galaxy, not counting moons or any other space bodies which could support life, at least a hundred billion galaxies... [lets estimate here with these rough numbers]( [URL] /?i=100+billion+*+100+billion+*+%281+%2F+1+trillion%29): *100 billion planets * 100 billion galaxies * ( 1 / 1 trillion chance of life) = 10 billion planets likely to have life.* [NEWLINE] [NEWLINE] With any estimate, even for just the observable universe, the likelihood that there *isn't* any other life bearing planets is near zero. The question becomes "<mask> is the probability that we will come in contact with one of these planets" and that's the basis of the Fermi Paradox. [NEWLINE] [NEWLINE] You<mask> have to account for the fact that<mask> you look out into space with a telescope, you're looking back in time. You're seeing light which has traveled for hundreds of thousands, and millions, and billions of lightyears just to reach us. The blackness of space which should be full of blinding light is the very edge of light and the birth of the universe. There are plenty of planets and stars and galaxies we do not know exist<mask> from our perspective, from the light that is reaching us, they don't<mask> exist. It's a time machine in a lens. The time it took for life to develop on earth could be masked to us on a planet whose light is older still<mask> now bears a vibrant lush nature. </s>
Label encoding: <s>Define extraterrestrial life? Do you mean life full stop, or intelligent life/more developed life? There's ~~plenty of good evidence that there's microbial life in space~~ evidence that life can survive the harshest conditions of space and theories about spacial origins, but it seems you're more talking about the Fermi Paradox, that somehow there's been no contact with any intelligent life, and yet the universe is so vast. [NEWLINE] [NEWLINE] If you're talking about the concept of the universe at large, it is infinite and so there's an infinite probability that there will be life. If you're talking about the observable universe. If there are hundreds of billions of stars and planets in this galaxy, not counting moons or any other space bodies which could support life, at least a hundred billion galaxies... [lets estimate here with these rough numbers]( [URL] /?i=100+billion+*+100+billion+*+%281+%2F+1+trillion%29): *100 billion planets * 100 billion galaxies * ( 1 / 1 trillion chance of life) = 10 billion planets likely to have life.* [NEWLINE] [NEWLINE] With any estimate, even for just the observable universe, the likelihood that there *isn't* any other life bearing planets is near zero. The question becomes " what is the probability that we will come in contact with one of these planets" and that's the basis of the Fermi Paradox. [NEWLINE] [NEWLINE] You also have to account for the fact that when you look out into space with a telescope, you're looking back in time. You're seeing light which has traveled for hundreds of thousands, and millions, and billions of lightyears just to reach us. The blackness of space which should be full of blinding light is the very edge of light and the birth of the universe. There are plenty of planets and stars and galaxies we do not know exist because from our perspective, from the light that is reaching us, they don't yet exist. It's a time machine in a lens. The time it took for life to develop on earth could be masked to us on a planet whose light is older still but now bears a vibrant lush nature. </s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, I am not a parent. Maybe that disqualifies me from making any comments about this matter in the first place. Either way, I am a fairly objective person and I can admit<mask> I am wrong. [NEWLINE] [NEWLINE] I do not buy into the whole argument of 'just<mask> our parents brought us into the world, we owe them our lives.' Whether a child was brought into the world by choice or not, I don't think that being born should impose a debt of respect on the child. [NEWLINE] [NEWLINE] Furthermore,<mask><mask> that this respect needs to be earned. I define respect in this context<mask>'regard for another person's rational ability, trusting that they can admit<mask> they are wrong and that their decisions are well-thought-out.' [NEWLINE] [NEWLINE] This is<mask><mask><mask> that giving the reason '<mask> I said<mask>'is a total cop out.<mask> the parent is not open to having a conversation about the reason for their actions, then I don't think they deserve the child's respect. [NEWLINE] [NEWLINE] Don't get me wrong,<mask><mask> it is crucial for a child to be told<mask> they are wrong<mask> that they don't grow up into narcissistic asshats.<mask>,<mask><mask> that they deserve a logical conversation with a parent until one side admits, of his own accord, that he is in the wrong. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Label encoding: <s>First off, I am not a parent. Maybe that disqualifies me from making any comments about this matter in the first place. Either way, I am a fairly objective person and I can admit when I am wrong. [NEWLINE] [NEWLINE] I do not buy into the whole argument of 'just because our parents brought us into the world, we owe them our lives.' Whether a child was brought into the world by choice or not, I don't think that being born should impose a debt of respect on the child. [NEWLINE] [NEWLINE] Furthermore, I think that this respect needs to be earned. I define respect in this context as'regard for another person's rational ability, trusting that they can admit when they are wrong and that their decisions are well-thought-out.' [NEWLINE] [NEWLINE] This is why I think that giving the reason'because I said so'is a total cop out. If the parent is not open to having a conversation about the reason for their actions, then I don't think they deserve the child's respect. [NEWLINE] [NEWLINE] Don't get me wrong, I think it is crucial for a child to be told when they are wrong so that they don't grow up into narcissistic asshats. However, I think that they deserve a logical conversation with a parent until one side admits, of his own accord, that he is in the wrong. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s><pad>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for responding. [NEWLINE] [NEWLINE] [STARTQ] The state/federal run court system is already<mask> 'biased'<mask> you fear a private system would be. [ENDQ] [NEWLINE] Yes. I hope I didn't intimate that<mask> AC is flawed<mask><mask> the liberal State is not flawed. I'm not arguing that Statism is perfect or even preferable, just that AC is flawed. [NEWLINE] [NEWLINE] [STARTQ] That is freedom. [ENDQ] [NEWLINE] Really,<mask>, businesses and consumers only have one choice: the set of State-endorsed courts. All of these are subject to a higher set of laws that regulate their conduct.<mask><mask><mask> I know no US state can pass laws that directly conflict with Federal laws or the constitution (well, they could pass the laws,<mask> they'd quickly be overturned by federal courts).<mask> in reality you have one monolithic legal system that dictates broad behavior, under which you have a choice of gradients for pursuing litigation. [NEWLINE] [NEWLINE] In a legal system lacking such a governing superstructure, I fail to see<mask> things wouldn't devolve into basically an exploitative game. Sure, you could choose which private court you want to settle a dispute in.<mask><mask> happens<mask> one party to a contract refuses to choose a non-corrupt court? And then<mask> happens<mask> said party is a high-value member of society, e.g. a monopolist or oligarch, and can't be reasonably "shunned' within a society? You end up with a completely stratified society<mask> justice is whatever outcome is agreeable for the party who can pay for judgment, meaning only those with wealth can obtain justice. Justice and wealth become one and the same. [NEWLINE] [NEWLINE] My argument is not<mask> much that AC is "bad" or "won't work", anything like that. Rather<mask><mask> ACistan simply evolves into the functional equivalent of a State,<mask> power is centralized and the majority of people end up subjects. The difference is that a liberal democratic state purportedly revolves around welfare, whereas ACistan revolves around wealth and property. I don't know which is preferable<mask><mask><mask> they're different versions of the same structure. Economic subjection is equally<mask> confining<mask> legal subjection.</s>
Label encoding: <s>Thanks for responding. [NEWLINE] [NEWLINE] [STARTQ] The state/federal run court system is already as 'biased' as you fear a private system would be. [ENDQ] [NEWLINE] Yes. I hope I didn't intimate that because AC is flawed I think the liberal State is not flawed. I'm not arguing that Statism is perfect or even preferable, just that AC is flawed. [NEWLINE] [NEWLINE] [STARTQ] That is freedom. [ENDQ] [NEWLINE] Really, though, businesses and consumers only have one choice: the set of State-endorsed courts. All of these are subject to a higher set of laws that regulate their conduct. As far as I know no US state can pass laws that directly conflict with Federal laws or the constitution (well, they could pass the laws, but they'd quickly be overturned by federal courts). So in reality you have one monolithic legal system that dictates broad behavior, under which you have a choice of gradients for pursuing litigation. [NEWLINE] [NEWLINE] In a legal system lacking such a governing superstructure, I fail to see how things wouldn't devolve into basically an exploitative game. Sure, you could choose which private court you want to settle a dispute in. But what happens when one party to a contract refuses to choose a non-corrupt court? And then what happens when said party is a high-value member of society, e.g. a monopolist or oligarch, and can't be reasonably "shunned' within a society? You end up with a completely stratified society where justice is whatever outcome is agreeable for the party who can pay for judgment, meaning only those with wealth can obtain justice. Justice and wealth become one and the same. [NEWLINE] [NEWLINE] My argument is not so much that AC is "bad" or "won't work", anything like that. Rather I think ACistan simply evolves into the functional equivalent of a State, where power is centralized and the majority of people end up subjects. The difference is that a liberal democratic state purportedly revolves around welfare, whereas ACistan revolves around wealth and property. I don't know which is preferable but I think they're different versions of the same structure. Economic subjection is equally as confining as legal subjection.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't mean breaking the fourth wall in that Frodo turns and talks into the camera, Ferris Beuhler style.  I just mean *I'm* reminded<mask> of<mask> it throws the story off. [NEWLINE] [NEWLINE] I keep coming back to this,<mask> one of my least favorite extended scenes is Gandalf riding to the Minas Tirith library to research<mask> he suspects Frodo's ring is the one ring, in Fellowship. [NEWLINE] [NEWLINE] In the theatrical cut, we get the impression that Gandalf suspects something at Bilbo's party,<mask> we're not sure<mask>.  Then we sort of forget about it<mask> we see Frodo going about normal life, until BAM, Gandalf's on the doorstep, wild-eyed, asking, "Is it secret?? Is it safe??"  We<mask> the viewer are surprised, thrown off guard.  Gandalf left the shire...<mask> is he back<mask> soon?  It's an "oh fuck" moment for Frodo, and for the audience.  Most importantly, **we get to discover<mask> the ring really is at the same time<mask> Frodo does.** [NEWLINE] [NEWLINE] This is a classic story-telling device in movies for a reason.  Think about Neo discovering<mask> the Matrix is.  Would that scene carry<mask> much weight<mask> we the audience already knew the truth, and it was just being explained to Neo?  Of course not.  It's exciting<mask> we're finding out along with Neo.  The extended cut of Fellowship ruins this. [NEWLINE] [NEWLINE] Plus, it ruins<mask> is otherwise an awesome reveal of Minas Tirith in Return of the King.  The audience is like, no big deal, already saw that shit two movies ago. [NEWLINE] [NEWLINE] This is<mask> you'll probably come back and say that it doesn't matter,<mask><mask> a book reader, you already know that Frodo's ring is the one ring. <mask> we have to judge a movie based on<mask> well it tells a story, not<mask> you<mask> an individual viewer already know.  In this instance, the extra scene is not only unnecessary and distracting, it's detrimental to the story telling.</s>
Label encoding: <s>I don't mean breaking the fourth wall in that Frodo turns and talks into the camera, Ferris Beuhler style.  I just mean *I'm* reminded because of how it throws the story off. [NEWLINE] [NEWLINE] I keep coming back to this, but one of my least favorite extended scenes is Gandalf riding to the Minas Tirith library to research when he suspects Frodo's ring is the one ring, in Fellowship. [NEWLINE] [NEWLINE] In the theatrical cut, we get the impression that Gandalf suspects something at Bilbo's party, but we're not sure what.  Then we sort of forget about it as we see Frodo going about normal life, until BAM, Gandalf's on the doorstep, wild-eyed, asking, "Is it secret?? Is it safe??"  We as the viewer are surprised, thrown off guard.  Gandalf left the shire... why is he back so soon?  It's an "oh fuck" moment for Frodo, and for the audience.  Most importantly, **we get to discover what the ring really is at the same time as Frodo does.** [NEWLINE] [NEWLINE] This is a classic story-telling device in movies for a reason.  Think about Neo discovering what the Matrix is.  Would that scene carry as much weight if we the audience already knew the truth, and it was just being explained to Neo?  Of course not.  It's exciting because we're finding out along with Neo.  The extended cut of Fellowship ruins this. [NEWLINE] [NEWLINE] Plus, it ruins what is otherwise an awesome reveal of Minas Tirith in Return of the King.  The audience is like, no big deal, already saw that shit two movies ago. [NEWLINE] [NEWLINE] This is where you'll probably come back and say that it doesn't matter, since as a book reader, you already know that Frodo's ring is the one ring.  But we have to judge a movie based on how well it tells a story, not what you as an individual viewer already know.  In this instance, the extra scene is not only unnecessary and distracting, it's detrimental to the story telling.</s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it cannot be manually driven I don't really see<mask> the point<mask> software failures, malfunctions or whatever would mean you can't go anywhere. [ENDQ] [NEWLINE] Just like things on regular cars can malfunction<mask> well to the point that the car can't be driven. [NEWLINE] [NEWLINE] [STARTQ] It seems like google is pumping a fair amount of money into this<mask> there are more important issues. Say<mask> they had focused on inventing a zero carbon emission car or whatever. [ENDQ] [NEWLINE] Google is a software company, meaning that the expertise of the company itself isn't to create a machine with lower emissions<mask> rather to create software which has various uses. Other car companies are working on creating zero-emission cars which one day may be able to use the software that Google creates<mask> that we can have zero-emitting self-driving cars. Saying that Google should be working on a zero-emission car instead of self-driving software is like saying that an electrician should fix your pipes<mask> a plumber rewires your house. [NEWLINE] [NEWLINE] [STARTQ] It just doesn't seem like something we need. [ENDQ] [NEWLINE] <mask> should all companies that make things that we don't need stop their production of these things? For example, Sharp makes TVs. We really don't need TVs,<mask> should Sharp stop making them? Should Bose stop making headphones<mask> we really don't need them? [NEWLINE] [NEWLINE] Aside from these points, there are various positives to self-driving cars. Once the software is perfected to the point that it is ready for consumers to purchase and its use is widespread, we will most likely see a decrease in car accidents<mask><mask><mask> of driver error. You won't have to worry about someone falling asleep at the wheel, someone getting road rage, or drunk drivers<mask> cars will be driving themselves. Aside from this,<mask> every car is connected through a computer software, each car can take different routes to the same destinations to lower traffic. For example, instead of having everyone cram onto a highway during rush hour going in or out of a major city, self-driving cars can all determine<mask> the best possible route is for them based on the routes of other self-driving cars.</s>
Label encoding: <s> [STARTQ] if it cannot be manually driven I don't really see why the point as software failures, malfunctions or whatever would mean you can't go anywhere. [ENDQ] [NEWLINE] Just like things on regular cars can malfunction as well to the point that the car can't be driven. [NEWLINE] [NEWLINE] [STARTQ] It seems like google is pumping a fair amount of money into this when there are more important issues. Say if they had focused on inventing a zero carbon emission car or whatever. [ENDQ] [NEWLINE] Google is a software company, meaning that the expertise of the company itself isn't to create a machine with lower emissions but rather to create software which has various uses. Other car companies are working on creating zero-emission cars which one day may be able to use the software that Google creates so that we can have zero-emitting self-driving cars. Saying that Google should be working on a zero-emission car instead of self-driving software is like saying that an electrician should fix your pipes while a plumber rewires your house. [NEWLINE] [NEWLINE] [STARTQ] It just doesn't seem like something we need. [ENDQ] [NEWLINE] So should all companies that make things that we don't need stop their production of these things? For example, Sharp makes TVs. We really don't need TVs, so should Sharp stop making them? Should Bose stop making headphones since we really don't need them? [NEWLINE] [NEWLINE] Aside from these points, there are various positives to self-driving cars. Once the software is perfected to the point that it is ready for consumers to purchase and its use is widespread, we will most likely see a decrease in car accidents as a result of driver error. You won't have to worry about someone falling asleep at the wheel, someone getting road rage, or drunk drivers since cars will be driving themselves. Aside from this, if every car is connected through a computer software, each car can take different routes to the same destinations to lower traffic. For example, instead of having everyone cram onto a highway during rush hour going in or out of a major city, self-driving cars can all determine what the best possible route is for them based on the routes of other self-driving cars.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your problem is fixable by yourself. Special Needs kids can't do anything about their problems. You complain that you're bored by the curriculum,<mask><mask> you don't understand (and I quickly learned<mask> a smart kid who stopped trying around 8th grade, now going into senior year) is that<mask> you don't learn<mask> to learn, you will be lost. You have to learn<mask> to take notes (sounds easy, not really) and<mask> to grasp concepts in different ways or else, like me, you will quickly find yourself behind and confused on<mask> other kids who aren't<mask> smart are getting better grades and even understand the concepts more than you do. You<mask> seemingly have the added problem of sounding kinda bratty about your "gifted" status. That will not bode well in your future.<mask> a kid who works 3 jobs, all very legitimate (one with a sports team, one with a startup, and one teaching at my religious institution), I feel that the reason my jobs love me<mask> my promptness, creativity, and reliability is that I am personable. I show my intelligence<mask> don't ever flaunt or boast about myself. You will learn that. It's just a matter of<mask>,<mask> hopefully you do before you go into the working world. I see kids who have a hubris about their intelligence in my jobs and they are the worst kind of people and everybody hates them. [NEWLINE] [NEWLINE] In short, you will have to learn at the very least for college<mask> not high school,<mask> to learn and take notes. I understand that your curriculum is not challenging enough for you,<mask> you are capable of teaching yourself<mask> you want, something special needs kids cannot do. You may like to think that it's unfair that you are expected to learn on the side and the other kids are not,<mask> you<mask>,<mask> you said, are "gifted". Use your gift that the special needs kids do not have. Don't complain, put your head down, and be the best you you can be. One day, you'll find the world catches the fuck up with you and<mask><mask> speeds right by you.</s><pad>
Label encoding: <s>Your problem is fixable by yourself. Special Needs kids can't do anything about their problems. You complain that you're bored by the curriculum, but what you don't understand (and I quickly learned as a smart kid who stopped trying around 8th grade, now going into senior year) is that if you don't learn how to learn, you will be lost. You have to learn how to take notes (sounds easy, not really) and how to grasp concepts in different ways or else, like me, you will quickly find yourself behind and confused on how other kids who aren't as smart are getting better grades and even understand the concepts more than you do. You also seemingly have the added problem of sounding kinda bratty about your "gifted" status. That will not bode well in your future. As a kid who works 3 jobs, all very legitimate (one with a sports team, one with a startup, and one teaching at my religious institution), I feel that the reason my jobs love me besides my promptness, creativity, and reliability is that I am personable. I show my intelligence but don't ever flaunt or boast about myself. You will learn that. It's just a matter of when, but hopefully you do before you go into the working world. I see kids who have a hubris about their intelligence in my jobs and they are the worst kind of people and everybody hates them. [NEWLINE] [NEWLINE] In short, you will have to learn at the very least for college if not high school, how to learn and take notes. I understand that your curriculum is not challenging enough for you, but you are capable of teaching yourself if you want, something special needs kids cannot do. You may like to think that it's unfair that you are expected to learn on the side and the other kids are not, but you also, as you said, are "gifted". Use your gift that the special needs kids do not have. Don't complain, put your head down, and be the best you you can be. One day, you'll find the world catches the fuck up with you and in fact speeds right by you.</s><pad>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, a fair warning. I will not engage with you<mask> you continue to use the tone you've presented in the body of your post. Speak with courtesy<mask> you wish us to speak. [NEWLINE] [NEWLINE] Most people with any sense hold views *given some exogenous assumptions about society*. Their position is usually not something like "capital punishment is wrong in any and all situations",<mask> closer to "capital punishment is wrong given the current systems and resources we have access to<mask> a society." This is *not* the same<mask> giving up one's ideals,<mask> gun control, social justice, environmentalism and<mask> on are not actually the ideals at play here. [NEWLINE] [NEWLINE] <mask> *is* at play are certain basal ideas about society. For example, a person supports gun control not<mask> they think guns are inherently bad<mask><mask> they think guns increase the rate of homicide in contemporary society (I'm am not saying that they do or do not, myself) and think that *homicide* is inherently bad. You're talking about policy prescriptions<mask> the relevant issue is people's underlying assumptions about<mask> is good and<mask> is bad. Surely there are some people who believe the greatest good lies in environmentalism or pacifism and would refuse to consider otherwise,<mask> these people are outliers. [NEWLINE] [NEWLINE] An analogy. I try to eat the best quality food whenever I can. I prepare most meals myself and haven't eaten fast food in nearly a decade.<mask> I purchase meat I do<mask> from an ethical butcher. This is not some lofty philosophical conviction I have. It is a response to certain axiomatic assumptions I make about myself and the world and my current situation. It is my *best response to my current environment*.<mask> that environment changed,<mask> the apocalypse occurred for example, I would have no qualms about butchering one of the wild jackrabbits that live around Calgary (<mask> I live) for food. [NEWLINE] [NEWLINE] Finally, I'm not exactly sure<mask> the point of this view is. Speaking honestly, cannibals would probably fare better than other people, all else being equal. That doesn't mean cannibalism has value in contemporary society. </s>
Label encoding: <s>First off, a fair warning. I will not engage with you if you continue to use the tone you've presented in the body of your post. Speak with courtesy if you wish us to speak. [NEWLINE] [NEWLINE] Most people with any sense hold views *given some exogenous assumptions about society*. Their position is usually not something like "capital punishment is wrong in any and all situations", but closer to "capital punishment is wrong given the current systems and resources we have access to as a society." This is *not* the same as giving up one's ideals, because gun control, social justice, environmentalism and so on are not actually the ideals at play here. [NEWLINE] [NEWLINE] What *is* at play are certain basal ideas about society. For example, a person supports gun control not because they think guns are inherently bad but because they think guns increase the rate of homicide in contemporary society (I'm am not saying that they do or do not, myself) and think that *homicide* is inherently bad. You're talking about policy prescriptions when the relevant issue is people's underlying assumptions about what is good and what is bad. Surely there are some people who believe the greatest good lies in environmentalism or pacifism and would refuse to consider otherwise, but these people are outliers. [NEWLINE] [NEWLINE] An analogy. I try to eat the best quality food whenever I can. I prepare most meals myself and haven't eaten fast food in nearly a decade. When I purchase meat I do so from an ethical butcher. This is not some lofty philosophical conviction I have. It is a response to certain axiomatic assumptions I make about myself and the world and my current situation. It is my *best response to my current environment*. If that environment changed, if the apocalypse occurred for example, I would have no qualms about butchering one of the wild jackrabbits that live around Calgary ( where I live) for food. [NEWLINE] [NEWLINE] Finally, I'm not exactly sure what the point of this view is. Speaking honestly, cannibals would probably fare better than other people, all else being equal. That doesn't mean cannibalism has value in contemporary society. </s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**Amelia (birth defect)**]( [URL] %20%28birth%20defect%29): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] [STARTQ] __Amelia__ (from [Greek]( [URL] ) ἀ- "lack of" plus μέλος (plural: μέλεα or μέλη) "limb") is the [birth defect]( [URL] ) of lacking one or more [limbs]( [URL] (anatomy\)). It can<mask> result in a shrunken or deformed limb. For example, a child might be born without an elbow or forearm. The term may be modified to indicate the number of legs or arms missing at birth, such<mask> tetra-amelia for the absence of all four limbs. A related term is [meromelia]( [URL] ), which is the partial absence of a limb or limbs. [ENDQ] [NEWLINE] [STARTQ] ==== [ENDQ] [NEWLINE] [STARTQ] [**Image**]( [URL].jpg) [^(i)]( [URL] :Amelia_right_forearm.jpg) [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Thalidomide]( [URL] ) ^| [^Meromelia]( [URL] ) ^| [^Lord ^Byron]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cjb5hv2) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cjb5hv2)^. ^Will ^<mask> ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Label encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**Amelia (birth defect)**]( [URL] %20%28birth%20defect%29): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] [STARTQ] __Amelia__ (from [Greek]( [URL] ) ἀ- "lack of" plus μέλος (plural: μέλεα or μέλη) "limb") is the [birth defect]( [URL] ) of lacking one or more [limbs]( [URL] (anatomy\)). It can also result in a shrunken or deformed limb. For example, a child might be born without an elbow or forearm. The term may be modified to indicate the number of legs or arms missing at birth, such as tetra-amelia for the absence of all four limbs. A related term is [meromelia]( [URL] ), which is the partial absence of a limb or limbs. [ENDQ] [NEWLINE] [STARTQ] ==== [ENDQ] [NEWLINE] [STARTQ] [**Image**]( [URL].jpg) [^(i)]( [URL] :Amelia_right_forearm.jpg) [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Thalidomide]( [URL] ) ^| [^Meromelia]( [URL] ) ^| [^Lord ^Byron]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cjb5hv2) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cjb5hv2)^. ^Will ^ also ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a drug and alcohol abuse counselor it's probably worth pointing out that the rate of violence and acts of aggression<mask> under the influence is far more likely with alcohol than virtually any other drug in existence. [NEWLINE] [NEWLINE] [STARTQ] Boyum, D., and Kleiman, M. Alcohol and other drugs. In Wilson, J.Q., and Petersilia, J., eds. Crime. [ENDQ] San Francisco: ICS Press, 1995. [NEWLINE] [NEWLINE] In terms of dependency, it's one of very few drugs which creates a level of dependency in which the withdrawal of the chemical can actually kill the person discontinuing its use. The same can't be said of heroin, cocaine, or methamphetamines,<mask> there's generally a cultural bias that these drugs are "more addictive" than alcohol. [NEWLINE] [NEWLINE] I'm in the same camp<mask> OP in terms of not looking to regulate alcohol any more than it already is. The vast majority of people who use alcohol do use it responsibly.<mask> I am opposed to is the idea that alcohol is somehow "different than" other drugs. It's not. [NEWLINE] [NEWLINE] <mask> I used to work with middle school students and I would ask them to identify all the drugs they could think of, alcohol rarely made the list. Ironically, tobacco would almost always make that list,<mask> alcohol was frequently absent. Often,<mask> I would suggest (at the end of the activity) that alcohol might be added to the list, there would often be protest from someone arguing that alcohol was not,<mask><mask>, a drug at all.<mask> it generated some interesting conversation, it's worth noting that people do tend to mentally hold alcohol<mask> being set apart from other substances. [NEWLINE] [NEWLINE] That's anecdotal,<mask> take it with a grain of salt,<mask><mask><mask> it does say something worth looking at about<mask> we view drugs and alcohol (at least in America).<mask> I'm certainly open to less regulation in terms of most substances, I would much prefer living in a world and society that views policy in terms of the evidence, rather than our feelings, intuitions, history, or cultural biases about a drug (or anything else for that matter).</s>
Label encoding: <s>As a drug and alcohol abuse counselor it's probably worth pointing out that the rate of violence and acts of aggression while under the influence is far more likely with alcohol than virtually any other drug in existence. [NEWLINE] [NEWLINE] [STARTQ] Boyum, D., and Kleiman, M. Alcohol and other drugs. In Wilson, J.Q., and Petersilia, J., eds. Crime. [ENDQ] San Francisco: ICS Press, 1995. [NEWLINE] [NEWLINE] In terms of dependency, it's one of very few drugs which creates a level of dependency in which the withdrawal of the chemical can actually kill the person discontinuing its use. The same can't be said of heroin, cocaine, or methamphetamines, though there's generally a cultural bias that these drugs are "more addictive" than alcohol. [NEWLINE] [NEWLINE] I'm in the same camp as OP in terms of not looking to regulate alcohol any more than it already is. The vast majority of people who use alcohol do use it responsibly. What I am opposed to is the idea that alcohol is somehow "different than" other drugs. It's not. [NEWLINE] [NEWLINE] When I used to work with middle school students and I would ask them to identify all the drugs they could think of, alcohol rarely made the list. Ironically, tobacco would almost always make that list, but alcohol was frequently absent. Often, when I would suggest (at the end of the activity) that alcohol might be added to the list, there would often be protest from someone arguing that alcohol was not, in fact, a drug at all. While it generated some interesting conversation, it's worth noting that people do tend to mentally hold alcohol as being set apart from other substances. [NEWLINE] [NEWLINE] That's anecdotal, so take it with a grain of salt, but I think it does say something worth looking at about how we view drugs and alcohol (at least in America). While I'm certainly open to less regulation in terms of most substances, I would much prefer living in a world and society that views policy in terms of the evidence, rather than our feelings, intuitions, history, or cultural biases about a drug (or anything else for that matter).</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hey, I'm a biostatistics researcher and I analyze this kind of stuff. I don't think this will change your view,<mask> I just want to walk through the stats part with you<mask> you seem quite interested in the hard, technical point of view.<mask><mask> you're making a common mistake - mixing absolute proportions with relative proportions. [NEWLINE] [NEWLINE] Let's say that we have a population of 1000 people. Let's make 33% of them 'Arabs', and 66% of them anything else. Three Arabs in this population are terrorists (which is ~1%, giving ALOT of credit), and there is 1 non-Arab terrorist. Yeah, I'm not being politically correct,<mask> I don't really give a shit. I'm gonna put them in a table like this. [NEWLINE] [NEWLINE] [URL].png [NEWLINE] [NEWLINE] <mask> we have 3 times<mask> many Arab terrorists<mask> Asians, Whites.. whatever. On top of that,<mask> you're an Arab, you're 6 times<mask> likely to be a terrorist than any other race. You can see<mask> from that first table,<mask> I can show you that by dividing each cell by the total number of people in that race - 3/333, 330/333, 1/667, 666/667. You can check the math<mask> you want. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [NEWLINE] <mask> wait, those percentages are pretty low, aren't they? Look just at each column individually.<mask> 0.9% of Arabs are terrorists, and 0.15% of other races are terrorists.<mask> I tell you someone is an Arab,<mask> much more likely are they to be a terrorist than an Asian, Caucasian, Mexican, Black? That's ~~0.15% minus 0.9%... About 0.6%.~~ 0.9% - 0.15%... **About 0.75%** (I got 0.6 earlier<mask> I can't math). That's all the information someone being Arabic gives you. Remember, this is<mask><mask> 1 in 100 Arabs are terrorists. In real life, the detection power - especially for a case worthy of FBI investigation - is even lower.</s>
Label encoding: <s>Hey, I'm a biostatistics researcher and I analyze this kind of stuff. I don't think this will change your view, but I just want to walk through the stats part with you because you seem quite interested in the hard, technical point of view. I think you're making a common mistake - mixing absolute proportions with relative proportions. [NEWLINE] [NEWLINE] Let's say that we have a population of 1000 people. Let's make 33% of them 'Arabs', and 66% of them anything else. Three Arabs in this population are terrorists (which is ~1%, giving ALOT of credit), and there is 1 non-Arab terrorist. Yeah, I'm not being politically correct, but I don't really give a shit. I'm gonna put them in a table like this. [NEWLINE] [NEWLINE] [URL].png [NEWLINE] [NEWLINE] So we have 3 times as many Arab terrorists as Asians, Whites.. whatever. On top of that, if you're an Arab, you're 6 times as likely to be a terrorist than any other race. You can see why from that first table, but I can show you that by dividing each cell by the total number of people in that race - 3/333, 330/333, 1/667, 666/667. You can check the math if you want. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [NEWLINE] But wait, those percentages are pretty low, aren't they? Look just at each column individually. So 0.9% of Arabs are terrorists, and 0.15% of other races are terrorists. If I tell you someone is an Arab, how much more likely are they to be a terrorist than an Asian, Caucasian, Mexican, Black? That's ~~0.15% minus 0.9%... About 0.6%.~~ 0.9% - 0.15%... **About 0.75%** (I got 0.6 earlier because I can't math). That's all the information someone being Arabic gives you. Remember, this is given that 1 in 100 Arabs are terrorists. In real life, the detection power - especially for a case worthy of FBI investigation - is even lower.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What are you 'not really'-ing here? The guy you're responding to is saying the comic is taking into account that any password made of common English words can be attacked with a dictionary search. Which is completely true. [NEWLINE] [NEWLINE] The comic is assuming any attack against the "horse staple" password uses a dictionary search across the 2048 most common words in the English language. Randall's point isn't that this is harder than the equivalent length password in random gibberish, it's that even something<mask> seemingly simple is beyond the point of getting any marginal return for additional complexity, simply by virtue of having four such words in the password. [NEWLINE] [NEWLINE] Let each of the four common words be drawn from a dictionary of size 2^11, or 2048 possibilities. Take four of these,<mask> the total search space becomes 2^44, or 1.76x10^13. Allow the brute force dictionary search to make 1000 password attempts per second. It will exhaust the possibility space in 1.76x10^10 seconds, which is more than 557 years. [NEWLINE] [NEWLINE] The point isn't that this is harder to brute force than random garbage, it's that it's *sufficiently* complex that it resists brute force attacks. The advantage of the "horse staple" algorithm is that it's easier to use and vastly less likely to be written down somewhere or to frustrate users into using common passwords, which are a far more real threat than a brute force attack running over the course of a hundred years. [NEWLINE] [NEWLINE] Telling people they should be using random garbage is all well and good for the small fraction who listen to you and manage to memorize that,<mask> it's counterproductive for all those who find that ridiculously user-unfriendly and end up going with "passw0rd" instead. And<mask> the comic notes is that "tr0oubador$" style passwords, which try to compromise memorability with complexity, are both harder to remember and *less* brute force resistant than "horse staple" passwords,<mask> the algorithm attacking the seemingly complex password uses an intelligent dictionary attack that mutates the words to resemble passwords containing letters and numbers.</s>
Label encoding: <s>What are you 'not really'-ing here? The guy you're responding to is saying the comic is taking into account that any password made of common English words can be attacked with a dictionary search. Which is completely true. [NEWLINE] [NEWLINE] The comic is assuming any attack against the "horse staple" password uses a dictionary search across the 2048 most common words in the English language. Randall's point isn't that this is harder than the equivalent length password in random gibberish, it's that even something so seemingly simple is beyond the point of getting any marginal return for additional complexity, simply by virtue of having four such words in the password. [NEWLINE] [NEWLINE] Let each of the four common words be drawn from a dictionary of size 2^11, or 2048 possibilities. Take four of these, so the total search space becomes 2^44, or 1.76x10^13. Allow the brute force dictionary search to make 1000 password attempts per second. It will exhaust the possibility space in 1.76x10^10 seconds, which is more than 557 years. [NEWLINE] [NEWLINE] The point isn't that this is harder to brute force than random garbage, it's that it's *sufficiently* complex that it resists brute force attacks. The advantage of the "horse staple" algorithm is that it's easier to use and vastly less likely to be written down somewhere or to frustrate users into using common passwords, which are a far more real threat than a brute force attack running over the course of a hundred years. [NEWLINE] [NEWLINE] Telling people they should be using random garbage is all well and good for the small fraction who listen to you and manage to memorize that, but it's counterproductive for all those who find that ridiculously user-unfriendly and end up going with "passw0rd" instead. And what the comic notes is that "tr0oubador$" style passwords, which try to compromise memorability with complexity, are both harder to remember and *less* brute force resistant than "horse staple" passwords, if the algorithm attacking the seemingly complex password uses an intelligent dictionary attack that mutates the words to resemble passwords containing letters and numbers.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> are your goals? [ENDQ] [NEWLINE] Currently, my primary goal is to get into medical school and keep engineering<mask> a backup. There's a good chance I might take a gap year and work an engineering job to save up some money and/or simply to strengthen my medical school application [NEWLINE] [NEWLINE] [STARTQ] <mask> you're going for premed, all medical schools care about are your GPA and MCAT. They don't care at all about<mask> many classes taken or<mask> many degrees you've gotten. [ENDQ] [NEWLINE] This is<mask> I'm torn about. On one hand, I can simply focus on doing the least<mask> best work I can to maximize my chances of getting into med school. Or I can explore my interests and work towards an additional goal of earning the degrees<mask> a form of personal satisfaction. I can just imagine right now<mask> amazing it would feel to graduate with 4 degrees in 4 years, even<mask> it doesn't amount to anything<mask><mask><mask> jobs go. The physical degree itself provides me with a concrete reward and structure for the hard work I will put in. This isn't even mentioning that I have a genuine interest in taking these courses, I'm not merely doing it for the piece of paper. Otherwise, I would simply get degrees in liberal arts or something [NEWLINE] [NEWLINE] [STARTQ] <mask> you're looking for work in engineering, internships are an absolute must...<mask> not internships, then at least undergraduate research is needed. [ENDQ] [NEWLINE] Yes, I am<mask> paying attention to these aspects. I got a biomedical engineering summer research position at my university<mask> I mainly designed a new mechatronic medical tool for a certain medical procedure<mask> I can use that experience under both engineering and bio/chem fields. [NEWLINE] [NEWLINE] I'm<mask> currently in the process of obtaining a pure bio/chem part-time research position in a lab during the school year (yes, this is on top of my insane amount of credits. I do have time<mask> I have 3 day weekends, no Friday classes).<mask>, I may have another research position lined up for next summer (still need confirmation)<mask> I am seeking engineering opportunities too just to see<mask> I can get.</s>
Label encoding: <s> [STARTQ] What are your goals? [ENDQ] [NEWLINE] Currently, my primary goal is to get into medical school and keep engineering as a backup. There's a good chance I might take a gap year and work an engineering job to save up some money and/or simply to strengthen my medical school application [NEWLINE] [NEWLINE] [STARTQ] If you're going for premed, all medical schools care about are your GPA and MCAT. They don't care at all about how many classes taken or how many degrees you've gotten. [ENDQ] [NEWLINE] This is what I'm torn about. On one hand, I can simply focus on doing the least but best work I can to maximize my chances of getting into med school. Or I can explore my interests and work towards an additional goal of earning the degrees as a form of personal satisfaction. I can just imagine right now how amazing it would feel to graduate with 4 degrees in 4 years, even if it doesn't amount to anything as far as jobs go. The physical degree itself provides me with a concrete reward and structure for the hard work I will put in. This isn't even mentioning that I have a genuine interest in taking these courses, I'm not merely doing it for the piece of paper. Otherwise, I would simply get degrees in liberal arts or something [NEWLINE] [NEWLINE] [STARTQ] If you're looking for work in engineering, internships are an absolute must... If not internships, then at least undergraduate research is needed. [ENDQ] [NEWLINE] Yes, I am also paying attention to these aspects. I got a biomedical engineering summer research position at my university where I mainly designed a new mechatronic medical tool for a certain medical procedure so I can use that experience under both engineering and bio/chem fields. [NEWLINE] [NEWLINE] I'm also currently in the process of obtaining a pure bio/chem part-time research position in a lab during the school year (yes, this is on top of my insane amount of credits. I do have time because I have 3 day weekends, no Friday classes). Additionally, I may have another research position lined up for next summer (still need confirmation) but I am seeking engineering opportunities too just to see what I can get.</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You semi-acknowledge that every generation considers the next one "soft".  I know my father's generation did, and now I'm a parent of a 17 year old, and they seem to have things easier that we did.  Of course, isn't that the point?  To make the world better for our children? [NEWLINE] [NEWLINE] <mask>, kids who have been working in HS need to be forced to undergo your "lesson" in<mask> to be productive? <mask>? [NEWLINE] [NEWLINE] It's hard to understand your "kids with their constant phones and social media" gripe. <mask> I was young, kids were constantly on their phones- I recall a lot of houses have a second "kids line"<mask> they were on the phone<mask> much.  Teens socialize.  It's the right and proper thing to do developmentally.  Would it be better for them to do it via landline phones instead of mobile or through social media? [NEWLINE] [NEWLINE] Now lets consider the impact of your "productive" work on the economy.  Putting aside the fact that it takes time and training to become good at construction, cleanup, social work or tutoring.  [And I have to pause to ask,<mask> you really think that it's a good idea to take an 18 year old with no training who is there under duress getting paid minimum wage (<mask><mask><mask> they had been working through school they might be making more) and put them with vulnerable people needing a social worker? ] <mask> do you think would happen to the people who would traditionally be doing those jobs at more than minimum wage?  Yup, they'd be fired (<mask> you specified this is work that needed to be done, not just make work). [NEWLINE] [NEWLINE] Finally,<mask> is the incentive for the workers to work well?  They can't get a raise.  They can't get fired.  They are just spending a year<mask> an indentured servitude to teach them... that the best way to be responsible is to have your freedom taken away from you? <mask> does this resemble "real life" in any way?</s>
Label encoding: <s>You semi-acknowledge that every generation considers the next one "soft".  I know my father's generation did, and now I'm a parent of a 17 year old, and they seem to have things easier that we did.  Of course, isn't that the point?  To make the world better for our children? [NEWLINE] [NEWLINE] So, kids who have been working in HS need to be forced to undergo your "lesson" in how to be productive?  Why? [NEWLINE] [NEWLINE] It's hard to understand your "kids with their constant phones and social media" gripe.  When I was young, kids were constantly on their phones- I recall a lot of houses have a second "kids line" because they were on the phone so much.  Teens socialize.  It's the right and proper thing to do developmentally.  Would it be better for them to do it via landline phones instead of mobile or through social media? [NEWLINE] [NEWLINE] Now lets consider the impact of your "productive" work on the economy.  Putting aside the fact that it takes time and training to become good at construction, cleanup, social work or tutoring.  [And I have to pause to ask, if you really think that it's a good idea to take an 18 year old with no training who is there under duress getting paid minimum wage ( even though if they had been working through school they might be making more) and put them with vulnerable people needing a social worker? ]  What do you think would happen to the people who would traditionally be doing those jobs at more than minimum wage?  Yup, they'd be fired ( since you specified this is work that needed to be done, not just make work). [NEWLINE] [NEWLINE] Finally, what is the incentive for the workers to work well?  They can't get a raise.  They can't get fired.  They are just spending a year as an indentured servitude to teach them... that the best way to be responsible is to have your freedom taken away from you?  How does this resemble "real life" in any way?</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to boil down your post down to 2 main points: [NEWLINE] [NEWLINE] 1 - You think it's okay to use stereotypes to create the basis for a character. [NEWLINE] [NEWLINE] 2 - You don't think you need to worry about the minority of your audience. [NEWLINE] [NEWLINE] For point 1, the problem isn't really stereotyping.  Yeah, stereotyping is bad, whatever.  The main thing is<mask>, it just feels like cheap storytelling.  Your main character should be built from the ground up, not off a stereotype. <mask> the character *happens* to be black, fine,<mask><mask> you go "Well, it's a gang,<mask> let's make them black" before considering the character's, well, character then you're building a character off of something that's already broken.  Same goes for representing women.  The woman probably doesn't *need* to be represented<mask> super sexy.  To represent her<mask> such<mask> her character wouldn't dress that way is just a cheap character build to get sales from teenage boys.  You *can* make sexy characters that are<mask> strong,<mask> it should be a part of their character. [NEWLINE] [NEWLINE] Compare the number of games out there with pointless<mask> sexy guy characters to the number with pointless<mask> sexy lady characters.  Clearly, there's some cheap character building going on with the women. [NEWLINE] [NEWLINE] [NEWLINE] <mask> for point 2.  You should include minorities in games for 2 reasons.  First off, it makes your fans happy. It's the same reason that TV shows and movies have minority characters. Most people don't mind,<mask> it gives those minorities something to finally relate to in a main character.  It feels nice. The second reason is that it allows for more interesting story-telling.<mask> your characters are all monochromatic, you lose a lot of potential drama and character growth.  For example, you could make a game about the civil war and have all-white troops,<mask> then it would basically be CoD again. Add some black troops to the units, and it becomes much more interesting story-telling.</s>
Label encoding: <s>I'm going to boil down your post down to 2 main points: [NEWLINE] [NEWLINE] 1 - You think it's okay to use stereotypes to create the basis for a character. [NEWLINE] [NEWLINE] 2 - You don't think you need to worry about the minority of your audience. [NEWLINE] [NEWLINE] For point 1, the problem isn't really stereotyping.  Yeah, stereotyping is bad, whatever.  The main thing is though, it just feels like cheap storytelling.  Your main character should be built from the ground up, not off a stereotype.  If the character *happens* to be black, fine, but if you go "Well, it's a gang, so let's make them black" before considering the character's, well, character then you're building a character off of something that's already broken.  Same goes for representing women.  The woman probably doesn't *need* to be represented as super sexy.  To represent her as such when her character wouldn't dress that way is just a cheap character build to get sales from teenage boys.  You *can* make sexy characters that are also strong, but it should be a part of their character. [NEWLINE] [NEWLINE] Compare the number of games out there with pointless but sexy guy characters to the number with pointless but sexy lady characters.  Clearly, there's some cheap character building going on with the women. [NEWLINE] [NEWLINE] [NEWLINE] As for point 2.  You should include minorities in games for 2 reasons.  First off, it makes your fans happy. It's the same reason that TV shows and movies have minority characters. Most people don't mind, but it gives those minorities something to finally relate to in a main character.  It feels nice. The second reason is that it allows for more interesting story-telling. If your characters are all monochromatic, you lose a lot of potential drama and character growth.  For example, you could make a game about the civil war and have all-white troops, but then it would basically be CoD again. Add some black troops to the units, and it becomes much more interesting story-telling.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> a higher wage does is it increases the buying power of those who get it. In this case, it's poor people. Poor people have a lot of pent-up demand.<mask> they had more money, they would SPEND it. [ENDQ] [NEWLINE] That doesn't make any sense.<mask> workers get more money, capitalists get less. Demand would not increase,<mask> redistributing income does not create purchasing power. It just moves it from one person to another. It's<mask> not poor people who would get the money. It would be skilled workers. [NEWLINE] [NEWLINE] [STARTQ] <mask> for unions, you completely missed the point of my example.<mask> unions raise the general wage that people expect for their labor, then even NON-union wages go up. [ENDQ] [NEWLINE] <mask> they don't raise the general wage. They only raise the wages of people in unions. [NEWLINE] [NEWLINE] [STARTQ] The reduction in the number of jobs you are concerned about is true,<mask> only in a limited number of circumstances.<mask> there is more demand than production capacity, then higher wages stunt jobs. [ENDQ] [NEWLINE] No, that's not<mask> that works.<mask> there's extra unemployment, then there is extra production capacity.<mask> higher wages are due to demand outstripping supply, then unions have nothing to do with it. Unions raise wages by lowering supply, not by increasing demand. [NEWLINE] [NEWLINE] [STARTQ] For all of the 2000's, wages were stagnant, [ENDQ] [NEWLINE] This is false. [NEWLINE] [NEWLINE] [STARTQ] and the growth of the economy was based entirely on people BORROWING money and spending it. Once the credit stopped, the economy crashed. [ENDQ] [NEWLINE] That's normal. [NEWLINE] [NEWLINE] [STARTQ] Spending has got to be balanced with pay.<mask> you want spending to go up, people have GOT to make more money. You can temporarily mess with that fundamental fact,<mask> only by paying a very heavy toll later on. [ENDQ] [NEWLINE] You're not talking about people making more money, you're talking about redistributing it from capitalists to workers in unions. And you're<mask> talking about redistributing some money from the poor to unionized workers.</s>
Label encoding: <s> [STARTQ] What a higher wage does is it increases the buying power of those who get it. In this case, it's poor people. Poor people have a lot of pent-up demand. If they had more money, they would SPEND it. [ENDQ] [NEWLINE] That doesn't make any sense. If workers get more money, capitalists get less. Demand would not increase, because redistributing income does not create purchasing power. It just moves it from one person to another. It's also not poor people who would get the money. It would be skilled workers. [NEWLINE] [NEWLINE] [STARTQ] As for unions, you completely missed the point of my example. When unions raise the general wage that people expect for their labor, then even NON-union wages go up. [ENDQ] [NEWLINE] But they don't raise the general wage. They only raise the wages of people in unions. [NEWLINE] [NEWLINE] [STARTQ] The reduction in the number of jobs you are concerned about is true, but only in a limited number of circumstances. If there is more demand than production capacity, then higher wages stunt jobs. [ENDQ] [NEWLINE] No, that's not how that works. If there's extra unemployment, then there is extra production capacity. If higher wages are due to demand outstripping supply, then unions have nothing to do with it. Unions raise wages by lowering supply, not by increasing demand. [NEWLINE] [NEWLINE] [STARTQ] For all of the 2000's, wages were stagnant, [ENDQ] [NEWLINE] This is false. [NEWLINE] [NEWLINE] [STARTQ] and the growth of the economy was based entirely on people BORROWING money and spending it. Once the credit stopped, the economy crashed. [ENDQ] [NEWLINE] That's normal. [NEWLINE] [NEWLINE] [STARTQ] Spending has got to be balanced with pay. If you want spending to go up, people have GOT to make more money. You can temporarily mess with that fundamental fact, but only by paying a very heavy toll later on. [ENDQ] [NEWLINE] You're not talking about people making more money, you're talking about redistributing it from capitalists to workers in unions. And you're also talking about redistributing some money from the poor to unionized workers.</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm an atheist. Lately, I've been thinking about<mask> it means to be human, and live life. I've come to the conclusion that a life without the idea of a higher power, and a greater meaning, is ultimately emptier, and less fulfilling in a sense. [NEWLINE] [NEWLINE] Without the idea of a god, or some sort of higher power, we are just coincidences of the development of the universe.<mask> we can give ourselves the illusion of meaning by developing relationships and keeping ourselves busy, we are no more than atoms and molecules that are the result of pure coincidence. [NEWLINE] [NEWLINE] There's more to the thought than that,<mask> perhaps someone can shed a little more light on the matter. CMV. [NEWLINE] [NEWLINE] Edit: I'm really thrilled with all of the responses I've gotten. This entire concept is extremely hard for me to wrap my head around, to be honest. [NEWLINE] [NEWLINE] I need more time to think about it,<mask><mask><mask> I may have been convinced that one can (and even with god, has to) create their own meaning behind life.<mask>,<mask><mask> I'm unconvinced that a life without god can be more fulfilling. It's actually pretty hard to understand even my own thoughts about the matter... [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than just downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I'm an atheist. Lately, I've been thinking about what it means to be human, and live life. I've come to the conclusion that a life without the idea of a higher power, and a greater meaning, is ultimately emptier, and less fulfilling in a sense. [NEWLINE] [NEWLINE] Without the idea of a god, or some sort of higher power, we are just coincidences of the development of the universe. Although we can give ourselves the illusion of meaning by developing relationships and keeping ourselves busy, we are no more than atoms and molecules that are the result of pure coincidence. [NEWLINE] [NEWLINE] There's more to the thought than that, but perhaps someone can shed a little more light on the matter. CMV. [NEWLINE] [NEWLINE] Edit: I'm really thrilled with all of the responses I've gotten. This entire concept is extremely hard for me to wrap my head around, to be honest. [NEWLINE] [NEWLINE] I need more time to think about it, but I think I may have been convinced that one can (and even with god, has to) create their own meaning behind life. However, I think I'm unconvinced that a life without god can be more fulfilling. It's actually pretty hard to understand even my own thoughts about the matter... [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than just downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see that you've changed your view already...<mask> my own position is that complete anonymity unleashes an incredible amount of artistic and intllectual energy. There are costs, of course. Reputation is a system of control, and<mask> people are controlled, that does have some benefits.<mask> most of those benefits only apply to physical interactions (<mask> it is possible to do someone lasting harm), not to exchanges of ideas and information. In public fora, being able to control people affects<mask> they say,<mask> they think, and<mask> they express themselves without having any corresponding benefit in terms of protecting them from each other. [NEWLINE] [NEWLINE] In any event, almost of all<mask> you are discussing is the result of real-identity or mixed eponymous/pseudonymous systems. On a truly anonymous bulletin board, it is impossible to bully someone<mask> you have no way to find him. I don't think it's a coincidence that the people who actually take being the target of verbal abuse seriously prefer to interact in pseudonymous or anonymous settings. [NEWLINE] [NEWLINE] Historically, periods of great cultural productivity were typically periods of unusual freedom from cultural control. People called Athens a cesspool of depravity and admired the disciplined societies of Sparta and Crete,<mask> two thousand years later we look at the Athenians<mask> the founders everything valuable in our civilization - logic, science, mathematics, philosophy, literature. You can say similar things about Renaissance Italy and Enlightenment Germany. In each case, the period of true intellectual flourishing was very short... fifty years or<mask>. It was a little window between<mask> one repressive regime gave up and another repressive regime was ushered in. [NEWLINE] [NEWLINE] <mask><mask> it's surprising that the freest places on the internet have produced the most astonishing contributions to public culture. Astonishing,<mask> not,<mask> I said, unexpected,<mask> this always happens<mask> you unleash people for a few years or a decade.<mask> sadly, instead of pushing anonymity further and further, the trend for the last decade has been to find ways to reproduce old mechanisms of social control on media that were originally mostly anonymous.</s>
Label encoding: <s>I see that you've changed your view already... but my own position is that complete anonymity unleashes an incredible amount of artistic and intllectual energy. There are costs, of course. Reputation is a system of control, and when people are controlled, that does have some benefits. But most of those benefits only apply to physical interactions ( where it is possible to do someone lasting harm), not to exchanges of ideas and information. In public fora, being able to control people affects what they say, how they think, and how they express themselves without having any corresponding benefit in terms of protecting them from each other. [NEWLINE] [NEWLINE] In any event, almost of all what you are discussing is the result of real-identity or mixed eponymous/pseudonymous systems. On a truly anonymous bulletin board, it is impossible to bully someone because you have no way to find him. I don't think it's a coincidence that the people who actually take being the target of verbal abuse seriously prefer to interact in pseudonymous or anonymous settings. [NEWLINE] [NEWLINE] Historically, periods of great cultural productivity were typically periods of unusual freedom from cultural control. People called Athens a cesspool of depravity and admired the disciplined societies of Sparta and Crete, but two thousand years later we look at the Athenians as the founders everything valuable in our civilization - logic, science, mathematics, philosophy, literature. You can say similar things about Renaissance Italy and Enlightenment Germany. In each case, the period of true intellectual flourishing was very short... fifty years or so. It was a little window between when one repressive regime gave up and another repressive regime was ushered in. [NEWLINE] [NEWLINE] I think it's surprising that the freest places on the internet have produced the most astonishing contributions to public culture. Astonishing, but not, as I said, unexpected, because this always happens when you unleash people for a few years or a decade. But sadly, instead of pushing anonymity further and further, the trend for the last decade has been to find ways to reproduce old mechanisms of social control on media that were originally mostly anonymous.</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right?! Applecare is the most amazing customer service I've ever experienced from any company. Ever. [NEWLINE] [NEWLINE] I cannot even say<mask> much applecare has helped me over the last few years. [NEWLINE] [NEWLINE] I had a sony vaoi before my macbook. I bought it<mask> my father bullied me out of getting an apple and I regretted it the moment the thing crapped out four months after getting it. Sent it back to them and they held it for FOUR MONTHS examining it. I had to borrow my mother's laptop<mask> I was a university student and had papers to write. I went weeks without communication and only after the seventh or eighth call asking for an update did the determine it was a dud and mailed me a new one - without the upgrades I'd paid for on the first one. The replacement's hard drive shit the bed my last week of university. I'd JUST finished my last paper (early too thankfully!)<mask> I could party with my friends. Instead I spent the last two days regurgitating the thirty way lit review in the library computer lab crying and looking vaguely homeless<mask> I hadn't showered or changed in 36 hours. I couldn't even bring myself to buy a computer for another six months<mask> I got<mask> angry even thinking about owning another POS like the Sony. [NEWLINE] [NEWLINE] My sister and parents bought me the macbook and it's been glorious. Had trouble with the the screens colours once - just brought it to the genius bar between grad school seminars. The chord ripped? No problem - they replaced it free of charge. Had to replace the battery which died this year and there was no apple stores anywhere in the vicinity - no problem. They had a list of local vendors who honour their warranties. Just dropped it off with them and apple took care of the cost. Kid I was babysitting ripped the headphones out and got the nub stuck in the jack - POST WARRANTY - and they still spent 20 minutes passing it between staffers at the store until one could muscle it out. [NEWLINE] [NEWLINE] Applecare is just bomb dot com.</s>
Label encoding: <s>Right?! Applecare is the most amazing customer service I've ever experienced from any company. Ever. [NEWLINE] [NEWLINE] I cannot even say how much applecare has helped me over the last few years. [NEWLINE] [NEWLINE] I had a sony vaoi before my macbook. I bought it because my father bullied me out of getting an apple and I regretted it the moment the thing crapped out four months after getting it. Sent it back to them and they held it for FOUR MONTHS examining it. I had to borrow my mother's laptop because I was a university student and had papers to write. I went weeks without communication and only after the seventh or eighth call asking for an update did the determine it was a dud and mailed me a new one - without the upgrades I'd paid for on the first one. The replacement's hard drive shit the bed my last week of university. I'd JUST finished my last paper (early too thankfully!) so I could party with my friends. Instead I spent the last two days regurgitating the thirty way lit review in the library computer lab crying and looking vaguely homeless because I hadn't showered or changed in 36 hours. I couldn't even bring myself to buy a computer for another six months because I got so angry even thinking about owning another POS like the Sony. [NEWLINE] [NEWLINE] My sister and parents bought me the macbook and it's been glorious. Had trouble with the the screens colours once - just brought it to the genius bar between grad school seminars. The chord ripped? No problem - they replaced it free of charge. Had to replace the battery which died this year and there was no apple stores anywhere in the vicinity - no problem. They had a list of local vendors who honour their warranties. Just dropped it off with them and apple took care of the cost. Kid I was babysitting ripped the headphones out and got the nub stuck in the jack - POST WARRANTY - and they still spent 20 minutes passing it between staffers at the store until one could muscle it out. [NEWLINE] [NEWLINE] Applecare is just bomb dot com.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is nothing<mask> a way to impose a set of moral values on an entire group of people.  It's incredibly arrogant to assume that<mask> we could just force people to reflect more they would see I'm right. [NEWLINE] [NEWLINE] This proposal is much the same<mask> the internal ultrasound laws, designed to make the procedure more uncomfortable in order to discoursge it.  There is no compelling reason for it, it does not improve the procedure in any way. [NEWLINE] [NEWLINE] This raises costs<mask> a doctor has to stand around and watch. <mask>, a piece of medical equipment needs to be developed, tested, approved, maintained, calibrated, and regulated.  It increases risk<mask> no matter<mask> failsafe it's made, operator errors happen.  And the more failsafe it's made the more expense is involved with the things I've listed above.  It adds to the cost of the procedure<mask> the patient will end up paying for the use of the machine<mask> a line item on their bill. [NEWLINE] [NEWLINE] Women are going to have abortions.  We can try to reduce it or mitigate it,<mask> it's going to happen<mask> we know<mask>.  The more uncomfortable, difficult, shameful we make it the more we encourage illegal abortion clinics. <mask> many of Gosnells' abortions would have been avoided<mask> abortions were a little more available and women didn't have to pass protesters on the sidewalk? [NEWLINE] [NEWLINE] Finally,<mask> makes you think that most women haven't already given the procedure a lot of thought?  You heard a few comments which offended your sensibilities,<mask> you haven't heard from the vast majority of women who agonize over the decision and feel it's the right thing to do.   They don't all take it lightly, and who are we to say they made the wrong decision? [NEWLINE] [NEWLINE] I suspect that most women who feel it's nothing more that a mole or a tumor are lying<mask> a defense mechanism. <mask> for those that sincerely feel that way, I'm glad they chose not to raise a baby. [NEWLINE] [NEWLINE] Edit:words and spaces</s>
Label encoding: <s>This is nothing but a way to impose a set of moral values on an entire group of people.  It's incredibly arrogant to assume that if we could just force people to reflect more they would see I'm right. [NEWLINE] [NEWLINE] This proposal is much the same as the internal ultrasound laws, designed to make the procedure more uncomfortable in order to discoursge it.  There is no compelling reason for it, it does not improve the procedure in any way. [NEWLINE] [NEWLINE] This raises costs because a doctor has to stand around and watch.  Also, a piece of medical equipment needs to be developed, tested, approved, maintained, calibrated, and regulated.  It increases risk because no matter how failsafe it's made, operator errors happen.  And the more failsafe it's made the more expense is involved with the things I've listed above.  It adds to the cost of the procedure because the patient will end up paying for the use of the machine as a line item on their bill. [NEWLINE] [NEWLINE] Women are going to have abortions.  We can try to reduce it or mitigate it, but it's going to happen because we know how.  The more uncomfortable, difficult, shameful we make it the more we encourage illegal abortion clinics.  How many of Gosnells' abortions would have been avoided if abortions were a little more available and women didn't have to pass protesters on the sidewalk? [NEWLINE] [NEWLINE] Finally, what makes you think that most women haven't already given the procedure a lot of thought?  You heard a few comments which offended your sensibilities, but you haven't heard from the vast majority of women who agonize over the decision and feel it's the right thing to do.   They don't all take it lightly, and who are we to say they made the wrong decision? [NEWLINE] [NEWLINE] I suspect that most women who feel it's nothing more that a mole or a tumor are lying as a defense mechanism.  As for those that sincerely feel that way, I'm glad they chose not to raise a baby. [NEWLINE] [NEWLINE] Edit:words and spaces</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1)<mask> is appropriate is not arbitrary, it is part of a long chain of stylistic development. Different types of clothing signal different things, partly<mask> specific types of clothing have an actual real and demonstrable psychological effect. That is, they produce an objective effect in an observer. [NEWLINE] [NEWLINE] 2) Looking good in a clothing requires quite a bit more than simply spending a lot of money on clothing, just<mask> being good at math requires a lot more than simply investing in a better calculator. Being able to spend that extra money helps,<mask> a person with a keen sense of fashion can literally go to the good will and put together a stunning outfit.<mask> you actually invest any time or energy into understanding fashion and style, this quickly becomes apparent. [NEWLINE] [NEWLINE] 3) That's a complete misunderstanding of<mask> I said. Fashion matters<mask> it is a way of signaling *real, tangible things* such<mask> intelligence, fitness, social skill, cultural awareness, attention to detail and wealth. Not everyone who doesn't care about fashion lacks those attributes,<mask> people who are exceptionally good at understanding fashion *do* have those attributes to various degrees. In that sense it is analogous to one's vocabulary. Not everyone with a small vocabulary is unintelligent,<mask> virtually everyone with a truly expansive vocabulary (and by truly expansive I mean both possessing a large vocabulary, and being able to use such a vocabulary correctly) *is* intelligent.<mask>,<mask> I see someone who dresses really sharp, I can be fairly confident that I now know a certain amount about them. I have more information than I would have had in the absence of their fashionability. Just<mask> the peacock is signaling real things about their fitness to the peahen with their outrageous tail-feather display (notably an ability to survive real dangers<mask> possessing such outrageous plumage, and a freedom from certain genetic defects),<mask> too does the well tailored man signal things to his employer by wearing a suit that flatters his figure, is well tailored, is well complemented by his tie, and is appropriate to the occasion. </s><pad>
Label encoding: <s>1) What is appropriate is not arbitrary, it is part of a long chain of stylistic development. Different types of clothing signal different things, partly because specific types of clothing have an actual real and demonstrable psychological effect. That is, they produce an objective effect in an observer. [NEWLINE] [NEWLINE] 2) Looking good in a clothing requires quite a bit more than simply spending a lot of money on clothing, just as being good at math requires a lot more than simply investing in a better calculator. Being able to spend that extra money helps, but a person with a keen sense of fashion can literally go to the good will and put together a stunning outfit. If you actually invest any time or energy into understanding fashion and style, this quickly becomes apparent. [NEWLINE] [NEWLINE] 3) That's a complete misunderstanding of what I said. Fashion matters because it is a way of signaling *real, tangible things* such as intelligence, fitness, social skill, cultural awareness, attention to detail and wealth. Not everyone who doesn't care about fashion lacks those attributes, but people who are exceptionally good at understanding fashion *do* have those attributes to various degrees. In that sense it is analogous to one's vocabulary. Not everyone with a small vocabulary is unintelligent, but virtually everyone with a truly expansive vocabulary (and by truly expansive I mean both possessing a large vocabulary, and being able to use such a vocabulary correctly) *is* intelligent. So, when I see someone who dresses really sharp, I can be fairly confident that I now know a certain amount about them. I have more information than I would have had in the absence of their fashionability. Just as the peacock is signaling real things about their fitness to the peahen with their outrageous tail-feather display (notably an ability to survive real dangers despite possessing such outrageous plumage, and a freedom from certain genetic defects), so too does the well tailored man signal things to his employer by wearing a suit that flatters his figure, is well tailored, is well complemented by his tie, and is appropriate to the occasion. </s><pad>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They don't have to win the gun battle, they just have to get a point across to trigger peaceful noncompliance, electoral action, or any number of other methods we have of forcing our political elite to actually care. It's not<mask> clearly illustrated in United States History<mask> it is in Latin America or Europe, mostly<mask> electoral action, peaceful noncompliance, and other methodologies are normally employed before things get bad enough to warrant violence. [NEWLINE] [NEWLINE] For example, the [Battle of Blair Mountain]( [URL] ) was a time<mask> 10,000 armed miners in West Virginia revolted against mine operators and lawmen in 1921. They clearly lost the battle,<mask> bombers and heavy machineguns were deployed against folks with at best hunting rifles.<mask> their objective, to remove legal barriers to unionizing half the state of West Virginia and make sure that SOMEONE outside the state was aware of the problem were successful by 1935. The fact that there was a battle at all was enough to redefine the issue in a way that four decades of peaceful protest and strike were unable. [NEWLINE] [NEWLINE] Similar basic principles operate today. The Dorner event was enough to demonstrate some of the basic concepts.<mask> of violence, national awareness was focused on a problem (LAPD sucking at life in general and racial profiling in particular). This is different<mask> it wasn't a group of people, and there was no backing community to define and push the narrative. [NEWLINE] [NEWLINE] The fact of the matter is that this kind of violence is clearly objectively bad. The fact that a person has to resort to it is enough to convince the American Public that the whole situation is fucked, and that change is absolutely necessary. It is extreme, it is inherently destructive, and it is incredibly rare. That give it power that otherwise better form of protest cannot match,<mask> it happened often then it wouldn't matter and<mask> people didn't die then it wouldn't be<mask> serious. Still, it has to be violence to a purpose to have full effect. Violence to just kill provides an automatic counterargument, again<mask> seen in the Dorner case.</s>
Label encoding: <s>They don't have to win the gun battle, they just have to get a point across to trigger peaceful noncompliance, electoral action, or any number of other methods we have of forcing our political elite to actually care. It's not as clearly illustrated in United States History as it is in Latin America or Europe, mostly because electoral action, peaceful noncompliance, and other methodologies are normally employed before things get bad enough to warrant violence. [NEWLINE] [NEWLINE] For example, the [Battle of Blair Mountain]( [URL] ) was a time where 10,000 armed miners in West Virginia revolted against mine operators and lawmen in 1921. They clearly lost the battle, as bombers and heavy machineguns were deployed against folks with at best hunting rifles. But their objective, to remove legal barriers to unionizing half the state of West Virginia and make sure that SOMEONE outside the state was aware of the problem were successful by 1935. The fact that there was a battle at all was enough to redefine the issue in a way that four decades of peaceful protest and strike were unable. [NEWLINE] [NEWLINE] Similar basic principles operate today. The Dorner event was enough to demonstrate some of the basic concepts. Because of violence, national awareness was focused on a problem (LAPD sucking at life in general and racial profiling in particular). This is different because it wasn't a group of people, and there was no backing community to define and push the narrative. [NEWLINE] [NEWLINE] The fact of the matter is that this kind of violence is clearly objectively bad. The fact that a person has to resort to it is enough to convince the American Public that the whole situation is fucked, and that change is absolutely necessary. It is extreme, it is inherently destructive, and it is incredibly rare. That give it power that otherwise better form of protest cannot match, if it happened often then it wouldn't matter and if people didn't die then it wouldn't be as serious. Still, it has to be violence to a purpose to have full effect. Violence to just kill provides an automatic counterargument, again as seen in the Dorner case.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>(Usa)<mask> a person with two young children... No. Before I had kids I thought this way. Most of the teachers I know<mask> certainly think they are hot shit and should be paid more.<mask>, now that I've got kids in school... No. I have many opinions on the issues,<mask> I'll look stick to one aspect that i can keep factual and not grounded in my own perception. I'll stick strictly to hours<mask> there being a myriad of other reasons there is no way in hell teachers should make doctors pay. [NEWLINE] [NEWLINE] They work partial days, around here the school day is 9-3 (that's 6 hours) and they<mask> are guarunteed an hour for lunch and two breaks during that 6 hours. Most of them<mask> don't work extra like they may claim. Most roll in the door at 8:55 and roll out at 3:15.<mask> they have anything to grade, it's done in front of the TV at night or who he the kids are in art or gym. Many of them week<mask> use their hour lunch to make photo copies or do other prep<mask> they won't stay a drop past 3:15. Then they<mask> get every holiday off<mask><mask> to a week for spring break and at least 2 weeks over Christmas and 2-3 months over the summer. Add to this that they push to use all the calamity days<mask> that they can have more paid days off. Its utterly ridiculous. 30k per year is more than generous for a job<mask> they only work 30 hours a week 8-9 months out of the year.  There are a lot of folks out there who work 50 hours a week 52 weeks a year who don't make a teachers pay. <mask> they actually worked year round and were *required* to stay for a full 8-9 hour day either grading, prepping, or doing other odd jobs around the building, I could see them making 50-60k.<mask> they basically have glorified part time jobs. They get benefits too, don't forget.</s>
Label encoding: <s>(Usa) As a person with two young children... No. Before I had kids I thought this way. Most of the teachers I know also certainly think they are hot shit and should be paid more. But, now that I've got kids in school... No. I have many opinions on the issues, but I'll look stick to one aspect that i can keep factual and not grounded in my own perception. I'll stick strictly to hours despite there being a myriad of other reasons there is no way in hell teachers should make doctors pay. [NEWLINE] [NEWLINE] They work partial days, around here the school day is 9-3 (that's 6 hours) and they also are guarunteed an hour for lunch and two breaks during that 6 hours. Most of them also don't work extra like they may claim. Most roll in the door at 8:55 and roll out at 3:15. If they have anything to grade, it's done in front of the TV at night or who he the kids are in art or gym. Many of them week also use their hour lunch to make photo copies or do other prep because they won't stay a drop past 3:15. Then they also get every holiday off in addition to a week for spring break and at least 2 weeks over Christmas and 2-3 months over the summer. Add to this that they push to use all the calamity days so that they can have more paid days off. Its utterly ridiculous. 30k per year is more than generous for a job where they only work 30 hours a week 8-9 months out of the year.  There are a lot of folks out there who work 50 hours a week 52 weeks a year who don't make a teachers pay.  If they actually worked year round and were *required* to stay for a full 8-9 hour day either grading, prepping, or doing other odd jobs around the building, I could see them making 50-60k. But they basically have glorified part time jobs. They get benefits too, don't forget.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd have definitely gone for paragon saves everyone Shep<mask> my 'final' run,<mask> not for the sour taste the ending left in my mouth. I'd actually planned for that to be the one I went with,<mask> given<mask> the whole stupid Crucible felt like it was just ignoring everything he'd done... yeah, didn't sit well with me. I wanted to leave the game feeling *something*<mask> vaguely discontented and annoyed with it, which is<mask> shitty-Shep happened. [NEWLINE] [NEWLINE] I do get<mask> you mean by the whole emotional payoff from saving people - I did really enjoy those moments and<mask><mask> those individual arcs were fantastically written and seriously impactful. I just think the ending of the game really let them down, and in a pretty disappointing manner, too.<mask><mask> I enjoy them<mask> individual character arcs,<mask><mask><mask> about the playthrough I want to remember<mask> my last, shitty-Shep that burnt pretty much every bridge it was possible to burn until only three choices were left is far more satisfying to me from a literary, storytelling perspective. [NEWLINE] [NEWLINE] Anyway, that's kind of<mask> I'd say there's no One Best Way - the right choice is one that satisfies the player emotionally. For some people that's saving everyone,<mask> quitting before the final sequence and reading a fan-made ending. For other people that's sticking it through to the end just<mask> they can say they finished it and then washing their hands of the series. For you it's to save everyone. For me it was making the World's Worst Shepard. And<mask> on and<mask> forth. [NEWLINE] [NEWLINE] <mask> I'm kinda glad to have changed your view. I totally get wanting people to see the things that made you happy,<mask> I'd suggest instead of being like 'no you must play THIS way and make THIS choice' you encourage people to try the game more than once<mask> that they can see<mask> would happen<mask> they made a different choice instead,<mask> opposed to railroading them down your route for their first run. </s>
Label encoding: <s>I'd have definitely gone for paragon saves everyone Shep as my 'final' run, if not for the sour taste the ending left in my mouth. I'd actually planned for that to be the one I went with, but given how the whole stupid Crucible felt like it was just ignoring everything he'd done... yeah, didn't sit well with me. I wanted to leave the game feeling *something* besides vaguely discontented and annoyed with it, which is why shitty-Shep happened. [NEWLINE] [NEWLINE] I do get what you mean by the whole emotional payoff from saving people - I did really enjoy those moments and I think those individual arcs were fantastically written and seriously impactful. I just think the ending of the game really let them down, and in a pretty disappointing manner, too. So while I enjoy them as individual character arcs, when I think about the playthrough I want to remember as my last, shitty-Shep that burnt pretty much every bridge it was possible to burn until only three choices were left is far more satisfying to me from a literary, storytelling perspective. [NEWLINE] [NEWLINE] Anyway, that's kind of why I'd say there's no One Best Way - the right choice is one that satisfies the player emotionally. For some people that's saving everyone, but quitting before the final sequence and reading a fan-made ending. For other people that's sticking it through to the end just so they can say they finished it and then washing their hands of the series. For you it's to save everyone. For me it was making the World's Worst Shepard. And so on and so forth. [NEWLINE] [NEWLINE] But I'm kinda glad to have changed your view. I totally get wanting people to see the things that made you happy, but I'd suggest instead of being like 'no you must play THIS way and make THIS choice' you encourage people to try the game more than once so that they can see what would happen if they made a different choice instead, as opposed to railroading them down your route for their first run. </s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not really sure<mask> you want to talk about more with this CMV, using prisoners or not using animals. Human testing is already a requisite prior to a drug coming to market. [Animal trials come first to prove safety and efficacy prior to using human subjects]( [URL].png). Prior to those trials there's a lot of in vitro studies to give us a guided estimate on<mask> to expect.<mask> more to your point, there is not drug on the market that hasn't been tested on humans before its release. Using animals is a rudimentary stepping stone given. [NEWLINE] [NEWLINE] In regards to your article on drugs, there are some limitations with its analysis. First it is listing side effects<mask> failing to categorize them. A side effect could be formation of a clot, a headache, sleep disturbances, etc. It's more likely animal testing didn't catch certain side effects (particularly more subjective ones)<mask> researchers can't exactly ask a mouse<mask> it is feeling.<mask>, it doesn't provide a link or any information on who conducted this study or<mask> it was done. It's all a little suspect all things considered. The article makes no mention of the FDA approval process for medications and fails to give its anecdotes full detail other than to push its agenda against animal testing. [NEWLINE] [NEWLINE] I work in medicine<mask> I'll say this, drugs are a complicated business. Yes, there have been many recalls and discontinuations of drugs brought to market. That is<mask><mask><mask> it can take 10-15 years to bring a drug to market, we still rarely have any idea of long-term side effects that can manifest.<mask>, some recalls and product discontinuations aren't necessarily<mask> a drug is dangerous. Some companies are finding their profits on a medication aren't worth its production or sometimes a drug receives bad press (even<mask> it is effective and safe). The fallout regarding HRT in post-menopausal women is something discovered years after it was adopted into guidelines by the medical community. We only discovered the risk post-hoc and theoretically it was a sound hypothesis initially. </s>
Label encoding: <s>Not really sure what you want to talk about more with this CMV, using prisoners or not using animals. Human testing is already a requisite prior to a drug coming to market. [Animal trials come first to prove safety and efficacy prior to using human subjects]( [URL].png). Prior to those trials there's a lot of in vitro studies to give us a guided estimate on what to expect. But more to your point, there is not drug on the market that hasn't been tested on humans before its release. Using animals is a rudimentary stepping stone given. [NEWLINE] [NEWLINE] In regards to your article on drugs, there are some limitations with its analysis. First it is listing side effects but failing to categorize them. A side effect could be formation of a clot, a headache, sleep disturbances, etc. It's more likely animal testing didn't catch certain side effects (particularly more subjective ones) as researchers can't exactly ask a mouse how it is feeling. Secondly, it doesn't provide a link or any information on who conducted this study or where it was done. It's all a little suspect all things considered. The article makes no mention of the FDA approval process for medications and fails to give its anecdotes full detail other than to push its agenda against animal testing. [NEWLINE] [NEWLINE] I work in medicine so I'll say this, drugs are a complicated business. Yes, there have been many recalls and discontinuations of drugs brought to market. That is because even though it can take 10-15 years to bring a drug to market, we still rarely have any idea of long-term side effects that can manifest. Secondly, some recalls and product discontinuations aren't necessarily because a drug is dangerous. Some companies are finding their profits on a medication aren't worth its production or sometimes a drug receives bad press (even if it is effective and safe). The fallout regarding HRT in post-menopausal women is something discovered years after it was adopted into guidelines by the medical community. We only discovered the risk post-hoc and theoretically it was a sound hypothesis initially. </s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A little quick math:  Average teacher's salary in the US is [$57,000]( [URL].asp?id=28).  Average doctor's salary is [$230,000]( [URL] /).  There are [3.3 million]( [URL].asp?id=28) teachers in the US. [NEWLINE] [NEWLINE] Cost of the pay increase is equal to the number of teachers times the salary difference:  3,300,000 *  ($230,000 - $57,000)  =  approx. $570,000,000,000 per year.  Divided by 120 million households, that comes out to $4,800 in additional taxes per year. <mask>,<mask> we know, most taxes are paid by the upper half of income earners<mask> the lower half essentially get a free ride. <mask>, more likely, the cost would be an additional $10k in taxes on average for the upper half of income earners, and no additional taxes for those in the lower half. [NEWLINE] [NEWLINE] <mask><mask> the people stuck with the bill would balk at paying an extra $10k in taxes every year. [NEWLINE] [NEWLINE] Edit:  And all this begs a couple of questions:  First, is it cost-effective?  Supposing there are enough highly competent individuals (equal in talent to an average doctor). <mask><mask> teacher talent is not the limiting factor for quality of education? <mask><mask>, instead, the main limitations to quality of education lie with the students themselves, and their families? <mask> that's the case, it doesn't matter<mask> good the teachers are.  Second, are there even enough highly talented individuals available to fill all these slots?  3.3 million is a lot of people. <mask> would they all come from?  There are only<mask> many brilliant people available.  Seems like for the plan to be effective, there would be a massive brain drain from Silicon Valley, corporate management, and<mask> the medical profession<mask> well.  Are you sure that's a desirable result?</s>
Label encoding: <s>A little quick math:  Average teacher's salary in the US is [$57,000]( [URL].asp?id=28).  Average doctor's salary is [$230,000]( [URL] /).  There are [3.3 million]( [URL].asp?id=28) teachers in the US. [NEWLINE] [NEWLINE] Cost of the pay increase is equal to the number of teachers times the salary difference:  3,300,000 *  ($230,000 - $57,000)  =  approx. $570,000,000,000 per year.  Divided by 120 million households, that comes out to $4,800 in additional taxes per year.  However, as we know, most taxes are paid by the upper half of income earners while the lower half essentially get a free ride.  So, more likely, the cost would be an additional $10k in taxes on average for the upper half of income earners, and no additional taxes for those in the lower half. [NEWLINE] [NEWLINE] I think the people stuck with the bill would balk at paying an extra $10k in taxes every year. [NEWLINE] [NEWLINE] Edit:  And all this begs a couple of questions:  First, is it cost-effective?  Supposing there are enough highly competent individuals (equal in talent to an average doctor).  What if teacher talent is not the limiting factor for quality of education?  What if, instead, the main limitations to quality of education lie with the students themselves, and their families?  If that's the case, it doesn't matter how good the teachers are.  Second, are there even enough highly talented individuals available to fill all these slots?  3.3 million is a lot of people.  Where would they all come from?  There are only so many brilliant people available.  Seems like for the plan to be effective, there would be a massive brain drain from Silicon Valley, corporate management, and indeed the medical profession as well.  Are you sure that's a desirable result?</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you about the boxing (and wrestling and<mask> on) issue; I've always seen the weight classes<mask> a little bizarre. I mean, this all goes back, in a way, to the Iliad,<mask> the other Greeks were *pissed* that Odysseus won the funeral horse-race by driving his chariot better,<mask> they thought the race should be all about whose *horses* were the fastest.<mask> I see the value of having one sport about fighting *well*<mask> everyone has the same amount of muscle, and another about putting on<mask> much muscle<mask> possible.<mask> in practice, it turns wrestling into competitive bulimia, which isn't my picture of an interesting competition. [NEWLINE] [NEWLINE] [STARTQ] <mask> in that case the only "real" thing left would be no-rules street fights, potentially with knives and guns; warfare; obtaining money by whatever means etc. Or seen from even higher, evolutionary perspective, to survive and reproduce; and from an even higher perspective there is no goal and purpose at all and we're just collections of complex molecules in an indifferent universe and all those things. [ENDQ] [NEWLINE] <mask><mask> that's taking things a littttle bit to the extreme. I would start by splitting up rule changes into six types. First,<mask> purpose does the rule serve? It can be to make the game better to play, better to watch, or better able to serve some ulterior purpose. Second,<mask> part of the game does the rule affect? It can either affect<mask> is legal within the game, or it can affect the circumstances under which games take place (in other words,<mask> is legal in a tournament or a league). [NEWLINE] [NEWLINE] This means that even in anarchy<mask> no one is controlling anything about the "tournament system", you could still have rules for<mask> to play *the game* that everyone agrees on,<mask> they think that game is the deepest game, or the most watchable game, or the game that gives the best exercise, and<mask> on.</s>
Label encoding: <s>I agree with you about the boxing (and wrestling and so on) issue; I've always seen the weight classes as a little bizarre. I mean, this all goes back, in a way, to the Iliad, where the other Greeks were *pissed* that Odysseus won the funeral horse-race by driving his chariot better, because they thought the race should be all about whose *horses* were the fastest. So I see the value of having one sport about fighting *well* where everyone has the same amount of muscle, and another about putting on as much muscle as possible. But in practice, it turns wrestling into competitive bulimia, which isn't my picture of an interesting competition. [NEWLINE] [NEWLINE] [STARTQ] But in that case the only "real" thing left would be no-rules street fights, potentially with knives and guns; warfare; obtaining money by whatever means etc. Or seen from even higher, evolutionary perspective, to survive and reproduce; and from an even higher perspective there is no goal and purpose at all and we're just collections of complex molecules in an indifferent universe and all those things. [ENDQ] [NEWLINE] I think that's taking things a littttle bit to the extreme. I would start by splitting up rule changes into six types. First, what purpose does the rule serve? It can be to make the game better to play, better to watch, or better able to serve some ulterior purpose. Second, what part of the game does the rule affect? It can either affect what is legal within the game, or it can affect the circumstances under which games take place (in other words, what is legal in a tournament or a league). [NEWLINE] [NEWLINE] This means that even in anarchy when no one is controlling anything about the "tournament system", you could still have rules for how to play *the game* that everyone agrees on, because they think that game is the deepest game, or the most watchable game, or the game that gives the best exercise, and so on.</s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your examples point at a different argument than your post title - they suggest a frustration with specific instances<mask> you wanted to be left alone for personal reasons (or<mask> you had your hands full) and weren't. That's different than saying that talking to strangers is never ok<mask> a Christian. [NEWLINE] [NEWLINE] The reason I point that out is that<mask><mask> most people would agree with you that proselytizing to people outside a synagogue, or during a holocaust ceremony, is rude. There's no "should be viewed<mask> rude" here - it is already. [NEWLINE] [NEWLINE] Other examples are more difficult - no one can tell that you are returning home for a funeral, for example. The right action in this instance is to back off and let a person in a sensitive spot take their time. The same thing applies with making a joke about cheating boyfriends in a conversation with someone that just got cheated on. It's not rude to make the joke,<mask> it is rude not to apologize or give the person some space. Likewise, someone proselytizing should back off<mask> someone is in a bad spot. [NEWLINE] [NEWLINE] Now to your original statement. My problem with it is that it suggests that a society is optimal<mask> people don't talk with strangers. You don't want to be bothered by Christians, marketers, vegans, or or environmentalists, and that's fair,<mask> opening a conversation with someone on the street, or in a bar, or wherever, should be socially acceptable. Full stop. [NEWLINE] [NEWLINE] <mask> someone doesn't want to talk, we should leave them alone,<mask> I want the initial opener to be ok. Otherwise, we lose the opportunity to make a connection, share information about a nearby pothole, or a guy at the bar who should be avoided. Christians and others might take advantage of this precept,<mask><mask><mask><mask> they back off<mask> you say "No, thank you," I don't think it should be considered rude. One sentence isn't harrassment. [NEWLINE] [NEWLINE] Reference: [URL] /</s>
Label encoding: <s>I think your examples point at a different argument than your post title - they suggest a frustration with specific instances where you wanted to be left alone for personal reasons (or because you had your hands full) and weren't. That's different than saying that talking to strangers is never ok as a Christian. [NEWLINE] [NEWLINE] The reason I point that out is that I think most people would agree with you that proselytizing to people outside a synagogue, or during a holocaust ceremony, is rude. There's no "should be viewed as rude" here - it is already. [NEWLINE] [NEWLINE] Other examples are more difficult - no one can tell that you are returning home for a funeral, for example. The right action in this instance is to back off and let a person in a sensitive spot take their time. The same thing applies with making a joke about cheating boyfriends in a conversation with someone that just got cheated on. It's not rude to make the joke, but it is rude not to apologize or give the person some space. Likewise, someone proselytizing should back off if someone is in a bad spot. [NEWLINE] [NEWLINE] Now to your original statement. My problem with it is that it suggests that a society is optimal when people don't talk with strangers. You don't want to be bothered by Christians, marketers, vegans, or or environmentalists, and that's fair, but opening a conversation with someone on the street, or in a bar, or wherever, should be socially acceptable. Full stop. [NEWLINE] [NEWLINE] If someone doesn't want to talk, we should leave them alone, but I want the initial opener to be ok. Otherwise, we lose the opportunity to make a connection, share information about a nearby pothole, or a guy at the bar who should be avoided. Christians and others might take advantage of this precept, but as long as they back off when you say "No, thank you," I don't think it should be considered rude. One sentence isn't harrassment. [NEWLINE] [NEWLINE] Reference: [URL] /</s>
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That was one of the best responses I've seen on a CMV in a long time. Thanks for your thoughtful input. [NEWLINE] [NEWLINE] Hopping on OP's thought process<mask> coming at it from a different angle, maybe you can assist<mask> you seem to be quite knowledgeable in this area. [NEWLINE] [NEWLINE] I struggle with the fact that our existence was no choice of ours at all. We were put on this earth by our parents decision to conceive us (sometimes consciously, sometimes not).<mask>, it is assumed that once we exist that we should strive to exist for<mask><mask><mask> possible. [NEWLINE] [NEWLINE] We have things that we are naturally made of that aide us on the path to existence for<mask><mask><mask> possible. Our Nervous systems tell us that something might cause us pain,<mask> we refrain from doing it. Our brains tell us we are hungry or that we need to sleep and<mask> we do it. Each of those actions are our animalistic or natural responses to a desire to exist. [NEWLINE] [NEWLINE] <mask><mask> separates us from other animals is the ability to have cognitive thought. With such thought we<mask> acquire the ability to not rely on our animalisitic responses. In other words<mask> we are hungry we can choose to not eat, even<mask> our body is reacting in a way that is telling us we should eat. [NEWLINE] [NEWLINE] <mask> we possess the ability to supersede our natural instincts<mask> is it assumed that we should still strive to exist for<mask><mask><mask> possible?<mask> our goal is not to live for<mask><mask><mask> possible, then<mask> should we exist at all<mask> such existence was never our doing to begin with?<mask> our goal for existence is simply to experience that with which existence comes with, is this a byproduct of our natural instinct or of our cognitive thought? I'd argue it is the latter, and<mask> that is the case then<mask> would the cognitive thought of wanting to experience supersede the cognitive thought of not wanting to exist? [NEWLINE] [NEWLINE] Any insight you might be able to provide, would be greatly appreciated. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>That was one of the best responses I've seen on a CMV in a long time. Thanks for your thoughtful input. [NEWLINE] [NEWLINE] Hopping on OP's thought process but coming at it from a different angle, maybe you can assist as you seem to be quite knowledgeable in this area. [NEWLINE] [NEWLINE] I struggle with the fact that our existence was no choice of ours at all. We were put on this earth by our parents decision to conceive us (sometimes consciously, sometimes not). Yet, it is assumed that once we exist that we should strive to exist for as long as possible. [NEWLINE] [NEWLINE] We have things that we are naturally made of that aide us on the path to existence for as long as possible. Our Nervous systems tell us that something might cause us pain, so we refrain from doing it. Our brains tell us we are hungry or that we need to sleep and so we do it. Each of those actions are our animalistic or natural responses to a desire to exist. [NEWLINE] [NEWLINE] However what separates us from other animals is the ability to have cognitive thought. With such thought we also acquire the ability to not rely on our animalisitic responses. In other words if we are hungry we can choose to not eat, even if our body is reacting in a way that is telling us we should eat. [NEWLINE] [NEWLINE] If we possess the ability to supersede our natural instincts why is it assumed that we should still strive to exist for as long as possible? If our goal is not to live for as long as possible, then why should we exist at all since such existence was never our doing to begin with? If our goal for existence is simply to experience that with which existence comes with, is this a byproduct of our natural instinct or of our cognitive thought? I'd argue it is the latter, and if that is the case then why would the cognitive thought of wanting to experience supersede the cognitive thought of not wanting to exist? [NEWLINE] [NEWLINE] Any insight you might be able to provide, would be greatly appreciated. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It doesn’t even matter<mask> much of her ex’s accusations are true or not,<mask> it’s nothing<mask> pure misogyny to use online harassment troops to punish a woman<mask> she didn’t meet your standards for a girlfriend. [ENDQ] [NEWLINE] I find myself agreeing with this. Whether or not an ex did you wrong, it's wrong to sic an online witch-hunt on them. I would seriously hope you agree. I don't necessarily agree that the boyfriend's motives were misogynistic,<mask> the extent to which the witch-hunt took hold definitely was. [NEWLINE] [NEWLINE] [STARTQ] Quinn’s ex and the harassers are accusing Quinn of an “ethics” violation, accusing her, no joke, of using sex to get a favorable review from Kotaku. [ENDQ] [NEWLINE] From<mask> I understand, the glowing review Quinn supposedly traded sex for actually was not that. Regardless, shouldn't the brunt of the public backlash fall on the irresponsible journalist(s) who were [allegedly] trading professional favors for sex? From<mask> I'm sitting, aren't they acting at least<mask> unethically<mask> Quinn?<mask> they, being men, aren't really catching the brunt of the public backlash, are they? [NEWLINE] [NEWLINE] [STARTQ] <mask> they are still reflective of the ugly attitudes about sex and power that permeate the gaming world and the larger geeky world that surrounds it. [ENDQ] [NEWLINE] Sorry, are you paying attention at all? I don't mean to be rude,<mask> I wouldn't blame you<mask> you weren't,<mask> among other things, people are circulating nude pictures of her, are saying she is setting women in the industry back, and are blaming her instead of the supposedly multiple journalists who traded their integrity for sex. Not to mention accusing her of rape (which is less misogynistic and more creepy and weird). Among other things. And frankly, that misogyny exists within the gaming community shouldn't be news to you; this is just an example of it, rather than proof.</s>
Label encoding: <s> [STARTQ] It doesn’t even matter how much of her ex’s accusations are true or not, because it’s nothing but pure misogyny to use online harassment troops to punish a woman because she didn’t meet your standards for a girlfriend. [ENDQ] [NEWLINE] I find myself agreeing with this. Whether or not an ex did you wrong, it's wrong to sic an online witch-hunt on them. I would seriously hope you agree. I don't necessarily agree that the boyfriend's motives were misogynistic, but the extent to which the witch-hunt took hold definitely was. [NEWLINE] [NEWLINE] [STARTQ] Quinn’s ex and the harassers are accusing Quinn of an “ethics” violation, accusing her, no joke, of using sex to get a favorable review from Kotaku. [ENDQ] [NEWLINE] From what I understand, the glowing review Quinn supposedly traded sex for actually was not that. Regardless, shouldn't the brunt of the public backlash fall on the irresponsible journalist(s) who were [allegedly] trading professional favors for sex? From where I'm sitting, aren't they acting at least as unethically as Quinn? But they, being men, aren't really catching the brunt of the public backlash, are they? [NEWLINE] [NEWLINE] [STARTQ] But they are still reflective of the ugly attitudes about sex and power that permeate the gaming world and the larger geeky world that surrounds it. [ENDQ] [NEWLINE] Sorry, are you paying attention at all? I don't mean to be rude, because I wouldn't blame you if you weren't, but among other things, people are circulating nude pictures of her, are saying she is setting women in the industry back, and are blaming her instead of the supposedly multiple journalists who traded their integrity for sex. Not to mention accusing her of rape (which is less misogynistic and more creepy and weird). Among other things. And frankly, that misogyny exists within the gaming community shouldn't be news to you; this is just an example of it, rather than proof.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My understanding is that the only situation your money actually is invested in the company, is<mask> you are purchasing brand new shares, which only happens<mask> the company is in a crisis and needs to acquire funding from investors [ENDQ] [NEWLINE] That is not true.<mask> anything, going public in a time of crisis would really not be the smartest choice. [NEWLINE] [NEWLINE] IPOs are immensely costly and time consuming.<mask>, the company can sell shares to raise funds directly through FPOs<mask> well after they have done an IPO<mask> money can still be raised directly for the company. [NEWLINE] [NEWLINE] Furthermore, even<mask> they weren't part of the IPO they are still owners of the company. They have bought a stake in ownership. You can't just dismiss that<mask> meaningless just<mask> the funds don't directly go to the company. [NEWLINE] [NEWLINE] <mask> all companies behaved<mask> such, then<mask> point would there be in buying a share? It would be absolutely meaningless. [NEWLINE] [NEWLINE] Second, a company *does* get hurt<mask> their stocks take a nosedive.<mask> post-IPO, most shares sold don't go to or from the company, it impacts it in other ways. [NEWLINE] [NEWLINE] A company's share price is an outward indicator showing the health of the company. Customers hear "Ford shares are taking a nosedive" and think, "Wow, maybe I should buy a Toyota instead. I want a company that's in good shape and reliable." [NEWLINE] [NEWLINE] Bankers hear the same and then they become less likely to give out loans. And given<mask> many executives are now paid in shares, it gives qualified talent less incentive to apply to work at the place. [NEWLINE] [NEWLINE] There are a number of reasons<mask> maximizing shareholder value is critical for a company. That doesn't mean,<mask>, they have to completely forgo quality, customer satisfaction, or work environment conditions. [NEWLINE] [NEWLINE] Those can<mask> be a priority. That doesn't mean maximizing shareholder value and pleasing the stockholders *shouldn't* be a priority<mask>. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] My understanding is that the only situation your money actually is invested in the company, is when you are purchasing brand new shares, which only happens when the company is in a crisis and needs to acquire funding from investors [ENDQ] [NEWLINE] That is not true. If anything, going public in a time of crisis would really not be the smartest choice. [NEWLINE] [NEWLINE] IPOs are immensely costly and time consuming. Besides, the company can sell shares to raise funds directly through FPOs as well after they have done an IPO so money can still be raised directly for the company. [NEWLINE] [NEWLINE] Furthermore, even if they weren't part of the IPO they are still owners of the company. They have bought a stake in ownership. You can't just dismiss that as meaningless just because the funds don't directly go to the company. [NEWLINE] [NEWLINE] If all companies behaved as such, then what point would there be in buying a share? It would be absolutely meaningless. [NEWLINE] [NEWLINE] Second, a company *does* get hurt if their stocks take a nosedive. While post-IPO, most shares sold don't go to or from the company, it impacts it in other ways. [NEWLINE] [NEWLINE] A company's share price is an outward indicator showing the health of the company. Customers hear "Ford shares are taking a nosedive" and think, "Wow, maybe I should buy a Toyota instead. I want a company that's in good shape and reliable." [NEWLINE] [NEWLINE] Bankers hear the same and then they become less likely to give out loans. And given how many executives are now paid in shares, it gives qualified talent less incentive to apply to work at the place. [NEWLINE] [NEWLINE] There are a number of reasons why maximizing shareholder value is critical for a company. That doesn't mean, however, they have to completely forgo quality, customer satisfaction, or work environment conditions. [NEWLINE] [NEWLINE] Those can also be a priority. That doesn't mean maximizing shareholder value and pleasing the stockholders *shouldn't* be a priority though. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have been reading plenty in the The Red Pill subreddit. It seems to be nothing more than a social equivalent of a far-right ideology. From<mask> I can gather, it's basically saying "Hey, 17-25 year old guy. You're down on your luck, have been having trouble with women. Your problems aren't your fault; they're there<mask> you're being socially governed by feminism!<mask> you just submit to our ideology, your life will be<mask> much better. You will be strong and successful! Sure, some parts will suck (like not being emotionally attached to your lovers),<mask> you will gain from it. Just<mask> fascism condemns democratic values<mask> weak, TRP condemns basic bonding and emotional attachment to women<mask> weak. Likewise, TRP often preaches a Social Darwinism (those who are successful deserve to be<mask><mask> it is nature's way and they are Alphas). Fascism, likewise, preaches that acquiring power is a sign of strength and those who have power deserve to have it. [NEWLINE] In short, just<mask> fascism appeals to the working class by asserting that they should emulate the capitalist class (or at least the portion of it that subscribes to their ideology), and blames its problems on outsiders, intellectuals, and those at the bottom of the heap, TRP preaches to those with lower-to-medium social status that their problems are due to outsiders (women), those with little social status (hamsters, uber-Betas, whatever), and intellectuals (feminists). Both promise a superficially better life<mask> at immense cost. Instead of assigning blame to problems appropriately, they attack the cornerstones of civilized society<mask> weak and preach their ideology<mask> the end-all-be-all,<mask> manifest that failure to subscribe is viewed<mask> willful ignorance. Human nature is denounced<mask> that of slightly more intelligent apes.<mask> please, CMV<mask> to<mask> TRP is to the social sphere<mask> to<mask> far-right ideology is to the political sphere. [NEWLINE] </s>
Label encoding: <s>I have been reading plenty in the The Red Pill subreddit. It seems to be nothing more than a social equivalent of a far-right ideology. From what I can gather, it's basically saying "Hey, 17-25 year old guy. You're down on your luck, have been having trouble with women. Your problems aren't your fault; they're there because you're being socially governed by feminism! If you just submit to our ideology, your life will be so much better. You will be strong and successful! Sure, some parts will suck (like not being emotionally attached to your lovers), but you will gain from it. Just as fascism condemns democratic values as weak, TRP condemns basic bonding and emotional attachment to women as weak. Likewise, TRP often preaches a Social Darwinism (those who are successful deserve to be so because it is nature's way and they are Alphas). Fascism, likewise, preaches that acquiring power is a sign of strength and those who have power deserve to have it. [NEWLINE] In short, just as fascism appeals to the working class by asserting that they should emulate the capitalist class (or at least the portion of it that subscribes to their ideology), and blames its problems on outsiders, intellectuals, and those at the bottom of the heap, TRP preaches to those with lower-to-medium social status that their problems are due to outsiders (women), those with little social status (hamsters, uber-Betas, whatever), and intellectuals (feminists). Both promise a superficially better life but at immense cost. Instead of assigning blame to problems appropriately, they attack the cornerstones of civilized society as weak and preach their ideology as the end-all-be-all, so manifest that failure to subscribe is viewed as willful ignorance. Human nature is denounced as that of slightly more intelligent apes. So please, CMV as to how TRP is to the social sphere as to what far-right ideology is to the political sphere. [NEWLINE] </s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hey, thanks for the response! I happen to completely agree with<mask> you're laying down - my personal view is in line with<mask> you're saying. With family background in health, and my own personal leanings, I don't share OP's original view.<mask>, I am curious to figure out whether I can respond to OP's question without limiting the scope to a "father vs. child" frame, without immediately dismissing premise 1 (<mask><mask> I personally do disagree with premise 1). [NEWLINE] [NEWLINE] <mask> my "devil's advocate" question is geared a bit differently.<mask> you accept premise 1<mask> a constraint (i.e.<mask> you look at this from the perspective of a strict Thatcherite/individualist, for lack of a more accurate term), then it seems to me that you inevitably end up with the quandary that OP presented.<mask> you posit that there must be completely equal rights/responsibilities for father and mother (which due to constraint 1 actually trump the future questions of the child's rights/responsibilities/entitlements/choices), you dictate the conclusion that OP presented. [NEWLINE] [NEWLINE] Fundamentally, I guess<mask> I'm wondering is this: is there an argument against OP's position (<mask> I've restated it) that doesn't demand a full reconsideration of the underlying assumption--even a complete reconsideration of your world view? Stated another way,<mask> you are a devoted Thatcherite/individualist, can you reconcile your devotion to individualism in a way that results in a different conclusion (different from "<mask> a prospective mother can abort, a prospective father should be able to opt out of support payments")? Or is it instead inevitable that Thatcherite/individualists will see things the way OP does, and that people from a more holistic perspective will see things your way, with neither able to reach the other's conclusions? [NEWLINE] [NEWLINE] I'm not sure I'm explaining my question clearly,<mask> feel free to share your thoughts either way.</s>
Label encoding: <s>Hey, thanks for the response! I happen to completely agree with what you're laying down - my personal view is in line with what you're saying. With family background in health, and my own personal leanings, I don't share OP's original view. However, I am curious to figure out whether I can respond to OP's question without limiting the scope to a "father vs. child" frame, without immediately dismissing premise 1 ( even though I personally do disagree with premise 1). [NEWLINE] [NEWLINE] So my "devil's advocate" question is geared a bit differently. If you accept premise 1 as a constraint (i.e. if you look at this from the perspective of a strict Thatcherite/individualist, for lack of a more accurate term), then it seems to me that you inevitably end up with the quandary that OP presented. If you posit that there must be completely equal rights/responsibilities for father and mother (which due to constraint 1 actually trump the future questions of the child's rights/responsibilities/entitlements/choices), you dictate the conclusion that OP presented. [NEWLINE] [NEWLINE] Fundamentally, I guess what I'm wondering is this: is there an argument against OP's position ( as I've restated it) that doesn't demand a full reconsideration of the underlying assumption--even a complete reconsideration of your world view? Stated another way, if you are a devoted Thatcherite/individualist, can you reconcile your devotion to individualism in a way that results in a different conclusion (different from " if a prospective mother can abort, a prospective father should be able to opt out of support payments")? Or is it instead inevitable that Thatcherite/individualists will see things the way OP does, and that people from a more holistic perspective will see things your way, with neither able to reach the other's conclusions? [NEWLINE] [NEWLINE] I'm not sure I'm explaining my question clearly, but feel free to share your thoughts either way.</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The internet (or at least Reddit) is filled with grammar Nazis, people who point out spelling and grammatical errors with a passion, no matter<mask> minute and insignificant. I hate seeing a legitimately interesting comment get 100 upvotes about the current topic and the comment below it get 500 upvotes for simply pointing out that he used the wrong "there". [NEWLINE] [NEWLINE] The point of language is to convey meaning in<mask> clear and easy way<mask> possible. That's<mask> we have structure. It wouldn't make sense to have no rules and no organization to English. I get that.<mask> is our understanding of<mask> you're trying to say really diminished<mask> you said "loose"<mask> you really meant "lose". At most, it took a half second to realize, "oh, he meant the other word". We're not writing a thesis for a phd, we're talking on an internet forum. It's the equivalent of correcting someone's grammar<mask> talking with casually talking with strangers. It's pointless and most people who do it are being petty. [NEWLINE] [NEWLINE] P.S. I'm sure there's some errors in this post. I tried. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>The internet (or at least Reddit) is filled with grammar Nazis, people who point out spelling and grammatical errors with a passion, no matter how minute and insignificant. I hate seeing a legitimately interesting comment get 100 upvotes about the current topic and the comment below it get 500 upvotes for simply pointing out that he used the wrong "there". [NEWLINE] [NEWLINE] The point of language is to convey meaning in as clear and easy way as possible. That's why we have structure. It wouldn't make sense to have no rules and no organization to English. I get that. But is our understanding of what you're trying to say really diminished because you said "loose" when you really meant "lose". At most, it took a half second to realize, "oh, he meant the other word". We're not writing a thesis for a phd, we're talking on an internet forum. It's the equivalent of correcting someone's grammar when talking with casually talking with strangers. It's pointless and most people who do it are being petty. [NEWLINE] [NEWLINE] P.S. I'm sure there's some errors in this post. I tried. [NEWLINE] [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's a different way to think about it. [NEWLINE] [NEWLINE] To a lot of people, the United States is a great place to live. Everyone has different expectations and desires, and for some, the United States meets those criteria. [NEWLINE] [NEWLINE] <mask> the United States does not meet your criteria (and I am not suggesting there is anything wrong with it), then you do not have to move there. A lot of the romanticism of living in the United States is not based on the criteria you mentioned (government problems, healthcare costs and social safety nets). Some people love the great expanse of a nation the United States is, with tons of variety<mask> it comes to land, people and cultures. To them, these factors are more important than<mask> their healthcare situation is handled. [NEWLINE] [NEWLINE] Take<mask> others have to say with a grain of salt and experience it for yourself (which you have,<mask> I read that you visited the United States). There is no one way to live your life, and<mask> the United States does not have<mask> you're looking for, then<mask> be it. For others, it does have<mask> they're looking for. [NEWLINE] [NEWLINE] We cannot generalize and assume that everyone wants to live somewhere based on<mask> some poll or<mask> our friends say.<mask><mask> you say that all you've ever heard is that people want to move to the United States, that doesn't mean those folks have the same expectations of their home country<mask> the next guy. [NEWLINE] [NEWLINE] Instead of talking about the issues with the United States and<mask> they make it a place you don't want to live, think more about<mask> people choose to say that the United States is a place they want to move to. They may not have the same expectations<mask> the rest of us. [NEWLINE] [NEWLINE] <mask> in closing, I do not want to change your view. Instead, I challenge others to think about each other and understand that everyone is different, and that their reasoning to move somewhere does not always match that of our own.</s>
Label encoding: <s>Here's a different way to think about it. [NEWLINE] [NEWLINE] To a lot of people, the United States is a great place to live. Everyone has different expectations and desires, and for some, the United States meets those criteria. [NEWLINE] [NEWLINE] If the United States does not meet your criteria (and I am not suggesting there is anything wrong with it), then you do not have to move there. A lot of the romanticism of living in the United States is not based on the criteria you mentioned (government problems, healthcare costs and social safety nets). Some people love the great expanse of a nation the United States is, with tons of variety when it comes to land, people and cultures. To them, these factors are more important than how their healthcare situation is handled. [NEWLINE] [NEWLINE] Take what others have to say with a grain of salt and experience it for yourself (which you have, as I read that you visited the United States). There is no one way to live your life, and if the United States does not have what you're looking for, then so be it. For others, it does have what they're looking for. [NEWLINE] [NEWLINE] We cannot generalize and assume that everyone wants to live somewhere based on what some poll or what our friends say. Even though you say that all you've ever heard is that people want to move to the United States, that doesn't mean those folks have the same expectations of their home country as the next guy. [NEWLINE] [NEWLINE] Instead of talking about the issues with the United States and why they make it a place you don't want to live, think more about why people choose to say that the United States is a place they want to move to. They may not have the same expectations as the rest of us. [NEWLINE] [NEWLINE] So in closing, I do not want to change your view. Instead, I challenge others to think about each other and understand that everyone is different, and that their reasoning to move somewhere does not always match that of our own.</s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I mean, I loved the calm music. It made me feel like Steve had some kind of bad past that he was escaping to begin MINECRAFT. I used to love the mobs, they made me scared,<mask> now they just seem flawed and childish. I used to love the Ender Dragon,<mask> now I just see it<mask> a type of clickbait that keeps you playing for hours killing endermen. I used to want to create the best garden after I made one on XBOX (my disk broke later) and remembered that I had more materials available on my PC copy. I used to love the mods,<mask> the fact that an API even<mask> basic<mask> Skyrim's hasn't been made after 3 years of being announced is annoying. I used to love fighting with swords,<mask> now I can only see the animation<mask> wimpy,<mask><mask> the sword just taps the enemy. I wanted to create the greatest looking house in survival,<mask> once I begin, I see it<mask> pointless. I used to want to explore the world,<mask> now I realize it makes my PC slow by making the save larger and having it use a lot of RAM (not exactly like that,<mask> exploring the world does make that world run slower). I wanted to create the largest village, then I realized there was no point to it<mask> villagers treat me like crap. I used to love the blocky style, first it looked like legos, then it looked like art, now it looks annoying. I used to love building,<mask> now I can't seem to know<mask> I want to build. Servers are full of annoying chat spamming kids acting like adults, and 8 year old moderators who block items until you pay them money. Creative servers that are advertised online by nice people get destroyed by trolls, and no servers allow you to help someone build something cool without a troll destroying it. Herobrine creepypasta was run too long. Admit it doesn't exist already!</s><pad>
Label encoding: <s>What I mean, I loved the calm music. It made me feel like Steve had some kind of bad past that he was escaping to begin MINECRAFT. I used to love the mobs, they made me scared, but now they just seem flawed and childish. I used to love the Ender Dragon, but now I just see it as a type of clickbait that keeps you playing for hours killing endermen. I used to want to create the best garden after I made one on XBOX (my disk broke later) and remembered that I had more materials available on my PC copy. I used to love the mods, but the fact that an API even as basic as Skyrim's hasn't been made after 3 years of being announced is annoying. I used to love fighting with swords, but now I can only see the animation as wimpy, as if the sword just taps the enemy. I wanted to create the greatest looking house in survival, but once I begin, I see it as pointless. I used to want to explore the world, but now I realize it makes my PC slow by making the save larger and having it use a lot of RAM (not exactly like that, but exploring the world does make that world run slower). I wanted to create the largest village, then I realized there was no point to it because villagers treat me like crap. I used to love the blocky style, first it looked like legos, then it looked like art, now it looks annoying. I used to love building, but now I can't seem to know what I want to build. Servers are full of annoying chat spamming kids acting like adults, and 8 year old moderators who block items until you pay them money. Creative servers that are advertised online by nice people get destroyed by trolls, and no servers allow you to help someone build something cool without a troll destroying it. Herobrine creepypasta was run too long. Admit it doesn't exist already!</s><pad>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's not about words other languages have borrowed from English (<mask> there are many)<mask> about the words English has, itself, borrowed. The high amount of words within the English language that resemble words in other languages means that a speaker is more likely to intuitively grasp<mask> a word means and<mask> it may be used, versus<mask> something were entirely foreign to them. [ENDQ] [NEWLINE] <mask> English had stuck to its Germanic roots, there would be more of that with Germanic languages and less of it with Latin languages.  I would say it's more a feature of certain languages being closely related to English than a consequence of English's inconsistency. [NEWLINE] [NEWLINE] [STARTQ] <mask> Mandarin Chinese purged all words from itself save that which it derived natively, then yes, I would say that. Fortunately, that is far from the case. Mandarin Chinese has expanded its conceptual horizons through the adoption of words like "Yuppie," "Cartoon," and "Cool" (Fonzie cool, not Mr. Freeze cool).<mask><mask> it's pretty neat that these words have crossed over. On the other side of things, we can thank Chinese for words like "brainwashing," "kowtow," and the phrase "long time no see."<mask>, I myself am fond of using the word "Mandarin" to describe far off functionaries with little understanding of on-the-ground realities<mask> that's getting a little fancy. [ENDQ] [NEWLINE] They use loanwords for food and pop culture stuff that Chinese has no equivalent of,<mask> they use native words for most things, even a lot of new science and technology vocabulary. [NEWLINE] [NEWLINE] [STARTQ] And,<mask>'s more, I'll bet Korean probably has its own weird peculiarities that leave even native speakers scratching their head. [ENDQ] [NEWLINE] Almost certainly. <mask> unless there's some weird "conservation of complexity" going on (and sometimes it does seem like that's the case), then there are surely languages with more peculiarities than others.</s>
Label encoding: <s> [STARTQ] It's not about words other languages have borrowed from English ( though there are many) but about the words English has, itself, borrowed. The high amount of words within the English language that resemble words in other languages means that a speaker is more likely to intuitively grasp what a word means and how it may be used, versus if something were entirely foreign to them. [ENDQ] [NEWLINE] If English had stuck to its Germanic roots, there would be more of that with Germanic languages and less of it with Latin languages.  I would say it's more a feature of certain languages being closely related to English than a consequence of English's inconsistency. [NEWLINE] [NEWLINE] [STARTQ] If Mandarin Chinese purged all words from itself save that which it derived natively, then yes, I would say that. Fortunately, that is far from the case. Mandarin Chinese has expanded its conceptual horizons through the adoption of words like "Yuppie," "Cartoon," and "Cool" (Fonzie cool, not Mr. Freeze cool). I think it's pretty neat that these words have crossed over. On the other side of things, we can thank Chinese for words like "brainwashing," "kowtow," and the phrase "long time no see." Also, I myself am fond of using the word "Mandarin" to describe far off functionaries with little understanding of on-the-ground realities but that's getting a little fancy. [ENDQ] [NEWLINE] They use loanwords for food and pop culture stuff that Chinese has no equivalent of, but they use native words for most things, even a lot of new science and technology vocabulary. [NEWLINE] [NEWLINE] [STARTQ] And, what's more, I'll bet Korean probably has its own weird peculiarities that leave even native speakers scratching their head. [ENDQ] [NEWLINE] Almost certainly.  But unless there's some weird "conservation of complexity" going on (and sometimes it does seem like that's the case), then there are surely languages with more peculiarities than others.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I do not see the appeal in Cake, whatever day it is. There is always way too much frosting, making the supposedly delicious treat leaving you overdosing with Sugar. Most of the time, whenever I am around and someone wants to whip out a cake, it is always some monstrosity of a pastry, like Carrot Cake, or with some ooey gooey disgusting filler that does not know<mask> it is fake or expired fruit.<mask> Cake makes you [feel bad]( [URL].jpg), too much pie is never a bad thing. Just ask [Piderman]( [URL].jpg)! [NEWLINE] [NEWLINE] In my view, [Pie is vastly superior]( [URL] /#!gallery/cff9). You can have a dessert that actually works with fruits, such<mask> Apple Pie, Pumpkin Pie, or my personal favorites, Blackberry and Olallieberry pie. We even eat Pie at [work]( [URL] )! Well, in between [BAWLS]( [URL] /) runs. [NEWLINE] [NEWLINE] <mask> CMV, guys, the only thing better than [Pie]( [URL].gif) is [Pie and Baseball]( [URL].jpg). [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I do not see the appeal in Cake, whatever day it is. There is always way too much frosting, making the supposedly delicious treat leaving you overdosing with Sugar. Most of the time, whenever I am around and someone wants to whip out a cake, it is always some monstrosity of a pastry, like Carrot Cake, or with some ooey gooey disgusting filler that does not know if it is fake or expired fruit. Where Cake makes you [feel bad]( [URL].jpg), too much pie is never a bad thing. Just ask [Piderman]( [URL].jpg)! [NEWLINE] [NEWLINE] In my view, [Pie is vastly superior]( [URL] /#!gallery/cff9). You can have a dessert that actually works with fruits, such as Apple Pie, Pumpkin Pie, or my personal favorites, Blackberry and Olallieberry pie. We even eat Pie at [work]( [URL] )! Well, in between [BAWLS]( [URL] /) runs. [NEWLINE] [NEWLINE] So CMV, guys, the only thing better than [Pie]( [URL].gif) is [Pie and Baseball]( [URL].jpg). [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you depend on an animal like a horse, cow, or working dog for your livelihood, then it makes sense to take that animal to the vet<mask> they get sick.<mask><mask> your animal is just a pet - meaning it provides little to no value beyond companionship and entertainment - then to spend large amounts of money to fix the pet's ailments is a waste of money.<mask><mask> that pet owners, upon purchase of the pet, need to accept the fact that the pet is going to die, and that this event will be sad. Throwing thousands of dollars after delaying that event is irrational.<mask> your pet dies and you can't live without a pet, get another one. [NEWLINE] [NEWLINE] I cannot convince myself that there is some moral imperative to fix pets. This does not mean that I don't think there's a moral imperative to feed them. I do believe we assume responsibility for the well-being of an animal once we take it under our care. I just don't believe that extends to medical treatment. The idea that a dog has some sort of right to medical treatment is totally absurd to me. [NEWLINE] [NEWLINE] CMV [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than just downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>If you depend on an animal like a horse, cow, or working dog for your livelihood, then it makes sense to take that animal to the vet if they get sick. But if your animal is just a pet - meaning it provides little to no value beyond companionship and entertainment - then to spend large amounts of money to fix the pet's ailments is a waste of money. I think that pet owners, upon purchase of the pet, need to accept the fact that the pet is going to die, and that this event will be sad. Throwing thousands of dollars after delaying that event is irrational. If your pet dies and you can't live without a pet, get another one. [NEWLINE] [NEWLINE] I cannot convince myself that there is some moral imperative to fix pets. This does not mean that I don't think there's a moral imperative to feed them. I do believe we assume responsibility for the well-being of an animal once we take it under our care. I just don't believe that extends to medical treatment. The idea that a dog has some sort of right to medical treatment is totally absurd to me. [NEWLINE] [NEWLINE] CMV [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than just downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>At issue is that your citizenship is, in some ways, a sort of contract. [NEWLINE] [NEWLINE] You agree to not violate the laws of your country, including operating against your nation's international interests, and they agree to provide you protection anywhere in the world on the basis of your passport. [NEWLINE] [NEWLINE] Police chasing you in Russia for a crime you didn't commit? No worries, mate -- just run into the nearest embassy and you'll have a fairly stout legal defense being made on your behalf. [NEWLINE] [NEWLINE] Happen to be vacationing in Thailand<mask> a tsunami hits? No problem, your passport means you go to the front of the line for help. [NEWLINE] [NEWLINE] Volunteering in Africa and come down with a strange disease? Not an issue, call the foreign office and you'll no longer be stuck in a back-water hospital. [NEWLINE] [NEWLINE] And the list goes on. [NEWLINE] [NEWLINE] The passport you hold means you are entitled on a large number of services. And in payment for those services you pay your taxes and obey the laws of your country. [NEWLINE] [NEWLINE] <mask> you go join a foreign Army without first revoking your citizenship, you are trying to have your cake and eat it too. You want to work against your country's national interests,<mask> still have them provide you protections around the world. [NEWLINE] [NEWLINE] That's not cricket. [NEWLINE] [NEWLINE] It is an entirely sensible policy to say: [NEWLINE] [NEWLINE] We do not want your actions overseas to in anyway cause blow-back to our international interests and relationships. To that end,<mask> you decide to provide material support to a foreign government either through service or some other means, without prior approval of our government, you're violating a law. The purpose of the law is to protect all of our country from becoming unintentionally embroiled in an international political controversy.<mask> you want to go fight for another nation without violating the law, you need to formally give up your citizenship first<mask> that your actions can in no way create issues for our government. </s>
Label encoding: <s>At issue is that your citizenship is, in some ways, a sort of contract. [NEWLINE] [NEWLINE] You agree to not violate the laws of your country, including operating against your nation's international interests, and they agree to provide you protection anywhere in the world on the basis of your passport. [NEWLINE] [NEWLINE] Police chasing you in Russia for a crime you didn't commit? No worries, mate -- just run into the nearest embassy and you'll have a fairly stout legal defense being made on your behalf. [NEWLINE] [NEWLINE] Happen to be vacationing in Thailand when a tsunami hits? No problem, your passport means you go to the front of the line for help. [NEWLINE] [NEWLINE] Volunteering in Africa and come down with a strange disease? Not an issue, call the foreign office and you'll no longer be stuck in a back-water hospital. [NEWLINE] [NEWLINE] And the list goes on. [NEWLINE] [NEWLINE] The passport you hold means you are entitled on a large number of services. And in payment for those services you pay your taxes and obey the laws of your country. [NEWLINE] [NEWLINE] When you go join a foreign Army without first revoking your citizenship, you are trying to have your cake and eat it too. You want to work against your country's national interests, but still have them provide you protections around the world. [NEWLINE] [NEWLINE] That's not cricket. [NEWLINE] [NEWLINE] It is an entirely sensible policy to say: [NEWLINE] [NEWLINE] We do not want your actions overseas to in anyway cause blow-back to our international interests and relationships. To that end, if you decide to provide material support to a foreign government either through service or some other means, without prior approval of our government, you're violating a law. The purpose of the law is to protect all of our country from becoming unintentionally embroiled in an international political controversy. If you want to go fight for another nation without violating the law, you need to formally give up your citizenship first so that your actions can in no way create issues for our government. </s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Have you tried eating most Asian food with chopsticks?  In most cases, it's actually a lot easier (once you've gotten the chopstick technique down) to eat the food with the sticks than it would be with a knife / fork.  I was raised in the west,<mask> lived with a Vietnamese family for some time,<mask> I've had ample experience with both utensil styles and can say that chopsticks are just easier to use in some cases. [NEWLINE] [NEWLINE] The vast majority of Asian dishes do not require a knife,<mask> they're typically designed with chopsticks / frying in mind and<mask> anything that would need to be cut in a western dish is generally already pre-cut into smaller, more manageable portions.  Most authentic Asian dishes are shared dishes.  With the typical, weekday, non-occasion meal (similar to something like meat and potatoes in the west) the family will sit down at the table, each with a small bowl of rice in front of them.  The mom will have cooked one or two plates of small, pre-cut meat, and another one or two plates of small, pre-cut veggies.  The family then takes<mask> they want from each plate<mask> they eat.  Using a spoon / fork to do<mask> would actually be more difficult, and considerably less elegant. [NEWLINE] [NEWLINE] Really,<mask> it comes down to is the culture and cuisine each utensil was designed for.  The knife and fork are Western utensils designed to deal with all the ripping, tearing, cutting, and stabbing required to eat most of the meat-centric dishes in the west.  The chopstick was designed to deal with all the small-portioned, rice / noodle centric dishes in the east.  It's certainly possible to eat most Asian dishes with a knife / fork / spoon,<mask> in some cases it's more difficult, and in most cases it's just not<mask> fun<mask> using chopsticks.</s>
Label encoding: <s>Have you tried eating most Asian food with chopsticks?  In most cases, it's actually a lot easier (once you've gotten the chopstick technique down) to eat the food with the sticks than it would be with a knife / fork.  I was raised in the west, but lived with a Vietnamese family for some time, so I've had ample experience with both utensil styles and can say that chopsticks are just easier to use in some cases. [NEWLINE] [NEWLINE] The vast majority of Asian dishes do not require a knife, as they're typically designed with chopsticks / frying in mind and thus anything that would need to be cut in a western dish is generally already pre-cut into smaller, more manageable portions.  Most authentic Asian dishes are shared dishes.  With the typical, weekday, non-occasion meal (similar to something like meat and potatoes in the west) the family will sit down at the table, each with a small bowl of rice in front of them.  The mom will have cooked one or two plates of small, pre-cut meat, and another one or two plates of small, pre-cut veggies.  The family then takes what they want from each plate as they eat.  Using a spoon / fork to do so would actually be more difficult, and considerably less elegant. [NEWLINE] [NEWLINE] Really, what it comes down to is the culture and cuisine each utensil was designed for.  The knife and fork are Western utensils designed to deal with all the ripping, tearing, cutting, and stabbing required to eat most of the meat-centric dishes in the west.  The chopstick was designed to deal with all the small-portioned, rice / noodle centric dishes in the east.  It's certainly possible to eat most Asian dishes with a knife / fork / spoon, but in some cases it's more difficult, and in most cases it's just not as fun as using chopsticks.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ]...<mask> basically everybody criminal ever. That does wonders in selling me on the idea of a democratic, uncontrollable currency, I'm sorry to say. [ENDQ] [NEWLINE] It's much better/worse than that. [NEWLINE] [NEWLINE] Services and extensions to the protocol that turn bitcoin into a laundering service are on the drawing boards. Zerocoin is one, there are several others. It is highly likely that one of these will become a standard part of the network and true anonymity in the act of spending will become<mask> easy<mask> the click of a mouse button. [NEWLINE] [NEWLINE] Fifty people spend a coin, the network mixes up all fifty of these coins internally, and spends them randomly to the fifty intended targets. The bitcoin you had went in, got mixed, and was spent by a complete stranger you've never met. In return, some other stranger's coin was spent on your behalf. [NEWLINE] [NEWLINE] Note<mask> that the *volume* of the money being laundered is the key limiting factor. A little bit is easy. A huge amount is extremely hard,<mask><mask> practically impossible, unless done slowly over a long period of time, bit by bit. The more value bitcoins have, and the more people who use them, the easier it becomes to launder larger amounts. [NEWLINE] [NEWLINE] I see this laundering service<mask> a major driver for adoption. Criminals love it, and criminals account for a lot of money. Bitcoin does not understand or care about 'criminal' activity - just like cash. All it knows are transactions. It is utterly neutral. This is<mask> there remains a role for regulation to play. [NEWLINE] [NEWLINE] The regulators are going to have their work cut out for them,<mask> a *legal* solution is impractical. It must be a *technical* solution. Bitcoin is out of the bottle and growing at a fantastic rate. Putting it back requires turning the net off. All the laws in the world will have just<mask> much effect on bitcoin<mask> copyright law does on piracy.</s>
Label encoding: <s> [STARTQ]... so basically everybody criminal ever. That does wonders in selling me on the idea of a democratic, uncontrollable currency, I'm sorry to say. [ENDQ] [NEWLINE] It's much better/worse than that. [NEWLINE] [NEWLINE] Services and extensions to the protocol that turn bitcoin into a laundering service are on the drawing boards. Zerocoin is one, there are several others. It is highly likely that one of these will become a standard part of the network and true anonymity in the act of spending will become as easy as the click of a mouse button. [NEWLINE] [NEWLINE] Fifty people spend a coin, the network mixes up all fifty of these coins internally, and spends them randomly to the fifty intended targets. The bitcoin you had went in, got mixed, and was spent by a complete stranger you've never met. In return, some other stranger's coin was spent on your behalf. [NEWLINE] [NEWLINE] Note however that the *volume* of the money being laundered is the key limiting factor. A little bit is easy. A huge amount is extremely hard, in fact practically impossible, unless done slowly over a long period of time, bit by bit. The more value bitcoins have, and the more people who use them, the easier it becomes to launder larger amounts. [NEWLINE] [NEWLINE] I see this laundering service as a major driver for adoption. Criminals love it, and criminals account for a lot of money. Bitcoin does not understand or care about 'criminal' activity - just like cash. All it knows are transactions. It is utterly neutral. This is why there remains a role for regulation to play. [NEWLINE] [NEWLINE] The regulators are going to have their work cut out for them, because a *legal* solution is impractical. It must be a *technical* solution. Bitcoin is out of the bottle and growing at a fantastic rate. Putting it back requires turning the net off. All the laws in the world will have just as much effect on bitcoin as copyright law does on piracy.</s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> type of insurance you're specifically referring to.<mask> insurance in general is used to cover losses - especially on things you cannot afford. [NEWLINE] [NEWLINE] <mask><mask> with you, that you would not insure something that is worth very little.<mask>, something like auto insurance could lead to protecting you from claims that would otherwise bankrupt you. [NEWLINE] [NEWLINE] You're measuring a cost benefit analysis. Odds of using vs benefit from using.<mask>, that is not the way it needs to be looked at. [NEWLINE] [NEWLINE] Let me demonstrate with a numerical example: [NEWLINE] [NEWLINE] Assume: [NEWLINE] your net worth is $100000 [NEWLINE] <mask> you get into a car accident, the damages will be over $100000 [NEWLINE] the probability of you getting into a car accident is.1% [NEWLINE] [NEWLINE] A car insurance company would charge you more than.1% * 100000 = $100 to insure you. Your implication (which<mask><mask> with) is that the more than $100 (let's assume $200 for the rest of my post) they will charge for insurance is higher than fair. [NEWLINE] [NEWLINE] Here is<mask> I don't agree with you: [NEWLINE] With Insurance: Net worth = $100000 - $200 = $99800 [NEWLINE] No Accident (99.9% chance): Net worth = $99800 [NEWLINE] Accident (.1% chance): Net worth = $99800 [NEWLINE] [NEWLINE] Without Insurance: Net worth = $100000 [NEWLINE] No Accident (99.9% chance): Net worth = $100000 [NEWLINE] Accident (.1% chance): Net worth = $0 [NEWLINE] [NEWLINE] Would you not rather have a guaranteed net worth of $99800 (instead of $100000) than risk the chance of being completely bankrupt? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] That all having been said,<mask> the numbers are the same and your net worth is 100 million instead of 100000 I can see your perspective<mask> the worst case scenario would not put a large dent in your lifestyle</s>
Label encoding: <s>I'm not sure what type of insurance you're specifically referring to. But insurance in general is used to cover losses - especially on things you cannot afford. [NEWLINE] [NEWLINE] I agree with you, that you would not insure something that is worth very little. However, something like auto insurance could lead to protecting you from claims that would otherwise bankrupt you. [NEWLINE] [NEWLINE] You're measuring a cost benefit analysis. Odds of using vs benefit from using. However, that is not the way it needs to be looked at. [NEWLINE] [NEWLINE] Let me demonstrate with a numerical example: [NEWLINE] [NEWLINE] Assume: [NEWLINE] your net worth is $100000 [NEWLINE] if you get into a car accident, the damages will be over $100000 [NEWLINE] the probability of you getting into a car accident is.1% [NEWLINE] [NEWLINE] A car insurance company would charge you more than.1% * 100000 = $100 to insure you. Your implication (which I agree with) is that the more than $100 (let's assume $200 for the rest of my post) they will charge for insurance is higher than fair. [NEWLINE] [NEWLINE] Here is why I don't agree with you: [NEWLINE] With Insurance: Net worth = $100000 - $200 = $99800 [NEWLINE] No Accident (99.9% chance): Net worth = $99800 [NEWLINE] Accident (.1% chance): Net worth = $99800 [NEWLINE] [NEWLINE] Without Insurance: Net worth = $100000 [NEWLINE] No Accident (99.9% chance): Net worth = $100000 [NEWLINE] Accident (.1% chance): Net worth = $0 [NEWLINE] [NEWLINE] Would you not rather have a guaranteed net worth of $99800 (instead of $100000) than risk the chance of being completely bankrupt? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] That all having been said, if the numbers are the same and your net worth is 100 million instead of 100000 I can see your perspective as the worst case scenario would not put a large dent in your lifestyle</s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's obviously true that women have more reproductive choices than men<mask> you must admit that this post is a massive exaggeration. You said [NEWLINE] [NEWLINE] [STARTQ] men have essentially no choice other then celibacy [ENDQ] [NEWLINE] <mask> that isn't true. Men can choose to use condoms or get a vasectomy.<mask>, almost all adoption and safe haven laws are written in a gender neutral fashion. Legally speaking, a woman can not unilaterally decide to give a child up for adoption. The father can sue for paternity and take custody of the child, at which point he can request child support from the mother. [NEWLINE] [NEWLINE] <mask>, in almost all states (all<mask> 3 or 4 I believe) a father can<mask> take advantage of safe haven laws<mask>, for instance, the mother were to disappear shortly after delivering the baby. Furthermore, the father can<mask> take custody of his child<mask> the mother leaves it at a safe haven without his permission and collect child support payments from her. [NEWLINE] [NEWLINE] Yes, the nature of birth means that it is far more likely for women to take advantage of these options than men.<mask>, that is usually<mask> it is rare for the father to end up with sole custody of a newborn baby with the mother nowhere to be found<mask>, you know, the kid just fell out of her vagina. [NEWLINE] [NEWLINE] It is<mask> true that it is possible for a woman to give birth to a child without informing the father that she was ever pregnant, in which case he would obviously not know that he could assert his parental rights.<mask>, there isn't really a way of fixing this short of interrogating women who just gave birth and forcing them to reveal the father's identity.<mask>,<mask> the father does happen to learn of the existence of his child at some later date he can often assert his parental rights and sue for custody. Adoption agencies are<mask> supposed to take reasonable measures to determine the paternity of children that they receive,<mask> of course this is usually essentially impossible.</s>
Label encoding: <s>It's obviously true that women have more reproductive choices than men but you must admit that this post is a massive exaggeration. You said [NEWLINE] [NEWLINE] [STARTQ] men have essentially no choice other then celibacy [ENDQ] [NEWLINE] But that isn't true. Men can choose to use condoms or get a vasectomy. Also, almost all adoption and safe haven laws are written in a gender neutral fashion. Legally speaking, a woman can not unilaterally decide to give a child up for adoption. The father can sue for paternity and take custody of the child, at which point he can request child support from the mother. [NEWLINE] [NEWLINE] Also, in almost all states (all but 3 or 4 I believe) a father can also take advantage of safe haven laws if, for instance, the mother were to disappear shortly after delivering the baby. Furthermore, the father can also take custody of his child if the mother leaves it at a safe haven without his permission and collect child support payments from her. [NEWLINE] [NEWLINE] Yes, the nature of birth means that it is far more likely for women to take advantage of these options than men. But, that is usually because it is rare for the father to end up with sole custody of a newborn baby with the mother nowhere to be found since, you know, the kid just fell out of her vagina. [NEWLINE] [NEWLINE] It is also true that it is possible for a woman to give birth to a child without informing the father that she was ever pregnant, in which case he would obviously not know that he could assert his parental rights. However, there isn't really a way of fixing this short of interrogating women who just gave birth and forcing them to reveal the father's identity. However, if the father does happen to learn of the existence of his child at some later date he can often assert his parental rights and sue for custody. Adoption agencies are also supposed to take reasonable measures to determine the paternity of children that they receive, though of course this is usually essentially impossible.</s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am 23, in my 4th year of college (not senior year, I came back for a certificate program) across 2 different schools. I went to a vocational school my last 2 years of high school<mask> they normally sent the trouble kids (drinkers, druggies and partiers)<mask> the home school didn't have to deal with them. And I am a teetotaler. I have tried alcohol in very small amounts and never liked it, and am actually opposed to its prevalence in society. Oh, and I do have social anxiety. [NEWLINE] [NEWLINE] With that background, I actually have to disagree with you. I have had many meaningful friendships and relationships without drinking. Some of my friends do drink, most are just courteous enough to not do it around me.<mask> they have a beer or whatever in my presence, I don't mind,<mask> they still know and understand that I wont join them. I'm<mask> not opposed to being a DD on occasion (usually birthdays or special events, not on a weekly basis). [NEWLINE] [NEWLINE] I do not associate with drunkards<mask> I find them to be poor company. I do not associate with people that do not support my decision not to drink. And I do not associate with people that do illicit drugs (not part of the main topic,<mask> still very related, at least in my area). [NEWLINE] With those rules in place, a lot of people are rejected<mask><mask> is left are the types of people I want to associate with. In my case, it ended up being people that I can fun with. Best of all, I can have an intense intellectual or philosophical conversation/debate with, something that I've never found possible with people who require drinks in every form of socialization. [NEWLINE] [NEWLINE] In other words, you just need to separate the wheat from the chaff; find the people that you want to socialize with instead of the ones that just want to drink.</s>
Label encoding: <s>I am 23, in my 4th year of college (not senior year, I came back for a certificate program) across 2 different schools. I went to a vocational school my last 2 years of high school where they normally sent the trouble kids (drinkers, druggies and partiers) so the home school didn't have to deal with them. And I am a teetotaler. I have tried alcohol in very small amounts and never liked it, and am actually opposed to its prevalence in society. Oh, and I do have social anxiety. [NEWLINE] [NEWLINE] With that background, I actually have to disagree with you. I have had many meaningful friendships and relationships without drinking. Some of my friends do drink, most are just courteous enough to not do it around me. If they have a beer or whatever in my presence, I don't mind, but they still know and understand that I wont join them. I'm also not opposed to being a DD on occasion (usually birthdays or special events, not on a weekly basis). [NEWLINE] [NEWLINE] I do not associate with drunkards as I find them to be poor company. I do not associate with people that do not support my decision not to drink. And I do not associate with people that do illicit drugs (not part of the main topic, but still very related, at least in my area). [NEWLINE] With those rules in place, a lot of people are rejected but what is left are the types of people I want to associate with. In my case, it ended up being people that I can fun with. Best of all, I can have an intense intellectual or philosophical conversation/debate with, something that I've never found possible with people who require drinks in every form of socialization. [NEWLINE] [NEWLINE] In other words, you just need to separate the wheat from the chaff; find the people that you want to socialize with instead of the ones that just want to drink.</s>
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Don't you worry that your life is the poorer for this? [ENDQ] [NEWLINE] Not at all.<mask><mask>, this is the first time my thoughts have ever even gone in that direction<mask> thumbs up for showing me a previously unknown perspective. :) [NEWLINE] [NEWLINE] [STARTQ] <mask> the people who posted homophobic crap instead kept it to themselves, perhaps you and they could be friends. [ENDQ] [NEWLINE] And<mask><mask> I'm bi, I would have been friends with people who think I should be a second class citizen or burn in hell. Naturally, I don't think I need those people to be my friends. Nor would they be<mask> that's<mask> they think of me and people like me. [NEWLINE] [NEWLINE] [STARTQ] Perhaps your friendship might change their homophobia, and either way, perhaps your relationship could be good<mask> the political differences. [ENDQ] [NEWLINE] I don't think I have the obligation to change their views.<mask>, having witnessed the nature and ferocity of their posts, I came to the conclusion they belong in the "not worth it" category and simply removed them (there's<mask> that option<mask> you don't have to de-friend someone, you can just hide their posts). Some political differences are more important than others and can affect relationships beyond repair. [NEWLINE] [NEWLINE] [STARTQ] Is this a fair assumption? I know multiple people who I consider friends (and without whom my life would be poorer) who cannot do this. [ENDQ] [NEWLINE] Have you discussed it with them? Have you tried to show them a different perspective? [NEWLINE] [NEWLINE] [STARTQ] in<mask> way you are actually living a better life by posting. [ENDQ] [NEWLINE] By not stressing over things like that. By not spending time deciding whether or not I'm too afraid to say<mask><mask><mask><mask> someone might not like me. By staying true to myself and by drawing the attention of others to issues that are important to me,<mask> allowing them to know me better and<mask> maybe change their views for the better<mask> well. [NEWLINE] [NEWLINE] Edit: a word </s><pad>
Label encoding: <s> [STARTQ] Don't you worry that your life is the poorer for this? [ENDQ] [NEWLINE] Not at all. In fact, this is the first time my thoughts have ever even gone in that direction so thumbs up for showing me a previously unknown perspective. :) [NEWLINE] [NEWLINE] [STARTQ] If the people who posted homophobic crap instead kept it to themselves, perhaps you and they could be friends. [ENDQ] [NEWLINE] And given that I'm bi, I would have been friends with people who think I should be a second class citizen or burn in hell. Naturally, I don't think I need those people to be my friends. Nor would they be if that's how they think of me and people like me. [NEWLINE] [NEWLINE] [STARTQ] Perhaps your friendship might change their homophobia, and either way, perhaps your relationship could be good despite the political differences. [ENDQ] [NEWLINE] I don't think I have the obligation to change their views. Also, having witnessed the nature and ferocity of their posts, I came to the conclusion they belong in the "not worth it" category and simply removed them (there's also that option where you don't have to de-friend someone, you can just hide their posts). Some political differences are more important than others and can affect relationships beyond repair. [NEWLINE] [NEWLINE] [STARTQ] Is this a fair assumption? I know multiple people who I consider friends (and without whom my life would be poorer) who cannot do this. [ENDQ] [NEWLINE] Have you discussed it with them? Have you tried to show them a different perspective? [NEWLINE] [NEWLINE] [STARTQ] in what way you are actually living a better life by posting. [ENDQ] [NEWLINE] By not stressing over things like that. By not spending time deciding whether or not I'm too afraid to say what I think because someone might not like me. By staying true to myself and by drawing the attention of others to issues that are important to me, thus allowing them to know me better and also maybe change their views for the better as well. [NEWLINE] [NEWLINE] Edit: a word </s><pad>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Think of the state of the universe<mask> one possible state. There are an infinite number of states and<mask> time is not quantized there are an infinite number of decisions made at any of the infinite time states. The fatal assumption you make is that the consciousness you experience exists<mask> the same entity in every state.<mask> this were true then anything that changes you would cause you to be a different person until there is no possible distinction between you and not you. In this sense the definition of you is ambiguous and can exist in every universe<mask> you have become everything. The truth of it is that you exist for only one instant in time and after that you fail to exist. The former you contributes to the definition of the future you and your consciousness from one universe to the next (and from one time instance to the next) is completely unique. You aren't who you were<mask> you began reading this comment anymore than<mask> you began taking this breath<mask> an infinite number of time instances have passed and<mask> was you has propagated through an infinite number of universes. The part of the universe which used to be considered alive in the previous state may or may not be alive in the next. Your consciousness exists only for one instance of time and only in one universe at a time.<mask> to say that the next state will<mask> contain something similar to you is likely to be true<mask> not guaranteed and just<mask> there are possible states that contain you doesn't mean they will be selected. Consider the states<mask> you've morphed into a cranberry.<mask> it may have been possible at some point for the state to be selected which would have resulted in you later becoming a cranberry instead of a human consciousness, it was not the case. Likewise it may have been possible to find ourselves in a state<mask> many more humans exist.<mask> we aren't in that universe. From any possible definition of who you are, it does not follow that you will continue in every next universe. </s>
Label encoding: <s>Think of the state of the universe as one possible state. There are an infinite number of states and as time is not quantized there are an infinite number of decisions made at any of the infinite time states. The fatal assumption you make is that the consciousness you experience exists as the same entity in every state. If this were true then anything that changes you would cause you to be a different person until there is no possible distinction between you and not you. In this sense the definition of you is ambiguous and can exist in every universe because you have become everything. The truth of it is that you exist for only one instant in time and after that you fail to exist. The former you contributes to the definition of the future you and your consciousness from one universe to the next (and from one time instance to the next) is completely unique. You aren't who you were when you began reading this comment anymore than when you began taking this breath as an infinite number of time instances have passed and what was you has propagated through an infinite number of universes. The part of the universe which used to be considered alive in the previous state may or may not be alive in the next. Your consciousness exists only for one instance of time and only in one universe at a time. So to say that the next state will also contain something similar to you is likely to be true but not guaranteed and just because there are possible states that contain you doesn't mean they will be selected. Consider the states where you've morphed into a cranberry. While it may have been possible at some point for the state to be selected which would have resulted in you later becoming a cranberry instead of a human consciousness, it was not the case. Likewise it may have been possible to find ourselves in a state where many more humans exist. But we aren't in that universe. From any possible definition of who you are, it does not follow that you will continue in every next universe. </s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think we need to break feminism down into two categories: mainstream and radical.  Other commenters have addressed the mainstream rather well,<mask> I'll address the radicals. [NEWLINE] [NEWLINE] Radical feminists are out their.  They are vocal and disproportionately influential.  They are sexist, often homophobic and transphibic.  They will actively fight against, suppress and oppress anyone advocating for men's issues.  They idealize a society<mask> women are dominant over men or even<mask> men do not exist.  They often advocate the use of violence to achieve their goals.  Even the more mellow radicals will support blatantly sexist legislation. [NEWLINE] [NEWLINE] Now<mask><mask> we can both agree that these people are hateful and seek to harm men.  The problem is that these hateful people are femenists just<mask> much<mask> the intelligent ones you seam to run with.  This means that a part of the feminist movement seeks to actively harm men.  Is it unreasonable or 'illegitimate' for another movement to form to stop such a hateful and vicious group? [NEWLINE] [NEWLINE] Please do not construe my statements<mask> "all femenists are bad".  Merely a subset is<mask> that subset must be vehemently opposed. [NEWLINE] [NEWLINE] <mask> : [NEWLINE] [NEWLINE] [STARTQ] The men's rights movement (which I truly hope only exists on the internet) is unnecessary. Men have held positions of power for millennia - is it really necessary for us to feel better about being a man? [ENDQ] [NEWLINE] This really bothers me.  It gets to a major problem with both feminism and the men's rights movement.  They lump people together by gender<mask> opposed to acknowledging them<mask> individuals.  Should the poor homeless man be okay with being refused charity<mask> he is a man just<mask> other men are in positions of power?  I say no.  No more<mask> than a woman should be okay with earning a lower wage for her work<mask> other women are getting women only scholarships. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I think we need to break feminism down into two categories: mainstream and radical.  Other commenters have addressed the mainstream rather well, so I'll address the radicals. [NEWLINE] [NEWLINE] Radical feminists are out their.  They are vocal and disproportionately influential.  They are sexist, often homophobic and transphibic.  They will actively fight against, suppress and oppress anyone advocating for men's issues.  They idealize a society where women are dominant over men or even where men do not exist.  They often advocate the use of violence to achieve their goals.  Even the more mellow radicals will support blatantly sexist legislation. [NEWLINE] [NEWLINE] Now I think we can both agree that these people are hateful and seek to harm men.  The problem is that these hateful people are femenists just as much as the intelligent ones you seam to run with.  This means that a part of the feminist movement seeks to actively harm men.  Is it unreasonable or 'illegitimate' for another movement to form to stop such a hateful and vicious group? [NEWLINE] [NEWLINE] Please do not construe my statements as "all femenists are bad".  Merely a subset is but that subset must be vehemently opposed. [NEWLINE] [NEWLINE] Also : [NEWLINE] [NEWLINE] [STARTQ] The men's rights movement (which I truly hope only exists on the internet) is unnecessary. Men have held positions of power for millennia - is it really necessary for us to feel better about being a man? [ENDQ] [NEWLINE] This really bothers me.  It gets to a major problem with both feminism and the men's rights movement.  They lump people together by gender as opposed to acknowledging them as individuals.  Should the poor homeless man be okay with being refused charity because he is a man just because other men are in positions of power?  I say no.  No more so than a woman should be okay with earning a lower wage for her work because other women are getting women only scholarships. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're comparing these fear mongers to fictional movies,<mask> there's a key difference.  Independence Day doesn't claim to be telling a true story.  These fear mongers<mask><mask><mask><mask> are presenting themselves<mask> experts.  See the difference?  It's the same reason a sci-fi movie about aliens visiting ancient Egypt would be goofy entertainment,<mask> the History Channel's Ancient Aliens is a disgraceful, misleading, embarrassing series.  It claims to be discussing a possibly legitimate truth;<mask> it transitions from mere entertainment to actively misleading the public for profit. [NEWLINE] [NEWLINE] Everyone is happy about their beliefs.  Racists, Nazis, ISIS, you name it.  Feeling good about your beliefs doesn't mean your propaganda should be "applauded."  Attitudes have an impact on society. [NEWLINE] [NEWLINE] <mask> parents stop trusting medical science<mask> of unqualified internet stars claiming to be "experts" and spouting superstitious, baseless misinformation, there will be tragic consequences.  Outbreaks of completely preventable infectious diseases.  Kids not getting treatment they need<mask> of fear of doctors.  In some cases hippie food movements driven by fear can have similarly tragic consequences (see the "raw milk" movement, which puts people at risk of contracting very dangerous infections from unpasteurized milk). [NEWLINE] [NEWLINE] It doesn't matter<mask> it makes people happy to be told outright lies for profit, that's not the point.  The point is that it's morally wrong to spread lies that put people in danger, especially children. [NEWLINE] [NEWLINE] <mask> a parent truly wants to feel better, they should listen to people who actually spend their lives trying to develop real, testable, concrete cures to disease.  People who spend their entire working careers in labs intensively studying the disorders that cause<mask> much suffering.  Not TV talking heads who act<mask><mask> the scientific establishment is some sort of corporate monster whose goal is to poison food and children.</s>
Label encoding: <s>You're comparing these fear mongers to fictional movies, but there's a key difference.  Independence Day doesn't claim to be telling a true story.  These fear mongers on the other hand are presenting themselves as experts.  See the difference?  It's the same reason a sci-fi movie about aliens visiting ancient Egypt would be goofy entertainment, but the History Channel's Ancient Aliens is a disgraceful, misleading, embarrassing series.  It claims to be discussing a possibly legitimate truth; therefore it transitions from mere entertainment to actively misleading the public for profit. [NEWLINE] [NEWLINE] Everyone is happy about their beliefs.  Racists, Nazis, ISIS, you name it.  Feeling good about your beliefs doesn't mean your propaganda should be "applauded."  Attitudes have an impact on society. [NEWLINE] [NEWLINE] If parents stop trusting medical science because of unqualified internet stars claiming to be "experts" and spouting superstitious, baseless misinformation, there will be tragic consequences.  Outbreaks of completely preventable infectious diseases.  Kids not getting treatment they need because of fear of doctors.  In some cases hippie food movements driven by fear can have similarly tragic consequences (see the "raw milk" movement, which puts people at risk of contracting very dangerous infections from unpasteurized milk). [NEWLINE] [NEWLINE] It doesn't matter if it makes people happy to be told outright lies for profit, that's not the point.  The point is that it's morally wrong to spread lies that put people in danger, especially children. [NEWLINE] [NEWLINE] If a parent truly wants to feel better, they should listen to people who actually spend their lives trying to develop real, testable, concrete cures to disease.  People who spend their entire working careers in labs intensively studying the disorders that cause so much suffering.  Not TV talking heads who act as if the scientific establishment is some sort of corporate monster whose goal is to poison food and children.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> does the threshold for having the mental ability to consent begin at 16-18? I understand it in a legal sense, minors can't legally consent until they turn a certain age,<mask> the justification for that seems to be the same argument, and it becomes cyclical. [ENDQ] [NEWLINE] <mask> some line needs to be drawn.  16-18 is reasonable, not perfect.  Sure, it will swallow (er, no pun here) teens who are probably responsible enough to have sex and it will allow some people who are kind of emotionally immature to have sex.  By and large,<mask>, it sufficiently covers the average teenager. [NEWLINE] [NEWLINE] The alternative is to do a fact-specific inquiry regarding the maturity and capacity to consent with each case.  The likely outcomes are: [NEWLINE] [NEWLINE] 1.  More drain on police resources and court congestion. [NEWLINE] 2.  Difficulty pin-pointing<mask> constitutes 'emotionally mature' or capable of consenting, and; [NEWLINE] 3.  Whatever contours we do use to define those terms will probably align fairly well with<mask> we tend to expect of older teens, not people who are just physically capable of having sex. [NEWLINE] [NEWLINE] It<mask> puts older people on notice.  I'm 27. <mask> do I know<mask> you're actually emotionally mature, and then your parents complain, and I get prosecuted?  People think it's unfair that folks who have sex with teens who misrepresent their ages are put in a compromised position (wow, it is really hard to avoid innuendo and double entendre.)  Would an 'emotional maturity' standard alleviate or exacerbate this and similar issues?  On the flip side, under present law, I know you're a teen,<mask> I'll stay the hell away,<mask> odds are much better that's illegal in my state. [NEWLINE] [NEWLINE] EDIT:  I<mask> think /u/smartlypretty touches on the overarching policy of minimizing exploitation.</s>
Label encoding: <s> [STARTQ] Why does the threshold for having the mental ability to consent begin at 16-18? I understand it in a legal sense, minors can't legally consent until they turn a certain age, but the justification for that seems to be the same argument, and it becomes cyclical. [ENDQ] [NEWLINE] Because some line needs to be drawn.  16-18 is reasonable, not perfect.  Sure, it will swallow (er, no pun here) teens who are probably responsible enough to have sex and it will allow some people who are kind of emotionally immature to have sex.  By and large, though, it sufficiently covers the average teenager. [NEWLINE] [NEWLINE] The alternative is to do a fact-specific inquiry regarding the maturity and capacity to consent with each case.  The likely outcomes are: [NEWLINE] [NEWLINE] 1.  More drain on police resources and court congestion. [NEWLINE] 2.  Difficulty pin-pointing what constitutes 'emotionally mature' or capable of consenting, and; [NEWLINE] 3.  Whatever contours we do use to define those terms will probably align fairly well with what we tend to expect of older teens, not people who are just physically capable of having sex. [NEWLINE] [NEWLINE] It also puts older people on notice.  I'm 27.  How do I know if you're actually emotionally mature, and then your parents complain, and I get prosecuted?  People think it's unfair that folks who have sex with teens who misrepresent their ages are put in a compromised position (wow, it is really hard to avoid innuendo and double entendre.)  Would an 'emotional maturity' standard alleviate or exacerbate this and similar issues?  On the flip side, under present law, I know you're a teen, so I'll stay the hell away, because odds are much better that's illegal in my state. [NEWLINE] [NEWLINE] EDIT:  I also think /u/smartlypretty touches on the overarching policy of minimizing exploitation.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You were banned<mask> you broke the rules of the subreddit. I don't know<mask> you were expecting to happen. [NEWLINE] [NEWLINE] [STARTQ] I replied to the thread with, "I don't know<mask> this one is necessarily sexist<mask> a woman could say this about a man. [ENDQ] [NEWLINE] First rule in the sidebar: [NEWLINE] [NEWLINE] [STARTQ] 1. RULE X: SRS is a circlejack and interrupting the circlejack is an easy way to get banned. For instance, commenters are not allowed to say "This post is not offensive" or "This is not SRS worthy." [ENDQ] [NEWLINE] SRS is not a place for intelligent discussion or discourse.  There are other subs for that. <mask> you have an honest question about<mask> something is offensive/unacceptable, your best bet is to post it to /r/socialjustice101, or /r/srsdiscussion,<mask> be aware that you may<mask> be banned from there<mask> you phrase your question in a combative or offensive manner,<mask> choose your words carefully. [NEWLINE] [NEWLINE] Edit: everyone is pestering me about<mask> /r/srsdiscussion isn't a real discussion, there isn't any dissent, etc.  Basically it's way too much for someone new to feminism and social justice to handle.  I understand<mask> you could get frustrated reading that stuff and it's very easy to inadvertently offend someone and get banned.  I have to get to sleep,<mask> my best advice would be to start by **lurking** /r/openbroke (or /r/circlebroke2) for threads relating to your "interest".  It's basically SRS lite; you won't get banned immediately for having a difference opinion and you'll find less hostility.  Just follow the damned rules there, too (there's less of them).  A word of warning,<mask> ; it will turn you into a smug asshole, like myself.</s>
Label encoding: <s>You were banned because you broke the rules of the subreddit. I don't know what you were expecting to happen. [NEWLINE] [NEWLINE] [STARTQ] I replied to the thread with, "I don't know if this one is necessarily sexist because a woman could say this about a man. [ENDQ] [NEWLINE] First rule in the sidebar: [NEWLINE] [NEWLINE] [STARTQ] 1. RULE X: SRS is a circlejack and interrupting the circlejack is an easy way to get banned. For instance, commenters are not allowed to say "This post is not offensive" or "This is not SRS worthy." [ENDQ] [NEWLINE] SRS is not a place for intelligent discussion or discourse.  There are other subs for that.  If you have an honest question about why something is offensive/unacceptable, your best bet is to post it to /r/socialjustice101, or /r/srsdiscussion, but be aware that you may also be banned from there if you phrase your question in a combative or offensive manner, so choose your words carefully. [NEWLINE] [NEWLINE] Edit: everyone is pestering me about how /r/srsdiscussion isn't a real discussion, there isn't any dissent, etc.  Basically it's way too much for someone new to feminism and social justice to handle.  I understand how you could get frustrated reading that stuff and it's very easy to inadvertently offend someone and get banned.  I have to get to sleep, but my best advice would be to start by **lurking** /r/openbroke (or /r/circlebroke2) for threads relating to your "interest".  It's basically SRS lite; you won't get banned immediately for having a difference opinion and you'll find less hostility.  Just follow the damned rules there, too (there's less of them).  A word of warning, though ; it will turn you into a smug asshole, like myself.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> dont they just roll into Iraq with all their tanks and missiles and trash the place until freedom and democracy is restored? [ENDQ] [NEWLINE] Didnt that happen? Iraq is a democracy right now, they just had an election. [NEWLINE] [NEWLINE] [STARTQ] Well obviously they're fighting guerilla style enemies, and no amount of planes and tanks can rout the kind of unconventional resistance fighters the US army would be up against<mask> there was a revolt. [ENDQ] [NEWLINE] <mask> it's hard for a conventional army to defeat a guerilla enemy, it has happened before in history. [NEWLINE] [NEWLINE] [STARTQ] <mask> would the missile targets be? There are no factories or HQ buildings to target<mask> you're fighting your own populace. [ENDQ] [NEWLINE] Of course there will be. The rebels wouldnt have super fancy corporate factories<mask> there will be weapons production sites. After all,<mask> will the rebels be getting more supplies? [NEWLINE] [NEWLINE] [STARTQ] <mask> the US population is over 3 hundred million strong and they have 89 guns each. [ENDQ] [NEWLINE] Not everyone will rise up against the government. Everyone from the upper middle class up would probably support the govt. [NEWLINE] [NEWLINE] [STARTQ] There's just no way the army and the police could control a fighting force that huge. [ENDQ] [NEWLINE] It can be done. [NEWLINE] [NEWLINE] [STARTQ] Obviously not all Americans would be able to fight<mask> still hundreds of millions of enemies is just too many for any army to fight long term. [ENDQ] [NEWLINE] You dont have to kill every enemy to win wars. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the conspiracy theorists are right, and everyone listens to them, then its goodnight washington for sure. [ENDQ] [NEWLINE] We are a civilized nation, and the procedure is to have civil discourse on the direction the country is heading.<mask> we fall to the point of civil war, then we<mask> a nation have already fallen.<mask> that happens the full weight of the US government will be used against the civilians, it will not be an easy victory for either side. [NEWLINE] </s>
Label encoding: <s> [STARTQ] why dont they just roll into Iraq with all their tanks and missiles and trash the place until freedom and democracy is restored? [ENDQ] [NEWLINE] Didnt that happen? Iraq is a democracy right now, they just had an election. [NEWLINE] [NEWLINE] [STARTQ] Well obviously they're fighting guerilla style enemies, and no amount of planes and tanks can rout the kind of unconventional resistance fighters the US army would be up against if there was a revolt. [ENDQ] [NEWLINE] While it's hard for a conventional army to defeat a guerilla enemy, it has happened before in history. [NEWLINE] [NEWLINE] [STARTQ] What would the missile targets be? There are no factories or HQ buildings to target if you're fighting your own populace. [ENDQ] [NEWLINE] Of course there will be. The rebels wouldnt have super fancy corporate factories but there will be weapons production sites. After all, where will the rebels be getting more supplies? [NEWLINE] [NEWLINE] [STARTQ] But the US population is over 3 hundred million strong and they have 89 guns each. [ENDQ] [NEWLINE] Not everyone will rise up against the government. Everyone from the upper middle class up would probably support the govt. [NEWLINE] [NEWLINE] [STARTQ] There's just no way the army and the police could control a fighting force that huge. [ENDQ] [NEWLINE] It can be done. [NEWLINE] [NEWLINE] [STARTQ] Obviously not all Americans would be able to fight but still hundreds of millions of enemies is just too many for any army to fight long term. [ENDQ] [NEWLINE] You dont have to kill every enemy to win wars. [NEWLINE] [NEWLINE] [STARTQ] So if the conspiracy theorists are right, and everyone listens to them, then its goodnight washington for sure. [ENDQ] [NEWLINE] We are a civilized nation, and the procedure is to have civil discourse on the direction the country is heading. If we fall to the point of civil war, then we as a nation have already fallen. When that happens the full weight of the US government will be used against the civilians, it will not be an easy victory for either side. [NEWLINE] </s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The most important thing in taking quotes out of the Koran is making sure you have them in the right context. [Here's a good wiki]( [URL] ) dealing with responses to often quoted verses that are part of the Koran. [NEWLINE] [NEWLINE] Have you read the Koran entirely?<mask> you haven't,<mask> can you be sure that the verses you are quoting aren't part of a story by Mohammed about<mask> not to do? Or maybe they are preceded by "don't". Or maybe they are invalidated by later sections. Or maybe, or maybe, or maybe. The point is, you can't be sure<mask> the heck the verses are talking about unless you actually take them in the context they are meant to be understood in, which includes not only reading the entire Koran,<mask> taking in historical context, religious interpretation, other religious works, and<mask> many other things. [NEWLINE] [NEWLINE] I'm a Christian myself, and the bible can be a pretty scary thing<mask> one is willing to pluck random verses out. Here's a sampling: [NEWLINE] [NEWLINE] Psalms 55:15 - Let death seize upon them, [and] let them go down quick into hell: for wickedness [is] in their dwellings, [and] among them. [NEWLINE] [NEWLINE] 1 Corinthians 16:22 -<mask> any man love not the Lord Jesus Christ, let him be Anathema Maranatha. (Anathema Maranatha means destroyed<mask> god comes) [NEWLINE] [NEWLINE] "I do not permit a woman to teach or to have authority over a man. She must be quiet." (1 Timothy 2:12) [NEWLINE] [NEWLINE] Blessed shall he be who takes your little ones and dashes them against the rock! Psalms 137:9 [NEWLINE] [NEWLINE] That last one is especially a good example - taken out of context, it sounds a lot like people who kill my children are blessed. Not a very Christian verse,<mask> one doesn't apply context to it.</s>
Label encoding: <s>The most important thing in taking quotes out of the Koran is making sure you have them in the right context. [Here's a good wiki]( [URL] ) dealing with responses to often quoted verses that are part of the Koran. [NEWLINE] [NEWLINE] Have you read the Koran entirely? If you haven't, how can you be sure that the verses you are quoting aren't part of a story by Mohammed about what not to do? Or maybe they are preceded by "don't". Or maybe they are invalidated by later sections. Or maybe, or maybe, or maybe. The point is, you can't be sure what the heck the verses are talking about unless you actually take them in the context they are meant to be understood in, which includes not only reading the entire Koran, but taking in historical context, religious interpretation, other religious works, and so many other things. [NEWLINE] [NEWLINE] I'm a Christian myself, and the bible can be a pretty scary thing if one is willing to pluck random verses out. Here's a sampling: [NEWLINE] [NEWLINE] Psalms 55:15 - Let death seize upon them, [and] let them go down quick into hell: for wickedness [is] in their dwellings, [and] among them. [NEWLINE] [NEWLINE] 1 Corinthians 16:22 - If any man love not the Lord Jesus Christ, let him be Anathema Maranatha. (Anathema Maranatha means destroyed when god comes) [NEWLINE] [NEWLINE] "I do not permit a woman to teach or to have authority over a man. She must be quiet." (1 Timothy 2:12) [NEWLINE] [NEWLINE] Blessed shall he be who takes your little ones and dashes them against the rock! Psalms 137:9 [NEWLINE] [NEWLINE] That last one is especially a good example - taken out of context, it sounds a lot like people who kill my children are blessed. Not a very Christian verse, if one doesn't apply context to it.</s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is no cure for depression. There are things to treat it<mask> clinical depression stays with you for life. The OP asked about suicide. Nothing in between. I was stating<mask><mask>. [NEWLINE] [NEWLINE] I had three friends kill themselves last year. One was after 5 back surgeries. The pain never went away and left her immobilized. She is free of pain. [NEWLINE] [NEWLINE] The other lost her son and was not the same. She could not function. She was on medication for depression<mask> it didn't work for her. [NEWLINE] [NEWLINE] The last was a brother of a friend who was addicted to drugs. has been in and out of rehab. Was homeless and finally killed himself. [NEWLINE] [NEWLINE] You really have to work to get to feeling okay. It takes a lot of effort and being depressed the last thing you want to do is get out of bed. People who have never experienced it, do not understand. They say: just get out of bed. head over to /r/depression and see<mask> we all go through. It is a struggle forever. At this point they have not found a way to repair misfiring neurotransmitters much like they haven't found a cure for diabetes and<mask> they have to take insulin. They can't wish it gone like misfiring neurotransmitters can't. It is a disease of the brain. [NEWLINE] [NEWLINE] <mask> you don't have money, family, friends...<mask> do you get help? i've been waiting for 11 months for help. I've been working for 18 years putting money into my disability. Now<mask> I need help, it takes 11 months? I went from making a lot of money<mask> a web dev to being homeless. My electricity is out and i have until dec to come up with money to pay rent. I<mask> have cancer. [NEWLINE] [NEWLINE] I can't physically work. You can't call in work and say you need a sick day<mask> you are depressed.</s>
Label encoding: <s>There is no cure for depression. There are things to treat it but clinical depression stays with you for life. The OP asked about suicide. Nothing in between. I was stating my opinion. [NEWLINE] [NEWLINE] I had three friends kill themselves last year. One was after 5 back surgeries. The pain never went away and left her immobilized. She is free of pain. [NEWLINE] [NEWLINE] The other lost her son and was not the same. She could not function. She was on medication for depression but it didn't work for her. [NEWLINE] [NEWLINE] The last was a brother of a friend who was addicted to drugs. has been in and out of rehab. Was homeless and finally killed himself. [NEWLINE] [NEWLINE] You really have to work to get to feeling okay. It takes a lot of effort and being depressed the last thing you want to do is get out of bed. People who have never experienced it, do not understand. They say: just get out of bed. head over to /r/depression and see what we all go through. It is a struggle forever. At this point they have not found a way to repair misfiring neurotransmitters much like they haven't found a cure for diabetes and therefore they have to take insulin. They can't wish it gone like misfiring neurotransmitters can't. It is a disease of the brain. [NEWLINE] [NEWLINE] when you don't have money, family, friends... how do you get help? i've been waiting for 11 months for help. I've been working for 18 years putting money into my disability. Now when I need help, it takes 11 months? I went from making a lot of money as a web dev to being homeless. My electricity is out and i have until dec to come up with money to pay rent. I also have cancer. [NEWLINE] [NEWLINE] I can't physically work. You can't call in work and say you need a sick day because you are depressed.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it actually does is put a lid on property values<mask> assessed for tax purposes. You have a house that has a Tax Assessment of $250k,<mask> is listed for $500k?<mask> it sells, you get $500k (less various fees). [ENDQ] [NEWLINE] The listing for $500k will reflect the new taxes the owner will have to pay. <mask> the new owner were to keep the 250k tax assessment, they'd be willing to pay more, up to the net present value of the difference in property tax. [NEWLINE] [NEWLINE] <mask> a 250k assessment difference translates to $5000 a year in property tax, that will have a NPV in the neighborhood of $100,000. <mask> the seller loses an effective revenue stream worth $100,000 by selling. <mask> the new owner were to get the lower assessment, they'd be willing to pay $600k instead of $500k. [NEWLINE] [NEWLINE] [STARTQ] <mask> is it you claim that it screws people who want to move in or out? By subjecting them to market price tax assessment<mask> they buy a new place?<mask> would that change<mask> the people who don't move were subjected to the same market forces? [ENDQ] [NEWLINE] <mask><mask> the same total property tax would be assessed, it means that new owners pay a far higher share of total tax than they would in a system<mask> everyone pays the same percent of the house's value in property tax. [NEWLINE] [NEWLINE] To use the numbers from your example,<mask> you have 2 identical houses next door to each other, one assessed at 250k, one at 500k, the person in the 250k assessed house is only paying 1/3 their fair share, and the person in the 500k assessed house is paying 2/3 their fair share.  I would raise the tax on the 250k assessed house to cut it on the 500k assessed house,<mask> taxes should treat similarly disposed people and assets similarly.</s>
Label encoding: <s> [STARTQ] What it actually does is put a lid on property values as assessed for tax purposes. You have a house that has a Tax Assessment of $250k, but is listed for $500k? When it sells, you get $500k (less various fees). [ENDQ] [NEWLINE] The listing for $500k will reflect the new taxes the owner will have to pay.  If the new owner were to keep the 250k tax assessment, they'd be willing to pay more, up to the net present value of the difference in property tax. [NEWLINE] [NEWLINE] If a 250k assessment difference translates to $5000 a year in property tax, that will have a NPV in the neighborhood of $100,000.  So the seller loses an effective revenue stream worth $100,000 by selling.  If the new owner were to get the lower assessment, they'd be willing to pay $600k instead of $500k. [NEWLINE] [NEWLINE] [STARTQ] How is it you claim that it screws people who want to move in or out? By subjecting them to market price tax assessment when they buy a new place? How would that change if the people who don't move were subjected to the same market forces? [ENDQ] [NEWLINE] Assuming that the same total property tax would be assessed, it means that new owners pay a far higher share of total tax than they would in a system where everyone pays the same percent of the house's value in property tax. [NEWLINE] [NEWLINE] To use the numbers from your example, if you have 2 identical houses next door to each other, one assessed at 250k, one at 500k, the person in the 250k assessed house is only paying 1/3 their fair share, and the person in the 500k assessed house is paying 2/3 their fair share.  I would raise the tax on the 250k assessed house to cut it on the 500k assessed house, because taxes should treat similarly disposed people and assets similarly.</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Basically, my point is that women and men will have dramatically different outcomes *from doing the same thing*. They will come away from the experience with different narratives. [NEWLINE] [NEWLINE] Let's assume your average couple who are<mask> metabolically average decide to go on a diet. By your math, he gets 1670 calories and she gets 1108. Okay, that means he gets 560 calories more food every day to achieve the same benefit. [NEWLINE] [NEWLINE] That's a nontrivial amount. His "extra" 560 calories are *more than half* her entire daily allotment.<mask> they eat 3 meals daily, his would be a satisfying 557 calories each and hers would be a bare 409. [NEWLINE] [NEWLINE] Depending on<mask> you measure (height or weight) she's 80 or 90% his size, and<mask> is her stomach.<mask> she is not allowed to eat 80 or 90% of the food he eats. She's eating 73% of<mask> he's getting. He gets more than 4 measures of everything to her 3. She's probably going to feel deprived and hungry constantly, in a way he rarely feels. Seriously, plate some food sometime that honestly reflects these distinctions. It's not silly, it's substantial. [NEWLINE] [NEWLINE] <mask> they go out and take a brisk 3 mile walk, he burns 240 calories, she burns 194. Again, they make the same effort<mask> he gets 20% more benefit. In 5 days, it's<mask><mask> he's exercised twice<mask> much<mask> she did,<mask> he actually didn't.. [NEWLINE] [NEWLINE] Bottom line: weight loss is much easier for guys. This is the fundamental reason<mask> women often feel like diet and exercise don't really work or are too hard to stick to. Similarly, this is the fundamental reason<mask> men often feel like weightloss isn't really a big deal, and women are whiners who are probably failing<mask> they are cheating on their diet or are lazy. </s><pad>
Label encoding: <s>Basically, my point is that women and men will have dramatically different outcomes *from doing the same thing*. They will come away from the experience with different narratives. [NEWLINE] [NEWLINE] Let's assume your average couple who are also metabolically average decide to go on a diet. By your math, he gets 1670 calories and she gets 1108. Okay, that means he gets 560 calories more food every day to achieve the same benefit. [NEWLINE] [NEWLINE] That's a nontrivial amount. His "extra" 560 calories are *more than half* her entire daily allotment. If they eat 3 meals daily, his would be a satisfying 557 calories each and hers would be a bare 409. [NEWLINE] [NEWLINE] Depending on how you measure (height or weight) she's 80 or 90% his size, and so is her stomach. But she is not allowed to eat 80 or 90% of the food he eats. She's eating 73% of what he's getting. He gets more than 4 measures of everything to her 3. She's probably going to feel deprived and hungry constantly, in a way he rarely feels. Seriously, plate some food sometime that honestly reflects these distinctions. It's not silly, it's substantial. [NEWLINE] [NEWLINE] When they go out and take a brisk 3 mile walk, he burns 240 calories, she burns 194. Again, they make the same effort but he gets 20% more benefit. In 5 days, it's as if he's exercised twice as much as she did, but he actually didn't.. [NEWLINE] [NEWLINE] Bottom line: weight loss is much easier for guys. This is the fundamental reason why women often feel like diet and exercise don't really work or are too hard to stick to. Similarly, this is the fundamental reason why men often feel like weightloss isn't really a big deal, and women are whiners who are probably failing because they are cheating on their diet or are lazy. </s><pad>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not saying this isn't a legitimate thing that happens,<mask> I mean is that it is incorrect.<mask> one is allowed to interpret a story in<mask> a many he (or she) wishes,<mask> then is the point of the author writing that story in the first place? [NEWLINE] [NEWLINE] For could the reader not have came to that same conclusion through a plethora of different methods. And<mask> this is to be said, this would make stories a very dull thing for it would be constant argument<mask> each individual forces the story to fit their biases and allow them to hear<mask> they want. [NEWLINE] [NEWLINE] <mask> often<mask> an author writes a book (especially in the case of Gatsby), it is to attempt to change the viewpoint of an entity (be it collective or individual). The lesson we learn from Gatsby is one of greed<mask> imagine<mask> some Wall Street executive took the book completely out of context to say maybe use it<mask> proof that corruption is good. [NEWLINE] [NEWLINE] That would be completely eliminating the point of the book which is to allow us to see the problems that arise from greed. [NEWLINE] [NEWLINE] Believe me, I get<mask> you are saying and<mask><mask> that we must view all stories with a certain degree of subjectivity or we would all have the same boring conclusion.<mask> I am trying to point out is that<mask> we can shift the themes and play around with the lessons a little, we must do<mask> sparingly. [NEWLINE] [NEWLINE] We cannot just vandalize an author's work by shifting his intended message<mask> drastically that it says whatever we wish it to say. At its most fundamental level, there are some parts of the story (could be the lessons or symbolism) that should not be changed<mask> they hold the most important or main lesson of the story. We are allowed to guess and change<mask> the lessons might mean<mask> never<mask> they are (at least not for the obvious and intended ones). </s>
Label encoding: <s>I'm not saying this isn't a legitimate thing that happens, what I mean is that it is incorrect. If one is allowed to interpret a story in however a many he (or she) wishes, what then is the point of the author writing that story in the first place? [NEWLINE] [NEWLINE] For could the reader not have came to that same conclusion through a plethora of different methods. And if this is to be said, this would make stories a very dull thing for it would be constant argument as each individual forces the story to fit their biases and allow them to hear what they want. [NEWLINE] [NEWLINE] Yet often when an author writes a book (especially in the case of Gatsby), it is to attempt to change the viewpoint of an entity (be it collective or individual). The lesson we learn from Gatsby is one of greed but imagine if some Wall Street executive took the book completely out of context to say maybe use it as proof that corruption is good. [NEWLINE] [NEWLINE] That would be completely eliminating the point of the book which is to allow us to see the problems that arise from greed. [NEWLINE] [NEWLINE] Believe me, I get what you are saying and I agree that we must view all stories with a certain degree of subjectivity or we would all have the same boring conclusion. What I am trying to point out is that although we can shift the themes and play around with the lessons a little, we must do so sparingly. [NEWLINE] [NEWLINE] We cannot just vandalize an author's work by shifting his intended message so drastically that it says whatever we wish it to say. At its most fundamental level, there are some parts of the story (could be the lessons or symbolism) that should not be changed as they hold the most important or main lesson of the story. We are allowed to guess and change what the lessons might mean but never what they are (at least not for the obvious and intended ones). </s>
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not all loud cars/bikes are loud by choice. [NEWLINE] [NEWLINE] Some simply end up that way<mask> you change anything in the car. Sometimes some of the cars are simply broken and they can't afford to fix them. (Mufflers go bad, Cats go bad, exhaust leaks...) [NEWLINE] [NEWLINE] Some cars are loud even tho the person has tried to quiet the car down, there are reasonable limits to<mask> can be accomplished. My brother and I both have this situation going on. [NEWLINE] [NEWLINE] I had to change the engine in my car, it's a 77. The new engine doesn't really belong in my car and<mask> everything needs to be custom made. I went the proper route<mask> I even needed to pass emissions. I have a Cat installed (actually one of the most restrictive cats you can buy) and a full length exhaust and even a real muffler (not a big coffee can or something designed to be loud.) [NEWLINE] [NEWLINE] My car is still loud<mask>, and you know<mask>, it drives me nuts. I feel like crap in the morning<mask> I start it up and it roars to life. I feel like crap<mask> I hate the noise it makes and<mask> it might disturb the other on the block. [NEWLINE] [NEWLINE] I guess<mask> that makes me an inconsiderate asshole, I have no other options. [NEWLINE] [NEWLINE] <mask> I rode a motorcylce, It was loud to begin with unfortunately. After getting run off the road multiple times I was told to try a louder exhaust. I did and it drove me insane,<mask> it did seem to work. I have no real study to show this is fact. [NEWLINE] [NEWLINE] I eventually got tired of the noise and being able to be heard a mile away<mask> I went back to stock. I started getting run off the road again and decided it was a lose/lose situation and sold the bike. [NEWLINE] [NEWLINE] Hope the insight helps.</s>
Label encoding: <s>Not all loud cars/bikes are loud by choice. [NEWLINE] [NEWLINE] Some simply end up that way if you change anything in the car. Sometimes some of the cars are simply broken and they can't afford to fix them. (Mufflers go bad, Cats go bad, exhaust leaks...) [NEWLINE] [NEWLINE] Some cars are loud even tho the person has tried to quiet the car down, there are reasonable limits to what can be accomplished. My brother and I both have this situation going on. [NEWLINE] [NEWLINE] I had to change the engine in my car, it's a 77. The new engine doesn't really belong in my car and so everything needs to be custom made. I went the proper route because I even needed to pass emissions. I have a Cat installed (actually one of the most restrictive cats you can buy) and a full length exhaust and even a real muffler (not a big coffee can or something designed to be loud.) [NEWLINE] [NEWLINE] My car is still loud however, and you know what, it drives me nuts. I feel like crap in the morning when I start it up and it roars to life. I feel like crap because I hate the noise it makes and how it might disturb the other on the block. [NEWLINE] [NEWLINE] I guess if that makes me an inconsiderate asshole, I have no other options. [NEWLINE] [NEWLINE] When I rode a motorcylce, It was loud to begin with unfortunately. After getting run off the road multiple times I was told to try a louder exhaust. I did and it drove me insane, but it did seem to work. I have no real study to show this is fact. [NEWLINE] [NEWLINE] I eventually got tired of the noise and being able to be heard a mile away so I went back to stock. I started getting run off the road again and decided it was a lose/lose situation and sold the bike. [NEWLINE] [NEWLINE] Hope the insight helps.</s>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Killing anything for the fun of it is drastically different than killing for a purpose. [ENDQ] [NEWLINE] Is having fun not a purpose? <mask><mask> I said "killing for pleasure"?  Surely we primarily eat meat for pleasure rather than for utility,<mask> it would be a far more efficient use of land, fertilizer, water, money, etc to eat only vegetables. [NEWLINE] [NEWLINE] [STARTQ] That said, my viewpoint is that man is a predator. There's an established food chain and to consider man immoral for killing animals to eat you'd<mask> have to consider all predators immoral and<mask><mask> the very balance of nature to be evil. [ENDQ] [NEWLINE] Many animals do not have a choice<mask> to eat meat.  In those cases it is not immoral for them to kill for food.  In the cases of animals that are perfectly capable of subsisting without killing, I would say that they are behaving in an immoral way,<mask> I don't blame them<mask> they probably lack the reasoning capacity to think about this situation from the other creature's point of view. [NEWLINE] [NEWLINE] [STARTQ] I don't think I'd ever try to convince anyone that their particular diet is somehow inferior to my own,<mask><mask> you eat<mask> you want and<mask> you think eating meat is inherently immoral then more power to you. [ENDQ] [NEWLINE] The reason I am doing the CMV is<mask> I want to be convinced that eating meat is morally acceptable,<mask><mask> you aren't going to try to convince me of that,<mask> are you posting? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask> the basis of your argument is flawed simply<mask> you fail to distinguish intent on the part of the person killing the animal. [ENDQ] [NEWLINE] Whether you are killing<mask> you like to kill things (e.g. hunting and not eating it) or you are killing<mask> you prefer the taste of meat, the intent is the same: you get pleasure from that choice<mask> you are doing it.</s>
Label encoding: <s> [STARTQ] Killing anything for the fun of it is drastically different than killing for a purpose. [ENDQ] [NEWLINE] Is having fun not a purpose?  What if I said "killing for pleasure"?  Surely we primarily eat meat for pleasure rather than for utility, since it would be a far more efficient use of land, fertilizer, water, money, etc to eat only vegetables. [NEWLINE] [NEWLINE] [STARTQ] That said, my viewpoint is that man is a predator. There's an established food chain and to consider man immoral for killing animals to eat you'd also have to consider all predators immoral and in fact the very balance of nature to be evil. [ENDQ] [NEWLINE] Many animals do not have a choice but to eat meat.  In those cases it is not immoral for them to kill for food.  In the cases of animals that are perfectly capable of subsisting without killing, I would say that they are behaving in an immoral way, but I don't blame them since they probably lack the reasoning capacity to think about this situation from the other creature's point of view. [NEWLINE] [NEWLINE] [STARTQ] I don't think I'd ever try to convince anyone that their particular diet is somehow inferior to my own, I think you eat what you want and if you think eating meat is inherently immoral then more power to you. [ENDQ] [NEWLINE] The reason I am doing the CMV is because I want to be convinced that eating meat is morally acceptable, so if you aren't going to try to convince me of that, why are you posting? [NEWLINE] [NEWLINE] [STARTQ] But I think the basis of your argument is flawed simply because you fail to distinguish intent on the part of the person killing the animal. [ENDQ] [NEWLINE] Whether you are killing because you like to kill things (e.g. hunting and not eating it) or you are killing because you prefer the taste of meat, the intent is the same: you get pleasure from that choice so you are doing it.</s>
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know much about body-related dysphoria that's not trans-related,<mask> I don't have an opinion on it beyond "everyone deserves to be happy about their body". [NEWLINE] [NEWLINE] I get the impression that you think of trans people<mask> "men who want to be/think that they're women" (and vice-versa)<mask> we tend to see ourselves<mask> "women everyone thinks of<mask> men" (and vice versa).<mask> you try looking at us the way we see ourselves, it might help to make sense of things. [NEWLINE] [NEWLINE] <mask> I say "the approaches you've thought of", I mean everything that's ever been aimed at turning trans people cis- talk therapy, psychiatric medication, hyperdosing with the hormones the patient's body is already producing, electroshock, psychosurgery, aversion therapy, CBT... They've all failed miserably (especially for the patients). Left untreated (or treated with the methods I've mentioned), trans people suffer from very high rates of things like anxiety, depression, addictions, discrimination, violence, un- or underemployment and homelessness. The attempted suicide rate is 41%. [NEWLINE] [NEWLINE] Transitioning,<mask><mask><mask><mask>, consistently produces healthy, functioning people with rates of the above problems consistent with the general population and a reported satisfaction rate in the mid-80% to high 90% range (depending on the metric used). Most dissatisfaction comes from being treated poorly by others- bullied, disowned, rejected, fired, evicted, assaulted, etc. [NEWLINE] [NEWLINE] (Am on mobile, can source links for this later). [NEWLINE] [NEWLINE] <mask> yeah, maybe we *could* try other methods to treat trans people,<mask><mask> nobody's found anything<mask> transitioning that has any positive effect on the condition, I don't see that there's much point<mask> we could be improving access and techniques instead. </s>
Label encoding: <s>I don't know much about body-related dysphoria that's not trans-related, so I don't have an opinion on it beyond "everyone deserves to be happy about their body". [NEWLINE] [NEWLINE] I get the impression that you think of trans people as "men who want to be/think that they're women" (and vice-versa) when we tend to see ourselves as "women everyone thinks of as men" (and vice versa). If you try looking at us the way we see ourselves, it might help to make sense of things. [NEWLINE] [NEWLINE] When I say "the approaches you've thought of", I mean everything that's ever been aimed at turning trans people cis- talk therapy, psychiatric medication, hyperdosing with the hormones the patient's body is already producing, electroshock, psychosurgery, aversion therapy, CBT... They've all failed miserably (especially for the patients). Left untreated (or treated with the methods I've mentioned), trans people suffer from very high rates of things like anxiety, depression, addictions, discrimination, violence, un- or underemployment and homelessness. The attempted suicide rate is 41%. [NEWLINE] [NEWLINE] Transitioning, on the other hand, consistently produces healthy, functioning people with rates of the above problems consistent with the general population and a reported satisfaction rate in the mid-80% to high 90% range (depending on the metric used). Most dissatisfaction comes from being treated poorly by others- bullied, disowned, rejected, fired, evicted, assaulted, etc. [NEWLINE] [NEWLINE] (Am on mobile, can source links for this later). [NEWLINE] [NEWLINE] So yeah, maybe we *could* try other methods to treat trans people, but since nobody's found anything besides transitioning that has any positive effect on the condition, I don't see that there's much point when we could be improving access and techniques instead. </s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The old tipping debate. I generally side with those that want to end the mostly American practice of tipping about 20% percent of the cost of the meal. That may never happen. [NEWLINE] [NEWLINE] <mask> bothers me, is the idea that the server's time is<mask> only<mask> valuable<mask> the cost of<mask>'s on the plate.<mask> I get a $8 plate of eggs at Waffle House I am to tip about 2 dollars.<mask> I am to get a $50 steak at the local trendy restaurant or steak house, I am to tip over $10? [NEWLINE] [NEWLINE] I understand that *generally* at a more expensive restaurant I will spend more time there making the servers work a longer time on my needs. And they are usually expected to be more knowledgeable about the menu. And the bus has to make thing look better and the bartender has to know more cocktails.... [NEWLINE] [NEWLINE] <mask> it seems to me, the tipping tradition has propped up a non-egalitarian system<mask> the young and attractive can rake in a decent pay check<mask> expensive restaurants can be exclusive about hiring, and others, especially the older who probably did not really want to be stuck server their whole career, are left taking the 15% on my cheap dinner. [NEWLINE] [NEWLINE] <mask> I go in simply for a beer and a dessert, I try to tip above 50%. For "affordable" dinners I try to tip 25-30%. For nicer dinners I usually end up tip 15-20%. Truly bad service and I will only tip 10-15%. And<mask> they have a truly helpful suggestion at that trendy restaurant I may tip over 20%..... [NEWLINE] [NEWLINE] <mask><mask> we should move to a system based more on<mask> long you are in the restaurant, and the amount of attention your visit actually needed, rather than a simple percentage completely isolated from the nuance of the service. [NEWLINE] [NEWLINE] <mask> the whole system just makes me feel guilty.</s>
Label encoding: <s>The old tipping debate. I generally side with those that want to end the mostly American practice of tipping about 20% percent of the cost of the meal. That may never happen. [NEWLINE] [NEWLINE] What bothers me, is the idea that the server's time is as only as valuable as the cost of what's on the plate. If I get a $8 plate of eggs at Waffle House I am to tip about 2 dollars. If I am to get a $50 steak at the local trendy restaurant or steak house, I am to tip over $10? [NEWLINE] [NEWLINE] I understand that *generally* at a more expensive restaurant I will spend more time there making the servers work a longer time on my needs. And they are usually expected to be more knowledgeable about the menu. And the bus has to make thing look better and the bartender has to know more cocktails.... [NEWLINE] [NEWLINE] But it seems to me, the tipping tradition has propped up a non-egalitarian system where the young and attractive can rake in a decent pay check because expensive restaurants can be exclusive about hiring, and others, especially the older who probably did not really want to be stuck server their whole career, are left taking the 15% on my cheap dinner. [NEWLINE] [NEWLINE] If I go in simply for a beer and a dessert, I try to tip above 50%. For "affordable" dinners I try to tip 25-30%. For nicer dinners I usually end up tip 15-20%. Truly bad service and I will only tip 10-15%. And if they have a truly helpful suggestion at that trendy restaurant I may tip over 20%..... [NEWLINE] [NEWLINE] I think we should move to a system based more on how long you are in the restaurant, and the amount of attention your visit actually needed, rather than a simple percentage completely isolated from the nuance of the service. [NEWLINE] [NEWLINE] But the whole system just makes me feel guilty.</s>
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[spring break]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;oq=spring+break&amp;aqs=chrome.0.69i59j0l5.1366j0j9&amp;sourceid=chrome&amp;es_sm=122&amp;ie=UTF-8) [NEWLINE] [NEWLINE] [carnival]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=carnival+rio&amp;revid=255166102) [NEWLINE] [NEWLINE] [mardi gras]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=mardi+gras) [NEWLINE] [NEWLINE] It has nothing to do with 'thinking differently'. It has to do with these 'holidays' frequently involve men and women wearing scantily clad outfits and costumes for each other.<mask> you have actually been around any of these parties, you'd know.</s><pad>
Label encoding: <s>[spring break]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;oq=spring+break&amp;aqs=chrome.0.69i59j0l5.1366j0j9&amp;sourceid=chrome&amp;es_sm=122&amp;ie=UTF-8) [NEWLINE] [NEWLINE] [carnival]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=carnival+rio&amp;revid=255166102) [NEWLINE] [NEWLINE] [mardi gras]( [URL] +break&amp;rlz=1CAASUD_enUS614US614&amp;es_sm=122&amp;source=lnms&amp;tbm=isch&amp;sa=X&amp;ei=wu7jVJ9c08axBNqLgNgG&amp;ved=0CAkQ_AUoAg&amp;biw=1366&amp;bih=657#tbm=isch&amp;q=mardi+gras) [NEWLINE] [NEWLINE] It has nothing to do with 'thinking differently'. It has to do with these 'holidays' frequently involve men and women wearing scantily clad outfits and costumes for each other. If you have actually been around any of these parties, you'd know.</s><pad>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The difference is whether rules are a starting point, or a result. [NEWLINE] [NEWLINE] Prescriptivism is looked down upon in Linguistic circles<mask> it is the linguistic equivalent of Begging the Question,<mask> the goal,<mask> Linguists see it, is to *find* the rules that govern language. [NEWLINE] [NEWLINE] To draw an analogy from physics, consider<mask> happened<mask> the first few dozen scientists produced data that contradicted the Newtonian Model of physics. [NEWLINE] [NEWLINE] A Prescriptivist would look at data (reliable, reproducible data, collected from the real world), and declare that they are, by definition, wrong<mask> they don't fit within the Newtonian model, and<mask> the Data should align itself with the Model. [NEWLINE] [NEWLINE] The Descriptivist looks at the data, then turns around to say that their *model* must be wrong, and that we need to find a new model,<mask> the data indicates that it isn't quite right, and we need to create a model that fits the (real world, reliable, reproducible) data. <mask> we get Relativistic Physics. [NEWLINE] [NEWLINE] The proper ~~deterministic~~ descriptivistic response to "Shops went the he to" would be to say "Wait a second, that flies in the face of everything we know.  Is the source reliable?  Can the results be duplicated?"  Without sufficient data, it'll likely be rejected<mask> "random noise," <mask> without agreement of acceptability from a community of cognitively normal, native speakers, it's not describing *language*<mask> describing a *sentence.* [NEWLINE] [NEWLINE] Heck, the very techniques to determine the rules rely on include things like "In this dialect, construction X is valid,<mask> construction Y isn't," <mask> without such a test, you cannot determine whether you have correctly described the rules of that dialect.</s>
Label encoding: <s>The difference is whether rules are a starting point, or a result. [NEWLINE] [NEWLINE] Prescriptivism is looked down upon in Linguistic circles because it is the linguistic equivalent of Begging the Question, because the goal, as Linguists see it, is to *find* the rules that govern language. [NEWLINE] [NEWLINE] To draw an analogy from physics, consider what happened when the first few dozen scientists produced data that contradicted the Newtonian Model of physics. [NEWLINE] [NEWLINE] A Prescriptivist would look at data (reliable, reproducible data, collected from the real world), and declare that they are, by definition, wrong because they don't fit within the Newtonian model, and therefore the Data should align itself with the Model. [NEWLINE] [NEWLINE] The Descriptivist looks at the data, then turns around to say that their *model* must be wrong, and that we need to find a new model, because the data indicates that it isn't quite right, and we need to create a model that fits the (real world, reliable, reproducible) data.  Thus we get Relativistic Physics. [NEWLINE] [NEWLINE] The proper ~~deterministic~~ descriptivistic response to "Shops went the he to" would be to say "Wait a second, that flies in the face of everything we know.  Is the source reliable?  Can the results be duplicated?"  Without sufficient data, it'll likely be rejected as "random noise,"  because without agreement of acceptability from a community of cognitively normal, native speakers, it's not describing *language* but describing a *sentence.* [NEWLINE] [NEWLINE] Heck, the very techniques to determine the rules rely on include things like "In this dialect, construction X is valid, but construction Y isn't,"  because without such a test, you cannot determine whether you have correctly described the rules of that dialect.</s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you misunderstand the concept of science. The term'scientific authority' is an oxymoron. Science is not about authority, it's about experimentation and deduction. It's about empirical evidence, open review, critical analysis, and reproducible results. [NEWLINE] [NEWLINE] Of course scientific expertise is valuable. Athletic ability is valuable too, it might help an athlete run a race more quickly and more adeptly than anyone else.<mask> I,<mask> I am a bit overweight and out of shape, can still make it down the track and cross the finish line, even<mask> I'm the last to get there, and panting and wheezing<mask> I arrive.<mask> I had no arms or legs, I might have to have help to make it, I might need someone to wheel me to the finish line,<mask> I could still potentially get there. Sure, I could just let the professional athlete do all the finish-line crossing and just watch from the bleachers.<mask> maybe I want to see<mask> it looks like from the vantage point of the finish line myself, and not just take his word for it. [NEWLINE] [NEWLINE] Scientific expertise gives a person an advantage in their field, not a monopoly on it. They might be the best of the best,<mask> they still make errors, and they still have limits. And more than that, they have the potential to be subjective, to be biased, and in some circumstances to be corruptible. Real science is about asking questions and not being satisfied with the answers you are given until it has been proven to you. Taking an expert's opinion in lieu of evidence, and letting them think for you instead of thinking for yourself, is all too reminiscent of the dogmatic system that scientific method was created to displace. [NEWLINE] [NEWLINE] Authority is *not* science. [NEWLINE] [NEWLINE] Oh, and nano-thermite [is a thing]( [URL] ).</s>
Label encoding: <s>I think you misunderstand the concept of science. The term'scientific authority' is an oxymoron. Science is not about authority, it's about experimentation and deduction. It's about empirical evidence, open review, critical analysis, and reproducible results. [NEWLINE] [NEWLINE] Of course scientific expertise is valuable. Athletic ability is valuable too, it might help an athlete run a race more quickly and more adeptly than anyone else. But I, though I am a bit overweight and out of shape, can still make it down the track and cross the finish line, even if I'm the last to get there, and panting and wheezing as I arrive. If I had no arms or legs, I might have to have help to make it, I might need someone to wheel me to the finish line, but I could still potentially get there. Sure, I could just let the professional athlete do all the finish-line crossing and just watch from the bleachers. But maybe I want to see what it looks like from the vantage point of the finish line myself, and not just take his word for it. [NEWLINE] [NEWLINE] Scientific expertise gives a person an advantage in their field, not a monopoly on it. They might be the best of the best, but they still make errors, and they still have limits. And more than that, they have the potential to be subjective, to be biased, and in some circumstances to be corruptible. Real science is about asking questions and not being satisfied with the answers you are given until it has been proven to you. Taking an expert's opinion in lieu of evidence, and letting them think for you instead of thinking for yourself, is all too reminiscent of the dogmatic system that scientific method was created to displace. [NEWLINE] [NEWLINE] Authority is *not* science. [NEWLINE] [NEWLINE] Oh, and nano-thermite [is a thing]( [URL] ).</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Saying that Vegetarianism/Veganism is "illogical" isn't exactly fair. Most people who are V/V, myself included, would freely admit that it is a personal or cultural decision they've made. I may not be able to "prove" that my lifestyle is fine,<mask> that's not really the point. [NEWLINE] [NEWLINE] <mask> for some points [NEWLINE] [NEWLINE] *The average lifespan of vegetarians may not be longer than that of meat-eaters,<mask> that probably says more about the people who are vegetarians. I believe the majority of vegetarians in the world are in India,<mask> lifespans are shorter than more developed parts of the world, which would potentially skew your average.<mask> sure, you can have a potentially healthy non-veg diet, and a potentially unhealthy veg diet. [NEWLINE] [NEWLINE] *Well,<mask> you want to use that argument, shouldn't we be eating some sort of Paleo diet? Berries, wild game, fruits, and no grains or domestic animals? [NEWLINE] [NEWLINE] *Hypothetical - Are there animals that you would refuse to eat on grounds of cruelty or ethics? Lots of people would refrain from eating horses, dogs, monkeys, etc. [NEWLINE] [NEWLINE] *We are capable of being omnivores,<mask> you can live on a plant-based diet. [NEWLINE] [NEWLINE] *Lots of people are V/V (or have reduced their meat consumption) for environmental reasons. You need far more energy to produce a pound of meat than<mask> you need to produce a pound of most plant-based food. The demand for fish has caused fish stocks to decline rapidly, and significantly disrupted many ecosystems. Examples available on request. [NEWLINE] [NEWLINE] *I'm not sure<mask> you mean here. Are you saying that I should eat a moderate amount of meat? Or that there's no harm in eating a moderate amount of any substance?</s>
Label encoding: <s>Saying that Vegetarianism/Veganism is "illogical" isn't exactly fair. Most people who are V/V, myself included, would freely admit that it is a personal or cultural decision they've made. I may not be able to "prove" that my lifestyle is fine, but that's not really the point. [NEWLINE] [NEWLINE] As for some points [NEWLINE] [NEWLINE] *The average lifespan of vegetarians may not be longer than that of meat-eaters, but that probably says more about the people who are vegetarians. I believe the majority of vegetarians in the world are in India, where lifespans are shorter than more developed parts of the world, which would potentially skew your average. But sure, you can have a potentially healthy non-veg diet, and a potentially unhealthy veg diet. [NEWLINE] [NEWLINE] *Well, if you want to use that argument, shouldn't we be eating some sort of Paleo diet? Berries, wild game, fruits, and no grains or domestic animals? [NEWLINE] [NEWLINE] *Hypothetical - Are there animals that you would refuse to eat on grounds of cruelty or ethics? Lots of people would refrain from eating horses, dogs, monkeys, etc. [NEWLINE] [NEWLINE] *We are capable of being omnivores, but you can live on a plant-based diet. [NEWLINE] [NEWLINE] *Lots of people are V/V (or have reduced their meat consumption) for environmental reasons. You need far more energy to produce a pound of meat than what you need to produce a pound of most plant-based food. The demand for fish has caused fish stocks to decline rapidly, and significantly disrupted many ecosystems. Examples available on request. [NEWLINE] [NEWLINE] *I'm not sure what you mean here. Are you saying that I should eat a moderate amount of meat? Or that there's no harm in eating a moderate amount of any substance?</s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have always firmly held the belief that marriage is something that celebrates heterosexual partnership [ENDQ] [NEWLINE] <mask><mask><mask> I'm curious<mask> to<mask> you feel marriage exclusively celebrates heterosexual partnerships,<mask> you can pinpoint<mask> you feel this way it might help in changing your view. [NEWLINE] [NEWLINE] [STARTQ] I would be extremely uncomfortable seeing men/women dancing/kissing people of the same sex at the wedding. [ENDQ] [NEWLINE] This is a personal feeling and something that I am unlikely to be able to change.<mask>, I would suggest that you consider<mask> hurt you would feel<mask> you went to a wedding and people felt it unpleasant to see you dance with your fiancée - it would feel awful. The easiest thing to do, honestly, is simply not to pay attention to them -<mask> I very much doubt you pay large attention to the dancing or kisses of straight couples around you at weddings. [NEWLINE] [NEWLINE] [STARTQ] one of these people is single, and like to try and seduce other women. [ENDQ] [NEWLINE] The simplest solution I can see to this is to speak to this person<mask> not make it in any way related to her orientation. Simply make it clear that you are hoping for a calm and tasteful wedding and that *nobody*<mask><mask> gender or orientation will be welcome<mask> they cannot respect the other guests. [NEWLINE] [NEWLINE] Something that might be worth thinking about is:<mask> are you getting married? Most people get married<mask> they want to have their relationship recognised by their loved ones. Now imagine<mask> you were invited to a wedding<mask> your fiancée was not - imagine<mask> disrespectful that would be. That is<mask> you are doing to these people - you are telling them that<mask> of something out of their control (their gender of their partner) their relationship is not'real' to you. You are saying that they do not count<mask> a real couple, that you do not respect them or their partner.</s><pad>
Label encoding: <s> [STARTQ] I have always firmly held the belief that marriage is something that celebrates heterosexual partnership [ENDQ] [NEWLINE] First of all I'm curious as to why you feel marriage exclusively celebrates heterosexual partnerships, if you can pinpoint why you feel this way it might help in changing your view. [NEWLINE] [NEWLINE] [STARTQ] I would be extremely uncomfortable seeing men/women dancing/kissing people of the same sex at the wedding. [ENDQ] [NEWLINE] This is a personal feeling and something that I am unlikely to be able to change. However, I would suggest that you consider how hurt you would feel if you went to a wedding and people felt it unpleasant to see you dance with your fiancée - it would feel awful. The easiest thing to do, honestly, is simply not to pay attention to them - since I very much doubt you pay large attention to the dancing or kisses of straight couples around you at weddings. [NEWLINE] [NEWLINE] [STARTQ] one of these people is single, and like to try and seduce other women. [ENDQ] [NEWLINE] The simplest solution I can see to this is to speak to this person but not make it in any way related to her orientation. Simply make it clear that you are hoping for a calm and tasteful wedding and that *nobody* regardless of gender or orientation will be welcome if they cannot respect the other guests. [NEWLINE] [NEWLINE] Something that might be worth thinking about is: why are you getting married? Most people get married because they want to have their relationship recognised by their loved ones. Now imagine if you were invited to a wedding but your fiancée was not - imagine how disrespectful that would be. That is what you are doing to these people - you are telling them that because of something out of their control (their gender of their partner) their relationship is not'real' to you. You are saying that they do not count as a real couple, that you do not respect them or their partner.</s><pad>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"murders committed in the name of faith" are rarely just about faith these days,<mask> usually have more to do with underlying issues of cultural imperialism, land and power, or historical political struggles. I of course would not contend that no one has ever killed or done wrong purely in the name of religion,<mask> people have<mask> done atrocious things in the name of democracy, and countless other institutions and forces that may or may not be positive. The point being, "religion = evil" is a simplistic view that simply is not supported by the historical record; religion has been at least partially responsible for great works of art, advancements in the natural sciences, gracious works of charity,<mask> well<mask> wars and oppressions. [NEWLINE] [NEWLINE] And, to go back to the "non-murderers club" example, the reason that example doesn't work is<mask> the vast majority of the population are not murderers. In lots of places, especially in the U.S., the vast majority of the population is religious. In these places atheism actually implies a different sort of lifestyle from the norm, and in a lot (<mask> not all) cases, a different set of values and beliefs. I understand that the term "atheism" in itself does not imply embracing certain beliefs,<mask> usually there are associated beliefs, such<mask> an emphasis in the power of science and reason, and a distrust of institutional religion. [NEWLINE] [NEWLINE] A better example than a "non-murderers club" would be a non-gun-carriers club in an area populated by gun-users. The analogy still isn't perfect,<mask><mask><mask> a club of this sort might seem weird depending on<mask> you live, people in the deep south of the U.S. for example, might embrace the non-gun-carrying lifestyle, which might be a departure from social norms.</s>
Label encoding: <s>"murders committed in the name of faith" are rarely just about faith these days, but usually have more to do with underlying issues of cultural imperialism, land and power, or historical political struggles. I of course would not contend that no one has ever killed or done wrong purely in the name of religion, but people have also done atrocious things in the name of democracy, and countless other institutions and forces that may or may not be positive. The point being, "religion = evil" is a simplistic view that simply is not supported by the historical record; religion has been at least partially responsible for great works of art, advancements in the natural sciences, gracious works of charity, as well as wars and oppressions. [NEWLINE] [NEWLINE] And, to go back to the "non-murderers club" example, the reason that example doesn't work is because the vast majority of the population are not murderers. In lots of places, especially in the U.S., the vast majority of the population is religious. In these places atheism actually implies a different sort of lifestyle from the norm, and in a lot ( but not all) cases, a different set of values and beliefs. I understand that the term "atheism" in itself does not imply embracing certain beliefs, but usually there are associated beliefs, such as an emphasis in the power of science and reason, and a distrust of institutional religion. [NEWLINE] [NEWLINE] A better example than a "non-murderers club" would be a non-gun-carriers club in an area populated by gun-users. The analogy still isn't perfect, but even though a club of this sort might seem weird depending on where you live, people in the deep south of the U.S. for example, might embrace the non-gun-carrying lifestyle, which might be a departure from social norms.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"That human life wouldn't be there<mask> you did not have sex. You at least some kind of responsability of everything you do, and<mask><mask> you did provoked the existence of the most important thing there is - a human life - you have responsibility over it." [NEWLINE] [NEWLINE] Okay,<mask><mask> can't the man just "have responsibility over it" by suggesting an abortion?<mask> the woman doesn't want to abort, that's fine,<mask> she shouldn't be able to force him to help her raise the child<mask> he doesn't want it. People do need to take responsibility for their actions,<mask><mask> there's an easy way to do it and there's a hard way to do it,<mask> force two people to do it the hard way<mask> one of them wants to opt for the easy way? [NEWLINE] [NEWLINE] Here: [NEWLINE] [NEWLINE] <mask> I'm a dude out driving my girlfriend's car for fun and I get a nail in a tire, I'll say, "Oh well--that's a possible consequence of driving. Better take responsibility for it and take care of that before I ruin the wheel, too."<mask> my girlfriend says, "No! I own the car and I want to leave the tire like it is!", then I've done my part by interjecting common sense; she can deal with the bent rims on her own<mask> the tire goes flat and she keeps driving on it. [NEWLINE] [NEWLINE] I'm too lazy to think of a better metaphor. I hope that one got my point across. [NEWLINE] [NEWLINE] EDIT: I would<mask><mask><mask> "a human life" is not necessarily "the most important thing there is",<mask> that doesn't really seem relevant to OP's CMV request. And<mask> I'm still lazy. [NEWLINE] [NEWLINE] EDIT #2: Obviously, I'm talking about situations in which the money and opportunity for abortion are available.</s>
Label encoding: <s>"That human life wouldn't be there if you did not have sex. You at least some kind of responsability of everything you do, and when what you did provoked the existence of the most important thing there is - a human life - you have responsibility over it." [NEWLINE] [NEWLINE] Okay, but why can't the man just "have responsibility over it" by suggesting an abortion? If the woman doesn't want to abort, that's fine, but she shouldn't be able to force him to help her raise the child if he doesn't want it. People do need to take responsibility for their actions, but when there's an easy way to do it and there's a hard way to do it, why force two people to do it the hard way when one of them wants to opt for the easy way? [NEWLINE] [NEWLINE] Here: [NEWLINE] [NEWLINE] If I'm a dude out driving my girlfriend's car for fun and I get a nail in a tire, I'll say, "Oh well--that's a possible consequence of driving. Better take responsibility for it and take care of that before I ruin the wheel, too." If my girlfriend says, "No! I own the car and I want to leave the tire like it is!", then I've done my part by interjecting common sense; she can deal with the bent rims on her own when the tire goes flat and she keeps driving on it. [NEWLINE] [NEWLINE] I'm too lazy to think of a better metaphor. I hope that one got my point across. [NEWLINE] [NEWLINE] EDIT: I would also argue that "a human life" is not necessarily "the most important thing there is", but that doesn't really seem relevant to OP's CMV request. And also I'm still lazy. [NEWLINE] [NEWLINE] EDIT #2: Obviously, I'm talking about situations in which the money and opportunity for abortion are available.</s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's go by your points one by one: [NEWLINE] [NEWLINE] 1. It's easy to say this<mask> casually<mask> your entire way of life and your actual life itself is not<mask>'s at risk. A war with North Korea could trigger some very hard times for both North and South. Any economic benefits to be reaped would likely be in the long term. It's easy to call for sacrifice<mask> you are not the one sacrificing. Who is to say unification will save lives? Violent, forced unification might cause extreme civil unrest, decades of terrorism, and continued or even accelerated loss of life on both sides. Unifying through full on war and unifying post-North Korean collapse are very different prospects that entail different likelihoods of success. [NEWLINE] [NEWLINE] 2. That is true. Many experts predict that unification could be an economic boon. The same Goldman Sachs report that you are unknowingly referencing<mask> states that this is under the assumption of most optimal circumstances (i.e. gradual and peaceful reunification). That circumstance is only most likely post natural internal North Korean collapse and not through violent takeover. North Korea is not ready to be a stable state<mask> taken over. [NEWLINE] [NEWLINE] 3. 60 years of consisted foreign policy doesn't change overnight. Their greatest concerns aren't cultural ones and more military. China,<mask> being a large coastal country, is landlocked militarily speaking by two great island chains - both of which are, in one way or another, under US control. The Korean peninsula is part of these chains (they form the mainland connection to the islands). North Korea falling further exacerbates America's cornering of China. It's a militarily strategic concern more than anything. [NEWLINE] [NEWLINE] Again,<mask><mask> it's important to emphasize this. It's easy to be hawkish<mask> it's not your life that's being thrown away. Some things to keep in mind.</s>
Label encoding: <s>Let's go by your points one by one: [NEWLINE] [NEWLINE] 1. It's easy to say this so casually when your entire way of life and your actual life itself is not what's at risk. A war with North Korea could trigger some very hard times for both North and South. Any economic benefits to be reaped would likely be in the long term. It's easy to call for sacrifice when you are not the one sacrificing. Who is to say unification will save lives? Violent, forced unification might cause extreme civil unrest, decades of terrorism, and continued or even accelerated loss of life on both sides. Unifying through full on war and unifying post-North Korean collapse are very different prospects that entail different likelihoods of success. [NEWLINE] [NEWLINE] 2. That is true. Many experts predict that unification could be an economic boon. The same Goldman Sachs report that you are unknowingly referencing also states that this is under the assumption of most optimal circumstances (i.e. gradual and peaceful reunification). That circumstance is only most likely post natural internal North Korean collapse and not through violent takeover. North Korea is not ready to be a stable state if taken over. [NEWLINE] [NEWLINE] 3. 60 years of consisted foreign policy doesn't change overnight. Their greatest concerns aren't cultural ones and more military. China, despite being a large coastal country, is landlocked militarily speaking by two great island chains - both of which are, in one way or another, under US control. The Korean peninsula is part of these chains (they form the mainland connection to the islands). North Korea falling further exacerbates America's cornering of China. It's a militarily strategic concern more than anything. [NEWLINE] [NEWLINE] Again, I think it's important to emphasize this. It's easy to be hawkish when it's not your life that's being thrown away. Some things to keep in mind.</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A fetus is a being which,<mask> technically human, cannot<mask> exercise most of the rights a human can. It cannot have freedom of speech, nor travel, nor religion, nor economic freedom, nor... you get the picture. There are, I suppose, really two broad forms of freedom: [NEWLINE] [NEWLINE] 1. Freedoms of doing [NEWLINE] [NEWLINE] 2. Freedoms of being [NEWLINE] [NEWLINE] Most freedoms of doing are by its very nature unavailable to the fetus and<mask> irrelevant. Any politician who claims to support the free speech, for instance, of a fetus, is transparently a charlatan. He (it's usually a he) seeks to take freedoms from women and give them to the fetus, with himself<mask> the controlling proxy. [NEWLINE] [NEWLINE] Freedoms of being are more to the point.<mask> freedoms can a fetus have in this regard? The freedom not to be destroyed, to receive nutrients, to exist in a peaceful and stress-free uterine environment, to develop into a healthy human being.<mask><mask> do you give a fetus more freedoms of being without taking them away from the woman carrying it? [NEWLINE] [NEWLINE] For taking freedom from women is the real aim of the "protectors" who claim to speak for the fetus. [NEWLINE] [NEWLINE] The only ways of maximizing freedoms for both women and the unborn is to offer women all the prenatal support they need: sufficient nutrition, safe environment (that includes a healthy planet, doesn't it?), and personal freedoms<mask> that<mask> they choose to bring forth a child, it will be one with a loving family, economic justice, and a free society in which to take part, and not a rapacious free-for-all basic human rights are only for the highest bidder. And women must still have reproductive choice: taking that away doesn't really help children. It merely makes for a worse world for those who are born.</s>
Label encoding: <s>A fetus is a being which, while technically human, cannot yet exercise most of the rights a human can. It cannot have freedom of speech, nor travel, nor religion, nor economic freedom, nor... you get the picture. There are, I suppose, really two broad forms of freedom: [NEWLINE] [NEWLINE] 1. Freedoms of doing [NEWLINE] [NEWLINE] 2. Freedoms of being [NEWLINE] [NEWLINE] Most freedoms of doing are by its very nature unavailable to the fetus and thus irrelevant. Any politician who claims to support the free speech, for instance, of a fetus, is transparently a charlatan. He (it's usually a he) seeks to take freedoms from women and give them to the fetus, with himself as the controlling proxy. [NEWLINE] [NEWLINE] Freedoms of being are more to the point. What freedoms can a fetus have in this regard? The freedom not to be destroyed, to receive nutrients, to exist in a peaceful and stress-free uterine environment, to develop into a healthy human being. But how do you give a fetus more freedoms of being without taking them away from the woman carrying it? [NEWLINE] [NEWLINE] For taking freedom from women is the real aim of the "protectors" who claim to speak for the fetus. [NEWLINE] [NEWLINE] The only ways of maximizing freedoms for both women and the unborn is to offer women all the prenatal support they need: sufficient nutrition, safe environment (that includes a healthy planet, doesn't it?), and personal freedoms so that if they choose to bring forth a child, it will be one with a loving family, economic justice, and a free society in which to take part, and not a rapacious free-for-all basic human rights are only for the highest bidder. And women must still have reproductive choice: taking that away doesn't really help children. It merely makes for a worse world for those who are born.</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have been reading pickup artist material on Reddit for a long time, and it has drastically improved my life. I have had better sex and deeper connections; I am a more confident and honest person; I have much more respect and much less resentment of women. I have<mask> made many women very happy and saved a lot of women a whole lot of trouble (by not getting too invested in them<mask> they don't really want me.) [NEWLINE] This morning, a former lover (who I am still very close friends with) sent me an article about<mask> a "seduction guide" was removed from kickstarter,<mask> it contained "misogynistic material" that promoted "violence towards women."<mask><mask> kickstarted and jezebel overreacted,<mask> I would like to discuss. [NEWLINE] In kickstarter's apology to the world, they cite this page from the reddit. Out of context, perhaps this might be taken to be negative--partic. the part<mask> it says, "Don't ask for permission. Be dominant. Force her to rebuff your advances." Out of context, this is def. bad advice.<mask>, the people who subscribe to the seduction subreddit are "nice guys," learning<mask> to express themselves,<mask> it's more complicated than just bad or good. [NEWLINE] The fact of the matter is that sexual expression is complicated and confusing for young men like me, who have been taught deference and complete respect for their whole lives. The classic advice, "just be honest; just be yourself" is not sufficient, and I have found pickup artist material helpful in showing me &lt;i [STARTQ] <mask> &lt;/i&gt; to be my honest self. [ENDQ] Given the context of the reddit post and the way it has helped people like me, do you still think it is overly aggressive/promoting violence against women?</s>
Label encoding: <s>I have been reading pickup artist material on Reddit for a long time, and it has drastically improved my life. I have had better sex and deeper connections; I am a more confident and honest person; I have much more respect and much less resentment of women. I have also made many women very happy and saved a lot of women a whole lot of trouble (by not getting too invested in them when they don't really want me.) [NEWLINE] This morning, a former lover (who I am still very close friends with) sent me an article about how a "seduction guide" was removed from kickstarter, because it contained "misogynistic material" that promoted "violence towards women." I think kickstarted and jezebel overreacted, but I would like to discuss. [NEWLINE] In kickstarter's apology to the world, they cite this page from the reddit. Out of context, perhaps this might be taken to be negative--partic. the part where it says, "Don't ask for permission. Be dominant. Force her to rebuff your advances." Out of context, this is def. bad advice. However, the people who subscribe to the seduction subreddit are "nice guys," learning how to express themselves, so it's more complicated than just bad or good. [NEWLINE] The fact of the matter is that sexual expression is complicated and confusing for young men like me, who have been taught deference and complete respect for their whole lives. The classic advice, "just be honest; just be yourself" is not sufficient, and I have found pickup artist material helpful in showing me &lt;i [STARTQ] how &lt;/i&gt; to be my honest self. [ENDQ] Given the context of the reddit post and the way it has helped people like me, do you still think it is overly aggressive/promoting violence against women?</s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I meant a "great job" at living morally [ENDQ] [NEWLINE] This means nothing<mask> this is subjective.  "Living morally" is not something you can objectively prove.  To a Christian, I'm living immorally<mask> I'm an atheist.  To a Muslim, both Christians and atheists are immoral.  You get the idea. [NEWLINE] [NEWLINE] [STARTQ] Anything "made in China" or India or Pakistan was likely made in a sweatshop, especially<mask> it's clothing.<mask> not "slave" labor, then conditions which are tantamount to slavery. [ENDQ] [NEWLINE] This is not true. <mask> the image of China<mask> the sweatshop land for capitalism is out of date now, for example (<mask> you need some further info, here's a good article: [URL] /) [NEWLINE] [NEWLINE] [STARTQ] China is no longer just the “workshop of the world” and a super-exploitable export platform for foreign capital — it is already one of the world’s most important consumer markets across a range of sectors including automobiles, smartphones, luxury goods, and fast food. The rising importance of the Chinese consumer market makes IP protection and investor arbitration a top priority for big global companies.<mask> foreign investors have long<mask> discovered that the illiberal and nationalist Chinese state, with its capricious legal system, is an unreliable protector of their interests in China. [ENDQ] [NEWLINE] <mask>, this isn't a matter of "morality"<mask> who's to say this "slavery" is "immoral"?  Arguing in terms of morality is pointless. [NEWLINE] [NEWLINE] [STARTQ] And<mask> we're just going to throw out the idea of morality in general then<mask> even bother expecting anybody to do the right thing, ever? [ENDQ] [NEWLINE] I do think throwing that idea out is a good idea,<mask> who's to say<mask>'s the "right thing" morally?  </s>
Label encoding: <s> [STARTQ] I meant a "great job" at living morally [ENDQ] [NEWLINE] This means nothing because this is subjective.  "Living morally" is not something you can objectively prove.  To a Christian, I'm living immorally because I'm an atheist.  To a Muslim, both Christians and atheists are immoral.  You get the idea. [NEWLINE] [NEWLINE] [STARTQ] Anything "made in China" or India or Pakistan was likely made in a sweatshop, especially if it's clothing. If not "slave" labor, then conditions which are tantamount to slavery. [ENDQ] [NEWLINE] This is not true.  Indeed the image of China as the sweatshop land for capitalism is out of date now, for example ( if you need some further info, here's a good article: [URL] /) [NEWLINE] [NEWLINE] [STARTQ] China is no longer just the “workshop of the world” and a super-exploitable export platform for foreign capital — it is already one of the world’s most important consumer markets across a range of sectors including automobiles, smartphones, luxury goods, and fast food. The rising importance of the Chinese consumer market makes IP protection and investor arbitration a top priority for big global companies. But foreign investors have long since discovered that the illiberal and nationalist Chinese state, with its capricious legal system, is an unreliable protector of their interests in China. [ENDQ] [NEWLINE] Besides, this isn't a matter of "morality" because who's to say this "slavery" is "immoral"?  Arguing in terms of morality is pointless. [NEWLINE] [NEWLINE] [STARTQ] And if we're just going to throw out the idea of morality in general then why even bother expecting anybody to do the right thing, ever? [ENDQ] [NEWLINE] I do think throwing that idea out is a good idea, because who's to say what's the "right thing" morally?  </s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It seems to me that the heart of human motivation consists in trying to gain good emotions and avoid bad emotions. [NEWLINE] [NEWLINE] For example, I go to the gym partly<mask> I want the good feeling of working out, partly<mask> I want the good feeling of being more attractive, and partly<mask> I want the good feelings of (potentially) finding a partner who finds me attractive. [NEWLINE] [NEWLINE] Similarly, I do a job to gain money that I can spend on things that I enjoy (good feelings) and paying bills (avoiding the bad feelings of being sued and going to jail). [NEWLINE] [NEWLINE] Apart from<mask> one is in the grip of a strong mood like anger (which I don't want to discuss<mask> it does not involve *EDIT: considered* actions), it seems to me that this is always the case? [NEWLINE] [NEWLINE] Can anyone change my view by coming up with examples of times<mask> people do things, and they don't expect to get a good feeling out of it or avoid a bad feeling? [NEWLINE] [NEWLINE] Thanks! [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>It seems to me that the heart of human motivation consists in trying to gain good emotions and avoid bad emotions. [NEWLINE] [NEWLINE] For example, I go to the gym partly because I want the good feeling of working out, partly because I want the good feeling of being more attractive, and partly because I want the good feelings of (potentially) finding a partner who finds me attractive. [NEWLINE] [NEWLINE] Similarly, I do a job to gain money that I can spend on things that I enjoy (good feelings) and paying bills (avoiding the bad feelings of being sued and going to jail). [NEWLINE] [NEWLINE] Apart from when one is in the grip of a strong mood like anger (which I don't want to discuss because it does not involve *EDIT: considered* actions), it seems to me that this is always the case? [NEWLINE] [NEWLINE] Can anyone change my view by coming up with examples of times when people do things, and they don't expect to get a good feeling out of it or avoid a bad feeling? [NEWLINE] [NEWLINE] Thanks! [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're opening another can of worms about the high school sports,<mask> yes...  <mask><mask><mask><mask> it is right at all to subject kids (those who can't make informed decisions<mask> they're minor) to such pressure to play and putting themselves at the risk of injury.   Again, different can of worms<mask> I'm even more adamantly opposed to football in high school. [NEWLINE] [NEWLINE] Going back up a bit, the other sports athletes are free to play<mask> sports they want to and the same thing exist, no one should be making profit off of their risk.   They're free to do<mask> they please and put themselves at risk,<mask> its shitty of people to make a profit off that.   Of course,<mask> stated originally, I'm a free market kinda guy and<mask> someone can get a university to pay them to coach some kids, I don't fault them for that.  Get paid<mask> you can get paid, fine.  That's<mask> the problem, to me, comes down to the fans that are willing to increase the demand<mask> much. [NEWLINE] [NEWLINE] [NEWLINE] Not impossible to go pro right out of high school,<mask> its my understanding that it is very unlikley.   Correct me<mask> I'm wrong about that, I don't follow it that close. [NEWLINE] [NEWLINE] <mask> they're receiving a scholarship<mask> are a star player, they're not getting paid their value.   Walk-ons, sure<mask><mask> with that.   Actually, all of them are free to play or not, they're hoping to get rich.  I still think fans should stop supporting their stupidity in working and being under-paid<mask> their value is. [NEWLINE] [NEWLINE] I don't support paying any of the athletes.  I support paying those athletes through minor leagues and removing the entire system outside of education.  </s>
Label encoding: <s>You're opening another can of worms about the high school sports, but yes...   I do not think it is right at all to subject kids (those who can't make informed decisions because they're minor) to such pressure to play and putting themselves at the risk of injury.   Again, different can of worms but I'm even more adamantly opposed to football in high school. [NEWLINE] [NEWLINE] Going back up a bit, the other sports athletes are free to play what sports they want to and the same thing exist, no one should be making profit off of their risk.   They're free to do as they please and put themselves at risk, but its shitty of people to make a profit off that.   Of course, as stated originally, I'm a free market kinda guy and if someone can get a university to pay them to coach some kids, I don't fault them for that.  Get paid how you can get paid, fine.  That's why the problem, to me, comes down to the fans that are willing to increase the demand so much. [NEWLINE] [NEWLINE] [NEWLINE] Not impossible to go pro right out of high school, but its my understanding that it is very unlikley.   Correct me if I'm wrong about that, I don't follow it that close. [NEWLINE] [NEWLINE] If they're receiving a scholarship but are a star player, they're not getting paid their value.   Walk-ons, sure I agree with that.   Actually, all of them are free to play or not, they're hoping to get rich.  I still think fans should stop supporting their stupidity in working and being under-paid what their value is. [NEWLINE] [NEWLINE] I don't support paying any of the athletes.  I support paying those athletes through minor leagues and removing the entire system outside of education.  </s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think your definition of overrated is correct.  Let me give you an example of a case<mask> it's easy to determine whether something is underrated or overrated. [NEWLINE] [NEWLINE] At the beginning of each sporting event season, various publications rank each team from first to last based on<mask> well they think they'll do.  At the end of the year, you can look at the final standings and see which teams were overrated and which were underrated.  A team that was predicted to come in 3rd<mask> actually came in 10th was overrated. In other words, overrating is only relevant based on future performance.  You<mask><mask> whatever ranking is given now is incorrect based on<mask> results will say in the future. Do you agree with this definition, at least for sports team rankings? [NEWLINE] [NEWLINE] Now, things get muddier<mask> you don't have a clear ranking system.  There's no final results for which movies or bands are best, instead<mask> we have is a bunch of rating systems that change over time. <mask> in these cases, I would take overrated to mean, "this band is ranked highly in x system now,<mask> in 20 years, it will be ranked much lower."  X can mean popular opinion polls, or rolling stones top lists or whatever you choose.  Saying the Beatles are overrated means that<mask> they consistently appear at the top of "best bands" lists in various media, in the future they will drop in popularity relative to other bands.  You are making a prediction about the future. [NEWLINE] [NEWLINE] <mask> I'm sure people do use overrated to mean, "Everyone thinks this is good<mask> I don't" that's not a good use for this word.  The word has "rate" in it which means that it only makes sense in the context of a rating scheme.</s>
Label encoding: <s>I don't think your definition of overrated is correct.  Let me give you an example of a case where it's easy to determine whether something is underrated or overrated. [NEWLINE] [NEWLINE] At the beginning of each sporting event season, various publications rank each team from first to last based on how well they think they'll do.  At the end of the year, you can look at the final standings and see which teams were overrated and which were underrated.  A team that was predicted to come in 3rd but actually came in 10th was overrated. In other words, overrating is only relevant based on future performance.  You argue that whatever ranking is given now is incorrect based on what results will say in the future. Do you agree with this definition, at least for sports team rankings? [NEWLINE] [NEWLINE] Now, things get muddier when you don't have a clear ranking system.  There's no final results for which movies or bands are best, instead what we have is a bunch of rating systems that change over time.  So in these cases, I would take overrated to mean, "this band is ranked highly in x system now, but in 20 years, it will be ranked much lower."  X can mean popular opinion polls, or rolling stones top lists or whatever you choose.  Saying the Beatles are overrated means that while they consistently appear at the top of "best bands" lists in various media, in the future they will drop in popularity relative to other bands.  You are making a prediction about the future. [NEWLINE] [NEWLINE] While I'm sure people do use overrated to mean, "Everyone thinks this is good but I don't" that's not a good use for this word.  The word has "rate" in it which means that it only makes sense in the context of a rating scheme.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's like the solipsism of ethics; it's useless. [ENDQ] [NEWLINE] It's not useless.  It's a starting point for discussion.  You either<mask><mask> morality is universal and spawn that discussion, or you agree that morality is not universal and proceed from there. [NEWLINE] [NEWLINE] [STARTQ] One of us could be right,<mask> the other wrong. We could both be wrong. We could both be partially right. We could both be right or wrong depending on context. [ENDQ] [NEWLINE] On the subject of morality<mask><mask> the notion of right and wrong is tenuous at best.  One can certainly be "right" and "wrong" in a factual sense,<mask> I would not blindly accept the extension of right and wrong to morality -- you'd probably have to make a *much* more compelling argument for that. [NEWLINE] [NEWLINE] <mask>, you are correct -- that doesn't necessarily mean we should just declare that morality is not universal and leave it at that.  Consensus is a tool for specifically avoiding inaction or blind disagreement. [NEWLINE] [NEWLINE] Rather than approaching the problem from trying to garner agreement on broad principles, e.g. morality revolving around human suffering, I prefer approaching from the opposite side, that of starting at nothing and building up to the broader items.  After all, words such<mask> "unreasonable and unnecessary" add even more ambiguity.  Who determines<mask> is "unreasonable and unnecessary"? [NEWLINE] [NEWLINE] [STARTQ] The two gladiators,<mask> consenting adults, are partially responsible for others who are acting outside the scope of our adherence to the concept of consent<mask> a moral virtue. [ENDQ] [NEWLINE] <mask><mask>.  People are ultimately responsible for their own actions.  You may be influenced, or ordered, or you may even be coerced,<mask> are you not responsible for your own actions?  Nuremberg defense and all that.  </s><pad>
Label encoding: <s> [STARTQ] It's like the solipsism of ethics; it's useless. [ENDQ] [NEWLINE] It's not useless.  It's a starting point for discussion.  You either argue that morality is universal and spawn that discussion, or you agree that morality is not universal and proceed from there. [NEWLINE] [NEWLINE] [STARTQ] One of us could be right, while the other wrong. We could both be wrong. We could both be partially right. We could both be right or wrong depending on context. [ENDQ] [NEWLINE] On the subject of morality I think the notion of right and wrong is tenuous at best.  One can certainly be "right" and "wrong" in a factual sense, but I would not blindly accept the extension of right and wrong to morality -- you'd probably have to make a *much* more compelling argument for that. [NEWLINE] [NEWLINE] However, you are correct -- that doesn't necessarily mean we should just declare that morality is not universal and leave it at that.  Consensus is a tool for specifically avoiding inaction or blind disagreement. [NEWLINE] [NEWLINE] Rather than approaching the problem from trying to garner agreement on broad principles, e.g. morality revolving around human suffering, I prefer approaching from the opposite side, that of starting at nothing and building up to the broader items.  After all, words such as "unreasonable and unnecessary" add even more ambiguity.  Who determines what is "unreasonable and unnecessary"? [NEWLINE] [NEWLINE] [STARTQ] The two gladiators, though consenting adults, are partially responsible for others who are acting outside the scope of our adherence to the concept of consent as a moral virtue. [ENDQ] [NEWLINE] I disagree.  People are ultimately responsible for their own actions.  You may be influenced, or ordered, or you may even be coerced, but are you not responsible for your own actions?  Nuremberg defense and all that.  </s><pad>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agreed given the current state of things,<mask> I'd argue all these points fall into "examples of<mask> we haven't got our shit figured out". [NEWLINE] [NEWLINE] <mask><mask> that once data is "out there" it's out of your direct control,<mask> it's not a black and white picture. There are different levels of "out there", and the user can absolutely have some control over that aspect.  Data shared to your close circle of friends it much less likely to end up being crawled by a bot than data shared publicly. [NEWLINE] [NEWLINE] Advertising revenue<mask> a driver for social services is the cause of a lot of the problems we're dealing with. It encourages centralised storage, a public-by-default approach to sharing, a static approach to temporal content. [NEWLINE] [NEWLINE] It seems to me that given enough time social is going to move from *service* to *layer*. [NEWLINE] [NEWLINE] A decentralized social layer not funded by advertising revenue (I don't propose a model here; the internet has a lot of shit to figure out regarding<mask> we fund things we deem valuable) and<mask> built to the priority of user experience could be fundamentally different from the services we use today. It would be decentralized, for one. With no need for public-by-default sharing for collecting advertising data and increasing ad impressions, it could encourage a private-by-default model to encourage users to be more selective with their sharing, and<mask> give them the ability to manage<mask> their content is available<mask> time moves on: e.g. new friends don't get to see content over a year old. [NEWLINE] [NEWLINE] <mask><mask> the popularity of apps like snapchat point to the demand for content to be more ephemeral, and that even the generation brought up always-connected aren't necessarily happy with the idea of static content<mask> a model for fleeting moments.</s>
Label encoding: <s>Agreed given the current state of things, though I'd argue all these points fall into "examples of how we haven't got our shit figured out". [NEWLINE] [NEWLINE] I agree that once data is "out there" it's out of your direct control, but it's not a black and white picture. There are different levels of "out there", and the user can absolutely have some control over that aspect.  Data shared to your close circle of friends it much less likely to end up being crawled by a bot than data shared publicly. [NEWLINE] [NEWLINE] Advertising revenue as a driver for social services is the cause of a lot of the problems we're dealing with. It encourages centralised storage, a public-by-default approach to sharing, a static approach to temporal content. [NEWLINE] [NEWLINE] It seems to me that given enough time social is going to move from *service* to *layer*. [NEWLINE] [NEWLINE] A decentralized social layer not funded by advertising revenue (I don't propose a model here; the internet has a lot of shit to figure out regarding how we fund things we deem valuable) and accordingly built to the priority of user experience could be fundamentally different from the services we use today. It would be decentralized, for one. With no need for public-by-default sharing for collecting advertising data and increasing ad impressions, it could encourage a private-by-default model to encourage users to be more selective with their sharing, and therefore give them the ability to manage how their content is available as time moves on: e.g. new friends don't get to see content over a year old. [NEWLINE] [NEWLINE] I think the popularity of apps like snapchat point to the demand for content to be more ephemeral, and that even the generation brought up always-connected aren't necessarily happy with the idea of static content as a model for fleeting moments.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In a secular society, such individuals are often dismissed<mask> cranks<mask> they are trying to prove/validate their own entire world view, (and usually in isolation from the works of others), rather than a particular/specific mechanism/conjecture. Religion has<mask> produced many great (and terrible) generalists and eclectic "Renaissance Men" -<mask> the laws of God/Truth were believed to operate "<mask> above<mask> below". [ENDQ] [NEWLINE] You had me until this point<mask><mask>. Secular society does not view science in a negative rational. [NEWLINE] [NEWLINE] [STARTQ] This contrasts with the incrementalist approach of modern scientists who generally don't dare to venture far from their narrow field of expertise. [ENDQ] [NEWLINE] With this we have<mask> seen a great expansion in our knowledge recently, with technology rapidly improving and more fundamental understandings by the minute. [NEWLINE] [NEWLINE] [STARTQ] Perhaps Newtown and the rest subconsciously cherry-picked/retained only the values of religion that motivated them to their great leaps. In any case, those values (belief in the existence of truth, certainty, purpose, meaning, right and wrong, good and evil) aren't offered by a modern secular society with the same degree of intensity and fervour. [ENDQ] [NEWLINE] They did,<mask> you have not watched this video please do, Neil Degrasse Tyson simplfies this better than I can without writing multiple pages of words. [NEWLINE] [NEWLINE] [STARTQ] causing enough motive in an individual to have enough arrogance or self-confidence that he can understand the mystery of the universe before he dies? [ENDQ] [NEWLINE] This made me think for a<mask>,<mask><mask><mask> I came up with a rational answer. Knowing the actually mystery behind the universe and saying you know the mystery behind the universe are two seperate things. I would rather die in ignorance than know I died living a lie. </s>
Label encoding: <s> [STARTQ] In a secular society, such individuals are often dismissed as cranks because they are trying to prove/validate their own entire world view, (and usually in isolation from the works of others), rather than a particular/specific mechanism/conjecture. Religion has thus produced many great (and terrible) generalists and eclectic "Renaissance Men" - because the laws of God/Truth were believed to operate " as above so below". [ENDQ] [NEWLINE] You had me until this point I think. Secular society does not view science in a negative rational. [NEWLINE] [NEWLINE] [STARTQ] This contrasts with the incrementalist approach of modern scientists who generally don't dare to venture far from their narrow field of expertise. [ENDQ] [NEWLINE] With this we have also seen a great expansion in our knowledge recently, with technology rapidly improving and more fundamental understandings by the minute. [NEWLINE] [NEWLINE] [STARTQ] Perhaps Newtown and the rest subconsciously cherry-picked/retained only the values of religion that motivated them to their great leaps. In any case, those values (belief in the existence of truth, certainty, purpose, meaning, right and wrong, good and evil) aren't offered by a modern secular society with the same degree of intensity and fervour. [ENDQ] [NEWLINE] They did, If you have not watched this video please do, Neil Degrasse Tyson simplfies this better than I can without writing multiple pages of words. [NEWLINE] [NEWLINE] [STARTQ] causing enough motive in an individual to have enough arrogance or self-confidence that he can understand the mystery of the universe before he dies? [ENDQ] [NEWLINE] This made me think for a while, but I think I came up with a rational answer. Knowing the actually mystery behind the universe and saying you know the mystery behind the universe are two seperate things. I would rather die in ignorance than know I died living a lie. </s>
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes,<mask> it *supposedly* already is. That's my point,<mask> the mindset of "I shouldn't have to pay for someone else's BIRTH CONTROL" is<mask> rampant, people tend to forget that it serves other purposes. [NEWLINE] [NEWLINE] This apparently even includes insurance companies. I've been *told* it should be covered,<mask> in reality it is not covered<mask> much<mask> any of my other medications<mask> of the overwhelming majority of people who feel it's any of their business<mask> I require it. [NEWLINE] [NEWLINE] Let me explain a bit further.. [NEWLINE] [NEWLINE] I have to use Seasonique/Amethia, which allows me to only have one period every three months.<mask> I have a period every month, I am in such excruciating pain *even<mask> I'm not menstruating* that I can barely go to work. [NEWLINE] [NEWLINE] Seasonique, without insurance, costs around $300 for a three month pack. (Which is outrageous,<mask> it is no different than any other medication.) There are other medications that are *monthly* packs that cost less than a third of<mask> Seasonique costs, and would work just fine for<mask> I need.<mask>, you are *not allowed* to pick up the medication any earlier than every 30 days, forcing you to take the placebos, which forces you to have a period every month. [NEWLINE] [NEWLINE] I attempted for several months to get them to allow me to pick it up a week in advance, and they refused. I spoke with my doctor, and he couldn't do anything about it. I've been told that I should be able to pick it up early<mask> my doctor says I can,<mask> nope. No such luck. [NEWLINE] [NEWLINE] Now I'm forced to spend $90 every three months instead of $5 or $0 every month. </s>
Label encoding: <s>Yes, but it *supposedly* already is. That's my point, when the mindset of "I shouldn't have to pay for someone else's BIRTH CONTROL" is so rampant, people tend to forget that it serves other purposes. [NEWLINE] [NEWLINE] This apparently even includes insurance companies. I've been *told* it should be covered, but in reality it is not covered as much as any of my other medications because of the overwhelming majority of people who feel it's any of their business why I require it. [NEWLINE] [NEWLINE] Let me explain a bit further.. [NEWLINE] [NEWLINE] I have to use Seasonique/Amethia, which allows me to only have one period every three months. If I have a period every month, I am in such excruciating pain *even when I'm not menstruating* that I can barely go to work. [NEWLINE] [NEWLINE] Seasonique, without insurance, costs around $300 for a three month pack. (Which is outrageous, as it is no different than any other medication.) There are other medications that are *monthly* packs that cost less than a third of what Seasonique costs, and would work just fine for what I need. However, you are *not allowed* to pick up the medication any earlier than every 30 days, forcing you to take the placebos, which forces you to have a period every month. [NEWLINE] [NEWLINE] I attempted for several months to get them to allow me to pick it up a week in advance, and they refused. I spoke with my doctor, and he couldn't do anything about it. I've been told that I should be able to pick it up early if my doctor says I can, but nope. No such luck. [NEWLINE] [NEWLINE] Now I'm forced to spend $90 every three months instead of $5 or $0 every month. </s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For a second let's treat comic books<mask><mask> they're the same thing<mask> books or movies. [NEWLINE] [NEWLINE] <mask> would you feel<mask> in book 5 of Harry Potter, JK Rowling decided she wanted to make Harry into a young black boy from the Bronx named Omar Holloway? [NEWLINE] [NEWLINE] I mean it's all fiction right?<mask> should it matter that she made a change like that<mask><mask><mask> the rest of the story remains the same? Well, with a book it's obvious,<mask> it's inconsistency in storytelling for<mask> will be seen<mask><mask> something done purely for the purposes of pandering to a different demographic. [NEWLINE] [NEWLINE] The people that are complaining about spiderman aren't racist and those complaining about Thor aren't sexist. They're the same people that complain<mask> batman is retconned<mask> it's the joker, the riddler, Ra's Al Ghul, and whoever the big villain in the new arc is that shot his parents. The same people that complain<mask> a character dies during an arc which takes place at the same time<mask> another<mask> he's alive. In practice it doesn't rattle the very foundation of the character,<mask> it's an unnecessary change that cheapens the artistic integrity of an artistic medium which is already viewed<mask> childish. [NEWLINE] [NEWLINE] Spiderman is a decades long saga that some readers have been following their entire lives.<mask> some people are able to embrace the clear distinction between comic book storytelling and just about every other type of storytelling and welcome the change, some are resistant<mask> they don't like seeing the artistic integrity of the story they love<mask> much changed to invite in newer readers that won't have<mask> strong a connection. [NEWLINE] [NEWLINE] <mask> an adult and occasional comic book reader I don't have an issue with it,<mask> can sympathize with those who do for the above reasons.</s>
Label encoding: <s>For a second let's treat comic books as if they're the same thing as books or movies. [NEWLINE] [NEWLINE] How would you feel if in book 5 of Harry Potter, JK Rowling decided she wanted to make Harry into a young black boy from the Bronx named Omar Holloway? [NEWLINE] [NEWLINE] I mean it's all fiction right? Why should it matter that she made a change like that as long as the rest of the story remains the same? Well, with a book it's obvious, because it's inconsistency in storytelling for what will be seen as as something done purely for the purposes of pandering to a different demographic. [NEWLINE] [NEWLINE] The people that are complaining about spiderman aren't racist and those complaining about Thor aren't sexist. They're the same people that complain when batman is retconned so it's the joker, the riddler, Ra's Al Ghul, and whoever the big villain in the new arc is that shot his parents. The same people that complain when a character dies during an arc which takes place at the same time as another where he's alive. In practice it doesn't rattle the very foundation of the character, but it's an unnecessary change that cheapens the artistic integrity of an artistic medium which is already viewed as childish. [NEWLINE] [NEWLINE] Spiderman is a decades long saga that some readers have been following their entire lives. While some people are able to embrace the clear distinction between comic book storytelling and just about every other type of storytelling and welcome the change, some are resistant because they don't like seeing the artistic integrity of the story they love so much changed to invite in newer readers that won't have as strong a connection. [NEWLINE] [NEWLINE] As an adult and occasional comic book reader I don't have an issue with it, but can sympathize with those who do for the above reasons.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0778, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't think that they should just be able to abort it. The usual exceptions I am ok with- rape, incest, [ENDQ] [NEWLINE] I don't understand<mask> you can say this and claim to be arguing it's<mask> of preserving the life of the fetus.<mask> does the conditions under which the fetus came to be matter to whether or not it should be protected. Your argument is basically treating pregnancy<mask> a punishment for choosing to have sex. Contraception can fail, pregnancy can occur even<mask> you take all the precautions to prevent it. Hell I know someone who was told by their doctor they were infertile, and then they got pregnant by their spouse! [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask> sex is fun and feels great, it's<mask> a responsibility. Women need to take charge of their own bodies and choices, and own up to the consequences<mask> they don't [ENDQ] [NEWLINE] And paying for an abortion **is taking charge of your body and your choices and owning up to the consequences**. I don't understand<mask> having an abortion is somehow *shirking responsibility*. That makes no sense to me. [NEWLINE] [NEWLINE] [STARTQ] <mask> no sane person would seriously suggest a "fourth trimester abortion" just<mask> it's inconvenient to raise a child. [ENDQ] [NEWLINE] Unlike a bundle of cells, or a fetus before 24 weeks, your newborn *has brain activity*, *is sentient*, has emotions, etc. Most people consider this to be a defining factor of personhood and<mask> an embryo, a fetus, etc. are not persons and<mask> are significantly different from your newborn. [NEWLINE] [NEWLINE] Ultimately,<mask><mask> pro-life people need to own up to their own arguments. It has nothing to do with the life or rights of the fetus and everything to do with controlling women having sex. Otherwise the arguments would be quite different.</s>
Label encoding: <s> [STARTQ] I don't think that they should just be able to abort it. The usual exceptions I am ok with- rape, incest, [ENDQ] [NEWLINE] I don't understand how you can say this and claim to be arguing it's because of preserving the life of the fetus. Why does the conditions under which the fetus came to be matter to whether or not it should be protected. Your argument is basically treating pregnancy as a punishment for choosing to have sex. Contraception can fail, pregnancy can occur even if you take all the precautions to prevent it. Hell I know someone who was told by their doctor they were infertile, and then they got pregnant by their spouse! [NEWLINE] [NEWLINE] [STARTQ] However, while sex is fun and feels great, it's also a responsibility. Women need to take charge of their own bodies and choices, and own up to the consequences if they don't [ENDQ] [NEWLINE] And paying for an abortion **is taking charge of your body and your choices and owning up to the consequences**. I don't understand how having an abortion is somehow *shirking responsibility*. That makes no sense to me. [NEWLINE] [NEWLINE] [STARTQ] Yet no sane person would seriously suggest a "fourth trimester abortion" just because it's inconvenient to raise a child. [ENDQ] [NEWLINE] Unlike a bundle of cells, or a fetus before 24 weeks, your newborn *has brain activity*, *is sentient*, has emotions, etc. Most people consider this to be a defining factor of personhood and thus an embryo, a fetus, etc. are not persons and thus are significantly different from your newborn. [NEWLINE] [NEWLINE] Ultimately, I think pro-life people need to own up to their own arguments. It has nothing to do with the life or rights of the fetus and everything to do with controlling women having sex. Otherwise the arguments would be quite different.</s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I hear most often is that they consider that the people from the United Stated are "arrogant". It's my belief that it's not really the arrogance that bothers them,<mask> the fact that whenever America puts its mind to something, it completely dominates. [ENDQ] [NEWLINE] You are sadly mistaken. The rest of the world isn't reeling in pain from the 92 dream team.  People are angry at the US for reasons such<mask> : [NEWLINE] [NEWLINE] 1) Thinking they are the best at things which they aren't the best at. (e.g. Healthcare) [NEWLINE] [NEWLINE] 2) Having a distorted view of their own history (e.g. the view that America single handedly won WWII). [NEWLINE] [NEWLINE] 3) Being largely ignorant/isolationist of<mask> goes on beyond their borders. [NEWLINE] [NEWLINE] 4) Their foreign policy. [NEWLINE] [NEWLINE] Just watch this: [URL] [NEWLINE] [NEWLINE] [STARTQ] They have a competition<mask> all of the teams are (more or less) on equal footing, and the United States sends a team of scrubs who are basically there to get waled on. [ENDQ] [NEWLINE] The US made it to the world cup. That's more than 177 teams. It's frankly insulting to those player to call them scrubs to get waled on. [NEWLINE] [NEWLINE] [STARTQ] I don't necessarily want any to try to change my view about whether or not the U.S. team would always win. I'd like to see<mask> you can change my view that it's better that the U.S. just stays out of the whole thing, and lets the rest of the world have something that they can be proud of. [ENDQ] [NEWLINE] I'd cite example 3. The US's reputation is hurt far more by belittling the one sport the rest of the world cares about, then by winning at it. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I hear most often is that they consider that the people from the United Stated are "arrogant". It's my belief that it's not really the arrogance that bothers them, but the fact that whenever America puts its mind to something, it completely dominates. [ENDQ] [NEWLINE] You are sadly mistaken. The rest of the world isn't reeling in pain from the 92 dream team.  People are angry at the US for reasons such as : [NEWLINE] [NEWLINE] 1) Thinking they are the best at things which they aren't the best at. (e.g. Healthcare) [NEWLINE] [NEWLINE] 2) Having a distorted view of their own history (e.g. the view that America single handedly won WWII). [NEWLINE] [NEWLINE] 3) Being largely ignorant/isolationist of what goes on beyond their borders. [NEWLINE] [NEWLINE] 4) Their foreign policy. [NEWLINE] [NEWLINE] Just watch this: [URL] [NEWLINE] [NEWLINE] [STARTQ] They have a competition where all of the teams are (more or less) on equal footing, and the United States sends a team of scrubs who are basically there to get waled on. [ENDQ] [NEWLINE] The US made it to the world cup. That's more than 177 teams. It's frankly insulting to those player to call them scrubs to get waled on. [NEWLINE] [NEWLINE] [STARTQ] I don't necessarily want any to try to change my view about whether or not the U.S. team would always win. I'd like to see if you can change my view that it's better that the U.S. just stays out of the whole thing, and lets the rest of the world have something that they can be proud of. [ENDQ] [NEWLINE] I'd cite example 3. The US's reputation is hurt far more by belittling the one sport the rest of the world cares about, then by winning at it. [NEWLINE] </s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many wealthy people have been *extremely* philanthropic.  You then dismiss their philanthropy on the grounds that they are "playing god and deciding<mask> people can best improve their lives". <mask><mask> that<mask> a wealthy person (or any person) wishes to spend his or her money to help others, it is not unreasonable that he or she gets to decide in<mask> way to help others.  And honestly, giving away all your money to the poor does not bring about any basic change.  It makes the poor temporarily less poor (and<mask> there are literally billions of poor people in the world, even the wealthiest person cannot elevate everybody out of poverty).  Andrew Carnegie built libraries, which are still in use and have benefited a great many people.  Bill Gates has worked to eradicate contagious diseases in India.  Many wealthy people have supported medical research.  These are very worthy goals which would not have been accomplished merely by giving away money to the poor.  It is<mask> not true to dismiss all wealth<mask> being ill-gotten ("rooted in violent historical conquest").  Some is, some isn't.  Some people create new inventions, new businesses, great works of art, and<mask> forth.  The wealth of Bill Gates may be rooted in commercial conquest,<mask> certainly not in violent historical conquest.  And realistically, pretty much everybody would like to be rich<mask> they could.  The difference between the rich and the poor is that the poor have not figured out<mask> to become rich.  There is no moral difference involved.  Rich people have the same virtues<mask> well<mask> the same failings<mask> poor people.  Both the rich and the poor have the capacity to be either selfish or generous, to be compassionate or cruel.  That's<mask> human beings are.  It is not dependent upon money.</s>
Label encoding: <s>Many wealthy people have been *extremely* philanthropic.  You then dismiss their philanthropy on the grounds that they are "playing god and deciding how people can best improve their lives".  I think that if a wealthy person (or any person) wishes to spend his or her money to help others, it is not unreasonable that he or she gets to decide in what way to help others.  And honestly, giving away all your money to the poor does not bring about any basic change.  It makes the poor temporarily less poor (and since there are literally billions of poor people in the world, even the wealthiest person cannot elevate everybody out of poverty).  Andrew Carnegie built libraries, which are still in use and have benefited a great many people.  Bill Gates has worked to eradicate contagious diseases in India.  Many wealthy people have supported medical research.  These are very worthy goals which would not have been accomplished merely by giving away money to the poor.  It is also not true to dismiss all wealth as being ill-gotten ("rooted in violent historical conquest").  Some is, some isn't.  Some people create new inventions, new businesses, great works of art, and so forth.  The wealth of Bill Gates may be rooted in commercial conquest, but certainly not in violent historical conquest.  And realistically, pretty much everybody would like to be rich if they could.  The difference between the rich and the poor is that the poor have not figured out how to become rich.  There is no moral difference involved.  Rich people have the same virtues as well as the same failings as poor people.  Both the rich and the poor have the capacity to be either selfish or generous, to be compassionate or cruel.  That's how human beings are.  It is not dependent upon money.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can you clarify<mask> you mean by "artistic value"? You say his work does not: [NEWLINE] [NEWLINE] [STARTQ] fulfill either a) sending a message or motif; or b) looking artistically pleasing and meaningful. [ENDQ] [NEWLINE] I mean there are some acceptable definitions of 'artistic value' that are true<mask><mask> your point. For example, Rothkos works have a pretty high dollar value, they've motivated a lot of useful thought and discussion, and from<mask> I understand (I don't have the skill to judge myself) his later work actually is obviously executed with clear professional skill and experience. [NEWLINE] [NEWLINE] I don't want to cheat my way to a response here by using one of those. Anyway it seems like the ideas of message, motif, meaning, or pleasure --ideas, all of them-- underlie your concept of "artistic value." [NEWLINE] [NEWLINE] Bear with me here,<mask> maybe there's simply more to art than meaning or intended purpose. Maybe 'effort expended' does not directly correspond to 'artistic value' (or creative value, or deliciousness of flavor.) Maybe there is something simpler, or purer, and those are simply things we're taught to add on to our experience of the world. [NEWLINE] [NEWLINE] For me, Rothko's works helped open up these avenues of thought, basically for the same basic reasons you've started this discussion. I've found this line of thought to be very useful to my perception of the world. [NEWLINE] [NEWLINE] We have our perception, and then we have our understanding of those perceptions. Sometimes you can bring<mask> many preconceptions with you to a new experience that you can miss out on just<mask> satisfying it is to be in the presence of a few huge, simple shapes...<mask> you're too busy looking around for the 'artistic value'.</s>
Label encoding: <s>Can you clarify what you mean by "artistic value"? You say his work does not: [NEWLINE] [NEWLINE] [STARTQ] fulfill either a) sending a message or motif; or b) looking artistically pleasing and meaningful. [ENDQ] [NEWLINE] I mean there are some acceptable definitions of 'artistic value' that are true but besides your point. For example, Rothkos works have a pretty high dollar value, they've motivated a lot of useful thought and discussion, and from what I understand (I don't have the skill to judge myself) his later work actually is obviously executed with clear professional skill and experience. [NEWLINE] [NEWLINE] I don't want to cheat my way to a response here by using one of those. Anyway it seems like the ideas of message, motif, meaning, or pleasure --ideas, all of them-- underlie your concept of "artistic value." [NEWLINE] [NEWLINE] Bear with me here, but maybe there's simply more to art than meaning or intended purpose. Maybe 'effort expended' does not directly correspond to 'artistic value' (or creative value, or deliciousness of flavor.) Maybe there is something simpler, or purer, and those are simply things we're taught to add on to our experience of the world. [NEWLINE] [NEWLINE] For me, Rothko's works helped open up these avenues of thought, basically for the same basic reasons you've started this discussion. I've found this line of thought to be very useful to my perception of the world. [NEWLINE] [NEWLINE] We have our perception, and then we have our understanding of those perceptions. Sometimes you can bring so many preconceptions with you to a new experience that you can miss out on just how satisfying it is to be in the presence of a few huge, simple shapes... because you're too busy looking around for the 'artistic value'.</s>
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, I don't aqree that we should be "governing" people at all. [NEWLINE] [NEWLINE] <mask> going on that premise,<mask><mask> that democracy sucks; totally overrated and makes the minority victim of the majority's stupidity. [NEWLINE] [NEWLINE] <mask>, the forms of governance in place are reflective of the economics on some level and<mask> there is still a competitive system<mask> people are gaming for advantage and pursuing money<mask> a commodity for its own sake, natural gravitations of power are still going to occur.  These gravitations of power (corporations and conglomerates) will fortify their base of power in the market, tending toward a monopoly with no government restriction (Im assuming by voluntaryism you dont think too much of govt interference in market??<mask> Im mistaken then no matter, these corporate entities will tend toward oligopoly or cartel which will have the same result<mask> monopoly on the back end).<mask> they consolidate power, they will keep propagating the myth that everyone is living in a voluntary system<mask> in reality people will be forced to engage in the market and be hopelessly dependant on it.  For without governmental fetters (again, Im assuming you eschew government market controls) all land would be bought up and people would be forced off it leaving them no choice to take whatever job they could in order to pay to live on this planet which was sold to large business interests "voluntarily." [NEWLINE] [NEWLINE] You may recognize this scenario<mask> similar to<mask> the world works today, except the myth that is propagated is democracy and not voluntaryism. <mask> I guarantee you, until we reach a post-scarcity, post-ownership[world, you can put in place whatever lofty ism you want and it will be the same rich and poor with a different set of lies. </s>
Label encoding: <s>Well, I don't aqree that we should be "governing" people at all. [NEWLINE] [NEWLINE] But going on that premise, I agree that democracy sucks; totally overrated and makes the minority victim of the majority's stupidity. [NEWLINE] [NEWLINE] However, the forms of governance in place are reflective of the economics on some level and if there is still a competitive system where people are gaming for advantage and pursuing money as a commodity for its own sake, natural gravitations of power are still going to occur.  These gravitations of power (corporations and conglomerates) will fortify their base of power in the market, tending toward a monopoly with no government restriction (Im assuming by voluntaryism you dont think too much of govt interference in market?? If Im mistaken then no matter, these corporate entities will tend toward oligopoly or cartel which will have the same result as monopoly on the back end). As they consolidate power, they will keep propagating the myth that everyone is living in a voluntary system while in reality people will be forced to engage in the market and be hopelessly dependant on it.  For without governmental fetters (again, Im assuming you eschew government market controls) all land would be bought up and people would be forced off it leaving them no choice to take whatever job they could in order to pay to live on this planet which was sold to large business interests "voluntarily." [NEWLINE] [NEWLINE] You may recognize this scenario as similar to how the world works today, except the myth that is propagated is democracy and not voluntaryism.  But I guarantee you, until we reach a post-scarcity, post-ownership[world, you can put in place whatever lofty ism you want and it will be the same rich and poor with a different set of lies. </s>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the connotation in society is generally that of purity. [ENDQ] [NEWLINE] I would say that's still subjective.<mask> is 'purity' and to whom? Is it not killing or stealing? Is it following a religious doctrine? Is it not being homosexual? Is not not masturbating? Is it not being Jewish (to go back to the WW2 example)? [NEWLINE] [NEWLINE] I'd say that'societies' definition is going to vary wildly depending on who,<mask>, and<mask> you ask it. It's still an arbitrary umbrella term at this point.<mask>,<mask> it is to be objective at some level,<mask> is that objectivity derived from? Is it just<mask> people generally consider to be bad? Or is there some 3rd party that determines this objective definition (similar to<mask> the religious view a God<mask> the absolute source of morality)? [NEWLINE] [NEWLINE] I will give you that the concept of being 'purely bad' is not realistic. Hitler was kind to children at the same time he was orchestrating one of the biggest exterminations in history (I really need to drop these WW2 examples sometime). Nothing is that clear cut.<mask> that doesn't stop me from considering him an evil person. [NEWLINE] [NEWLINE] [STARTQ] That may be true in your circumstance,<mask> is it true from a societal perspective? [ENDQ] [NEWLINE] I'm not really sure<mask> you're asking. It's true that some groups of people may toss around the term 'evil'<mask> a dismissive statement.<mask>,<mask><mask> that would happen, even<mask> the word didn't exist.<mask> someone dismisses a person<mask> they consider them evil, then the judgment has already been made in their mind. Calling them evil is just a reflection of that. Personally, I don't hand that term down lightly,<mask> I can see<mask> others might.</s>
Label encoding: <s> [STARTQ] the connotation in society is generally that of purity. [ENDQ] [NEWLINE] I would say that's still subjective. What is 'purity' and to whom? Is it not killing or stealing? Is it following a religious doctrine? Is it not being homosexual? Is not not masturbating? Is it not being Jewish (to go back to the WW2 example)? [NEWLINE] [NEWLINE] I'd say that'societies' definition is going to vary wildly depending on who, where, and when you ask it. It's still an arbitrary umbrella term at this point. But, if it is to be objective at some level, where is that objectivity derived from? Is it just what people generally consider to be bad? Or is there some 3rd party that determines this objective definition (similar to how the religious view a God as the absolute source of morality)? [NEWLINE] [NEWLINE] I will give you that the concept of being 'purely bad' is not realistic. Hitler was kind to children at the same time he was orchestrating one of the biggest exterminations in history (I really need to drop these WW2 examples sometime). Nothing is that clear cut. But that doesn't stop me from considering him an evil person. [NEWLINE] [NEWLINE] [STARTQ] That may be true in your circumstance, but is it true from a societal perspective? [ENDQ] [NEWLINE] I'm not really sure what you're asking. It's true that some groups of people may toss around the term 'evil' as a dismissive statement. But, I think that would happen, even if the word didn't exist. If someone dismisses a person because they consider them evil, then the judgment has already been made in their mind. Calling them evil is just a reflection of that. Personally, I don't hand that term down lightly, but I can see how others might.</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There has been almost no better time and place to raise children in the *history* of humanity than today. There are enormous institutional problems today, of course,<mask> think about the following. [NEWLINE] [NEWLINE] Prejudice based on race, gender, and sexual orientation is the lowest it has been in centuries.<mask> institutional problems leading to income inequality we have a (somewhat effective) social safety net in place and regulations that prevent the horrible abuses towards workers that existed in the in the 19th century. Due to the development of nuclear weapons, total war is no longer a threat to the world (for the foreseeable future). Education is cheaper and more accessible than ever before thanks to the internet and I would expect that this would extend into colleges by the time your children are 18. Georgia Tech is currently trying a pilot program to offer an online Masters in Computer Science for just a few thousand dollars. We have better access to live saving medicine than ever before in history. We have a greater understanding of mental illness to the likelihood that your children will be saddled with crippling depression or anxiety is limited. [NEWLINE] [NEWLINE] I could go on and on. [NEWLINE] [NEWLINE] My real point is that<mask> you truly believe that it is irresponsible to have a child today then (barring two arguments I'll get to in a moment) it must have been irresponsible to have a child for *most of human history*. You might agree that that is the case,<mask> at least you must acknowledge that it is a pretty extreme view. [NEWLINE] [NEWLINE] The only arguments that work for modern times<mask> not past decades or centuries are overpopulation and climate change.<mask><mask> you can make a valid argument that we should be having fewer children<mask> of these problems (<mask> certainly not zero children)<mask> that isn't the argument you seem to be making. </s>
Label encoding: <s>There has been almost no better time and place to raise children in the *history* of humanity than today. There are enormous institutional problems today, of course, but think about the following. [NEWLINE] [NEWLINE] Prejudice based on race, gender, and sexual orientation is the lowest it has been in centuries. Despite institutional problems leading to income inequality we have a (somewhat effective) social safety net in place and regulations that prevent the horrible abuses towards workers that existed in the in the 19th century. Due to the development of nuclear weapons, total war is no longer a threat to the world (for the foreseeable future). Education is cheaper and more accessible than ever before thanks to the internet and I would expect that this would extend into colleges by the time your children are 18. Georgia Tech is currently trying a pilot program to offer an online Masters in Computer Science for just a few thousand dollars. We have better access to live saving medicine than ever before in history. We have a greater understanding of mental illness to the likelihood that your children will be saddled with crippling depression or anxiety is limited. [NEWLINE] [NEWLINE] I could go on and on. [NEWLINE] [NEWLINE] My real point is that if you truly believe that it is irresponsible to have a child today then (barring two arguments I'll get to in a moment) it must have been irresponsible to have a child for *most of human history*. You might agree that that is the case, but at least you must acknowledge that it is a pretty extreme view. [NEWLINE] [NEWLINE] The only arguments that work for modern times but not past decades or centuries are overpopulation and climate change. I think you can make a valid argument that we should be having fewer children because of these problems ( but certainly not zero children) but that isn't the argument you seem to be making. </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As sexual activity is not inherently harmful to children, there is no basis on which to label all sexual relations between children and adults<mask> abusive. [NEWLINE] [NEWLINE] [STARTQ] A Dutch study published in 1987 found that a sample of boys in paedophilic relationships felt positively about them. And a major<mask> still controversial 1998-2000 meta-study suggests –<mask> J Michael Bailey of Northwestern University, Chicago, says – that such relationships, entered into voluntarily, are "nearly uncorrelated with undesirable outcomes". [ENDQ] [NEWLINE] [STARTQ] Most people find that idea impossible.<mask> writing last year in the peer-reviewed Archives of Sexual Behaviour, Bailey said that<mask> he<mask> found the notion "disturbing", he was forced to recognise that **"persuasive evidence for the harmfulness of paedophilic relationships does not<mask> exist".** [ENDQ] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] A substantial number of people who<mask> children had sex with adults feel positively about the experience, and do not regard it<mask> abusive in any capacity. [NEWLINE] [NEWLINE] Long-Range Effects of Child and Adolescent Sexual Experiences Positive Review", Allie C. Kilpatrick. [NEWLINE] [NEWLINE] &gt;This book will be disturbing to many readers. The assumption that all children are "damaged" by their experiences is challenged by Kilpatrick's finding that 38% of the adult respondents reported the sexual experiences<mask> children to be "pleasant"<mask> only 25% reported them to be "unpleasant." Kilpatrick<mask> found that,<mask> the majority of the women stated that the experience was initiated by the partner, for many (23% of the children 0-14 years and 39% of adolescents 15-17 years) the women reported having been the initiator. Another surprising finding was that only 4% of the respondents reported that they would have liked to have had counseling.</s>
Label encoding: <s>As sexual activity is not inherently harmful to children, there is no basis on which to label all sexual relations between children and adults as abusive. [NEWLINE] [NEWLINE] [STARTQ] A Dutch study published in 1987 found that a sample of boys in paedophilic relationships felt positively about them. And a major if still controversial 1998-2000 meta-study suggests – as J Michael Bailey of Northwestern University, Chicago, says – that such relationships, entered into voluntarily, are "nearly uncorrelated with undesirable outcomes". [ENDQ] [NEWLINE] [STARTQ] Most people find that idea impossible. But writing last year in the peer-reviewed Archives of Sexual Behaviour, Bailey said that while he also found the notion "disturbing", he was forced to recognise that **"persuasive evidence for the harmfulness of paedophilic relationships does not yet exist".** [ENDQ] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] A substantial number of people who as children had sex with adults feel positively about the experience, and do not regard it as abusive in any capacity. [NEWLINE] [NEWLINE] Long-Range Effects of Child and Adolescent Sexual Experiences Positive Review", Allie C. Kilpatrick. [NEWLINE] [NEWLINE] &gt;This book will be disturbing to many readers. The assumption that all children are "damaged" by their experiences is challenged by Kilpatrick's finding that 38% of the adult respondents reported the sexual experiences as children to be "pleasant" while only 25% reported them to be "unpleasant." Kilpatrick also found that, although the majority of the women stated that the experience was initiated by the partner, for many (23% of the children 0-14 years and 39% of adolescents 15-17 years) the women reported having been the initiator. Another surprising finding was that only 4% of the respondents reported that they would have liked to have had counseling.</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> it is a bit patronizing to suggest people re-evaluate their preferences [ENDQ] [NEWLINE] I may be misunderstanding<mask> you are saying here,<mask> re-evaluating something is almost never a bad idea. Re-evaluating a situation doesn't mean there has to be change, it is just giving a second thought to the matter. Sometimes you may find that your initial reaction was based on a strange internal-bias that really shouldn't inform your decision making, other times your initial reaction can be totally correct. For your smoking example it could take ten seconds to re-evaluate, 'I won't date someone who does something with huge long-term health implications, smoking has those'.<mask> sometimes it goes further, you have to investigate whether you are consistent: 'Do I reject all potential partners who indulge in actions with large long-term health implications?<mask> not then<mask>?'. [NEWLINE] [NEWLINE] Sure, it is insulting<mask> you have put long days of thought into a position and someone just offhandedly says 'you should re-evaluate that',<mask> generally they have no idea<mask> much effort you have put in. It would be more helpful<mask> they asked questions like '<mask> did you come to this position?' and try to understand your reasoning and help investigate whether it is consistent,<mask> that takes a lot of effort and sometimes people just aren't willing to engage in it. [NEWLINE] [NEWLINE] Personally I like to have my preferences questioned<mask> it prompts me to examine whether I have a good reason for my choices or have been restricting my actions based on entirely arbitrary principles. I am eternally grateful for one of my friends who doesn't hesitate to ask '<mask> do you believe that?'<mask> they helped me examine biases I didn't  even suspect I possessed and freed me from entirely artificial restraints. </s>
Label encoding: <s> [STARTQ] I think it is a bit patronizing to suggest people re-evaluate their preferences [ENDQ] [NEWLINE] I may be misunderstanding what you are saying here, but re-evaluating something is almost never a bad idea. Re-evaluating a situation doesn't mean there has to be change, it is just giving a second thought to the matter. Sometimes you may find that your initial reaction was based on a strange internal-bias that really shouldn't inform your decision making, other times your initial reaction can be totally correct. For your smoking example it could take ten seconds to re-evaluate, 'I won't date someone who does something with huge long-term health implications, smoking has those'. But sometimes it goes further, you have to investigate whether you are consistent: 'Do I reject all potential partners who indulge in actions with large long-term health implications? If not then why?'. [NEWLINE] [NEWLINE] Sure, it is insulting if you have put long days of thought into a position and someone just offhandedly says 'you should re-evaluate that', but generally they have no idea how much effort you have put in. It would be more helpful if they asked questions like'how did you come to this position?' and try to understand your reasoning and help investigate whether it is consistent, but that takes a lot of effort and sometimes people just aren't willing to engage in it. [NEWLINE] [NEWLINE] Personally I like to have my preferences questioned because it prompts me to examine whether I have a good reason for my choices or have been restricting my actions based on entirely arbitrary principles. I am eternally grateful for one of my friends who doesn't hesitate to ask'why do you believe that?' because they helped me examine biases I didn't  even suspect I possessed and freed me from entirely artificial restraints. </s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is more to morality than depriving others of something. [NEWLINE] [NEWLINE] Let's start by imagining a situation<mask> everyone circumvents ads. In such a situation, a creator spends time making stuff and people consume it freely. The creator gets no remuneration<mask> he provided something valuable enough for consumers to choose to consume it. Everyone got value out of the deal except the one person who deserved it. [NEWLINE] [NEWLINE] Clearly that is not moral. [NEWLINE] [NEWLINE] <mask> the question becomes:<mask> enough people pay for a good, does it eventually become moral to consume it without remunerating the creator? [NEWLINE] [NEWLINE] Let's say the cutoff is 50%.<mask> I'm watching the ads and you're not, is it immoral for me to start circumventing them? It was moral for you, it doesn't seem fair for it to suddenly become immoral for me, particularly<mask> I've been remunerating the creator from the beginning<mask> you weren't. [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] Your point that<mask> there is no direct consequence on the creator whether you don't consume their content or consume it without remunerating them ignores the ripple effect such a behaviour has (see above). [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] <mask>, consuming the content without paying is not "voting with your wallet". Voting with your wallet is something<mask> you pay once the price is low enough to be fair to you. [NEWLINE] [NEWLINE] Essentially,<mask> I want something<mask> am not willing to pay the price it is sold at, I can pay with my wallet by waiting for it to get to a pricepoint I am willing to pay for. [NEWLINE] [NEWLINE] In your case, no matter<mask> much the "price" lowers, it won't matter, you've already consumed the content. [NEWLINE] [NEWLINE] It is not voting with your wallet.</s>
Label encoding: <s>There is more to morality than depriving others of something. [NEWLINE] [NEWLINE] Let's start by imagining a situation where everyone circumvents ads. In such a situation, a creator spends time making stuff and people consume it freely. The creator gets no remuneration yet he provided something valuable enough for consumers to choose to consume it. Everyone got value out of the deal except the one person who deserved it. [NEWLINE] [NEWLINE] Clearly that is not moral. [NEWLINE] [NEWLINE] So the question becomes: if enough people pay for a good, does it eventually become moral to consume it without remunerating the creator? [NEWLINE] [NEWLINE] Let's say the cutoff is 50%. If I'm watching the ads and you're not, is it immoral for me to start circumventing them? It was moral for you, it doesn't seem fair for it to suddenly become immoral for me, particularly since I've been remunerating the creator from the beginning while you weren't. [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] Your point that because there is no direct consequence on the creator whether you don't consume their content or consume it without remunerating them ignores the ripple effect such a behaviour has (see above). [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] Lastly, consuming the content without paying is not "voting with your wallet". Voting with your wallet is something where you pay once the price is low enough to be fair to you. [NEWLINE] [NEWLINE] Essentially, if I want something but am not willing to pay the price it is sold at, I can pay with my wallet by waiting for it to get to a pricepoint I am willing to pay for. [NEWLINE] [NEWLINE] In your case, no matter how much the "price" lowers, it won't matter, you've already consumed the content. [NEWLINE] [NEWLINE] It is not voting with your wallet.</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's<mask> much rhetoric behind this statement that no one seems to just stop and look at<mask>'s actually happening. Most parents don't force Barbies or GI Joes on their kids, it just kind of happens in most cases. [ENDQ] [NEWLINE] Does it really? [NEWLINE] [NEWLINE] This is the part<mask> I start to lose respect for you,<mask> I wonder<mask> you actually read my post: [NEWLINE] [NEWLINE] [STARTQ] <mask> apparently, it's a big issue that boys like trucks and girls like dolls - for some reason that I'm not sure I quite understand. [ENDQ] [NEWLINE] [Is this a good enough reason?]( [URL].com/index.php?db=comics&amp;id=1883#comic) [NEWLINE] [NEWLINE] Not a good enough reason to assume it's a problem,<mask> maybe a good enough reason to make sure the girl actually wants the doll<mask> she actually wants it, instead of just wanting the doll<mask> "Boys like trucks, girls like dolls" has been: [NEWLINE] [NEWLINE] * Modeled by parents [NEWLINE] * Modeled by box art [NEWLINE] * Modeled by advertising [NEWLINE] * All-<mask> -*enforced* by toy stores, with a "boys" section and a "girls" section [NEWLINE] [NEWLINE] To say nothing of TV, movies, and<mask> on,<mask> the above are points we can at least reasonably address. [NEWLINE] [NEWLINE] Which brings me back to: Does it really *just happen*? Or is there *massive* societal influence here? [NEWLINE] [NEWLINE] <mask> you really think it *just happens,* let me ask you something else: *<mask> * do you think it happens? Do you think that there's something biological that's happening to kids who haven't even hit puberty to make them pick one of these or the other?</s>
Label encoding: <s> [STARTQ] There's so much rhetoric behind this statement that no one seems to just stop and look at what's actually happening. Most parents don't force Barbies or GI Joes on their kids, it just kind of happens in most cases. [ENDQ] [NEWLINE] Does it really? [NEWLINE] [NEWLINE] This is the part where I start to lose respect for you, because I wonder if you actually read my post: [NEWLINE] [NEWLINE] [STARTQ] But apparently, it's a big issue that boys like trucks and girls like dolls - for some reason that I'm not sure I quite understand. [ENDQ] [NEWLINE] [Is this a good enough reason?]( [URL].com/index.php?db=comics&amp;id=1883#comic) [NEWLINE] [NEWLINE] Not a good enough reason to assume it's a problem, but maybe a good enough reason to make sure the girl actually wants the doll because she actually wants it, instead of just wanting the doll because "Boys like trucks, girls like dolls" has been: [NEWLINE] [NEWLINE] * Modeled by parents [NEWLINE] * Modeled by box art [NEWLINE] * Modeled by advertising [NEWLINE] * All- but -*enforced* by toy stores, with a "boys" section and a "girls" section [NEWLINE] [NEWLINE] To say nothing of TV, movies, and so on, but the above are points we can at least reasonably address. [NEWLINE] [NEWLINE] Which brings me back to: Does it really *just happen*? Or is there *massive* societal influence here? [NEWLINE] [NEWLINE] If you really think it *just happens,* let me ask you something else: * Why * do you think it happens? Do you think that there's something biological that's happening to kids who haven't even hit puberty to make them pick one of these or the other?</s>
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] *Not<mask> much<mask> my father's friends did by pointing out every woman they saw they'd like to fuck.* [ENDQ] [NEWLINE] <mask> does that have anything to do with rape?<mask> the hell is wrong with a guy saying he wants to have sex with women? [NEWLINE] [NEWLINE] [STARTQ] *There are a lot more social constructs at play than just feminism.<mask><mask><mask><mask> that feminism plays ANY role in it, I'd say its more a problem with the masculine social construct.* [ENDQ] [NEWLINE] This is victim blaming plain and simple. You are saying that the feminist's negative portrayal of men and masculinity is correct and that men must act more like women in order to solve their problems. [NEWLINE] [NEWLINE] That is an inherently sexist point of view. [NEWLINE] [NEWLINE] [STARTQ] *All MRA arguements, yours included, frame feminism<mask> the bad guy.<mask>?<mask> not just say "rape is bad, we shouldn't make a joke of any victim."* [ENDQ] [NEWLINE] <mask> feminists don't leave men alone.<mask> feminists have historically hidden and marginalized male victims and denied them access to any help.<mask> feminists lie about rape statistics.<mask> feminists promote the lie that rape is a gendered crime.<mask> feminists encourage our society to view women<mask> victims and men<mask> perpetrators.<mask> feminists ARE the bad guy. [NEWLINE] [NEWLINE] 1.<mask> feminists never said another word about men. [NEWLINE] 2. And never tried to pass another law that infringed on the rights of men or demanded that men pay to subsidize their lifestyles. [NEWLINE] [NEWLINE] Then we wouldn't have a problem with them.<mask><mask> is, feminists attack masculinity, men, boys, maleness on a constant basis. Feminists have been waging a gender war on men for the past 50 years. We're just beginning to fight back. </s>
Label encoding: <s> [STARTQ] *Not as much as my father's friends did by pointing out every woman they saw they'd like to fuck.* [ENDQ] [NEWLINE] How does that have anything to do with rape? What the hell is wrong with a guy saying he wants to have sex with women? [NEWLINE] [NEWLINE] [STARTQ] *There are a lot more social constructs at play than just feminism. In fact I disagree that feminism plays ANY role in it, I'd say its more a problem with the masculine social construct.* [ENDQ] [NEWLINE] This is victim blaming plain and simple. You are saying that the feminist's negative portrayal of men and masculinity is correct and that men must act more like women in order to solve their problems. [NEWLINE] [NEWLINE] That is an inherently sexist point of view. [NEWLINE] [NEWLINE] [STARTQ] *All MRA arguements, yours included, frame feminism as the bad guy. Why? why not just say "rape is bad, we shouldn't make a joke of any victim."* [ENDQ] [NEWLINE] Because feminists don't leave men alone. Because feminists have historically hidden and marginalized male victims and denied them access to any help. Because feminists lie about rape statistics. Because feminists promote the lie that rape is a gendered crime. Because feminists encourage our society to view women as victims and men as perpetrators. Because feminists ARE the bad guy. [NEWLINE] [NEWLINE] 1. If feminists never said another word about men. [NEWLINE] 2. And never tried to pass another law that infringed on the rights of men or demanded that men pay to subsidize their lifestyles. [NEWLINE] [NEWLINE] Then we wouldn't have a problem with them. But as is, feminists attack masculinity, men, boys, maleness on a constant basis. Feminists have been waging a gender war on men for the past 50 years. We're just beginning to fight back. </s>
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do we know it is not a white swan that got soaked in tar? [ENDQ] [NEWLINE] This is a natural problem with<mask> observations work, and has been the cause of many false theories throughout time. I feel that this is covered pretty well by<mask> truth is defined,<mask>. All hypothesis', theories and laws are based on<mask> we,<mask> humans, have observed. Are these observations perhaps false/imperfect? Yes,<mask> we never claimed they were not - they are simply the best we have for now.<mask> contradicting observations, or a theory that explains the observations better, come along, we're more than ready to scrap our previous ideas. [NEWLINE] [NEWLINE] In conclusion, the swan might be white -<mask> until observations happen that show it is, it is considered black in our view. This means that we might not have a perfect view of the world -<mask> we never claim to have one. [NEWLINE] [NEWLINE] Science is an adaptive field - its ultimate goal is to uncover every truth about the world we live in, and it attempts to do<mask> through proposing ideas, testing them, and then refining or scrapping them. Unfortunately, we are limited in the sense that we can only base our theories on<mask> we experience, and<mask> we experience is limited by our ability to correctly observe. There have been written many, many books around this concept,<mask> I'm afraid I do not know enough about it to have in-depth discussions on the topic. [NEWLINE] [NEWLINE] [STARTQ] The question is meant to raise the other question of who decides whether something is silly or unreasonable? [ENDQ] [NEWLINE] Such things<mask> "silly" or "unreasonable" generally do not exist in science. For<mask><mask><mask> whatever you suggest is falsifiable and testable, it is science.</s>
Label encoding: <s> [STARTQ] How do we know it is not a white swan that got soaked in tar? [ENDQ] [NEWLINE] This is a natural problem with how observations work, and has been the cause of many false theories throughout time. I feel that this is covered pretty well by how truth is defined, however. All hypothesis', theories and laws are based on what we, as humans, have observed. Are these observations perhaps false/imperfect? Yes, but we never claimed they were not - they are simply the best we have for now. If contradicting observations, or a theory that explains the observations better, come along, we're more than ready to scrap our previous ideas. [NEWLINE] [NEWLINE] In conclusion, the swan might be white - but until observations happen that show it is, it is considered black in our view. This means that we might not have a perfect view of the world - but we never claim to have one. [NEWLINE] [NEWLINE] Science is an adaptive field - its ultimate goal is to uncover every truth about the world we live in, and it attempts to do so through proposing ideas, testing them, and then refining or scrapping them. Unfortunately, we are limited in the sense that we can only base our theories on what we experience, and what we experience is limited by our ability to correctly observe. There have been written many, many books around this concept, but I'm afraid I do not know enough about it to have in-depth discussions on the topic. [NEWLINE] [NEWLINE] [STARTQ] The question is meant to raise the other question of who decides whether something is silly or unreasonable? [ENDQ] [NEWLINE] Such things as "silly" or "unreasonable" generally do not exist in science. For as long as whatever you suggest is falsifiable and testable, it is science.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You kind of answered your own question. There are<mask> many different levels of psychiatric disorder that it just can't be limited on a blanket basis.  Should someone who is overtly violent and threatening be limited? Sure, everyone can agree on that.<mask><mask> about the other cases? [NEWLINE] [NEWLINE] Seasonal depression (which is caused by vitamin deficiencies) is a mood disorder. Should someone be limited due to feeling down every now and then? [NEWLINE] [NEWLINE] My wife was diagnosed<mask> bipolar<mask> she was in Highschool (<mask> hormones are at their craziest) and placed on all sorts of medications for it. 10 years later, she decides to get off the medication and come off birth control, and voila! She has absolutely no issues whatsoever. Should she be barred? [NEWLINE] [NEWLINE] PTSD has hundreds of symptoms. I have no problems with a normal life. Play with my kid, work, deal with people daily, my only actual "symptom" of PTSD is I don't like large/loud gatherings of people. Should I be barred from owning firearms<mask> I don't like malls or concerts? Not to mention that soldiers are not the leading group of PTSD sufferers (that title belongs to car accident victims). [NEWLINE] [NEWLINE] <mask> about a woman who has has PTSD being abused by her ex-husband? Should she be barred from protecting herself (or her kids)<mask> he violates that restraining order and tries to break into her apartment to attack her? [NEWLINE] [NEWLINE] Tl;dr version: The severity of mental illnesses can vary<mask> much (and be<mask> over diagnosed) that a blanket law barring ownership would do much more harm than good. A case by case basis,<mask> is already available (and rarely used) by law is a better system. </s>
Label encoding: <s>You kind of answered your own question. There are so many different levels of psychiatric disorder that it just can't be limited on a blanket basis.  Should someone who is overtly violent and threatening be limited? Sure, everyone can agree on that. But what about the other cases? [NEWLINE] [NEWLINE] Seasonal depression (which is caused by vitamin deficiencies) is a mood disorder. Should someone be limited due to feeling down every now and then? [NEWLINE] [NEWLINE] My wife was diagnosed as bipolar when she was in Highschool ( when hormones are at their craziest) and placed on all sorts of medications for it. 10 years later, she decides to get off the medication and come off birth control, and voila! She has absolutely no issues whatsoever. Should she be barred? [NEWLINE] [NEWLINE] PTSD has hundreds of symptoms. I have no problems with a normal life. Play with my kid, work, deal with people daily, my only actual "symptom" of PTSD is I don't like large/loud gatherings of people. Should I be barred from owning firearms because I don't like malls or concerts? Not to mention that soldiers are not the leading group of PTSD sufferers (that title belongs to car accident victims). [NEWLINE] [NEWLINE] What about a woman who has has PTSD being abused by her ex-husband? Should she be barred from protecting herself (or her kids) if he violates that restraining order and tries to break into her apartment to attack her? [NEWLINE] [NEWLINE] Tl;dr version: The severity of mental illnesses can vary so much (and be so over diagnosed) that a blanket law barring ownership would do much more harm than good. A case by case basis, as is already available (and rarely used) by law is a better system. </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm on reddit on my phone.... And get response notifications. Thanks for making judgements about my life. [NEWLINE] Growing thick skin is a good idea,<mask> learning to deal with people trying to hurt you is better. It will make you a better person. It's not like I'm saying it's something you should have to go through, just that you probably will have to, and it's best to learn to handle it. Yes, you do deal with bullies your whole life. At work, in your family(hopefully not,<mask> many of my friends are not<mask> lucky), around town, some people have the need to bring down others, and some are very persistent. The solution isn't<mask> easy<mask> disconnecting your devices,<mask><mask>.<mask>, that's not a bad start. The bullies I deal with an adult are<mask> much worse than high school. Instead of petty rumors, one, for example, just recently has got a substantial amount of people convinced my good friend and housemate is a child molester, just to satisfy their own personal vendetta. That's a bully. [NEWLINE] [NEWLINE] <mask><mask> bullying has lifelong consequences,<mask> that's up to the person being bullied. You can let it drag you down, or you can let it make you stronger. A child might not be able to understand that<mask> other people say about them doesn't matter, an adult should. [NEWLINE] [NEWLINE] I'm not trying to be all, tough it up kid,<mask> at the same time I believe it's a normal part of growing up. It's not fair,<mask> that's one of the first lessons on<mask> life isn't fair. It's an important lesson. Overcoming adversity is crucial to being a strong person.</s>
Label encoding: <s>I'm on reddit on my phone.... And get response notifications. Thanks for making judgements about my life. [NEWLINE] Growing thick skin is a good idea, but learning to deal with people trying to hurt you is better. It will make you a better person. It's not like I'm saying it's something you should have to go through, just that you probably will have to, and it's best to learn to handle it. Yes, you do deal with bullies your whole life. At work, in your family(hopefully not, but many of my friends are not so lucky), around town, some people have the need to bring down others, and some are very persistent. The solution isn't as easy as disconnecting your devices, I agree. However, that's not a bad start. The bullies I deal with an adult are so much worse than high school. Instead of petty rumors, one, for example, just recently has got a substantial amount of people convinced my good friend and housemate is a child molester, just to satisfy their own personal vendetta. That's a bully. [NEWLINE] [NEWLINE] I agree bullying has lifelong consequences, but that's up to the person being bullied. You can let it drag you down, or you can let it make you stronger. A child might not be able to understand that what other people say about them doesn't matter, an adult should. [NEWLINE] [NEWLINE] I'm not trying to be all, tough it up kid, but at the same time I believe it's a normal part of growing up. It's not fair, but that's one of the first lessons on why life isn't fair. It's an important lesson. Overcoming adversity is crucial to being a strong person.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The best way I heard someone put this was: "<mask> you think women have it better just<mask> you want to fuck them?" [NEWLINE] [NEWLINE] Most of your points boil down to "women are sexually desirable to men,<mask> women have it better." And<mask> you think about it, this is only really an issue in a context<mask> men happen to have the majority of the power. Women are generally attracted to men, too.<mask> is the converse not true?<mask> you're an attractive man, don't you have just<mask> much power over women? And<mask> about unattractive women? Do you think between a man and a woman of equal physical attractiveness, the woman has more opportunities and faces less discrimination? [NEWLINE] [NEWLINE] Men today typically get paid more; occupy more positions of power; are catered to by default in society. In just about every measure that matters I can think of, men do better than women (I guess women have it slightly better<mask> it comes to LGBT issues,<mask> that's the only one I can think of off the top of my head). Women aren't seen<mask> "sexually desirable"<mask> society values women more than men; women are seen<mask> sexually desirable<mask> straight men have most of the power, and straight men desire women, and<mask> women are seen more sexually<mask> that's<mask> straight men want. That's<mask> the idea of objectification comes from. [NEWLINE] [NEWLINE] The best counterexample to this is,<mask> I mentioned, an ugly woman.<mask> many ugly women do you see in media?<mask> are they depicted? Are they "valued" in the same way you think attractive women are valued?<mask> do you treat a lack of physical attractiveness in a woman<mask> opposed to a man? </s>
Label encoding: <s>The best way I heard someone put this was: " So you think women have it better just because you want to fuck them?" [NEWLINE] [NEWLINE] Most of your points boil down to "women are sexually desirable to men, therefore women have it better." And if you think about it, this is only really an issue in a context where men happen to have the majority of the power. Women are generally attracted to men, too. Why is the converse not true? If you're an attractive man, don't you have just as much power over women? And what about unattractive women? Do you think between a man and a woman of equal physical attractiveness, the woman has more opportunities and faces less discrimination? [NEWLINE] [NEWLINE] Men today typically get paid more; occupy more positions of power; are catered to by default in society. In just about every measure that matters I can think of, men do better than women (I guess women have it slightly better when it comes to LGBT issues, but that's the only one I can think of off the top of my head). Women aren't seen as "sexually desirable" because society values women more than men; women are seen as sexually desirable because straight men have most of the power, and straight men desire women, and therefore women are seen more sexually because that's what straight men want. That's where the idea of objectification comes from. [NEWLINE] [NEWLINE] The best counterexample to this is, as I mentioned, an ugly woman. How many ugly women do you see in media? How are they depicted? Are they "valued" in the same way you think attractive women are valued? How do you treat a lack of physical attractiveness in a woman as opposed to a man? </s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Men's Rights supports some great issues, such<mask> father's rights in custody battles, equity in monetary settlements for divorce proceedings, getting the FBI to recognize rapped men<mask> victims of rape rather than assault, and<mask> forth. [NEWLINE] [NEWLINE] It<mask> has it's share of crazy nutters who are doing whatever they can in order to get news coverage. And people who simply are filled with hatred for others and have found a cover for their bigotry. [NEWLINE] [NEWLINE] In this way they are no different from feminism. [NEWLINE] [NEWLINE] People forget that feminism and women's rights  has a huge past history of prominent members stating with absolute clarity that their primary issue is that they are nut jobs who hate men. [NEWLINE] [NEWLINE] There's Andrea Dworkin's famous "I want to see a man beaten to a bloody pulp with a high-heel shoved in his mouth, like an apple in the mouth of a pig." and Susan Brownmiller's "Rape is nothing more or less than a conscious process of intimidation by which all men keep all women in a state of fear." And Catherine MacKinnon's claim that all consensual heterosexual sex is rape<mask> women<mask> a group are not in a position of social equality<mask> no woman can actually consent to sex.  And Sally Gearhart stating that men should be killed off and only a small breeding stock maintained to guarantee survival of the species. And Catherine Comins saying that unjust accusations of rape are a good thing. And on and on and on. [NEWLINE] [NEWLINE] These weren't statements from one-off individuals. They were prominent and visible members of the feminist movement. [NEWLINE] [NEWLINE] <mask><mask> you can forgive feminism it's idiots retrospectively,<mask> do you find it hard to forgive men's rights their idiots concurrently?</s>
Label encoding: <s>Men's Rights supports some great issues, such as father's rights in custody battles, equity in monetary settlements for divorce proceedings, getting the FBI to recognize rapped men as victims of rape rather than assault, and so forth. [NEWLINE] [NEWLINE] It also has it's share of crazy nutters who are doing whatever they can in order to get news coverage. And people who simply are filled with hatred for others and have found a cover for their bigotry. [NEWLINE] [NEWLINE] In this way they are no different from feminism. [NEWLINE] [NEWLINE] People forget that feminism and women's rights  has a huge past history of prominent members stating with absolute clarity that their primary issue is that they are nut jobs who hate men. [NEWLINE] [NEWLINE] There's Andrea Dworkin's famous "I want to see a man beaten to a bloody pulp with a high-heel shoved in his mouth, like an apple in the mouth of a pig." and Susan Brownmiller's "Rape is nothing more or less than a conscious process of intimidation by which all men keep all women in a state of fear." And Catherine MacKinnon's claim that all consensual heterosexual sex is rape because women as a group are not in a position of social equality so no woman can actually consent to sex.  And Sally Gearhart stating that men should be killed off and only a small breeding stock maintained to guarantee survival of the species. And Catherine Comins saying that unjust accusations of rape are a good thing. And on and on and on. [NEWLINE] [NEWLINE] These weren't statements from one-off individuals. They were prominent and visible members of the feminist movement. [NEWLINE] [NEWLINE] So if you can forgive feminism it's idiots retrospectively, why do you find it hard to forgive men's rights their idiots concurrently?</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That's the conversation we should be having, not about<mask> awful and sexist this guy is for his shirt. [ENDQ] [NEWLINE] Aside from a few Twitter crazies, that's exactly the conversation that people were trying to have! The reaction to the shirt was, "Whoa, that's not okay; wearing a shirt that features naked women in bondage is not something that's part of a work environment welcoming to men and women.<mask> this company would let an employee *go on international media wearing that shirt,* then<mask> sort of corporate culture are they creating the rest of the time?" And the response to them was, "<mask> are you bullying this man for bringing us science and space? You're weak and ridiculous<mask> you feel that this shirt could be part of a culture that discourages women's participation." [NEWLINE] [NEWLINE] <mask> a lot of women would feel uncomfortable around that shirt in a professional environment. *I* would feel uncomfortable around that shirt,<mask> my boss or coworker were wearing it<mask> at work.<mask> are those feelings of discomfort invalid? At<mask> point would you, personally, believe that I'm allowed to feel uncomfortable by workplace attire that aggressively sexualizes women?<mask> the naked women are wearing leashes?<mask> the person wearing the shirt is watching porn?<mask> the women on the shirt are carrying signs that say "one of the primary functions of women is to look sexy"? [NEWLINE] [NEWLINE] The fact that shirt exists is not sexist. The fact the shirt was designed by a woman has nothing whatsoever to do with whether it is sexist or not. And the issue isn't whether anyone ever wanted to oppress women. The issue is raising awareness of<mask> certain actions work environments that unintentionally make women (or other groups) feel unwelcome.</s>
Label encoding: <s> [STARTQ] That's the conversation we should be having, not about how awful and sexist this guy is for his shirt. [ENDQ] [NEWLINE] Aside from a few Twitter crazies, that's exactly the conversation that people were trying to have! The reaction to the shirt was, "Whoa, that's not okay; wearing a shirt that features naked women in bondage is not something that's part of a work environment welcoming to men and women. If this company would let an employee *go on international media wearing that shirt,* then what sort of corporate culture are they creating the rest of the time?" And the response to them was, " Why are you bullying this man for bringing us science and space? You're weak and ridiculous if you feel that this shirt could be part of a culture that discourages women's participation." [NEWLINE] [NEWLINE] But a lot of women would feel uncomfortable around that shirt in a professional environment. *I* would feel uncomfortable around that shirt, if my boss or coworker were wearing it while at work. Why are those feelings of discomfort invalid? At what point would you, personally, believe that I'm allowed to feel uncomfortable by workplace attire that aggressively sexualizes women? When the naked women are wearing leashes? When the person wearing the shirt is watching porn? When the women on the shirt are carrying signs that say "one of the primary functions of women is to look sexy"? [NEWLINE] [NEWLINE] The fact that shirt exists is not sexist. The fact the shirt was designed by a woman has nothing whatsoever to do with whether it is sexist or not. And the issue isn't whether anyone ever wanted to oppress women. The issue is raising awareness of why certain actions work environments that unintentionally make women (or other groups) feel unwelcome.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] To me, it's the expected outcome of the action. That's<mask> determines morality, and<mask> determines<mask> I should do. [ENDQ] [NEWLINE] Then<mask> is your stance on ignorant negligence? Namely, is someone held accountable<mask> they speed in a particular neighborhood unaware of the speed limit and end up hitting a person? In situations<mask> people are not making informed decisions they cannot make the best decisions. [NEWLINE] [NEWLINE] [STARTQ] To me, "morality" and "<mask> I should do" are synonyms. They seem like they aren't, to you,<mask><mask> it's<mask> easy to come up with situations<mask> the thing you think you should do is immoral by your standards. [ENDQ] [NEWLINE] Correct, they are not the same for me. "Ethically good" decisions are not, to me, synonymous with "the best option available". [NEWLINE] [NEWLINE] [STARTQ] Then really think about whether it's better to continue using whatever moral system it is you're using now, or switch to a consequentialist one. I promise consequentialism is great! We have cookies, and mandatory polio vaccines! [ENDQ] [NEWLINE] I love talking about these things<mask> I do enjoy probing my own belief systems. I'd say that my view on morality is probably very similar to consequentialism<mask>, clearly, I do not uphold ethical "ideals" for their own sake<mask><mask> of its effect on persons around me. [NEWLINE] [NEWLINE] In the Gestapo example, robbing a person of an informed decision is wrong.<mask>, not allowing him the great possibility of harming the people hiding in my house is far more important due to its greater impact on a person. [NEWLINE] [NEWLINE] I'm not sure<mask> that's called consequentialism,<mask> ultimately<mask><mask> the only difference my be semantic wording.</s>
Label encoding: <s> [STARTQ] To me, it's the expected outcome of the action. That's what determines morality, and what determines what I should do. [ENDQ] [NEWLINE] Then what is your stance on ignorant negligence? Namely, is someone held accountable when they speed in a particular neighborhood unaware of the speed limit and end up hitting a person? In situations where people are not making informed decisions they cannot make the best decisions. [NEWLINE] [NEWLINE] [STARTQ] To me, "morality" and " what I should do" are synonyms. They seem like they aren't, to you, given that it's so easy to come up with situations where the thing you think you should do is immoral by your standards. [ENDQ] [NEWLINE] Correct, they are not the same for me. "Ethically good" decisions are not, to me, synonymous with "the best option available". [NEWLINE] [NEWLINE] [STARTQ] Then really think about whether it's better to continue using whatever moral system it is you're using now, or switch to a consequentialist one. I promise consequentialism is great! We have cookies, and mandatory polio vaccines! [ENDQ] [NEWLINE] I love talking about these things because I do enjoy probing my own belief systems. I'd say that my view on morality is probably very similar to consequentialism as, clearly, I do not uphold ethical "ideals" for their own sake but because of its effect on persons around me. [NEWLINE] [NEWLINE] In the Gestapo example, robbing a person of an informed decision is wrong. However, not allowing him the great possibility of harming the people hiding in my house is far more important due to its greater impact on a person. [NEWLINE] [NEWLINE] I'm not sure if that's called consequentialism, but ultimately I think the only difference my be semantic wording.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The weight of your pack varies constantly depending on the conditions and length of time you're out. 80lbs isn't pushing it for the kids I'm thinking of; you build up to it. [NEWLINE] [NEWLINE] I mean I specifically said the hard line would be exactly<mask> you said a woman likely could not do. Like, we used identical examples. At NO POINT in any of my posts here did I advocate for lower standards anywhere. In a life or death career like the military or fire brigade or whatever that would be a crime against humanity. [NEWLINE] [NEWLINE] <mask> there's far more than 1/20 women capable of carrying 150 lbs on a trail run.<mask><mask> I would say the hardest part would be convincing them they could,<mask> the human (yes even female) body was designed to handle that particular strain extremely well. [NEWLINE] [NEWLINE] Women aren't a separate, hollow boned avian species. Note the lack of feathers and more developed brains. :P You are right that men are physically much stronger than women and much better at improving upon that quickly,<mask> there are loads of things women could do in the military and you're not giving them nearly enough credit. [NEWLINE] [NEWLINE] And actually snipers usually work in teams of two in modern warfare. They don't go out in platoons unless they're sent with a guard. You might have to carry another person out of the shit,<mask> I do think some women are capable of that. Put them in teams of similar weight class and you would see less issues. Another 150 lbs for half a mile with adrenaline and the fear of death pumping through you would be shockingly light.<mask> you are damn right they should be able to do it and prove it before they're given the job.</s>
Label encoding: <s>The weight of your pack varies constantly depending on the conditions and length of time you're out. 80lbs isn't pushing it for the kids I'm thinking of; you build up to it. [NEWLINE] [NEWLINE] I mean I specifically said the hard line would be exactly what you said a woman likely could not do. Like, we used identical examples. At NO POINT in any of my posts here did I advocate for lower standards anywhere. In a life or death career like the military or fire brigade or whatever that would be a crime against humanity. [NEWLINE] [NEWLINE] But there's far more than 1/20 women capable of carrying 150 lbs on a trail run. In fact I would say the hardest part would be convincing them they could, but the human (yes even female) body was designed to handle that particular strain extremely well. [NEWLINE] [NEWLINE] Women aren't a separate, hollow boned avian species. Note the lack of feathers and more developed brains. :P You are right that men are physically much stronger than women and much better at improving upon that quickly, but there are loads of things women could do in the military and you're not giving them nearly enough credit. [NEWLINE] [NEWLINE] And actually snipers usually work in teams of two in modern warfare. They don't go out in platoons unless they're sent with a guard. You might have to carry another person out of the shit, but I do think some women are capable of that. Put them in teams of similar weight class and you would see less issues. Another 150 lbs for half a mile with adrenaline and the fear of death pumping through you would be shockingly light. But you are damn right they should be able to do it and prove it before they're given the job.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask>, that bit has mostly to do with the government manipulating language. [ENDQ] [NEWLINE] No it doesn't.  He spends five minutes highlighting euphemisms before identifying them<mask> politically correct language.  He then launches into PC language specifically about race and references said language in the context of sexuality and physicality. [NEWLINE] [NEWLINE] <mask><mask>, by time *and* number of examples that bit has more to do with political correctness outside of politics than political correctness within it. [NEWLINE] [NEWLINE] [Here]( [URL] ;v=Pc0ZHsoHAlE#t=631) is a link to the full recording of the event set to the start of his bit on political correctness. [NEWLINE] [NEWLINE] [STARTQ] PTSD really is a better term than "shell shock"<mask> it is psychologically the same thing that happens to people in any traumatic situation [ENDQ] [NEWLINE] Carlin's point is that "post-traumatic stress disorder" conveys neither the raw, visceral nature nor the gravity of the problem - that's<mask> he mentions Vietnam veterans in conjunction with that portion of the bit.  The sentiment expresses his problem with political correctness<mask> a whole. [NEWLINE] [NEWLINE] [STARTQ] George Carlin would never respect you for simply taking his viewpoints and repeating them<mask><mask> they were axioms. [ENDQ] [NEWLINE] Carlin likely wouldn't respect someone using his name<mask> an appeal to authority - he was very much anti-authoritarian. [NEWLINE] [NEWLINE] [STARTQ] do you see<mask> I took the spirit of<mask> Carlin was saying and applied it to my life and this situation and even used it to argue against his own points? [ENDQ] [NEWLINE] You didn't.  Your viewpoints are diametrically opposed.  You advocate for whitewashing language.  He campaigned against that.</s>
Label encoding: <s> [STARTQ] In fact, that bit has mostly to do with the government manipulating language. [ENDQ] [NEWLINE] No it doesn't.  He spends five minutes highlighting euphemisms before identifying them as politically correct language.  He then launches into PC language specifically about race and references said language in the context of sexuality and physicality. [NEWLINE] [NEWLINE] In fact, by time *and* number of examples that bit has more to do with political correctness outside of politics than political correctness within it. [NEWLINE] [NEWLINE] [Here]( [URL] ;v=Pc0ZHsoHAlE#t=631) is a link to the full recording of the event set to the start of his bit on political correctness. [NEWLINE] [NEWLINE] [STARTQ] PTSD really is a better term than "shell shock" because it is psychologically the same thing that happens to people in any traumatic situation [ENDQ] [NEWLINE] Carlin's point is that "post-traumatic stress disorder" conveys neither the raw, visceral nature nor the gravity of the problem - that's why he mentions Vietnam veterans in conjunction with that portion of the bit.  The sentiment expresses his problem with political correctness as a whole. [NEWLINE] [NEWLINE] [STARTQ] George Carlin would never respect you for simply taking his viewpoints and repeating them as if they were axioms. [ENDQ] [NEWLINE] Carlin likely wouldn't respect someone using his name as an appeal to authority - he was very much anti-authoritarian. [NEWLINE] [NEWLINE] [STARTQ] do you see how I took the spirit of what Carlin was saying and applied it to my life and this situation and even used it to argue against his own points? [ENDQ] [NEWLINE] You didn't.  Your viewpoints are diametrically opposed.  You advocate for whitewashing language.  He campaigned against that.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why would you want this view to be changed? [NEWLINE] [NEWLINE] No one is obligated to like anyone else.  No one is entitled to your body. [NEWLINE] [NEWLINE] <mask> you are forced to be romantic or sexual partners against your will - that would be abuse at best, rape at worst. [NEWLINE] [NEWLINE] <mask> you are coerced into a relationship via shame (you didn't date me<mask> I am &lt;X [STARTQ],  you are &lt;X&gt; phobic). That would be a form of emotional abuse. [ENDQ] [NEWLINE] Now<mask> you are seeing ads or posts that say "fat is beautiful",  understand that these are simply  things with a twofold goal: [NEWLINE] 1) to make people think about their preferences and take a second look, rather than dismissing people out of hand due to weight<mask> the media says "skinny is the only beautiful" [NEWLINE] 2)<mask> a self esteem boost for people who are fat.  to let them know that there are folks out there that will find them beautiful. [NEWLINE] [NEWLINE] That's a pretty positive and innocuous goal,<mask><mask>. [NEWLINE] [NEWLINE] [NEWLINE] Now,<mask><mask><mask><mask>.    <mask> you've just shot someone down, and they want to know<mask>.   There's no winning answer to this. [NEWLINE] [NEWLINE] They are disappointed that they are turned down, <mask> they are insecure about being &lt;fat|trans|tall|skinny|short&gt;, it would not be unsurprising that they lash out at you<mask> you cite that<mask> the reason.    It's better to be tactful and say "You're nice<mask> I'm just not feeling a spark" and try to extract yourself from the situation.</s>
Label encoding: <s>Why would you want this view to be changed? [NEWLINE] [NEWLINE] No one is obligated to like anyone else.  No one is entitled to your body. [NEWLINE] [NEWLINE] If you are forced to be romantic or sexual partners against your will - that would be abuse at best, rape at worst. [NEWLINE] [NEWLINE] If you are coerced into a relationship via shame (you didn't date me because I am &lt;X [STARTQ],  you are &lt;X&gt; phobic). That would be a form of emotional abuse. [ENDQ] [NEWLINE] Now If you are seeing ads or posts that say "fat is beautiful",  understand that these are simply  things with a twofold goal: [NEWLINE] 1) to make people think about their preferences and take a second look, rather than dismissing people out of hand due to weight because the media says "skinny is the only beautiful" [NEWLINE] 2) as a self esteem boost for people who are fat.  to let them know that there are folks out there that will find them beautiful. [NEWLINE] [NEWLINE] That's a pretty positive and innocuous goal, I think. [NEWLINE] [NEWLINE] [NEWLINE] Now, on the other hand.     If you've just shot someone down, and they want to know why.   There's no winning answer to this. [NEWLINE] [NEWLINE] They are disappointed that they are turned down,  if they are insecure about being &lt;fat|trans|tall|skinny|short&gt;, it would not be unsurprising that they lash out at you if you cite that as the reason.    It's better to be tactful and say "You're nice but I'm just not feeling a spark" and try to extract yourself from the situation.</s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Food companies are businesses that sell products that people buy.<mask> people keep buying a product that's not good for them companies will keep selling those products.<mask> people stop buying a certain product<mask> they don't like it companies will change it, case in point, Kraft Dinner who is removing artificial colouring from their products after an online petition gained a lot of signatures. Another example is the gluten free movement. Companies started making gluten free products and labeling products that don't contain gluten in the list place<mask> gluten free<mask> that's<mask> people want. [NEWLINE] [NEWLINE] People blame food companies for lots of things such<mask> not using a standardized service size and claim that that is deceptive.<mask><mask><mask> that it is more convenient to compare products<mask> they had the same serving size,<mask> it is not necessary. You only need [elemental-school-level math]( [URL] ) to figure it out. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Food companies are businesses that sell products that people buy. If people keep buying a product that's not good for them companies will keep selling those products. If people stop buying a certain product because they don't like it companies will change it, case in point, Kraft Dinner who is removing artificial colouring from their products after an online petition gained a lot of signatures. Another example is the gluten free movement. Companies started making gluten free products and labeling products that don't contain gluten in the list place as gluten free because that's what people want. [NEWLINE] [NEWLINE] People blame food companies for lots of things such as not using a standardized service size and claim that that is deceptive. While I agree that it is more convenient to compare products if they had the same serving size, but it is not necessary. You only need [elemental-school-level math]( [URL] ) to figure it out. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your "own beliefs" are a result of growing up in a society that has taught you these beliefs [ENDQ] [NEWLINE] Yes, they are a start -<mask> we have the ability to think critically and alter our positions. That's<mask> ethics evolves - people reject and alter the positions commonly accepted by society. In my country, homosexuality was banned until 1986. This year Marriage Equality passed. The older generation didn't like it -<mask> that didn't matter,<mask> more people had rejected their view<mask> immoral. [NEWLINE] [NEWLINE] [STARTQ] <mask> you are "deriving your morals from reason", you are simply extending your pre-existing moral teachings to different situations [ENDQ] [NEWLINE] Or you are testing your pre-existing moral teachings to different situations, to test their validity. To<mask> extent your morality comes from reason or from<mask> you've been taught is blurry at best -<mask> to simply reduce ethics to "society taught you this" is ridiculous. [NEWLINE] [NEWLINE] [STARTQ] <mask> you're the only one following these morals then you're just putting a set of restrictions on yourself for nothing [ENDQ] [NEWLINE] Following your own morality is its own reward - Think of the Germans who hid Jews during WWII, were they doing it to keep themselves self? For some indirect self-interested reason? No. They did it<mask> they believed it was the right thing to do. [NEWLINE] [NEWLINE] [STARTQ] Murder is "wrong"<mask><mask> it is "right" then you are at risk of being murdered [ENDQ] [NEWLINE] Then<mask> would I decry murder occurring in Africa,<mask> it clearly has no impact on me? There is more to morality than self-interest. The German example illustrates this well - there was no self-interest for them. They did it<mask> they believed it was right.</s>
Label encoding: <s> [STARTQ] Your "own beliefs" are a result of growing up in a society that has taught you these beliefs [ENDQ] [NEWLINE] Yes, they are a start - but we have the ability to think critically and alter our positions. That's how ethics evolves - people reject and alter the positions commonly accepted by society. In my country, homosexuality was banned until 1986. This year Marriage Equality passed. The older generation didn't like it - but that didn't matter, because more people had rejected their view as immoral. [NEWLINE] [NEWLINE] [STARTQ] If you are "deriving your morals from reason", you are simply extending your pre-existing moral teachings to different situations [ENDQ] [NEWLINE] Or you are testing your pre-existing moral teachings to different situations, to test their validity. To what extent your morality comes from reason or from what you've been taught is blurry at best - but to simply reduce ethics to "society taught you this" is ridiculous. [NEWLINE] [NEWLINE] [STARTQ] If you're the only one following these morals then you're just putting a set of restrictions on yourself for nothing [ENDQ] [NEWLINE] Following your own morality is its own reward - Think of the Germans who hid Jews during WWII, were they doing it to keep themselves self? For some indirect self-interested reason? No. They did it because they believed it was the right thing to do. [NEWLINE] [NEWLINE] [STARTQ] Murder is "wrong" because if it is "right" then you are at risk of being murdered [ENDQ] [NEWLINE] Then why would I decry murder occurring in Africa, as it clearly has no impact on me? There is more to morality than self-interest. The German example illustrates this well - there was no self-interest for them. They did it because they believed it was right.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've made a common error here. Soldiers do protect your freedom,<mask> not by going to war. The United States has *overwhelming military dominance* over the entire world.<mask> does that mean? It means that it would take several world powers to come close to challenging us and victory would be incredibly unlikely. The only real threats we could run into are Russia and China, and a war with us would be tremendously costly for either of them. [NEWLINE] [NEWLINE] Take a look at this [breakdown of the top 10 military powers]( [URL] ). China's ranking in naval power seems really high,<mask> that's<mask> they own a lot of frigates. Our significant air superiority would seem to make this somewhat irrelevant. [NEWLINE] [NEWLINE] *Going to war* doesn't make us safer.<mask><mask>, in some instances, it might make us less safe. It's pretty easy to<mask><mask> the war in Iraq and the subsequent rise of ISIS was a pretty bad scenario that we ought to have avoided. In this case our military activities abroad probably actually made us less safe at home. They just shot someone associated with ISIS in Boston last weekend and arrested two more. It's bad. [NEWLINE] [NEWLINE] This does not,<mask>, detract from the fact that overwhelming military dominance keeps us safe from foreign invasion. ISIS or Al Qaeda may be able to convince the odd individual to do something horrific,<mask> they don't have the resources to march an army over and take possession of our territory. Anything they do has to be quick and stealthy and<mask> soon<mask> we discover them we take them out. [NEWLINE] [NEWLINE] <mask>?<mask> we have overwhelming military dominance. *That* is<mask>'s meant by the idea that the military protects our freedom. </s>
Label encoding: <s>You've made a common error here. Soldiers do protect your freedom, but not by going to war. The United States has *overwhelming military dominance* over the entire world. What does that mean? It means that it would take several world powers to come close to challenging us and victory would be incredibly unlikely. The only real threats we could run into are Russia and China, and a war with us would be tremendously costly for either of them. [NEWLINE] [NEWLINE] Take a look at this [breakdown of the top 10 military powers]( [URL] ). China's ranking in naval power seems really high, but that's because they own a lot of frigates. Our significant air superiority would seem to make this somewhat irrelevant. [NEWLINE] [NEWLINE] *Going to war* doesn't make us safer. In fact, in some instances, it might make us less safe. It's pretty easy to argue that the war in Iraq and the subsequent rise of ISIS was a pretty bad scenario that we ought to have avoided. In this case our military activities abroad probably actually made us less safe at home. They just shot someone associated with ISIS in Boston last weekend and arrested two more. It's bad. [NEWLINE] [NEWLINE] This does not, however, detract from the fact that overwhelming military dominance keeps us safe from foreign invasion. ISIS or Al Qaeda may be able to convince the odd individual to do something horrific, but they don't have the resources to march an army over and take possession of our territory. Anything they do has to be quick and stealthy and as soon as we discover them we take them out. [NEWLINE] [NEWLINE] Why? Because we have overwhelming military dominance. *That* is what's meant by the idea that the military protects our freedom. </s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason we created laws for working conditions is that we realize that it's unethical to force people to work in dangerous conditions for overlong hours for low compensation. We see<mask> bad child labor is.<mask> should only our citizens get these rights<mask> we can help it. It seems to me that the only reason we don't do this is that our things will cost more.<mask> we can't put a price on people's lives and wellbeing. [NEWLINE] [NEWLINE] The subject's complicated,<mask> I believe it would incentivize creating American jobs.<mask> there would be the same standards, costs would be similar, and America with the lower transportation costs would seem like a better option. We will<mask> make more money off the taxes. Money isn't the main reason,<mask> it is a benefit. [NEWLINE] [NEWLINE] Edit: I guess I more meant that the American companies should pay for the upgrades [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>The reason we created laws for working conditions is that we realize that it's unethical to force people to work in dangerous conditions for overlong hours for low compensation. We see how bad child labor is. Why should only our citizens get these rights if we can help it. It seems to me that the only reason we don't do this is that our things will cost more. But we can't put a price on people's lives and wellbeing. [NEWLINE] [NEWLINE] The subject's complicated, but I believe it would incentivize creating American jobs. Because there would be the same standards, costs would be similar, and America with the lower transportation costs would seem like a better option. We will also make more money off the taxes. Money isn't the main reason, but it is a benefit. [NEWLINE] [NEWLINE] Edit: I guess I more meant that the American companies should pay for the upgrades [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a few layers involved. We all have a defined genetic sex, which may or may not match the gendered way out brains develop which together with cultural influences guides our gender performance - the ways in which we indicate our gender to others. These layers of gender identity can all match up or they can differ. [NEWLINE] [NEWLINE] Not everyone has XX or XY cells,<mask> there's a way in which gender can differ from<mask>'s most common right at the base of our biology. Some people<mask> they develop are not capable of responding to hormonal cues to become male and<mask> they outwardly remain female. Some people develop an anatomical gender that is somewhere intermediate between male and female. The way the brain develops affects our perception of our own gender and<mask> it's been observed that the brains of transgender individuals and people whose psychological gender identification is congruent with their biological sex have a number of differences. Finally, there is gender performance which is different from culture to culture and time to time. Some cultures have defined roles, clothing and activities for men and women. Others add additional roles reflecting genders that are in between or apart from male and female. Over time, gender roles and the expression of gender identity have changed. It used to be in European societies that hose, high-heeled shoes and long hair indicated a masculine look. The policing of gender roles changes. Androgyny was more acceptable in the popular culture of many places in the 1980s than it is now. [NEWLINE] [NEWLINE] <mask> I mean to say with all this is that gender is complicated and the idea that genetic sex must match with biological sex must match with a particular kind of gender performance is just wrong and doesn't reflect the entire experience of human history.</s>
Label encoding: <s>There are a few layers involved. We all have a defined genetic sex, which may or may not match the gendered way out brains develop which together with cultural influences guides our gender performance - the ways in which we indicate our gender to others. These layers of gender identity can all match up or they can differ. [NEWLINE] [NEWLINE] Not everyone has XX or XY cells, so there's a way in which gender can differ from what's most common right at the base of our biology. Some people as they develop are not capable of responding to hormonal cues to become male and so they outwardly remain female. Some people develop an anatomical gender that is somewhere intermediate between male and female. The way the brain develops affects our perception of our own gender and so it's been observed that the brains of transgender individuals and people whose psychological gender identification is congruent with their biological sex have a number of differences. Finally, there is gender performance which is different from culture to culture and time to time. Some cultures have defined roles, clothing and activities for men and women. Others add additional roles reflecting genders that are in between or apart from male and female. Over time, gender roles and the expression of gender identity have changed. It used to be in European societies that hose, high-heeled shoes and long hair indicated a masculine look. The policing of gender roles changes. Androgyny was more acceptable in the popular culture of many places in the 1980s than it is now. [NEWLINE] [NEWLINE] What I mean to say with all this is that gender is complicated and the idea that genetic sex must match with biological sex must match with a particular kind of gender performance is just wrong and doesn't reflect the entire experience of human history.</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This to me relates well to the concept of anonymity. Something that an under-moderated forum is has<mask> its strong suite<mask> over-moderated forums have responsibility<mask> their strong suite. [NEWLINE] [NEWLINE] Whether or not there is bias or a lack of interesting material still exists in either category. It is very subjective on these measures<mask> to which type of moderation comes out on top. [NEWLINE] [NEWLINE] <mask> sticking to<mask> I broke under/over to be about..anonymity/responsibility which is better? Which is more relevant and which see's the best gains in knowledge? [NEWLINE] [NEWLINE] An anon forum takes a lot of effort to come across gold. Not saying they don't get there,<mask> the output of shit to find the diamonds is massive and generally creates a lot of wasted energy. [NEWLINE] [NEWLINE] A responsible forum requires people to have strong conviction<mask> they wish to push forward an agenda of some type. Which is much closer to reality in my mind. [NEWLINE] [NEWLINE] I do not agree with anon behavior,<mask><mask> it brings out a very dark side of humanity and it's not even a particularly relevant part most of the time. [NEWLINE] [NEWLINE] To cast this into a different forum style, consider it applied politics.<mask><mask> you government was under-moderated or over-moderated? One is certainly more entertaining, and open to radical influence the other is potentially a lot more transparent<mask> decidedly slower to act. [NEWLINE] [NEWLINE] I'd like to point out that anon is a very recent phenomena, dating back to urban development. Prior to which we lived in communities and couldn't afford to flame our neighbors<mask> they'd likely use real flames against us in greater force. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>This to me relates well to the concept of anonymity. Something that an under-moderated forum is has as its strong suite while over-moderated forums have responsibility as their strong suite. [NEWLINE] [NEWLINE] Whether or not there is bias or a lack of interesting material still exists in either category. It is very subjective on these measures as to which type of moderation comes out on top. [NEWLINE] [NEWLINE] So sticking to what I broke under/over to be about..anonymity/responsibility which is better? Which is more relevant and which see's the best gains in knowledge? [NEWLINE] [NEWLINE] An anon forum takes a lot of effort to come across gold. Not saying they don't get there, but the output of shit to find the diamonds is massive and generally creates a lot of wasted energy. [NEWLINE] [NEWLINE] A responsible forum requires people to have strong conviction if they wish to push forward an agenda of some type. Which is much closer to reality in my mind. [NEWLINE] [NEWLINE] I do not agree with anon behavior, I think it brings out a very dark side of humanity and it's not even a particularly relevant part most of the time. [NEWLINE] [NEWLINE] To cast this into a different forum style, consider it applied politics. What if you government was under-moderated or over-moderated? One is certainly more entertaining, and open to radical influence the other is potentially a lot more transparent but decidedly slower to act. [NEWLINE] [NEWLINE] I'd like to point out that anon is a very recent phenomena, dating back to urban development. Prior to which we lived in communities and couldn't afford to flame our neighbors because they'd likely use real flames against us in greater force. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, first off there is [a long history of human cultures dealing with more than 2 genders]( [URL] #History),<mask> that part of your argument is just factually incorrect. <mask>,<mask><mask> approaching this from a standpoint of looking for evidence to prove a claim is misguided.  Trans people are not trying to pretend to be another gender and pull the wool over everyone else's eyes;<mask><mask>, in the suicide rate among trans people is unbelievably high compared to those conforming to the current male/female standard and even to LGBT people.  They don't wish they were trans, they wish they were born in the body they feel their minds were meant for, and that dysmorphia is incredibly hard to deal with. [NEWLINE] [NEWLINE] Beyond all that, there *is* [mounting scientific evidence]( [URL] #.VW0e0M9Viko) that<mask> causes people to be trans is not some desire for attention or to be different,<mask> actually represented by physiological differences in the brain from non-trans people. [NEWLINE] [NEWLINE] Look, I used to think about it like you do,<mask> I don't look down on you for being ill-informed.  Quite the contrary; I applaud your desire to test the opinions you have and see<mask> they hold true.  In my case, after talking to some trans people online and researching the matter, I've come to a place<mask> I don't feel it really even my place to tell someone<mask> they can or cannot identify<mask>,<mask><mask><mask> it's genuine.  There's no good reason to mock someone whom we may perceive<mask> weird, and no good reason to try and convince them to be not weird to us.</s>
Label encoding: <s>Well, first off there is [a long history of human cultures dealing with more than 2 genders]( [URL] #History), so that part of your argument is just factually incorrect.  Secondly, I think approaching this from a standpoint of looking for evidence to prove a claim is misguided.  Trans people are not trying to pretend to be another gender and pull the wool over everyone else's eyes; in fact, in the suicide rate among trans people is unbelievably high compared to those conforming to the current male/female standard and even to LGBT people.  They don't wish they were trans, they wish they were born in the body they feel their minds were meant for, and that dysmorphia is incredibly hard to deal with. [NEWLINE] [NEWLINE] Beyond all that, there *is* [mounting scientific evidence]( [URL] #.VW0e0M9Viko) that what causes people to be trans is not some desire for attention or to be different, but actually represented by physiological differences in the brain from non-trans people. [NEWLINE] [NEWLINE] Look, I used to think about it like you do, so I don't look down on you for being ill-informed.  Quite the contrary; I applaud your desire to test the opinions you have and see if they hold true.  In my case, after talking to some trans people online and researching the matter, I've come to a place where I don't feel it really even my place to tell someone what they can or cannot identify as, so long as it's genuine.  There's no good reason to mock someone whom we may perceive as weird, and no good reason to try and convince them to be not weird to us.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I posted was a sociological article.<mask> such it uses some terminology in a context which is unique to the field. For example, a traditional man is somebody who is more likely to hurt people<mask> a traditional man is aggressive and domineering. This is not all men by ANY means whatsoever,<mask> it is the image of masculinity that society pressures men to live up to. Traditional masculinity is hegemonic masculinity, a search term that might get you some useful articles to help in understanding the concept. [NEWLINE] [NEWLINE] A lot of the other issues you take umbrage with are,<mask><mask>, a result of a poor reading of<mask> is being said. For instance, that first passage you reference is not saying men are more likely to be violent<mask> they are more likely to be violent. It is saying that threats of violence are more credible coming from men<mask> of the socializing institutions which imagine the ideal man<mask> aggressive and the ideal woman<mask> submissive.<mask> a man threatens a woman, we view that<mask> a serious threat<mask> we have an image of<mask> men ought to be.<mask> a woman threatens a man, we don't take that threat seriously<mask> that contrasts with traditional feminine gender roles. Threats of violence from men are more credible and serious<mask> we expect them to be more credible and serious. That's all. [NEWLINE] [NEWLINE] I grant you that this article takes a lot of things for granted, and<mask> you have a problem with<mask> this article defines traditional gender roles there is a wealth of academic literature of that subject. I suppose I shouldn't have assumed that a sociological article in a nonsociological subreddit would have been read from a sociological lens,<mask>.</s>
Label encoding: <s>What I posted was a sociological article. As such it uses some terminology in a context which is unique to the field. For example, a traditional man is somebody who is more likely to hurt people because a traditional man is aggressive and domineering. This is not all men by ANY means whatsoever, but it is the image of masculinity that society pressures men to live up to. Traditional masculinity is hegemonic masculinity, a search term that might get you some useful articles to help in understanding the concept. [NEWLINE] [NEWLINE] A lot of the other issues you take umbrage with are, I think, a result of a poor reading of what is being said. For instance, that first passage you reference is not saying men are more likely to be violent because they are more likely to be violent. It is saying that threats of violence are more credible coming from men because of the socializing institutions which imagine the ideal man as aggressive and the ideal woman as submissive. When a man threatens a woman, we view that as a serious threat because we have an image of how men ought to be. When a woman threatens a man, we don't take that threat seriously because that contrasts with traditional feminine gender roles. Threats of violence from men are more credible and serious because we expect them to be more credible and serious. That's all. [NEWLINE] [NEWLINE] I grant you that this article takes a lot of things for granted, and if you have a problem with how this article defines traditional gender roles there is a wealth of academic literature of that subject. I suppose I shouldn't have assumed that a sociological article in a nonsociological subreddit would have been read from a sociological lens, though.</s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. English,<mask><mask><mask> any scholar can tell, is changing just<mask> it has been<mask> its wee beginnings in Pre-Proto-Indo-European thousands and thousands of years ago. There is certainly evidence that it is still changing. [NEWLINE] [NEWLINE] <mask><mask> we need to separate "written English" and "the English language" in order to deal with your argument. "written English" is a human invention: it is constructed, taught, guided, and reliant on materials and resources. It's a result of writing's usefulness<mask> a method of transferring ideas, and it will continue to function<mask><mask><mask> it's practical. [NEWLINE] [NEWLINE] 2. The goal of language is communication, yes. Language change is a universal, almost-constant process. English will probably inevitably split into a variety of mutually unintelligible new languages, just<mask> Latin, Chinese, Arabic, or any other language in history did. It is not a bad thing, and does not significantly impede the process of societal and scientific change. Consistency in language is not a result of grammar nazis' collective efforts,<mask> rather a result of the necessity of communication. [NEWLINE] [NEWLINE] The video is a further example of the spoken vs. written phenomenon. Spoken language is<mask> it is and will probably always be,<mask><mask><mask> we know: a changing thing that diverges and becomes mutually unintelligible with distance (social, economic, or geographic) and time. The almost unintelligibility of the person in the video is a result of hundreds of years of separation, and I expect in the next couple hundred years "the English language" will be<mask> much an arbitrary political construct<mask> Arabic.</s>
Label encoding: <s>1. English, as far as any scholar can tell, is changing just as it has been since its wee beginnings in Pre-Proto-Indo-European thousands and thousands of years ago. There is certainly evidence that it is still changing. [NEWLINE] [NEWLINE] I think we need to separate "written English" and "the English language" in order to deal with your argument. "written English" is a human invention: it is constructed, taught, guided, and reliant on materials and resources. It's a result of writing's usefulness as a method of transferring ideas, and it will continue to function as long as it's practical. [NEWLINE] [NEWLINE] 2. The goal of language is communication, yes. Language change is a universal, almost-constant process. English will probably inevitably split into a variety of mutually unintelligible new languages, just as Latin, Chinese, Arabic, or any other language in history did. It is not a bad thing, and does not significantly impede the process of societal and scientific change. Consistency in language is not a result of grammar nazis' collective efforts, but rather a result of the necessity of communication. [NEWLINE] [NEWLINE] The video is a further example of the spoken vs. written phenomenon. Spoken language is what it is and will probably always be, as far as we know: a changing thing that diverges and becomes mutually unintelligible with distance (social, economic, or geographic) and time. The almost unintelligibility of the person in the video is a result of hundreds of years of separation, and I expect in the next couple hundred years "the English language" will be as much an arbitrary political construct as Arabic.</s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In practice, that isn't<mask> happens.  Of course they are going to stop more Arabs than Israelis on average simply<mask> Arabs are more likely to fit suspicious criteria (they're probably more nervous in that situation, more guarded, etc.). [NEWLINE] [NEWLINE] More importantly,<mask> you remove the same practice from this context and place it in another, the problems you see don't arise.  It is more than likely that an anti-Israeli terrorist would be an Arab,<mask> you can't lay down a similar hard and fast rule concerning an American criminal.  We justify racial profiling by saying that a person with a specific characteristic is individually more likely to have committed a crime,<mask> crime is not primarily committed by those with that characteristic. <mask> you account for poverty, questionable laws (drug policy) and the fact that increased scrutiny of any group will uncover more crime, you're essentially enacting a self-fulfilling prophecy. [NEWLINE] [NEWLINE] <mask> crime-wise, this is a detrimental policy. [NEWLINE] [NEWLINE] Terrorism-wise, others have pointed out that most acts of terrorism have not been committed by Muslims.  Furthermore, there are enough Muslim terrorists that sending a Saudi or Syrian or Iranian would just be a bad idea. <mask> I were the bad guys, I'd try to enlist a white American convert.  Failing that, I'd get a Chechen (there are some that look a lot more European than Tsarnaev) a Muslim from the former Yugoslavia, a European convert, an Asian convert... [NEWLINE] [NEWLINE] There are<mask> many ways to subvert racial profiling that it becomes detrimental to security.  Especially<mask> you're turning a blind eye to those least likely to be detected.</s>
Label encoding: <s>In practice, that isn't what happens.  Of course they are going to stop more Arabs than Israelis on average simply because Arabs are more likely to fit suspicious criteria (they're probably more nervous in that situation, more guarded, etc.). [NEWLINE] [NEWLINE] More importantly, if you remove the same practice from this context and place it in another, the problems you see don't arise.  It is more than likely that an anti-Israeli terrorist would be an Arab, but you can't lay down a similar hard and fast rule concerning an American criminal.  We justify racial profiling by saying that a person with a specific characteristic is individually more likely to have committed a crime, though crime is not primarily committed by those with that characteristic.  When you account for poverty, questionable laws (drug policy) and the fact that increased scrutiny of any group will uncover more crime, you're essentially enacting a self-fulfilling prophecy. [NEWLINE] [NEWLINE] So crime-wise, this is a detrimental policy. [NEWLINE] [NEWLINE] Terrorism-wise, others have pointed out that most acts of terrorism have not been committed by Muslims.  Furthermore, there are enough Muslim terrorists that sending a Saudi or Syrian or Iranian would just be a bad idea.  If I were the bad guys, I'd try to enlist a white American convert.  Failing that, I'd get a Chechen (there are some that look a lot more European than Tsarnaev) a Muslim from the former Yugoslavia, a European convert, an Asian convert... [NEWLINE] [NEWLINE] There are so many ways to subvert racial profiling that it becomes detrimental to security.  Especially if you're turning a blind eye to those least likely to be detected.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>He issued a declaration freeing the chattel slaves held in those territories currently in rebellion against the US.  He<mask> oversaw the middle of the ratification process for the 13th Amendment which did outlaw (most) slavery in the US ("except<mask> a punishment for crime whereof the party shall have been duly convicted"). [NEWLINE] [NEWLINE] <mask><mask> it was expedient to do<mask> he was fine with slavery.  The Emancipation Proclamation only applied to states in rebellion, states that stayed in the Union<mask> had slaves were allowed to keep them - and<mask> he had no real legal authority over the states currently in rebellion it hardly mattered<mask> he said about the law there. [NEWLINE] [NEWLINE] And I'm not sure he deserves all the credit for the 13th Amendment, especially<mask> he,<mask> President, had no role in voting on it or ratifying it.  He didn't draft it or propose it either.  He did, eventually, endorse the amendment<mask> it took a threatened 3rd party run at the White House on an anti-slavery platform to push him to do<mask> (and once reelected made it a major part of his legislative agenda - after the Senate had already proposed and passed it). [NEWLINE] [NEWLINE] Not to mention that he was fine with involuntary service - overseeing forced conscription of over 40,000 solders and an additional 120,000 paid substitutes for draftees  (leading to riots) and illegally and unconstitutionally suspending *habeas corpus* to allow that draft to function. [NEWLINE] [NEWLINE] The end of end of chattel slavery in the US began under Lincoln,<mask> he was hardly a stalwart champion of freedom and liberty.</s>
Label encoding: <s>He issued a declaration freeing the chattel slaves held in those territories currently in rebellion against the US.  He also oversaw the middle of the ratification process for the 13th Amendment which did outlaw (most) slavery in the US ("except as a punishment for crime whereof the party shall have been duly convicted"). [NEWLINE] [NEWLINE] But when it was expedient to do so he was fine with slavery.  The Emancipation Proclamation only applied to states in rebellion, states that stayed in the Union but had slaves were allowed to keep them - and since he had no real legal authority over the states currently in rebellion it hardly mattered what he said about the law there. [NEWLINE] [NEWLINE] And I'm not sure he deserves all the credit for the 13th Amendment, especially as he, as President, had no role in voting on it or ratifying it.  He didn't draft it or propose it either.  He did, eventually, endorse the amendment but it took a threatened 3rd party run at the White House on an anti-slavery platform to push him to do so (and once reelected made it a major part of his legislative agenda - after the Senate had already proposed and passed it). [NEWLINE] [NEWLINE] Not to mention that he was fine with involuntary service - overseeing forced conscription of over 40,000 solders and an additional 120,000 paid substitutes for draftees  (leading to riots) and illegally and unconstitutionally suspending *habeas corpus* to allow that draft to function. [NEWLINE] [NEWLINE] The end of end of chattel slavery in the US began under Lincoln, but he was hardly a stalwart champion of freedom and liberty.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're sort of ignoring a really important feature of the gun - it can (in theory) harm or kill people from quite a distance, on short notice.  This makes running away a bad option, and everyone understands this.  With the lighter method, you have to get within arm's reach of the person to make good on the threat.  People will start trying to get away the second you squirt fluid on them, maybe take their clothes off, and generally it's going to be messy.  Shooting someone is a lot quicker and simpler. [NEWLINE] [NEWLINE] Furthermore, most robbers aren't terribly interested in hurting people.  They're willing to do it,<mask> it comes right down to it -<mask> 99/100 times they'd rather not shoot anyone.  The mental cost of lighting another human being on fire to the robber is much higher.  It's a terrible and slow way to die.<mask> they don't have the stomach for shooting someone, they CERTAINLY don't have the stomach to burn someone alive. [NEWLINE] [NEWLINE] Your point about the news spreading<mask> makes it a much less attractive option.  A 7-11 clerk gets shot or has a gun pointed at them?  NBD, you are at an average level of priority for the cops.  Happens all the time. [NEWLINE] [NEWLINE] You threaten to burn someone alive?  You're definitely going to make the news and the cops will have to try a lot harder to catch you.  A normal criminal is one thing, a super-deranged psychopath who might burn you alive is another.  You're making your life much harder, even assuming you carry off the robbery. </s>
Label encoding: <s>You're sort of ignoring a really important feature of the gun - it can (in theory) harm or kill people from quite a distance, on short notice.  This makes running away a bad option, and everyone understands this.  With the lighter method, you have to get within arm's reach of the person to make good on the threat.  People will start trying to get away the second you squirt fluid on them, maybe take their clothes off, and generally it's going to be messy.  Shooting someone is a lot quicker and simpler. [NEWLINE] [NEWLINE] Furthermore, most robbers aren't terribly interested in hurting people.  They're willing to do it, if it comes right down to it - but 99/100 times they'd rather not shoot anyone.  The mental cost of lighting another human being on fire to the robber is much higher.  It's a terrible and slow way to die. If they don't have the stomach for shooting someone, they CERTAINLY don't have the stomach to burn someone alive. [NEWLINE] [NEWLINE] Your point about the news spreading also makes it a much less attractive option.  A 7-11 clerk gets shot or has a gun pointed at them?  NBD, you are at an average level of priority for the cops.  Happens all the time. [NEWLINE] [NEWLINE] You threaten to burn someone alive?  You're definitely going to make the news and the cops will have to try a lot harder to catch you.  A normal criminal is one thing, a super-deranged psychopath who might burn you alive is another.  You're making your life much harder, even assuming you carry off the robbery. </s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"women make 70%<mask> men make" sounds just<mask> misleading<mask> "It's 100% due to the choices women make",<mask> at least one is objectively true<mask> lacking context. The chance that this whole issue can be broken down into one cause is stupid. [NEWLINE] [NEWLINE] American society, at the very least, still has a ways to go before gender is completely equal. We're doing pretty well, and government helps to keep obvious breaches in pay from happening,<mask> that doesn't necessarily mean that women are equally given the choice to be promoted. I feel like the huge dominance of men in government shows just<mask> little confidence America has with women in positions of authority. [NEWLINE] [NEWLINE] And then there's the issue of<mask> careers women choose to go into. You're right, it *is* technically their choice to go into the career paths which pay less.<mask> these choices are probably influenced by our society<mask> well<mask> the individual cultures within each industry. The argument here isn't really over whether women are *choosing*, it's the disparity in<mask> they are treated based on the choices they make. Both men and women can choose to become programmers,<mask> only women have to worry about the reported misogynistic culture of that industry. [NEWLINE] [NEWLINE] Maybe that's not something you think society should have to try and change; whether we can or should try to change it is debatable, especially<mask> we are making arguments about inherent differences in men and women.<mask> don't reduce it to some stupid argument about society having absolutely no influence over individuals' choices, or I'll just have to turn it into a stupider argument about<mask> no one has free choice in any case.</s><pad>
Label encoding: <s>"women make 70% what men make" sounds just as misleading as "It's 100% due to the choices women make", but at least one is objectively true despite lacking context. The chance that this whole issue can be broken down into one cause is stupid. [NEWLINE] [NEWLINE] American society, at the very least, still has a ways to go before gender is completely equal. We're doing pretty well, and government helps to keep obvious breaches in pay from happening, but that doesn't necessarily mean that women are equally given the choice to be promoted. I feel like the huge dominance of men in government shows just how little confidence America has with women in positions of authority. [NEWLINE] [NEWLINE] And then there's the issue of what careers women choose to go into. You're right, it *is* technically their choice to go into the career paths which pay less. But these choices are probably influenced by our society as well as the individual cultures within each industry. The argument here isn't really over whether women are *choosing*, it's the disparity in how they are treated based on the choices they make. Both men and women can choose to become programmers, but only women have to worry about the reported misogynistic culture of that industry. [NEWLINE] [NEWLINE] Maybe that's not something you think society should have to try and change; whether we can or should try to change it is debatable, especially if we are making arguments about inherent differences in men and women. But don't reduce it to some stupid argument about society having absolutely no influence over individuals' choices, or I'll just have to turn it into a stupider argument about how no one has free choice in any case.</s><pad>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe you're discussing two separate issues here. [NEWLINE] [NEWLINE] The first is whether the program you described should be considered "Computer Science". Many school have "Computer Science" programs that are actually "Software/Computer Engineering" programs, or a mix of both.<mask><mask><mask><mask>, classes like algorithms, data structures and formal languages should clearly fall under the "Computer Science" side of things. [NEWLINE] [NEWLINE] The second issue, which<mask><mask> is closer to<mask> you're discussing, is whether "Computer Science" is a science, or a "math"<mask> you describe. I would<mask><mask> your classification is not entirely informative.<mask> much<mask> Math is "axiom-based" (rather than theory+evidence based<mask> in Science), and<mask> much<mask> we have a tendency to link all axiom-based things to math (i.e. CS is a branch of math, logic is a branch of math, etc), we don't often call "axiom-based" things Math, even<mask> we regard it<mask> being derivative of math. [NEWLINE] [NEWLINE] A good comparison is statistics, which is "axiom-based" just like the mathy parts of Computer Science. I don't think we call Statistics "Statistical Mathematics" except<mask> we're emphasizing the mathematical/proof-based parts of Statistics (<mask><mask> we call is "Mathematic Statistics" in those cases).<mask><mask><mask><mask>, we do sometimes see departments labeling the major or department<mask> "Statistical Sciences". This is a close parallel to Computer Science. [NEWLINE] [NEWLINE] <mask> the bone you're picking is less with Computer Science,<mask> more with<mask> we differentiate between Math and Science. </s>
Label encoding: <s>I believe you're discussing two separate issues here. [NEWLINE] [NEWLINE] The first is whether the program you described should be considered "Computer Science". Many school have "Computer Science" programs that are actually "Software/Computer Engineering" programs, or a mix of both. On the other hand, classes like algorithms, data structures and formal languages should clearly fall under the "Computer Science" side of things. [NEWLINE] [NEWLINE] The second issue, which I think is closer to what you're discussing, is whether "Computer Science" is a science, or a "math" as you describe. I would argue that your classification is not entirely informative. As much as Math is "axiom-based" (rather than theory+evidence based as in Science), and as much as we have a tendency to link all axiom-based things to math (i.e. CS is a branch of math, logic is a branch of math, etc), we don't often call "axiom-based" things Math, even if we regard it as being derivative of math. [NEWLINE] [NEWLINE] A good comparison is statistics, which is "axiom-based" just like the mathy parts of Computer Science. I don't think we call Statistics "Statistical Mathematics" except when we're emphasizing the mathematical/proof-based parts of Statistics ( I think we call is "Mathematic Statistics" in those cases). On the other hand, we do sometimes see departments labeling the major or department as "Statistical Sciences". This is a close parallel to Computer Science. [NEWLINE] [NEWLINE] So the bone you're picking is less with Computer Science, but more with how we differentiate between Math and Science. </s>
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> that an accusation proven false still hangs around someone's neck. [ENDQ] [NEWLINE]..and<mask>, an accusation wrongly considered false leaves many predators to keep on molesting victims. [NEWLINE] [NEWLINE] Guess<mask> the statistics say is many hundreds of times more likely? False accusations of sexual misconduct are<mask> low<mask> false accusations of any other type of crime - anywhere from 2% to 8%.<mask> rape and sexual assault are typically THE most underreported, underprosecuted, and underconvicted crimes in almost every country in the world. The chances of an actual guilty rapist seeing any jail time whatsoever is about 3%. [NEWLINE] [NEWLINE] With this in mind, can we continue to be<mask> exclusively concerned about the fate of the accused, to the exclusion of everything else? [NEWLINE] [NEWLINE] There is some seriously Sharia shit going on in this thread. MANY girls over SEVERAL years apparently complained to one another that the doctor was creepy and weird during the exam. And<mask> that commenter and everyone upvoting him and most replies to him dismisses their statements<mask> "girls that age can be<mask> WEIRD"?! [NEWLINE] [NEWLINE] Sure, let's not make very highly publicised accusations against the doctor and let's not ruin his life over<mask> might be nothing. [NEWLINE] [NEWLINE] <mask> a LOT more importantly, let's not dismiss female testimony<mask> automatically invalid and untrustworthy either. Let's launch investigations, be open to the idea that sexual misconduct occurs (<mask> it DOES!), and be vigilant against the stereotype that women always lie. [NEWLINE] [NEWLINE] *There is no reason other than abject misogyny for commenters on this forum to assume those girls were ALL lying, year after year after year.* </s>
Label encoding: <s> [STARTQ] but that an accusation proven false still hangs around someone's neck. [ENDQ] [NEWLINE]..and also, an accusation wrongly considered false leaves many predators to keep on molesting victims. [NEWLINE] [NEWLINE] Guess what the statistics say is many hundreds of times more likely? False accusations of sexual misconduct are as low as false accusations of any other type of crime - anywhere from 2% to 8%. But rape and sexual assault are typically THE most underreported, underprosecuted, and underconvicted crimes in almost every country in the world. The chances of an actual guilty rapist seeing any jail time whatsoever is about 3%. [NEWLINE] [NEWLINE] With this in mind, can we continue to be so exclusively concerned about the fate of the accused, to the exclusion of everything else? [NEWLINE] [NEWLINE] There is some seriously Sharia shit going on in this thread. MANY girls over SEVERAL years apparently complained to one another that the doctor was creepy and weird during the exam. And yet that commenter and everyone upvoting him and most replies to him dismisses their statements as "girls that age can be so WEIRD"?! [NEWLINE] [NEWLINE] Sure, let's not make very highly publicised accusations against the doctor and let's not ruin his life over what might be nothing. [NEWLINE] [NEWLINE] But a LOT more importantly, let's not dismiss female testimony as automatically invalid and untrustworthy either. Let's launch investigations, be open to the idea that sexual misconduct occurs ( because it DOES!), and be vigilant against the stereotype that women always lie. [NEWLINE] [NEWLINE] *There is no reason other than abject misogyny for commenters on this forum to assume those girls were ALL lying, year after year after year.* </s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand "Realty unless noted", I can enjoy Toy Story, the universe set itself with that toys can come to life. Its great its cute, they play with idea. And I am a fan of old school science fiction,<mask> they play with novel ideas and authors try to predict<mask> humans interact with them.<mask> you ever read Roadside Picnic, that is by far one of favorite short stories involving aliens. [NEWLINE] [NEWLINE] I suppose my issue with Batman is<mask> I understand the need and enjoyment for escapist literature, don't get me wrong I love that<mask> well. Its my experience with the fandom and I suppose its love in pop culture, that drives me nuts. He's dark and he's brooding and the guy's a bad ass, it turns me off<mask> people put batman on a pedestal.<mask> its hard for a guy like a me to take it seriously.<mask> I suppose geeks will be geeks, and I should let them be. [NEWLINE] [NEWLINE] Guess I'm jaded. I'm fine with crazy stuff in entertainment,<mask><mask> really enjoy something and I want to share it I want the people to act like people. It seems like in the superhero genre, you have loop holes then gods and more bizarre shit, ect ect. It jumps the shark for me. It seems like lazy writing.<mask> whatever, I suppose Batman or the superhero genre is right there with Entourage, Sex and the City, and 50 shades of Grey. Just without the nerd culture stigma anymore. [NEWLINE] [NEWLINE] Well I still think batman's dumb,<mask> I guess now I can see the appeal. Delta for you. ∆</s>
Label encoding: <s>I understand "Realty unless noted", I can enjoy Toy Story, the universe set itself with that toys can come to life. Its great its cute, they play with idea. And I am a fan of old school science fiction, where they play with novel ideas and authors try to predict how humans interact with them. If you ever read Roadside Picnic, that is by far one of favorite short stories involving aliens. [NEWLINE] [NEWLINE] I suppose my issue with Batman is while I understand the need and enjoyment for escapist literature, don't get me wrong I love that as well. Its my experience with the fandom and I suppose its love in pop culture, that drives me nuts. He's dark and he's brooding and the guy's a bad ass, it turns me off how people put batman on a pedestal. But its hard for a guy like a me to take it seriously. But I suppose geeks will be geeks, and I should let them be. [NEWLINE] [NEWLINE] Guess I'm jaded. I'm fine with crazy stuff in entertainment, but if really enjoy something and I want to share it I want the people to act like people. It seems like in the superhero genre, you have loop holes then gods and more bizarre shit, ect ect. It jumps the shark for me. It seems like lazy writing. But whatever, I suppose Batman or the superhero genre is right there with Entourage, Sex and the City, and 50 shades of Grey. Just without the nerd culture stigma anymore. [NEWLINE] [NEWLINE] Well I still think batman's dumb, but I guess now I can see the appeal. Delta for you. ∆</s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me see<mask> I can address some of his points. [NEWLINE] [NEWLINE] * Space programs can warn us of natural disasters like asteroids, solar flares, etc. [NEWLINE] [NEWLINE] Fair enough.<mask><mask><mask><mask> I'm aware, that's already being done by NASA. [NEWLINE] [NEWLINE] * Without a definite goal in mind for the space program, there's no force operating on "the economic pipeline of America." [NEWLINE] [NEWLINE] Bullshit.<mask> NASA may have the potential to draw kids into science, it's definitely not the only thing that can.<mask>'s needed there is better marketing, not new programs for the sake of inspiring 8-year-old wonder. It's worth noting that very few people want to go into the same career at age 18 that they did at age 8.<mask>,<mask><mask><mask> I'm aware, we don't really have a major deficit of STEM-majors at the moment. [NEWLINE] [NEWLINE] * We should be investing in science for economic reasons and to compete on the world stage. [NEWLINE] [NEWLINE] True, I'm just not convinced that money should go to space exploration. Keep in mind that the countries he noted<mask> drivers of technological innovations (Germany, Switzerland) don't have very big space programs. They just have high-quality educational systems. And the amount of money spent on the LHC over several years is something like half the current annual NASA budget.<mask> it's not like they're in the poor house. [NEWLINE] [NEWLINE] * NASA costs half a penny on the dollar. [NEWLINE] [NEWLINE] This seems like his best argument to me,<mask> it's unclear whether he's referring to revenue earned by NASA or the economic growth it indirectly stimulates.</s>
Label encoding: <s>Let me see if I can address some of his points. [NEWLINE] [NEWLINE] * Space programs can warn us of natural disasters like asteroids, solar flares, etc. [NEWLINE] [NEWLINE] Fair enough. But as far as I'm aware, that's already being done by NASA. [NEWLINE] [NEWLINE] * Without a definite goal in mind for the space program, there's no force operating on "the economic pipeline of America." [NEWLINE] [NEWLINE] Bullshit. While NASA may have the potential to draw kids into science, it's definitely not the only thing that can. What's needed there is better marketing, not new programs for the sake of inspiring 8-year-old wonder. It's worth noting that very few people want to go into the same career at age 18 that they did at age 8. Besides, as far as I'm aware, we don't really have a major deficit of STEM-majors at the moment. [NEWLINE] [NEWLINE] * We should be investing in science for economic reasons and to compete on the world stage. [NEWLINE] [NEWLINE] True, I'm just not convinced that money should go to space exploration. Keep in mind that the countries he noted as drivers of technological innovations (Germany, Switzerland) don't have very big space programs. They just have high-quality educational systems. And the amount of money spent on the LHC over several years is something like half the current annual NASA budget. So it's not like they're in the poor house. [NEWLINE] [NEWLINE] * NASA costs half a penny on the dollar. [NEWLINE] [NEWLINE] This seems like his best argument to me, but it's unclear whether he's referring to revenue earned by NASA or the economic growth it indirectly stimulates.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just a heads up, none of these are nostalgic for me-- I only started playing them a few months ago on the Eshop. [NEWLINE] [NEWLINE] I really want to understand<mask> the hubbub is about for SMW? I don't know<mask> it's just me,<mask> it is not<mask> fun<mask> SMB3, it just seems messy. I want to say that it is probably just nostalgia, people reminiscing about their childhood or something,<mask> I don't see it<mask> the best Mario game ever made (<mask> that award had to go to the retro Marios, I would give it to SMB3 anyway). [NEWLINE] [NEWLINE] Maybe I just suck at understanding games<mask> art,<mask> I do not understand the cult following behind SMW, and especially<mask> it is a superior game to SMB3. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Just a heads up, none of these are nostalgic for me-- I only started playing them a few months ago on the Eshop. [NEWLINE] [NEWLINE] I really want to understand what the hubbub is about for SMW? I don't know if it's just me, but it is not as fun as SMB3, it just seems messy. I want to say that it is probably just nostalgia, people reminiscing about their childhood or something, but I don't see it as the best Mario game ever made ( if that award had to go to the retro Marios, I would give it to SMB3 anyway). [NEWLINE] [NEWLINE] Maybe I just suck at understanding games as art, but I do not understand the cult following behind SMW, and especially how it is a superior game to SMB3. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In my experience,  I will say that most of<mask> you are experiencing/describing is<mask> way more common in younger people. [NEWLINE] [NEWLINE] <mask> you get older (and hopefully, a bit more secure in yourself and your relationships) it's less of an issue.  Jealousy and the constant hunt for new sexual partners isn't nearly<mask> big of a priority once you've matured yourself and your relationship. [NEWLINE] [NEWLINE] <mask>, the idea that one gets a boost by being in the company of intelligent, attractive, likeable people who<mask> happen to enjoy your company is certainly not sexist or uncommon (nor inherently "bad").  This is true regardless is that company is of the opposite sex or not. <mask> a male,<mask> your best male friend suddenly had a vagina it wouldn't change the context of your relationship. [NEWLINE] [NEWLINE] Anything people have here will be anecdotal evidence. [NEWLINE] [NEWLINE] Furthermore, automatically dismissing any sort of friendship with the opposite sex simply based on your relationship status ("No, I'm sorry, I don't want to be your friend<mask> I'm in a committed relationship") is pretty absurd. People base relationships (including friendships) on all sort of things - some deep and meaningful, others superficial and/or selfish.  Choosing to maintain a friendship with someone<mask> of witty banter, shared interest, mutual attraction, and sexual tension isn't the worst someone can do.  At the end of the day, those are just feelings that are dealt with<mask> (acted upon or not, like every other urge you may get).  You're saying a lot more about the individual people than you are about the context of their friendship.</s>
Label encoding: <s>In my experience,  I will say that most of what you are experiencing/describing is also way more common in younger people. [NEWLINE] [NEWLINE] As you get older (and hopefully, a bit more secure in yourself and your relationships) it's less of an issue.  Jealousy and the constant hunt for new sexual partners isn't nearly as big of a priority once you've matured yourself and your relationship. [NEWLINE] [NEWLINE] Additionally, the idea that one gets a boost by being in the company of intelligent, attractive, likeable people who also happen to enjoy your company is certainly not sexist or uncommon (nor inherently "bad").  This is true regardless is that company is of the opposite sex or not.  As a male, if your best male friend suddenly had a vagina it wouldn't change the context of your relationship. [NEWLINE] [NEWLINE] Anything people have here will be anecdotal evidence. [NEWLINE] [NEWLINE] Furthermore, automatically dismissing any sort of friendship with the opposite sex simply based on your relationship status ("No, I'm sorry, I don't want to be your friend because I'm in a committed relationship") is pretty absurd. People base relationships (including friendships) on all sort of things - some deep and meaningful, others superficial and/or selfish.  Choosing to maintain a friendship with someone because of witty banter, shared interest, mutual attraction, and sexual tension isn't the worst someone can do.  At the end of the day, those are just feelings that are dealt with accordingly (acted upon or not, like every other urge you may get).  You're saying a lot more about the individual people than you are about the context of their friendship.</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There were 8.5 million armed service members in the US during the Vietnam War, 3.5 million of which were deployed to Vietnam and the surrounding areas. Either stat is significantly more than the current size of the US military, which is less than 2.5 million including reserves. You haven't explained at all<mask>,<mask> the US decided to me in a similar conflict today, they wouldn't start conscripting. [ENDQ] [NEWLINE] That's 8.5 million individuals who served during the Vietnam War era which officially lasted from 1964 to 1973, not 8.5 million simultaneously serving in the military. [NEWLINE] [NEWLINE] Those 3.5 million were total troops rotated through Vietnam during those 8 years, for which the average tour length was a six months to a year. [NEWLINE] [NEWLINE] <mask><mask>, *peak* troop strength in Vietnam was at ~540,000 in 1968. [NEWLINE] [NEWLINE] Even with the draft at the height of the Vietnam War, total US active duty strength in the late 60s peaked at 3.5 million, with many troops being sent to West Germany or South Korea, which were at the height of the Cold War. [NEWLINE] [NEWLINE] <mask> not only are your numbers over-inflated with regards to # drafted for Vietnam, it's<mask> forgetting those committed to the Cold War elsewhere. [NEWLINE] [NEWLINE] <mask><mask>, military doctrine has changed considerably<mask> the Vietnam War era - a draft isn't going to happen<mask><mask><mask> our active duty forces are sufficient in size for to fulfill our National Security Strategy<mask> outlined every four years by the president. Our current NSS is significantly smaller than the "win two wars simultaneously" doctrine of the Cold War era</s>
Label encoding: <s> [STARTQ] There were 8.5 million armed service members in the US during the Vietnam War, 3.5 million of which were deployed to Vietnam and the surrounding areas. Either stat is significantly more than the current size of the US military, which is less than 2.5 million including reserves. You haven't explained at all why, if the US decided to me in a similar conflict today, they wouldn't start conscripting. [ENDQ] [NEWLINE] That's 8.5 million individuals who served during the Vietnam War era which officially lasted from 1964 to 1973, not 8.5 million simultaneously serving in the military. [NEWLINE] [NEWLINE] Those 3.5 million were total troops rotated through Vietnam during those 8 years, for which the average tour length was a six months to a year. [NEWLINE] [NEWLINE] In fact, *peak* troop strength in Vietnam was at ~540,000 in 1968. [NEWLINE] [NEWLINE] Even with the draft at the height of the Vietnam War, total US active duty strength in the late 60s peaked at 3.5 million, with many troops being sent to West Germany or South Korea, which were at the height of the Cold War. [NEWLINE] [NEWLINE] So not only are your numbers over-inflated with regards to # drafted for Vietnam, it's also forgetting those committed to the Cold War elsewhere. [NEWLINE] [NEWLINE] In addition, military doctrine has changed considerably since the Vietnam War era - a draft isn't going to happen so long as our active duty forces are sufficient in size for to fulfill our National Security Strategy as outlined every four years by the president. Our current NSS is significantly smaller than the "win two wars simultaneously" doctrine of the Cold War era</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's clearly some middle ground,<mask> you leave sandwiches for a<mask> and want them cut.<mask> about a party<mask> you want cut sandwiches for several hours?<mask> about taking food to work for lunch<mask> it has to last for hours? [NEWLINE] [NEWLINE] [STARTQ] I don't see<mask> a triangular cut would be any messier than a rectangular cut<mask> you'd still be burying your face is sandwichness<mask> you eat. [ENDQ] [NEWLINE] The width varies,<mask><mask> pushing it into your mouth it's easy to push the corner in then push a bit too much in and have the edge smear. It's a common and annoying problem. A triangle is just too unpredictable. [NEWLINE] [NEWLINE] [URL].jpg [NEWLINE] [NEWLINE] See this image say-<mask> she pushes it just slightly to the left or right the corner of her mouth will be smudged with food. With rectangles by contrast you can design thin, easily eatable strips of sandwich. [NEWLINE] [NEWLINE] With the sandwich stacking the images you compare are of vastly different sandwiches. [NEWLINE] [NEWLINE] [URL].jpg [NEWLINE] [NEWLINE] <mask> it really isn't hard to stack rectangles well. And<mask> you do<mask> then<mask> people remove sandwiches you don't get the mess that is<mask> normal. With your artistic platter formation after a few sandwiches are removed from it they start to fall apart, flop over each other, leak fluids. It's artistic until someone eats it. With a rectangle? You just remove a sandwich from the top and repeat. I admit that triangles can be pretty<mask><mask> you want more than looks the rectangle is the way to go. [NEWLINE] [NEWLINE] Thanks with the transportation and dipping, and thanks for the delta.</s><pad>
Label encoding: <s>There's clearly some middle ground, where you leave sandwiches for a while and want them cut. What about a party where you want cut sandwiches for several hours? What about taking food to work for lunch where it has to last for hours? [NEWLINE] [NEWLINE] [STARTQ] I don't see how a triangular cut would be any messier than a rectangular cut since you'd still be burying your face is sandwichness as you eat. [ENDQ] [NEWLINE] The width varies, so when pushing it into your mouth it's easy to push the corner in then push a bit too much in and have the edge smear. It's a common and annoying problem. A triangle is just too unpredictable. [NEWLINE] [NEWLINE] [URL].jpg [NEWLINE] [NEWLINE] See this image say- if she pushes it just slightly to the left or right the corner of her mouth will be smudged with food. With rectangles by contrast you can design thin, easily eatable strips of sandwich. [NEWLINE] [NEWLINE] With the sandwich stacking the images you compare are of vastly different sandwiches. [NEWLINE] [NEWLINE] [URL].jpg [NEWLINE] [NEWLINE] But it really isn't hard to stack rectangles well. And if you do so then when people remove sandwiches you don't get the mess that is so normal. With your artistic platter formation after a few sandwiches are removed from it they start to fall apart, flop over each other, leak fluids. It's artistic until someone eats it. With a rectangle? You just remove a sandwich from the top and repeat. I admit that triangles can be pretty but if you want more than looks the rectangle is the way to go. [NEWLINE] [NEWLINE] Thanks with the transportation and dipping, and thanks for the delta.</s><pad>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Could you elaborate a little bit on your definition of harm?<mask><mask> that is essential to really discuss your view, and it can mean a great variety of things<mask> harm is a very vague term. Here are some questions that could help you flesh out your terms a little. [NEWLINE] [NEWLINE] *<mask> does your definition of physical harm begin and end? Does someone engaging in a self destructive habit qualify (i.e. eating too much, smoking, recreational drugs, not getting enough exercise)?<mask> about two people who are have consented to fighting eachother? Or<mask> about instances<mask> person is into kinky sex that involves pain? [NEWLINE] * Are you counting emotional harm?<mask> you determine who is necessarily harming who in these types of situations? [NEWLINE] * Do you include psychological harm (i.e.<mask> your children are exposed to this type of behavior it can lead to developmental issues down the line)? [NEWLINE] * Spiritual harm (i.e. people who believe they will be condemned to hell<mask> they don't fight against an action)? [NEWLINE] * Economic harm? (i.e. slander can cause a person to get fired. Sometimes personalities clash and a person is fired without just cause) [NEWLINE] *<mask> about instances<mask> someone does not believe they are being harmed (this can be found in instances of domestic abuse)? [NEWLINE] *<mask> about situations<mask> harm is more nuanced and subtle? For instance, would you consider a homophobic person making flippant homophobic comments around a gay adolescent to be harmful? [NEWLINE] [NEWLINE] <mask> you don't have answers to these questions it will be difficult to discuss your views very thoroughly,<mask> hopefully it provides some food for thought.</s>
Label encoding: <s>Could you elaborate a little bit on your definition of harm? I think that is essential to really discuss your view, and it can mean a great variety of things because harm is a very vague term. Here are some questions that could help you flesh out your terms a little. [NEWLINE] [NEWLINE] * Where does your definition of physical harm begin and end? Does someone engaging in a self destructive habit qualify (i.e. eating too much, smoking, recreational drugs, not getting enough exercise)? What about two people who are have consented to fighting eachother? Or what about instances where person is into kinky sex that involves pain? [NEWLINE] * Are you counting emotional harm? How you determine who is necessarily harming who in these types of situations? [NEWLINE] * Do you include psychological harm (i.e. if your children are exposed to this type of behavior it can lead to developmental issues down the line)? [NEWLINE] * Spiritual harm (i.e. people who believe they will be condemned to hell if they don't fight against an action)? [NEWLINE] * Economic harm? (i.e. slander can cause a person to get fired. Sometimes personalities clash and a person is fired without just cause) [NEWLINE] * What about instances where someone does not believe they are being harmed (this can be found in instances of domestic abuse)? [NEWLINE] * What about situations where harm is more nuanced and subtle? For instance, would you consider a homophobic person making flippant homophobic comments around a gay adolescent to be harmful? [NEWLINE] [NEWLINE] If you don't have answers to these questions it will be difficult to discuss your views very thoroughly, so hopefully it provides some food for thought.</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you donate your body to science, most of the time you will be a cadaver for medical students.  In most places, after a year the schools give the bodies back to family and pay for a burial (or whatever funeral rites the family/deceased asked for). [NEWLINE] [NEWLINE] There are some people whose bodies never make it back to their families,<mask> generally<mask> you donate your body to science your family still has *something* to put in a grave/cremate and put in an urn. [NEWLINE] [NEWLINE] I understand that this probably isn't going to convince anyone not to donate their bodies, I'm just letting you know that even<mask> you do there's more to it. <mask>,<mask> you're serious about it, get it in legal writing.  Contact a nearby medical school and they'll be happy to set you up for donation upon death.  Even<mask> your family doesn't get any part of you returned to them it's pretty common for them to offer to pay for the cost of a memorial service. [NEWLINE] [NEWLINE] <mask><mask><mask> an actual grave goes, that's choice personal,<mask> graves aren't really about the dead.  They're about giving the living a place to grieve and a sense of finality.  For some people maintaining a grave of a loved one is important to them<mask> it shows that the deceased hasn't been forgotten.  Beyond the people who will actually know you, graves are important to historians who get tons of data from there. <mask> your grave may not be important in 100 years,<mask> in 500 or 1,000 years it might be to someone.</s>
Label encoding: <s>If you donate your body to science, most of the time you will be a cadaver for medical students.  In most places, after a year the schools give the bodies back to family and pay for a burial (or whatever funeral rites the family/deceased asked for). [NEWLINE] [NEWLINE] There are some people whose bodies never make it back to their families, but generally if you donate your body to science your family still has *something* to put in a grave/cremate and put in an urn. [NEWLINE] [NEWLINE] I understand that this probably isn't going to convince anyone not to donate their bodies, I'm just letting you know that even if you do there's more to it.  Also, if you're serious about it, get it in legal writing.  Contact a nearby medical school and they'll be happy to set you up for donation upon death.  Even if your family doesn't get any part of you returned to them it's pretty common for them to offer to pay for the cost of a memorial service. [NEWLINE] [NEWLINE] As far as an actual grave goes, that's choice personal, but graves aren't really about the dead.  They're about giving the living a place to grieve and a sense of finality.  For some people maintaining a grave of a loved one is important to them because it shows that the deceased hasn't been forgotten.  Beyond the people who will actually know you, graves are important to historians who get tons of data from there.  So your grave may not be important in 100 years, but in 500 or 1,000 years it might be to someone.</s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The notion of a right existing independent of the recognition of that right is really not a particularly difficult concept, and it has great utility. [NEWLINE] [NEWLINE] <mask> someone believes that there are liberties that people should enjoy<mask><mask> the actual circumstances of the government they exist under, then they are acknowledging the per-existence of rights to some degree or another. [NEWLINE] [NEWLINE] <mask> it is unarguably true that a person's rights are realized only to the extent that their current situation allows for such rights to be recognized; the notion that such a person should have particular liberties need in no way be tied to the actual reality. [NEWLINE] [NEWLINE] <mask>, it is nearly impossible to talk about reasons for wanting to change some existing government structure without running into the notion of rights. [NEWLINE] [NEWLINE] That citizens hold responsibilities<mask> well,<mask> you point out, is not arguable.<mask>, it is not the case that one depends upon or is contingent upon the other. [NEWLINE] [NEWLINE] I have a responsibility to be engaged in the civic life of my community<mask><mask><mask> my local government is violating some right or another, either of myself or another.<mask>, the greater the degree of any such violations the more incumbent it is upon me to be involved! It is precisely<mask> our rights are most threatened that our collective responsibilities are the most burdensome. [NEWLINE] [NEWLINE] You are not wrong to presume that responsibilities exist. You are mistaken in thinking that rights are contingent upon them. That is not the case. I can abrogate my responsibilities independent of any violations of my rights. Likewise, I can see my rights trampled irrespective of my having in no way failed in my responsibilities. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>The notion of a right existing independent of the recognition of that right is really not a particularly difficult concept, and it has great utility. [NEWLINE] [NEWLINE] If someone believes that there are liberties that people should enjoy regardless of the actual circumstances of the government they exist under, then they are acknowledging the per-existence of rights to some degree or another. [NEWLINE] [NEWLINE] While it is unarguably true that a person's rights are realized only to the extent that their current situation allows for such rights to be recognized; the notion that such a person should have particular liberties need in no way be tied to the actual reality. [NEWLINE] [NEWLINE] Indeed, it is nearly impossible to talk about reasons for wanting to change some existing government structure without running into the notion of rights. [NEWLINE] [NEWLINE] That citizens hold responsibilities as well, as you point out, is not arguable. However, it is not the case that one depends upon or is contingent upon the other. [NEWLINE] [NEWLINE] I have a responsibility to be engaged in the civic life of my community regardless of if my local government is violating some right or another, either of myself or another. Indeed, the greater the degree of any such violations the more incumbent it is upon me to be involved! It is precisely when our rights are most threatened that our collective responsibilities are the most burdensome. [NEWLINE] [NEWLINE] You are not wrong to presume that responsibilities exist. You are mistaken in thinking that rights are contingent upon them. That is not the case. I can abrogate my responsibilities independent of any violations of my rights. Likewise, I can see my rights trampled irrespective of my having in no way failed in my responsibilities. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I work a rotating shift schedule in a refinery/chemical plant. Most units in the plant work the same sort of schedule that I do: two 12-hour days, then two 12-hour nights, and then 4 days off. 5am-5pm, or 5pm-5am. [NEWLINE] [NEWLINE] <mask> we had operators working only night shifts, they wouldn't develop a lot of the key skills and knowledge that is needed to work this job. During the day operations, the plant does maintenance and construction work. We<mask> interact with contractors, engineers, management, and all sorts of others who play a key role in keeping the plant running smoothly. [NEWLINE] [NEWLINE] <mask> I only ever worked nights, I would never see<mask> goes on<mask> a pump is taken out of service, or<mask> work is required to open up and repair a large gas compressor. Then,<mask> an emergency of some sort, be it a leak, or a process unit upset were to happen at night, I wouldn't have enough actual experience with the equipment to know<mask>'s<mask>. All the night shift really does is monitor things to ensure smooth operations, and<mask> that was the only shift I worked, I wouldn't be able to do much more than say "hey that thing is leaking". [NEWLINE] [NEWLINE] We<mask> get paid quite well to compensate for having to work a schedule like this, and everyone working alongside me in plant operations knew exactly<mask> they were getting into<mask> they applied for this job. The plant can't shut down nightly, and we need experienced employees at night, and the only way to really get experience in this job is to<mask> work days.</s>
Label encoding: <s>I work a rotating shift schedule in a refinery/chemical plant. Most units in the plant work the same sort of schedule that I do: two 12-hour days, then two 12-hour nights, and then 4 days off. 5am-5pm, or 5pm-5am. [NEWLINE] [NEWLINE] If we had operators working only night shifts, they wouldn't develop a lot of the key skills and knowledge that is needed to work this job. During the day operations, the plant does maintenance and construction work. We also interact with contractors, engineers, management, and all sorts of others who play a key role in keeping the plant running smoothly. [NEWLINE] [NEWLINE] If I only ever worked nights, I would never see what goes on when a pump is taken out of service, or what work is required to open up and repair a large gas compressor. Then, if an emergency of some sort, be it a leak, or a process unit upset were to happen at night, I wouldn't have enough actual experience with the equipment to know what's what. All the night shift really does is monitor things to ensure smooth operations, and if that was the only shift I worked, I wouldn't be able to do much more than say "hey that thing is leaking". [NEWLINE] [NEWLINE] We also get paid quite well to compensate for having to work a schedule like this, and everyone working alongside me in plant operations knew exactly what they were getting into when they applied for this job. The plant can't shut down nightly, and we need experienced employees at night, and the only way to really get experience in this job is to also work days.</s>
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand<mask> you're coming from, I really do,<mask> I still have a problem with the expression "social responsibility"<mask>, for me, it implies a generalization that is too far from being realistic and humane to be acceptable. [NEWLINE] [NEWLINE] A child or a teenager should be considered socially responsible for reporting something that they might not even understand? [NEWLINE] [NEWLINE] Someone who remembers saying no quite a few times<mask> was too drunk and didn't have the strength to physically resist the assault (<mask> isn't bruised), having no chance whatsoever to prove lack of consent, should be considered socially responsible? [NEWLINE] [NEWLINE] [NEWLINE] <mask> about someone who is having trouble understanding that they were not at fault<mask> they just keep wondering<mask> it was something that *they* did? Maybe they should have dressed more appropriately, or maybe they shouldn't have smiled<mask> much and definitely they should not have flirted back. They should have taken a taxi home. They put themselves in that situation. It can take a serious amount of time for someone to realize it was not their fault,<mask> can they be responsible for reporting it<mask> they can't be sure themselves they weren't to blame. [NEWLINE] [NEWLINE] [NEWLINE] This kind of experience is huge on the victims.<mask> they are sure about the aggressor's identity, went through medical exams<mask> DNA and other evidence was collected and/or are strong enough to have their story questioned hundreds of times then please do press charges. It might help someone else down the line.<mask><mask> they aren't, that shouldn't have to be an extra weight for them to carry, they are not responsible for anything, they are the victims. </s>
Label encoding: <s>I understand where you're coming from, I really do, but I still have a problem with the expression "social responsibility" because, for me, it implies a generalization that is too far from being realistic and humane to be acceptable. [NEWLINE] [NEWLINE] A child or a teenager should be considered socially responsible for reporting something that they might not even understand? [NEWLINE] [NEWLINE] Someone who remembers saying no quite a few times but was too drunk and didn't have the strength to physically resist the assault ( so isn't bruised), having no chance whatsoever to prove lack of consent, should be considered socially responsible? [NEWLINE] [NEWLINE] [NEWLINE] What about someone who is having trouble understanding that they were not at fault because they just keep wondering if it was something that *they* did? Maybe they should have dressed more appropriately, or maybe they shouldn't have smiled as much and definitely they should not have flirted back. They should have taken a taxi home. They put themselves in that situation. It can take a serious amount of time for someone to realize it was not their fault, how can they be responsible for reporting it if they can't be sure themselves they weren't to blame. [NEWLINE] [NEWLINE] [NEWLINE] This kind of experience is huge on the victims. If they are sure about the aggressor's identity, went through medical exams where DNA and other evidence was collected and/or are strong enough to have their story questioned hundreds of times then please do press charges. It might help someone else down the line. But if they aren't, that shouldn't have to be an extra weight for them to carry, they are not responsible for anything, they are the victims. </s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The most important aspect of a constitution is stability and consistency regarding the laws which govern everyone. Without those two things, a constitution is meaningless. [NEWLINE] [NEWLINE] <mask> abandoning whatever the constitution was intended to mean, on a whim, renders that constitution pointless. The fundamental principles of a nation should NOT just change on the fly unless there is extremely good reason to do<mask>. [NEWLINE] [NEWLINE] The other important reason to have a constitution (which is related to the above), is to take power away from people. The founding fathers realized that people are corruptible, and have a propensity to become tyrants. A constitution limits their power, and gives the people a comparative point of reference: [NEWLINE] [NEWLINE] "Hey, here's<mask> our constitution says, here's<mask> our government/leaders say, and those two things do not add up. Time to get rid of the government/leaders". [NEWLINE] [NEWLINE] Without a consistent reference point that can be evaluated equally by each generation, the government is free to evolve and morph into something ugly. [NEWLINE] [NEWLINE] Can it evolve into something good? Sure.<mask> the problem with ugly evolution is that it's always one-way. Once a government gains power, it rarely ever lets go of it. Even with our current constitution, the Federal government has WAY WAY WAY too much power. [NEWLINE] [NEWLINE] 1. It's mandating that you own a private sector service (health insurance) [NEWLINE] 2. The NSA... exists [NEWLINE] 3. A judge recently ruled that the CIA is above the law [NEWLINE] [NEWLINE] There are numerous examples of<mask> big and ugly the federal government has gotten,<mask> a constitution intended to help keep it in check. </s>
Label encoding: <s>The most important aspect of a constitution is stability and consistency regarding the laws which govern everyone. Without those two things, a constitution is meaningless. [NEWLINE] [NEWLINE] Thus abandoning whatever the constitution was intended to mean, on a whim, renders that constitution pointless. The fundamental principles of a nation should NOT just change on the fly unless there is extremely good reason to do so. [NEWLINE] [NEWLINE] The other important reason to have a constitution (which is related to the above), is to take power away from people. The founding fathers realized that people are corruptible, and have a propensity to become tyrants. A constitution limits their power, and gives the people a comparative point of reference: [NEWLINE] [NEWLINE] "Hey, here's what our constitution says, here's what our government/leaders say, and those two things do not add up. Time to get rid of the government/leaders". [NEWLINE] [NEWLINE] Without a consistent reference point that can be evaluated equally by each generation, the government is free to evolve and morph into something ugly. [NEWLINE] [NEWLINE] Can it evolve into something good? Sure. But the problem with ugly evolution is that it's always one-way. Once a government gains power, it rarely ever lets go of it. Even with our current constitution, the Federal government has WAY WAY WAY too much power. [NEWLINE] [NEWLINE] 1. It's mandating that you own a private sector service (health insurance) [NEWLINE] 2. The NSA... exists [NEWLINE] 3. A judge recently ruled that the CIA is above the law [NEWLINE] [NEWLINE] There are numerous examples of how big and ugly the federal government has gotten, despite a constitution intended to help keep it in check. </s>
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Useless" is a pretty harsh term. I've stricken that from my original post. And replaced it with something that<mask><mask> better describes<mask> I'm coming from: [NEWLINE] [NEWLINE] [STARTQ] Belief in God(s) is ~~fairly useless~~ incredibly illogical, and is not necessary to provide the benefits attributed to belief in it(them) [ENDQ] [NEWLINE] I personally believe that the belief is "useless" in the sense that the same benefits that your friend experienced from belief could have been accomplished without having to believe in something that has no evidence behind it. There is plenty of non-superstitious literature that contains exactly the same moral and emotional life lessons that the Bible conveys. [NEWLINE] [NEWLINE] For instance,<mask> I used to be depressed and anoyed all the time at others, and really just didn't want to be in society, a thorough read through of David Foster Wallace's "This is Water" speech gave me everything I needed to start leading a happier, healthier, and more selfless life, without believing in anything supernatural. [NEWLINE] [NEWLINE] I would aslo<mask><mask> believing in the *rest* of the things in the Bible/Christianity could be dangerous to your friend going forward,<mask><mask> the immediate benefits toward his health were very nice for him (<mask> he starts to *<mask> * believe in prejudice against gays, misogyny, and apathy toward slavery, things which I would say the Bible most certainly supports at some points, I would say he was not<mask> well off anymore from his belief,<mask><mask> their initial help to him). [NEWLINE] [NEWLINE] Do you see<mask> I'm saying? </s>
Label encoding: <s>"Useless" is a pretty harsh term. I've stricken that from my original post. And replaced it with something that I think better describes where I'm coming from: [NEWLINE] [NEWLINE] [STARTQ] Belief in God(s) is ~~fairly useless~~ incredibly illogical, and is not necessary to provide the benefits attributed to belief in it(them) [ENDQ] [NEWLINE] I personally believe that the belief is "useless" in the sense that the same benefits that your friend experienced from belief could have been accomplished without having to believe in something that has no evidence behind it. There is plenty of non-superstitious literature that contains exactly the same moral and emotional life lessons that the Bible conveys. [NEWLINE] [NEWLINE] For instance, when I used to be depressed and anoyed all the time at others, and really just didn't want to be in society, a thorough read through of David Foster Wallace's "This is Water" speech gave me everything I needed to start leading a happier, healthier, and more selfless life, without believing in anything supernatural. [NEWLINE] [NEWLINE] I would aslo argue that believing in the *rest* of the things in the Bible/Christianity could be dangerous to your friend going forward, even though the immediate benefits toward his health were very nice for him ( If he starts to * also * believe in prejudice against gays, misogyny, and apathy toward slavery, things which I would say the Bible most certainly supports at some points, I would say he was not so well off anymore from his belief, regardless of their initial help to him). [NEWLINE] [NEWLINE] Do you see what I'm saying? </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for the information! I have another question:<mask> can these HFT/Market-making firms be small? Or in other words,<mask> much money do they have tied up against the sell orders, and stock against the buy orders, and<mask> many securities do they cover like this for? [NEWLINE] [NEWLINE] Continuing with using Microsoft stock<mask> an example,<mask> somebody comes to the market wanting to buy small amount, say $1,000 of it, at whatever the current market price is, then presumably a market-making firm will instantly complete that trade out of the stock they have on hand.<mask><mask><mask><mask>,<mask> somebody wants to buy a large amount, say $100,000,000 of Microsoft stock, then I suppose that is too big for these firms to handle, and whoever is buying that will have to go find somebody who has that much stock and find out<mask> price they are willing to sell it at.<mask><mask> is the actual amount, or order of magnitude of amount, that these firms cover the market with? [NEWLINE] [NEWLINE] <mask> a million dollars is a good amount to cover a single security with, then, with 2,700 stocks listed on the NASDAQ, wouldn't there have to be 2.7bn of market cap covering the market for each market-making company? And that's only one company on one exchange,<mask> there are several of each, right?<mask> 5 or<mask> companies were doing market making on each security in each market, then wouldn't there have to be tens or maybe hundreds of billions of dollars tied up in that? Or am I way off on something?</s>
Label encoding: <s>Thanks for the information! I have another question: How can these HFT/Market-making firms be small? Or in other words, how much money do they have tied up against the sell orders, and stock against the buy orders, and how many securities do they cover like this for? [NEWLINE] [NEWLINE] Continuing with using Microsoft stock as an example, if somebody comes to the market wanting to buy small amount, say $1,000 of it, at whatever the current market price is, then presumably a market-making firm will instantly complete that trade out of the stock they have on hand. On the other hand, if somebody wants to buy a large amount, say $100,000,000 of Microsoft stock, then I suppose that is too big for these firms to handle, and whoever is buying that will have to go find somebody who has that much stock and find out what price they are willing to sell it at. But what is the actual amount, or order of magnitude of amount, that these firms cover the market with? [NEWLINE] [NEWLINE] If a million dollars is a good amount to cover a single security with, then, with 2,700 stocks listed on the NASDAQ, wouldn't there have to be 2.7bn of market cap covering the market for each market-making company? And that's only one company on one exchange, when there are several of each, right? If 5 or so companies were doing market making on each security in each market, then wouldn't there have to be tens or maybe hundreds of billions of dollars tied up in that? Or am I way off on something?</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Another aspect to consider is the increasing militarization of police, and advancements in technology.  I don't know the actual statistics of<mask> many bad cops there are, or<mask> it's gone up or down. <mask> even<mask> the percentage had gone down, the change in<mask> tools the police use<mask> affects<mask> much of a problem it is. [NEWLINE] [NEWLINE] Before tasers, there was a much bigger gap between the non-violent methods a cop could use, or going for their gun<mask> they were prepared to kill somebody.  Tasers were advertised<mask> a non-lethal option, to reduce deaths even<mask> things got to that extreme point<mask> a cop would have had to be prepared to kill before.  Instead, tasers have become a compliance tool, used on a much wider array of situations<mask> police would have (or should have) never considered reaching for a gun.  And now tasers have proven to be simply "less lethal" rather than non-lethal,<mask> they can still have fatal effects<mask> used on someone with health problems, or<mask> abused with excessive use. [NEWLINE] [NEWLINE] Other tools like LRADs are used<mask> crowd control.  In a crowd control situation, they could affect many non-violent protesters who would have complied with police directions.  Whether a non-violent protester should expect some risk of having some unruly people in the crowd and getting treated<mask> is a separate debate. <mask> the police should<mask> expect there to be innocents in the collateral damage, and<mask> they shouldn't employ tools like the LRAD's which can cause *permanent* hearing damage.</s>
Label encoding: <s>Another aspect to consider is the increasing militarization of police, and advancements in technology.  I don't know the actual statistics of how many bad cops there are, or if it's gone up or down.  But even if the percentage had gone down, the change in what tools the police use also affects how much of a problem it is. [NEWLINE] [NEWLINE] Before tasers, there was a much bigger gap between the non-violent methods a cop could use, or going for their gun when they were prepared to kill somebody.  Tasers were advertised as a non-lethal option, to reduce deaths even when things got to that extreme point where a cop would have had to be prepared to kill before.  Instead, tasers have become a compliance tool, used on a much wider array of situations where police would have (or should have) never considered reaching for a gun.  And now tasers have proven to be simply "less lethal" rather than non-lethal, as they can still have fatal effects when used on someone with health problems, or when abused with excessive use. [NEWLINE] [NEWLINE] Other tools like LRADs are used as crowd control.  In a crowd control situation, they could affect many non-violent protesters who would have complied with police directions.  Whether a non-violent protester should expect some risk of having some unruly people in the crowd and getting treated accordingly is a separate debate.  But the police should also expect there to be innocents in the collateral damage, and so they shouldn't employ tools like the LRAD's which can cause *permanent* hearing damage.</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, you can.<mask><mask> OP he refers to the Christian god. The Christian god IS "God".<mask><mask> the Christianity, the Christian god IS responsible for every human achievement and everything else. [NEWLINE] [NEWLINE] <mask> there is an assumption that was for the god or gods not under the Christian religion(Pascal's Wager) then the claim is no longer viable under either religion or science. [NEWLINE] [NEWLINE] <mask> : [NEWLINE] [STARTQ] I believe that [ENDQ] [NEWLINE] Key word is OP 'believes' that this is the case(not unlike religion). I'm not assuming,<mask> I'm sure the OP is not a philosopher or a scientist given the questions and use of the word believe. They used their preconceptions and understanding of religion(Christianity) to counter it under those pretenses. [NEWLINE] [NEWLINE] You have a very strong objective point.<mask> this is a subjective matter, and are using the wrong objective idea for this perspective. [NEWLINE] [NEWLINE] <mask> you wanted to play devil's advocate for his/her situation, and in his/her situation the Christian god is relatively defined. I, personally neither subscribe to that or know enough about the Christian religion to provide points to change his/her view. [NEWLINE] [NEWLINE] <mask> it would go something like this: [NEWLINE] [NEWLINE] "Of course god exists, and of course humans need him<mask> they would not exist without him, and/or they would not function righteously" [NEWLINE] [NEWLINE] or [NEWLINE] [NEWLINE] "<mask> this god were to exist, then the small scale that is humanity's pathetic existence would be helplessly dependent on a deity's intervention<mask> of [insert bible quote about needing god]"</s>
Label encoding: <s>Yes, you can. According to OP he refers to the Christian god. The Christian god IS "God". According to the Christianity, the Christian god IS responsible for every human achievement and everything else. [NEWLINE] [NEWLINE] If there is an assumption that was for the god or gods not under the Christian religion(Pascal's Wager) then the claim is no longer viable under either religion or science. [NEWLINE] [NEWLINE] However : [NEWLINE] [STARTQ] I believe that [ENDQ] [NEWLINE] Key word is OP 'believes' that this is the case(not unlike religion). I'm not assuming, but I'm sure the OP is not a philosopher or a scientist given the questions and use of the word believe. They used their preconceptions and understanding of religion(Christianity) to counter it under those pretenses. [NEWLINE] [NEWLINE] You have a very strong objective point. But this is a subjective matter, and are using the wrong objective idea for this perspective. [NEWLINE] [NEWLINE] If you wanted to play devil's advocate for his/her situation, and in his/her situation the Christian god is relatively defined. I, personally neither subscribe to that or know enough about the Christian religion to provide points to change his/her view. [NEWLINE] [NEWLINE] But it would go something like this: [NEWLINE] [NEWLINE] "Of course god exists, and of course humans need him because they would not exist without him, and/or they would not function righteously" [NEWLINE] [NEWLINE] or [NEWLINE] [NEWLINE] " If this god were to exist, then the small scale that is humanity's pathetic existence would be helplessly dependent on a deity's intervention because of [insert bible quote about needing god]"</s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From that video: [NEWLINE] [NEWLINE] [STARTQ] Pedophilia: A psychiatric disorder characterized by primary or exclusive sexual interest toward children. [ENDQ] [NEWLINE] I feel that not only is this kind of wrong (the sexual interest part),<mask> it is kind of... weird.<mask><mask> its NOT a primary/exclusive interest towards children?<mask> then? Is it still a disorder, or is it something different entirely? [NEWLINE] [NEWLINE] I<mask> feel like a lot of the stuff was just not related?<mask> does my height and handedness have to do with my sexual interests? [NEWLINE] [NEWLINE] Another point, on Pedophiles triggering the "sex systems" over the "parental systems", once again,<mask><mask> that isn't the case? [NEWLINE] [NEWLINE] One more: [NEWLINE] [NEWLINE] [STARTQ] We need to be able to be dispassionate and clinical and think rationally. [ENDQ] [NEWLINE] That's not going to help. We aren't robots. We are people, with emotions and we need to have emotions in our interactions. Being cold like this is only going to push people away from getting "help".<mask> you could go to someone that could help you every time,<mask> was cold, uncaring and inhuman about their ways, OR you could go to someone that might help you,<mask> was always kind, warm and nice about their ways, which would you choose? I don't know about you,<mask> I'd go for the kind, warm and nice person, 10/10 times. [NEWLINE] [NEWLINE] This guy leaves a lot of questions unanswered, doesn't seem to really provide a valid solution and just seems to be talking with a lot of fluff.</s>
Label encoding: <s>From that video: [NEWLINE] [NEWLINE] [STARTQ] Pedophilia: A psychiatric disorder characterized by primary or exclusive sexual interest toward children. [ENDQ] [NEWLINE] I feel that not only is this kind of wrong (the sexual interest part), but it is kind of... weird. What if its NOT a primary/exclusive interest towards children? What then? Is it still a disorder, or is it something different entirely? [NEWLINE] [NEWLINE] I also feel like a lot of the stuff was just not related? What does my height and handedness have to do with my sexual interests? [NEWLINE] [NEWLINE] Another point, on Pedophiles triggering the "sex systems" over the "parental systems", once again, what if that isn't the case? [NEWLINE] [NEWLINE] One more: [NEWLINE] [NEWLINE] [STARTQ] We need to be able to be dispassionate and clinical and think rationally. [ENDQ] [NEWLINE] That's not going to help. We aren't robots. We are people, with emotions and we need to have emotions in our interactions. Being cold like this is only going to push people away from getting "help". If you could go to someone that could help you every time, but was cold, uncaring and inhuman about their ways, OR you could go to someone that might help you, but was always kind, warm and nice about their ways, which would you choose? I don't know about you, but I'd go for the kind, warm and nice person, 10/10 times. [NEWLINE] [NEWLINE] This guy leaves a lot of questions unanswered, doesn't seem to really provide a valid solution and just seems to be talking with a lot of fluff.</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> the over sexualization of women on Reddit is repulsive [ENDQ] [NEWLINE] <mask><mask> exactly constitutes over sexualization and<mask> women are more sexualized on reddit than elsewhere is up for up debate, that's your opinion and you're entitled to it. [NEWLINE] [STARTQ] The "front page" of the internet should not have obviously sexual pictures of women<mask> it wants to continue holding that title. [ENDQ] It's not like there's some top down authority that declared the front page of the internet. It's the front page<mask><mask> many people frequent it. Unfortunately (fortunately?) sex is<mask> interests people. Women aren't any less complicity of that. They're just underrepresented in front page comments. [NEWLINE] [NEWLINE] You say the sexualization of women detracts from the quality of front page discussions. This is probably true,<mask> even it wasn't there, front page discussions would probably suck anyway. It's really not a good place for quality comments or content. For alternative, look for subreddits on topics that interest you. Try the sfw porn network. Look for moderate sized subreddits that have discussions that stay on topic. I find the a good critical mass for a subreddit is 10k subscribers. That usually allows for a good discussion. There are some larger subreddits that I work<mask> they are well moderated. /r/DepthHub, /r/AskHistorians, /r/TrueReddit, /r/AskWomen are some that are interesting to me.<mask> time goes on, I find myself unsubscribing from more and more default subs, and my experience has only been improved by it. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I think the over sexualization of women on Reddit is repulsive [ENDQ] [NEWLINE] Although what exactly constitutes over sexualization and if women are more sexualized on reddit than elsewhere is up for up debate, that's your opinion and you're entitled to it. [NEWLINE] [STARTQ] The "front page" of the internet should not have obviously sexual pictures of women if it wants to continue holding that title. [ENDQ] It's not like there's some top down authority that declared the front page of the internet. It's the front page because so many people frequent it. Unfortunately (fortunately?) sex is what interests people. Women aren't any less complicity of that. They're just underrepresented in front page comments. [NEWLINE] [NEWLINE] You say the sexualization of women detracts from the quality of front page discussions. This is probably true, but even it wasn't there, front page discussions would probably suck anyway. It's really not a good place for quality comments or content. For alternative, look for subreddits on topics that interest you. Try the sfw porn network. Look for moderate sized subreddits that have discussions that stay on topic. I find the a good critical mass for a subreddit is 10k subscribers. That usually allows for a good discussion. There are some larger subreddits that I work because they are well moderated. /r/DepthHub, /r/AskHistorians, /r/TrueReddit, /r/AskWomen are some that are interesting to me. As time goes on, I find myself unsubscribing from more and more default subs, and my experience has only been improved by it. [NEWLINE] </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1.) $0.60 is not much extra per six pack, and is probably dwarfed by the alcohol/sales/other tax you're already paying for it.  No one is happy to see prices rise,<mask> it's death and well... You know. [NEWLINE] [NEWLINE] 2.)No one will know you're protesting the tax just<mask> you stop recycling,<mask> it will not be an effective protest.<mask><mask>,<mask> you continue to drink you'll still be paying the tax you just won't be recycling. [NEWLINE] [NEWLINE] 3.)  Addressing: [NEWLINE] [STARTQ] I find it infuriating the the government has essentially created a needless job for people via a tax...  You then have others who exploit this to earn a living. [ENDQ] [NEWLINE] [NEWLINE] Of which I'm not sure<mask><mask> with any... It's not a needless job, for example.  Sure, many people will actually look through people's recycling bins and take out bottles that were already going to be recycled<mask> they can make money off the deposit, which doesn't help anyone,<mask> many other bottles wind up along the side of the road or on lawns.  I've seen homeless people collect these bottles plenty of times, and in doing<mask> they're picking up litter and putting resources back into the recycling stream and the only reason they're doing it<mask> of the deposit.  I frankly don't think it's fair that to say they are exploiting this system that most people ignore;<mask> anything by providing a petty $0.05 per bottle they pick up, the government is exploiting *them*. [NEWLINE] [NEWLINE] edit:formatting</s>
Label encoding: <s>1.) $0.60 is not much extra per six pack, and is probably dwarfed by the alcohol/sales/other tax you're already paying for it.  No one is happy to see prices rise, but it's death and well... You know. [NEWLINE] [NEWLINE] 2.)No one will know you're protesting the tax just because you stop recycling, so it will not be an effective protest. In fact, if you continue to drink you'll still be paying the tax you just won't be recycling. [NEWLINE] [NEWLINE] 3.)  Addressing: [NEWLINE] [STARTQ] I find it infuriating the the government has essentially created a needless job for people via a tax...  You then have others who exploit this to earn a living. [ENDQ] [NEWLINE] [NEWLINE] Of which I'm not sure I agree with any... It's not a needless job, for example.  Sure, many people will actually look through people's recycling bins and take out bottles that were already going to be recycled so they can make money off the deposit, which doesn't help anyone, but many other bottles wind up along the side of the road or on lawns.  I've seen homeless people collect these bottles plenty of times, and in doing so they're picking up litter and putting resources back into the recycling stream and the only reason they're doing it because of the deposit.  I frankly don't think it's fair that to say they are exploiting this system that most people ignore; if anything by providing a petty $0.05 per bottle they pick up, the government is exploiting *them*. [NEWLINE] [NEWLINE] edit:formatting</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is an interesting point.<mask><mask><mask><mask>, all of those bands did have major hits that reached the top of the pop charts and sold out stadiums doing it. In general,<mask><mask> that the only rock bands that can still sell out stadiums without being part of a festival are legacy acts or reunion tours. [NEWLINE] [NEWLINE] <mask>, by your logic, we should be seeing bigger pockets of innovation and movements in rock that don't have mainstream exposure.<mask><mask><mask> there are definitely trends that are playing out in rock right now ([The Return of the Male Piano Balladeer]( [URL] /), [the female takeover of Punk Rock]( [URL] )) I would<mask><mask> even those have a tendency to either be throwbacks to an older style of music, or an interspersion of EDM, hip-hop, and R&amp;B into rock. One could point to the evolution of Noise<mask> a genre<mask> an innovation in rock,<mask> [<mask> this article points out]( [URL] ;CNDID=31368913&amp;mbid=nl_072415Daily&amp;CNDID=31368913&amp;spMailingID=7931543&amp;spUserID=MTA1MTUwNzk4Njc5S0&amp;spJobID=723208489&amp;spReportId=NzIzMjA4NDg5S0), part of<mask> is allowing Noise to gain ground is its back-and-forth with modern (not particularly rock-influenced) pop.</s>
Label encoding: <s>This is an interesting point. On the other hand, all of those bands did have major hits that reached the top of the pop charts and sold out stadiums doing it. In general, I think that the only rock bands that can still sell out stadiums without being part of a festival are legacy acts or reunion tours. [NEWLINE] [NEWLINE] Also, by your logic, we should be seeing bigger pockets of innovation and movements in rock that don't have mainstream exposure. While I think there are definitely trends that are playing out in rock right now ([The Return of the Male Piano Balladeer]( [URL] /), [the female takeover of Punk Rock]( [URL] )) I would argue that even those have a tendency to either be throwbacks to an older style of music, or an interspersion of EDM, hip-hop, and R&amp;B into rock. One could point to the evolution of Noise as a genre as an innovation in rock, but [ as this article points out]( [URL] ;CNDID=31368913&amp;mbid=nl_072415Daily&amp;CNDID=31368913&amp;spMailingID=7931543&amp;spUserID=MTA1MTUwNzk4Njc5S0&amp;spJobID=723208489&amp;spReportId=NzIzMjA4NDg5S0), part of what is allowing Noise to gain ground is its back-and-forth with modern (not particularly rock-influenced) pop.</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Nintendo has three games that get more attention than the others. Mario Kart, Smash Bros, and Mario Party, none of these games existed for the first 18 months of the consoles life. [ENDQ] [NEWLINE] There are two big ones missing from this: Zelda and the core Mario games. [NEWLINE] [NEWLINE] Wii U did launch with a new Mario: Super Mario Bros U, and a second core Mario, Super Mario 3D World, launched a<mask> back<mask> well. [NEWLINE] [NEWLINE] <mask> Nintendo has a relatively small number of tentpoles<mask>, it makes sense that they're conservative about releasing them too quickly.  Nintendo has never released a truly bad Mario or Zelda tentpole game for example.  And keeping that quality means the games aren't done quickly. [NEWLINE] [NEWLINE] <mask> to the third party developers question,<mask><mask> Nintendo is playing it smart. <mask> they tried to copy Sony and Microsoft, they're basically making a commodity: a powerful computer with good graphics to plug into your TV.  Making a console with wildly different controls puts them in a spot of selling something that's unique and has potential to be a blockbuster in the way the Wii was. <mask><mask> they did botch the launch marketing of the console<mask>, no question. [NEWLINE] [NEWLINE] There's a business case for making a quirky console, especially<mask> you have the kind of brand power Nintendo does.  The Wii U did miss the mark in some important respects,<mask> that doesn't mean Nintendo should stop making consoles, especially with the success they've had with the DS and 3DS.  One flop does not kill a model.</s>
Label encoding: <s> [STARTQ] Nintendo has three games that get more attention than the others. Mario Kart, Smash Bros, and Mario Party, none of these games existed for the first 18 months of the consoles life. [ENDQ] [NEWLINE] There are two big ones missing from this: Zelda and the core Mario games. [NEWLINE] [NEWLINE] Wii U did launch with a new Mario: Super Mario Bros U, and a second core Mario, Super Mario 3D World, launched a while back as well. [NEWLINE] [NEWLINE] Because Nintendo has a relatively small number of tentpoles though, it makes sense that they're conservative about releasing them too quickly.  Nintendo has never released a truly bad Mario or Zelda tentpole game for example.  And keeping that quality means the games aren't done quickly. [NEWLINE] [NEWLINE] As to the third party developers question, I think Nintendo is playing it smart.  If they tried to copy Sony and Microsoft, they're basically making a commodity: a powerful computer with good graphics to plug into your TV.  Making a console with wildly different controls puts them in a spot of selling something that's unique and has potential to be a blockbuster in the way the Wii was.  I think they did botch the launch marketing of the console though, no question. [NEWLINE] [NEWLINE] There's a business case for making a quirky console, especially when you have the kind of brand power Nintendo does.  The Wii U did miss the mark in some important respects, but that doesn't mean Nintendo should stop making consoles, especially with the success they've had with the DS and 3DS.  One flop does not kill a model.</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're not wrong, a big part of it is that teachers have a **continuing contract** (not 'tenure', at least in my state,<mask> the difference is semantic) is<mask> they have strong unions. Be careful,<mask>, they're not federal employees, they're employees of the school district which is an extension of the *state* government. [NEWLINE] [NEWLINE] I kinda want to preface this by saying that you probably don't understand<mask> having a continuing contract/tenure really means.<mask> the administration does its part and keeps records of teacher performance, behavior, and opportunities to correct themselves, it is  hard to rightfully terminate teacher employment. Every state has it's own guidelines about<mask> a teacher'a continuing contract can be terminated for,<mask> they're largely the same, and they cover the bases. Incompetence, insubordination, moral turpitude, etc. All it really means is that teacher contracts cannot be terminated on a whim. There must be sufficient cause, due notice, etc., and you should keep records of that in case the teacher tries to fight it. [NEWLINE] [NEWLINE] And that's really it. It's protection for teachers who didn't do anything wrong, for whom sufficient funding is available, and can at least improve<mask> given criticism. It's some level of job security. [NEWLINE] [NEWLINE] <mask>, just an FYI, there are **not** set curricula. There are state-mandated curriculum *guidelines*, which can be pretty vague, or at least open-ended in order to protect teacher autonomy.</s>
Label encoding: <s>You're not wrong, a big part of it is that teachers have a **continuing contract** (not 'tenure', at least in my state, though the difference is semantic) is because they have strong unions. Be careful, though, they're not federal employees, they're employees of the school district which is an extension of the *state* government. [NEWLINE] [NEWLINE] I kinda want to preface this by saying that you probably don't understand what having a continuing contract/tenure really means. If the administration does its part and keeps records of teacher performance, behavior, and opportunities to correct themselves, it is  hard to rightfully terminate teacher employment. Every state has it's own guidelines about what a teacher'a continuing contract can be terminated for, but they're largely the same, and they cover the bases. Incompetence, insubordination, moral turpitude, etc. All it really means is that teacher contracts cannot be terminated on a whim. There must be sufficient cause, due notice, etc., and you should keep records of that in case the teacher tries to fight it. [NEWLINE] [NEWLINE] And that's really it. It's protection for teachers who didn't do anything wrong, for whom sufficient funding is available, and can at least improve when given criticism. It's some level of job security. [NEWLINE] [NEWLINE] Also, just an FYI, there are **not** set curricula. There are state-mandated curriculum *guidelines*, which can be pretty vague, or at least open-ended in order to protect teacher autonomy.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ I checked out the news page and,<mask> not 100% perfect, it actually seems better than CNN in terms of clickbait/native advertisement. Most seem like well structured/legitimate articles. I suppose<mask> you are looking to BuzzFeed<mask> a new source and stick to the '/news' section it's probably good. The editor in chief seems to have a lot of experience and has built a pretty large newsroom. It's definitely not separate in most people's minds (including mine),<mask> I suppose that's just an ingrained bias in me from all the silly quizzes posted about on facebook. [NEWLINE] [NEWLINE] I can see your comparison with Reddit. It's a little different<mask> Reddit isn't trying to pass itself of<mask> a news site<mask><mask> a content aggregator,<mask> I'd be lying<mask> I didn't say I get a good chunk of my news from here. Heck, sometimes<mask> I'm lazy I don't even read articles and just get the news from random people on the internet in the comment section. And I suppose all the hard hitting, important news I find on reddit is sandwiched between cute cat photos,<mask> I don't say it hurts journalism.<mask><mask> people can keep the two separate even<mask> it's all on the front page. [NEWLINE] [NEWLINE] I'd say I'm still not going over to BuzzFeed, and will still not hold it in much regard,<mask> definitely view it in a different light now. Maybe after time I can come to see BuzzFeed news separately from the rest of the site. Thanks for giving me some good perspective!</s>
Label encoding: <s>∆ I checked out the news page and, while not 100% perfect, it actually seems better than CNN in terms of clickbait/native advertisement. Most seem like well structured/legitimate articles. I suppose if you are looking to BuzzFeed as a new source and stick to the '/news' section it's probably good. The editor in chief seems to have a lot of experience and has built a pretty large newsroom. It's definitely not separate in most people's minds (including mine), but I suppose that's just an ingrained bias in me from all the silly quizzes posted about on facebook. [NEWLINE] [NEWLINE] I can see your comparison with Reddit. It's a little different because Reddit isn't trying to pass itself of as a news site but as a content aggregator, but I'd be lying if I didn't say I get a good chunk of my news from here. Heck, sometimes if I'm lazy I don't even read articles and just get the news from random people on the internet in the comment section. And I suppose all the hard hitting, important news I find on reddit is sandwiched between cute cat photos, but I don't say it hurts journalism. I think people can keep the two separate even when it's all on the front page. [NEWLINE] [NEWLINE] I'd say I'm still not going over to BuzzFeed, and will still not hold it in much regard, but definitely view it in a different light now. Maybe after time I can come to see BuzzFeed news separately from the rest of the site. Thanks for giving me some good perspective!</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I certainly do not think that the fact that we know and love Spoiler Redacted makes<mask> happened to her somehow more horrifying than<mask> happened to Jeyne, or that Jeyne and Lollys were somehow "ok" to rape<mask> they're minor characters. [ENDQ] [NEWLINE] <mask> you might not<mask><mask> it's clear that others do.  There was no call to action for the rape of Lollys in the books or the show, and no call to action for the rape of Jeyne Poole<mask> it happened in the book. [NEWLINE] [NEWLINE] The whole talk of narrative and context just sounds *off* to me.  Rape, or any crime or tragedy that happens, always lacks context<mask> it happens in reality.  It always happens to someone who's story seemed like it was going to be some other story, and their lives before and after aren't an arc that sets up and then resolves the victimization in some appropriate way.  It just happens, and they deal with it (or not).  I'll never forget the story of<mask> one woman made her rapist breakfast ( [URL] /).  The story of her<mask> a victim just didn't make any sense, didn't fit,<mask> she changed it to another, more'suitable' story. [NEWLINE] [NEWLINE] <mask> sure,<mask><mask> it is really important that stories we tell about victimization should be well handled,<mask><mask><mask> that includes stories in which there's no setup, no resolution, no consequences for the perpetrator or the world at large,<mask> that's<mask> it happens in real life.</s>
Label encoding: <s> [STARTQ] I certainly do not think that the fact that we know and love Spoiler Redacted makes what happened to her somehow more horrifying than what happened to Jeyne, or that Jeyne and Lollys were somehow "ok" to rape because they're minor characters. [ENDQ] [NEWLINE] While you might not I think it's clear that others do.  There was no call to action for the rape of Lollys in the books or the show, and no call to action for the rape of Jeyne Poole when it happened in the book. [NEWLINE] [NEWLINE] The whole talk of narrative and context just sounds *off* to me.  Rape, or any crime or tragedy that happens, always lacks context when it happens in reality.  It always happens to someone who's story seemed like it was going to be some other story, and their lives before and after aren't an arc that sets up and then resolves the victimization in some appropriate way.  It just happens, and they deal with it (or not).  I'll never forget the story of how one woman made her rapist breakfast ( [URL] /).  The story of her as a victim just didn't make any sense, didn't fit, so she changed it to another, more'suitable' story. [NEWLINE] [NEWLINE] So sure, I think it is really important that stories we tell about victimization should be well handled, but I think that includes stories in which there's no setup, no resolution, no consequences for the perpetrator or the world at large, because that's how it happens in real life.</s>
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; = to you. Hi klw, thanks for diving a bit deeper into<mask> hyperrealism really is. I'll say that my view *has* changed based on discussion yesterday,<mask><mask> I would've responded to you before that happened is:<mask><mask><mask> that the process is certainly beneficial to the artist, it should be irrelevant to the viewer. In other words, pointless to the viewer, not the artist.<mask> I were to see the photograph of the red head, it would be about her.<mask><mask><mask><mask><mask> I see the hyperrealistic drawing of said red head, the art is no longer about her,<mask> about the process of having created the piece to begin with (the photograph could easily be adjusted to match the colors he used, and I would say the other changes he made don't change the subject matter substantially<mask><mask> ).<mask> I originally argued, I don't think it should be about the process,<mask> about the subject matter. [NEWLINE] [NEWLINE] Well consider my view changed.<mask><mask> now that art cannot be viewed without context,<mask> it's an essential part, whatever the context is.<mask><mask> I concede now that hyperrealism isn't pointless, I might still<mask><mask> it's *merely* a contemplation on technique (at least for the viewer;<mask> you say, it can server many purposes for the artist).<mask> such, I find it rather uninteresting, boring and unimaginative,<mask> that would simply fall in the realm of my personal opinions. [NEWLINE] [NEWLINE] Thanks for the reply!</s>
Label encoding: <s>&amp;#8710; = to you. Hi klw, thanks for diving a bit deeper into what hyperrealism really is. I'll say that my view *has* changed based on discussion yesterday, but what I would've responded to you before that happened is: while I agree that the process is certainly beneficial to the artist, it should be irrelevant to the viewer. In other words, pointless to the viewer, not the artist. If I were to see the photograph of the red head, it would be about her. If on the other hand I see the hyperrealistic drawing of said red head, the art is no longer about her, but about the process of having created the piece to begin with (the photograph could easily be adjusted to match the colors he used, and I would say the other changes he made don't change the subject matter substantially IMHO ). As I originally argued, I don't think it should be about the process, but about the subject matter. [NEWLINE] [NEWLINE] Well consider my view changed. I think now that art cannot be viewed without context, so it's an essential part, whatever the context is. So while I concede now that hyperrealism isn't pointless, I might still argue that it's *merely* a contemplation on technique (at least for the viewer; as you say, it can server many purposes for the artist). As such, I find it rather uninteresting, boring and unimaginative, but that would simply fall in the realm of my personal opinions. [NEWLINE] [NEWLINE] Thanks for the reply!</s>
Loss: tensor(0.0720, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Many rape victims report experiencing sexual pleasure and orgasm<mask><mask><mask> of bodily processes outside of their control. This does not mean that they are sexually attracted to their rapists. [ENDQ] [NEWLINE] You CMV, ∆.  I figured that satisfying sex (which would include orgasm) with either gender should mean that they're sexually attracted,<mask> this is an exception that totally turns this around.  Thank you very much. [NEWLINE] [NEWLINE] <mask><mask> my V has been C'd, could you tell me<mask> exactly sexual attraction is, and<mask> it occurs?  That would help me get an even better grip on<mask>, for example, a straight girl receiving oral sex from a girl would not be bisexual, even<mask> she enjoyed the event. [NEWLINE] [NEWLINE] [STARTQ] <mask> you experience romantic love differently, in that you romantically love those that you find physically unattractive, then you simply have a different sexual identity than I do. [ENDQ] [NEWLINE] <mask> I did not think that my current boyfriend was physically attractive at all before we dated, I was in denial of<mask> important physical attraction is to romantic love earlier. [NEWLINE] [NEWLINE] [STARTQ] You telling me, "No, you're not really straight, you do like members of the same sex, you just don't want to admit it" [ENDQ] [NEWLINE] I clarified this in other comments,<mask> it is not necessarily that you are actively denying it to yourself, just that you would be attracted to your undesired sex<mask> you grew up in an unbiased society. [NEWLINE] [NEWLINE] I can't respond to much else of<mask> you said,<mask> you changed by view about it!</s>
Label encoding: <s> [STARTQ] Many rape victims report experiencing sexual pleasure and orgasm as a result of bodily processes outside of their control. This does not mean that they are sexually attracted to their rapists. [ENDQ] [NEWLINE] You CMV, ∆.  I figured that satisfying sex (which would include orgasm) with either gender should mean that they're sexually attracted, but this is an exception that totally turns this around.  Thank you very much. [NEWLINE] [NEWLINE] Even though my V has been C'd, could you tell me what exactly sexual attraction is, and how it occurs?  That would help me get an even better grip on why, for example, a straight girl receiving oral sex from a girl would not be bisexual, even if she enjoyed the event. [NEWLINE] [NEWLINE] [STARTQ] If you experience romantic love differently, in that you romantically love those that you find physically unattractive, then you simply have a different sexual identity than I do. [ENDQ] [NEWLINE] Although I did not think that my current boyfriend was physically attractive at all before we dated, I was in denial of how important physical attraction is to romantic love earlier. [NEWLINE] [NEWLINE] [STARTQ] You telling me, "No, you're not really straight, you do like members of the same sex, you just don't want to admit it" [ENDQ] [NEWLINE] I clarified this in other comments, but it is not necessarily that you are actively denying it to yourself, just that you would be attracted to your undesired sex if you grew up in an unbiased society. [NEWLINE] [NEWLINE] I can't respond to much else of what you said, since you changed by view about it!</s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wrote themselves into a corner? Every scene in Game of Thrones is METICULOUSLY planned and something<mask> big<mask> Sansa marrying into the family that killed hers has to have been set in motion for a long time. [NEWLINE] [NEWLINE] <mask> there was no reaction to the rapes of Dany and Cercei. Do you think the character who told her own brother "Lay a hand on me again and you will no longer have a hand" would FALL IN LOVE with her rapist? Cercei destroys anyone who hurts her or even poses a threat<mask><mask> the one person she trusted the most brutally violated that trust, there was no change in the relationship. Furthermore, Drogo and Jaime are portrayed<mask> caring and loving with the women they raped and are painted in a sympathetic light for viewers. Every action in GOT has a reaction, all of the deaths and tortures leave lasting effects on the characters (Theon isn't the same person, Arya became a killer, Sansa grew up, Robb became a leader) unless we're talking about rape. Then it's never discussed or leaves an effect on the characters. [NEWLINE] [NEWLINE] Fans are pissed<mask> GOT has proved that they don't take rape seriously and<mask> should they expect any different for Sansa? [NEWLINE] [NEWLINE] <mask><mask><mask><mask> a minor character wouldn't show<mask> much of an effect from rape, their character development would be minimal either way, they're not a major player. [NEWLINE] [NEWLINE] <mask><mask> rape should be present in TV shows and definitely Game of Thrones. They just handle it<mask> poorly. </s>
Label encoding: <s>Wrote themselves into a corner? Every scene in Game of Thrones is METICULOUSLY planned and something as big as Sansa marrying into the family that killed hers has to have been set in motion for a long time. [NEWLINE] [NEWLINE] But there was no reaction to the rapes of Dany and Cercei. Do you think the character who told her own brother "Lay a hand on me again and you will no longer have a hand" would FALL IN LOVE with her rapist? Cercei destroys anyone who hurts her or even poses a threat but when the one person she trusted the most brutally violated that trust, there was no change in the relationship. Furthermore, Drogo and Jaime are portrayed as caring and loving with the women they raped and are painted in a sympathetic light for viewers. Every action in GOT has a reaction, all of the deaths and tortures leave lasting effects on the characters (Theon isn't the same person, Arya became a killer, Sansa grew up, Robb became a leader) unless we're talking about rape. Then it's never discussed or leaves an effect on the characters. [NEWLINE] [NEWLINE] Fans are pissed because GOT has proved that they don't take rape seriously and why should they expect any different for Sansa? [NEWLINE] [NEWLINE] As far as why a minor character wouldn't show as much of an effect from rape, their character development would be minimal either way, they're not a major player. [NEWLINE] [NEWLINE] I think rape should be present in TV shows and definitely Game of Thrones. They just handle it so poorly. </s>
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You seem to be ignoring<mask> the purpose of morality is (<mask> a trick that some species evolve to help them live in societies can be considered to have a "purpose", of course). [NEWLINE] [NEWLINE] We label certain behaviors with different labels<mask> they have different impacts on our ability to live together in societies. [NEWLINE] [NEWLINE] People that do "bad" things with common motives that are sometimes beneficial (e.g. greed can be quite beneficial<mask> people use it to motivate themselves to produce benefits for others in exchange for goods and services) get one label (in this case, let's call it "bad", or "wrong"). [NEWLINE] [NEWLINE] People that do even very similar things for uncommon motives that are likely to result in extraordinarily negative outcomes for society, without significant redeeming qualities get another label, in this case "evil". [NEWLINE] [NEWLINE] There's nothing metaphysical about evil. It's entirely a practical quality of morality that causes us to revile such actions and motives. We do it<mask> it advantages us to treat those things very negatively, in order to discourage them. [NEWLINE] [NEWLINE] The concept definitely exists, and it has great social benefits. It's utterly absurd to say that "evil doesn't exist", and in the long run, it would be socially harmful to adopt that view widely. The myths that we invent to rationalize this into something metaphysical have similar purposes,<mask> people don't just naturally fall into this rationalistic way of looking at the world. [NEWLINE] [NEWLINE] Instead, I propose you see evil for<mask> it is: harmful to our species' long term survival and success. </s><pad>
Label encoding: <s>You seem to be ignoring what the purpose of morality is ( if a trick that some species evolve to help them live in societies can be considered to have a "purpose", of course). [NEWLINE] [NEWLINE] We label certain behaviors with different labels because they have different impacts on our ability to live together in societies. [NEWLINE] [NEWLINE] People that do "bad" things with common motives that are sometimes beneficial (e.g. greed can be quite beneficial if people use it to motivate themselves to produce benefits for others in exchange for goods and services) get one label (in this case, let's call it "bad", or "wrong"). [NEWLINE] [NEWLINE] People that do even very similar things for uncommon motives that are likely to result in extraordinarily negative outcomes for society, without significant redeeming qualities get another label, in this case "evil". [NEWLINE] [NEWLINE] There's nothing metaphysical about evil. It's entirely a practical quality of morality that causes us to revile such actions and motives. We do it because it advantages us to treat those things very negatively, in order to discourage them. [NEWLINE] [NEWLINE] The concept definitely exists, and it has great social benefits. It's utterly absurd to say that "evil doesn't exist", and in the long run, it would be socially harmful to adopt that view widely. The myths that we invent to rationalize this into something metaphysical have similar purposes, because people don't just naturally fall into this rationalistic way of looking at the world. [NEWLINE] [NEWLINE] Instead, I propose you see evil for what it is: harmful to our species' long term survival and success. </s><pad>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is most of the trauma simply caused by peoples' reactions, rather than the crime itself? [ENDQ] [NEWLINE] It's both. The crime creates the first trauma. People's reactions to that crime recreate that trauma, plus multiply trauma on top of it. [NEWLINE] [NEWLINE] For example, imagine telling the survivor of an attempted homicide that they need to try strangulation again, with someone they love? It can feel wonderful, after all,<mask> the risks associated with it. [NEWLINE] [NEWLINE] Alternatively, imagine telling anyone who nearly died in a car crash, that wanting to drive again was proof they alone were responsible for the accident? Or that there was no accident at all? [NEWLINE] [NEWLINE] Both metaphors apply to the experience of being a rape victim. [NEWLINE] [NEWLINE] Then there's the way in which society in general handles sex. In some places, it's a taboo, which means it's everywhere,<mask> you're not allowed to speak of it. [NEWLINE] [NEWLINE] Imagine<mask> instead of sex, it was spiders. Imagine a world<mask> people hide spiders underneath their clothes. You can see them moving, sometimes. Sometimes, they want you to see them moving. [NEWLINE] [NEWLINE] Imagine the shadows of spiders sold perfume, or fashion, or cars. Imagine<mask> there were webs all over billboards and magazine stands, all to sell things to people who like spiders.<mask> you're not allowed to mention it.<mask> you do, it must mean you're obsessed with spiders. And<mask> you're obsessed with spiders, people can't be blamed<mask> their spiders bite you... [NEWLINE] [NEWLINE] Have any of these metaphors helped? </s>
Label encoding: <s> [STARTQ] Is most of the trauma simply caused by peoples' reactions, rather than the crime itself? [ENDQ] [NEWLINE] It's both. The crime creates the first trauma. People's reactions to that crime recreate that trauma, plus multiply trauma on top of it. [NEWLINE] [NEWLINE] For example, imagine telling the survivor of an attempted homicide that they need to try strangulation again, with someone they love? It can feel wonderful, after all, despite the risks associated with it. [NEWLINE] [NEWLINE] Alternatively, imagine telling anyone who nearly died in a car crash, that wanting to drive again was proof they alone were responsible for the accident? Or that there was no accident at all? [NEWLINE] [NEWLINE] Both metaphors apply to the experience of being a rape victim. [NEWLINE] [NEWLINE] Then there's the way in which society in general handles sex. In some places, it's a taboo, which means it's everywhere, but you're not allowed to speak of it. [NEWLINE] [NEWLINE] Imagine if instead of sex, it was spiders. Imagine a world where people hide spiders underneath their clothes. You can see them moving, sometimes. Sometimes, they want you to see them moving. [NEWLINE] [NEWLINE] Imagine the shadows of spiders sold perfume, or fashion, or cars. Imagine if there were webs all over billboards and magazine stands, all to sell things to people who like spiders. But you're not allowed to mention it. If you do, it must mean you're obsessed with spiders. And if you're obsessed with spiders, people can't be blamed if their spiders bite you... [NEWLINE] [NEWLINE] Have any of these metaphors helped? </s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I had the OP of that thread **tagged**<mask> an Islamophobic white nationalist before<mask> it sucks to see that actual bigots can get away with using the sub<mask> a political soapbox [ENDQ] [NEWLINE] I really hope you will reconsider your decision to use tags, they are inherently racist.  The beauty of a pseudo anonymous online discussion group is that ideas become separate from people.  Your ideas can stand or fall by themselves without inherent bias of race/religion/gender/political-identity getting mixed in. [NEWLINE] [NEWLINE] [STARTQ] the whole 4chan vs. Tumblr thing showed that there were a lot of 4chan.org/pol/ users there.. a place that's legitimately fucked up. [ENDQ] [NEWLINE] 4chan is ranked 475 on alexia for the US, that's higher than Jezebel.com  and Dailykos.com, it's safe to say people are aware of the site.  The 4chan vs Tumblr was a big thing on Tumblr<mask> it was on TIA.  The central joke about it,<mask> Tumblr was talking about it before hand, was that Tumblr was going to spam gay porn or kittens on 4chan.  4chan is already spammed everyday with gay porn and kittens; or at least it was a few years back<mask> I visited it. <mask> don't think that visiting /pol/ is the same<mask> supporting the views on /pol/;<mask> I used to go to /b/, it was for the same reason I go to TIA to see people say crazy stuff on the internet.</s>
Label encoding: <s> [STARTQ] I had the OP of that thread **tagged** as an Islamophobic white nationalist before so it sucks to see that actual bigots can get away with using the sub as a political soapbox [ENDQ] [NEWLINE] I really hope you will reconsider your decision to use tags, they are inherently racist.  The beauty of a pseudo anonymous online discussion group is that ideas become separate from people.  Your ideas can stand or fall by themselves without inherent bias of race/religion/gender/political-identity getting mixed in. [NEWLINE] [NEWLINE] [STARTQ] the whole 4chan vs. Tumblr thing showed that there were a lot of 4chan.org/pol/ users there.. a place that's legitimately fucked up. [ENDQ] [NEWLINE] 4chan is ranked 475 on alexia for the US, that's higher than Jezebel.com  and Dailykos.com, it's safe to say people are aware of the site.  The 4chan vs Tumblr was a big thing on Tumblr therefore it was on TIA.  The central joke about it, because Tumblr was talking about it before hand, was that Tumblr was going to spam gay porn or kittens on 4chan.  4chan is already spammed everyday with gay porn and kittens; or at least it was a few years back when I visited it.  Also don't think that visiting /pol/ is the same as supporting the views on /pol/; while I used to go to /b/, it was for the same reason I go to TIA to see people say crazy stuff on the internet.</s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is exactly<mask> we say you shouldn't have sex with drunk people... it's hard to tell<mask> they actually wanted sex or not.  Sometimes they go on auto pilot.  Sometimes they get scared of saying no and drunk brain says "it's safest to go along with it."  Sometimes you're drunk and they consented to makeouts and you don't notice they stopped going along with it<mask> it passed that point<mask> they were too drunk to say no.  You don't know.  That's<mask> drunk sex is like drunk driving... maybe something bad will happen, maybe not,<mask> there's not enough control to be sure.  Doing it a few times is risky and a bad plan.  Doing it constantly is the sexual equivalent of reckless endangerment. [NEWLINE] [NEWLINE] The general rubric is "afterwords, is one of the people involved undergoing rape trauma?  Is someone seriously hurt by<mask> happened"  Much like you don't know<mask> you're going to hit someone<mask> you drive drunk, you just have to look afterwords and see<mask> the car was damaged or<mask> there's injured bystanders. [NEWLINE] [NEWLINE] And yes, the whole "it must have been the man's fault" thing is fucked up<mask> well.  It's a separate issue... by default we assume it's the man's fault.  This is a serious problem that is not limited to sex<mask> drunk.  I've seen cases<mask> it's absolutely clear cut and absolutely the woman who did it and<mask> the man gets blamed.  It's horrific.</s>
Label encoding: <s>This is exactly why we say you shouldn't have sex with drunk people... it's hard to tell if they actually wanted sex or not.  Sometimes they go on auto pilot.  Sometimes they get scared of saying no and drunk brain says "it's safest to go along with it."  Sometimes you're drunk and they consented to makeouts and you don't notice they stopped going along with it when it passed that point but they were too drunk to say no.  You don't know.  That's why drunk sex is like drunk driving... maybe something bad will happen, maybe not, but there's not enough control to be sure.  Doing it a few times is risky and a bad plan.  Doing it constantly is the sexual equivalent of reckless endangerment. [NEWLINE] [NEWLINE] The general rubric is "afterwords, is one of the people involved undergoing rape trauma?  Is someone seriously hurt by what happened"  Much like you don't know if you're going to hit someone if you drive drunk, you just have to look afterwords and see if the car was damaged or if there's injured bystanders. [NEWLINE] [NEWLINE] And yes, the whole "it must have been the man's fault" thing is fucked up as well.  It's a separate issue... by default we assume it's the man's fault.  This is a serious problem that is not limited to sex while drunk.  I've seen cases where it's absolutely clear cut and absolutely the woman who did it and yet the man gets blamed.  It's horrific.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>,<mask>, the fetus is a person, then all of the pro-choice arguments become pathetic, and weaponized against them. After all,<mask> a women has the right to bodily autonomy, then that includes the female babies, and their right to not get killed.<mask> people should have the right to live their lives<mask> they want it, then surely that includes a fetuses right to live? Pregnancies can pose a health risk to the mother,<mask> abortions pose a hell of a bigger risk to the child, and<mask> on. [ENDQ] [NEWLINE] I'm only going to address this point,<mask> I do not entirely agree with your part about pro-life, either. [NEWLINE] [NEWLINE] I'm going to mainly address the following line: [NEWLINE] [NEWLINE] [STARTQ] <mask> people should have the right to live their lives<mask> they want it, then surely that includes a fetuses right to live? [ENDQ] [NEWLINE] The main issue with this is that I do not, and neither does anyone else, have the right to force you to provide parts of your body to keep myself alive. [NEWLINE] [NEWLINE] <mask> the argument is that a fetus is a person, and should have the same rights<mask> anyone else, I'm completely OK with that,<mask> do not give it the additional right to force someone else to keep it alive by using that person's body. [NEWLINE] [NEWLINE] <mask><mask>, I believe *this* is the only relevant part of the entire debate. [NEWLINE] [NEWLINE] I do not have the right to rig a woman up<mask> my own personal life support,<mask><mask> should a fetus?</s>
Label encoding: <s> [STARTQ] If, however, the fetus is a person, then all of the pro-choice arguments become pathetic, and weaponized against them. After all, if a women has the right to bodily autonomy, then that includes the female babies, and their right to not get killed. If people should have the right to live their lives how they want it, then surely that includes a fetuses right to live? Pregnancies can pose a health risk to the mother, but abortions pose a hell of a bigger risk to the child, and so on. [ENDQ] [NEWLINE] I'm only going to address this point, though I do not entirely agree with your part about pro-life, either. [NEWLINE] [NEWLINE] I'm going to mainly address the following line: [NEWLINE] [NEWLINE] [STARTQ] If people should have the right to live their lives how they want it, then surely that includes a fetuses right to live? [ENDQ] [NEWLINE] The main issue with this is that I do not, and neither does anyone else, have the right to force you to provide parts of your body to keep myself alive. [NEWLINE] [NEWLINE] If the argument is that a fetus is a person, and should have the same rights as anyone else, I'm completely OK with that, but do not give it the additional right to force someone else to keep it alive by using that person's body. [NEWLINE] [NEWLINE] In fact, I believe *this* is the only relevant part of the entire debate. [NEWLINE] [NEWLINE] I do not have the right to rig a woman up as my own personal life support, so why should a fetus?</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is a common practice in the introduction of legeslation to include completely unrelated issues in an attempt to get them passed<mask> part of the main issue. For example, one party might want to pass law A,<mask> not law B. A second party wants to pass lab B,<mask> not law A.<mask><mask> the first part introduces a bill on law A the second party will tack on law B in hopes that the first wants law A to pass<mask> badly that they'll be willing to allow law B to slip through. A recent example was anti abortion legislation being grouped in with motorcycle safety laws. [NEWLINE] [NEWLINE] I find this disruptive to the process. By forcing politicians to weigh the merits of completely unrelated issues against each other, the consideration given to individual issues is diluted and good laws get shot down to avoid the passing of bad, or vise versa. Change My View. [NEWLINE] [NEWLINE] **EDIT:** My view has been changed. Thanks to /u/thedeeno for pointing out that this is<mask> minority interests get their views noticed and pushed through<mask> little interest.<mask> to /u/floorberry for pointing out that this is a way to get around using a LOT of time on small and (to the greater body) unimportant topics such<mask> relatively small allocations for infrastructure upkeep. /u/treseritops made a similar point very well.<mask>, thanks to /u/auandi for pointing out that the House *already* has a rule against this,<mask> it's effectively castrated by the normal political structure. </s>
Label encoding: <s>It is a common practice in the introduction of legeslation to include completely unrelated issues in an attempt to get them passed as part of the main issue. For example, one party might want to pass law A, but not law B. A second party wants to pass lab B, but not law A. So when the first part introduces a bill on law A the second party will tack on law B in hopes that the first wants law A to pass so badly that they'll be willing to allow law B to slip through. A recent example was anti abortion legislation being grouped in with motorcycle safety laws. [NEWLINE] [NEWLINE] I find this disruptive to the process. By forcing politicians to weigh the merits of completely unrelated issues against each other, the consideration given to individual issues is diluted and good laws get shot down to avoid the passing of bad, or vise versa. Change My View. [NEWLINE] [NEWLINE] **EDIT:** My view has been changed. Thanks to /u/thedeeno for pointing out that this is how minority interests get their views noticed and pushed through despite little interest. Also to /u/floorberry for pointing out that this is a way to get around using a LOT of time on small and (to the greater body) unimportant topics such as relatively small allocations for infrastructure upkeep. /u/treseritops made a similar point very well. Also, thanks to /u/auandi for pointing out that the House *already* has a rule against this, but it's effectively castrated by the normal political structure. </s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>and [this thesis]( [URL].cgi?article=1240&amp;context=facpubs&amp;sei-redir=1&amp;referer=http%3A%2F%2Fscholar.google.com.au%2Fscholar_url%3Fhl%3Den%26q%3Dhttp%3A%2F%2Fdigitalcommons.law.msu.edu%2Fcgi%2Fviewcontent.cgi%253Farticle%253D1240%2526context%253Dfacpubs%26sa%3DX%26scisig%3DAAGBfm0R0nhptUXyYNOKVCS7ei_ZBYTnJw%26oi%3Dscholarr%26ei%3DGpWqUoeLBsnikAWj_oC4Bw%26ved%3D0CC4QgAMoATAA#search=%22http%3A%2F%2Fdigitalcommons.law.msu.edu%2Fcgi%2Fviewcontent.cgi%3Farticle%3D1240%26context%3Dfacpubs%22) shows that athletes grades are very often inflated to keep them enrolled and<mask> on the team. [NEWLINE] [NEWLINE] I didn't know that about the Big Ten schools sharing facilities<mask>,<mask> at least there is some proven academic benefit to sports programs.</s>
Label encoding: <s>and [this thesis]( [URL].cgi?article=1240&amp;context=facpubs&amp;sei-redir=1&amp;referer=http%3A%2F%2Fscholar.google.com.au%2Fscholar_url%3Fhl%3Den%26q%3Dhttp%3A%2F%2Fdigitalcommons.law.msu.edu%2Fcgi%2Fviewcontent.cgi%253Farticle%253D1240%2526context%253Dfacpubs%26sa%3DX%26scisig%3DAAGBfm0R0nhptUXyYNOKVCS7ei_ZBYTnJw%26oi%3Dscholarr%26ei%3DGpWqUoeLBsnikAWj_oC4Bw%26ved%3D0CC4QgAMoATAA#search=%22http%3A%2F%2Fdigitalcommons.law.msu.edu%2Fcgi%2Fviewcontent.cgi%3Farticle%3D1240%26context%3Dfacpubs%22) shows that athletes grades are very often inflated to keep them enrolled and therefore on the team. [NEWLINE] [NEWLINE] I didn't know that about the Big Ten schools sharing facilities though, so at least there is some proven academic benefit to sports programs.</s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] which in the end is the same money you'd be paying on rent. [ENDQ] [NEWLINE] Not really. A portion of your monthly mortgage payment goes toward interest,<mask> a portion goes toward the principal of the loan. This amount that goes toward the principal is still in your name and is a positive asset.<mask> you sell the house, whatever you've paid toward principal goes into your pocket<mask> cash. (Assuming no change in the value of the house). Payments are mostly interest with a bit of principal at the beginning of the mortgage,<mask> the ratio changes over time and by the end it's mostly principal with a bit of interest.<mask> you can pay extra toward the principal, that ratio changes quite a bit faster. [NEWLINE] [NEWLINE] On a 300k loan for 30 years at a fixed 4.5% interest rate, that means at the end you will have paid 247k in interest.<mask><mask> you'll be paying about 1500 a month, on average only 686 of that will be interest and the remaining 814 will be building your assets. [NEWLINE] [NEWLINE] The costs are fixed and can't be arbitrarily raised over that timespan by your landlord *and* you'll have 300k in assets to your name<mask> all is said and done.<mask> you're going to stay in the same place for a long time, buying is almost always much better financially. [NEWLINE] [NEWLINE] This glosses over some nuance - such<mask> transaction, insurance, and maintenance costs -<mask> these typically don't completely erase the financial advantage over the long term of buying rather then renting.</s>
Label encoding: <s> [STARTQ] which in the end is the same money you'd be paying on rent. [ENDQ] [NEWLINE] Not really. A portion of your monthly mortgage payment goes toward interest, but a portion goes toward the principal of the loan. This amount that goes toward the principal is still in your name and is a positive asset. If you sell the house, whatever you've paid toward principal goes into your pocket as cash. (Assuming no change in the value of the house). Payments are mostly interest with a bit of principal at the beginning of the mortgage, but the ratio changes over time and by the end it's mostly principal with a bit of interest. If you can pay extra toward the principal, that ratio changes quite a bit faster. [NEWLINE] [NEWLINE] On a 300k loan for 30 years at a fixed 4.5% interest rate, that means at the end you will have paid 247k in interest. Even though you'll be paying about 1500 a month, on average only 686 of that will be interest and the remaining 814 will be building your assets. [NEWLINE] [NEWLINE] The costs are fixed and can't be arbitrarily raised over that timespan by your landlord *and* you'll have 300k in assets to your name when all is said and done. If you're going to stay in the same place for a long time, buying is almost always much better financially. [NEWLINE] [NEWLINE] This glosses over some nuance - such as transaction, insurance, and maintenance costs - but these typically don't completely erase the financial advantage over the long term of buying rather then renting.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The primary problem with cherry-picking one's views from any given source of guidance is that it invalidates the original source<mask> a justification for holding that view. [NEWLINE] [NEWLINE] Let's say I cite the divinity of the Bible for justification that murder is wrong. God says<mask>,<mask> it must be true.<mask>, let's<mask> say that I love eating lobster and claim that that part of the bible "is a metaphor or something". *This invalidates the divinity of the Bible<mask> a justification for my views.* I can no longer claim that I believe that murder is wrong<mask> the Bible says<mask>,<mask> I've shown clear disregard for the actual infallible divinity of the book itself. [NEWLINE] [NEWLINE] Does this make murder not wrong,<mask>? No, definitely not. Does this even make any arguments written *in* the Bible invalid? No, it doesn't. All it means is that I cannot claim the Bible<mask> the justification for my belief. [NEWLINE] [NEWLINE] The problem here isn't the cherry-picking of views out of the Bible (or any other source). **The problem is the association of those views *with* the source and the use of the source<mask> justification for those views.** Are your views cherry-picked from a variety of different religious and political viewpoints? That's fine, have fun! Just don't go claiming those original sources<mask> reasons for you to hold a view unless you hold the original source<mask> valid in its entirety and actually believe "<mask> [the source] says<mask> " is a valid justification.</s>
Label encoding: <s>The primary problem with cherry-picking one's views from any given source of guidance is that it invalidates the original source as a justification for holding that view. [NEWLINE] [NEWLINE] Let's say I cite the divinity of the Bible for justification that murder is wrong. God says so, therefore it must be true. However, let's also say that I love eating lobster and claim that that part of the bible "is a metaphor or something". *This invalidates the divinity of the Bible as a justification for my views.* I can no longer claim that I believe that murder is wrong because the Bible says so, because I've shown clear disregard for the actual infallible divinity of the book itself. [NEWLINE] [NEWLINE] Does this make murder not wrong, however? No, definitely not. Does this even make any arguments written *in* the Bible invalid? No, it doesn't. All it means is that I cannot claim the Bible as the justification for my belief. [NEWLINE] [NEWLINE] The problem here isn't the cherry-picking of views out of the Bible (or any other source). **The problem is the association of those views *with* the source and the use of the source as justification for those views.** Are your views cherry-picked from a variety of different religious and political viewpoints? That's fine, have fun! Just don't go claiming those original sources as reasons for you to hold a view unless you hold the original source as valid in its entirety and actually believe " because [the source] says so " is a valid justification.</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My only issue with the rape scenes I'm aware of (haven't seen the last two seasons) is that they were consensual in the books--and turning them into rapes for the show completely changes the characters involved. [NEWLINE] [NEWLINE] The first was Drogo and Daenerys.  In the books,<mask> he could just have his way with her, Drogo actually seduces Daenerys, and the way it ultimately plays out does a lot to define both characters--it makes him more than a brutal barbarian king stereotype, and her more than a weak, passive victim or slave, and sets up their relationship in a way the show never does. [NEWLINE] [NEWLINE] Then, much later, the show has Jaime rape Cersei,<mask> in the book, Cersei is half into it and half doing it to help maintain her control over him, and certainly goes along with it.  Turning it into a rape makes her appear weak in a way the character in the book isn't until she's descended deeper into paranoia, and it makes Jaime seem brutal and cruel in a way he never is in the books, and generally isn't in the show. [NEWLINE] [NEWLINE] I don't object to depictions of rape<mask> it serves a narrative purpose--telling you something about the characters, the setting, the society, whatever. <mask> I feel like these two rapes were added to the series just for the sake of being edgy and maybe even precisely to generate this kind of controversy, and that they serve the narrative far better the way they play out in the books.</s>
Label encoding: <s>My only issue with the rape scenes I'm aware of (haven't seen the last two seasons) is that they were consensual in the books--and turning them into rapes for the show completely changes the characters involved. [NEWLINE] [NEWLINE] The first was Drogo and Daenerys.  In the books, although he could just have his way with her, Drogo actually seduces Daenerys, and the way it ultimately plays out does a lot to define both characters--it makes him more than a brutal barbarian king stereotype, and her more than a weak, passive victim or slave, and sets up their relationship in a way the show never does. [NEWLINE] [NEWLINE] Then, much later, the show has Jaime rape Cersei, where in the book, Cersei is half into it and half doing it to help maintain her control over him, and certainly goes along with it.  Turning it into a rape makes her appear weak in a way the character in the book isn't until she's descended deeper into paranoia, and it makes Jaime seem brutal and cruel in a way he never is in the books, and generally isn't in the show. [NEWLINE] [NEWLINE] I don't object to depictions of rape where it serves a narrative purpose--telling you something about the characters, the setting, the society, whatever.  But I feel like these two rapes were added to the series just for the sake of being edgy and maybe even precisely to generate this kind of controversy, and that they serve the narrative far better the way they play out in the books.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not sure<mask> you'll get this or<mask> you can reply<mask> reddit is going through maintenance randomly here. [NEWLINE] [NEWLINE] [STARTQ] is that non-child-related life goals are extrinsically motivated. [ENDQ] [NEWLINE] I'd be interested to hear<mask>,<mask> that would certainly be interesting and would put a lot of this into dispute. [NEWLINE] [NEWLINE] [STARTQ] I'd choose my career and life abroad. They were more fulfilling. [ENDQ] [NEWLINE] Again, I want to preface to say I'm not trying to be pokey for the sake of being a jerk. <mask> to clarify... you would<mask> you had a time machine, choose not to have your daughter. [NEWLINE] [NEWLINE] And there were no physical or mental issues to deal with, and your relationship is stable? [NEWLINE] [NEWLINE] <mask> it really is just about traveling more and your career? [NEWLINE] [NEWLINE] It's very surprising to me.  And like I said, I don't consider either choice a matter of selfishness,<mask> all choices are rationally selfish (including having children, including not having children, and even helping people in other countries.)  I understand both you and your husband are accomplished and have obtained several of your life goals.  I don't know<mask> you have one singular goal you're still chasing<mask>.  Is there a "unicorn goal" out there? [NEWLINE] [NEWLINE] Even<mask> you have a reason you think fits in with my attempts to control, I'd appreciate it.  Otherwise we'll go on your word and go from there (<mask> you're inclined to continue the debate).</s>
Label encoding: <s>Not sure if you'll get this or if you can reply since reddit is going through maintenance randomly here. [NEWLINE] [NEWLINE] [STARTQ] is that non-child-related life goals are extrinsically motivated. [ENDQ] [NEWLINE] I'd be interested to hear why, because that would certainly be interesting and would put a lot of this into dispute. [NEWLINE] [NEWLINE] [STARTQ] I'd choose my career and life abroad. They were more fulfilling. [ENDQ] [NEWLINE] Again, I want to preface to say I'm not trying to be pokey for the sake of being a jerk.  So to clarify... you would if you had a time machine, choose not to have your daughter. [NEWLINE] [NEWLINE] And there were no physical or mental issues to deal with, and your relationship is stable? [NEWLINE] [NEWLINE] So it really is just about traveling more and your career? [NEWLINE] [NEWLINE] It's very surprising to me.  And like I said, I don't consider either choice a matter of selfishness, since all choices are rationally selfish (including having children, including not having children, and even helping people in other countries.)  I understand both you and your husband are accomplished and have obtained several of your life goals.  I don't know if you have one singular goal you're still chasing though.  Is there a "unicorn goal" out there? [NEWLINE] [NEWLINE] Even if you have a reason you think fits in with my attempts to control, I'd appreciate it.  Otherwise we'll go on your word and go from there ( if you're inclined to continue the debate).</s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> constitutes "not even trying"? Fetal development is intimately related to parental health, which is a large aspect of medicine (read: all of it). [ENDQ] [NEWLINE] <mask> you're going to<mask><mask> research into the health of the mother equals research into the health of the fetus I'm not even gonna bother arguing. That's like arguing research into people's health is<mask> research into tapeworms' health. [NEWLINE] [NEWLINE] [STARTQ] There have<mask> been drug trials for miscarriage preventions. [ENDQ] [NEWLINE] I<mask> googled this and I have to tell you, almost all of the top results are drugs which CAUSE a miscarriage, or at least allow an already dead embryo to be expelled from the body faster. [NEWLINE] [NEWLINE] [STARTQ] Cowardice/laziness of others doesn't change the morality of an action. I wouldn't try to stop a murderer; I would call the cops. No one can do that for abortion. [ENDQ] [NEWLINE] You'd still be morally obligated to stop a murder yourself,<mask>. [NEWLINE] [NEWLINE] <mask><mask> the best way of putting it I've heard is this: suppose you work in an in-vitro fertilization clinic, and somehow the building gets set on fire. In one wing there's the lab with all the embryos; about a hundred of them. In the other wing is the break room, which has somebody's kid who they brought to work. You only have enough time to save one or the other. [NEWLINE] [NEWLINE] <mask>, which would you save, a hundred embryos or one child? </s>
Label encoding: <s> [STARTQ] What constitutes "not even trying"? Fetal development is intimately related to parental health, which is a large aspect of medicine (read: all of it). [ENDQ] [NEWLINE] If you're going to argue that research into the health of the mother equals research into the health of the fetus I'm not even gonna bother arguing. That's like arguing research into people's health is also research into tapeworms' health. [NEWLINE] [NEWLINE] [STARTQ] There have also been drug trials for miscarriage preventions. [ENDQ] [NEWLINE] I also googled this and I have to tell you, almost all of the top results are drugs which CAUSE a miscarriage, or at least allow an already dead embryo to be expelled from the body faster. [NEWLINE] [NEWLINE] [STARTQ] Cowardice/laziness of others doesn't change the morality of an action. I wouldn't try to stop a murderer; I would call the cops. No one can do that for abortion. [ENDQ] [NEWLINE] You'd still be morally obligated to stop a murder yourself, though. [NEWLINE] [NEWLINE] IMO the best way of putting it I've heard is this: suppose you work in an in-vitro fertilization clinic, and somehow the building gets set on fire. In one wing there's the lab with all the embryos; about a hundred of them. In the other wing is the break room, which has somebody's kid who they brought to work. You only have enough time to save one or the other. [NEWLINE] [NEWLINE] So, which would you save, a hundred embryos or one child? </s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are we conscious<mask> of our brain? [NEWLINE] [NEWLINE] This concept used to be written in stone. [NEWLINE] [NEWLINE] "You" is the information stored and processed in your neocortex.<mask> the neocortex is disabled, there is,<mask><mask><mask> early 21st century neuro-science can tell, *no way* you can think, dream, remember, observe or process anything. [NEWLINE] [NEWLINE] I used to be pretty convinced like you in the non existence of the "soul".<mask><mask><mask><mask><mask>, the term is<mask> tainted by religious connotations that I'd rather not use it. [NEWLINE] [NEWLINE] Then I started hearing some ER stories from...ER medical staff and patients, read a few, watched a few and had a few conversations with ER medical staff. [NEWLINE] [NEWLINE] A person that would have been declared "dead" 10-15 years ago in an Emergency Room can be reanimated today, relatively easily. [NEWLINE] [NEWLINE] Some come back with "stories". OK. Dreams are stories too. [NEWLINE] [NEWLINE] Some come back with stories<mask> their neocortex was flat ( due to for example bacterial meningitis, cerebrovascular accident etc). This is *impossible*. Let me emphasize this: it cannot be. [NEWLINE] [NEWLINE] <mask> it has been witnessed and documented by many medical facilities. [NEWLINE] [NEWLINE] <mask> are we really conscious<mask> of our brain or<mask> some argue *<mask> * out brains? [NEWLINE] [NEWLINE] There is for the first time a real discussion in the medical/scientific community about this even<mask> the subject seems to be still taboo. [NEWLINE] </s>
Label encoding: <s>Are we conscious because of our brain? [NEWLINE] [NEWLINE] This concept used to be written in stone. [NEWLINE] [NEWLINE] "You" is the information stored and processed in your neocortex. When the neocortex is disabled, there is, as far as early 21st century neuro-science can tell, *no way* you can think, dream, remember, observe or process anything. [NEWLINE] [NEWLINE] I used to be pretty convinced like you in the non existence of the "soul". As a matter of fact, the term is so tainted by religious connotations that I'd rather not use it. [NEWLINE] [NEWLINE] Then I started hearing some ER stories from...ER medical staff and patients, read a few, watched a few and had a few conversations with ER medical staff. [NEWLINE] [NEWLINE] A person that would have been declared "dead" 10-15 years ago in an Emergency Room can be reanimated today, relatively easily. [NEWLINE] [NEWLINE] Some come back with "stories". OK. Dreams are stories too. [NEWLINE] [NEWLINE] Some come back with stories as their neocortex was flat ( due to for example bacterial meningitis, cerebrovascular accident etc). This is *impossible*. Let me emphasize this: it cannot be. [NEWLINE] [NEWLINE] Yet it has been witnessed and documented by many medical facilities. [NEWLINE] [NEWLINE] So are we really conscious because of our brain or as some argue * despite * out brains? [NEWLINE] [NEWLINE] There is for the first time a real discussion in the medical/scientific community about this even if the subject seems to be still taboo. [NEWLINE] </s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No one should be punished for making a decision like that, it's harder to make then it looks.<mask> a right that I was born with, that can't be given or taken away by any government, is the right to my own body. I can decide to donate kidneys to keep someone alive, and I can<mask> refuse to donate kidneys to keep someone alive. Likewise<mask> a little think pops up in my stomach asking to use my organs to keep itself alive I can say no. [NEWLINE] [NEWLINE] This is one of those things that I don't care<mask> other peoples opinions are,<mask> I get pregnant and the thought of going through a pregnancy and birth disgust me, and I absolutely do not want to do it, I wont.<mask> the government made it illegal I would simply find a combination of drugs and alcohol that would do it for me. I wouldn't care,<mask> to hell<mask> I am going to subject my body to something I don't want,<mask> someone else whose life is not going to be affected *at all* gets squeamish. [NEWLINE] [NEWLINE] That is unless that person is willing to pay me enough money to cover **everything** through out a pregnancy, and a little bonus for going through labor, plus<mask> I'm not the type of person who would be able to give up any baby I gave birth to, would then be willing to pay for **everything** until that child turns 18, other then that<mask> it would be my life its affecting,<mask> I want an abortion, I'm having one.</s>
Label encoding: <s>No one should be punished for making a decision like that, it's harder to make then it looks. Also a right that I was born with, that can't be given or taken away by any government, is the right to my own body. I can decide to donate kidneys to keep someone alive, and I can also refuse to donate kidneys to keep someone alive. Likewise if a little think pops up in my stomach asking to use my organs to keep itself alive I can say no. [NEWLINE] [NEWLINE] This is one of those things that I don't care what other peoples opinions are, if I get pregnant and the thought of going through a pregnancy and birth disgust me, and I absolutely do not want to do it, I wont. If the government made it illegal I would simply find a combination of drugs and alcohol that would do it for me. I wouldn't care, because to hell if I am going to subject my body to something I don't want, because someone else whose life is not going to be affected *at all* gets squeamish. [NEWLINE] [NEWLINE] That is unless that person is willing to pay me enough money to cover **everything** through out a pregnancy, and a little bonus for going through labor, plus since I'm not the type of person who would be able to give up any baby I gave birth to, would then be willing to pay for **everything** until that child turns 18, other then that since it would be my life its affecting, if I want an abortion, I'm having one.</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The first two paragraphs are simply to set up the idea that "truth" is only one consideration you have to make<mask> considering<mask> to say. I would<mask><mask> say you should lie in both situations. This is, of course, opposed to OP's view<mask> this is r/changemyview, not r/agreewithme. [NEWLINE] [NEWLINE] In the second, I'd like you to consider your professional area of expertise. I'd then like to ask you to think about<mask> to convey the most complex idea in it you understand- with the catch that you have to convey it with no mistakes in 30 seconds. Unless you picked a very simple idea, you are going to have to leave out things, and explain inaccurately. Take Newtonian physics, for example. We teach that they are correct in high school. They aren't- they're approximations. Nonetheless, you do not explain quantum mechanics to a fourteen year old right off the bat,<mask> it won't actually aid understanding. [NEWLINE] [NEWLINE] <mask> considering explaining things to adults, you must realistically consider<mask> much time they're going to spend thinking about<mask> you say, and<mask> intelligent and/or open-minded the average person is. There's a reason all the logical arguments in the world didn't have much effect on long-term opinions about gay marriage, [<mask> door-to-door campaigning<mask> people simply met gay couples did.]( [URL] ) People are not logic machines. Everything you say conveys a message- shape it wisely. </s>
Label encoding: <s>The first two paragraphs are simply to set up the idea that "truth" is only one consideration you have to make when considering what to say. I would in fact say you should lie in both situations. This is, of course, opposed to OP's view because this is r/changemyview, not r/agreewithme. [NEWLINE] [NEWLINE] In the second, I'd like you to consider your professional area of expertise. I'd then like to ask you to think about how to convey the most complex idea in it you understand- with the catch that you have to convey it with no mistakes in 30 seconds. Unless you picked a very simple idea, you are going to have to leave out things, and explain inaccurately. Take Newtonian physics, for example. We teach that they are correct in high school. They aren't- they're approximations. Nonetheless, you do not explain quantum mechanics to a fourteen year old right off the bat, because it won't actually aid understanding. [NEWLINE] [NEWLINE] When considering explaining things to adults, you must realistically consider how much time they're going to spend thinking about what you say, and how intelligent and/or open-minded the average person is. There's a reason all the logical arguments in the world didn't have much effect on long-term opinions about gay marriage, [ but door-to-door campaigning where people simply met gay couples did.]( [URL] ) People are not logic machines. Everything you say conveys a message- shape it wisely. </s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're not miscalculating the danger,<mask> there is none,<mask> you're madly miscalculating the benefit. [NEWLINE] [NEWLINE] You are suggesting that people would be more open and more accepting of each other<mask> only people interacted with more random people like they used to. [NEWLINE] [NEWLINE] That suggests that back<mask> people did tend to interact with more random people, they were more open and accepting of each other. The evidence does not support your idea.<mask> all we needed was to go back to that to get a new and never-before-seen-in-humanity level of public understanding, then we would have had that understanding back the first time we had more public discourse. [NEWLINE] [NEWLINE] You're<mask> overestimating the benefit people individually get. You figure that<mask> someone meets a woman or a black person and sees that they are actually an interesting complex person, they will realize that they are mischaracterizing women or black people and their mind will open. [NEWLINE] <mask> tends to actually happen is that they note "wow, this female/black person is not like most female/black people", and mark them<mask> an anomaly - per the common terribly mistaken phrase "I don't think of you<mask> black". [NEWLINE] They don't tend to actually change the general prejudice they were carrying. [NEWLINE] [NEWLINE] Again, you can see this in history. [NEWLINE] [NEWLINE] Don't forget that those Athenians discoursing in the public marketplace held slaves, and that<mask> all citizens got to vote, few residents got to be citizens.  </s>
Label encoding: <s>You're not miscalculating the danger, as there is none, but you're madly miscalculating the benefit. [NEWLINE] [NEWLINE] You are suggesting that people would be more open and more accepting of each other if only people interacted with more random people like they used to. [NEWLINE] [NEWLINE] That suggests that back when people did tend to interact with more random people, they were more open and accepting of each other. The evidence does not support your idea. If all we needed was to go back to that to get a new and never-before-seen-in-humanity level of public understanding, then we would have had that understanding back the first time we had more public discourse. [NEWLINE] [NEWLINE] You're also overestimating the benefit people individually get. You figure that if someone meets a woman or a black person and sees that they are actually an interesting complex person, they will realize that they are mischaracterizing women or black people and their mind will open. [NEWLINE] What tends to actually happen is that they note "wow, this female/black person is not like most female/black people", and mark them as an anomaly - per the common terribly mistaken phrase "I don't think of you as black". [NEWLINE] They don't tend to actually change the general prejudice they were carrying. [NEWLINE] [NEWLINE] Again, you can see this in history. [NEWLINE] [NEWLINE] Don't forget that those Athenians discoursing in the public marketplace held slaves, and that while all citizens got to vote, few residents got to be citizens.  </s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It seems, in that last bit, like you want bisexuality to be a person being equally attracted to both sexes. It's not necessarily that at all. It just means that you have *some* sexual attraction to each sex. Your attraction to one or the other could be significantly stronger, or one could be aromantic and the other could be romantic, or any number of other variations and nuances. That a woman's sexual interest in other women comes<mask> a peripheral to her interest in men doesn't necessarily mean she's not "really" bisexual.<mask> someone who isn't them I don't see that you really have much business declaring that their proclaimed sexuality is false. [NEWLINE] [NEWLINE] And<mask><mask><mask> anti-gay speech goes, that gets thrown at us just<mask> readily, especially<mask> you're a guy. Aside from most queer folks and a few allies, society is just going to see you<mask> gay<mask> you're male. It doesn't matter that you have sex or relationships with women, closeted men have sex and relationships with women all the time. And<mask> your mannerisms, speech patterns, or interests aren't exactly typically male, forget about it. We get all the bullshit supposedly designated for gay men plus a load of bullshit of our own. There are no bi-specific insults<mask> nobody who says that shit cares enough to get it right. They can just<mask> easily call us faggots too, it's not the having sex with or being attracted to women part they take issue with anyway. </s>
Label encoding: <s>It seems, in that last bit, like you want bisexuality to be a person being equally attracted to both sexes. It's not necessarily that at all. It just means that you have *some* sexual attraction to each sex. Your attraction to one or the other could be significantly stronger, or one could be aromantic and the other could be romantic, or any number of other variations and nuances. That a woman's sexual interest in other women comes as a peripheral to her interest in men doesn't necessarily mean she's not "really" bisexual. As someone who isn't them I don't see that you really have much business declaring that their proclaimed sexuality is false. [NEWLINE] [NEWLINE] And as far as anti-gay speech goes, that gets thrown at us just as readily, especially if you're a guy. Aside from most queer folks and a few allies, society is just going to see you as gay if you're male. It doesn't matter that you have sex or relationships with women, closeted men have sex and relationships with women all the time. And if your mannerisms, speech patterns, or interests aren't exactly typically male, forget about it. We get all the bullshit supposedly designated for gay men plus a load of bullshit of our own. There are no bi-specific insults because nobody who says that shit cares enough to get it right. They can just as easily call us faggots too, it's not the having sex with or being attracted to women part they take issue with anyway. </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How is treating every woman like a past sexual assault victim fair to the women who like it<mask> a man puts his hand on the small of her back<mask> they're talking in a bar? Who thinks a man who doesn't have chronic hover hand or can actually read signals is sexy? [NEWLINE] [NEWLINE] I want a woman to tell me<mask> she wants. You just have to do it then, not 20 minutes after the fact.<mask> men walk up and start talking to you and try to make that first light contact they're communicating their interest. Its always better<mask> the woman communicates back.<mask> she's interested<mask> doesn't want the physical touch indicators take his hand away and call him a cheeky bastard. Keep it playful. [NEWLINE] [NEWLINE] Remember, communication, especially in the moment, is a two way street and we don't read minds. [NEWLINE] [NEWLINE] Stand around any bar and watch the Hookup Culture in action. The freedom of sexuality that was fought for is here, and it's reflected in every club and bar after 11 pm. Game and it's evolution were a natural result of the Hookup Culture, just like Chivalry was a natural result of requiring the parent's permission before marrying a girl. Sexual strategies will always exist. This one hasn't reached its final form<mask>,<mask> with enough refinement like our interactions here, I believe we can reach the peak of safety and fulfilling sexuality. Thank you for engaging with me openly about this. **You have made a difference.** [NEWLINE] [NEWLINE] Edit: Formatting</s>
Label encoding: <s>How is treating every woman like a past sexual assault victim fair to the women who like it when a man puts his hand on the small of her back while they're talking in a bar? Who thinks a man who doesn't have chronic hover hand or can actually read signals is sexy? [NEWLINE] [NEWLINE] I want a woman to tell me what she wants. You just have to do it then, not 20 minutes after the fact. When men walk up and start talking to you and try to make that first light contact they're communicating their interest. Its always better when the woman communicates back. If she's interested but doesn't want the physical touch indicators take his hand away and call him a cheeky bastard. Keep it playful. [NEWLINE] [NEWLINE] Remember, communication, especially in the moment, is a two way street and we don't read minds. [NEWLINE] [NEWLINE] Stand around any bar and watch the Hookup Culture in action. The freedom of sexuality that was fought for is here, and it's reflected in every club and bar after 11 pm. Game and it's evolution were a natural result of the Hookup Culture, just like Chivalry was a natural result of requiring the parent's permission before marrying a girl. Sexual strategies will always exist. This one hasn't reached its final form yet, but with enough refinement like our interactions here, I believe we can reach the peak of safety and fulfilling sexuality. Thank you for engaging with me openly about this. **You have made a difference.** [NEWLINE] [NEWLINE] Edit: Formatting</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We're not simple animals. We can't point to our animal nature in defense of beastly decisions. I might have teeth and nails and testosterone<mask> I can still be charged with murder. [NEWLINE] [NEWLINE] [STARTQ] our species is relying on it to exist [ENDQ] [NEWLINE] we're not talking about the species, we're talking about civilisation. And with 7 billion people on the planet we don't seem to be at risk of extinction due to child support issues,<mask> I'm unclear<mask> you're attempting to say. [NEWLINE] [NEWLINE] Perhaps you're referencing something like the [social contract]( [URL] )? Well of-course yes - everybody contributes, childless or otherwise,<mask> we don't contribute to your children, we contribute to a society<mask> your children can be raised with freedom. [NEWLINE] [NEWLINE] I'm not paying taxes to keep your kid in school, I'm paying taxes<mask> society demands an education system (one that I myself enjoyed), and one I continue to enjoy, by interacting with other educated people. I don't care (nor am I required to) about *your* kids... I can't<mask> I have nothing<mask> a fleeting relationship with them -<mask> can I be asked to care about someone I've never met? -<mask> I do care about the society we both live in: I do and must contribute indirectly,<mask> not to raising any one else's kids any more than I do another's health care, I contribute to a society<mask> these values are held high,<mask> I choose to live in that society.</s>
Label encoding: <s>We're not simple animals. We can't point to our animal nature in defense of beastly decisions. I might have teeth and nails and testosterone but I can still be charged with murder. [NEWLINE] [NEWLINE] [STARTQ] our species is relying on it to exist [ENDQ] [NEWLINE] we're not talking about the species, we're talking about civilisation. And with 7 billion people on the planet we don't seem to be at risk of extinction due to child support issues, so I'm unclear what you're attempting to say. [NEWLINE] [NEWLINE] Perhaps you're referencing something like the [social contract]( [URL] )? Well of-course yes - everybody contributes, childless or otherwise, but we don't contribute to your children, we contribute to a society where your children can be raised with freedom. [NEWLINE] [NEWLINE] I'm not paying taxes to keep your kid in school, I'm paying taxes because society demands an education system (one that I myself enjoyed), and one I continue to enjoy, by interacting with other educated people. I don't care (nor am I required to) about *your* kids... I can't because I have nothing but a fleeting relationship with them - how can I be asked to care about someone I've never met? - but I do care about the society we both live in: I do and must contribute indirectly, but not to raising any one else's kids any more than I do another's health care, I contribute to a society where these values are held high, because I choose to live in that society.</s>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Edit: YOU HAVE CHANGED MY VIEW. (Only regarding hunting for food) [NEWLINE] [NEWLINE] This isn't about Cecil the lion<mask> it did get me thinking.....people who hunt must be at least a bit sadistic. Hear me out - I'm not here to make a dig at hunters. I don't eat meat myself<mask> I actually believe hunting game<mask> a source of meat is much more ethical than eating mass produced factory farmed animals. Game animals have at least had a natural life without chemicals, genetic engineering and is the ultimate in animal welfare. You can't beat having the life nature dictates.<mask> I'm in conflict with<mask> is actually involved in hunting. [NEWLINE] Lets take an example of a hunting a deer. A healthy meat with a plentiful population to chose from.<mask> you go out into its habitat, learn its habits, track it and then finally you see it. Quietly in its happily eating some food. Minding its own business. Full of life and<mask> majestic in its natural habitat. This stunning beast is a pleasure to see.  I believe many hunters genuinely have a lot of respect for the animals they hunt.<mask> then full of all that respect BOOM! They take its life. Just like that. All that effort to understand it, follow it and watch it they extinguished  in a second.  Cold hard logic can<mask> justify<mask> the animal is killed<mask> deep down the pleasure hunters feel<mask> they've killed an animal doesn't make sense to me. [NEWLINE] </s>
Label encoding: <s>Edit: YOU HAVE CHANGED MY VIEW. (Only regarding hunting for food) [NEWLINE] [NEWLINE] This isn't about Cecil the lion but it did get me thinking.....people who hunt must be at least a bit sadistic. Hear me out - I'm not here to make a dig at hunters. I don't eat meat myself but I actually believe hunting game as a source of meat is much more ethical than eating mass produced factory farmed animals. Game animals have at least had a natural life without chemicals, genetic engineering and is the ultimate in animal welfare. You can't beat having the life nature dictates. So I'm in conflict with what is actually involved in hunting. [NEWLINE] Lets take an example of a hunting a deer. A healthy meat with a plentiful population to chose from. So you go out into its habitat, learn its habits, track it and then finally you see it. Quietly in its happily eating some food. Minding its own business. Full of life and so majestic in its natural habitat. This stunning beast is a pleasure to see.  I believe many hunters genuinely have a lot of respect for the animals they hunt. But then full of all that respect BOOM! They take its life. Just like that. All that effort to understand it, follow it and watch it they extinguished  in a second.  Cold hard logic can indeed justify why the animal is killed but deep down the pleasure hunters feel when they've killed an animal doesn't make sense to me. [NEWLINE] </s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It would be more accurate to say that every person has the potential to be valuable to society. From that, OP argues that society needs to enable those that are throwing away their potential. I don't believe that the homeless and unemployed have a great incentive to contribute to society.<mask><mask> that they have a lot to gain,<mask><mask> you're out on the street, your priorities shift drastically. Imagine not knowing<mask> your next meal will be. This is now your primary concern, and the hiring process is too long to provide you with the money to get food by tonight. [NEWLINE] [NEWLINE] Let's say you're one of those that are informed of the services available in your community. You know<mask> the soup kitchens are,<mask> you're getting two meals a day.<mask><mask> about shelter?<mask> will you sleep? The nearby homeless shelter has a lengthy application process which is partial to women and children,<mask><mask> a single man, you aren't hopeful. Next, clothing. You'll need nice clothes<mask> you want to go in for any interviews,<mask> do you really want to spend that money? Bare necessities are more pressing concerns, and hell, a job would be a waste of time anyway. Working 8 hours a day<mask> you only get paid every two weeks? That's time better spent scrounging. [NEWLINE] [NEWLINE] Anyway, my point is that the homeless have a lot to gain by working,<mask> they're too concerned with their immediate situation to be able to do anything about it. </s>
Label encoding: <s>It would be more accurate to say that every person has the potential to be valuable to society. From that, OP argues that society needs to enable those that are throwing away their potential. I don't believe that the homeless and unemployed have a great incentive to contribute to society. I agree that they have a lot to gain, but when you're out on the street, your priorities shift drastically. Imagine not knowing when your next meal will be. This is now your primary concern, and the hiring process is too long to provide you with the money to get food by tonight. [NEWLINE] [NEWLINE] Let's say you're one of those that are informed of the services available in your community. You know where the soup kitchens are, so you're getting two meals a day. So what about shelter? Where will you sleep? The nearby homeless shelter has a lengthy application process which is partial to women and children, so as a single man, you aren't hopeful. Next, clothing. You'll need nice clothes if you want to go in for any interviews, but do you really want to spend that money? Bare necessities are more pressing concerns, and hell, a job would be a waste of time anyway. Working 8 hours a day when you only get paid every two weeks? That's time better spent scrounging. [NEWLINE] [NEWLINE] Anyway, my point is that the homeless have a lot to gain by working, but they're too concerned with their immediate situation to be able to do anything about it. </s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. I am fine with them being run for profit,<mask> I wouldn't say I DO want it. Maybe they could be run<mask> public utilities, with a limited profit or such,<mask> I'm not sure exactly<mask> I want them to do. [NEWLINE] [NEWLINE] 2. I guess. I'd prefer<mask> the government ran it. [NEWLINE] [NEWLINE] 3. Due diligence would be good. [NEWLINE] [NEWLINE] 4. No,<mask> they can't afford their debts then they should be reduced.<mask> they can afford their debts then they can't opt out. [NEWLINE] [NEWLINE] 5. Yes,<mask> for secured loans I'd be fine with banks retaining a stake in the money from selling the possessions. This would only affect people who couldn't afford their loans. [NEWLINE] [NEWLINE] [STARTQ] <mask> you tell people that they'll keep their homes/cars/whatevers, even<mask> they can't pay off the loan, you will be incentivizing them to take out way, way more risky loans,<mask> they know they aren't bearing the full costs of their actions. [ENDQ] [NEWLINE] Do you have a study that shows this would happen, and isn't currently happening? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> the people taking the loans out aren't taking the losses, who is? Either the banks (which means they'll just stop the practice altogether) or the government (higher taxes/less money around for welfare programs/something else?). [ENDQ] [NEWLINE] I'd still be fine with them making loans to people who could actually afford to pay their debts.</s>
Label encoding: <s>1. I am fine with them being run for profit, but I wouldn't say I DO want it. Maybe they could be run as public utilities, with a limited profit or such, but I'm not sure exactly what I want them to do. [NEWLINE] [NEWLINE] 2. I guess. I'd prefer if the government ran it. [NEWLINE] [NEWLINE] 3. Due diligence would be good. [NEWLINE] [NEWLINE] 4. No, if they can't afford their debts then they should be reduced. If they can afford their debts then they can't opt out. [NEWLINE] [NEWLINE] 5. Yes, though for secured loans I'd be fine with banks retaining a stake in the money from selling the possessions. This would only affect people who couldn't afford their loans. [NEWLINE] [NEWLINE] [STARTQ] If you tell people that they'll keep their homes/cars/whatevers, even if they can't pay off the loan, you will be incentivizing them to take out way, way more risky loans, since they know they aren't bearing the full costs of their actions. [ENDQ] [NEWLINE] Do you have a study that shows this would happen, and isn't currently happening? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] If the people taking the loans out aren't taking the losses, who is? Either the banks (which means they'll just stop the practice altogether) or the government (higher taxes/less money around for welfare programs/something else?). [ENDQ] [NEWLINE] I'd still be fine with them making loans to people who could actually afford to pay their debts.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This could be region specific,<mask> folks around<mask> I live (Northern Suburbs of the City of Atlanta) stop and chat with one another all the time. It's difficult to walk on the local college campus without being dragged into some conversation or other. That's actually a problem, normally<mask> you're walking you are *going somewhere* and *need to be there at a specific time*. Getting drawn into a conversation that requires significant time to come to a bigger realization and a better understanding of the world isn't necessarily the best idea in those situations. After all, you have to choose between meeting your engagements (getting to class/work/parties on time) or having a chat. The same sort of thing can happen (<mask> happens less often) on Kennesaw Main Street or the Marietta Square.<mask> someone who has been there, chatting with people in public can become very expensive rather quickly in terms of time and effort. Is it worth the benefits for the costs in time and effort? That's a the kind of thing<mask> each person has a different answer. It's worth it for some folks and not worth it to others. [NEWLINE] [NEWLINE] Last year I was up in Connecticut to see my sister, and I'm used to small talk on the street and such. Saying "hello" and "have a nice day" are essentially involuntary responses for me at this point. People looked at me like I was crazy.<mask>, it was definitely a local culture specific thing.</s>
Label encoding: <s>This could be region specific, but folks around where I live (Northern Suburbs of the City of Atlanta) stop and chat with one another all the time. It's difficult to walk on the local college campus without being dragged into some conversation or other. That's actually a problem, normally when you're walking you are *going somewhere* and *need to be there at a specific time*. Getting drawn into a conversation that requires significant time to come to a bigger realization and a better understanding of the world isn't necessarily the best idea in those situations. After all, you have to choose between meeting your engagements (getting to class/work/parties on time) or having a chat. The same sort of thing can happen ( but happens less often) on Kennesaw Main Street or the Marietta Square. As someone who has been there, chatting with people in public can become very expensive rather quickly in terms of time and effort. Is it worth the benefits for the costs in time and effort? That's a the kind of thing where each person has a different answer. It's worth it for some folks and not worth it to others. [NEWLINE] [NEWLINE] Last year I was up in Connecticut to see my sister, and I'm used to small talk on the street and such. Saying "hello" and "have a nice day" are essentially involuntary responses for me at this point. People looked at me like I was crazy. So, it was definitely a local culture specific thing.</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This really isn't medically accurate<mask>. There are all sorts of congenital intersex conditions. Take congenital androgen insensitivity syndrome,<mask> someone who is genetically male will develop essentially<mask> a normal female (except be unable to reproduce). [ENDQ] [NEWLINE] That makes them a male with a congenital defect, not a female. It doesn't change their gender anymore than Vitiligo changes someone's race. [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It's a very forced choice one has to make then. Choosing<mask> to respond behaviorally and medically to profound physical or hormonal abnormalities is a precondition to living a tolerably normal life, not a lifestyle decision. [ENDQ] [NEWLINE] Societal norms aren't Divine Mandate.<mask> you're talking about is someone making a choice to fit into society, not adherence to God's will, and that's perfectly fine- people are free to live<mask> they choose,<mask> again, it doesn't make it any less of a sin. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [STARTQ] One could<mask><mask> in cases of profound gender dysphoria, there's a disconnect between the way the brain and body have developed. You can even experimentally create gender dysphoria and doctors used to do it routinely in cases<mask> babies were born with ambiguous genitalia by arbitrarily assigning them a gender with disastrous consequences. [ENDQ] [NEWLINE] [NEWLINE] I'll ask<mask> again, "are you classifying someone being Transgender<mask> having a birth defect"?</s>
Label encoding: <s> [STARTQ] This really isn't medically accurate though. There are all sorts of congenital intersex conditions. Take congenital androgen insensitivity syndrome, where someone who is genetically male will develop essentially as a normal female (except be unable to reproduce). [ENDQ] [NEWLINE] That makes them a male with a congenital defect, not a female. It doesn't change their gender anymore than Vitiligo changes someone's race. [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It's a very forced choice one has to make then. Choosing how to respond behaviorally and medically to profound physical or hormonal abnormalities is a precondition to living a tolerably normal life, not a lifestyle decision. [ENDQ] [NEWLINE] Societal norms aren't Divine Mandate. What you're talking about is someone making a choice to fit into society, not adherence to God's will, and that's perfectly fine- people are free to live however they choose, but again, it doesn't make it any less of a sin. [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] [STARTQ] One could argue that in cases of profound gender dysphoria, there's a disconnect between the way the brain and body have developed. You can even experimentally create gender dysphoria and doctors used to do it routinely in cases where babies were born with ambiguous genitalia by arbitrarily assigning them a gender with disastrous consequences. [ENDQ] [NEWLINE] [NEWLINE] I'll ask yet again, "are you classifying someone being Transgender as having a birth defect"?</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't understand your reasoning. It would have been perfectly okay to refuse you service,<mask> not okay to provide you limited service? That doesn't make any sense. [ENDQ] [NEWLINE] Would you go buy a product off a shelf that was only half there? [NEWLINE] [NEWLINE] <mask> they weren't willing, or were unable, to provide the service I was looking for, they could've said<mask>.<mask><mask>,<mask> they *did* say<mask>, I was perfectly willing to go somewhere else. The manager chose to intervene and say that the whole menu would be available. [NEWLINE] [NEWLINE] I don't see the problem. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] And unless it was prix fixe, there is no reason for you to expect a discount on a partial menu. Do you expect a discount<mask> a restaurant runs out of an item? [ENDQ] [NEWLINE] <mask> you call Store X, and ask them: Do you have the new iWidget 6 in stock, and they say yes. Then you drive there and ask again: Do you have the new iWidget 6 in stock, and they say yes again.<mask> you tell them I want an iWidget 6, and they tell you: "We'll it's late, and we've already got the iWidget 6's put up for the night,<mask> you can have an iWidget 4 for the same price", you're telling my you'd be okay with that? [NEWLINE] [NEWLINE] I don't think<mask>. </s>
Label encoding: <s> [STARTQ] I don't understand your reasoning. It would have been perfectly okay to refuse you service, but not okay to provide you limited service? That doesn't make any sense. [ENDQ] [NEWLINE] Would you go buy a product off a shelf that was only half there? [NEWLINE] [NEWLINE] If they weren't willing, or were unable, to provide the service I was looking for, they could've said so. In fact, when they *did* say so, I was perfectly willing to go somewhere else. The manager chose to intervene and say that the whole menu would be available. [NEWLINE] [NEWLINE] I don't see the problem. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [STARTQ] And unless it was prix fixe, there is no reason for you to expect a discount on a partial menu. Do you expect a discount when a restaurant runs out of an item? [ENDQ] [NEWLINE] If you call Store X, and ask them: Do you have the new iWidget 6 in stock, and they say yes. Then you drive there and ask again: Do you have the new iWidget 6 in stock, and they say yes again. So you tell them I want an iWidget 6, and they tell you: "We'll it's late, and we've already got the iWidget 6's put up for the night, but you can have an iWidget 4 for the same price", you're telling my you'd be okay with that? [NEWLINE] [NEWLINE] I don't think so. </s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Life, is going to kill a 7,046,000,000 (billions) people or more, and that's<mask> there are no more new people, then that numbers rises, <mask> exterminating any fraction of the population, up to and including 100% should be "ethical," and for any arbitrary reason. <mask>, now, not only is lying ethical,<mask> mass murder is too? [NEWLINE] [NEWLINE] <mask> about, I change your view, that lying and murder are both unethical.  That <mask>  makes the noble lie dangerous is that it first  supposes a superior or  noble  status,<mask> the person being lied to is inferior, and does not deserve or can not handle truths.   Second, it is a tool for tyranny upon the those "inferiors."   Always has been, always will be. [NEWLINE] [NEWLINE] Instead of trying to save the world, <mask> about you regain your integrity, and make amends to your friends you have lied to? [NEWLINE] [NEWLINE] <mask>  can you save the world,<mask> you  succumb to corruption and disintegration?   Unless perhaps you envision yourself a martyr, a willing sacrifice to the world.  I would like to change your view on that,<mask> that is the case, that the world is not worth your loss.  You should save it for something important. [NEWLINE] [NEWLINE] <mask>, I'd like to leave you with a short video.  [Saving the planet]( [URL] )</s>
Label encoding: <s>Life, is going to kill a 7,046,000,000 (billions) people or more, and that's if there are no more new people, then that numbers rises,  therefore exterminating any fraction of the population, up to and including 100% should be "ethical," and for any arbitrary reason.  So, now, not only is lying ethical, but mass murder is too? [NEWLINE] [NEWLINE] How about, I change your view, that lying and murder are both unethical.  That  what  makes the noble lie dangerous is that it first  supposes a superior or  noble  status, while the person being lied to is inferior, and does not deserve or can not handle truths.   Second, it is a tool for tyranny upon the those "inferiors."   Always has been, always will be. [NEWLINE] [NEWLINE] Instead of trying to save the world,  how about you regain your integrity, and make amends to your friends you have lied to? [NEWLINE] [NEWLINE] How  can you save the world, if you  succumb to corruption and disintegration?   Unless perhaps you envision yourself a martyr, a willing sacrifice to the world.  I would like to change your view on that, if that is the case, that the world is not worth your loss.  You should save it for something important. [NEWLINE] [NEWLINE] Lastly, I'd like to leave you with a short video.  [Saving the planet]( [URL] )</s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask>, that article was a case of people leaving carts on the public roads that everyone uses, which isn't actually their property -<mask> you removed the government, then those roads would be owned by someone, who would then tell all the meat-sellers to get off their land (and possibly sue them for littering it and trespassing, to boot). The land next to that road would<mask> be privately owned and could possibly be bought up for use by the various cart-owners, which would inevitably create the various shops that the article mentions later on. [ENDQ] [NEWLINE] Or they could be charged rent..... [NEWLINE] [NEWLINE] [STARTQ] Yeah, nothing makes you look worse than increasing the wealth of the area. And<mask> exactly will people see through the bullshit inevitably spouted by private corporations,<mask> not by politicians?<mask> you're saying sounds like a whole lot of conspiratorialism, and not a lot of actual evidence or examples. [ENDQ] [NEWLINE] [Lyndon B. Johnson.]( [URL].php/article/40889) [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the death of the various specialised butcher shops are<mask> of the existence of supermarkets [ENDQ] [NEWLINE] <mask> it's not. The article states it occurred<mask> street carts were made illegal by the Government [NEWLINE] [NEWLINE] [STARTQ] <mask> apparently this is<mask> there are a limited number of permits issued for hot-dog stands [ENDQ] [NEWLINE] Exactly, the government is restricting access to a market. That is not a 'free-market.' That is a restricted market.</s>
Label encoding: <s> [STARTQ] First of all, that article was a case of people leaving carts on the public roads that everyone uses, which isn't actually their property - if you removed the government, then those roads would be owned by someone, who would then tell all the meat-sellers to get off their land (and possibly sue them for littering it and trespassing, to boot). The land next to that road would also be privately owned and could possibly be bought up for use by the various cart-owners, which would inevitably create the various shops that the article mentions later on. [ENDQ] [NEWLINE] Or they could be charged rent..... [NEWLINE] [NEWLINE] [STARTQ] Yeah, nothing makes you look worse than increasing the wealth of the area. And why exactly will people see through the bullshit inevitably spouted by private corporations, but not by politicians? What you're saying sounds like a whole lot of conspiratorialism, and not a lot of actual evidence or examples. [ENDQ] [NEWLINE] [Lyndon B. Johnson.]( [URL].php/article/40889) [NEWLINE] [NEWLINE] [STARTQ] I think the death of the various specialised butcher shops are because of the existence of supermarkets [ENDQ] [NEWLINE] But it's not. The article states it occurred when street carts were made illegal by the Government [NEWLINE] [NEWLINE] [STARTQ] So apparently this is because there are a limited number of permits issued for hot-dog stands [ENDQ] [NEWLINE] Exactly, the government is restricting access to a market. That is not a 'free-market.' That is a restricted market.</s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think its worse<mask> you just dont know. Is that job interviewer racist or you didnt meet the qualifications for a job? Knowing one way or the other allows you to make corrections. Do you go back to school or dismiss that interviewer<mask> racist and move on? [NEWLINE] [NEWLINE] In the past the sign would tell you, whites only,<mask> you were not white you would apply somewhere else. These days the racism is there,<mask> no sign to warn you ahead of time.<mask> you are gonna be racist, own up to it. Put the whites only sign back up<mask> I know exactly<mask> type of person I am dealing with from the beginning. Wouldnt you rather know that the business owner of that cafe is a racist and spend your money with someone that doesnt see race? Lets say you go to a cafe with your<mask> that is a different race than you, service is slow, food is crappy, you end up with less than a fun time. Now the rest of your day is filled with thoughts of a crappy meal. Or, you see the whites only sign, go next door to a non racist, get regular service, meal is hot, the rest of your day is better. Multiply that by the number of times you interact with people in a day times every day for the rest of your life. [NEWLINE] [NEWLINE] Just to be clear,<mask><mask> that judging someone or a group of people on the color of their skin is disgusting. We are all human.</s>
Label encoding: <s>I think its worse because you just dont know. Is that job interviewer racist or you didnt meet the qualifications for a job? Knowing one way or the other allows you to make corrections. Do you go back to school or dismiss that interviewer as racist and move on? [NEWLINE] [NEWLINE] In the past the sign would tell you, whites only, if you were not white you would apply somewhere else. These days the racism is there, but no sign to warn you ahead of time. If you are gonna be racist, own up to it. Put the whites only sign back up so I know exactly what type of person I am dealing with from the beginning. Wouldnt you rather know that the business owner of that cafe is a racist and spend your money with someone that doesnt see race? Lets say you go to a cafe with your SO that is a different race than you, service is slow, food is crappy, you end up with less than a fun time. Now the rest of your day is filled with thoughts of a crappy meal. Or, you see the whites only sign, go next door to a non racist, get regular service, meal is hot, the rest of your day is better. Multiply that by the number of times you interact with people in a day times every day for the rest of your life. [NEWLINE] [NEWLINE] Just to be clear, I think that judging someone or a group of people on the color of their skin is disgusting. We are all human.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's still a level of trust between him and my stepmom (I just call her that,<mask> them not being married), and they're both faithful towards each other,<mask> they're still very much independent aside from that [ENDQ] [NEWLINE] This is<mask> my marriage to my husband (note: i'm male) works. We trust each other a lot. We love each other deeply. We live together and have done<mask> for nine years, and a lot of our life is spent together.<mask> we each have our own lives, and it's important to each of us that the other have strong connections outside the relationship. [NEWLINE] [NEWLINE] I wouldn't describe our relationship<mask> not serious; we're *married*, after all, and we've made binding lifelong commitments to each other that we intend to keep, and we love each other. [NEWLINE] [NEWLINE] <mask>... our relationship is going to be stronger<mask> we have our own independent spaces<mask> we grow and learn and do things that we then bring back to each other, and our relationship is going to be stronger<mask> we have strong emotional bonds to other people to help sustain us. [NEWLINE] [NEWLINE] This is<mask> I *expect* relationships to work. (And, no, my parents' relationships didn't work this way; my mother was married four times to abusive asshoels and had a string of terrible boyfriends in between the marriages; i've seen relationships just<mask> dysfuncitonal<mask> the ones you are describing) [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] There's still a level of trust between him and my stepmom (I just call her that, despite them not being married), and they're both faithful towards each other, but they're still very much independent aside from that [ENDQ] [NEWLINE] This is how my marriage to my husband (note: i'm male) works. We trust each other a lot. We love each other deeply. We live together and have done so for nine years, and a lot of our life is spent together. But we each have our own lives, and it's important to each of us that the other have strong connections outside the relationship. [NEWLINE] [NEWLINE] I wouldn't describe our relationship as not serious; we're *married*, after all, and we've made binding lifelong commitments to each other that we intend to keep, and we love each other. [NEWLINE] [NEWLINE] But... our relationship is going to be stronger if we have our own independent spaces where we grow and learn and do things that we then bring back to each other, and our relationship is going to be stronger if we have strong emotional bonds to other people to help sustain us. [NEWLINE] [NEWLINE] This is how I *expect* relationships to work. (And, no, my parents' relationships didn't work this way; my mother was married four times to abusive asshoels and had a string of terrible boyfriends in between the marriages; i've seen relationships just as dysfuncitonal as the ones you are describing) [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry abacabbmk, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=abacabbmk+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \)) [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=abacabbmk+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry abacabbmk, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=abacabbmk+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \)) [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=abacabbmk+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do you see them ALL driving? Or is there a smaller fraction of minorities that take these courses, which would apply even more to a months-long course? [ENDQ] [NEWLINE] [Minorities are less likely to have a driving license]( [URL].pdf),<mask> even under the current system [minorities vote less]( [URL] /). Ultimately, an educated vote for a representative is going to do a lot more to help them, than voting for a person that lies (for example<mask> we want to fix the infrastructure we need a decent Gas tax, which we currently don't have, in part due to fear of "more taxes). [NEWLINE] [NEWLINE] [STARTQ] To say nothing of the fact that driving does not directly dictate public policy and law; voting does. [ENDQ] [NEWLINE] Then wouldn't you consider voting to be more important. [NEWLINE] [NEWLINE] [STARTQ] Sure there is, otherwise it wouldn't affect them the most. Unless, of course, you're saying they're fundamentally less fit to vote, which is a whole different issue. [ENDQ] [NEWLINE] The reason<mask> they'll be affected the most is<mask> they are the ones more likely to have two-jobs and a child and simply no desire to put up with anything else on their plate (like courses), or have a full-time job, are currently going to a higher education institution and have a child - which once again puts too much on their plate to care for taking an additional course. Not<mask> they are "less fit to vote". </s>
Label encoding: <s> [STARTQ] Do you see them ALL driving? Or is there a smaller fraction of minorities that take these courses, which would apply even more to a months-long course? [ENDQ] [NEWLINE] [Minorities are less likely to have a driving license]( [URL].pdf), however even under the current system [minorities vote less]( [URL] /). Ultimately, an educated vote for a representative is going to do a lot more to help them, than voting for a person that lies (for example if we want to fix the infrastructure we need a decent Gas tax, which we currently don't have, in part due to fear of "more taxes). [NEWLINE] [NEWLINE] [STARTQ] To say nothing of the fact that driving does not directly dictate public policy and law; voting does. [ENDQ] [NEWLINE] Then wouldn't you consider voting to be more important. [NEWLINE] [NEWLINE] [STARTQ] Sure there is, otherwise it wouldn't affect them the most. Unless, of course, you're saying they're fundamentally less fit to vote, which is a whole different issue. [ENDQ] [NEWLINE] The reason why they'll be affected the most is because they are the ones more likely to have two-jobs and a child and simply no desire to put up with anything else on their plate (like courses), or have a full-time job, are currently going to a higher education institution and have a child - which once again puts too much on their plate to care for taking an additional course. Not because they are "less fit to vote". </s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hmm, it seems like we're not quite on the same page here.<mask> you look at the numbers, there is a higher percentage of men in STEM related fields than women.<mask><mask> we're saying about men and women being equal<mask> it comes to the brain is true, then there should be no disparity. It should be 50% men, 50% women,<mask> it's not.<mask><mask> we know that men and women are equal, then<mask>'s going on? There's simply a societal pressure to not pursue some of these fields. [NEWLINE] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] This article talks about<mask> under 20% of Computer Science degrees are going to women.<mask> you really think about it, it simply doesn't make sense. The only answer to<mask> things like this are happening, is<mask> in American society, girls don't take computer science classes. [NEWLINE] [NEWLINE] One of the keys to fixing that is getting those numbers up artificially (through grants, job spots, etc.)<mask> much<mask> possible,<mask> young, impressionable children can have an idol in that certain field. This strategy will bring us overall better scientists, computer scientists, engineers, psysicists, etc. over time which is better for everyone. [NEWLINE] [NEWLINE] This isn't a conversation about "healthy spaces" for men.<mask> every space is a "healthy space" for men and any that aren't should<mask> be destigmatized. There shouldn't be "unhealthy spaces" for anyone.</s>
Label encoding: <s>Hmm, it seems like we're not quite on the same page here. When you look at the numbers, there is a higher percentage of men in STEM related fields than women. If what we're saying about men and women being equal when it comes to the brain is true, then there should be no disparity. It should be 50% men, 50% women, but it's not. But since we know that men and women are equal, then what's going on? There's simply a societal pressure to not pursue some of these fields. [NEWLINE] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] This article talks about how under 20% of Computer Science degrees are going to women. When you really think about it, it simply doesn't make sense. The only answer to why things like this are happening, is because in American society, girls don't take computer science classes. [NEWLINE] [NEWLINE] One of the keys to fixing that is getting those numbers up artificially (through grants, job spots, etc.) as much as possible, so young, impressionable children can have an idol in that certain field. This strategy will bring us overall better scientists, computer scientists, engineers, psysicists, etc. over time which is better for everyone. [NEWLINE] [NEWLINE] This isn't a conversation about "healthy spaces" for men. But every space is a "healthy space" for men and any that aren't should also be destigmatized. There shouldn't be "unhealthy spaces" for anyone.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you are suggesting that in order not to increase the price of a meal in order to pay waiters a guaranteed minimum wage, you should increase the price of a meal automatically by 25% in order to pay a decent tip to a server [ENDQ] [NEWLINE] Not exactly. Servers already get an automatic minimum wage (contrary to<mask> you hear on reddit.)<mask> their tips don't meet that amout, the employer, by law, is obligated to subsidize their pay.<mask>, many states don't allow a separate wage for tipped employees at all (California for instance,)<mask> the server is earning more than minimum wage already. My problem with simply raising the wage and the prices correspondingly only benefits the employer, not the server.<mask> volume is extremely high and the server had to work much harder than usual, the server's pay stays the same<mask> the employer receives more profit.<mask> the increase is a separate line item, the server is directly compensated for their additional effort. [NEWLINE] [NEWLINE] [STARTQ] again<mask><mask> it's telling that the American'server' is preferred to a more European 'waiter') [ENDQ] [NEWLINE] The term "server" is relatively new, much<mask> "flight attendant". We used to use the terms "waiter" and "waitress"<mask> have abandoned those primarily<mask> they are a bit sexist to some. Semantically,<mask>, I don't see a fundamental difference between "waiting" on someone and "serving" them. [NEWLINE] </s>
Label encoding: <s> [STARTQ] you are suggesting that in order not to increase the price of a meal in order to pay waiters a guaranteed minimum wage, you should increase the price of a meal automatically by 25% in order to pay a decent tip to a server [ENDQ] [NEWLINE] Not exactly. Servers already get an automatic minimum wage (contrary to what you hear on reddit.) If their tips don't meet that amout, the employer, by law, is obligated to subsidize their pay. Also, many states don't allow a separate wage for tipped employees at all (California for instance,) so the server is earning more than minimum wage already. My problem with simply raising the wage and the prices correspondingly only benefits the employer, not the server. If volume is extremely high and the server had to work much harder than usual, the server's pay stays the same but the employer receives more profit. If the increase is a separate line item, the server is directly compensated for their additional effort. [NEWLINE] [NEWLINE] [STARTQ] again I think it's telling that the American'server' is preferred to a more European 'waiter') [ENDQ] [NEWLINE] The term "server" is relatively new, much as "flight attendant". We used to use the terms "waiter" and "waitress" but have abandoned those primarily because they are a bit sexist to some. Semantically, though, I don't see a fundamental difference between "waiting" on someone and "serving" them. [NEWLINE] </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>but the american economic system is capitalist, and the soviet union was not communist. capitalism is the private ownership of the means of production and the exploitation of labor power for the purposes of capital accumulation, it really doesn't matter<mask> many social programs or regulations are involved<mask><mask><mask> those prerequisites are maintained. communism is a classless, marketless, non-hierarchical society. it's not a "no true scotsman" argument to point out that the definitions you are using are totally and childishly ignorant cliches. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask>, even<mask> you were right, you're point is<mask>, exactly? 50 million dead by their own governments isn't enough to reject a premise? [ENDQ] [NEWLINE] <mask><mask> it is certainly enough to reject the premise that the vanguardist and authoritarian regimes masquerading<mask> communism for PR purposes were defensible in practice, and nothing more than that. [NEWLINE] [NEWLINE] [STARTQ] <mask>, you aren't right. I don't know<mask> you're getting the 40-50 Million figure from,<mask> every single source I can find puts the toll at at least 85 million, which is double your total, with most of them extending well beyond 100 million. Once again, this isn't including war deaths. It seems like you're trying to write off 70 million dead from starvation<mask> "incompetence" to defend a failed ideology. [ENDQ] [NEWLINE] Hint: Your sources are crap.</s>
Label encoding: <s>but the american economic system is capitalist, and the soviet union was not communist. capitalism is the private ownership of the means of production and the exploitation of labor power for the purposes of capital accumulation, it really doesn't matter how many social programs or regulations are involved as long as those prerequisites are maintained. communism is a classless, marketless, non-hierarchical society. it's not a "no true scotsman" argument to point out that the definitions you are using are totally and childishly ignorant cliches. [NEWLINE] [NEWLINE] [STARTQ] First of all, even if you were right, you're point is what, exactly? 50 million dead by their own governments isn't enough to reject a premise? [ENDQ] [NEWLINE] I think it is certainly enough to reject the premise that the vanguardist and authoritarian regimes masquerading as communism for PR purposes were defensible in practice, and nothing more than that. [NEWLINE] [NEWLINE] [STARTQ] Secondly, you aren't right. I don't know where you're getting the 40-50 Million figure from, but every single source I can find puts the toll at at least 85 million, which is double your total, with most of them extending well beyond 100 million. Once again, this isn't including war deaths. It seems like you're trying to write off 70 million dead from starvation as "incompetence" to defend a failed ideology. [ENDQ] [NEWLINE] Hint: Your sources are crap.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] All of your points are still ignoring the fact that<mask> the day black Americans were forced over to this country they were not considered people. [ENDQ] [NEWLINE] They'd be doing *<mask> * well otherwise.  By every metric, black Americans fare better than black citizens of any African country you'd care to name.  This is not to say that taking them over was right - it most certainly was not -<mask> rather that saying the problems in the black American community can all be laid down to the external influence of the white man is foolish. [NEWLINE] [NEWLINE] [STARTQ] And that very same logic is still present today, otherwise black bodies wouldn't be getting killed by systems that were put in place meant to protect them. [ENDQ] [NEWLINE] White bodies, too, man. ["Adjusted to take into account the racial breakdown of the U.S. population, he said black men are 3.5 times more likely to be killed by police than white men.<mask><mask> adjusted to take into account the racial breakdown in violent crime, the data actually show that police are less likely to kill black suspects than white ones."]( [URL] )  I wonder<mask> many Asians get shot by police in America. [NEWLINE] [NEWLINE] You've spared me plenty of time<mask> far.  "I don't have time" seems to often translate into "I don't have further arguments."  We're in CMV; this is a relatively new position for me and I'm very open to being persuaded again.</s>
Label encoding: <s> [STARTQ] All of your points are still ignoring the fact that since the day black Americans were forced over to this country they were not considered people. [ENDQ] [NEWLINE] They'd be doing * so * well otherwise.  By every metric, black Americans fare better than black citizens of any African country you'd care to name.  This is not to say that taking them over was right - it most certainly was not - but rather that saying the problems in the black American community can all be laid down to the external influence of the white man is foolish. [NEWLINE] [NEWLINE] [STARTQ] And that very same logic is still present today, otherwise black bodies wouldn't be getting killed by systems that were put in place meant to protect them. [ENDQ] [NEWLINE] White bodies, too, man. ["Adjusted to take into account the racial breakdown of the U.S. population, he said black men are 3.5 times more likely to be killed by police than white men. But also adjusted to take into account the racial breakdown in violent crime, the data actually show that police are less likely to kill black suspects than white ones."]( [URL] )  I wonder how many Asians get shot by police in America. [NEWLINE] [NEWLINE] You've spared me plenty of time so far.  "I don't have time" seems to often translate into "I don't have further arguments."  We're in CMV; this is a relatively new position for me and I'm very open to being persuaded again.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> do the limits of this authority end? This rationale could, and has been easily used to justify fascism. The question remains,<mask> does the authority of government to protect civil society stop, and the rights of the individual to freely choose their actions begin? That's been the big question throughout this entire thread, and<mask> solving one aspect of the problem, you've brought in a completely new, and<mask> argue more dangerous aspect.<mask> pragmatically, you've convinced me, you lack an objective limiting system for your arguments, in other words, you have no system, no "line in the sand", to determine to<mask> extend the government has the authority to limit freedoms to ensure a civil society. This,<mask> evidenced by history, quite often leads to totalitarianism and/or fascism. [NEWLINE] <mask>, your argument seems to me to just be justifying free agents getting in line with a government plan for society. It's a plan<mask><mask> with,<mask> that doesn't really factor in. [NEWLINE] That being said... [NEWLINE] [NEWLINE] Δ That was incredibly worded and a very, very good point. You convinced me! /r/threadkillers<mask>, I don't know<mask> you were thinking about his or not,<mask> a good addition to your post would be the *duty* of individual citizens to help create a civil society. That provides a nice individualist counterpart to your systematic argument. [NEWLINE] ΔΔΔ (does this extra count?)</s>
Label encoding: <s> Where do the limits of this authority end? This rationale could, and has been easily used to justify fascism. The question remains, when does the authority of government to protect civil society stop, and the rights of the individual to freely choose their actions begin? That's been the big question throughout this entire thread, and while solving one aspect of the problem, you've brought in a completely new, and if argue more dangerous aspect. While pragmatically, you've convinced me, you lack an objective limiting system for your arguments, in other words, you have no system, no "line in the sand", to determine to what extend the government has the authority to limit freedoms to ensure a civil society. This, as evidenced by history, quite often leads to totalitarianism and/or fascism. [NEWLINE] Also, your argument seems to me to just be justifying free agents getting in line with a government plan for society. It's a plan I agree with, but that doesn't really factor in. [NEWLINE] That being said... [NEWLINE] [NEWLINE] Δ That was incredibly worded and a very, very good point. You convinced me! /r/threadkillers Also, I don't know if you were thinking about his or not, but a good addition to your post would be the *duty* of individual citizens to help create a civil society. That provides a nice individualist counterpart to your systematic argument. [NEWLINE] ΔΔΔ (does this extra count?)</s>
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it is logical for the parenthesis to come first,<mask> it is the most critical component. [ENDQ] [NEWLINE] Left to right writing conventions don't really have all that much to do with the order of importance. [NEWLINE] [NEWLINE] [STARTQ] <mask> it fuses with whatever punctuation that the sentence contains [ENDQ] [NEWLINE] Both orientations do this. [NEWLINE] [NEWLINE] [STARTQ] Most humans are right handed,<mask> they would naturally tilt their heads to the right. [ENDQ] [NEWLINE] Most people are capable of parsing emoticons without head tilting.<mask> this wasn't the case, the dominate eye would likely play a greater role than dominate hand.<mask>, the point about mirroring is strange,<mask> the emoticon has no distinction between right and left. The "photo" version would be identical to the "mirror" version. The other representation is upside down. [NEWLINE] [NEWLINE] [STARTQ] On the keyboard itself, it is much more natural to go from a higher key to a lower key rather than the other way around. [ENDQ] [NEWLINE] Unsubstantiated. Most words require movements in both directions.<mask> the one feels more natural than the other, this is likely due to your personal habits. [NEWLINE] [NEWLINE] Both standard smiley faces are perfectly valid and just<mask> readily interpreted. [NEWLINE] [NEWLINE] ;) [NEWLINE] [NEWLINE] (; [NEWLINE] [NEWLINE] The wink,<mask><mask><mask><mask>, likely contributes to the common preference for placing the eyes first. With eyes following mouth, the semicolon takes on a sad angle.</s>
Label encoding: <s> [STARTQ] so it is logical for the parenthesis to come first, as it is the most critical component. [ENDQ] [NEWLINE] Left to right writing conventions don't really have all that much to do with the order of importance. [NEWLINE] [NEWLINE] [STARTQ] because it fuses with whatever punctuation that the sentence contains [ENDQ] [NEWLINE] Both orientations do this. [NEWLINE] [NEWLINE] [STARTQ] Most humans are right handed, so they would naturally tilt their heads to the right. [ENDQ] [NEWLINE] Most people are capable of parsing emoticons without head tilting. If this wasn't the case, the dominate eye would likely play a greater role than dominate hand. Also, the point about mirroring is strange, as the emoticon has no distinction between right and left. The "photo" version would be identical to the "mirror" version. The other representation is upside down. [NEWLINE] [NEWLINE] [STARTQ] On the keyboard itself, it is much more natural to go from a higher key to a lower key rather than the other way around. [ENDQ] [NEWLINE] Unsubstantiated. Most words require movements in both directions. If the one feels more natural than the other, this is likely due to your personal habits. [NEWLINE] [NEWLINE] Both standard smiley faces are perfectly valid and just as readily interpreted. [NEWLINE] [NEWLINE] ;) [NEWLINE] [NEWLINE] (; [NEWLINE] [NEWLINE] The wink, on the other hand, likely contributes to the common preference for placing the eyes first. With eyes following mouth, the semicolon takes on a sad angle.</s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a fellow UK resident and healthcare worker,<mask><mask> with you about<mask> fundamentally important it is to have care that is given based on need rather than on the ability to pay. I won't pretend to have a perfect understanding of the state of play in America,<mask> I understand that<mask> being one of the most expensive healthcare systems in regards to the patients, it's<mask> FAR from the best. [NEWLINE] [NEWLINE] <mask> I take from Steven Levitt's statement is that the USA cannot expect to provide universal healthcare based on its current costs and profits. The reality is that in all healthcare systems, the costs of treatments, facilities, and even training doctors needs to be made up for somewhere down the road. You cannot expect to cover the costs of treating one of the most obese (requiring a lot of expensive treatments), populous (meaning there are a lot more people to distribute limited resources between) and underpaid (limiting their ability to both contribute through their taxes and to contribute to either preventative or curative treatment) demographics of patients completely through taxes and government subsidies. [NEWLINE] [NEWLINE] The long and short of it is that healthcare is always going to involve some supply and demand. At present, universal healthcare would likely struggle to supply enough resources to meet an increasingly growing demand. I imagine that juggling the budget somewhat from more bloated cash sinkholes like the US military might help to ease the pressure,<mask> in the current global climate that's not going to happen.</s>
Label encoding: <s>As a fellow UK resident and healthcare worker, I agree with you about how fundamentally important it is to have care that is given based on need rather than on the ability to pay. I won't pretend to have a perfect understanding of the state of play in America, though I understand that despite being one of the most expensive healthcare systems in regards to the patients, it's also FAR from the best. [NEWLINE] [NEWLINE] What I take from Steven Levitt's statement is that the USA cannot expect to provide universal healthcare based on its current costs and profits. The reality is that in all healthcare systems, the costs of treatments, facilities, and even training doctors needs to be made up for somewhere down the road. You cannot expect to cover the costs of treating one of the most obese (requiring a lot of expensive treatments), populous (meaning there are a lot more people to distribute limited resources between) and underpaid (limiting their ability to both contribute through their taxes and to contribute to either preventative or curative treatment) demographics of patients completely through taxes and government subsidies. [NEWLINE] [NEWLINE] The long and short of it is that healthcare is always going to involve some supply and demand. At present, universal healthcare would likely struggle to supply enough resources to meet an increasingly growing demand. I imagine that juggling the budget somewhat from more bloated cash sinkholes like the US military might help to ease the pressure, but in the current global climate that's not going to happen.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] TL;DR I see<mask> it is consistent within the system to try older children<mask> adults.<mask><mask> it's still a bit vengeful<mask> it's punitive,<mask> hey, that's the system. I still fundamentally disagree with the system,<mask> that doesn't have bearing in this argument<mask> stated. [NEWLINE] [NEWLINE] The definition of juveniles versus children is a good point. I hate it, and<mask><mask> it's stupid still,<mask> that has to do with<mask> I feel the judicial system<mask> unfair<mask> a whole rather then the argument at hand. It makes sense. [NEWLINE] [NEWLINE] I can<mask> see<mask> the system would view them<mask> beyond rehabilitated,<mask> once again, this is a difference with<mask> I view the system, not specifically about children. [NEWLINE] [NEWLINE] In the case of teenagers with violent offenses near adulthood, I see little difference between trying an 18 year old<mask> an adult vs a 16 year old, and your reasoning makes sense for that case. 93%[1] of the cases that get transferred to criminal court are 15 and older. It's consistent. [NEWLINE] [NEWLINE] That being said, the other 7% I would have a problem with. "Adult" acts by much younger teenagers gets too much into the territory of people not being fully developed.<mask><mask> it's even more society's obligation to try harder to rehabilitate, and should stay in criminal court. [NEWLINE] [NEWLINE] [NEWLINE] [1] [URL] #Demographics</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] TL;DR I see how it is consistent within the system to try older children as adults. I think it's still a bit vengeful since it's punitive, but hey, that's the system. I still fundamentally disagree with the system, but that doesn't have bearing in this argument as stated. [NEWLINE] [NEWLINE] The definition of juveniles versus children is a good point. I hate it, and I think it's stupid still, but that has to do with how I feel the judicial system as unfair as a whole rather then the argument at hand. It makes sense. [NEWLINE] [NEWLINE] I can also see how the system would view them as beyond rehabilitated, but once again, this is a difference with how I view the system, not specifically about children. [NEWLINE] [NEWLINE] In the case of teenagers with violent offenses near adulthood, I see little difference between trying an 18 year old as an adult vs a 16 year old, and your reasoning makes sense for that case. 93%[1] of the cases that get transferred to criminal court are 15 and older. It's consistent. [NEWLINE] [NEWLINE] That being said, the other 7% I would have a problem with. "Adult" acts by much younger teenagers gets too much into the territory of people not being fully developed. I think it's even more society's obligation to try harder to rehabilitate, and should stay in criminal court. [NEWLINE] [NEWLINE] [NEWLINE] [1] [URL] #Demographics</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The example I just gave focuses on utilizing technology to deliver the lecture.  The ted talk I linked involves a brief story (you can read more about it in an article somewhere<mask> I can't look it up right now) about an elementary school remedial math class<mask> they replaced the teacher with the khan academy.  The rule in class was something along the lines of "<mask> you can't figure something out ask your neighbor,<mask> you both can't figure it out together, THEN you can ask the teacher."  There was no set curriculum for the entire classroom, just go through the site for 90 minutes each day or whatever, and by the end of the year, all these students were doing college level trig. <mask> that isn't a perfect example of<mask> individualized study can do for students, I don't know<mask> is. [NEWLINE] [NEWLINE] Now, not all subjects are<mask> simple and straight-forward<mask> math.  Most of the learning in English comes from discussion and peer review, and it's hard to teach team sports without a team,<mask> there is no reason<mask> we need to restrict all of school to this structure. [NEWLINE] [NEWLINE] Again, in a perfect world, we would have both a computer aid for the student<mask> well<mask> maybe a 1:5 ratio in the classroom<mask> a whole. <mask> lacking that, we could have a pretty damn effective classroom with 30 students per teacher<mask> we would properly utilize technology.</s>
Label encoding: <s>The example I just gave focuses on utilizing technology to deliver the lecture.  The ted talk I linked involves a brief story (you can read more about it in an article somewhere but I can't look it up right now) about an elementary school remedial math class where they replaced the teacher with the khan academy.  The rule in class was something along the lines of " if you can't figure something out ask your neighbor, if you both can't figure it out together, THEN you can ask the teacher."  There was no set curriculum for the entire classroom, just go through the site for 90 minutes each day or whatever, and by the end of the year, all these students were doing college level trig.  If that isn't a perfect example of what individualized study can do for students, I don't know what is. [NEWLINE] [NEWLINE] Now, not all subjects are as simple and straight-forward as math.  Most of the learning in English comes from discussion and peer review, and it's hard to teach team sports without a team, but there is no reason why we need to restrict all of school to this structure. [NEWLINE] [NEWLINE] Again, in a perfect world, we would have both a computer aid for the student as well as maybe a 1:5 ratio in the classroom as a whole.  But lacking that, we could have a pretty damn effective classroom with 30 students per teacher if we would properly utilize technology.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ah you're one of THOSE people. [NEWLINE] [NEWLINE] We pay taxes to pay for things controlled/funded by the government like roads, public transit and various other schemes like education and<mask> not. [NEWLINE] [NEWLINE] By voting a country/state whatever comes to a general consensus<mask> to<mask> their tax money pays for.<mask> you don't like it, you vote for someone else and hope that enough people agree with you<mask> your money can go toward<mask> you want it to go toward. Simply put - democracy, in ridiculously simple terms. [NEWLINE] [NEWLINE] <mask> those taxes are a generally agreed upon thing by the people. And,<mask> part of a civilised society, sometimes you have to follow along with things you don't particularly like<mask> it has been generally agreed by the public that that's<mask> it should be. [NEWLINE] [NEWLINE] I personally don't like that fact that in part, my tax money is paying for David Cameron to twat around chatting bullshit,<mask> I'm not going to sit here saying that I'm being forced to pay for him to do it under threat of violence. It's democracy -<mask> come May I'll be putting my vote in to have my say on<mask> my tax money pays for and I'll be crossing my fingers that people have the same opinions<mask> me. [NEWLINE] [NEWLINE] <mask> you don't like that... well, I suppose you could move to a dictatorship or something, I'm sure you'll like that a lot more!</s>
Label encoding: <s>Ah you're one of THOSE people. [NEWLINE] [NEWLINE] We pay taxes to pay for things controlled/funded by the government like roads, public transit and various other schemes like education and what not. [NEWLINE] [NEWLINE] By voting a country/state whatever comes to a general consensus as to what their tax money pays for. If you don't like it, you vote for someone else and hope that enough people agree with you so your money can go toward what you want it to go toward. Simply put - democracy, in ridiculously simple terms. [NEWLINE] [NEWLINE] So those taxes are a generally agreed upon thing by the people. And, as part of a civilised society, sometimes you have to follow along with things you don't particularly like because it has been generally agreed by the public that that's how it should be. [NEWLINE] [NEWLINE] I personally don't like that fact that in part, my tax money is paying for David Cameron to twat around chatting bullshit, but I'm not going to sit here saying that I'm being forced to pay for him to do it under threat of violence. It's democracy - so come May I'll be putting my vote in to have my say on what my tax money pays for and I'll be crossing my fingers that people have the same opinions as me. [NEWLINE] [NEWLINE] If you don't like that... well, I suppose you could move to a dictatorship or something, I'm sure you'll like that a lot more!</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It seems to me your chances of getting the flu are higher WITH the shot<mask> you are absolutely exposing yourself to the flu versus maybe not being exposed at all. [ENDQ] [NEWLINE] Well,<mask><mask> the [CDC]( [URL] ) you can't get sick from the flu shot<mask> "Flu vaccines that are administered with a needle are currently made in two ways: the vaccine is made either with a) flu vaccine viruses that have been ‘inactivated’ and are<mask> not infectious, or b) with no flu vaccine viruses at all (which is the case for recombinant influenza vaccine). In randomized, blinded studies,<mask> some people got flu shots and others got saltwater shots, the only differences in symptoms was increased soreness in the arm and redness at the injection site among people who got the flu shot. There were no differences in terms of body aches, fever, cough, runny nose or sore throat."<mask><mask> you get sick after getting a flu shot, it's not<mask> the shot gave you the flu. It's<mask> your immune system was weakened by the shot and you were exposed to something else (maybe even another strain of the flu that wasn't covered by the vaccine) which was able to attack your immune system in its weakened state.<mask>,<mask> you're not exposed to the flu<mask> you don't get the vaccine, then<mask> do people who don't get the vaccine catch it? </s>
Label encoding: <s> [STARTQ] It seems to me your chances of getting the flu are higher WITH the shot since you are absolutely exposing yourself to the flu versus maybe not being exposed at all. [ENDQ] [NEWLINE] Well, according to the [CDC]( [URL] ) you can't get sick from the flu shot because "Flu vaccines that are administered with a needle are currently made in two ways: the vaccine is made either with a) flu vaccine viruses that have been ‘inactivated’ and are therefore not infectious, or b) with no flu vaccine viruses at all (which is the case for recombinant influenza vaccine). In randomized, blinded studies, where some people got flu shots and others got saltwater shots, the only differences in symptoms was increased soreness in the arm and redness at the injection site among people who got the flu shot. There were no differences in terms of body aches, fever, cough, runny nose or sore throat." So if you get sick after getting a flu shot, it's not because the shot gave you the flu. It's because your immune system was weakened by the shot and you were exposed to something else (maybe even another strain of the flu that wasn't covered by the vaccine) which was able to attack your immune system in its weakened state. Also, if you're not exposed to the flu if you don't get the vaccine, then how do people who don't get the vaccine catch it? </s>
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a few problems with your perspective that I would like to point out: [NEWLINE] [NEWLINE] 1) Allowing people to deteriorate and starve to death will significantly spike crime rates.  It's pretty obvious that the poorer someone is, statistically speaking, the more likely they are to engage in crime.  Desperation can drive people to do things they wouldn't otherwise do. <mask> well, people aren't<mask> emphathetic towards others<mask> they feel that they have been screwed over themselves.  The basic idea is that<mask> would they want to live peacefully in society with you<mask> they get nothing from it?  They wouldn't.  They won't sit back peacefully. <mask> you're suggesting would cause much more violence and instability, which is bad for everyone, including you. [NEWLINE] [NEWLINE] 2) Your view completely ignores any and all environmental factors.  Jobs aren't always easy to come by. <mask> you're born into a ghetto, it becomes much more difficult to get out of it than it would be for a middle-class person to stay middle-class.  We should have equality of opportunity,<mask> de facto, that doesn't exist.  It's our responsibility<mask> a society to put safety nets in place to try and ensure that everyone has a fair shot at improving their situation.  Trying to ignore all societal factors (like racism)<mask> well<mask> inequality of opportunity is a recipe for social disaster.</s>
Label encoding: <s>There's a few problems with your perspective that I would like to point out: [NEWLINE] [NEWLINE] 1) Allowing people to deteriorate and starve to death will significantly spike crime rates.  It's pretty obvious that the poorer someone is, statistically speaking, the more likely they are to engage in crime.  Desperation can drive people to do things they wouldn't otherwise do.  As well, people aren't as emphathetic towards others if they feel that they have been screwed over themselves.  The basic idea is that why would they want to live peacefully in society with you when they get nothing from it?  They wouldn't.  They won't sit back peacefully.  What you're suggesting would cause much more violence and instability, which is bad for everyone, including you. [NEWLINE] [NEWLINE] 2) Your view completely ignores any and all environmental factors.  Jobs aren't always easy to come by.  If you're born into a ghetto, it becomes much more difficult to get out of it than it would be for a middle-class person to stay middle-class.  We should have equality of opportunity, but de facto, that doesn't exist.  It's our responsibility as a society to put safety nets in place to try and ensure that everyone has a fair shot at improving their situation.  Trying to ignore all societal factors (like racism) as well as inequality of opportunity is a recipe for social disaster.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0685, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I see this very often,<mask> this is a very inaccurate criticism of AA. College admissions officers don't choose unqualified minorities over qualified non-minorities. You may<mask><mask> the person picked may be less qualified than the white person picked,<mask> the minority who is enrolled is not unqualified to be there. They still have to meet college admission standards. [ENDQ] [NEWLINE] <mask><mask> we are stying the same thing with different words. Less qualified *is* unqualified. [NEWLINE] [NEWLINE] [STARTQ] Which individual,<mask> taking into account their racial status, hurts more? [ENDQ] [NEWLINE] I'd argue both are harmed equally. Again, the number helped *cannot* exceed those who are unfairly penalized. [NEWLINE] [NEWLINE] [STARTQ] You need to understand that by only looking at things from an individual level you are ignoring<mask> racial discrimination affects large groups of people. Individual white people are discriminated against,<mask> opposed to entire swaths of the black community. [ENDQ] [NEWLINE] Affirmative action is still discrimination against wide swaths of people.<mask> you choose to justify it doesn't change<mask> it is. [NEWLINE] [NEWLINE] The only difference is, instead of individual racists, you have the power of law on your side. Your side spends my tax monies in ways that are designed to hurt me by those who hate me and I have no choice<mask> to participate. Minorities at least have the option of avoiding business with individual bigots.  </s><pad>
Label encoding: <s> [STARTQ] I see this very often, but this is a very inaccurate criticism of AA. College admissions officers don't choose unqualified minorities over qualified non-minorities. You may argue that the person picked may be less qualified than the white person picked, but the minority who is enrolled is not unqualified to be there. They still have to meet college admission standards. [ENDQ] [NEWLINE] I think we are stying the same thing with different words. Less qualified *is* unqualified. [NEWLINE] [NEWLINE] [STARTQ] Which individual, when taking into account their racial status, hurts more? [ENDQ] [NEWLINE] I'd argue both are harmed equally. Again, the number helped *cannot* exceed those who are unfairly penalized. [NEWLINE] [NEWLINE] [STARTQ] You need to understand that by only looking at things from an individual level you are ignoring how racial discrimination affects large groups of people. Individual white people are discriminated against, as opposed to entire swaths of the black community. [ENDQ] [NEWLINE] Affirmative action is still discrimination against wide swaths of people. However you choose to justify it doesn't change what it is. [NEWLINE] [NEWLINE] The only difference is, instead of individual racists, you have the power of law on your side. Your side spends my tax monies in ways that are designed to hurt me by those who hate me and I have no choice but to participate. Minorities at least have the option of avoiding business with individual bigots.  </s><pad>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's always problematic to compare tragedies,<mask> to the extent that one massacre can be said to be worse than another, the Holocaust has every aggravating factor. [NEWLINE] [NEWLINE] *Scale:  Check.  It's huge. [NEWLINE] [NEWLINE] *Genocidal:  Check.  The intent was to exterminate an entire race of people.  And it was born out of pure one-sided racism, not out of a preexisting violent conflict between two ethnic groups. [NEWLINE] [NEWLINE] *Torture: Check.  Deliberate torment, slavery, and suffering were inflicted beyond efficiency or the requirements of the mass murder. [NEWLINE] [NEWLINE] *Deliberate: Check.  Things didn't sort of "get out of hand"; they were deliberately plotted out.  A lot of people (from SS to Hitler) could have said no to the scheme or just to their own participation, slept on it, and decided to proceed. [NEWLINE] [NEWLINE] *Widespread complicity: Check.  This wasn't just "The SS run amok"; you had a whole host of military and civilian cooperation with the process. [NEWLINE] [NEWLINE] *Refusal to rescue: you have refugees literally being sent back by uninvolved parties. [NEWLINE] [NEWLINE] *Horrible ingenuity: Check.  Some brilliant people were clearly involved in the invention and problem-solving here.  It's not merely a massacre, it's a masterpiece of horror. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>It's always problematic to compare tragedies, but to the extent that one massacre can be said to be worse than another, the Holocaust has every aggravating factor. [NEWLINE] [NEWLINE] *Scale:  Check.  It's huge. [NEWLINE] [NEWLINE] *Genocidal:  Check.  The intent was to exterminate an entire race of people.  And it was born out of pure one-sided racism, not out of a preexisting violent conflict between two ethnic groups. [NEWLINE] [NEWLINE] *Torture: Check.  Deliberate torment, slavery, and suffering were inflicted beyond efficiency or the requirements of the mass murder. [NEWLINE] [NEWLINE] *Deliberate: Check.  Things didn't sort of "get out of hand"; they were deliberately plotted out.  A lot of people (from SS to Hitler) could have said no to the scheme or just to their own participation, slept on it, and decided to proceed. [NEWLINE] [NEWLINE] *Widespread complicity: Check.  This wasn't just "The SS run amok"; you had a whole host of military and civilian cooperation with the process. [NEWLINE] [NEWLINE] *Refusal to rescue: you have refugees literally being sent back by uninvolved parties. [NEWLINE] [NEWLINE] *Horrible ingenuity: Check.  Some brilliant people were clearly involved in the invention and problem-solving here.  It's not merely a massacre, it's a masterpiece of horror. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I hope you read this. I have thought about this long and hard and had an epiphany about it a little<mask> ago that has given me great comfort. The epiphany was the realization that being dead is literally nothing I will ever have to experience. Existing and being alive are synonymous, and you have to exist in order to experience. I know this sound obvious,<mask> don't let that make you just brush it off. It was through really internalizing these thoughts and<mask> "not existing" implies that I finally became at peace. [NEWLINE] [NEWLINE] There's a fundamental flaw that I can sense underlies your thinking, that I used to have<mask> well. And that flaw is you're still treating yourself<mask> an eternal thing that will one day be in a state called "dead". That's<mask> you don't like the thought of not being able to think, and see, and all that.<mask> you truly grasp<mask> it means to "not exist", you will see that it's not that it's just "not that bad", it's literally nothing that you will ever have to deal with in any way. Being afraid of being dead is like being afraid of going somewhere that's impossible to get to. You will ONLY ever be alive from your point of view. [NEWLINE] [NEWLINE] I don't know<mask> I did a good job conveying my point,<mask> please ask questions<mask> I can clarify it better.</s>
Label encoding: <s>I hope you read this. I have thought about this long and hard and had an epiphany about it a little while ago that has given me great comfort. The epiphany was the realization that being dead is literally nothing I will ever have to experience. Existing and being alive are synonymous, and you have to exist in order to experience. I know this sound obvious, but don't let that make you just brush it off. It was through really internalizing these thoughts and what "not existing" implies that I finally became at peace. [NEWLINE] [NEWLINE] There's a fundamental flaw that I can sense underlies your thinking, that I used to have as well. And that flaw is you're still treating yourself as an eternal thing that will one day be in a state called "dead". That's why you don't like the thought of not being able to think, and see, and all that. When you truly grasp what it means to "not exist", you will see that it's not that it's just "not that bad", it's literally nothing that you will ever have to deal with in any way. Being afraid of being dead is like being afraid of going somewhere that's impossible to get to. You will ONLY ever be alive from your point of view. [NEWLINE] [NEWLINE] I don't know if I did a good job conveying my point, so please ask questions so I can clarify it better.</s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For me personally, the worst "bad" teacher I had was<mask> one who had favorites and I happened to get on her good side<mask> my grades were quite good. Friends of mine in the class did very badly for work I considered to be about equal to my own. The playing favorites thing is a big issue to,<mask> I don't know<mask> it can be addressed. Could more rigorous student evaluations maybe be a part of teacher performance reviews? [NEWLINE] [NEWLINE] About "good" teachers, unequivocally yes for many of them. My favorite teacher of all time taught me more about his subject than any other class I'd taken at that point and<mask> taught be about life and<mask> I was passionate about. Doubtless, he's the kind of teacher that would survive<mask><mask> the evaluation method and that is a very good thing. Other teachers that I would maybe consider "good" with a slight catch were the ones who taught me a lot about life and not quite<mask> much<mask> they probably should have about the subject. I loved them and their classes<mask> don't know<mask> long they would last<mask> they had to teach to the test. It's a balance and I don't know which side is more important. [NEWLINE] [NEWLINE] I have not had that specific experience with a test,<mask> I have encountered tests that were not at all<mask> I expected after attending all classes/lectures. Have you had that experience? </s>
Label encoding: <s>For me personally, the worst "bad" teacher I had was also one who had favorites and I happened to get on her good side so my grades were quite good. Friends of mine in the class did very badly for work I considered to be about equal to my own. The playing favorites thing is a big issue to, though I don't know how it can be addressed. Could more rigorous student evaluations maybe be a part of teacher performance reviews? [NEWLINE] [NEWLINE] About "good" teachers, unequivocally yes for many of them. My favorite teacher of all time taught me more about his subject than any other class I'd taken at that point and also taught be about life and what I was passionate about. Doubtless, he's the kind of teacher that would survive regardless of the evaluation method and that is a very good thing. Other teachers that I would maybe consider "good" with a slight catch were the ones who taught me a lot about life and not quite so much as they probably should have about the subject. I loved them and their classes but don't know how long they would last if they had to teach to the test. It's a balance and I don't know which side is more important. [NEWLINE] [NEWLINE] I have not had that specific experience with a test, though I have encountered tests that were not at all what I expected after attending all classes/lectures. Have you had that experience? </s>
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I should have written that better.<mask> I meant was, he didn't continue the momentum. Tons of people were really excited for Obama. They were excited for change.<mask> he became president that excitement died down<mask> he focused less on the public and more on his policy. The problem with that is that Congress doesn't care<mask> public favor he had before.<mask> they can block everything he tries to pass, they will. Take TPP for example.<mask> most people knew about it, they would hate it.<mask> most people don't, and they don't care. They feel<mask><mask> they don't have a say in politics.<mask> Congress and the president can do whatever they want and the people who will oppose them are a minority. [NEWLINE] [NEWLINE] The difference with Bernie is that he isn't running just<mask> he can try to get bills<mask> Congress. He wants to be president to inspire the population<mask> that people are more involved in politics.<mask><mask> Congress blocks a bill that the population wants, people will get angry, and people will vote. Then the members of Congress who helped block the bill will have a much harder time getting re-elected. To quote Bernie, "<mask> millions of people stand up and fight, they win." That's his goal; not to push bills through Congress,<mask> to get the population involved in their government<mask> that our voices and votes will decide which bills go through. </s>
Label encoding: <s>I should have written that better. What I meant was, he didn't continue the momentum. Tons of people were really excited for Obama. They were excited for change. When he became president that excitement died down as he focused less on the public and more on his policy. The problem with that is that Congress doesn't care what public favor he had before. If they can block everything he tries to pass, they will. Take TPP for example. If most people knew about it, they would hate it. But most people don't, and they don't care. They feel as though they don't have a say in politics. So Congress and the president can do whatever they want and the people who will oppose them are a minority. [NEWLINE] [NEWLINE] The difference with Bernie is that he isn't running just so he can try to get bills though Congress. He wants to be president to inspire the population so that people are more involved in politics. So when Congress blocks a bill that the population wants, people will get angry, and people will vote. Then the members of Congress who helped block the bill will have a much harder time getting re-elected. To quote Bernie, " When millions of people stand up and fight, they win." That's his goal; not to push bills through Congress, but to get the population involved in their government so that our voices and votes will decide which bills go through. </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm not sure<mask> you want an indirect solution that may or may not solve the problem....Indirect solutions can easily backfire. [ENDQ] [NEWLINE] I find this particularly amusing<mask> I say the exact same thing very often, with one difference:  The minimum wage is the "indirect solution" I'm referring to.  The fundamental problem that minimum wage tries to solve is that people should have enough money to survive;  it seems obvious that the solution is giving people enough money to survive. <mask> min wage does instead is: [NEWLINE] "transfer wealth from arbitrary consumers (through higher prices) and arbitrary companies to people who generally (<mask> not always) coincide with those who actually need financial help.  Don't help people who are unable to find a job, and in the process, make it more attractive for companies to lay off people in favor of automation". [NEWLINE] [NEWLINE] That's about<mask> indirect<mask> it gets.  The direct (and common-sense) solution is a basic income,<mask><mask><mask> that's likely politically infeasible for now, more robust welfare (effectively a basic income + means-testing).  To be fair, in the absence of robust welfare, min wage is a stop-gap,<mask> holding it up<mask><mask> to an "indirect" solution is just plain wrong;  min-wage is a poster-child for a suboptimal, indirect solution.</s>
Label encoding: <s> [STARTQ] I'm not sure why you want an indirect solution that may or may not solve the problem....Indirect solutions can easily backfire. [ENDQ] [NEWLINE] I find this particularly amusing since I say the exact same thing very often, with one difference:  The minimum wage is the "indirect solution" I'm referring to.  The fundamental problem that minimum wage tries to solve is that people should have enough money to survive;  it seems obvious that the solution is giving people enough money to survive.  What min wage does instead is: [NEWLINE] "transfer wealth from arbitrary consumers (through higher prices) and arbitrary companies to people who generally ( but not always) coincide with those who actually need financial help.  Don't help people who are unable to find a job, and in the process, make it more attractive for companies to lay off people in favor of automation". [NEWLINE] [NEWLINE] That's about as indirect as it gets.  The direct (and common-sense) solution is a basic income, but given that that's likely politically infeasible for now, more robust welfare (effectively a basic income + means-testing).  To be fair, in the absence of robust welfare, min wage is a stop-gap, but holding it up in contrast to an "indirect" solution is just plain wrong;  min-wage is a poster-child for a suboptimal, indirect solution.</s>
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See, there's two types of thinkers in the world. [NEWLINE] [NEWLINE] There's logical thinkers, who work pretty much the same way you and I work, and then there's emotional thinkers, who think in the same way your friend (and probably who made this video) think. [NEWLINE] [NEWLINE] We can't 'feel' things the way they perceive this, the same way they can't feel things the same way we do. [NEWLINE] [NEWLINE] For some people 'Think about the children' is a valid, logical point. For me it isn't. [NEWLINE] [NEWLINE] <mask><mask><mask> emotions are not a huge part of<mask> you perceive the world, that's fine,<mask><mask> you do perceive the world emotionally, then this kind of reson is going to make several valid points to you. [NEWLINE] You can't deny that hanging with your friends is different than having a skype conference chat, and this is<mask> the video is all about. [NEWLINE] [NEWLINE] Technologically speaking, we're moving forward. Technology is then creating a sort of nostalgia in people who notice that children today are growing in a different way than they did (and<mask><mask> that's fine, the world is not the same,<mask> I don't expect them to be the same),<mask> for some people that brings an emotional loss of some kind. That's it, I don't think there's something that can change your view<mask> it's mostly a personal preference.</s>
Label encoding: <s>See, there's two types of thinkers in the world. [NEWLINE] [NEWLINE] There's logical thinkers, who work pretty much the same way you and I work, and then there's emotional thinkers, who think in the same way your friend (and probably who made this video) think. [NEWLINE] [NEWLINE] We can't 'feel' things the way they perceive this, the same way they can't feel things the same way we do. [NEWLINE] [NEWLINE] For some people 'Think about the children' is a valid, logical point. For me it isn't. [NEWLINE] [NEWLINE] As long as emotions are not a huge part of how you perceive the world, that's fine, but if you do perceive the world emotionally, then this kind of reson is going to make several valid points to you. [NEWLINE] You can't deny that hanging with your friends is different than having a skype conference chat, and this is what the video is all about. [NEWLINE] [NEWLINE] Technologically speaking, we're moving forward. Technology is then creating a sort of nostalgia in people who notice that children today are growing in a different way than they did (and I think that's fine, the world is not the same, so I don't expect them to be the same), but for some people that brings an emotional loss of some kind. That's it, I don't think there's something that can change your view since it's mostly a personal preference.</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I haven't posted anything on my FB profile in a long time,<mask> I don't think anyone should dictate<mask> I do and don't post. It is, after all, *my* account. [NEWLINE] [NEWLINE] [STARTQ] By doing<mask>, they run a grave risk of alienating friends, family, or coworkers. [ENDQ] [NEWLINE] <mask> anyone will end up alienated over my views on politics -<mask> they can't deal with the fact that people have different opinions - maybe we shouldn't be friends on social media (or in real life) after all. And like you said, some issues are important and it's worth it. I deleted some people who constantly posted homophobic stuff all over my feed, for example. [NEWLINE] [NEWLINE] [STARTQ] Is this rank foolishness on your part, or is there some gain that I am failing to appreciate? [ENDQ] [NEWLINE] Generally<mask> I post anywhere online, I don't think about any sort of personal gain in the way you seem to suggest. I share and post things that interest me<mask> I want to discuss them, draw people's attention to them or simply make my view known. I do that under the assumption that people can take differences in opinion in a civilized manner.<mask> they can't, they are free not to engage me or remove me from their social media. I retain the same right. That sounds much more civil than someone else telling me<mask> I should or shouldn't post. </s>
Label encoding: <s>I haven't posted anything on my FB profile in a long time, but I don't think anyone should dictate what I do and don't post. It is, after all, *my* account. [NEWLINE] [NEWLINE] [STARTQ] By doing so, they run a grave risk of alienating friends, family, or coworkers. [ENDQ] [NEWLINE] If anyone will end up alienated over my views on politics - if they can't deal with the fact that people have different opinions - maybe we shouldn't be friends on social media (or in real life) after all. And like you said, some issues are important and it's worth it. I deleted some people who constantly posted homophobic stuff all over my feed, for example. [NEWLINE] [NEWLINE] [STARTQ] Is this rank foolishness on your part, or is there some gain that I am failing to appreciate? [ENDQ] [NEWLINE] Generally when I post anywhere online, I don't think about any sort of personal gain in the way you seem to suggest. I share and post things that interest me because I want to discuss them, draw people's attention to them or simply make my view known. I do that under the assumption that people can take differences in opinion in a civilized manner. If they can't, they are free not to engage me or remove me from their social media. I retain the same right. That sounds much more civil than someone else telling me what I should or shouldn't post. </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just a few lines in the tax return?<mask> would that work? [NEWLINE] [NEWLINE] And<mask> counts<mask> "future earnings" in this scheme?<mask> I file jointly, does it include my future spouse's earnings?<mask> about capital gains on inheritances or other investments? [NEWLINE] [NEWLINE] And it can't be just the W-2,<mask> that doesn't account for small business income, etc. [NEWLINE] [NEWLINE] I'm afraid it's going to have to be your entire tax return... don't see any way around that. Every year. For the rest of your life. [NEWLINE] [NEWLINE] And leaving aside the privacy concerns about that, and the complications of defining exactly<mask> constitutes "future earnings", and the additional paperwork, and the effort it takes for the investors to parse all of those tax returns<mask><mask> whatever the complicated rules are... [NEWLINE] [NEWLINE] There are all the concerns that others bring up.<mask> should people that have been successful in their careers pay way more in order to subsidize people that have gotten degrees in underwater basketweaving and never amount to much of anything. [NEWLINE] [NEWLINE] <mask> any scheme like this is going to have to have other people pay for the mistakes of others, rather than having those others pay for them. The total amount collected will end up being higher (<mask> risky investments are always more expensive, and<mask> of the above administrative burdens), and it will be paid by only a fraction of the earners.</s>
Label encoding: <s>Just a few lines in the tax return? How would that work? [NEWLINE] [NEWLINE] And what counts as "future earnings" in this scheme? If I file jointly, does it include my future spouse's earnings? How about capital gains on inheritances or other investments? [NEWLINE] [NEWLINE] And it can't be just the W-2, because that doesn't account for small business income, etc. [NEWLINE] [NEWLINE] I'm afraid it's going to have to be your entire tax return... don't see any way around that. Every year. For the rest of your life. [NEWLINE] [NEWLINE] And leaving aside the privacy concerns about that, and the complications of defining exactly what constitutes "future earnings", and the additional paperwork, and the effort it takes for the investors to parse all of those tax returns according to whatever the complicated rules are... [NEWLINE] [NEWLINE] There are all the concerns that others bring up. Why should people that have been successful in their careers pay way more in order to subsidize people that have gotten degrees in underwater basketweaving and never amount to much of anything. [NEWLINE] [NEWLINE] Because any scheme like this is going to have to have other people pay for the mistakes of others, rather than having those others pay for them. The total amount collected will end up being higher ( because risky investments are always more expensive, and because of the above administrative burdens), and it will be paid by only a fraction of the earners.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is an interesting question to an economist such<mask> myself.  Lets run through the scenario of<mask> happens to the guy who has the same exact GPA and degree<mask> you<mask> he cheated off of you.  On paper, you have the same knowledge and qualification.  Unbeknownst to the firm (one that requires a college degree) that hires one of you, the cheater does not have the knowledge that you posses.  Suppose that the firm hires him instead of you.  He attempts to perform the job (Design bridges, mix chemicals, do<mask> he does with his degree),<mask> is unable to do<mask><mask> he does not know<mask>.  The firm, not seeing a passable quality of labor from this guy, fires him.  The firm then returns to the pool of potential candidates.  You are selected for the position. <mask> you have the actual knowledge that the degree says you do, you perform you job functions regularly and the firm is content. [NEWLINE] [NEWLINE] The key observation to take away from this game is that either you or the cheater could be randomly selected by an employer and the outcome would be the same (degree dependent.  Some jobs are/might be doable without a degree).  The issue is not that cheating devalues your degree, it's that the amount of people that go through the same program that you do devalues it.</s>
Label encoding: <s>This is an interesting question to an economist such as myself.  Lets run through the scenario of what happens to the guy who has the same exact GPA and degree as you because he cheated off of you.  On paper, you have the same knowledge and qualification.  Unbeknownst to the firm (one that requires a college degree) that hires one of you, the cheater does not have the knowledge that you posses.  Suppose that the firm hires him instead of you.  He attempts to perform the job (Design bridges, mix chemicals, do what he does with his degree), but is unable to do so because he does not know how.  The firm, not seeing a passable quality of labor from this guy, fires him.  The firm then returns to the pool of potential candidates.  You are selected for the position.  Because you have the actual knowledge that the degree says you do, you perform you job functions regularly and the firm is content. [NEWLINE] [NEWLINE] The key observation to take away from this game is that either you or the cheater could be randomly selected by an employer and the outcome would be the same (degree dependent.  Some jobs are/might be doable without a degree).  The issue is not that cheating devalues your degree, it's that the amount of people that go through the same program that you do devalues it.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When you say that porn's effect on society has been negative, you need to articulate<mask> kind of society you envision.  Society invented porn and proliferated it liberally<mask> porn served a cultural purpose.  No one ordered people to watch porn or participate in it.  We choose to watch and to create it, and porn is extremely diverse and getting even more<mask> with the advent of the internet and a digital camera in every hand.  Society is not harmed by porn any more than it is harmed by Christmas or anything else that has been produced and consumed at large.  It's not that society has been corrupted by this mysterious entity called porn which crash-landed on Earth from beyond.  Society *produced* porn<mask> a fiction,<mask> an art,<mask> something to be used and even privately (arguably publicly) celebrated. <mask> you blame porn for society, you're essentially putting the cart before the horse.  Society would have to be remedied<mask><mask> to no longer have a need for consuming the kind of porn you find offensive.  Countries that ban pornography are denying creature comforts in the same way that communism did.  And for<mask>? For one person or one pedantic individual's view of<mask> society ought to be? [NEWLINE] [NEWLINE] TL;DR Porn follows society,<mask> society is not the innocent entity being corrupted, it's *actually* the porn itself.</s>
Label encoding: <s>When you say that porn's effect on society has been negative, you need to articulate what kind of society you envision.  Society invented porn and proliferated it liberally because porn served a cultural purpose.  No one ordered people to watch porn or participate in it.  We choose to watch and to create it, and porn is extremely diverse and getting even more so with the advent of the internet and a digital camera in every hand.  Society is not harmed by porn any more than it is harmed by Christmas or anything else that has been produced and consumed at large.  It's not that society has been corrupted by this mysterious entity called porn which crash-landed on Earth from beyond.  Society *produced* porn as a fiction, as an art, as something to be used and even privately (arguably publicly) celebrated.  When you blame porn for society, you're essentially putting the cart before the horse.  Society would have to be remedied so as to no longer have a need for consuming the kind of porn you find offensive.  Countries that ban pornography are denying creature comforts in the same way that communism did.  And for what? For one person or one pedantic individual's view of how society ought to be? [NEWLINE] [NEWLINE] TL;DR Porn follows society, therefore society is not the innocent entity being corrupted, it's *actually* the porn itself.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Going along with this, I wish it were possible for every child to have an equivalent upbringing, with access to equal resources (food, water, shelter, education, etc...). I'm not a fan of our "your life depends upon who your parents happen to be and<mask> much money they have" system, which leaves millions of kids mired in hopeless poverty and gives a precious few an absurd amount of wealth and opportunity simply for having been born to the right parents. [NEWLINE] [NEWLINE] Unfortunately, I don't really see<mask> this could be fixed without taking all kids away from their parents and raising them communally. Wealthier parents aren't going to voluntarily give up their nice homes and expensive lifestyles just<mask> that they can live with their kids in whatever housing the worst-off parents can afford (which would be necessary to keep families together),<mask> kids would need to be taken away from parents and given equal (or at least equivalent) upbringings in order to make that kind of system work. Obviously that kind of system has a lot of problems with it, probably ones that are even worse than those of our current "system." [NEWLINE] [NEWLINE] Any ideas on<mask> you would organize society to fix this issue<mask> you were behind the veil of ignorance? I can't really think of any policies that would achieve the goals of (1) increased equality of opportunity for kids and (2) keeping families together.</s>
Label encoding: <s>Going along with this, I wish it were possible for every child to have an equivalent upbringing, with access to equal resources (food, water, shelter, education, etc...). I'm not a fan of our "your life depends upon who your parents happen to be and how much money they have" system, which leaves millions of kids mired in hopeless poverty and gives a precious few an absurd amount of wealth and opportunity simply for having been born to the right parents. [NEWLINE] [NEWLINE] Unfortunately, I don't really see how this could be fixed without taking all kids away from their parents and raising them communally. Wealthier parents aren't going to voluntarily give up their nice homes and expensive lifestyles just so that they can live with their kids in whatever housing the worst-off parents can afford (which would be necessary to keep families together), so kids would need to be taken away from parents and given equal (or at least equivalent) upbringings in order to make that kind of system work. Obviously that kind of system has a lot of problems with it, probably ones that are even worse than those of our current "system." [NEWLINE] [NEWLINE] Any ideas on how you would organize society to fix this issue if you were behind the veil of ignorance? I can't really think of any policies that would achieve the goals of (1) increased equality of opportunity for kids and (2) keeping families together.</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do *you* read french? Have you viewed Wolinski's disgusting "satires", even before the anti-islam craze? People of the generation of my parents suffered from them,<mask> it gave sexism a left-wing cover. Charlie Hebdo has always been a putrid magazine, and yes<mask>, it is comparable to the crap anti-semitic publications, maybe not those of WWII,<mask> definitely those of the 1890s. Those magazines used "satire" to entice the mobs to go against the pro-Dreyfus and the jews in general; just<mask> Charlie Hebdo used "satire" to entice people to go after muslim people. [NEWLINE] [NEWLINE] Even to hail Charlie Hebdo<mask> satire is wrong to me. Satire,<mask> its inception during the Enlightement, was about making fun of the powerful, of the have-it-all, of the ruling people. "satire" of poor, landless people, be it overt or throught ridiculous religious stereotype, is plain mockery. Le Canard Enchaîné does satire, Charlie Hebdo never did. [NEWLINE] [NEWLINE] Charlie Hebdo always claimed it did it to fight religious extremism;<mask> it never honestly aimed that. It mocked religion and religious people for easy sales, sure,<mask> never tried to actually fight extremism. It never actually tried to engage extremism or religion in any intellectual ways.</s>
Label encoding: <s>Do *you* read french? Have you viewed Wolinski's disgusting "satires", even before the anti-islam craze? People of the generation of my parents suffered from them, because it gave sexism a left-wing cover. Charlie Hebdo has always been a putrid magazine, and yes indeed, it is comparable to the crap anti-semitic publications, maybe not those of WWII, but definitely those of the 1890s. Those magazines used "satire" to entice the mobs to go against the pro-Dreyfus and the jews in general; just as Charlie Hebdo used "satire" to entice people to go after muslim people. [NEWLINE] [NEWLINE] Even to hail Charlie Hebdo as satire is wrong to me. Satire, since its inception during the Enlightement, was about making fun of the powerful, of the have-it-all, of the ruling people. "satire" of poor, landless people, be it overt or throught ridiculous religious stereotype, is plain mockery. Le Canard Enchaîné does satire, Charlie Hebdo never did. [NEWLINE] [NEWLINE] Charlie Hebdo always claimed it did it to fight religious extremism; but it never honestly aimed that. It mocked religion and religious people for easy sales, sure, but never tried to actually fight extremism. It never actually tried to engage extremism or religion in any intellectual ways.</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What words did I put on the mouth of the designer and wearer? I am repeating words the people who raised all this ruckus first wrote, and am trying to see<mask> it fits some logical criteria. [NEWLINE] [NEWLINE] Please, do inspect the images. I'm a comic artist. These clothes were normal superhero wear in the nineties. No private parts are uncovered, no woman appears distressed<mask> of some slavery issue.<mask><mask>, you'll find they fit most "sexy" Halloween costumes, in that they imply<mask> do not reveal. They do that by being shiny and skin tight.<mask> anything, all these women represent trash action heroines of the nineties. [NEWLINE] [NEWLINE] <mask> can I prove any of these fictional women had a choice? Well the designer's word is law in that occasion. She designed them to look like this, and gave the shirt to her friend<mask> a present.<mask> she wasn't his friend, then maybe one could claim a Trojan Horse in the form of a gift, that he thought was one thing at a glance, and under careful inspection would look like something damning.<mask> in that case, again, the maliciousness doesn't fall on the wearer<mask> on the artist. [NEWLINE] [NEWLINE] By the way, a woman who looks at you with a grim expression and is brandishing brutal weapons would have to be acting<mask> she were really thinking "hee-hee".</s>
Label encoding: <s>What words did I put on the mouth of the designer and wearer? I am repeating words the people who raised all this ruckus first wrote, and am trying to see if it fits some logical criteria. [NEWLINE] [NEWLINE] Please, do inspect the images. I'm a comic artist. These clothes were normal superhero wear in the nineties. No private parts are uncovered, no woman appears distressed because of some slavery issue. In fact, you'll find they fit most "sexy" Halloween costumes, in that they imply but do not reveal. They do that by being shiny and skin tight. If anything, all these women represent trash action heroines of the nineties. [NEWLINE] [NEWLINE] How can I prove any of these fictional women had a choice? Well the designer's word is law in that occasion. She designed them to look like this, and gave the shirt to her friend as a present. If she wasn't his friend, then maybe one could claim a Trojan Horse in the form of a gift, that he thought was one thing at a glance, and under careful inspection would look like something damning. But in that case, again, the maliciousness doesn't fall on the wearer but on the artist. [NEWLINE] [NEWLINE] By the way, a woman who looks at you with a grim expression and is brandishing brutal weapons would have to be acting if she were really thinking "hee-hee".</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the conflict is not<mask> much between a logical and an emotional side of you,<mask> between a semantic and a practical side. [NEWLINE] [NEWLINE] You read a story with a strong beautiful female OC in it, that you can identify with, and recall reading something about these people being "Mary Sues" and that being bad,<mask> *in practice* you still like it,<mask> you feel shameful about that. [NEWLINE] [NEWLINE] The problem with this is that it reverses the causality of<mask> certain tropes are considered bad in the first place. [NEWLINE] [NEWLINE] People first find a trope unenjoyable, and then *start calling it bad for that reason alone*.<mask> a supposed example is not,<mask><mask>, unenjoyable, then it loses the justification for being lumped with that "bad characters" group in the first place. [NEWLINE] [NEWLINE] The reason<mask> Mary Sue exists<mask> an avoided trope, is that they tend to provide boringly easy plots, and<mask> writer avatars, receive gifts that the *writer* would enjoy, rather than the reader. [NEWLINE] [NEWLINE] <mask> you read a semantically "Mary Sue" character who is<mask><mask> a successfully written audience avatar, and serves in-universe<mask> an explicit beacon of goodness with the conflict still staying relevant, then there is not much point in calling that a Mary Sue, let alone shaming you for liking for it. [NEWLINE] </s>
Label encoding: <s>I think the conflict is not so much between a logical and an emotional side of you, as between a semantic and a practical side. [NEWLINE] [NEWLINE] You read a story with a strong beautiful female OC in it, that you can identify with, and recall reading something about these people being "Mary Sues" and that being bad, yet *in practice* you still like it, so you feel shameful about that. [NEWLINE] [NEWLINE] The problem with this is that it reverses the causality of why certain tropes are considered bad in the first place. [NEWLINE] [NEWLINE] People first find a trope unenjoyable, and then *start calling it bad for that reason alone*. If a supposed example is not, in fact, unenjoyable, then it loses the justification for being lumped with that "bad characters" group in the first place. [NEWLINE] [NEWLINE] The reason why Mary Sue exists as an avoided trope, is that they tend to provide boringly easy plots, and as writer avatars, receive gifts that the *writer* would enjoy, rather than the reader. [NEWLINE] [NEWLINE] If you read a semantically "Mary Sue" character who is in fact a successfully written audience avatar, and serves in-universe as an explicit beacon of goodness with the conflict still staying relevant, then there is not much point in calling that a Mary Sue, let alone shaming you for liking for it. [NEWLINE] </s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It sounds like the crux of your disagreement is that you think that a democracy is defined<mask> a state in which citizens vote directly on policy, instead of representatives.  You're explicitly contradicted by every dictionary and encyclopedia that I've ever read. <mask> are you basing your definition on?  Everything I'm able to find indicates that representative democracy is very much a type of democracy (which should be clear from the fact that the root word of the phrase is "democracy"). [NEWLINE] [NEWLINE] [URL] [NEWLINE] [URL] [NEWLINE] [URL] [NEWLINE] [NEWLINE] (yes yes, I know that Wikipedia isn't in and of itself a definitional source,<mask> the citation links for things like definitions are always useful). [NEWLINE] [NEWLINE] [URL] [NEWLINE] [URL].com/dictionary/democracy [NEWLINE] [URL] [NEWLINE] [NEWLINE] The term you're looking for, wherein citizens directly propose and vote on policy, is called "direct democracy", which is another form of democracy (just<mask> representative democracy is).  Language is of course descriptivist to a certain degree,<mask><mask> I start telling ~~their people~~ people they're wrong<mask> they're calling their orange-colored citrus fruits "oranges" instead of "bananas" (which<mask><mask> they should be called), it would be safe to say that I'd be "wrong".  In large part<mask> I'd be contradicted by everything from dictionaries to encyclopedias to overwhelming common usage.</s>
Label encoding: <s>It sounds like the crux of your disagreement is that you think that a democracy is defined as a state in which citizens vote directly on policy, instead of representatives.  You're explicitly contradicted by every dictionary and encyclopedia that I've ever read.  What are you basing your definition on?  Everything I'm able to find indicates that representative democracy is very much a type of democracy (which should be clear from the fact that the root word of the phrase is "democracy"). [NEWLINE] [NEWLINE] [URL] [NEWLINE] [URL] [NEWLINE] [URL] [NEWLINE] [NEWLINE] (yes yes, I know that Wikipedia isn't in and of itself a definitional source, but the citation links for things like definitions are always useful). [NEWLINE] [NEWLINE] [URL] [NEWLINE] [URL].com/dictionary/democracy [NEWLINE] [URL] [NEWLINE] [NEWLINE] The term you're looking for, wherein citizens directly propose and vote on policy, is called "direct democracy", which is another form of democracy (just as representative democracy is).  Language is of course descriptivist to a certain degree, but if I start telling ~~their people~~ people they're wrong because they're calling their orange-colored citrus fruits "oranges" instead of "bananas" (which I think they should be called), it would be safe to say that I'd be "wrong".  In large part because I'd be contradicted by everything from dictionaries to encyclopedias to overwhelming common usage.</s>
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is more behind school than you might imagine. The school is interessted in making their pupils into functioning members of society, they don't teach you math<mask> you can solve these very specific equations you learn during highschool, they want you to understand the greater logic behind it. The same goes for sport, the school isn't interessted in teaching you<mask> to play soccer just for the sake of it or<mask> they want to make you into a star soccer player,<mask> it has other benefits too: [NEWLINE] [NEWLINE] - People who are doing sport are generaly fitter than those who don't, that doesn't only decrease obesity<mask><mask> makes for better students overall [NEWLINE] - People who attend sports regularly without being forced to (like they are<mask> attending schools) will learn to form habits and stick to them<mask> they are benefitial,<mask><mask> its uncomfortable from time to time (For example<mask> its raining outside, you really aren't in the mood for it or something like that) you are not giving up<mask> easily [NEWLINE] - You learn things like teamwork or the ability to make quick decisions or overall skills that are usefull<mask> playing sports,<mask> are<mask> usefull somewhere else [NEWLINE] - The social factors is<mask> pretty huge, the school doesn't want 1000 students who are foreign to one another, its generally better for everyone<mask> students are binding, sport is excellent for that</s>
Label encoding: <s>There is more behind school than you might imagine. The school is interessted in making their pupils into functioning members of society, they don't teach you math so you can solve these very specific equations you learn during highschool, they want you to understand the greater logic behind it. The same goes for sport, the school isn't interessted in teaching you how to play soccer just for the sake of it or because they want to make you into a star soccer player, but it has other benefits too: [NEWLINE] [NEWLINE] - People who are doing sport are generaly fitter than those who don't, that doesn't only decrease obesity but also makes for better students overall [NEWLINE] - People who attend sports regularly without being forced to (like they are when attending schools) will learn to form habits and stick to them when they are benefitial, even though its uncomfortable from time to time (For example when its raining outside, you really aren't in the mood for it or something like that) you are not giving up as easily [NEWLINE] - You learn things like teamwork or the ability to make quick decisions or overall skills that are usefull when playing sports, but are also usefull somewhere else [NEWLINE] - The social factors is also pretty huge, the school doesn't want 1000 students who are foreign to one another, its generally better for everyone when students are binding, sport is excellent for that</s>
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Sexual jealousy could be innate. The male wants to pass on his genes.<mask> his mate is having sex with another male not only could she get pregnant with another male's child,<mask> he will be expending energy to raise that child. [ENDQ] [NEWLINE] Exactly. This is the man's point of view.<mask> would the woman point of view be?<mask> we see reproduction<mask> the final aim, then she should have sex with<mask> many different partners<mask> possible,<mask> her genes have the highest chance of survival. This is<mask><mask><mask> it is manmade. [NEWLINE] [NEWLINE] [STARTQ] Some animals have pack system<mask> who gets to mate (and who doesn't) is very important. Animals will fight - sometimes to the death - for the right to mate. Trying to mate with the pack leader's harem can be risky. In other words, animals show sexual jealousy. [ENDQ] [NEWLINE] From a biology point of view, we share more DNA with bonobos than with most other species. I can't remember the exact figure,<mask> bonobos mate to mate 1 to 4 times per hour with up to a dozen partners. [NEWLINE] [NEWLINE] [STARTQ] Or in response to your last sentence, Man didn't invent sexual jealousy to be sure who their kids are. Mother Nature invented sexual jealousy to be sure who's the father of the kids. [ENDQ] [NEWLINE] The father is not essential for the kids, is it?</s>
Label encoding: <s> [STARTQ] Sexual jealousy could be innate. The male wants to pass on his genes. If his mate is having sex with another male not only could she get pregnant with another male's child, but he will be expending energy to raise that child. [ENDQ] [NEWLINE] Exactly. This is the man's point of view. What would the woman point of view be? If we see reproduction as the final aim, then she should have sex with as many different partners as possible, so her genes have the highest chance of survival. This is why I think it is manmade. [NEWLINE] [NEWLINE] [STARTQ] Some animals have pack system where who gets to mate (and who doesn't) is very important. Animals will fight - sometimes to the death - for the right to mate. Trying to mate with the pack leader's harem can be risky. In other words, animals show sexual jealousy. [ENDQ] [NEWLINE] From a biology point of view, we share more DNA with bonobos than with most other species. I can't remember the exact figure, but bonobos mate to mate 1 to 4 times per hour with up to a dozen partners. [NEWLINE] [NEWLINE] [STARTQ] Or in response to your last sentence, Man didn't invent sexual jealousy to be sure who their kids are. Mother Nature invented sexual jealousy to be sure who's the father of the kids. [ENDQ] [NEWLINE] The father is not essential for the kids, is it?</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is entirely possible your son was misdiagnosed. It is<mask> entirely possible that autism is over-diagnosed in general.<mask> they're unrelated issues.<mask> your son was misdiagnosed,  it would not be indicative of autism being over-diagnosed. [NEWLINE] [NEWLINE] Now, I want to draw your attention to something you said: [NEWLINE] [STARTQ] Lots of kids get placed in the PDD-NOS category,<mask> it gets them social services that will help with developmental delays,<mask><mask> their cause. ABA (Advanced Behavioral something...) is actually pretty awesome, and I feel like it would benefit every kid,<mask><mask> delay or diagnosis. [ENDQ] [NEWLINE] From this standpoint, is a potential for over-diagnosing autism necessarily a bad thing? The goal is to try and help<mask> many people<mask> possible. The only way to do this is to err on the safe side and diagnose kids early and readily in case they truly do need the help. [NEWLINE] [NEWLINE] This sort of thing is not an exact science, it isn't like there is some clear identifier for autism and there are a lot of variables to take into account.<mask> a diagnoses is primarily symptom-based instead of based on physical evidence (like literally finding the cancer or a blood test confirming diabetes) you run the risk of misdiagnosing someone. People can display symptoms of some malady without necessarily suffering from it.</s>
Label encoding: <s>It is entirely possible your son was misdiagnosed. It is also entirely possible that autism is over-diagnosed in general. But they're unrelated issues. If your son was misdiagnosed,  it would not be indicative of autism being over-diagnosed. [NEWLINE] [NEWLINE] Now, I want to draw your attention to something you said: [NEWLINE] [STARTQ] Lots of kids get placed in the PDD-NOS category, as it gets them social services that will help with developmental delays, regardless of their cause. ABA (Advanced Behavioral something...) is actually pretty awesome, and I feel like it would benefit every kid, regardless of delay or diagnosis. [ENDQ] [NEWLINE] From this standpoint, is a potential for over-diagnosing autism necessarily a bad thing? The goal is to try and help as many people as possible. The only way to do this is to err on the safe side and diagnose kids early and readily in case they truly do need the help. [NEWLINE] [NEWLINE] This sort of thing is not an exact science, it isn't like there is some clear identifier for autism and there are a lot of variables to take into account. When a diagnoses is primarily symptom-based instead of based on physical evidence (like literally finding the cancer or a blood test confirming diabetes) you run the risk of misdiagnosing someone. People can display symptoms of some malady without necessarily suffering from it.</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I admit I wasn't crying out for touch-ups to some late PS3 releases,<mask> now that I've played them I'm very glad they exist. Sometimes we don't always know<mask> we want. [NEWLINE] [NEWLINE] <mask> for<mask> often... hard to say. It depends largely on the game and standards of the day. I'd say<mask> the game already plays at a smooth consistent 60 frames per second and at least 720p, it's probably fine<mask> it is. I'd say it's not worth remastering anything at those standards until we can reliably get 4k gaming at a minimum of 60 frames per second<mask> a standard.<mask> 7th gen, 6th gen, or portable games that still run at lesser framerates, with graphical tearing, no anti-aliasing, sub-HD etc. I could see almost any game like that justifiably being remastered today. [NEWLINE] [NEWLINE] Edit: Oh, and here in UK the remastered versions of games have been releasing at about 2/3rd the normal full price, and usually have a price drop after the first month. I picked up 3 remastered games for £64. Considering most games in UK release between £45 and £55 (and the latter is getting more common, thanks Call Of fucking Duty); that's pretty good value... worked out at just over £21 each on average.</s>
Label encoding: <s>I admit I wasn't crying out for touch-ups to some late PS3 releases, but now that I've played them I'm very glad they exist. Sometimes we don't always know what we want. [NEWLINE] [NEWLINE] As for how often... hard to say. It depends largely on the game and standards of the day. I'd say if the game already plays at a smooth consistent 60 frames per second and at least 720p, it's probably fine as it is. I'd say it's not worth remastering anything at those standards until we can reliably get 4k gaming at a minimum of 60 frames per second as a standard. But 7th gen, 6th gen, or portable games that still run at lesser framerates, with graphical tearing, no anti-aliasing, sub-HD etc. I could see almost any game like that justifiably being remastered today. [NEWLINE] [NEWLINE] Edit: Oh, and here in UK the remastered versions of games have been releasing at about 2/3rd the normal full price, and usually have a price drop after the first month. I picked up 3 remastered games for £64. Considering most games in UK release between £45 and £55 (and the latter is getting more common, thanks Call Of fucking Duty); that's pretty good value... worked out at just over £21 each on average.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Were you ever shamed<mask> a child to act properly?  Would you ever shame a child into acting properly?  This is a serious question and I'd like you to answer it. [NEWLINE] [NEWLINE] I DON'T disagree with you about people losing their jobs from off-handed comments.  I'm actually TOTALLY with you on that stuff. <mask><mask> the pervasiveness of social media and people's natural tendency to let their guard down gets them into shit they otherwise would never have been in.  People will adapt and just not do it<mask> much. [NEWLINE] [NEWLINE] Doxxing and harassing a person who made a joke you could see Sarah Silverman doing is ridiculous.  Furthermore, it's not showing that shaming, in and of itself, is wrong or censorious.  The cases you bring up are cases of brigading against individuals who were not acting in an official capacity representing their brand. <mask><mask> CEOs are ALWAYS in that capacity, and no one should bat an eye for that Mozilla guy. <mask> you're a high-ranking official, you are representing your company, just like they used to say on school field trips "You're representing your school,<mask> behave!"  Boycotts are a real tool of social change, and they helped end apartheid.  Jesus, apartheid - the world community shamed South Africa out of that.</s>
Label encoding: <s>Were you ever shamed as a child to act properly?  Would you ever shame a child into acting properly?  This is a serious question and I'd like you to answer it. [NEWLINE] [NEWLINE] I DON'T disagree with you about people losing their jobs from off-handed comments.  I'm actually TOTALLY with you on that stuff.  I think the pervasiveness of social media and people's natural tendency to let their guard down gets them into shit they otherwise would never have been in.  People will adapt and just not do it as much. [NEWLINE] [NEWLINE] Doxxing and harassing a person who made a joke you could see Sarah Silverman doing is ridiculous.  Furthermore, it's not showing that shaming, in and of itself, is wrong or censorious.  The cases you bring up are cases of brigading against individuals who were not acting in an official capacity representing their brand.  I think CEOs are ALWAYS in that capacity, and no one should bat an eye for that Mozilla guy.  If you're a high-ranking official, you are representing your company, just like they used to say on school field trips "You're representing your school, so behave!"  Boycotts are a real tool of social change, and they helped end apartheid.  Jesus, apartheid - the world community shamed South Africa out of that.</s>
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The first is the utter lack of empathy. Having a child with sever medical problems is incredibly taxing. It could happen to anyone through no fault of their own, and<mask> such we have decided<mask> a society to spread that risk around to limit its severity. [ENDQ] [NEWLINE] <mask> much are you willing to spend<mask> that a child grows to be three days older? One million? Three hundred million? Empathy has its limits.<mask> a society we have to determine an amount we are willing to spend on different scenarios. It may be cruel,<mask><mask> is allowing a heart-stricken parent sign off on for an incredible amount of debt to see their little one not die for another day. [NEWLINE] [NEWLINE] [STARTQ] The second flaw is ignoring the fallout from not assisting in any way. You're taking one of the most powerful motivating forces in a humans life, the well being of their child, and placing them in a situation<mask> they have no legal recourse to saving it. That would potentially result in lots of illegal activity<mask> the only means of providing support. Desperate people do desperate things. [ENDQ] [NEWLINE] ∆ [NEWLINE] [NEWLINE] You make an interesting point. I do believe that there could be a slight increase from illegal activities, I don't think it would have an impact on the scale that would negate the benefit of the program.<mask> many crimes could the parent commit before they are found? [NEWLINE] </s>
Label encoding: <s> [STARTQ] The first is the utter lack of empathy. Having a child with sever medical problems is incredibly taxing. It could happen to anyone through no fault of their own, and as such we have decided as a society to spread that risk around to limit its severity. [ENDQ] [NEWLINE] How much are you willing to spend so that a child grows to be three days older? One million? Three hundred million? Empathy has its limits. As a society we have to determine an amount we are willing to spend on different scenarios. It may be cruel, but so is allowing a heart-stricken parent sign off on for an incredible amount of debt to see their little one not die for another day. [NEWLINE] [NEWLINE] [STARTQ] The second flaw is ignoring the fallout from not assisting in any way. You're taking one of the most powerful motivating forces in a humans life, the well being of their child, and placing them in a situation where they have no legal recourse to saving it. That would potentially result in lots of illegal activity as the only means of providing support. Desperate people do desperate things. [ENDQ] [NEWLINE] ∆ [NEWLINE] [NEWLINE] You make an interesting point. I do believe that there could be a slight increase from illegal activities, I don't think it would have an impact on the scale that would negate the benefit of the program. How many crimes could the parent commit before they are found? [NEWLINE] </s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's for a moment forget about the actual compositional qualities of trap music and think more about its application. [NEWLINE] [NEWLINE] From my experience, the only time I ever really hear trap music (<mask> a person who would not choose to listen to trap<mask> a genre frequently) is<mask> I'm in a club that's playing it, or<mask> someone puts it on at a house gathering or something along those lines. [NEWLINE] [NEWLINE] Trap is loaded with extremely simple verbal hooks, which are absolutely great to shout out with your friends after a few beers or<mask> you're in a party environment. [NEWLINE] [NEWLINE] I have the exact same opinion with Oasis, I find their music to be bland and unendurable<mask> I listen to it whilst sober,<mask> I will always grab a friend and sing along<mask> it's played in a social situation. [NEWLINE] [NEWLINE] Now imagine that scenario,<mask> with an individual being exposed to trap more and more frequently, until the point<mask> it "grows on them" for lack of a better term. The first time I heard Sugar Coated Sour by Dillinger Escape Plan, I thought Mathcore was just pretentious noise and unlistenable,<mask> now it's my favourite genre. [NEWLINE] [NEWLINE] People may understand that the music itself isn't the most groundbreaking step on into songwriting,<mask> it fits its purpose, and that's<mask> it's good.</s>
Label encoding: <s>Let's for a moment forget about the actual compositional qualities of trap music and think more about its application. [NEWLINE] [NEWLINE] From my experience, the only time I ever really hear trap music ( as a person who would not choose to listen to trap as a genre frequently) is when I'm in a club that's playing it, or when someone puts it on at a house gathering or something along those lines. [NEWLINE] [NEWLINE] Trap is loaded with extremely simple verbal hooks, which are absolutely great to shout out with your friends after a few beers or when you're in a party environment. [NEWLINE] [NEWLINE] I have the exact same opinion with Oasis, I find their music to be bland and unendurable if I listen to it whilst sober, but I will always grab a friend and sing along if it's played in a social situation. [NEWLINE] [NEWLINE] Now imagine that scenario, but with an individual being exposed to trap more and more frequently, until the point where it "grows on them" for lack of a better term. The first time I heard Sugar Coated Sour by Dillinger Escape Plan, I thought Mathcore was just pretentious noise and unlistenable, but now it's my favourite genre. [NEWLINE] [NEWLINE] People may understand that the music itself isn't the most groundbreaking step on into songwriting, but it fits its purpose, and that's why it's good.</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0906, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Easements [ENDQ] [NEWLINE] This is a non-starter.  Cats are considered property.  Easements are not in place for you to utilize my property for the enjoyment of your property. [NEWLINE] [NEWLINE] [STARTQ] Precedence and societal consensus. It's likely that the cat population has been around for much longer than you, the house owner. [ENDQ] [NEWLINE] <mask><mask><mask> reasoning this is a poor one. This is de facto  Appeal to Tradition, traditionally Animal Control is tasked with dealing with stray animals, which<mask> you let your cat wander it is. [NEWLINE] [NEWLINE] [STARTQ] societal consensus [ENDQ] [NEWLINE] Again you are correct. <mask><mask> societal consensus your cat<mask> it is roaming is a stray and<mask> such is under the purview of Animal Control. [NEWLINE] [NEWLINE] [STARTQ] Limited liability ----<mask> many cities have a policy of tolerating them, and in those cities the cats have a right to exist just like any other inhabitant. [ENDQ] [NEWLINE] <mask><mask> I'd need a citation here. [NEWLINE] [NEWLINE] <mask> you have a cat and let it roam it is a stray cat<mask> it is outside of your supervision.  You do not posses the rights to let your domestic animals use my property.  It is that simple. [NEWLINE] [NEWLINE] <mask> I have problems with cats, I trap them and turn them over to Animal Control. <mask>? <mask> they destroy wildlife and property.  </s>
Label encoding: <s> [STARTQ] Easements [ENDQ] [NEWLINE] This is a non-starter.  Cats are considered property.  Easements are not in place for you to utilize my property for the enjoyment of your property. [NEWLINE] [NEWLINE] [STARTQ] Precedence and societal consensus. It's likely that the cat population has been around for much longer than you, the house owner. [ENDQ] [NEWLINE] As far as reasoning this is a poor one. This is de facto  Appeal to Tradition, traditionally Animal Control is tasked with dealing with stray animals, which when you let your cat wander it is. [NEWLINE] [NEWLINE] [STARTQ] societal consensus [ENDQ] [NEWLINE] Again you are correct.  According to societal consensus your cat when it is roaming is a stray and as such is under the purview of Animal Control. [NEWLINE] [NEWLINE] [STARTQ] Limited liability ---- But many cities have a policy of tolerating them, and in those cities the cats have a right to exist just like any other inhabitant. [ENDQ] [NEWLINE] I think I'd need a citation here. [NEWLINE] [NEWLINE] If you have a cat and let it roam it is a stray cat while it is outside of your supervision.  You do not posses the rights to let your domestic animals use my property.  It is that simple. [NEWLINE] [NEWLINE] If I have problems with cats, I trap them and turn them over to Animal Control.  Why?  Because they destroy wildlife and property.  </s>
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll tell you<mask> I'm coming from. I have<mask> experimented. I had ideas about heterosexuality being mostly a social construct and I didn't want to live a crippled life due to heterosexist socialization. After several attempts and a lot of thinking, reluctantly I had to accept that I'm hetero. I'm not capable of getting aroused by female genitals. I'm not grossed out by them they're just neutral to me. I don't mind being intimate with women<mask> it doesn't turn me on either. It just feels "friendly". [NEWLINE] [NEWLINE] Now I have chosen a lot of other things. For example I used to be ashamed of my fetishes<mask> I was young.<mask> my basic sexuality is "hetero" and that's not a choice. [NEWLINE] [NEWLINE] I'm not 100% convinced of anything<mask> it comes to these issues<mask> I don't think there are any bulletproof answers.<mask> your story, and my story both indicate to me that we have certain inclinations that may be partially genetically influenced and possibly take a more solid form<mask> we have our first sexual experiences. Then we can grow<mask> humans and develop and we can embrace or suppress desires.<mask> I don't think you can choose<mask> you are sexually attracted to. [NEWLINE] [NEWLINE] <mask>.. i guess the answer is, it depends on<mask> you mean by sexuality. :)</s>
Label encoding: <s>I'll tell you where I'm coming from. I have also experimented. I had ideas about heterosexuality being mostly a social construct and I didn't want to live a crippled life due to heterosexist socialization. After several attempts and a lot of thinking, reluctantly I had to accept that I'm hetero. I'm not capable of getting aroused by female genitals. I'm not grossed out by them they're just neutral to me. I don't mind being intimate with women but it doesn't turn me on either. It just feels "friendly". [NEWLINE] [NEWLINE] Now I have chosen a lot of other things. For example I used to be ashamed of my fetishes when I was young. But my basic sexuality is "hetero" and that's not a choice. [NEWLINE] [NEWLINE] I'm not 100% convinced of anything when it comes to these issues because I don't think there are any bulletproof answers. But your story, and my story both indicate to me that we have certain inclinations that may be partially genetically influenced and possibly take a more solid form when we have our first sexual experiences. Then we can grow as humans and develop and we can embrace or suppress desires. But I don't think you can choose what you are sexually attracted to. [NEWLINE] [NEWLINE] So.. i guess the answer is, it depends on what you mean by sexuality. :)</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off you seem to have ignored the larger part of my argument<mask> ok [NEWLINE] [NEWLINE] [STARTQ] Absolutely minstrel shows were and are racist,<mask> the actions of this kid and Ms. Hough are in no way similar. [ENDQ] [NEWLINE] the wearer's intentions are irrelevant. Just<mask> you ignorantly offend someone doesn't make<mask> you say/do less offensive. [NEWLINE] sure being ignorant and genuinely not meaning to offend someone is slightly more forgivable,<mask> doesn't mean you get to run around and continue doing ignorant things. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] By saying 'this is a part of my person that you aren't allowed to talk about, joke about, appreciate, or associate with', you are making skin color out to be more determining than it actually is. I don't see skin color<mask> any real factor for one's character, ie my being white doesn't automatically lead to x,y,z nor should I be lead to think<mask>. [ENDQ] [NEWLINE] [NEWLINE] maybe for you,<mask> you haven't experienced being part of a marginalized group based on your skin color.<mask> for people of color not<mask> much. Just<mask> you say it isnt that way doesnt mean that it is actual reality. We are not determining that skin color is more than it is, it has *already been done for us* [NEWLINE] and wanting to paint yourself darker is literally reinforcing that. </s>
Label encoding: <s>First off you seem to have ignored the larger part of my argument but ok [NEWLINE] [NEWLINE] [STARTQ] Absolutely minstrel shows were and are racist, but the actions of this kid and Ms. Hough are in no way similar. [ENDQ] [NEWLINE] the wearer's intentions are irrelevant. Just because you ignorantly offend someone doesn't make what you say/do less offensive. [NEWLINE] sure being ignorant and genuinely not meaning to offend someone is slightly more forgivable, but doesn't mean you get to run around and continue doing ignorant things. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] By saying 'this is a part of my person that you aren't allowed to talk about, joke about, appreciate, or associate with', you are making skin color out to be more determining than it actually is. I don't see skin color as any real factor for one's character, ie my being white doesn't automatically lead to x,y,z nor should I be lead to think so. [ENDQ] [NEWLINE] [NEWLINE] maybe for you, because you haven't experienced being part of a marginalized group based on your skin color. But for people of color not so much. Just because you say it isnt that way doesnt mean that it is actual reality. We are not determining that skin color is more than it is, it has *already been done for us* [NEWLINE] and wanting to paint yourself darker is literally reinforcing that. </s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> saying stuff like 'that's emotional bullshit' is probably really hurtful to someone who was deeply affected like some people who already posted in this thread. Most men may be fine after incidences like this, some aren't. You should read some of these people's posts<mask> man, they're eye opening. [ENDQ] [NEWLINE] You're right. I just get<mask> tired of Reddit circle jerking that video pretending that is<mask> everyone would normally react. I need to control my reactions to these things. [NEWLINE] [NEWLINE] <mask><mask> really bugs me is this actor is talking about statutory rape and not actual rape. Notice that not once in the video did he ever says no to the adult. He even said he physically enjoyed it.<mask> I'm not saying that some boys couldn't react poorly. Some would,<mask> it isn't standard(maybe it is this way<mask> of society,<mask> the average teenage boy still doesn't feel taken advantage of). It just spreads the propaganda that teenage boys wouldn't enjoy sex with a hot adult girl (and vise versa), and they are actually being taken advantage of. It continues this idea that is okay to send adults to jail for years, let them actually get raped in jail multiple times, and put them on the sex offenders list for the rest of their life, all<mask> of something that in all likelihood the boy actually enjoyed.</s>
Label encoding: <s> [STARTQ] Although saying stuff like 'that's emotional bullshit' is probably really hurtful to someone who was deeply affected like some people who already posted in this thread. Most men may be fine after incidences like this, some aren't. You should read some of these people's posts though man, they're eye opening. [ENDQ] [NEWLINE] You're right. I just get so tired of Reddit circle jerking that video pretending that is how everyone would normally react. I need to control my reactions to these things. [NEWLINE] [NEWLINE] What also really bugs me is this actor is talking about statutory rape and not actual rape. Notice that not once in the video did he ever says no to the adult. He even said he physically enjoyed it. While I'm not saying that some boys couldn't react poorly. Some would, but it isn't standard(maybe it is this way because of society, but the average teenage boy still doesn't feel taken advantage of). It just spreads the propaganda that teenage boys wouldn't enjoy sex with a hot adult girl (and vise versa), and they are actually being taken advantage of. It continues this idea that is okay to send adults to jail for years, let them actually get raped in jail multiple times, and put them on the sex offenders list for the rest of their life, all because of something that in all likelihood the boy actually enjoyed.</s>
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can't just pull up one of the densest, driest passages you could find and use it<mask> a defining example of philosophy<mask> a whole. [NEWLINE] [NEWLINE] Like /u/MackDaddyVelli already mentioned in another comment, [NEWLINE] [NEWLINE] [STARTQ] The reason you have a hard time understanding the quote (and I do, too) is<mask> it's from Heidegger and Heidegger is *hard*. [ENDQ] [NEWLINE] The type of philsophy that could (and should) be taught to high school students is Intro/101 stuff, like "is the green I see the same<mask> the green you see?", brains in vats, etc. [NEWLINE] [NEWLINE] <mask> that material may be widely mocked in more sophisticated philosophical circles, it is *exactly* the kind of accessible material that could act<mask> a "philosophical stater-kit" for young minds, setting them off on a path of introspection and self-examination that was previously foreign to them. [NEWLINE] [NEWLINE] <mask> you start teaching kids math, you don't start with calculus; similarly,<mask> we were to introduce a philosophy curriculum to children, you would absolutely not see any Heidegger in the first class. Maybe you wouldn't see him at all, at any point - it's simply not *necessary* to reach that level of complexity just to get kids thinking analytically.</s>
Label encoding: <s>You can't just pull up one of the densest, driest passages you could find and use it as a defining example of philosophy as a whole. [NEWLINE] [NEWLINE] Like /u/MackDaddyVelli already mentioned in another comment, [NEWLINE] [NEWLINE] [STARTQ] The reason you have a hard time understanding the quote (and I do, too) is because it's from Heidegger and Heidegger is *hard*. [ENDQ] [NEWLINE] The type of philsophy that could (and should) be taught to high school students is Intro/101 stuff, like "is the green I see the same as the green you see?", brains in vats, etc. [NEWLINE] [NEWLINE] While that material may be widely mocked in more sophisticated philosophical circles, it is *exactly* the kind of accessible material that could act as a "philosophical stater-kit" for young minds, setting them off on a path of introspection and self-examination that was previously foreign to them. [NEWLINE] [NEWLINE] When you start teaching kids math, you don't start with calculus; similarly, if we were to introduce a philosophy curriculum to children, you would absolutely not see any Heidegger in the first class. Maybe you wouldn't see him at all, at any point - it's simply not *necessary* to reach that level of complexity just to get kids thinking analytically.</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I really think of Thomas<mask> a wild card.  I mean, he and Scalia are collectively the idiot manchildren of the Supreme Court, and nothing's going to change that,<mask> sometimes Thomas tosses in a real curveball.  Sometimes it's good, usually it's bad,<mask><mask> you can really count on is that it'll be *crazy*, like his lone dissent in Brown v. Entertainment Merchants Association (in which he invented, from whole cloth and without precedent, the idea that free speech did not extend to any speech involving children) or Smith v. Cain (<mask> he basically asserted that no amount of prosecutorial misconduct can *ever* be enough to reverse a conviction). [NEWLINE] [NEWLINE] Regardless,<mask>, Thomas's claims of "originalism" are window-dressing for a fairly complicated worldview that has a lot of nuance: it's steeped in this respect for Southern tradition, it has a complicated understanding of<mask> authority is vested and for<mask> purposes, it offers some respect to civil rights even<mask> it undermines legal mechanisms for addressing violations of those rights...  Scalia,<mask><mask><mask><mask>, uses his "originalism" to promote a strict hierarchy of RICH PEOPLE &gt; LAW ENFORCEMENT &gt; LEGISLATORS &gt; POOR PEOPLE, which makes his decisions *waaay* safer bets.</s>
Label encoding: <s>I really think of Thomas as a wild card.  I mean, he and Scalia are collectively the idiot manchildren of the Supreme Court, and nothing's going to change that, but sometimes Thomas tosses in a real curveball.  Sometimes it's good, usually it's bad, but what you can really count on is that it'll be *crazy*, like his lone dissent in Brown v. Entertainment Merchants Association (in which he invented, from whole cloth and without precedent, the idea that free speech did not extend to any speech involving children) or Smith v. Cain ( where he basically asserted that no amount of prosecutorial misconduct can *ever* be enough to reverse a conviction). [NEWLINE] [NEWLINE] Regardless, though, Thomas's claims of "originalism" are window-dressing for a fairly complicated worldview that has a lot of nuance: it's steeped in this respect for Southern tradition, it has a complicated understanding of where authority is vested and for what purposes, it offers some respect to civil rights even while it undermines legal mechanisms for addressing violations of those rights...  Scalia, on the other hand, uses his "originalism" to promote a strict hierarchy of RICH PEOPLE &gt; LAW ENFORCEMENT &gt; LEGISLATORS &gt; POOR PEOPLE, which makes his decisions *waaay* safer bets.</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is he somehow prevented from insisting upon a condom,<mask>? [ENDQ] [NEWLINE] No, of course not. The only reason<mask> I brought that up was<mask> I was expecting you to extrapolate that there are people out there who think taking a bc pill means no baby. Hell, there are people out there who think a bc pill will protect them from STD's, believe it or not. Guys should just use a condom by default,<mask> only to protect themselves,<mask> really to help protect both people involved. [NEWLINE] [NEWLINE] I was raising an example of<mask> better education and respect to the situation was needed, and simply forgot to expand upon it. Sorry about that. [NEWLINE] [NEWLINE] Anyway, preach personal responsibility on both sides of the sex line. That would be a better solution that presenting an argument which appears to be based solely on the fact that it is the woman who has to carry the child. [NEWLINE] [NEWLINE] To that end, doesn't saying "guys shouldn't have sex with a girl they wouldn't want to have a baby with" absolve the girl of knowing<mask> exactly she's getting in to<mask> she agrees to have consensual unprotected sex? Again, irrational. No, it is the responsibility of *both* parties equally to know the score, the consequences, and<mask> to deal with those consequences like adults. This can only be solved through better education. </s>
Label encoding: <s> [STARTQ] Is he somehow prevented from insisting upon a condom, though? [ENDQ] [NEWLINE] No, of course not. The only reason why I brought that up was because I was expecting you to extrapolate that there are people out there who think taking a bc pill means no baby. Hell, there are people out there who think a bc pill will protect them from STD's, believe it or not. Guys should just use a condom by default, if only to protect themselves, but really to help protect both people involved. [NEWLINE] [NEWLINE] I was raising an example of why better education and respect to the situation was needed, and simply forgot to expand upon it. Sorry about that. [NEWLINE] [NEWLINE] Anyway, preach personal responsibility on both sides of the sex line. That would be a better solution that presenting an argument which appears to be based solely on the fact that it is the woman who has to carry the child. [NEWLINE] [NEWLINE] To that end, doesn't saying "guys shouldn't have sex with a girl they wouldn't want to have a baby with" absolve the girl of knowing what exactly she's getting in to when she agrees to have consensual unprotected sex? Again, irrational. No, it is the responsibility of *both* parties equally to know the score, the consequences, and how to deal with those consequences like adults. This can only be solved through better education. </s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can lead a horse to water,<mask> you can't make it drink. [NEWLINE] [NEWLINE] I've pointed out that the subreddit is more than representative of the movement<mask> a whole — it basically *is* the entire movement. [NEWLINE] [NEWLINE] I could fill a book with threads and examples of<mask> anti-feminist and anti-women (those are not the same thing) the MRM is and<mask> pervasive those beliefs are and<mask> they come from the top down. [NEWLINE] [NEWLINE] I could do an entire *chapter* composed entirely of quotes of MRAs advocating disproportionate violence<mask> women are involved in an altercation. [NEWLINE] [NEWLINE] A chapter wouldn't be enough room. [NEWLINE] [NEWLINE] I've pointed out that the MRM clearly states its goal is to oppose feminism. It's the first link in their sidebar. I cannot link to it<mask> most subreddits auto-ban links to the website it's on.<mask>?<mask> it's a site famous for doxxing. You'll have to go look yourself, sorry. [NEWLINE] [NEWLINE] I've pointed out that the entirety of the celebrity level leadership of the MRM is outspokenly anti-feminist. [NEWLINE] [NEWLINE] (These last two points are more meta-relevant to this thread itself rather than the previous reply specifically, sorry. Just letting it all out in one place.) [NEWLINE] [NEWLINE] Ain't much more I can do here.</s>
Label encoding: <s>You can lead a horse to water, but you can't make it drink. [NEWLINE] [NEWLINE] I've pointed out that the subreddit is more than representative of the movement as a whole — it basically *is* the entire movement. [NEWLINE] [NEWLINE] I could fill a book with threads and examples of how anti-feminist and anti-women (those are not the same thing) the MRM is and how pervasive those beliefs are and how they come from the top down. [NEWLINE] [NEWLINE] I could do an entire *chapter* composed entirely of quotes of MRAs advocating disproportionate violence when women are involved in an altercation. [NEWLINE] [NEWLINE] A chapter wouldn't be enough room. [NEWLINE] [NEWLINE] I've pointed out that the MRM clearly states its goal is to oppose feminism. It's the first link in their sidebar. I cannot link to it because most subreddits auto-ban links to the website it's on. Why? Because it's a site famous for doxxing. You'll have to go look yourself, sorry. [NEWLINE] [NEWLINE] I've pointed out that the entirety of the celebrity level leadership of the MRM is outspokenly anti-feminist. [NEWLINE] [NEWLINE] (These last two points are more meta-relevant to this thread itself rather than the previous reply specifically, sorry. Just letting it all out in one place.) [NEWLINE] [NEWLINE] Ain't much more I can do here.</s>
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So correct me<mask> I'm wrong, you are just wondering<mask> it would be morally wrong or right for parents/caretakers to be able to choose to euthanize their child? Yea including the child complicates things<mask> it is a very complicated question. On the one hand, parents should have the right to do whatever they want with their own child<mask> they see fit,<mask><mask> is the line crossed?<mask> a parent should have the right to euthanize their child<mask> can't they beat them? or kill them?<mask> you are only taking the parents views in consideration,<mask> not just let them do<mask> ever they want to their child? The reason they can't do those things always comes down to<mask> it's not in the best interest of the child. Like it or not, that's<mask> this question will come down to, there isn't really any other way to talk about it without including the child. You even stated in your original post that the child and the parents might be better off, you included the child.<mask> I don't see<mask> it is largely separate from your question. The only reason most parents would choose to euthanize their child would be for the child's own good,<mask> it is<mask><mask> a core part of the argument. Sorry<mask> I am not helping you answer your question at all.</s>
Label encoding: <s>So correct me if I'm wrong, you are just wondering if it would be morally wrong or right for parents/caretakers to be able to choose to euthanize their child? Yea including the child complicates things but it is a very complicated question. On the one hand, parents should have the right to do whatever they want with their own child as they see fit, but when is the line crossed? If a parent should have the right to euthanize their child why can't they beat them? or kill them? If you are only taking the parents views in consideration, why not just let them do what ever they want to their child? The reason they can't do those things always comes down to because it's not in the best interest of the child. Like it or not, that's what this question will come down to, there isn't really any other way to talk about it without including the child. You even stated in your original post that the child and the parents might be better off, you included the child. So I don't see how it is largely separate from your question. The only reason most parents would choose to euthanize their child would be for the child's own good, so it is IMO a core part of the argument. Sorry if I am not helping you answer your question at all.</s>
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Karate Instructor here (<mask> that means anything). [NEWLINE] [NEWLINE] You really need to understand that<mask> an instructor explains a technique, he is trying to simplify or dumb it down<mask> that the average person he's explaining it to will understand it. We can't simply open our brains and give you the knowledge,<mask> we have to give it to you in pieces. We try to show you one specific defense for a very specific attack,<mask> in real life, there are dozens of possible variables that will lead to one of several possible options.<mask>, we can't just teach you by saying: [NEWLINE] [NEWLINE] "Ok,<mask><mask> he grabs you like this, you do this,<mask><mask> he grabs like this you do this,<mask> then<mask> his other foot is out you do this, unless YOUR foot is on the other side and then you do this. Of course,<mask> he's coming in with the other hand you're going to have to do this.....etc" [NEWLINE] [NEWLINE] <mask>, we take it one step at a time and the plan is that over time you gather the experience and reflexes to respond properly to any attack. This process takes a very long time, and the first step is doing "canned" techniques.<mask> you do each of the hundreds of canned techniques 1000 times, you have a good chance of reacting appropriately<mask> the time comes.</s>
Label encoding: <s>Karate Instructor here ( if that means anything). [NEWLINE] [NEWLINE] You really need to understand that when an instructor explains a technique, he is trying to simplify or dumb it down so that the average person he's explaining it to will understand it. We can't simply open our brains and give you the knowledge, so we have to give it to you in pieces. We try to show you one specific defense for a very specific attack, because in real life, there are dozens of possible variables that will lead to one of several possible options. However, we can't just teach you by saying: [NEWLINE] [NEWLINE] "Ok, so if he grabs you like this, you do this, but if he grabs like this you do this, but then if his other foot is out you do this, unless YOUR foot is on the other side and then you do this. Of course, if he's coming in with the other hand you're going to have to do this.....etc" [NEWLINE] [NEWLINE] So, we take it one step at a time and the plan is that over time you gather the experience and reflexes to respond properly to any attack. This process takes a very long time, and the first step is doing "canned" techniques. If you do each of the hundreds of canned techniques 1000 times, you have a good chance of reacting appropriately when the time comes.</s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've been living in rented accommodation<mask> I moved out of home at eighteen. Housing is super expensive<mask> I live (London) and I don't imagine I'll ever be able to afford to buy my own place. This doesn't bother me.<mask> a renter I have fewer rights<mask><mask> fewer responsibilities.<mask> the pipes burst I don't have to pay a plumber to fix them.<mask> the house burns down I'm not ruined. There is much less paperwork and legal obligations. I'm<mask> not tied to one place - I can easily move for work, to be closer to family, or just<mask> I'm bored and want to try somewhere new. [NEWLINE] [NEWLINE] The two great drawbacks are roommates (who can be very hit-and-miss) and the fact that at the end of a mortgage you have an asset to your name,<mask> with renting it's money gone forever.<mask> for the majority of my life it won't make a difference either way and at least this way I'm not tied into hundreds of thousands of pounds of debt. It's not<mask><mask> I can take a house with me<mask> I die<mask> it'll hardly matter in the end. [NEWLINE] [NEWLINE] I thought I might get a good debate on this<mask> reddit skews American and I believe homeowning is a bigger deal in the US. CMV!</s>
Label encoding: <s>I've been living in rented accommodation since I moved out of home at eighteen. Housing is super expensive where I live (London) and I don't imagine I'll ever be able to afford to buy my own place. This doesn't bother me. As a renter I have fewer rights but also fewer responsibilities. If the pipes burst I don't have to pay a plumber to fix them. If the house burns down I'm not ruined. There is much less paperwork and legal obligations. I'm also not tied to one place - I can easily move for work, to be closer to family, or just because I'm bored and want to try somewhere new. [NEWLINE] [NEWLINE] The two great drawbacks are roommates (who can be very hit-and-miss) and the fact that at the end of a mortgage you have an asset to your name, while with renting it's money gone forever. But for the majority of my life it won't make a difference either way and at least this way I'm not tied into hundreds of thousands of pounds of debt. It's not as though I can take a house with me when I die so it'll hardly matter in the end. [NEWLINE] [NEWLINE] I thought I might get a good debate on this since reddit skews American and I believe homeowning is a bigger deal in the US. CMV!</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From<mask> I've read in this post, this is a build-your-own society hypothetical not relevant to the current state<mask> the U.S. wouldn't be open to deconstructing existing legal infrastructure. [NEWLINE] [NEWLINE] <mask><mask> your point about making the legal system comprehensible to the every-person is an interesting concept,<mask> it's Utopian. I'd like to<mask><mask> this isn't a possibility<mask> of the nature of humanity. [NEWLINE] [NEWLINE] We'd first have to look at<mask> the nation was founded and who was in power to create the basic infrastructure that would eventually build to a complex infrastructure. We'd have to see their influences and their motivations in creating this new state.<mask><mask> corruption will always play a huge part in the creation of nations<mask> humans are more apt to look after themselves first in a spirit of survival and pleasure seeking, and then look at community interests<mask> an afterthought. [NEWLINE] [NEWLINE] <mask>,<mask> creating a legal system is such a complex and long-term project, we'd have to look at<mask> it would be possible to build a comprehensible and comprehensive legal system without the idiosyncrasies and terrible writing it contains. [NEWLINE] [NEWLINE] I don't think your argument is plausible<mask><mask><mask> there's a very small chance that the individuals building the system will be capable of meeting the qualifications you outline. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>From what I've read in this post, this is a build-your-own society hypothetical not relevant to the current state because the U.S. wouldn't be open to deconstructing existing legal infrastructure. [NEWLINE] [NEWLINE] I think your point about making the legal system comprehensible to the every-person is an interesting concept, but it's Utopian. I'd like to argue that this isn't a possibility because of the nature of humanity. [NEWLINE] [NEWLINE] We'd first have to look at how the nation was founded and who was in power to create the basic infrastructure that would eventually build to a complex infrastructure. We'd have to see their influences and their motivations in creating this new state. I think corruption will always play a huge part in the creation of nations because humans are more apt to look after themselves first in a spirit of survival and pleasure seeking, and then look at community interests as an afterthought. [NEWLINE] [NEWLINE] Also, because creating a legal system is such a complex and long-term project, we'd have to look at how it would be possible to build a comprehensible and comprehensive legal system without the idiosyncrasies and terrible writing it contains. [NEWLINE] [NEWLINE] I don't think your argument is plausible because I think there's a very small chance that the individuals building the system will be capable of meeting the qualifications you outline. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0697, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0865, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I suppose my problem with this answer is that without believing in god, the argument that one can create meaning through creation is invalid. [ENDQ] [NEWLINE] Are you perhaps trying to find some extra gravity in the concepts of meaning and fulfillment themselves? Sometimes it can be like growing up in a community obsessed with lapel pins, and people who don't have lapel pins, or who chose their pins poorly, are wracked with guilt and shame. A foreigner seeing someone stressed and depressed over the lack of a lapel pin would think he was bonkers. [NEWLINE] [NEWLINE] The comparison to God is to point out that the poor fellah must have exactly the same problem<mask> you,<mask> by being God he doesn't have anyone else to believe in.<mask><mask> can we somehow get meaning by having a middle-man?<mask> you can see<mask> the concept doesn't make sense, perhaps it can unchain your sense of<mask> meaning is supposed to be. There's no bottle of MeaningStuff^TM that any god can add to your life. [NEWLINE] [NEWLINE] One other way of looking at it is whether it matters or not that the universe has no fixed point of reference. It means there's no such thing<mask> the absolute speed of a star or planet,<mask> it doesn't mean they aren't moving, they're just moving relative to something else.</s>
Label encoding: <s> [STARTQ] I suppose my problem with this answer is that without believing in god, the argument that one can create meaning through creation is invalid. [ENDQ] [NEWLINE] Are you perhaps trying to find some extra gravity in the concepts of meaning and fulfillment themselves? Sometimes it can be like growing up in a community obsessed with lapel pins, and people who don't have lapel pins, or who chose their pins poorly, are wracked with guilt and shame. A foreigner seeing someone stressed and depressed over the lack of a lapel pin would think he was bonkers. [NEWLINE] [NEWLINE] The comparison to God is to point out that the poor fellah must have exactly the same problem as you, because by being God he doesn't have anyone else to believe in. So how can we somehow get meaning by having a middle-man? If you can see how the concept doesn't make sense, perhaps it can unchain your sense of what meaning is supposed to be. There's no bottle of MeaningStuff^TM that any god can add to your life. [NEWLINE] [NEWLINE] One other way of looking at it is whether it matters or not that the universe has no fixed point of reference. It means there's no such thing as the absolute speed of a star or planet, but it doesn't mean they aren't moving, they're just moving relative to something else.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In THIS language right NOW the word faggot has been stripped of all meaning other than a derogatory insult. [ENDQ] [NEWLINE] You've solved your own issue with this sentence; It has been stripped of all meaning - (usually) including the association with homosexuality.<mask> I call someone a "fucker", I am using the word in a derogatory manner, which is goes against the meaning of the word (ie. someone who fucks -<mask> can that possibly be an insult to say "that guy has a lot of sex"?). [NEWLINE] [NEWLINE] It's<mask> the word has shock value... it has stigma attached to it. [NEWLINE] [NEWLINE] It's<mask> rappers use "nigger"<mask> much. Obviously<mask> they're black, using "nigger"<mask> a pejorative is only insulting themselves,<mask><mask> do they do it?<mask> they've disassociated the word from its literal meaning, and are using it<mask> of the stigma attached. Stigma gets through; stigma punctuates the expletive.<mask> I say "that darned Johnson kid", it lacks the punch of saying "that fucking little shit Johnson kid". [NEWLINE] [NEWLINE] Faggot is being used the same way. Shock value and stigma - nothing to do with actual homosexuality anymore (at least, in the cases<mask> it's not being used<mask> a hate term). [NEWLINE] </s>
Label encoding: <s> [STARTQ] In THIS language right NOW the word faggot has been stripped of all meaning other than a derogatory insult. [ENDQ] [NEWLINE] You've solved your own issue with this sentence; It has been stripped of all meaning - (usually) including the association with homosexuality. If I call someone a "fucker", I am using the word in a derogatory manner, which is goes against the meaning of the word (ie. someone who fucks - how can that possibly be an insult to say "that guy has a lot of sex"?). [NEWLINE] [NEWLINE] It's because the word has shock value... it has stigma attached to it. [NEWLINE] [NEWLINE] It's why rappers use "nigger" so much. Obviously if they're black, using "nigger" as a pejorative is only insulting themselves, so why do they do it? Because they've disassociated the word from its literal meaning, and are using it because of the stigma attached. Stigma gets through; stigma punctuates the expletive. If I say "that darned Johnson kid", it lacks the punch of saying "that fucking little shit Johnson kid". [NEWLINE] [NEWLINE] Faggot is being used the same way. Shock value and stigma - nothing to do with actual homosexuality anymore (at least, in the cases where it's not being used as a hate term). [NEWLINE] </s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all,<mask> would you ever argue with a cop in the first place?<mask> authority over criminal law, or public policy, do you think some patrol cop has? Arguing with a cop is like arguing with the guy at the drive thru<mask> you don't like the way McDonald's is altering their existing franchise agreements. Even<mask> you're right, even it you "win" the argument, it's irrelevant<mask> the person you're arguing with has no authority to effect change. [NEWLINE] [NEWLINE] Second, I'll say again, cops have no real authority. Their job is to take suspected criminals into custody, and present them to the people who *do* have the authority: Prosecutors &amp; Judges.<mask> a cop has reasonable suspicion that you've committed a crime, you're going to jail and that's all there is to it; the only question is are you going to go the easy way, or the hard way.<mask> you're innocent, or<mask> the cop gets out of line, then you have an entire legal system to avail yourself of,<mask> arguing with some cop on the side of the road, who not only doesn't care<mask> you're guilty of innocent,<mask> who doesn't have the authority to do anything about it even<mask> they did, is only going to make a bad situation worse.<mask> bother?</s><pad>
Label encoding: <s>First of all, why would you ever argue with a cop in the first place? What authority over criminal law, or public policy, do you think some patrol cop has? Arguing with a cop is like arguing with the guy at the drive thru because you don't like the way McDonald's is altering their existing franchise agreements. Even if you're right, even it you "win" the argument, it's irrelevant because the person you're arguing with has no authority to effect change. [NEWLINE] [NEWLINE] Second, I'll say again, cops have no real authority. Their job is to take suspected criminals into custody, and present them to the people who *do* have the authority: Prosecutors &amp; Judges. If a cop has reasonable suspicion that you've committed a crime, you're going to jail and that's all there is to it; the only question is are you going to go the easy way, or the hard way. If you're innocent, or if the cop gets out of line, then you have an entire legal system to avail yourself of, but arguing with some cop on the side of the road, who not only doesn't care if you're guilty of innocent, but who doesn't have the authority to do anything about it even if they did, is only going to make a bad situation worse. Why bother?</s><pad>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1) This is something I have a problem with, the very notion of Monarchy is that it must be traditional,<mask><mask> this is true then<mask> can it represent the people? Monarchy may work in those places,<mask> in the U.S. and many other nations it would create tremendous rifts between those who felt part of that culture and those who do not. [NEWLINE] [NEWLINE] I never claimed that monarchy is right for EVERY country.<mask> your OP was not limited to USA. I was under the impression that, quote, "U.S. is an especially good **example**." [NEWLINE] [NEWLINE] For countries with traditional monarchy - monarchy may be best. For countries with anti-monarchy history and traditions... not<mask> much. The argument still works against your broad assertion that democracy is the best with no qualifications. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> a good example of this is the way American celebrities sometimes act<mask> statesmen, like<mask> Steven Seagal called for that talk between the U.S. and Russia [ENDQ] [NEWLINE] <mask><mask> you would agree that a visit from the Queen of England will carry a lot more weight and cachet than a visit from Steven Seagal. [NEWLINE] [NEWLINE] The best analogue that USA can muster  is sending a respected retied president.<mask> i would still say that reigning queen can be more effective.</s>
Label encoding: <s>1) This is something I have a problem with, the very notion of Monarchy is that it must be traditional, but if this is true then how can it represent the people? Monarchy may work in those places, but in the U.S. and many other nations it would create tremendous rifts between those who felt part of that culture and those who do not. [NEWLINE] [NEWLINE] I never claimed that monarchy is right for EVERY country. However your OP was not limited to USA. I was under the impression that, quote, "U.S. is an especially good **example**." [NEWLINE] [NEWLINE] For countries with traditional monarchy - monarchy may be best. For countries with anti-monarchy history and traditions... not so much. The argument still works against your broad assertion that democracy is the best with no qualifications. [NEWLINE] [NEWLINE] [STARTQ] I think a good example of this is the way American celebrities sometimes act as statesmen, like when Steven Seagal called for that talk between the U.S. and Russia [ENDQ] [NEWLINE] I think you would agree that a visit from the Queen of England will carry a lot more weight and cachet than a visit from Steven Seagal. [NEWLINE] [NEWLINE] The best analogue that USA can muster  is sending a respected retied president. But i would still say that reigning queen can be more effective.</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They demonstrate an aversion to stimuli that cause them to suffer.  Pleasurable and aversive states do not require much thought, and there is sound reasoning to suppose that pleasurable and aversive stimuli are common to almost all animals. [NEWLINE] [NEWLINE] <mask>, an explicit desire to continue living is much more advanced that mere pain avoidance and pleasure seeking,<mask> it is contingent on numerous cogntive tasks that currently seem beyond the abilities of most animals.  It requires a sense of self persisting over time, the understanding of temporality (past, present future experiences) etc. [NEWLINE] [NEWLINE] We can easily state that an animal prefers not to be shocked or burned or bled by observing animal behaviour.  It is easy to see that pain and pleasure are simple enough to be experienced my the overwhelming majority of animals. <mask>, it is not<mask> clear that they have reasons beyond that, or abstract reasons for behaviour, like preserving oneself.  It may be (and likely is) that they avoid threats merely<mask> their instincts scream that they do, and that avoiding such threats is a way of escaping self-imposed suffering (fear and distress) that instinct provides.  It is not clear that they think "that will kill me, and I wish to continue living",<mask> instead "shit, run!!!!" or something to that effect.</s>
Label encoding: <s>They demonstrate an aversion to stimuli that cause them to suffer.  Pleasurable and aversive states do not require much thought, and there is sound reasoning to suppose that pleasurable and aversive stimuli are common to almost all animals. [NEWLINE] [NEWLINE] However, an explicit desire to continue living is much more advanced that mere pain avoidance and pleasure seeking, as it is contingent on numerous cogntive tasks that currently seem beyond the abilities of most animals.  It requires a sense of self persisting over time, the understanding of temporality (past, present future experiences) etc. [NEWLINE] [NEWLINE] We can easily state that an animal prefers not to be shocked or burned or bled by observing animal behaviour.  It is easy to see that pain and pleasure are simple enough to be experienced my the overwhelming majority of animals.  However, it is not so clear that they have reasons beyond that, or abstract reasons for behaviour, like preserving oneself.  It may be (and likely is) that they avoid threats merely because their instincts scream that they do, and that avoiding such threats is a way of escaping self-imposed suffering (fear and distress) that instinct provides.  It is not clear that they think "that will kill me, and I wish to continue living", but instead "shit, run!!!!" or something to that effect.</s>
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] He knows about it and could stop it<mask> chooses not to. [ENDQ] [NEWLINE] [NEWLINE] [STARTQ] He knows about it<mask> is unable to stop it. [ENDQ] [NEWLINE] [NEWLINE] [STARTQ] He doesn't know about it. [ENDQ] [NEWLINE] He knows about it<mask> is unable to stop it<mask> God have previously granted humankind free will. To interfere beyond<mask> God have done<mask> already is to diminish the gift of free will. [NEWLINE] [NEWLINE] An all powerful god can't exactly go back on his previous promise/gift that easily simply by the virtue of God being all powerful. [NEWLINE] [NEWLINE] To simply conjure up a healthy candy for the daughter's wish is only to satisfy a desire rather than to really satisfy the daughter's choice. [NEWLINE] [NEWLINE] Omnipotent does not mean knowing that everything will happen this way. It can mean to know of ALL possibilities entirely, including those dictated by man's free will. [NEWLINE] [NEWLINE] An audience at a basketball free throw will know of multiple possible outcomes, that the ball will go nothing<mask> net, that the ball can slip out of the hand, that the ball will bounce off the rim, etc. The audience already knows these possibilities without dictating the basketball shooter's action<mask> pre-destiny. [NEWLINE] [NEWLINE] (hope my explanation can help clarify some. I'm a little sleep deprived<mask> it may not be completely coherent.)</s>
Label encoding: <s> [STARTQ] He knows about it and could stop it but chooses not to. [ENDQ] [NEWLINE] [NEWLINE] [STARTQ] He knows about it but is unable to stop it. [ENDQ] [NEWLINE] [NEWLINE] [STARTQ] He doesn't know about it. [ENDQ] [NEWLINE] He knows about it but is unable to stop it because God have previously granted humankind free will. To interfere beyond what God have done so already is to diminish the gift of free will. [NEWLINE] [NEWLINE] An all powerful god can't exactly go back on his previous promise/gift that easily simply by the virtue of God being all powerful. [NEWLINE] [NEWLINE] To simply conjure up a healthy candy for the daughter's wish is only to satisfy a desire rather than to really satisfy the daughter's choice. [NEWLINE] [NEWLINE] Omnipotent does not mean knowing that everything will happen this way. It can mean to know of ALL possibilities entirely, including those dictated by man's free will. [NEWLINE] [NEWLINE] An audience at a basketball free throw will know of multiple possible outcomes, that the ball will go nothing but net, that the ball can slip out of the hand, that the ball will bounce off the rim, etc. The audience already knows these possibilities without dictating the basketball shooter's action as pre-destiny. [NEWLINE] [NEWLINE] (hope my explanation can help clarify some. I'm a little sleep deprived so it may not be completely coherent.)</s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0719, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0653, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Anyone who argues that abortion is a necessary safeguard against a life of suffering and disability is<mask><mask> the unborn child is not<mask> a living human being.<mask> this is exactly the point that they must prove before they can even begin to make such claims. Disability isn't the issue, it's humanity. We do not kill people for their disabilities, period.<mask>, unless we're not human beings before we're born, our disabilities should no more disqualify us from life before birth than they do after birth. [ENDQ] [NEWLINE] [STARTQ] Furthermore, this pressure to abort handicapped babies is built largely on conjecture, on the mere "likelihood" that a child has some kind of disability. Often, the tests prove wrong, and more often still, these children,<mask> allowed to live, end up with lives of joy and happiness that far exceeds those of their "more healthy" peers. Suffering and hardship are not bad things. They are means to a greater end, a crucial part of the human journey. Anyone who tries to eliminate suffering by killing the "sufferers" is establishing a horrific trend. It is not for us to decide who has a life worth living and who doesn't, and we certainly wouldn't want someone else making that decision for us! [ENDQ] [NEWLINE] [The Case Against Abortion]( [URL] /) [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Anyone who argues that abortion is a necessary safeguard against a life of suffering and disability is assuming that the unborn child is not yet a living human being. But this is exactly the point that they must prove before they can even begin to make such claims. Disability isn't the issue, it's humanity. We do not kill people for their disabilities, period. Therefore, unless we're not human beings before we're born, our disabilities should no more disqualify us from life before birth than they do after birth. [ENDQ] [NEWLINE] [STARTQ] Furthermore, this pressure to abort handicapped babies is built largely on conjecture, on the mere "likelihood" that a child has some kind of disability. Often, the tests prove wrong, and more often still, these children, if allowed to live, end up with lives of joy and happiness that far exceeds those of their "more healthy" peers. Suffering and hardship are not bad things. They are means to a greater end, a crucial part of the human journey. Anyone who tries to eliminate suffering by killing the "sufferers" is establishing a horrific trend. It is not for us to decide who has a life worth living and who doesn't, and we certainly wouldn't want someone else making that decision for us! [ENDQ] [NEWLINE] [The Case Against Abortion]( [URL] /) [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If EVERYONE lived in an HOA, it would negate the benefits of an HOA<mask> you'd have neighbors who didn't care about their property (and many of them wouldn't care about the tickets or fines they received,<mask> it would continue until their house was foreclosed on, which would affect everyone in the HOA). HOAs work<mask> the only people agreeing to them are those who commit to the idea of caring about<mask> their property looks. [NEWLINE] [NEWLINE] I like my neighborhood, which has an HOA. Everyone's property looks great (especially compared to nearby neighborhoods not in an HOA).<mask> part of the benefit of the HOA is that it discourages the people I don't want<mask> neighbors from moving in. Not that I'm saying anyone who doesn't want to live in an HOA doesn't care about their property,<mask> there are certainly a ton of people who don't care, are too lazy, or have different opinions of<mask> is acceptable/good looking. Prior to moving to my current location, I had more than my fair share of slobs who didn't care--they let their lawns die, junked cars everywhere, they sometimes had unlawful tenants, etc. I don't want them<mask> my neighbors, and thankfully the HOA has kept them out of the neighborhood. [NEWLINE] </s>
Label encoding: <s>If EVERYONE lived in an HOA, it would negate the benefits of an HOA because you'd have neighbors who didn't care about their property (and many of them wouldn't care about the tickets or fines they received, so it would continue until their house was foreclosed on, which would affect everyone in the HOA). HOAs work because the only people agreeing to them are those who commit to the idea of caring about how their property looks. [NEWLINE] [NEWLINE] I like my neighborhood, which has an HOA. Everyone's property looks great (especially compared to nearby neighborhoods not in an HOA). But part of the benefit of the HOA is that it discourages the people I don't want as neighbors from moving in. Not that I'm saying anyone who doesn't want to live in an HOA doesn't care about their property, but there are certainly a ton of people who don't care, are too lazy, or have different opinions of what is acceptable/good looking. Prior to moving to my current location, I had more than my fair share of slobs who didn't care--they let their lawns die, junked cars everywhere, they sometimes had unlawful tenants, etc. I don't want them as my neighbors, and thankfully the HOA has kept them out of the neighborhood. [NEWLINE] </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> does "community's expectation" mean? [ENDQ] [NEWLINE] In this case our community has made it clear that strong pornography is not something we should stock. The community uses the library board and mayor to deliver this message<mask> we<mask> read the newspapers and libraries nearby have had to fight battles over this issue. We do stock books of a sexual nature and movies containing nudity<mask> we don't stock anything that is clearly identified<mask> pornography, even<mask> it's by the same guy who wrote Watchmen. [NEWLINE] [NEWLINE] We do not restrict children from browsing and checking out books from the adult collection and there's probably concern we might illegally convey material restricted to adults to children,<mask> I'm speculating there. [NEWLINE] [NEWLINE] [STARTQ] <mask> I would say is that I support entirely content-neutral ways of acquiring content for the library. [ENDQ] [NEWLINE] <mask> does that mean<mask> not buying books purely at random? Any other system is going to have to take into account the content of the books. I mean, I'm at a loss for<mask> you would do this. [NEWLINE] [NEWLINE] Remember a library's job is to be a curator. There are nearly an infinite  number of books available to choose from and our patrons expect us to use our expertise to whittle down that to a more reasonable selection and that means we have to judge the content of the books.</s>
Label encoding: <s> [STARTQ] What does "community's expectation" mean? [ENDQ] [NEWLINE] In this case our community has made it clear that strong pornography is not something we should stock. The community uses the library board and mayor to deliver this message but we also read the newspapers and libraries nearby have had to fight battles over this issue. We do stock books of a sexual nature and movies containing nudity but we don't stock anything that is clearly identified as pornography, even if it's by the same guy who wrote Watchmen. [NEWLINE] [NEWLINE] We do not restrict children from browsing and checking out books from the adult collection and there's probably concern we might illegally convey material restricted to adults to children, but I'm speculating there. [NEWLINE] [NEWLINE] [STARTQ] What I would say is that I support entirely content-neutral ways of acquiring content for the library. [ENDQ] [NEWLINE] What does that mean if not buying books purely at random? Any other system is going to have to take into account the content of the books. I mean, I'm at a loss for how you would do this. [NEWLINE] [NEWLINE] Remember a library's job is to be a curator. There are nearly an infinite  number of books available to choose from and our patrons expect us to use our expertise to whittle down that to a more reasonable selection and that means we have to judge the content of the books.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Two reasons<mask> grammar nazis are *saving* language and promoting globalization: [NEWLINE] [NEWLINE] 1. Consistency. English has been<mask> it is for a very long time.<mask>, there are a lot of resources that are still correct- it doesn't matter<mask> it's from 1930 or 2005, it's probably at least 90% correct. Changing the language causes problems for people in remote places who may only have old textbooks and learning supplies-<mask> they use a book from 1980, right now that's not a problem,<mask><mask> we gradually change the language,<mask> long until it's outdated and useless? [NEWLINE] [NEWLINE] 2. The goal of language is communication. Change is not a consistent, constant, universal thing- promoting or allowing change on a large scale will, over time, result in a variety of mutually unintelligible new pseudo-English languages that will limit people's communication ability to their region. Keeping language consistent, via people like grammar nazis, ensures that the English spoken in place A is at least 90% the same<mask> the English spoken in place B. [NEWLINE] [NEWLINE] <mask>, a quick example of<mask> happens<mask> you *don't* have a limiting factor keeping English spelling/pronunciation consistent and mutually intelligible: [NEWLINE] [NEWLINE] [Listen to it without looking at the subtitles]( [URL] ) [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Two reasons why grammar nazis are *saving* language and promoting globalization: [NEWLINE] [NEWLINE] 1. Consistency. English has been as it is for a very long time. Therefore, there are a lot of resources that are still correct- it doesn't matter if it's from 1930 or 2005, it's probably at least 90% correct. Changing the language causes problems for people in remote places who may only have old textbooks and learning supplies- if they use a book from 1980, right now that's not a problem, but if we gradually change the language, how long until it's outdated and useless? [NEWLINE] [NEWLINE] 2. The goal of language is communication. Change is not a consistent, constant, universal thing- promoting or allowing change on a large scale will, over time, result in a variety of mutually unintelligible new pseudo-English languages that will limit people's communication ability to their region. Keeping language consistent, via people like grammar nazis, ensures that the English spoken in place A is at least 90% the same as the English spoken in place B. [NEWLINE] [NEWLINE] Lastly, a quick example of what happens when you *don't* have a limiting factor keeping English spelling/pronunciation consistent and mutually intelligible: [NEWLINE] [NEWLINE] [Listen to it without looking at the subtitles]( [URL] ) [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>,<mask> you get wrong is idea that life was bad before capitalism. [ENDQ] [NEWLINE] This is laughable. Growing up in the middle ages you would have been at the mercy of feudal lords who wouldve demanded a huge amount in taxes, you would be constantly manipulated and terrorized by the church. You would almost certainly be a farmer like the vast majority of the population. 99% of<mask> you own is from within 10 miles of your home. You would've almost certainly have grown up being beaten abused by not only your parents,<mask> by other adults. Growing up you would have<mask> witnessed constant cruelty to animals,<mask> well<mask> constant cruelty to other humans.<mask> you are 6 years old you watch a man being hanged for stealing from a local lord. [NEWLINE] [NEWLINE] Arthritis, parasites, migraines, and dysentery would have been a constant problem<mask> you spent almost every day with some nagging physical issue that you have no idea about. Your teeth would have been a constant worry<mask> your high in grain diet caused cavities that fester until you finally are in such pain you pull the tooth entirely. [NEWLINE] [NEWLINE] And thats the life you have<mask> there are no wars, and no epidemic diseases. Once those start happening your quality of life drops even further. The middle ages was not the shire. </s>
Label encoding: <s> [STARTQ] However, where you get wrong is idea that life was bad before capitalism. [ENDQ] [NEWLINE] This is laughable. Growing up in the middle ages you would have been at the mercy of feudal lords who wouldve demanded a huge amount in taxes, you would be constantly manipulated and terrorized by the church. You would almost certainly be a farmer like the vast majority of the population. 99% of what you own is from within 10 miles of your home. You would've almost certainly have grown up being beaten abused by not only your parents, but by other adults. Growing up you would have also witnessed constant cruelty to animals, as well as constant cruelty to other humans. When you are 6 years old you watch a man being hanged for stealing from a local lord. [NEWLINE] [NEWLINE] Arthritis, parasites, migraines, and dysentery would have been a constant problem as you spent almost every day with some nagging physical issue that you have no idea about. Your teeth would have been a constant worry as your high in grain diet caused cavities that fester until you finally are in such pain you pull the tooth entirely. [NEWLINE] [NEWLINE] And thats the life you have if there are no wars, and no epidemic diseases. Once those start happening your quality of life drops even further. The middle ages was not the shire. </s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0682, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That works<mask> Lasik is not a procedure that's non-essential: You can choose to wear glasses without it seriously impeding your life.<mask> such private Lasik surgeries don't have their customer over a barrel,<mask> it's beneficial for surgeries to find ways to reduce costs and improve their procedures and results in order to attract more people to consider Lasik worth the cost to not have to wear glasses again (or at least until your eyesight starts failing with age). [NEWLINE] [NEWLINE] <mask> it comes to life-saving medications and procedures<mask>, your choices are limited to pay up, or die/be severely afflicted. People don't really have the option to decline the treatment with minimal impact like they do with Lasik. You can't decide to carry on with a failed liver, or cancer, or blindness without it either killing you, or seriously impeding your life without medical intervention.<mask> such the pressure to reduce the cost of treatment is nowhere near<mask> great. [NEWLINE] [NEWLINE] One could<mask><mask> this is a great reason to make essential medical care a human right,<mask> human rights are generally put in place to protect life and prevent suffering. Having to wear glasses isn't going to cause a massive amount of suffering,<mask> having to decline an operation to save your eyesight<mask> you can't afford it most certainly is.</s>
Label encoding: <s>That works because Lasik is not a procedure that's non-essential: You can choose to wear glasses without it seriously impeding your life. As such private Lasik surgeries don't have their customer over a barrel, so it's beneficial for surgeries to find ways to reduce costs and improve their procedures and results in order to attract more people to consider Lasik worth the cost to not have to wear glasses again (or at least until your eyesight starts failing with age). [NEWLINE] [NEWLINE] When it comes to life-saving medications and procedures though, your choices are limited to pay up, or die/be severely afflicted. People don't really have the option to decline the treatment with minimal impact like they do with Lasik. You can't decide to carry on with a failed liver, or cancer, or blindness without it either killing you, or seriously impeding your life without medical intervention. As such the pressure to reduce the cost of treatment is nowhere near as great. [NEWLINE] [NEWLINE] One could argue that this is a great reason to make essential medical care a human right, since human rights are generally put in place to protect life and prevent suffering. Having to wear glasses isn't going to cause a massive amount of suffering, but having to decline an operation to save your eyesight because you can't afford it most certainly is.</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ask a computer security expert to tell you<mask> "secure" means. Now ask another one. And another. There is no universal definition of "secure" in computer security. All security is defined against some abstract security model that is specific to the issue at hand. The security model for network protocols is wildly different than the security model for web systems. There are dozens of different adversary models (a part of a security model) in cryptographic security. Is this a weakness of computer security<mask> a field? Or does it mean that "secure" is a very hard thing to define properly and there is a lot of controversy over<mask> it means? [NEWLINE] [NEWLINE] Feminism is a very fractured movement, even within just academia.<mask><mask> you actually read feminist papers and essays you'll find that the definition of "patriarchy" converges reasonably closely to<mask> I've posted. You won't find too many academics claiming that there is a patriarchy<mask><mask> the linguistic definition. [NEWLINE] [NEWLINE] Dictionaries are **not** accepted<mask> definitive in many circles. Dictionaries present general definitions<mask> they are understood by laypeople. They don't work for all contexts. In my field I'd be laughed out of the room<mask> I suddenly started using dictionary definitions of every term rather than the appropriate jargon specific to my field. </s>
Label encoding: <s>Ask a computer security expert to tell you what "secure" means. Now ask another one. And another. There is no universal definition of "secure" in computer security. All security is defined against some abstract security model that is specific to the issue at hand. The security model for network protocols is wildly different than the security model for web systems. There are dozens of different adversary models (a part of a security model) in cryptographic security. Is this a weakness of computer security as a field? Or does it mean that "secure" is a very hard thing to define properly and there is a lot of controversy over what it means? [NEWLINE] [NEWLINE] Feminism is a very fractured movement, even within just academia. But if you actually read feminist papers and essays you'll find that the definition of "patriarchy" converges reasonably closely to what I've posted. You won't find too many academics claiming that there is a patriarchy according to the linguistic definition. [NEWLINE] [NEWLINE] Dictionaries are **not** accepted as definitive in many circles. Dictionaries present general definitions as they are understood by laypeople. They don't work for all contexts. In my field I'd be laughed out of the room if I suddenly started using dictionary definitions of every term rather than the appropriate jargon specific to my field. </s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Morality is all about protecting and promoting life (<mask> people have many different definitions of<mask> that is).<mask> is life good? You are free to disagree,<mask> this will eventually lead to your death, and remove that perspective from the universe. [NEWLINE] [NEWLINE] Moral rules are not arbitrary,<mask> are based on helping us to continue to exist and to grow. A society that saw murdering each other<mask> moral would not long last, compared to a society that saw protecting each other<mask> moral. An individual who does not value their own life will be more likely to lose it than one who does,<mask> there's likely a selection process in terms of<mask> views of morality can exist for any prolonged period of time, in an individual or in a society. [NEWLINE] [NEWLINE] I'd<mask> say that<mask> morality really is all about promoting and protecting life, then it is<mask> (at least partially) hard-wired into us. (Almost) all living things act to protect and spread life, working to maintain their existence and to have descendants. I don't think you need to view morality<mask> a magical rule... instead<mask><mask> it makes sense to have the term morality to describe a portion of our understanding of reality, which reflects<mask> we do certain things, and perhaps can<mask> be used to guide our understanding<mask> we study it.</s>
Label encoding: <s>Morality is all about protecting and promoting life ( though people have many different definitions of what that is). Why is life good? You are free to disagree, but this will eventually lead to your death, and remove that perspective from the universe. [NEWLINE] [NEWLINE] Moral rules are not arbitrary, but are based on helping us to continue to exist and to grow. A society that saw murdering each other as moral would not long last, compared to a society that saw protecting each other as moral. An individual who does not value their own life will be more likely to lose it than one who does, so there's likely a selection process in terms of what views of morality can exist for any prolonged period of time, in an individual or in a society. [NEWLINE] [NEWLINE] I'd also say that if morality really is all about promoting and protecting life, then it is also (at least partially) hard-wired into us. (Almost) all living things act to protect and spread life, working to maintain their existence and to have descendants. I don't think you need to view morality as a magical rule... instead I think it makes sense to have the term morality to describe a portion of our understanding of reality, which reflects why we do certain things, and perhaps can also be used to guide our understanding if we study it.</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0901, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a business model that works for some kinds of software, sure,<mask> not even for all forms of software. [NEWLINE] [NEWLINE] Say you develop a game, it takes 3 years, ten employees, and about a million dollars to produce.<mask> do you make money from it? You sell it to consumers.<mask> there is no legal protection against piracy, and only one person pays for it,<mask> all of the other consumers get free copies from the first consumer, the developer is never going to see even a fraction of the million dollars they spent on making the game. The developer folds, and the employees no longer have jobs. In this system, there is no incentive for the developer to make any more games ever again. This occurs<mask><mask> the quality of the game. [NEWLINE] [NEWLINE] Compare this to the situation<mask> the developer's intellectual property is protected by law (<mask> it is today), and they have legal recourse against people who pirate the game without paying for it. A large proportion of consumers buy the game, and<mask> more buy it 6 months later<mask> the price decreases. The developer makes its money back along with some return on its investment, and it goes on to make more games in the future. [NEWLINE] [NEWLINE] Please explain to me<mask> the removal of copyright laws isn't going to produce the first scenario.</s>
Label encoding: <s>This is a business model that works for some kinds of software, sure, but not even for all forms of software. [NEWLINE] [NEWLINE] Say you develop a game, it takes 3 years, ten employees, and about a million dollars to produce. How do you make money from it? You sell it to consumers. If there is no legal protection against piracy, and only one person pays for it, while all of the other consumers get free copies from the first consumer, the developer is never going to see even a fraction of the million dollars they spent on making the game. The developer folds, and the employees no longer have jobs. In this system, there is no incentive for the developer to make any more games ever again. This occurs regardless of the quality of the game. [NEWLINE] [NEWLINE] Compare this to the situation where the developer's intellectual property is protected by law ( as it is today), and they have legal recourse against people who pirate the game without paying for it. A large proportion of consumers buy the game, and yet more buy it 6 months later when the price decreases. The developer makes its money back along with some return on its investment, and it goes on to make more games in the future. [NEWLINE] [NEWLINE] Please explain to me why the removal of copyright laws isn't going to produce the first scenario.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> respect is free (this is<mask> I was arguing against).<mask> not give it away free?<mask> should anyone not have respect<mask> it's such a free resource? [ENDQ] [NEWLINE] I'm going ahead and giving a delta for that &amp;#8710;<mask> it changed my perspective. I still believe being respectful is free,<mask> you make a valid point that achieving a state were you're owed respect isn't necessarily. [NEWLINE] [NEWLINE] [STARTQ] Disagree.<mask> it comes to me showing deference to someone, it generally means I wasn't going to make that decision.<mask> I just give my respect to say the oldest person in the room,<mask><mask> they are wrong? [ENDQ] [NEWLINE] <mask><mask> were not talking about the same thing. I mean deference "respect, esteem or consideration", not necessarily<mask> doing everything someone tells you. For instance, I'd say I owed to my parents' to consider their opinion on a matter seriously, something I wouldn't do for just anybody. [NEWLINE] [NEWLINE] Again, I'm not saying they can't ever be wrong. [NEWLINE] [NEWLINE] [STARTQ] <mask> someone is given deference just based on their experience or position, bad decisions are made. [ENDQ] [NEWLINE] I don't think this is<mask> clear-cut. Bad decisions are made due to inexperience all the time. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Because respect is free (this is what I was arguing against). Why not give it away free? Why should anyone not have respect if it's such a free resource? [ENDQ] [NEWLINE] I'm going ahead and giving a delta for that &amp;#8710; since it changed my perspective. I still believe being respectful is free, but you make a valid point that achieving a state were you're owed respect isn't necessarily. [NEWLINE] [NEWLINE] [STARTQ] Disagree. If it comes to me showing deference to someone, it generally means I wasn't going to make that decision. If I just give my respect to say the oldest person in the room, what if they are wrong? [ENDQ] [NEWLINE] I think were not talking about the same thing. I mean deference "respect, esteem or consideration", not necessarily as doing everything someone tells you. For instance, I'd say I owed to my parents' to consider their opinion on a matter seriously, something I wouldn't do for just anybody. [NEWLINE] [NEWLINE] Again, I'm not saying they can't ever be wrong. [NEWLINE] [NEWLINE] [STARTQ] When someone is given deference just based on their experience or position, bad decisions are made. [ENDQ] [NEWLINE] I don't think this is as clear-cut. Bad decisions are made due to inexperience all the time. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's nothing unethical about taking the money people are willing to pay you. [ENDQ] [STARTQ] &gt; "Your money or your life" isn't unethical either then. After all, you gave them a choice and they choose freely. [ENDQ] [NEWLINE] Eh...<mask><mask> that's kinda different there isn't it? [NEWLINE] [NEWLINE] The money someone is paid for there job is offered, without any form of coersion,<mask> there may be some negotiation [NEWLINE] [NEWLINE] Your money or your life is straight up intimidation - "Give it to me, or I'll kill you and probably take it anyway" or "<mask> you just give it to me now, I won't hurt you to get it" are pretty much saying the same thing [NEWLINE] [NEWLINE] Not that I don't agree that the little guys get squeezed in a company [NEWLINE] [NEWLINE] There are plenty of companies here in the UK,<mask> the day to day employees have had no pay rise or have been laid off in their droves,<mask> the guy at the other end of the spectrum takes his/her entire annual salary again<mask> a bonus - I'm sorry,<mask><mask> you're already earning £1,000,000 per year, you shouldn't be taking a further £1,000,000<mask> a bonus after you've laid off hundreds of workers. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] There's nothing unethical about taking the money people are willing to pay you. [ENDQ] [STARTQ] &gt; "Your money or your life" isn't unethical either then. After all, you gave them a choice and they choose freely. [ENDQ] [NEWLINE] Eh... I think that's kinda different there isn't it? [NEWLINE] [NEWLINE] The money someone is paid for there job is offered, without any form of coersion, although there may be some negotiation [NEWLINE] [NEWLINE] Your money or your life is straight up intimidation - "Give it to me, or I'll kill you and probably take it anyway" or " If you just give it to me now, I won't hurt you to get it" are pretty much saying the same thing [NEWLINE] [NEWLINE] Not that I don't agree that the little guys get squeezed in a company [NEWLINE] [NEWLINE] There are plenty of companies here in the UK, where the day to day employees have had no pay rise or have been laid off in their droves, while the guy at the other end of the spectrum takes his/her entire annual salary again as a bonus - I'm sorry, but if you're already earning £1,000,000 per year, you shouldn't be taking a further £1,000,000 as a bonus after you've laid off hundreds of workers. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0766, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll start by saying that<mask><mask> with OP that a trophy for participating is absurd. <mask>,<mask><mask> some form of acknowledgment for being in the league at all is cool.  I remember one  league I was in<mask> every player received a patch with the league logo and the year.  It's fun to go back years later and see a physical reminder of the activities of your youth, especially<mask> you're the type of kid who will never go on to play in a more competitive format.  I ended up playing college soccer,<mask> I still like seeing my patches from<mask> I was like 8 years old. [NEWLINE] [NEWLINE] The one saving grace,<mask><mask>, is that<mask><mask><mask> there's a big, badass trophy for the winners and everyone else gets small, shitty trophies, any kid with even a little competition in his DNA will be pissed about the small trophy and want the big one,<mask> I don't think it impacts the kids mentally<mask> much<mask> it impacts the parents who see the stupid symbolism. [NEWLINE] [NEWLINE] <mask><mask> that it sets the wrong example, and I'd rather more parents had to have the conversation started by their kid saying, "<mask><mask> do *they* get a trophy?"<mask> it's probably not<mask> bad<mask> people like you and<mask><mask> it is.</s>
Label encoding: <s>I'll start by saying that I agree with OP that a trophy for participating is absurd.  However, I think some form of acknowledgment for being in the league at all is cool.  I remember one  league I was in where every player received a patch with the league logo and the year.  It's fun to go back years later and see a physical reminder of the activities of your youth, especially if you're the type of kid who will never go on to play in a more competitive format.  I ended up playing college soccer, but I still like seeing my patches from when I was like 8 years old. [NEWLINE] [NEWLINE] The one saving grace, I think, is that as long as there's a big, badass trophy for the winners and everyone else gets small, shitty trophies, any kid with even a little competition in his DNA will be pissed about the small trophy and want the big one, so I don't think it impacts the kids mentally as much as it impacts the parents who see the stupid symbolism. [NEWLINE] [NEWLINE] I agree that it sets the wrong example, and I'd rather more parents had to have the conversation started by their kid saying, " But why do *they* get a trophy?" but it's probably not as bad as people like you and I think it is.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For a long time I've regarded most modern art<mask> "art for *artists'* sake". I've known many people who've studied art and now call themselves artists, quite a few who do this full time and exhibit. Usually they're lovely people. It may well be that I'm missing something,<mask><mask> I hear artists talk about their work it's usually the most circuitous, non-committal language I've ever heard.<mask><mask> someone may have done a study on the language used by artists, and<mask> not then linguists - get on it. The common denominator seems to be a desire to please the audience, the key part here being that the audience is *other artists*. [NEWLINE] [NEWLINE] Perhaps I've become too cynical, and there are of course exceptions to<mask> I'm saying,<mask><mask><mask> most artists have turned exploratory play into a nebulous form of pseudo-intellectual navel gazing and utterly tedious group conformity (ironic<mask><mask> art is'meant' to push boundaries). They're mainly trying to impress each other, and use references that only other 'in' people can grasp (<mask> that's often buried<mask> deep that few people do).<mask><mask> the pursuit of meaning in art is often simply 'emperor's new clothes'. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>For a long time I've regarded most modern art as "art for *artists'* sake". I've known many people who've studied art and now call themselves artists, quite a few who do this full time and exhibit. Usually they're lovely people. It may well be that I'm missing something, but when I hear artists talk about their work it's usually the most circuitous, non-committal language I've ever heard. I think someone may have done a study on the language used by artists, and if not then linguists - get on it. The common denominator seems to be a desire to please the audience, the key part here being that the audience is *other artists*. [NEWLINE] [NEWLINE] Perhaps I've become too cynical, and there are of course exceptions to what I'm saying, but IMO most artists have turned exploratory play into a nebulous form of pseudo-intellectual navel gazing and utterly tedious group conformity (ironic given that art is'meant' to push boundaries). They're mainly trying to impress each other, and use references that only other 'in' people can grasp ( but that's often buried so deep that few people do). IMO the pursuit of meaning in art is often simply 'emperor's new clothes'. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Even though it seems like individually there is "nothing" any one person can do,<mask> a critical mass of people started doing *something* about it, then policies can change, things can get better, and people's personal lives can get better. [NEWLINE] [NEWLINE] I live in a mountain valley area<mask> there are often weather inversions that cause the air quality and particles to get horrendously bad during the winter months. <mask> just 5% of drivers decide to stay home or carpool, another 5% of drivers decide to cut out unnecessary side trips and only commute to/from work, and 10% of homeowners with wood log burning fireplaces decide to not burn wood, that has a net positive impact on the air quality.  Every particle that doesn't get thrown into the air is better for the overall health of people living in the valley. [NEWLINE] [NEWLINE] This is the same argument people have against "voting."<mask> nobody voted, then yeah it's useless.  You're throwing away your own vote,<mask> there's no reason to claim that voting is completely pointless.   It's the aggregate effect of voting in general that affects outcomes.  In the case of environmental impact, it's even more direct and immediate than voting.  Every small thing that every person does has an impact.</s>
Label encoding: <s>Even though it seems like individually there is "nothing" any one person can do, if a critical mass of people started doing *something* about it, then policies can change, things can get better, and people's personal lives can get better. [NEWLINE] [NEWLINE] I live in a mountain valley area where there are often weather inversions that cause the air quality and particles to get horrendously bad during the winter months.  If just 5% of drivers decide to stay home or carpool, another 5% of drivers decide to cut out unnecessary side trips and only commute to/from work, and 10% of homeowners with wood log burning fireplaces decide to not burn wood, that has a net positive impact on the air quality.  Every particle that doesn't get thrown into the air is better for the overall health of people living in the valley. [NEWLINE] [NEWLINE] This is the same argument people have against "voting." If nobody voted, then yeah it's useless.  You're throwing away your own vote, but there's no reason to claim that voting is completely pointless.   It's the aggregate effect of voting in general that affects outcomes.  In the case of environmental impact, it's even more direct and immediate than voting.  Every small thing that every person does has an impact.</s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's a nice idea in theory,<mask> sometimes hat and vitriol is the only way to get something shamed away quick enough.  Frankly, it's surprising that violence is<mask> rarely on the table.  It wouldn't be an issue<mask> there was no time-table,<mask> laws and rules are constantly being made which have huge effects on LGBT peoples.  Waiting for tolerance would more likely see the good laws ignored and the bad laws passed. [NEWLINE] [NEWLINE] ENDA is a very good example right now, which passed the senate,<mask> very well may not even be brought up in the house.  All it does is provide employment protection to LGBTs,<mask> people say that it infringes on their rights to hire who they want for religious purposes (aka, legal discrimination). [NEWLINE] [NEWLINE] In the state I'm currently in (Florida), there are no state laws preventing me from being fired or evicted simply<mask> I'm gay.  It is unconstitutional for me to get married or have a civil union, and until 2010 even to adopt a child.  There is still no hate crime law based on gender identity,<mask> transgender individuals are screwed in that note.  Attempts to change this situation almost always has to go through the courts,<mask> it would not pass in the proper way.</s>
Label encoding: <s>It's a nice idea in theory, but sometimes hat and vitriol is the only way to get something shamed away quick enough.  Frankly, it's surprising that violence is so rarely on the table.  It wouldn't be an issue if there was no time-table, but laws and rules are constantly being made which have huge effects on LGBT peoples.  Waiting for tolerance would more likely see the good laws ignored and the bad laws passed. [NEWLINE] [NEWLINE] ENDA is a very good example right now, which passed the senate, but very well may not even be brought up in the house.  All it does is provide employment protection to LGBTs, but people say that it infringes on their rights to hire who they want for religious purposes (aka, legal discrimination). [NEWLINE] [NEWLINE] In the state I'm currently in (Florida), there are no state laws preventing me from being fired or evicted simply because I'm gay.  It is unconstitutional for me to get married or have a civil union, and until 2010 even to adopt a child.  There is still no hate crime law based on gender identity, so transgender individuals are screwed in that note.  Attempts to change this situation almost always has to go through the courts, as it would not pass in the proper way.</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0625, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The Vulcans were (will be?) a pretty civilized and advanced race,<mask><mask> you're on to something. [NEWLINE] [NEWLINE] Nerd jokes aside<mask> you're describing is a coping mechanism, many people will develop their own unique way of regaining or approximating physical, emotional, or psychological homeostasis after a traumatic event (which can include these "Coming to Jesus" moments) or timespan. In many cases it's perfectly normal and healthy. [NEWLINE] [NEWLINE] <mask> do<mask> works for you. I suspect that, with time, your natural human emotional responses will creep their way back into your life.<mask> this time around they'll be tempered by the control and discipline you're gaining now. Emotion is far from a flaw; it is a valuable asset to the human condition<mask> it can be afforded the discipline of maturity. Or perhaps the maturity of discipline. I dont mean this to be patronizing in any way,<mask> that's<mask> "maturing" is all about--and you're quite fortunate to be doing it at a relatively young age<mask> many never really do. [NEWLINE] [NEWLINE] Good luck. Talk to a friend or consider checking out a therapist<mask> it really gets too bad,<mask> it seems to me like you're handling things just fine. [NEWLINE] [NEWLINE] Edit: Rule 1 compliance</s>
Label encoding: <s>The Vulcans were (will be?) a pretty civilized and advanced race, I think you're on to something. [NEWLINE] [NEWLINE] Nerd jokes aside what you're describing is a coping mechanism, many people will develop their own unique way of regaining or approximating physical, emotional, or psychological homeostasis after a traumatic event (which can include these "Coming to Jesus" moments) or timespan. In many cases it's perfectly normal and healthy. [NEWLINE] [NEWLINE] So do what works for you. I suspect that, with time, your natural human emotional responses will creep their way back into your life. But this time around they'll be tempered by the control and discipline you're gaining now. Emotion is far from a flaw; it is a valuable asset to the human condition when it can be afforded the discipline of maturity. Or perhaps the maturity of discipline. I dont mean this to be patronizing in any way, but that's what "maturing" is all about--and you're quite fortunate to be doing it at a relatively young age when many never really do. [NEWLINE] [NEWLINE] Good luck. Talk to a friend or consider checking out a therapist if it really gets too bad, but it seems to me like you're handling things just fine. [NEWLINE] [NEWLINE] Edit: Rule 1 compliance</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The core gameplay is nearly identical (2142, Bad Company 1/2, BF 3/4). Large team battles of capture and holding points. [ENDQ] [NEWLINE] That's like saying every game with a CTF game mode is identical, or that every racing game is the same... Come on. Yes, conquest<mask> a game mode hasn't changed,<mask> a shitton of other game modes have been added<mask>, not to mention Hardcore and other modifications you can choose. And all that beside the fact that the way vehicles are utilized changes the game from every map to another, let alone every game to the next. BF 2 had no play-controlled artillery, for example,<mask> 1942 and 4, the bookends of the series, do. [NEWLINE] [NEWLINE] [STARTQ] CoD has spawned through plenty of time periods [ENDQ] [NEWLINE] Two? A decade in the future isn't a different time period, and Black Ops purports to be Vietnam-era<mask> it's completely anachronistic. [NEWLINE] [NEWLINE] [STARTQ] changed engines [ENDQ] [NEWLINE] Not really<mask> MW unless I'm mistaken. [NEWLINE] [NEWLINE] [STARTQ] All CS:S was, was a facelift to the original CS. [ENDQ] [NEWLINE] My point exactly: CoD is *the exact same thing*, except in 3 times<mask> many editions in 5 fewer years.</s>
Label encoding: <s> [STARTQ] The core gameplay is nearly identical (2142, Bad Company 1/2, BF 3/4). Large team battles of capture and holding points. [ENDQ] [NEWLINE] That's like saying every game with a CTF game mode is identical, or that every racing game is the same... Come on. Yes, conquest as a game mode hasn't changed, but a shitton of other game modes have been added since, not to mention Hardcore and other modifications you can choose. And all that beside the fact that the way vehicles are utilized changes the game from every map to another, let alone every game to the next. BF 2 had no play-controlled artillery, for example, while 1942 and 4, the bookends of the series, do. [NEWLINE] [NEWLINE] [STARTQ] CoD has spawned through plenty of time periods [ENDQ] [NEWLINE] Two? A decade in the future isn't a different time period, and Black Ops purports to be Vietnam-era but it's completely anachronistic. [NEWLINE] [NEWLINE] [STARTQ] changed engines [ENDQ] [NEWLINE] Not really since MW unless I'm mistaken. [NEWLINE] [NEWLINE] [STARTQ] All CS:S was, was a facelift to the original CS. [ENDQ] [NEWLINE] My point exactly: CoD is *the exact same thing*, except in 3 times as many editions in 5 fewer years.</s>
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Be renamed officially? No.<mask> I'm proposing is that society be more accepting of pedophiles. Pedophiles are already demonized. That's one of the reasons a pedophile has such a hard time seeking pro-active treatment,<mask> most treatments are geared towards offenders<mask> it is, and reporting requirements are a further disincentive to those who haven't. *All* pedophiles are demonized, including the ones who never offend,<mask> simply admit that they have that attraction.  And the demonization keeps most who don't want to offend in the shadows, and I'm sure you can see<mask> that is problematic for them<mask> they need or desire help. [NEWLINE] [NEWLINE] <mask> yes, we should recognize each and every pedophile who doesn't offend. We should commend them for that - especially<mask> they have to do<mask> with almost zero outside support, and we should be encouraging them to be open about their attraction, and we should be offering them judgement free therapy<mask> they can better deal with it and live healthy, productive lives. [NEWLINE] [NEWLINE] And again, unless you believe that black people have a natural desire to be jobless or a deadbeat parent, that is a horrible analogy. Whether they act on it or not, pedophiles have a natural attraction to children. </s>
Label encoding: <s>Be renamed officially? No. What I'm proposing is that society be more accepting of pedophiles. Pedophiles are already demonized. That's one of the reasons a pedophile has such a hard time seeking pro-active treatment, since most treatments are geared towards offenders as it is, and reporting requirements are a further disincentive to those who haven't. *All* pedophiles are demonized, including the ones who never offend, but simply admit that they have that attraction.  And the demonization keeps most who don't want to offend in the shadows, and I'm sure you can see why that is problematic for them if they need or desire help. [NEWLINE] [NEWLINE] So yes, we should recognize each and every pedophile who doesn't offend. We should commend them for that - especially since they have to do so with almost zero outside support, and we should be encouraging them to be open about their attraction, and we should be offering them judgement free therapy so they can better deal with it and live healthy, productive lives. [NEWLINE] [NEWLINE] And again, unless you believe that black people have a natural desire to be jobless or a deadbeat parent, that is a horrible analogy. Whether they act on it or not, pedophiles have a natural attraction to children. </s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have been struggling with this<mask> one part of my brain recognizes that people are not<mask> fortunate<mask> I am, don't have the resources or the physical/emotional/social support to sustain themselves and<mask>not, and the other part is deeply mistrustful of what these people have done to end up<mask> they are. [ENDQ] [NEWLINE] An old Cherokee chief was teaching his grandson about life... [NEWLINE] [NEWLINE] "A fight is going on inside me," he said to the boy. [NEWLINE] "It is a terrible fight and it is between two wolves. [NEWLINE] [NEWLINE] "One is evil - he is anger, envy, sorrow, regret, greed, arrogance, self-pity, guilt, resentment, inferiority, lies, false pride, superiority, self-doubt, and ego. [NEWLINE] [NEWLINE] "The other is good - he is joy, peace, love, hope, serenity, humility, kindness, benevolence, empathy, generosity, truth, compassion, and faith. [NEWLINE] [NEWLINE] "This same fight is going on inside you - and inside every other person, too." [NEWLINE] [NEWLINE] The grandson thought about it for a minute and then asked his grandfather, [NEWLINE] "Which wolf will win?" [NEWLINE] [NEWLINE] The old chief simply replied, [NEWLINE] "The one you feed." [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I have been struggling with this because one part of my brain recognizes that people are not as fortunate as I am, don't have the resources or the physical/emotional/social support to sustain themselves and whatnot, and the other part is deeply mistrustful of what these people have done to end up where they are. [ENDQ] [NEWLINE] An old Cherokee chief was teaching his grandson about life... [NEWLINE] [NEWLINE] "A fight is going on inside me," he said to the boy. [NEWLINE] "It is a terrible fight and it is between two wolves. [NEWLINE] [NEWLINE] "One is evil - he is anger, envy, sorrow, regret, greed, arrogance, self-pity, guilt, resentment, inferiority, lies, false pride, superiority, self-doubt, and ego. [NEWLINE] [NEWLINE] "The other is good - he is joy, peace, love, hope, serenity, humility, kindness, benevolence, empathy, generosity, truth, compassion, and faith. [NEWLINE] [NEWLINE] "This same fight is going on inside you - and inside every other person, too." [NEWLINE] [NEWLINE] The grandson thought about it for a minute and then asked his grandfather, [NEWLINE] "Which wolf will win?" [NEWLINE] [NEWLINE] The old chief simply replied, [NEWLINE] "The one you feed." [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0689, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm Australian, and<mask><mask> there's some pretty crazy pants things going on in the USA, like having the second highest per capita incarceration rate in the world,<mask> don't think I'm coming at this from a U-S-A U-S-A perspective, I'm not. The USA has much to improve on. [NEWLINE] [NEWLINE] <mask> to use the words *police state* is completely hyperbolic. [NEWLINE] [NEWLINE] [STARTQ] The Soviet government implemented mass destruction of pre-revolutionary and foreign books and journals from libraries. Only "special collections" (spetskhran), accessible by special permits granted by the KGB, contained old and politically incorrect material.[2] Towards the end of Soviet rule, perestroika led to loosened restrictions on information and publishing. [ENDQ] [URL] [NEWLINE] [NEWLINE] [STARTQ] thousands of political prisoners throughout Germany—and from 1941, throughout the occupied territories under the Night and Fog Decree—simply disappeared<mask> in Gestapo custody. [ENDQ] [URL] [NEWLINE] [NEWLINE] Sure the USA isn't perfect, sometimes acts against its own constitution,<mask><mask> it was a real police state your post would have resulted in you being dragged off to the Gulag. That hasn't happened and that is all the evidence you need to know you don't live in a police state.</s>
Label encoding: <s>I'm Australian, and I think there's some pretty crazy pants things going on in the USA, like having the second highest per capita incarceration rate in the world, so don't think I'm coming at this from a U-S-A U-S-A perspective, I'm not. The USA has much to improve on. [NEWLINE] [NEWLINE] But to use the words *police state* is completely hyperbolic. [NEWLINE] [NEWLINE] [STARTQ] The Soviet government implemented mass destruction of pre-revolutionary and foreign books and journals from libraries. Only "special collections" (spetskhran), accessible by special permits granted by the KGB, contained old and politically incorrect material.[2] Towards the end of Soviet rule, perestroika led to loosened restrictions on information and publishing. [ENDQ] [URL] [NEWLINE] [NEWLINE] [STARTQ] thousands of political prisoners throughout Germany—and from 1941, throughout the occupied territories under the Night and Fog Decree—simply disappeared while in Gestapo custody. [ENDQ] [URL] [NEWLINE] [NEWLINE] Sure the USA isn't perfect, sometimes acts against its own constitution, but if it was a real police state your post would have resulted in you being dragged off to the Gulag. That hasn't happened and that is all the evidence you need to know you don't live in a police state.</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0830, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You seem to be operating under the assumption that art and music are the causes of crime and strife<mask> opposed to merely a reflection of it. Basically<mask> you seem to be saying is, hey world just don't talk about, bring up or acknowledge these problems and they will cease to exist. Furthermore it appears that you are suggesting negative human behaviors such<mask> aggression, greed, disrespect and violence are unique to poverty areas. Which is<mask> incorrect. I understand it can be difficult to translate these behaviors from one context to another,<mask> any act of crime in a ghetto has an equivalent found in suburbs and high society. They are just in different context and perceived differently. There are cultural norms that are destructive across every social class.<mask> this cmv sounds like you are trying to characterize an entire culture, good and bad by cherry picking specific items such<mask> particular verses of rap songs.<mask> not only are you generalizing<mask> focusing on the reflection of society instead of the core causes. In other words, looking at the wrong things, in the wrong order and for the wrong reasons. Do you feel greed and violence us unique to urban American culture? Hopefully you will say no, in which case okay -<mask> then do those traits persist in Syria, or on Wall Street, or in mobil Alabama?</s>
Label encoding: <s>You seem to be operating under the assumption that art and music are the causes of crime and strife as opposed to merely a reflection of it. Basically what you seem to be saying is, hey world just don't talk about, bring up or acknowledge these problems and they will cease to exist. Furthermore it appears that you are suggesting negative human behaviors such as aggression, greed, disrespect and violence are unique to poverty areas. Which is also incorrect. I understand it can be difficult to translate these behaviors from one context to another, but any act of crime in a ghetto has an equivalent found in suburbs and high society. They are just in different context and perceived differently. There are cultural norms that are destructive across every social class. But this cmv sounds like you are trying to characterize an entire culture, good and bad by cherry picking specific items such as particular verses of rap songs. So not only are you generalizing but focusing on the reflection of society instead of the core causes. In other words, looking at the wrong things, in the wrong order and for the wrong reasons. Do you feel greed and violence us unique to urban American culture? Hopefully you will say no, in which case okay - why then do those traits persist in Syria, or on Wall Street, or in mobil Alabama?</s>
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Sine you are Christian, I'm even more shocked that you think it's perfectly find to go around doubting other Christian's faiths and calling them bad Christians. [ENDQ] [NEWLINE] I<mask> don't go around sinning and then saying "<mask> I'm a good Christian". [NEWLINE] [NEWLINE] For the record, I consider myself a practicing, flawed Christian.  I wouldn't say I'm a "good" Christian; and<mask> we're in a discussion of right and wrong,<mask><mask> it's fair game to point out that engaging in the behavior at hand is inconsistent with one who practices the faith. [NEWLINE] [NEWLINE] [STARTQ] Pot is hardly the only product covered in blood that is used in everyday American life.<mask>, legalize it and it won't be a violent product anymore. [ENDQ] [NEWLINE] And we can talk about all those things, too!  *Just<mask> other things are wrong doesn't make this wrong thing okay*. <mask>, weed is<mask>'s currently under discussion. <mask> for your legalization argument, that's not the current state of the world,<mask> until that changes, the situation holds true.  I wouldn't mind taking a hit on a bud -<mask> I can purchase it legally,<mask> I don't have the moral dilemma of<mask> that joint costs tangled up in it.</s>
Label encoding: <s> [STARTQ] Sine you are Christian, I'm even more shocked that you think it's perfectly find to go around doubting other Christian's faiths and calling them bad Christians. [ENDQ] [NEWLINE] I also don't go around sinning and then saying " But I'm a good Christian". [NEWLINE] [NEWLINE] For the record, I consider myself a practicing, flawed Christian.  I wouldn't say I'm a "good" Christian; and since we're in a discussion of right and wrong, I think it's fair game to point out that engaging in the behavior at hand is inconsistent with one who practices the faith. [NEWLINE] [NEWLINE] [STARTQ] Pot is hardly the only product covered in blood that is used in everyday American life. Besides, legalize it and it won't be a violent product anymore. [ENDQ] [NEWLINE] And we can talk about all those things, too!  *Just because other things are wrong doesn't make this wrong thing okay*.  However, weed is what's currently under discussion.  As for your legalization argument, that's not the current state of the world, so until that changes, the situation holds true.  I wouldn't mind taking a hit on a bud - when I can purchase it legally, because I don't have the moral dilemma of what that joint costs tangled up in it.</s>
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Property and possession are not really inherently meaningful terms without a government definition and enforcement. [NEWLINE] [NEWLINE] Imagine a caveman killing a tiger, and skinning it. Does he really have a property right in that skin,<mask><mask><mask> any other caveman with a bigger club can come and take it? [NEWLINE] [NEWLINE] Subsequently, government arose have defined property to regulate<mask> things are distributed between people. We made law: that<mask> you kill a wild animal, the carcass and pelt are your property: [NEWLINE] [NEWLINE] [URL]._Post [NEWLINE] [NEWLINE] <mask> you can see, property is not a meaningful concept unless government defines and enforces<mask> property is. [NEWLINE] [NEWLINE] Later, governments<mask> found that it is useful to extend property laws  to non-tangible things, like works of art or inventions. This was done simply to encourage people to create more art and to invent more.  Same way, the government gave property right in wolf furs in order to encourage wolf hunting. [NEWLINE] [NEWLINE] TL;DR: Property/ownership/stealing with respect to tangible property is exactly<mask> meaningful<mask> Property/ownership/stealing with respect to intellectual property - both are artificially created social constricts. <mask> you have no problem with one, you should be OK with the other. [NEWLINE] </s>
Label encoding: <s>Property and possession are not really inherently meaningful terms without a government definition and enforcement. [NEWLINE] [NEWLINE] Imagine a caveman killing a tiger, and skinning it. Does he really have a property right in that skin, as long as any other caveman with a bigger club can come and take it? [NEWLINE] [NEWLINE] Subsequently, government arose have defined property to regulate how things are distributed between people. We made law: that if you kill a wild animal, the carcass and pelt are your property: [NEWLINE] [NEWLINE] [URL]._Post [NEWLINE] [NEWLINE] As you can see, property is not a meaningful concept unless government defines and enforces what property is. [NEWLINE] [NEWLINE] Later, governments also found that it is useful to extend property laws  to non-tangible things, like works of art or inventions. This was done simply to encourage people to create more art and to invent more.  Same way, the government gave property right in wolf furs in order to encourage wolf hunting. [NEWLINE] [NEWLINE] TL;DR: Property/ownership/stealing with respect to tangible property is exactly as meaningful as Property/ownership/stealing with respect to intellectual property - both are artificially created social constricts.  if you have no problem with one, you should be OK with the other. [NEWLINE] </s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a philosophical question with a personal answer. I would like all of my time to be considered free time. I just happen to want to do something that requires a commitment of working 40 hours a week... [NEWLINE] [NEWLINE] The labor movement in the United States in the late 19th century used the slogan "8 hours for work, 8 hours for sleep, 8 hours for<mask> we will." This has of course become more like 9 hours for work, 2 hours for commuting, 2 hours for self maintenance, 2 hours for eating, 6 hours for sleeping, and 3 hours for whatever else you want to do. That 3 hours probably doesn't even happen for most people... [NEWLINE] [NEWLINE] <mask> my base is that all of my time should be spent<mask> I choose, I instead try to work efficiently.<mask> do you want to spend your energy?<mask> do you recharge? Are you actually working effectively<mask> you work for 10-12 hours each day? I only have about 5 hours of flow in me. [NEWLINE] [NEWLINE] <mask>.<mask> the fuck do you want money? Money is useful<mask> it can solve problems. Money is useful<mask> it frees up your time. Don't make the mistake of being too heads down and shortsighted. [A nice little story about that]( [URL] /).</s>
Label encoding: <s>This is a philosophical question with a personal answer. I would like all of my time to be considered free time. I just happen to want to do something that requires a commitment of working 40 hours a week... [NEWLINE] [NEWLINE] The labor movement in the United States in the late 19th century used the slogan "8 hours for work, 8 hours for sleep, 8 hours for what we will." This has of course become more like 9 hours for work, 2 hours for commuting, 2 hours for self maintenance, 2 hours for eating, 6 hours for sleeping, and 3 hours for whatever else you want to do. That 3 hours probably doesn't even happen for most people... [NEWLINE] [NEWLINE] Since my base is that all of my time should be spent however I choose, I instead try to work efficiently. Where do you want to spend your energy? How do you recharge? Are you actually working effectively when you work for 10-12 hours each day? I only have about 5 hours of flow in me. [NEWLINE] [NEWLINE] Also. Why the fuck do you want money? Money is useful because it can solve problems. Money is useful because it frees up your time. Don't make the mistake of being too heads down and shortsighted. [A nice little story about that]( [URL] /).</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They're actually classified<mask> a hate *site.* Hate groups have to actually do something in the real world to earn the official hate group designation from the SPLC. The MRM<mask> far hasn't murdered anyone that we know of, and they don't have regular IRL meetings like Stormfront or other white rights groups (your basic hate groups) do. [NEWLINE] [NEWLINE] The SPLC recognized r/mensrights, along with other mainstream figureheads of the MRM like AVfM,<mask> one of the leading *misogynistic* sites on the internet. [NEWLINE] [NEWLINE] Now, the definition of misogyny is hatred of women,<mask> people may technically colloquially refer to them<mask> a hate group.<mask> try to do it and they'll throw a dummy spit. [NEWLINE] [NEWLINE] <mask><mask> that their institutionalized misogyny is very worrying,<mask> is its expression<mask> far in doxxing and harassing women, and wishing women would be murdered — [this link features a small selection of murderous wishes from /r/mensrights moderator AnnArchist]( [URL] /). I say small<mask> I've seen more than that,<mask> I don't make a habit of collecting death threats that MRAs make. Hopefully this is something they grow out of. I hope they do before someone gets hurt.</s>
Label encoding: <s>They're actually classified as a hate *site.* Hate groups have to actually do something in the real world to earn the official hate group designation from the SPLC. The MRM so far hasn't murdered anyone that we know of, and they don't have regular IRL meetings like Stormfront or other white rights groups (your basic hate groups) do. [NEWLINE] [NEWLINE] The SPLC recognized r/mensrights, along with other mainstream figureheads of the MRM like AVfM, as one of the leading *misogynistic* sites on the internet. [NEWLINE] [NEWLINE] Now, the definition of misogyny is hatred of women, so people may technically colloquially refer to them as a hate group. But try to do it and they'll throw a dummy spit. [NEWLINE] [NEWLINE] I agree that their institutionalized misogyny is very worrying, as is its expression so far in doxxing and harassing women, and wishing women would be murdered — [this link features a small selection of murderous wishes from /r/mensrights moderator AnnArchist]( [URL] /). I say small because I've seen more than that, but I don't make a habit of collecting death threats that MRAs make. Hopefully this is something they grow out of. I hope they do before someone gets hurt.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You must know<mask> well<mask> I do<mask> that junking a car gets one significantly less money than the value of the car.<mask><mask> I would only get 200$ from the person, then that would get me more money (and possibly even save me money<mask> the junkyard charged to take it). [NEWLINE] [NEWLINE] I'm not saying that it occurs often, I'm saying it can occur. In the right situation, one may only care about making 500 quick dollars right now and neglect to think about raised premiums in the future. (Honestly, whoever would engage in this on purpose would probably not be in the right state of mind in the first place.) And<mask><mask> that you could definitely pick<mask> damage is done.<mask> can be seen [here]( [URL].aspx), there are plenty of scams that can be pulled in order to cause damage to certain parts of your vehicle. [NEWLINE] [NEWLINE] I'm not worried about insurance fraud with already broken parts of the car; I unfortunately agree that this would happen either way. [NEWLINE] [NEWLINE] I understand that you don't think people are putting their vehicles in harm's way,<mask> that doesn't negate the fact that some people do.<mask><mask> that the current system rewards people who think they can make fast money by allowing damage to their car.</s>
Label encoding: <s>You must know as well as I do though that junking a car gets one significantly less money than the value of the car. So if I would only get 200$ from the person, then that would get me more money (and possibly even save me money if the junkyard charged to take it). [NEWLINE] [NEWLINE] I'm not saying that it occurs often, I'm saying it can occur. In the right situation, one may only care about making 500 quick dollars right now and neglect to think about raised premiums in the future. (Honestly, whoever would engage in this on purpose would probably not be in the right state of mind in the first place.) And I think that you could definitely pick where damage is done. As can be seen [here]( [URL].aspx), there are plenty of scams that can be pulled in order to cause damage to certain parts of your vehicle. [NEWLINE] [NEWLINE] I'm not worried about insurance fraud with already broken parts of the car; I unfortunately agree that this would happen either way. [NEWLINE] [NEWLINE] I understand that you don't think people are putting their vehicles in harm's way, but that doesn't negate the fact that some people do. I think that the current system rewards people who think they can make fast money by allowing damage to their car.</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not going to attempt to change your view that it is good for politicians to change their stances.  Instead, I want to change your view about<mask> people are upset at Hillary (it's Hillary you're thinking of, right?) [NEWLINE] [NEWLINE] People aren't upset at Hillary for changing her mind.  I don't even think people are upset that her public stance changed is likely done out of political expediency,<mask> perhaps many are disappointed in that (like I was<mask> Obama changed his public stance likely out of political expedience).  Instead, people view Hillary<mask> exploiting the hard work of gay activists for her own political gain. [NEWLINE] [NEWLINE] <mask> do I mean.  See the following video: [equal]( [URL] ).  Notice that Hillary keeps talking about the hard work it has taken in order to get gay rights<mask><mask><mask> it has. <mask> the video does is to get the viewer to psychologically associate Hillary with that hard work, entreating the viewer to erroneously infer that she has been part of that hard work. <mask> she wasn't part of that hard work,<mask> she only changed her mind on the issue in 2013. <mask>, she is exploiting the hard work of gay rights activist for her own political gain; this is<mask> upset people. </s>
Label encoding: <s>I'm not going to attempt to change your view that it is good for politicians to change their stances.  Instead, I want to change your view about why people are upset at Hillary (it's Hillary you're thinking of, right?) [NEWLINE] [NEWLINE] People aren't upset at Hillary for changing her mind.  I don't even think people are upset that her public stance changed is likely done out of political expediency, although perhaps many are disappointed in that (like I was when Obama changed his public stance likely out of political expedience).  Instead, people view Hillary as exploiting the hard work of gay activists for her own political gain. [NEWLINE] [NEWLINE] What do I mean.  See the following video: [equal]( [URL] ).  Notice that Hillary keeps talking about the hard work it has taken in order to get gay rights as far as it has.  What the video does is to get the viewer to psychologically associate Hillary with that hard work, entreating the viewer to erroneously infer that she has been part of that hard work.  But she wasn't part of that hard work, since she only changed her mind on the issue in 2013.  Thus, she is exploiting the hard work of gay rights activist for her own political gain; this is what upset people. </s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Youre missing the point. I don't listen to rap<mask> I wanna listen to something deep. [ENDQ] [NEWLINE] "CMV, Rap isn't deep." [NEWLINE] [NEWLINE] "I don't listen to rap<mask> I want something deep." [NEWLINE] [NEWLINE] Maybe, just maybe. The fact that you actively avoid engaging with deep rap and certainly don't look for depth<mask> you listen to rap, might have something to do with you not seeing it's depth? [NEWLINE] [NEWLINE] [STARTQ] I certainly wouldn't look towards murs for depth. [ENDQ] [NEWLINE] Then you haven't listened to his stuff. Go listen to The Biggest Lie, from Felt. [NEWLINE] [NEWLINE] [STARTQ] I listen to rap to listen to dope rhymes and delivery. [ENDQ] [NEWLINE] With art, you receive<mask> you put in.<mask> you don't look for depth, you might not find it. The same goes for poetry.<mask> you're reading "The Soul Select's Her Own Society" without actively analyzing, you'll probably find it plain and quaint. [NEWLINE] [NEWLINE] [STARTQ] Wayne is recognized<mask> talented by most rap fans who listen to underground rap music. [ENDQ] [NEWLINE] This made me laugh pretty hard. Wayne is looked down upon by the underground rap scene. He is about<mask> much of a Mainstream sell out<mask> anyone. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Youre missing the point. I don't listen to rap when I wanna listen to something deep. [ENDQ] [NEWLINE] "CMV, Rap isn't deep." [NEWLINE] [NEWLINE] "I don't listen to rap when I want something deep." [NEWLINE] [NEWLINE] Maybe, just maybe. The fact that you actively avoid engaging with deep rap and certainly don't look for depth when you listen to rap, might have something to do with you not seeing it's depth? [NEWLINE] [NEWLINE] [STARTQ] I certainly wouldn't look towards murs for depth. [ENDQ] [NEWLINE] Then you haven't listened to his stuff. Go listen to The Biggest Lie, from Felt. [NEWLINE] [NEWLINE] [STARTQ] I listen to rap to listen to dope rhymes and delivery. [ENDQ] [NEWLINE] With art, you receive what you put in. If you don't look for depth, you might not find it. The same goes for poetry. If you're reading "The Soul Select's Her Own Society" without actively analyzing, you'll probably find it plain and quaint. [NEWLINE] [NEWLINE] [STARTQ] Wayne is recognized as talented by most rap fans who listen to underground rap music. [ENDQ] [NEWLINE] This made me laugh pretty hard. Wayne is looked down upon by the underground rap scene. He is about as much of a Mainstream sell out as anyone. [NEWLINE] </s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The point is that there should not be any artificial restrictions in education, employment etc. Criteria for entry may not be chosen such that they disadvantage specific groups more than the general population, e.g. only prohibiting dreadlocks<mask> a hair style in an office, unnecessarily strict overtime schedules, a minimum height for no legitimate reason (women are shorter on average) etc. They must<mask> make reasonable accommodations, e.g.<mask> there are candidates with disabilities or illnesses. [ENDQ] [NEWLINE] I'm ambivalent on the whole issue of whether employers should be able to discriminate on some of the grounds you mentioned,<mask><mask><mask> these days most wouldn't bother (and those that do can still do<mask> even under anti-discrimination legislation). It seems an illiberal measure,<mask> I can see<mask> the law is in place. [NEWLINE] [NEWLINE] This kind of thing isn't really<mask> I had in mind for the whole topic<mask>. [NEWLINE] [NEWLINE] [STARTQ] <mask> do you mean by this? [ENDQ] [NEWLINE] Success at school (and in life) is a complex mix of social, genetic, economic, behavioural and other factors that could only realistically be controlled by institutionalising every child<mask> that variables can be controlled for. That's<mask><mask><mask> equality of outcome is malign, and not merely undesirable. </s>
Label encoding: <s> [STARTQ] The point is that there should not be any artificial restrictions in education, employment etc. Criteria for entry may not be chosen such that they disadvantage specific groups more than the general population, e.g. only prohibiting dreadlocks as a hair style in an office, unnecessarily strict overtime schedules, a minimum height for no legitimate reason (women are shorter on average) etc. They must also make reasonable accommodations, e.g. when there are candidates with disabilities or illnesses. [ENDQ] [NEWLINE] I'm ambivalent on the whole issue of whether employers should be able to discriminate on some of the grounds you mentioned, because I think these days most wouldn't bother (and those that do can still do so even under anti-discrimination legislation). It seems an illiberal measure, although I can see why the law is in place. [NEWLINE] [NEWLINE] This kind of thing isn't really what I had in mind for the whole topic though. [NEWLINE] [NEWLINE] [STARTQ] What do you mean by this? [ENDQ] [NEWLINE] Success at school (and in life) is a complex mix of social, genetic, economic, behavioural and other factors that could only realistically be controlled by institutionalising every child so that variables can be controlled for. That's why I think equality of outcome is malign, and not merely undesirable. </s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you not think that biological factors might influence behaviour? There are many differences in male and female neuroanatomy. I list some bell curves here and many citations in the root text post itself. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] I don't think that it's actually caused by mathematical disability in women,<mask> an aggregate of other abilities that cause them to select other disciplines. I believe this<mask> women who elect to pursue mathematics are just<mask> successful and talented<mask> men. [NEWLINE] [NEWLINE] Women are better with empathy, better with language, on psychometric evaluations (<mask> again, obviously,<mask> much is culture?) The physical structures of women's brains are physically different from male brains at a measurable level. The structures associated with language are physically larger.<mask><mask> we had a society with more women studying language than men, should we be at all surprised? [NEWLINE] [NEWLINE] In short, I don't think that there is a biological reason for women to not like math,<mask> rather, for them to like other subjects more. [NEWLINE] [NEWLINE] <mask> anyways, my main message was,<mask> much is culture,<mask> much is biology?<mask> would we possibly measure it? Could biological differences account for the entire difference? I believe not. Could they account for the majority of the difference? I believe<mask>. </s>
Label encoding: <s>Do you not think that biological factors might influence behaviour? There are many differences in male and female neuroanatomy. I list some bell curves here and many citations in the root text post itself. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] I don't think that it's actually caused by mathematical disability in women, but an aggregate of other abilities that cause them to select other disciplines. I believe this because women who elect to pursue mathematics are just as successful and talented as men. [NEWLINE] [NEWLINE] Women are better with empathy, better with language, on psychometric evaluations ( but again, obviously, how much is culture?) The physical structures of women's brains are physically different from male brains at a measurable level. The structures associated with language are physically larger. So if we had a society with more women studying language than men, should we be at all surprised? [NEWLINE] [NEWLINE] In short, I don't think that there is a biological reason for women to not like math, but rather, for them to like other subjects more. [NEWLINE] [NEWLINE] But anyways, my main message was, how much is culture, how much is biology? How would we possibly measure it? Could biological differences account for the entire difference? I believe not. Could they account for the majority of the difference? I believe so. </s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here are a few tracks that are truly representative of the Dubstep sound. [NEWLINE] [NEWLINE] [Skream - Dutch Flowerz]( [URL] ) - Lots of funky Dub influences [NEWLINE] [NEWLINE] [Darkstar - Round Ours]( [URL] ) - Very emotionally driven. [NEWLINE] [NEWLINE] [Kahn - Dread]( [URL] ) - Very sparse, with clear Dub/Reggae sounds. [NEWLINE] [NEWLINE] [Skream Ft. Cluekid - Sandsnake]( [URL] ) - Very dark, massive, wobbly sound. [NEWLINE] [NEWLINE] [Silkie - Selva Nova]( [URL] ) - Reminiscent of Old School Jungle, only slower. [NEWLINE] [NEWLINE] Notice<mask> each track sounds fairly unique. This is<mask> dubstep is an incredibly versatile genre, and there's a sound out there for everyone. Please don't let the newer artists and labels with a more westernized sound turn you off,<mask> Skrillex and Monstercat Records are<mask> far from the true Dubstep sound that it should just be categorized<mask> its own genre. The unofficial name is Brostep,<mask> that's kind of a derogatory term honestly. I'd much prefer it be called Metalstep or Heavystep or something. [NEWLINE] [NEWLINE] Listen with an open mind and read about the culture behind dubstep. It's very interesting.</s>
Label encoding: <s>Here are a few tracks that are truly representative of the Dubstep sound. [NEWLINE] [NEWLINE] [Skream - Dutch Flowerz]( [URL] ) - Lots of funky Dub influences [NEWLINE] [NEWLINE] [Darkstar - Round Ours]( [URL] ) - Very emotionally driven. [NEWLINE] [NEWLINE] [Kahn - Dread]( [URL] ) - Very sparse, with clear Dub/Reggae sounds. [NEWLINE] [NEWLINE] [Skream Ft. Cluekid - Sandsnake]( [URL] ) - Very dark, massive, wobbly sound. [NEWLINE] [NEWLINE] [Silkie - Selva Nova]( [URL] ) - Reminiscent of Old School Jungle, only slower. [NEWLINE] [NEWLINE] Notice how each track sounds fairly unique. This is because dubstep is an incredibly versatile genre, and there's a sound out there for everyone. Please don't let the newer artists and labels with a more westernized sound turn you off, because Skrillex and Monstercat Records are so far from the true Dubstep sound that it should just be categorized as its own genre. The unofficial name is Brostep, but that's kind of a derogatory term honestly. I'd much prefer it be called Metalstep or Heavystep or something. [NEWLINE] [NEWLINE] Listen with an open mind and read about the culture behind dubstep. It's very interesting.</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Issue is that Choate's educators really don't make anything near that amount, nor does BB&amp;N.<mask> we want to bring their salaries to 250k, we're literally multiplying their salary 5-10 times. [ENDQ] [NEWLINE] Right. <mask> we can't even convince the rich that the best teachers in the country are important to obtain (which would presumably involve better pay), then<mask> is the convincing case that some teachers are better than others? [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask> a child from Choate saw their educator works for a large salary (and that gives them an additional incentive),<mask><mask> that would isolate them into wanting to work at a private education setting.<mask> a Choate child saw their teacher worked for 250k, then they got a job in a public education school, the Choate teacher is still making over 7 times<mask> a traditional educator would make their first year (35k). I'm not sure<mask> that would do anything more than make private schools more, well, private. [ENDQ] [NEWLINE] <mask> you want something that will really bring excellent candidates into teaching without just using the profit motive. <mask><mask>, this will be really hard.   Maybe religion?  Can we start a cult that values teachers and gives them harems?</s>
Label encoding: <s> [STARTQ] Issue is that Choate's educators really don't make anything near that amount, nor does BB&amp;N. If we want to bring their salaries to 250k, we're literally multiplying their salary 5-10 times. [ENDQ] [NEWLINE] Right.  If we can't even convince the rich that the best teachers in the country are important to obtain (which would presumably involve better pay), then what is the convincing case that some teachers are better than others? [NEWLINE] [NEWLINE] [STARTQ] But, if a child from Choate saw their educator works for a large salary (and that gives them an additional incentive), I think that would isolate them into wanting to work at a private education setting. If a Choate child saw their teacher worked for 250k, then they got a job in a public education school, the Choate teacher is still making over 7 times what a traditional educator would make their first year (35k). I'm not sure if that would do anything more than make private schools more, well, private. [ENDQ] [NEWLINE] So you want something that will really bring excellent candidates into teaching without just using the profit motive.  I agree, this will be really hard.   Maybe religion?  Can we start a cult that values teachers and gives them harems?</s>
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0775, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>/u/WhaleMeatFantasy is actually on to something. This is a very famous philosophical issue called the [problem of induction]( [URL] ) proposed by David Hume. The more specific problem is the [uniformity of nature]( [URL] ) which is<mask> you are assuming: that natural laws are uniform across time and space. This axiom is necessary for us to extrapolate into the unobservable past.<mask> it's just that, an axiom: it's not provable. We assume it<mask> we haven't observed changes in natural law,<mask> that doesn't mean they don't exist. After all, we have been competent observers for a tiny, tiny, miniscule fraction of the universe's existence.<mask> there was some reversal or shift in natural law it's actually quite probable we *wouldn't* observe it. In essence, we assume it must be<mask><mask> we've never seen otherwise,<mask> philosophically the thinking used to get there is not sound. [NEWLINE] [NEWLINE] Mind you, I don't believe this, and<mask> we ever did observe a reversal of natural law I'd be shocked to put it mildly,<mask> Ken Ham's philosophy *is* technically right. Ironically I don't think he knows that himself. </s>
Label encoding: <s>/u/WhaleMeatFantasy is actually on to something. This is a very famous philosophical issue called the [problem of induction]( [URL] ) proposed by David Hume. The more specific problem is the [uniformity of nature]( [URL] ) which is what you are assuming: that natural laws are uniform across time and space. This axiom is necessary for us to extrapolate into the unobservable past. But it's just that, an axiom: it's not provable. We assume it because we haven't observed changes in natural law, but that doesn't mean they don't exist. After all, we have been competent observers for a tiny, tiny, miniscule fraction of the universe's existence. If there was some reversal or shift in natural law it's actually quite probable we *wouldn't* observe it. In essence, we assume it must be so because we've never seen otherwise, but philosophically the thinking used to get there is not sound. [NEWLINE] [NEWLINE] Mind you, I don't believe this, and if we ever did observe a reversal of natural law I'd be shocked to put it mildly, but Ken Ham's philosophy *is* technically right. Ironically I don't think he knows that himself. </s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your reasoning is precisely<mask> anti-discrimination laws had to be put in place. [ENDQ] [NEWLINE] Herein lies the rub; this employer is just trying to do a good job of managing his employees and company. In the US, that discrimination is rewarded,<mask> it is very difficult to prove sexual discrimination in hiring without significant insider knowledge of past hiring practices. Beyond that, the fact that there are inequal laws regarding parental leave and family law essentially sets the precedent at the government level, implying that it is ok to treat one sex differently than another<mask> it comes to parental rights and family leave. [NEWLINE] [NEWLINE] <mask> OP is certainly taking a risk in his hiring practices, it seems his opinion is based on<mask><mask><mask> cost effective for his business; OP clearly does not see enforcing the law<mask> one of his primary responsibilities. [NEWLINE] [NEWLINE] You can call OP un-American or whatever,<mask> bottom line, it's the government's job to enforce non-discrimination, and<mask> you set up laws favorable to one sex over another<mask> it comes to employment leave, you are essentially telling employers "make sure you treat this class of employees differently,<mask> make sure you don't get caught doing it unless it's specifically legally allowed, or we'll allow them to sue you."</s>
Label encoding: <s> [STARTQ] Your reasoning is precisely why anti-discrimination laws had to be put in place. [ENDQ] [NEWLINE] Herein lies the rub; this employer is just trying to do a good job of managing his employees and company. In the US, that discrimination is rewarded, because it is very difficult to prove sexual discrimination in hiring without significant insider knowledge of past hiring practices. Beyond that, the fact that there are inequal laws regarding parental leave and family law essentially sets the precedent at the government level, implying that it is ok to treat one sex differently than another when it comes to parental rights and family leave. [NEWLINE] [NEWLINE] While OP is certainly taking a risk in his hiring practices, it seems his opinion is based on what is more cost effective for his business; OP clearly does not see enforcing the law as one of his primary responsibilities. [NEWLINE] [NEWLINE] You can call OP un-American or whatever, but bottom line, it's the government's job to enforce non-discrimination, and when you set up laws favorable to one sex over another when it comes to employment leave, you are essentially telling employers "make sure you treat this class of employees differently, but make sure you don't get caught doing it unless it's specifically legally allowed, or we'll allow them to sue you."</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There really isnt any 'free market' on the planet for any industry,<mask> the US healthcare system is one of the most free on the planet. Most other countries have drastically higher regulations and socialism, and almost all have lower prices and better outcomes. [ENDQ] [NEWLINE] The US health care market may be relatively more free than health care in socialized countries in certain ways,<mask> it is still one of the most unfree industries in the country. [NEWLINE] [NEWLINE] To see relatively more free markets in health care, you have to venture outside first-world countries, to places like Mexico, Brazil, Costa Rica, and Thailand, and examine the private care systems there which attract international patients. People from first-world countries commonly engage in medical tourism to these and other countries in order to get<mask> they want. People from the US travel internationally to get better price on medical procedures,<mask> people from countries with socialized medicine do it to avoid long wait times. [NEWLINE] [NEWLINE] It is in these markets that we find the closest examples of free market health care in the world: prices are transparent, consumers shop around to find quality providers at a good price, and they pay out of pocket,<mask> there are no middle-men like governments or insurance companies to create overhead.</s><pad>
Label encoding: <s> [STARTQ] There really isnt any 'free market' on the planet for any industry, but the US healthcare system is one of the most free on the planet. Most other countries have drastically higher regulations and socialism, and almost all have lower prices and better outcomes. [ENDQ] [NEWLINE] The US health care market may be relatively more free than health care in socialized countries in certain ways, but it is still one of the most unfree industries in the country. [NEWLINE] [NEWLINE] To see relatively more free markets in health care, you have to venture outside first-world countries, to places like Mexico, Brazil, Costa Rica, and Thailand, and examine the private care systems there which attract international patients. People from first-world countries commonly engage in medical tourism to these and other countries in order to get what they want. People from the US travel internationally to get better price on medical procedures, while people from countries with socialized medicine do it to avoid long wait times. [NEWLINE] [NEWLINE] It is in these markets that we find the closest examples of free market health care in the world: prices are transparent, consumers shop around to find quality providers at a good price, and they pay out of pocket, so there are no middle-men like governments or insurance companies to create overhead.</s><pad>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're right.  In a career, it is wise to compare yourself to your peers,<mask><mask> you fall too far behind you could be replaced by someone better and be out of a job. [NEWLINE] [NEWLINE] <mask>,<mask><mask> the old adage "don't compare yourself to others" has a broader scope than you are using.  Lets say you're 25, and relatively new to your career.  There is another guy at the place<mask> you work who is 50 years old, and has been in this career for nearly 30 years.  He knows all the ins and outs.  You could never hope to be<mask> good<mask> this guy,<mask> he's been working this job<mask> much longer and has<mask> much more experience.  Is it appropriate to compare yourself to him? <mask> good will it do to compare yourself to him? [NEWLINE] [NEWLINE] Likewise, you're sitting at home, and you look at your neighbor - he has a much nicer house, car, clothes, and everything.  Well, it turns out he's a very successful doctor and just makes a lot more money than you.  Is it appropriate to compare<mask> he has to<mask> you have? <mask> good will it do to compare yourself to him?</s>
Label encoding: <s>I think you're right.  In a career, it is wise to compare yourself to your peers, since if you fall too far behind you could be replaced by someone better and be out of a job. [NEWLINE] [NEWLINE] However, I think the old adage "don't compare yourself to others" has a broader scope than you are using.  Lets say you're 25, and relatively new to your career.  There is another guy at the place where you work who is 50 years old, and has been in this career for nearly 30 years.  He knows all the ins and outs.  You could never hope to be as good as this guy, because he's been working this job so much longer and has so much more experience.  Is it appropriate to compare yourself to him?  What good will it do to compare yourself to him? [NEWLINE] [NEWLINE] Likewise, you're sitting at home, and you look at your neighbor - he has a much nicer house, car, clothes, and everything.  Well, it turns out he's a very successful doctor and just makes a lot more money than you.  Is it appropriate to compare what he has to what you have?  What good will it do to compare yourself to him?</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] [URL] [ENDQ] [NEWLINE] This goes to a subredditdrama thread about ViolentAcrez being targeted by Adrian Chen, a writer for a non-reddit website which isn't affiliated with SRS. It does link to an SRS thread in which SRS indicates it's happy to see him suffer,<mask> is not evidence that SRS is behind Mr. Chen's investigation. [NEWLINE] [NEWLINE] [STARTQ] And Acebulf's line of comments in that CMV topic provide a good picture of unsavory events SRS has been involved in: here. [ENDQ] [NEWLINE] This is just a bunch of people talking about<mask> they dislike SRS and speculating that SRS is behind various things, not evidence. [NEWLINE] [NEWLINE] Specifically,<mask> clearly I wasn't specific enough,<mask> I'm asking for is a link to the topic (or screenshots of it) that SRS is alleged to have mass-commented on encouraging suicide, or a link (or screenshot) to direct evidence that SRS<mask> a sub has encouraged or facilitated doxxing, not links to doxxing that is speculated to have been related in some way to individual members of SRS (<mask> that type of evidence could actually indict all of Reddit,<mask> everyone here is a member of Reddit).</s>
Label encoding: <s> [STARTQ] [URL] [ENDQ] [NEWLINE] This goes to a subredditdrama thread about ViolentAcrez being targeted by Adrian Chen, a writer for a non-reddit website which isn't affiliated with SRS. It does link to an SRS thread in which SRS indicates it's happy to see him suffer, but is not evidence that SRS is behind Mr. Chen's investigation. [NEWLINE] [NEWLINE] [STARTQ] And Acebulf's line of comments in that CMV topic provide a good picture of unsavory events SRS has been involved in: here. [ENDQ] [NEWLINE] This is just a bunch of people talking about how they dislike SRS and speculating that SRS is behind various things, not evidence. [NEWLINE] [NEWLINE] Specifically, since clearly I wasn't specific enough, what I'm asking for is a link to the topic (or screenshots of it) that SRS is alleged to have mass-commented on encouraging suicide, or a link (or screenshot) to direct evidence that SRS as a sub has encouraged or facilitated doxxing, not links to doxxing that is speculated to have been related in some way to individual members of SRS ( since that type of evidence could actually indict all of Reddit, since everyone here is a member of Reddit).</s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm enjoying the discussion either way,<mask> out of curiosity, do you really believe in consensual genocide? [NEWLINE] [NEWLINE] [NEWLINE] I'd be interested to hear your arguments for the sustainability of a species. You've said genocide is wrong,<mask> not<mask> sustaining humanity is right. [NEWLINE] [NEWLINE] [NEWLINE] Regarding your point re:immaturity-- [NEWLINE] I like to think that antinatilism itself is not immature. Those who think life is not worth living may or may not be making an immature assumption.<mask> antinatilism (or at least my argument) does not say those people are right. It just says we shouldn't force them to endure lives<mask><mask> *our own* views of life<mask> worthwhile. [NEWLINE] [NEWLINE] <mask> that begs an interesting question: is a preference for nonexistence *objectively less correct/valuable/moral* than a preference for existence? Does a person favoring nonexistence have less of a right to live<mask><mask> his/her view than a person favoring existence does? I can't think of<mask> they would be less deserving of autonomy,<mask>,<mask> they are, this could definitely knock down my argument. [NEWLINE] I'd like to hear some reasons- can you (or anyone) support this idea?</s>
Label encoding: <s>I'm enjoying the discussion either way, but out of curiosity, do you really believe in consensual genocide? [NEWLINE] [NEWLINE] [NEWLINE] I'd be interested to hear your arguments for the sustainability of a species. You've said genocide is wrong, but not why sustaining humanity is right. [NEWLINE] [NEWLINE] [NEWLINE] Regarding your point re:immaturity-- [NEWLINE] I like to think that antinatilism itself is not immature. Those who think life is not worth living may or may not be making an immature assumption. But antinatilism (or at least my argument) does not say those people are right. It just says we shouldn't force them to endure lives according to *our own* views of life as worthwhile. [NEWLINE] [NEWLINE] But that begs an interesting question: is a preference for nonexistence *objectively less correct/valuable/moral* than a preference for existence? Does a person favoring nonexistence have less of a right to live according to his/her view than a person favoring existence does? I can't think of why they would be less deserving of autonomy, but, if they are, this could definitely knock down my argument. [NEWLINE] I'd like to hear some reasons- can you (or anyone) support this idea?</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, you're right. That was just poor phrasing. It was an afterthought. My point focused on<mask> it was possible to discern between human life and human matter. That opinion was to add something,<mask> a side note. That said,<mask><mask> any abortion in which the fetus doesn't endanger the mother and has a chance at a relatively normal,<mask> short life, is wrong.<mask><mask> I can stand by that phrasing. [NEWLINE] That said, I have a few objections to your reasoning. First, using metaphors like that desensitizes the subject matter. I can see your reasoning,<mask> it probably makes most people forget you're talking about human life and not facial hair growth. That makes it easier to accept that defining stubbles has the same importance<mask> defining human life,<mask> both should be thought with the same emphasis. [NEWLINE] The second objection: people will look at anyone who holds a different idea like their crazy; that has nothing to due with the idea's value. [NEWLINE] The final objection: you identified<mask> you say is a flawed logic,<mask> you didn't address the questions I posed.<mask> there is a grey area,<mask> can one person or a government decide on<mask> is permissible or not. [NEWLINE] </s>
Label encoding: <s>Yes, you're right. That was just poor phrasing. It was an afterthought. My point focused on how it was possible to discern between human life and human matter. That opinion was to add something, as a side note. That said, I think any abortion in which the fetus doesn't endanger the mother and has a chance at a relatively normal, however short life, is wrong. I think I can stand by that phrasing. [NEWLINE] That said, I have a few objections to your reasoning. First, using metaphors like that desensitizes the subject matter. I can see your reasoning, but it probably makes most people forget you're talking about human life and not facial hair growth. That makes it easier to accept that defining stubbles has the same importance as defining human life, so both should be thought with the same emphasis. [NEWLINE] The second objection: people will look at anyone who holds a different idea like their crazy; that has nothing to due with the idea's value. [NEWLINE] The final objection: you identified what you say is a flawed logic, but you didn't address the questions I posed. If there is a grey area, how can one person or a government decide on what is permissible or not. [NEWLINE] </s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>(<mask> this is definitely off topic from the question that OP asked, I'll bite, slightly...) [NEWLINE] [NEWLINE] <mask> would you kill an aspect of Yahweh himself? Especially<mask> the aspect is,<mask> it is in many sects of Judaism, a metaphysical concept that represents a person or idea that is a corrupting force in someone's life? To many 'Ha-Satan' is simply a force with no true free will or personality that acts<mask> a necessary cog in their God's plan. [NEWLINE] [NEWLINE] Oddly enough, outside of Hassidic and Haredi youths getting into fights in the streets, the most aggressive and militant modern Jews are the Secular Jewish Zionists... Which are primarily atheist or agnostic. These are people who act on Jewish culture, rather than religion or faith. [NEWLINE] [NEWLINE] I'm not arguing that people don't do<mask> you described,<mask> I do think that this generally is mostly on the people responsible, not their religion. Judaism is a religion<mask> all things are up for debate and all sides should be argued.<mask> later stories contain tales of brutal wars and even genocides, many of the core books lay out specific rules on giving mercy to your enemies in any way possible. </s>
Label encoding: <s>( While this is definitely off topic from the question that OP asked, I'll bite, slightly...) [NEWLINE] [NEWLINE] Why would you kill an aspect of Yahweh himself? Especially if the aspect is, as it is in many sects of Judaism, a metaphysical concept that represents a person or idea that is a corrupting force in someone's life? To many 'Ha-Satan' is simply a force with no true free will or personality that acts as a necessary cog in their God's plan. [NEWLINE] [NEWLINE] Oddly enough, outside of Hassidic and Haredi youths getting into fights in the streets, the most aggressive and militant modern Jews are the Secular Jewish Zionists... Which are primarily atheist or agnostic. These are people who act on Jewish culture, rather than religion or faith. [NEWLINE] [NEWLINE] I'm not arguing that people don't do what you described, but I do think that this generally is mostly on the people responsible, not their religion. Judaism is a religion where all things are up for debate and all sides should be argued. While later stories contain tales of brutal wars and even genocides, many of the core books lay out specific rules on giving mercy to your enemies in any way possible. </s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In *practice* I largely agree with you,<mask> I have to disagree on the underpinnings of your argument. Namely this: [NEWLINE] [NEWLINE] [STARTQ] intrinsic worth of human life [ENDQ] [NEWLINE] Any intrinsic value to human life is an arbitrary construct that we humans have created in order to provide rationale for our actions. Let me rephrase that<mask><mask><mask> it's really important: Our concept of 'rights' and 'intrinsic value of humans' is an after-the-fact justification to explain the way that human societies tend to behave. [NEWLINE] [NEWLINE] Humans are animals, just like any other. We seem to be (<mask><mask><mask> we can observe) at the extreme end of the distribution<mask> it comes to some behaviors,<mask> we are creatures of biology and our social constructs are rooted in behaviors that increased the survivability and fitness of our ancestors. ('Fitness' here means ability to pass on genes.) Traits like empathy, trust, altruism, etc... these are all things that may harm individual fitness,<mask><mask> operating in a group can increase the overall fitness of that group. I believe that this is the origin of our concepts about any 'intrinsic value' to life.</s>
Label encoding: <s>In *practice* I largely agree with you, however I have to disagree on the underpinnings of your argument. Namely this: [NEWLINE] [NEWLINE] [STARTQ] intrinsic worth of human life [ENDQ] [NEWLINE] Any intrinsic value to human life is an arbitrary construct that we humans have created in order to provide rationale for our actions. Let me rephrase that because I think it's really important: Our concept of 'rights' and 'intrinsic value of humans' is an after-the-fact justification to explain the way that human societies tend to behave. [NEWLINE] [NEWLINE] Humans are animals, just like any other. We seem to be ( as far as we can observe) at the extreme end of the distribution when it comes to some behaviors, but we are creatures of biology and our social constructs are rooted in behaviors that increased the survivability and fitness of our ancestors. ('Fitness' here means ability to pass on genes.) Traits like empathy, trust, altruism, etc... these are all things that may harm individual fitness, but when operating in a group can increase the overall fitness of that group. I believe that this is the origin of our concepts about any 'intrinsic value' to life.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have been getting this response over and over and I have to just point out: [NEWLINE] [NEWLINE] I started my post saying that I don't agree with bashing/bullying/oppressing anyone. I firmly believe in respect.<mask> I do believe no one has the right, and shouldn't, hurl insults and hurt people's feelings intentionally, I don't think there's anything wrong with noting that some people are ass holes, and trying to be encouraging and helpful to those in need of help and support. [NEWLINE] [NEWLINE] Just<mask> people shouldn't be ass holes, doesn't mean they aren't going to be. And<mask><mask><mask>, getting hung up on<mask> the ass holes are ass holes is less beneficial to society.<mask> people focus on<mask> the oppressors are oppressing, they only end up sharing more hate. Victors win, and victors don't listen to the haters. I would love to see equal rights for all sexes and genders, for all races and ethnicities, respect for all.<mask> the same way an older sibling "wins"<mask> he gets his little brother to cry from picking on him,<mask> do the oppressing forces win<mask> all they get are negative reactions from the people they dislike.</s>
Label encoding: <s>I have been getting this response over and over and I have to just point out: [NEWLINE] [NEWLINE] I started my post saying that I don't agree with bashing/bullying/oppressing anyone. I firmly believe in respect. While I do believe no one has the right, and shouldn't, hurl insults and hurt people's feelings intentionally, I don't think there's anything wrong with noting that some people are ass holes, and trying to be encouraging and helpful to those in need of help and support. [NEWLINE] [NEWLINE] Just because people shouldn't be ass holes, doesn't mean they aren't going to be. And in my opinion, getting hung up on how the ass holes are ass holes is less beneficial to society. when people focus on how the oppressors are oppressing, they only end up sharing more hate. Victors win, and victors don't listen to the haters. I would love to see equal rights for all sexes and genders, for all races and ethnicities, respect for all. But the same way an older sibling "wins" when he gets his little brother to cry from picking on him, so do the oppressing forces win when all they get are negative reactions from the people they dislike.</s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem I see most of the time with this is that time seems to be relative after 5pm.  I would in IT and hear all the time<mask> people stay to all hours of the night.  I typically come in at 715 or 730<mask> I am out like a cannon at 5.  For years I heard<mask> I needed to stay later like everyone else.  They all came in after 8 and some closer to 9. <mask> I thought there was no way I was staying to 6 or 7,<mask> it would try it.  I started staying later only to find out those same people were all gone by 530. <mask><mask> you come in late and stay the appropriate amount of time after you should leave you are good. <mask> my experience is that time seems to double for those that stay after 5. <mask> remember, for a 8 hour day with a lunch<mask> you come in at 10 you have to leave at 7. <mask> I find is at best those coming in late are staying until 6, with a lunch. [NEWLINE] [NEWLINE] Tl;dr I have the opposite problem at my work and that tells me he's probably not stay long enough to make up for coming in later,</s>
Label encoding: <s>The problem I see most of the time with this is that time seems to be relative after 5pm.  I would in IT and hear all the time how people stay to all hours of the night.  I typically come in at 715 or 730 so I am out like a cannon at 5.  For years I heard how I needed to stay later like everyone else.  They all came in after 8 and some closer to 9.  So I thought there was no way I was staying to 6 or 7, but it would try it.  I started staying later only to find out those same people were all gone by 530.  So if you come in late and stay the appropriate amount of time after you should leave you are good.  But my experience is that time seems to double for those that stay after 5.  So remember, for a 8 hour day with a lunch if you come in at 10 you have to leave at 7.  What I find is at best those coming in late are staying until 6, with a lunch. [NEWLINE] [NEWLINE] Tl;dr I have the opposite problem at my work and that tells me he's probably not stay long enough to make up for coming in later,</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can't convince you to *love* your parents. It doesn't work that way.<mask> I can encourage you to do is to think twice about cutting them out. [NEWLINE] [NEWLINE] Frankly, it's healthy (to a point) to push them away;<mask> they've done their job half-well, you'll be ready to face things on your own. And everyone's parents are bound to have a different sense of the world.<mask><mask> their sense of things is warped, don't forget that yours might be too.<mask> Nietzsche once said, every philosophy is the philosophy of some stage of life. [NEWLINE] [NEWLINE] Parenthood is not a reciprocal arrangement. You are not bound to care about them the same way they care about you.<mask> you may understand someday<mask> you are a parent. Try and imagine it. [NEWLINE] [NEWLINE] For now, you can't force yourself to care. Much less, none of us can do that for you.<mask> think of it<mask> a favor to people who have done a lot for you. Call them regularly, just to say hello and catch up. Later on you'll be glad you did, and in the meantime, you'll be giving them a comfort that you will someday appreciate. </s>
Label encoding: <s>I can't convince you to *love* your parents. It doesn't work that way. What I can encourage you to do is to think twice about cutting them out. [NEWLINE] [NEWLINE] Frankly, it's healthy (to a point) to push them away; if they've done their job half-well, you'll be ready to face things on your own. And everyone's parents are bound to have a different sense of the world. But if their sense of things is warped, don't forget that yours might be too. As Nietzsche once said, every philosophy is the philosophy of some stage of life. [NEWLINE] [NEWLINE] Parenthood is not a reciprocal arrangement. You are not bound to care about them the same way they care about you. But you may understand someday when you are a parent. Try and imagine it. [NEWLINE] [NEWLINE] For now, you can't force yourself to care. Much less, none of us can do that for you. But think of it as a favor to people who have done a lot for you. Call them regularly, just to say hello and catch up. Later on you'll be glad you did, and in the meantime, you'll be giving them a comfort that you will someday appreciate. </s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think *most* people view them<mask> *completely* insane.  I personally know many Mormons, and they are definitely not insane even a little bit.  I went to a Christian church last Sunday with a friend of mine, and I can totally see<mask><mask> many people believe in that stuff.  The pastor was literally telling them that questioning the authenticity of the Bible was immoral.  There was a whole community of people agreeing that it is wrong, even evil, to *think* about their religious beliefs.  I can't judge people who were born into organized religions.  I'm sure I'd be religious too<mask> I was taught from a young age that one of the worst things I could do in life was to let down my guard and allow myself to contemplate the possibility that somebody made it up. [NEWLINE] [NEWLINE] Sure, you can tell a Mormon that Joseph Smith was a fraud,<mask> they *know* he wasn't, and<mask> they stop to think about<mask> sure they are, they are committing a serious sin. <mask> you ever catch yourself even considering whether your friend is basing these claims on historical fact, you are letting down your family, your God, and most importantly, yourself.</s>
Label encoding: <s>I don't think *most* people view them as *completely* insane.  I personally know many Mormons, and they are definitely not insane even a little bit.  I went to a Christian church last Sunday with a friend of mine, and I can totally see how so many people believe in that stuff.  The pastor was literally telling them that questioning the authenticity of the Bible was immoral.  There was a whole community of people agreeing that it is wrong, even evil, to *think* about their religious beliefs.  I can't judge people who were born into organized religions.  I'm sure I'd be religious too if I was taught from a young age that one of the worst things I could do in life was to let down my guard and allow myself to contemplate the possibility that somebody made it up. [NEWLINE] [NEWLINE] Sure, you can tell a Mormon that Joseph Smith was a fraud, but they *know* he wasn't, and if they stop to think about how sure they are, they are committing a serious sin.  If you ever catch yourself even considering whether your friend is basing these claims on historical fact, you are letting down your family, your God, and most importantly, yourself.</s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I guess I brought my basketball to the ice rink. I have to say that you do bring up good questions that got me thinking, specifically in regards to culture<mask>, I cannot think of a culture foreign or domestic that has existed in a different context that had a differing outcome in regards to power, morality and views on sexuality.<mask> we are to talk about human nature surly we must talk about cultures that have existed, cultures that are merely postulated might<mask> well be aliens. [NEWLINE] [NEWLINE] Lets talk about war, every culture has been involved in the act of waging warfare.<mask> can any moral human being justify the taking of another human life? Through pride of their own superiority? For economic gain?<mask> can anyone say that one human has a right to put their fellow countrymen/woman at risk? We went to war to spread our ideology "freedom, capitalism" the Soviets did the same, Germans and the Japanese.<mask> human nature is good<mask> are we<mask> easily swayed through ideologies to commit such violence? [NEWLINE] [NEWLINE] The only culture I can think of that might have been different are the Minoans<mask>, I'm unsure<mask> of the lack of certain historical evidence. Again leaving us to postulate.</s>
Label encoding: <s>Well I guess I brought my basketball to the ice rink. I have to say that you do bring up good questions that got me thinking, specifically in regards to culture however, I cannot think of a culture foreign or domestic that has existed in a different context that had a differing outcome in regards to power, morality and views on sexuality. If we are to talk about human nature surly we must talk about cultures that have existed, cultures that are merely postulated might as well be aliens. [NEWLINE] [NEWLINE] Lets talk about war, every culture has been involved in the act of waging warfare. How can any moral human being justify the taking of another human life? Through pride of their own superiority? For economic gain? How can anyone say that one human has a right to put their fellow countrymen/woman at risk? We went to war to spread our ideology "freedom, capitalism" the Soviets did the same, Germans and the Japanese. If human nature is good how are we so easily swayed through ideologies to commit such violence? [NEWLINE] [NEWLINE] The only culture I can think of that might have been different are the Minoans but, I'm unsure because of the lack of certain historical evidence. Again leaving us to postulate.</s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason condoms are readily availible<mask> a lot of feminine birth control is not is simple. Condoms are really cheap in comparison, and<mask> prevents the majority of STI's. In my experience, the university almost never go for the good brands, and people are better off buying their own at a drug store.<mask>, and I can't stress this enough, Universities etc are not *required* to provide condoms. They do<mask> in part<mask> the cost of the condoms are quite a bit less than the cost of a few cases of accidental pregnancies would tie up in resources (the girl and boy might drop out, or require extensive counselling). [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, things like birth control pills have a higher associated risk. The university, by providing these things, take on a responsibility for peopel who use them. They are unlikely to take on the liability for birth control pills,<mask> it's a lot more risky. The worst that can happen with a condom is<mask> the person is allergic to latex, whereas birth control pills affect the inside of the body and a lot more can go wrong. Plus it's a hell of a lot more expensive and requires more maintenance, and advice from a doctor.</s>
Label encoding: <s>The reason condoms are readily availible where a lot of feminine birth control is not is simple. Condoms are really cheap in comparison, and also prevents the majority of STI's. In my experience, the university almost never go for the good brands, and people are better off buying their own at a drug store. Also, and I can't stress this enough, Universities etc are not *required* to provide condoms. They do so in part because the cost of the condoms are quite a bit less than the cost of a few cases of accidental pregnancies would tie up in resources (the girl and boy might drop out, or require extensive counselling). [NEWLINE] [NEWLINE] On the other hand, things like birth control pills have a higher associated risk. The university, by providing these things, take on a responsibility for peopel who use them. They are unlikely to take on the liability for birth control pills, since it's a lot more risky. The worst that can happen with a condom is if the person is allergic to latex, whereas birth control pills affect the inside of the body and a lot more can go wrong. Plus it's a hell of a lot more expensive and requires more maintenance, and advice from a doctor.</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the standard of living has increased<mask> much over the past decades that comparing the middle class in 1990 to 2015 is meaningless without comparing standard of living with it. [ENDQ] [NEWLINE] There is an argument to be made that the standard of living for the middle class has been increased<mask> becoming less and less attainable<mask><mask><mask> of our fascination with the lifestyles of the rich and the famous. The middle class have usually<mask>pired to become upper-middle class or wealthy, and the rise in inequality has lead to an inflation in<mask> those terms mean as well, and<mask> is aspired to. [NEWLINE] [NEWLINE] Lifestyle inflation hasn't occurred in a vacuum and has likely been heavily influenced by wealth distribution and income inequality,<mask> instead of beating that dead horse I am attempting to examine a different angle. [NEWLINE] [NEWLINE] *Not<mask> ninja edit:<mask><mask> that<mask> the overall pie is larger and you are getting more pie<mask><mask> your slice is proportionally smaller than it once was this can be a good thing,<mask> it can<mask> lead to some really big issues.<mask> that one guy with 99% is doing with it likely isn't "for the greater good"<mask> acquiring that much means something went very wrong somewhere in the process.</s>
Label encoding: <s> [STARTQ] the standard of living has increased so much over the past decades that comparing the middle class in 1990 to 2015 is meaningless without comparing standard of living with it. [ENDQ] [NEWLINE] There is an argument to be made that the standard of living for the middle class has been increased while becoming less and less attainable as a result of our fascination with the lifestyles of the rich and the famous. The middle class have usually aspired to become upper-middle class or wealthy, and the rise in inequality has lead to an inflation in what those terms mean as well, and what is aspired to. [NEWLINE] [NEWLINE] Lifestyle inflation hasn't occurred in a vacuum and has likely been heavily influenced by wealth distribution and income inequality, but instead of beating that dead horse I am attempting to examine a different angle. [NEWLINE] [NEWLINE] *Not so ninja edit: I agree that if the overall pie is larger and you are getting more pie even though your slice is proportionally smaller than it once was this can be a good thing, but it can also lead to some really big issues. What that one guy with 99% is doing with it likely isn't "for the greater good" because acquiring that much means something went very wrong somewhere in the process.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it was that little and the person using it used only very little per month I imagine it wouldn't be worth installing and maintaining for that person. [ENDQ] [NEWLINE] Like I mentioned,<mask> the infrastructure was already existing,<mask> would it be<mask> hard to govern? We already have laws demanding that utilities must be provided and maintained even for rural areas, it isn't that much different once we have an infrastructure going. You could just<mask> well<mask><mask> to provide electricity for a rural farmland costs a ton, and<mask> it wouldn't be worth installing. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> you sorely underestimate the amount of people who are online for a good chunk of each day. [ENDQ] [NEWLINE] Again, this is not about<mask> "this generation does" or<mask> "everyone is now using everything online". Everyone can turn on their air-conditioning 24/7<mask> they can afford to,<mask><mask> I choose to be frugal<mask> does it matter<mask> much *other people* choose to use? Everyone can use a Tb per month for all I care,<mask><mask> all I plan to use it for is reading the news and sending emails,<mask> does it matter<mask> much data *other people* choose to use?</s>
Label encoding: <s> [STARTQ] If it was that little and the person using it used only very little per month I imagine it wouldn't be worth installing and maintaining for that person. [ENDQ] [NEWLINE] Like I mentioned, if the infrastructure was already existing, why would it be so hard to govern? We already have laws demanding that utilities must be provided and maintained even for rural areas, it isn't that much different once we have an infrastructure going. You could just as well argue that to provide electricity for a rural farmland costs a ton, and therefore it wouldn't be worth installing. [NEWLINE] [NEWLINE] [STARTQ] I think you sorely underestimate the amount of people who are online for a good chunk of each day. [ENDQ] [NEWLINE] Again, this is not about what "this generation does" or how "everyone is now using everything online". Everyone can turn on their air-conditioning 24/7 if they can afford to, but if I choose to be frugal why does it matter how much *other people* choose to use? Everyone can use a Tb per month for all I care, but if all I plan to use it for is reading the news and sending emails, why does it matter how much data *other people* choose to use?</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I see at least one of these STEMs are the best threads a month.<mask> I have never seen one going the other way. This leads me to think STEM people are really insecure and/or arrogant. [ENDQ] [NEWLINE] I'm not STEM at all.<mask><mask>, it's the first thing I wrote. I put in that disclaimer specifically<mask> I knew somebody would work in an ad hominem like you did, putting into question the moral fibre of every STEM person. [NEWLINE] [NEWLINE] [STARTQ] Do you have any source for this<mask> 'I remember in high school...'? Your high school experience is not a valid argument.<mask> you had studied more about logical fallacies in 'inferior' subjects like philosophy you'd know that. [ENDQ] [NEWLINE] Dude, it's not possible to spend one day on reddit without quickly learning everything about fallacies. It's one of the most frequently used concepts on this site. I know my anecdotes aren't valid.<mask> I would be prepared to wager huge sums of money that the lessons they have given me are correct: namely that the best do STEM. It's not possible to do well in STEM without thinking a lot. That's not the case for humanities.</s>
Label encoding: <s> [STARTQ] I see at least one of these STEMs are the best threads a month. But I have never seen one going the other way. This leads me to think STEM people are really insecure and/or arrogant. [ENDQ] [NEWLINE] I'm not STEM at all. In fact, it's the first thing I wrote. I put in that disclaimer specifically because I knew somebody would work in an ad hominem like you did, putting into question the moral fibre of every STEM person. [NEWLINE] [NEWLINE] [STARTQ] Do you have any source for this besides 'I remember in high school...'? Your high school experience is not a valid argument. If you had studied more about logical fallacies in 'inferior' subjects like philosophy you'd know that. [ENDQ] [NEWLINE] Dude, it's not possible to spend one day on reddit without quickly learning everything about fallacies. It's one of the most frequently used concepts on this site. I know my anecdotes aren't valid. However I would be prepared to wager huge sums of money that the lessons they have given me are correct: namely that the best do STEM. It's not possible to do well in STEM without thinking a lot. That's not the case for humanities.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I never said to starve yourself. A lot of people have an expectation that they need to be full after you eat. That's not true. Once you learn<mask> your body needs and doesn't need, hunger just becomes a sensation. Like being a little too hot or a little too cold. You're uncomfortable, yeah.<mask> you're not at risk of dying instantly. [NEWLINE] [NEWLINE] The best thing to do for your metabolism is to eat many small meals throughout today. [NEWLINE] [NEWLINE] Here is my honest to God diet for today. [NEWLINE] [NEWLINE] 0530: Banana [NEWLINE] [NEWLINE] 0700: Bowl of Honey Nut Cheerios with 2% milk and a light drizzle of honey [NEWLINE] [NEWLINE] 1030: Small bowl of light yogurt and granola [NEWLINE] [NEWLINE] 1130: I will have celery and carrot sticks with hummus [NEWLINE] [NEWLINE] 1230: Cottage cheese with a light drizzle of honey [NEWLINE] [NEWLINE] After work, I'm not sure<mask> I'll be eating. I have some ground turkey<mask> leftovers. Maybe I'll eat that. [NEWLINE] [NEWLINE] I've been "hungry"<mask> I woke up.<mask> I know I'm getting enough calories<mask> I don't overeat and stuff my stomach.</s>
Label encoding: <s>I never said to starve yourself. A lot of people have an expectation that they need to be full after you eat. That's not true. Once you learn what your body needs and doesn't need, hunger just becomes a sensation. Like being a little too hot or a little too cold. You're uncomfortable, yeah. But you're not at risk of dying instantly. [NEWLINE] [NEWLINE] The best thing to do for your metabolism is to eat many small meals throughout today. [NEWLINE] [NEWLINE] Here is my honest to God diet for today. [NEWLINE] [NEWLINE] 0530: Banana [NEWLINE] [NEWLINE] 0700: Bowl of Honey Nut Cheerios with 2% milk and a light drizzle of honey [NEWLINE] [NEWLINE] 1030: Small bowl of light yogurt and granola [NEWLINE] [NEWLINE] 1130: I will have celery and carrot sticks with hummus [NEWLINE] [NEWLINE] 1230: Cottage cheese with a light drizzle of honey [NEWLINE] [NEWLINE] After work, I'm not sure what I'll be eating. I have some ground turkey as leftovers. Maybe I'll eat that. [NEWLINE] [NEWLINE] I've been "hungry" since I woke up. But I know I'm getting enough calories so I don't overeat and stuff my stomach.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is due to the fact that they were already in a position to give that much to donate. It's like seeing a kid go through your collection of video games (<mask> you've collected them all from early on up to<mask> you're older); it's not that the figure is amazing or it's the norm, it's just that the people aren't used to seeing it that much. This, I believe, is a direct consequence of being generous, and not always intentional. [ENDQ] [NEWLINE] This makes no sense to me.<mask> I see that someone donates, let's say, $10,000 - that's not<mask> I would think of<mask> an astronomical donation. It's large,<mask> not legendary.<mask>, donating anything close to that sum of money would be absolutely infeasible for me.  It's think it's the same for most people, especially<mask> you're talking about big "million dollar" donations like OP was hinting at.<mask> can people be wowed at that<mask> they were already in a position to donate a million dollars? It's chump change for a billionaire,<mask> more than most people will ever see in their lives. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] This is due to the fact that they were already in a position to give that much to donate. It's like seeing a kid go through your collection of video games ( if you've collected them all from early on up to when you're older); it's not that the figure is amazing or it's the norm, it's just that the people aren't used to seeing it that much. This, I believe, is a direct consequence of being generous, and not always intentional. [ENDQ] [NEWLINE] This makes no sense to me. When I see that someone donates, let's say, $10,000 - that's not what I would think of as an astronomical donation. It's large, but not legendary. However, donating anything close to that sum of money would be absolutely infeasible for me.  It's think it's the same for most people, especially when you're talking about big "million dollar" donations like OP was hinting at. How can people be wowed at that because they were already in a position to donate a million dollars? It's chump change for a billionaire, but more than most people will ever see in their lives. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I state this<mask><mask><mask> movies in general are made/released in the context of their time and that people who see them decades in the future might not really understand them or their importance/quality. [NEWLINE] [NEWLINE] I can give 2 examples: some years ago I saw Citizen kane for the first time, I understand<mask> a movie being realeased in the time of musicals and westerns it must have been an amazing breath of fresh air,<mask> decades away it's looks to me just<mask> an ok movie,<mask> I refrained from rating<mask> it would be unfair to give it a relativelly slow rating. [NEWLINE] [NEWLINE] Another example was the blues brothers: saw it weeks ago and thought it was mainly boring and childish, after seeing the reviews (and trying to understand the high rating) I realized at it's time with those actors, those cameos it must have been something great, just not from my actual point of view... [NEWLINE] (this could probably be said about most movies of the 80s) [NEWLINE] [NEWLINE] [NEWLINE] <mask> basicly that's it: movies/tv shows should be judged in the context of their time and rating them decades in the future will result in unfair ratings.</s>
Label encoding: <s>I state this because I think movies in general are made/released in the context of their time and that people who see them decades in the future might not really understand them or their importance/quality. [NEWLINE] [NEWLINE] I can give 2 examples: some years ago I saw Citizen kane for the first time, I understand as a movie being realeased in the time of musicals and westerns it must have been an amazing breath of fresh air, but decades away it's looks to me just as an ok movie, so I refrained from rating as it would be unfair to give it a relativelly slow rating. [NEWLINE] [NEWLINE] Another example was the blues brothers: saw it weeks ago and thought it was mainly boring and childish, after seeing the reviews (and trying to understand the high rating) I realized at it's time with those actors, those cameos it must have been something great, just not from my actual point of view... [NEWLINE] (this could probably be said about most movies of the 80s) [NEWLINE] [NEWLINE] [NEWLINE] SO basicly that's it: movies/tv shows should be judged in the context of their time and rating them decades in the future will result in unfair ratings.</s>
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not saying that earning more than my limit should be made illegal, or taxed, or punished. I just think there should be more social stigma towards and less admiration for, these people. [NEWLINE] [NEWLINE] My point is, there isn't any extra good that comes<mask> those people have a shit ton of money. [NEWLINE] [NEWLINE] I<mask> don't think that limiting<mask> much we build and achieve would neccessarily be a bad thing.  Remember the matrix?  "Every mammal on this planet instinctively develops a natural equilibrium with the surrounding environment<mask> you humans do not. You move to an area and you multiply and multiply until every natural resource is consumed and the only way you can survive is to spread to another area. There is another organism on this planet that follows the same pattern. Do you know<mask> it is? A virus." [NEWLINE] [NEWLINE] <mask><mask> do we stop? [NEWLINE] [NEWLINE] The space program is funded on some tiny fraction of the federal budget. The majority of people who start working at Nasa are G8,<mask> make ~75-100k/yr.<mask> I don't see<mask> limiting millionares would prevent people from reaching space, or doing other cool stuff. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I'm not saying that earning more than my limit should be made illegal, or taxed, or punished. I just think there should be more social stigma towards and less admiration for, these people. [NEWLINE] [NEWLINE] My point is, there isn't any extra good that comes when those people have a shit ton of money. [NEWLINE] [NEWLINE] I also don't think that limiting how much we build and achieve would neccessarily be a bad thing.  Remember the matrix?  "Every mammal on this planet instinctively develops a natural equilibrium with the surrounding environment but you humans do not. You move to an area and you multiply and multiply until every natural resource is consumed and the only way you can survive is to spread to another area. There is another organism on this planet that follows the same pattern. Do you know what it is? A virus." [NEWLINE] [NEWLINE] So how do we stop? [NEWLINE] [NEWLINE] The space program is funded on some tiny fraction of the federal budget. The majority of people who start working at Nasa are G8, so make ~75-100k/yr. So I don't see how limiting millionares would prevent people from reaching space, or doing other cool stuff. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, in regards to your second point, at first it just seemed unfair to me that on one side you've got a company that hired developers to make the software<mask> this is their full-time job and depend on its success to get paid; and on the other side you've got someone or a group of people that<mask> a hobby create a completely free product without any costs attached. They have their full-time job they depend on too,<mask> to imagine someone comes and does it for free, they would probably lose it. And its different from another competitor coming with a product with lower price<mask> its basically hobbyists that don't depend on it to survive, its like they aren't even playing the same game. [NEWLINE] [NEWLINE] <mask> yeah, obviously companies need to adapt to the market and adjust their offerings to keep their advantage. And<mask> they can't handle a free alternative, then probably their software isn't much worth at all. That's<mask> we see<mask> much of SaaS, Support,... nowadays, and it makes sense they open source at least some of their code,<mask> in the end everyone benefits from it and shared knowledge is<mask> a society should be all about.</s>
Label encoding: <s>Well, in regards to your second point, at first it just seemed unfair to me that on one side you've got a company that hired developers to make the software where this is their full-time job and depend on its success to get paid; and on the other side you've got someone or a group of people that as a hobby create a completely free product without any costs attached. They have their full-time job they depend on too, so to imagine someone comes and does it for free, they would probably lose it. And its different from another competitor coming with a product with lower price because its basically hobbyists that don't depend on it to survive, its like they aren't even playing the same game. [NEWLINE] [NEWLINE] But yeah, obviously companies need to adapt to the market and adjust their offerings to keep their advantage. And if they can't handle a free alternative, then probably their software isn't much worth at all. That's why we see so much of SaaS, Support,... nowadays, and it makes sense they open source at least some of their code, because in the end everyone benefits from it and shared knowledge is what a society should be all about.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You mention that young people use this phrase to undermine the authority of their elders<mask><mask><mask> good reason.<mask>, it's worth noting the phrase primarily serves<mask> a counterpoint to "respect your elders".<mask> the phrase "respect your elders" is used by an older person to convince a younger person to listen to them, it is typically used<mask> a logically equivalent statement to "<mask> I said<mask> ". This is a classic begging the question fallacy, and young people are right to be skeptical of anyone using statements like these.<mask> someone has to resort to such empty arguments to justify<mask> they are saying, then I'd say that's more than enough reason to challenge them.<mask> an older person actually has a reason for<mask> they are telling a younger person, they should just say that reason instead.<mask> they don't, then they shouldn't be saying it in the first place. Young people have stronger critical thinking skills than they are given credit for, and they are more likely to listen to authority figures that give them reason to listen to<mask> they say than simply demand respect. The latter is met with instant distrust<mask> it implies that they have no reason for<mask> they are saying.</s>
Label encoding: <s>You mention that young people use this phrase to undermine the authority of their elders in spite of good reason. However, it's worth noting the phrase primarily serves as a counterpoint to "respect your elders". When the phrase "respect your elders" is used by an older person to convince a younger person to listen to them, it is typically used as a logically equivalent statement to " because I said so ". This is a classic begging the question fallacy, and young people are right to be skeptical of anyone using statements like these. If someone has to resort to such empty arguments to justify what they are saying, then I'd say that's more than enough reason to challenge them. If an older person actually has a reason for what they are telling a younger person, they should just say that reason instead. If they don't, then they shouldn't be saying it in the first place. Young people have stronger critical thinking skills than they are given credit for, and they are more likely to listen to authority figures that give them reason to listen to what they say than simply demand respect. The latter is met with instant distrust because it implies that they have no reason for what they are saying.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0668, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The average wage of a US citizen in most markets is above $20 /hr. [NEWLINE] [NEWLINE] [(Source)]( [URL].release/empsit.t19.htm) [NEWLINE] [NEWLINE] The average comic book provides ~15 minutes of entertainment,<mask><mask> it's being enjoyed and only assuming one or two reads. [NEWLINE] [NEWLINE] Considering that money is essentially just a medium that we use to value our time, and that for an average citizen they make ~$5 in 15 minutes of work, (low end of average, $20/hr, /4 (for 15-minute increments) I would say that they're priced just about right. [NEWLINE] [NEWLINE] There may be an argument that the average consumer of comic books is on the lower end of the earning scale, (which I'm not claiming,<mask> it would be interesting to see data to that point) which would reduce the cost/enjoyment ratio possibly enough to put it to a negative overall,<mask> this is<mask> to be proven. [NEWLINE] [NEWLINE] From just crunching the numbers, it appears<mask><mask> the price is set proportionally to both affordability, and the higher end of<mask> the market will bear. [NEWLINE] </s>
Label encoding: <s>The average wage of a US citizen in most markets is above $20 /hr. [NEWLINE] [NEWLINE] [(Source)]( [URL].release/empsit.t19.htm) [NEWLINE] [NEWLINE] The average comic book provides ~15 minutes of entertainment, assuming that it's being enjoyed and only assuming one or two reads. [NEWLINE] [NEWLINE] Considering that money is essentially just a medium that we use to value our time, and that for an average citizen they make ~$5 in 15 minutes of work, (low end of average, $20/hr, /4 (for 15-minute increments) I would say that they're priced just about right. [NEWLINE] [NEWLINE] There may be an argument that the average consumer of comic books is on the lower end of the earning scale, (which I'm not claiming, but it would be interesting to see data to that point) which would reduce the cost/enjoyment ratio possibly enough to put it to a negative overall, but this is yet to be proven. [NEWLINE] [NEWLINE] From just crunching the numbers, it appears as though the price is set proportionally to both affordability, and the higher end of what the market will bear. [NEWLINE] </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask> they don't make them to try to change peoples beliefs. [ENDQ] [NEWLINE] <mask> do you draw the line?  I imagine you can manipulate children's beliefs not just through telling them things in absolutes.  You could probably predispose a child to religious belief<mask> you get them to take part in ceremonies akin to Churchgoing or strictly non-rational thought.  Something like singing/dancing for the sun to come out in a big group, saying wishes with their eyes closed before bed or getting the kids together to listen to an authority figure tell them a moral or life lesson.  These activities aren't Christian in nature<mask> they fit the tone and lend themselves to a religious mentality. [NEWLINE] [NEWLINE] The point here is that you can't raise a child in a vacuum and give them the perfect upbringing to lead to this idea of perfect objectivity.  You can't have reasonable legislation on child indoctrination<mask> it's an extremely nebulous concept.  Pretty much anything could be construed<mask> an attempt to indoctrinate a child. [NEWLINE] [NEWLINE] <mask><mask> that forcing your beliefs on your child is bad<mask> it's<mask> absurd to legislate against it.</s>
Label encoding: <s> [STARTQ] As long as they don't make them to try to change peoples beliefs. [ENDQ] [NEWLINE] Where do you draw the line?  I imagine you can manipulate children's beliefs not just through telling them things in absolutes.  You could probably predispose a child to religious belief if you get them to take part in ceremonies akin to Churchgoing or strictly non-rational thought.  Something like singing/dancing for the sun to come out in a big group, saying wishes with their eyes closed before bed or getting the kids together to listen to an authority figure tell them a moral or life lesson.  These activities aren't Christian in nature but they fit the tone and lend themselves to a religious mentality. [NEWLINE] [NEWLINE] The point here is that you can't raise a child in a vacuum and give them the perfect upbringing to lead to this idea of perfect objectivity.  You can't have reasonable legislation on child indoctrination because it's an extremely nebulous concept.  Pretty much anything could be construed as an attempt to indoctrinate a child. [NEWLINE] [NEWLINE] I agree that forcing your beliefs on your child is bad but it's also absurd to legislate against it.</s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] By leader do you only the president or prime minister should possess this privilege.<mask><mask>, we are dealing with a very small sample size<mask> we have to be willing to accept that there aren't many examples. [ENDQ] [NEWLINE] I would say any elected official acting in an executive capacity.  Not sure<mask> I feel about cabinet ministers in a parliamentary system. [NEWLINE] [NEWLINE] <mask><mask> we have a small sample size to work with for examples. [NEWLINE] [NEWLINE] [STARTQ] Depends on the system in place. In the US the prosecution of presidents would lead to the same party maintaining power.<mask> it wouldn't necessarily have to transfer between adversaries. [ENDQ] [NEWLINE] A one party state is not a democracy.  That's exactly the sort of result I'm worried about. <mask> you can't transition to another party, you don't have democracy.  Power *has* to be able to transfer to an adversary to be meaningful. [NEWLINE] [NEWLINE] [STARTQ] Egypt wasn't exactly stable to begin with. [ENDQ] [NEWLINE] They were stable,<mask> undemocratic.  Their experiment with democracy failed<mask><mask> in part due to the prosecutions of Mubarak and Morsi.  Now they're back to military rule.</s>
Label encoding: <s> [STARTQ] By leader do you only the president or prime minister should possess this privilege. If so, we are dealing with a very small sample size so we have to be willing to accept that there aren't many examples. [ENDQ] [NEWLINE] I would say any elected official acting in an executive capacity.  Not sure how I feel about cabinet ministers in a parliamentary system. [NEWLINE] [NEWLINE] I agree we have a small sample size to work with for examples. [NEWLINE] [NEWLINE] [STARTQ] Depends on the system in place. In the US the prosecution of presidents would lead to the same party maintaining power. So it wouldn't necessarily have to transfer between adversaries. [ENDQ] [NEWLINE] A one party state is not a democracy.  That's exactly the sort of result I'm worried about.  If you can't transition to another party, you don't have democracy.  Power *has* to be able to transfer to an adversary to be meaningful. [NEWLINE] [NEWLINE] [STARTQ] Egypt wasn't exactly stable to begin with. [ENDQ] [NEWLINE] They were stable, but undemocratic.  Their experiment with democracy failed I think in part due to the prosecutions of Mubarak and Morsi.  Now they're back to military rule.</s>
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask><mask> the ultimately superior strategy would simply be to enter the store in question, obtain a can of deodorant or hair spray from a shelving unit, then use a pre-bought lighter to threaten a flamethrowering in order to obtain cash from the register. Not only does this mean you need only the initial capital to obtain a lighter (rather than the relatively costly supersoaker),<mask><mask> is it safer for the user and<mask> the victim - in the worst case scenario one can only inflict 3rd or perhaps 2nd degree burns instead of the rather dangerous use of a flammable liquid. [NEWLINE] [NEWLINE] Deodorant serves the same purpose, has decreased risk for both sides of the transaction, requires less initial investment, and the precompressed nature of the gas means that there is no wind-up/pump-up time. Furthermore, it is far less traceable<mask> any purchase of a supersoaker within a surrounding radius across a short time window would be investigated by the police force. [NEWLINE] [NEWLINE] <mask> your idea is on the right track, there are more optimal ways to execute the same strategy. </s>
Label encoding: <s>I would argue that the ultimately superior strategy would simply be to enter the store in question, obtain a can of deodorant or hair spray from a shelving unit, then use a pre-bought lighter to threaten a flamethrowering in order to obtain cash from the register. Not only does this mean you need only the initial capital to obtain a lighter (rather than the relatively costly supersoaker), but also is it safer for the user and indeed the victim - in the worst case scenario one can only inflict 3rd or perhaps 2nd degree burns instead of the rather dangerous use of a flammable liquid. [NEWLINE] [NEWLINE] Deodorant serves the same purpose, has decreased risk for both sides of the transaction, requires less initial investment, and the precompressed nature of the gas means that there is no wind-up/pump-up time. Furthermore, it is far less traceable since any purchase of a supersoaker within a surrounding radius across a short time window would be investigated by the police force. [NEWLINE] [NEWLINE] While your idea is on the right track, there are more optimal ways to execute the same strategy. </s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are some essentially interests of fairness here. Someone that lives in a property they own is not *responsible* for absurd bubble housing price increases created almost *entirely* by speculators, the speculators are responsible for those. [NEWLINE] [NEWLINE] There's<mask> no real argument that I can think of<mask> to<mask> governments should get a windfall<mask><mask><mask> speculation, especially at the hands of the majority of people not responsible for the bubble. [NEWLINE] [NEWLINE] Does housing prices going up change the amount of services that government needs to provide? Not at all. It changes, to a degree,<mask> *expensive* some of those services are to provide,<mask> cost of living might increase for new public servants moving to the state,<mask> again, the people benefiting from this are largely the speculators that drive up the prices. [NEWLINE] [NEWLINE] Until and unless someone actually *sells* their house, they profit in no way from these increased theoretical prices.<mask>, there's really no fair way to set the "market price" of a property except to sell it. Until that happens, the supposed increased "market value" is entirely a matter of speculation. </s>
Label encoding: <s>There are some essentially interests of fairness here. Someone that lives in a property they own is not *responsible* for absurd bubble housing price increases created almost *entirely* by speculators, the speculators are responsible for those. [NEWLINE] [NEWLINE] There's also no real argument that I can think of as to why governments should get a windfall because of this speculation, especially at the hands of the majority of people not responsible for the bubble. [NEWLINE] [NEWLINE] Does housing prices going up change the amount of services that government needs to provide? Not at all. It changes, to a degree, how *expensive* some of those services are to provide, because cost of living might increase for new public servants moving to the state, but again, the people benefiting from this are largely the speculators that drive up the prices. [NEWLINE] [NEWLINE] Until and unless someone actually *sells* their house, they profit in no way from these increased theoretical prices. Indeed, there's really no fair way to set the "market price" of a property except to sell it. Until that happens, the supposed increased "market value" is entirely a matter of speculation. </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're looking at cars from a completely logical perspective.  Logically, you are correct, manual tranmissions and coupes make absolutely no sense.  You're ignoring the emotional element of driving. Cars were exhilarating once. Part of the "driving experience" was the way they "felt." The manual transmission gave the driver a chance to be part of the machine,<mask><mask> none of the action would be possible without the talented conductor in driver's seat. Coupes triggered emotions by the way they looked. The coupe allowed auto designers the freedom to build cars that captured the imagination.  By removing these "inefficiencies", you're left with something sterile, benign, and essentially an appliance. Efficiency is boring. Sure a GTR is faster than almost everything around a track<mask> most people don't live on race tracks.<mask> you're not turning hot laps, the GTR is essentially a big heavy automatic coupe that is completely indifferent to whoever is sitting in the drivers seat. Drive enough cars and you'll realize that the cars that put the biggest smile on your face are usually the ones that are not pushing the envelope. </s>
Label encoding: <s>You're looking at cars from a completely logical perspective.  Logically, you are correct, manual tranmissions and coupes make absolutely no sense.  You're ignoring the emotional element of driving. Cars were exhilarating once. Part of the "driving experience" was the way they "felt." The manual transmission gave the driver a chance to be part of the machine, as if none of the action would be possible without the talented conductor in driver's seat. Coupes triggered emotions by the way they looked. The coupe allowed auto designers the freedom to build cars that captured the imagination.  By removing these "inefficiencies", you're left with something sterile, benign, and essentially an appliance. Efficiency is boring. Sure a GTR is faster than almost everything around a track but most people don't live on race tracks. When you're not turning hot laps, the GTR is essentially a big heavy automatic coupe that is completely indifferent to whoever is sitting in the drivers seat. Drive enough cars and you'll realize that the cars that put the biggest smile on your face are usually the ones that are not pushing the envelope. </s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you misread<mask> I mean, the 9-5 would not be a job, I was relating the reverse boot camp to a job in the sense that they could live at home and be with their families,<mask> for some unspecified amount of time they would attend this re-integration program, *much like a job*. [NEWLINE] [NEWLINE] And just saying the "vast majority of veterans do not commit suicide,"<mask><mask><mask>, is not constructive. That's like saying, well, the vast majority of kids who are bullied don't go on gun rampages,<mask> we just shouldn't worry about the possibility that they might. [NEWLINE] [NEWLINE] I am curious<mask><mask> you would think making more mental health care programs available might work. Any expansion on that? I've read a few replies already that say that there are already *enough* mental health programs available, people just don't take advantage of them. In my mind, that means that 'forcing' them to take advantage of them is the best way to ensure that everyone who needs help (even<mask> they don't initiatively realize it) will receive it. [NEWLINE] [NEWLINE] Thoughts?</s>
Label encoding: <s>I think you misread what I mean, the 9-5 would not be a job, I was relating the reverse boot camp to a job in the sense that they could live at home and be with their families, but for some unspecified amount of time they would attend this re-integration program, *much like a job*. [NEWLINE] [NEWLINE] And just saying the "vast majority of veterans do not commit suicide," in my opinion, is not constructive. That's like saying, well, the vast majority of kids who are bullied don't go on gun rampages, so we just shouldn't worry about the possibility that they might. [NEWLINE] [NEWLINE] I am curious though how you would think making more mental health care programs available might work. Any expansion on that? I've read a few replies already that say that there are already *enough* mental health programs available, people just don't take advantage of them. In my mind, that means that 'forcing' them to take advantage of them is the best way to ensure that everyone who needs help (even if they don't initiatively realize it) will receive it. [NEWLINE] [NEWLINE] Thoughts?</s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right. [NEWLINE] [NEWLINE] <mask> they forced those who posted there to show facial features, it would be kinda messed up. [NEWLINE] [NEWLINE] <mask> they can't force them to do that, for the simple reason that it's impossible for them to force anyone into posting anything there. [NEWLINE] [NEWLINE] There are probably lots of subreddits dedicated to this category of posts. Any given OP may choose to post to any of these—<mask> they find the conditions in this one too invasive, there is nothing forcing them to reveal all of part of their face. [NEWLINE] [NEWLINE] <mask><mask><mask>, the only people who post to this subreddit are the particular people who are okay with showing all or part of their face. [NEWLINE] [NEWLINE] In general principle, yes, forcing people to show their face in stuff like this is invasive and wrong.<mask><mask> you look at any individual OP, they are not under any pressure to post there at all. At most, the OP in question may prefer to post to that subreddit,<mask> this isn't a strong enough reason. In the end, anyone who posts there is okay with it, and the people who are not post to other subreddits. </s>
Label encoding: <s>You're right. [NEWLINE] [NEWLINE] If they forced those who posted there to show facial features, it would be kinda messed up. [NEWLINE] [NEWLINE] But they can't force them to do that, for the simple reason that it's impossible for them to force anyone into posting anything there. [NEWLINE] [NEWLINE] There are probably lots of subreddits dedicated to this category of posts. Any given OP may choose to post to any of these— if they find the conditions in this one too invasive, there is nothing forcing them to reveal all of part of their face. [NEWLINE] [NEWLINE] As a result, the only people who post to this subreddit are the particular people who are okay with showing all or part of their face. [NEWLINE] [NEWLINE] In general principle, yes, forcing people to show their face in stuff like this is invasive and wrong. But if you look at any individual OP, they are not under any pressure to post there at all. At most, the OP in question may prefer to post to that subreddit, but this isn't a strong enough reason. In the end, anyone who posts there is okay with it, and the people who are not post to other subreddits. </s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] An 8' tall 400 pound behemoth [ENDQ] [NEWLINE] Linemen are the only guys that large on the field. [NEWLINE] [NEWLINE] [STARTQ] <mask> there's very little "skill" involved in it. It's just a big muscly guy with a probable steroid addiction running really fast and possibly using his hands to catch a ball. [ENDQ] [NEWLINE] <mask> this is the case then any body can get together a group of fat guys and win a game,<mask> it doesn't work like that. There is huge amounts of skill in reading defenses or offenses and this is especially true for the line. A good line will make or break a team. [NEWLINE] [NEWLINE] Sure you can catch a ball, now try evading a 6 foot 190 pound guy and catching that ball 30 yards out. [NEWLINE] [NEWLINE] [STARTQ] This is<mask> you fail. [ENDQ] [NEWLINE] No, I'm sorry that you have a juvenile world view and miss out on the intricacies of other sports. Notice<mask> no one is shitting on futbol,<mask> here you are shitting on Americans and our football. Grow up. [NEWLINE] [NEWLINE] And<mask> do baseball players have to do with this at all?</s>
Label encoding: <s> [STARTQ] An 8' tall 400 pound behemoth [ENDQ] [NEWLINE] Linemen are the only guys that large on the field. [NEWLINE] [NEWLINE] [STARTQ] But there's very little "skill" involved in it. It's just a big muscly guy with a probable steroid addiction running really fast and possibly using his hands to catch a ball. [ENDQ] [NEWLINE] If this is the case then any body can get together a group of fat guys and win a game, but it doesn't work like that. There is huge amounts of skill in reading defenses or offenses and this is especially true for the line. A good line will make or break a team. [NEWLINE] [NEWLINE] Sure you can catch a ball, now try evading a 6 foot 190 pound guy and catching that ball 30 yards out. [NEWLINE] [NEWLINE] [STARTQ] This is why you fail. [ENDQ] [NEWLINE] No, I'm sorry that you have a juvenile world view and miss out on the intricacies of other sports. Notice how no one is shitting on futbol, but here you are shitting on Americans and our football. Grow up. [NEWLINE] [NEWLINE] And what do baseball players have to do with this at all?</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0643, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Basically you can(/should be able to) cycle [ENDQ] [NEWLINE] That doesn't work in all climates.<mask> you lived in a very cold or very warm place, you you wouldn't want to cycle. [NEWLINE] [NEWLINE] [STARTQ] or get public transport to shops and places of work. [ENDQ] [NEWLINE] Public transport takes ages to get to the same place you're going to, and it often doesn't take you there directly,<mask> through many intermediate steps. Furthermore, it doesn't take you there<mask> you need to go,<mask> it takes you more or less in the vicinity. [NEWLINE] [NEWLINE] Granted,<mask><mask> you propose were ever to become a reality, the network of public transport would become more extended,<mask> I still doubt that it would be extended enough to provide a viable alternative. [NEWLINE] [NEWLINE] And<mask> should the elderly and people with disabilities do,<mask> the bus drops them three blocks away from<mask> they need to be? [NEWLINE] [NEWLINE] Plus don't forget that not everyone lives in a city, and a bus cannot go to and from a country town every ten minutes. That would greatly limit the possibility of movement of people living in towns and villages.</s>
Label encoding: <s> [STARTQ] Basically you can(/should be able to) cycle [ENDQ] [NEWLINE] That doesn't work in all climates. If you lived in a very cold or very warm place, you you wouldn't want to cycle. [NEWLINE] [NEWLINE] [STARTQ] or get public transport to shops and places of work. [ENDQ] [NEWLINE] Public transport takes ages to get to the same place you're going to, and it often doesn't take you there directly, but through many intermediate steps. Furthermore, it doesn't take you there where you need to go, but it takes you more or less in the vicinity. [NEWLINE] [NEWLINE] Granted, if what you propose were ever to become a reality, the network of public transport would become more extended, but I still doubt that it would be extended enough to provide a viable alternative. [NEWLINE] [NEWLINE] And what should the elderly and people with disabilities do, when the bus drops them three blocks away from where they need to be? [NEWLINE] [NEWLINE] Plus don't forget that not everyone lives in a city, and a bus cannot go to and from a country town every ten minutes. That would greatly limit the possibility of movement of people living in towns and villages.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a serious problem with this explanation. [NEWLINE] [NEWLINE] [STARTQ] It's about a culture that says it abhors it,<mask> failing to change social norms and institutions that actually help to cultivate sexual assault. [ENDQ] [NEWLINE] Compared to<mask> other culture? Have there been studies about other cultures that have existed throughout history, concerning rape statistics? Does this particular First Western World Culture have higher rape statistics than the rest?<mask> not,<mask> is only this called a rape culture? Or is every culture<mask> far called a 'rape culture'? [NEWLINE] [NEWLINE] Has the revision culture suggested by feminism even been tested anywhere? Or are we taking it at some peoples' word without checking<mask> it's any better? [NEWLINE] [NEWLINE] [Remember, there have been similar rhetorics about violence in male-dominated cultures,<mask> it has been indicated that violence follows humanity<mask> a whole, not just men.]( [URL] ;utm_source=io9_facebook&amp;utm_medium=socialflow). [NEWLINE] [NEWLINE] My point is, until an alternative has been tried and tested, we can't be talking about rape culture<mask> there's no alternative to compare it to.</s>
Label encoding: <s>I have a serious problem with this explanation. [NEWLINE] [NEWLINE] [STARTQ] It's about a culture that says it abhors it, while failing to change social norms and institutions that actually help to cultivate sexual assault. [ENDQ] [NEWLINE] Compared to what other culture? Have there been studies about other cultures that have existed throughout history, concerning rape statistics? Does this particular First Western World Culture have higher rape statistics than the rest? If not, why is only this called a rape culture? Or is every culture so far called a 'rape culture'? [NEWLINE] [NEWLINE] Has the revision culture suggested by feminism even been tested anywhere? Or are we taking it at some peoples' word without checking if it's any better? [NEWLINE] [NEWLINE] [Remember, there have been similar rhetorics about violence in male-dominated cultures, but it has been indicated that violence follows humanity as a whole, not just men.]( [URL] ;utm_source=io9_facebook&amp;utm_medium=socialflow). [NEWLINE] [NEWLINE] My point is, until an alternative has been tried and tested, we can't be talking about rape culture because there's no alternative to compare it to.</s>
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are many people that share the same view<mask> you. Sometimes techniques like this remove a portion of the game, and a significant portion of speedrunning is<mask> the entertainment factor for the spectators too. There are<mask> people that find the most entertaining run the one that pulls all the stops to get to the end<mask> quickly<mask> possible. [NEWLINE] The solution that the communities for the various games generally come up with is to have multiple categories. For example: [NEWLINE] In the Legend of Zelda: Ocarina of Time, there were initially just two categories, any% and 100%. (any% being completion of the game,<mask> fast<mask> possible- this meant clearing all dungeons, at the time). Eventually, glitches and exploits were found, and now there are at least 3 big categories: any%, 100%, and MST (completing all the dungeons,<mask> there were exploits found to skip from the first dungeon all the way to the final boss!). [NEWLINE] [NEWLINE] The MST category avoids the exploitative nature of a glitch in the any% run and more of the game is shown (and more skill? Depends on who you ask).</s>
Label encoding: <s>There are many people that share the same view as you. Sometimes techniques like this remove a portion of the game, and a significant portion of speedrunning is also the entertainment factor for the spectators too. There are also people that find the most entertaining run the one that pulls all the stops to get to the end as quickly as possible. [NEWLINE] The solution that the communities for the various games generally come up with is to have multiple categories. For example: [NEWLINE] In the Legend of Zelda: Ocarina of Time, there were initially just two categories, any% and 100%. (any% being completion of the game, as fast as possible- this meant clearing all dungeons, at the time). Eventually, glitches and exploits were found, and now there are at least 3 big categories: any%, 100%, and MST (completing all the dungeons, as there were exploits found to skip from the first dungeon all the way to the final boss!). [NEWLINE] [NEWLINE] The MST category avoids the exploitative nature of a glitch in the any% run and more of the game is shown (and more skill? Depends on who you ask).</s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The practical application of utilitarianism isn't particularly relevant to its status<mask> a universal moral axiom. [ENDQ] [NEWLINE] Sure  it is. In the sense that<mask> it can be shown that it will consistently be unable to find maximum utility in theory, then it can never find maximum utility in practice. [NEWLINE] [NEWLINE] For example - let me construct a scenario<mask> utility to me, utility to others and utility destroyed can be figured out before the action takes place. And let us assume only minimum choices to simplify matters further. Now the halting problem guarantees that the first part is impossible in all cases,<mask> say we only encounter problems which can be solved without running into the halting problem. Now to justify any action, I only need to lie about such a thing having a maximal utility for that action to be moral. [NEWLINE] [NEWLINE] [STARTQ] The crux is that,<mask><mask> the methodology, it must be a means to maximizing utility, should that methodology remain ethical. [ENDQ] [NEWLINE] This is a little bit strange, the phrasing -<mask> we maximise utility, the action must necessarily be moral, no? [NEWLINE] [NEWLINE] Foe example - it is a p</s>
Label encoding: <s> [STARTQ] The practical application of utilitarianism isn't particularly relevant to its status as a universal moral axiom. [ENDQ] [NEWLINE] Sure  it is. In the sense that if it can be shown that it will consistently be unable to find maximum utility in theory, then it can never find maximum utility in practice. [NEWLINE] [NEWLINE] For example - let me construct a scenario where utility to me, utility to others and utility destroyed can be figured out before the action takes place. And let us assume only minimum choices to simplify matters further. Now the halting problem guarantees that the first part is impossible in all cases, but say we only encounter problems which can be solved without running into the halting problem. Now to justify any action, I only need to lie about such a thing having a maximal utility for that action to be moral. [NEWLINE] [NEWLINE] [STARTQ] The crux is that, regardless of the methodology, it must be a means to maximizing utility, should that methodology remain ethical. [ENDQ] [NEWLINE] This is a little bit strange, the phrasing - if we maximise utility, the action must necessarily be moral, no? [NEWLINE] [NEWLINE] Foe example - it is a p</s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm trying to understand your point of view here,<mask> frankly it seems rather ad hoc.  It's<mask><mask> you are saying the actual act itself is not immoral, just the mental decision to act, only<mask> it gets you the conclusion you desire. [NEWLINE] [NEWLINE] By<mask> standard are you using to determine which immoral acts are<mask> immoral acts and which ones are not actually immoral acts at all,<mask> rather 100% only immoral in the mental decision?  From my perspective you appear to have arbitrarily placed cheating in its own special category simply<mask> it allows you to concoct a giant loophole to excuse certain behaviors you wished were not immoral. [NEWLINE] [NEWLINE] You may have a point that<mask> I make a sacred vow to someone, the decision to break that vow is immoral. <mask> I don't see the logic in boldly stating the the actual breaking of the vow itself is not immoral. [NEWLINE] [NEWLINE] It's wrong to hurt other people.  It's wrong to help someone else hurt other people.  The actual act of cheating risks hurting another party greatly. <mask>, helping someone perform the actual act of cheating is<mask> immoral, right?</s>
Label encoding: <s>I'm trying to understand your point of view here, but frankly it seems rather ad hoc.  It's as if you are saying the actual act itself is not immoral, just the mental decision to act, only because it gets you the conclusion you desire. [NEWLINE] [NEWLINE] By what standard are you using to determine which immoral acts are indeed immoral acts and which ones are not actually immoral acts at all, but rather 100% only immoral in the mental decision?  From my perspective you appear to have arbitrarily placed cheating in its own special category simply because it allows you to concoct a giant loophole to excuse certain behaviors you wished were not immoral. [NEWLINE] [NEWLINE] You may have a point that if I make a sacred vow to someone, the decision to break that vow is immoral.  But I don't see the logic in boldly stating the the actual breaking of the vow itself is not immoral. [NEWLINE] [NEWLINE] It's wrong to hurt other people.  It's wrong to help someone else hurt other people.  The actual act of cheating risks hurting another party greatly.  Thus, helping someone perform the actual act of cheating is therefore immoral, right?</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would not agree that not paying taxes would be theft to society<mask> you can't be liable for a contract that is not valid and the social contract is not a valid contract<mask> it is not possible to consent to it.  Other implied contracts can be consented to, such<mask> going in to a restaurant and ordering dinner under the implication you will pay,<mask> you can avoid doing<mask>. <mask> you cannot avoid being born. [NEWLINE] [NEWLINE] There is a second aspect to my point 2 that defines<mask> a valid contract is.  I neglected to put in here<mask> I should have.  It's basically that an opt-out contract is invalid. <mask><mask> you grow up in a neighbourhood<mask> a local car dealer says "every year after age 18 in order to live here you consent to buy one car from me", that is not a valid contract<mask> you did not agree to it.  It is opt-out (invalid) vs. opt-in (valid).  This explains<mask> the person has no obligation to move out of the country at age 18 or whatever, in order to avoid the social contract.</s>
Label encoding: <s>I would not agree that not paying taxes would be theft to society because you can't be liable for a contract that is not valid and the social contract is not a valid contract because it is not possible to consent to it.  Other implied contracts can be consented to, such as going in to a restaurant and ordering dinner under the implication you will pay, because you can avoid doing so.  However you cannot avoid being born. [NEWLINE] [NEWLINE] There is a second aspect to my point 2 that defines what a valid contract is.  I neglected to put in here but I should have.  It's basically that an opt-out contract is invalid.  So if you grow up in a neighbourhood where a local car dealer says "every year after age 18 in order to live here you consent to buy one car from me", that is not a valid contract since you did not agree to it.  It is opt-out (invalid) vs. opt-in (valid).  This explains why the person has no obligation to move out of the country at age 18 or whatever, in order to avoid the social contract.</s>
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Eh, I feel like there's plenty of reason to be butthurt about, say, the treatment of Native Americans, for example. I'd expect many of their people to be doing very well for themselves<mask> America hadn't repeatedly abused its powers and broke multiple treaties in order to grab more land. [NEWLINE] [NEWLINE] Similarly, I feel there's a lot of reason for black Americans to be angry at the white majority,<mask> they wouldn't have such problems with crushing poverty, undereducation, and racism had it not been for the system of slavery that brought them over en masse to America, and the system of government that upheld grossly unfavorable systems like segregation and jim crow to keep them down even after they were freed. [NEWLINE] [NEWLINE] I feel like these problems are all rooted in atrocities and oppression that took place hundreds of years ago, and still have a huge effect on the people in question today.<mask> it's not that hard to see<mask> more recent atrocities would<mask> have a huge effect on people today,<mask> the economic and psychological factors are not quite<mask> obvious among Jewish people<mask> they are among, say, black people.</s>
Label encoding: <s>Eh, I feel like there's plenty of reason to be butthurt about, say, the treatment of Native Americans, for example. I'd expect many of their people to be doing very well for themselves if America hadn't repeatedly abused its powers and broke multiple treaties in order to grab more land. [NEWLINE] [NEWLINE] Similarly, I feel there's a lot of reason for black Americans to be angry at the white majority, since they wouldn't have such problems with crushing poverty, undereducation, and racism had it not been for the system of slavery that brought them over en masse to America, and the system of government that upheld grossly unfavorable systems like segregation and jim crow to keep them down even after they were freed. [NEWLINE] [NEWLINE] I feel like these problems are all rooted in atrocities and oppression that took place hundreds of years ago, and still have a huge effect on the people in question today. So it's not that hard to see why more recent atrocities would also have a huge effect on people today, though the economic and psychological factors are not quite as obvious among Jewish people as they are among, say, black people.</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Where exactly would you like me to gather reliable statistics on<mask> much pot is cultivated peacefully versus non peacefully? [NEWLINE] [NEWLINE] Taking your anecdote at face value, let's say your scenario is 100 percent accurate.  No one involved in your friends extended circle of bud enthusiasts ever committed or were the victim of a violent crime<mask> a consequence of the product they were buying and selling. [NEWLINE] [NEWLINE] <mask> would any of those individuals have done had they been stolen from?<mask> would they have defended their property? <mask> would they have done<mask> caught by law enforcement?  Even<mask> they would have peacefully surrendered, the moral wrong of contributing to their likely incarceration would exist for any of their buyers; I don't feel the need to point out the implications<mask> they had resisted arrest. [NEWLINE] [NEWLINE] Dispensaries are for legal or semi legal environments, which don't exist in OPs circumstances. [NEWLINE] [NEWLINE] No one wants that kind of environment,<mask> it doesn't stop plenty of people.... [NEWLINE] [NEWLINE] Oh, and feel free to explain to me directly<mask> having the view that being a good Christian and engaging in morally questionable activities is contradictory is satirical.</s>
Label encoding: <s>Where exactly would you like me to gather reliable statistics on how much pot is cultivated peacefully versus non peacefully? [NEWLINE] [NEWLINE] Taking your anecdote at face value, let's say your scenario is 100 percent accurate.  No one involved in your friends extended circle of bud enthusiasts ever committed or were the victim of a violent crime as a consequence of the product they were buying and selling. [NEWLINE] [NEWLINE] What would any of those individuals have done had they been stolen from? How would they have defended their property?  What would they have done if caught by law enforcement?  Even if they would have peacefully surrendered, the moral wrong of contributing to their likely incarceration would exist for any of their buyers; I don't feel the need to point out the implications if they had resisted arrest. [NEWLINE] [NEWLINE] Dispensaries are for legal or semi legal environments, which don't exist in OPs circumstances. [NEWLINE] [NEWLINE] No one wants that kind of environment, but it doesn't stop plenty of people.... [NEWLINE] [NEWLINE] Oh, and feel free to explain to me directly why having the view that being a good Christian and engaging in morally questionable activities is contradictory is satirical.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I don't know much about the Idaho Stop law in particular, I know the following is true on<mask> cycles are treated the same<mask> cars. [NEWLINE] [NEWLINE] 1. Predictability. By making cyclists have to follow the same laws<mask> cars, there is no question on<mask> a cyclist is going to behave. This makes both the cyclist and the driver safer,<mask> they always know<mask> to expect. [NEWLINE] [NEWLINE] 2. Judgement of cyclists. The Idaho Stop law relies on the judgement of bike riders to determine<mask> it is or is not safe to go. Not everybody can judge this. And kids who would lack the ability to judge<mask> ride bikes. [NEWLINE] [NEWLINE] <mask>, anecdotal evidence. I don't live in an area with anything like the Idaho Stop law,<mask> cyclists seem to follow it's provisions<mask><mask> doing<mask> is illegal<mask> I live. I have almost gotten run over multiple times in the crosswalk<mask> I had the right of way to cross by a cyclist who thought they should be allowed to go. Idaho Stop laws would only make that situation worse, and allow cyclists to override the right of way for pedestrians. </s>
Label encoding: <s>While I don't know much about the Idaho Stop law in particular, I know the following is true on why cycles are treated the same as cars. [NEWLINE] [NEWLINE] 1. Predictability. By making cyclists have to follow the same laws as cars, there is no question on how a cyclist is going to behave. This makes both the cyclist and the driver safer, as they always know what to expect. [NEWLINE] [NEWLINE] 2. Judgement of cyclists. The Idaho Stop law relies on the judgement of bike riders to determine when it is or is not safe to go. Not everybody can judge this. And kids who would lack the ability to judge also ride bikes. [NEWLINE] [NEWLINE] Also, anecdotal evidence. I don't live in an area with anything like the Idaho Stop law, but cyclists seem to follow it's provisions even though doing so is illegal where I live. I have almost gotten run over multiple times in the crosswalk when I had the right of way to cross by a cyclist who thought they should be allowed to go. Idaho Stop laws would only make that situation worse, and allow cyclists to override the right of way for pedestrians. </s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do you think there is any possible way the US could invade Afghanistan<mask> there were 5 million mostly untrained al qaeda guerrilla fighters? [ENDQ] [NEWLINE] That's an interesting point. Invade, yes, of course. Win, no (whatever "win' means anyway.) In a civil war there would no option of declaring yourself the winner and leave the country<mask> the war goes badly. [NEWLINE] [NEWLINE] The Taliban won in a political sense. The war was politically unsustainable in the US. The rules in a civil war would be different and I'd say hard to predict.<mask> the world economy would probably immediately crash hard and who knows<mask> consequences that would have. Other superpowers would not just stand idly by. [NEWLINE] [NEWLINE] The Chinese (covertly) joining forces with the US govt against the uprising...? Seems outlandish.<mask><mask> you think about it in economical terms it makes a lot of sense... [NEWLINE] [NEWLINE] The Russians covertly supporting the rebellion? An ensuing global conflict? [NEWLINE] [NEWLINE] Food for thought. In any case, a modern civil war in the US would not end in any way that we think it will.</s>
Label encoding: <s> [STARTQ] Do you think there is any possible way the US could invade Afghanistan if there were 5 million mostly untrained al qaeda guerrilla fighters? [ENDQ] [NEWLINE] That's an interesting point. Invade, yes, of course. Win, no (whatever "win' means anyway.) In a civil war there would no option of declaring yourself the winner and leave the country if the war goes badly. [NEWLINE] [NEWLINE] The Taliban won in a political sense. The war was politically unsustainable in the US. The rules in a civil war would be different and I'd say hard to predict. Besides the world economy would probably immediately crash hard and who knows what consequences that would have. Other superpowers would not just stand idly by. [NEWLINE] [NEWLINE] The Chinese (covertly) joining forces with the US govt against the uprising...? Seems outlandish. But if you think about it in economical terms it makes a lot of sense... [NEWLINE] [NEWLINE] The Russians covertly supporting the rebellion? An ensuing global conflict? [NEWLINE] [NEWLINE] Food for thought. In any case, a modern civil war in the US would not end in any way that we think it will.</s>
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this has something to do with Americans basically misusing the terms "liberal" and "conservative," politically speaking that is. In reality, American political parties all tend to fall under the radical-Liberal category,<mask><mask> the framework set by a political scientist by the name of Robert Dahl. [NEWLINE] [NEWLINE] Socially, we Americans are spot on with our idea of liberal vs. conservative differentiation,<mask> I don't know that the U.S. has any significant *politically* conservative demographic. Actual political conservatism looks much more like a monarchy<mask> I'm not mistaken (it emphasizes *class*<mask> the means by which a person's political influence is determined).<mask>, radical-Liberalism can be seen in the American obsession with personal liberties and the value of the rational individual<mask> a political actor. [NEWLINE] [NEWLINE] <mask>, it sounds weird to call Romney a liberal,<mask><mask><mask> conventions within the science of politics, he technically is a liberal (<mask> he is socially conservative). I don't know much about McCain's views,<mask> I can't really label him. [NEWLINE] [NEWLINE] edited: a thing and then some other stuff</s>
Label encoding: <s>I think this has something to do with Americans basically misusing the terms "liberal" and "conservative," politically speaking that is. In reality, American political parties all tend to fall under the radical-Liberal category, according to the framework set by a political scientist by the name of Robert Dahl. [NEWLINE] [NEWLINE] Socially, we Americans are spot on with our idea of liberal vs. conservative differentiation, but I don't know that the U.S. has any significant *politically* conservative demographic. Actual political conservatism looks much more like a monarchy if I'm not mistaken (it emphasizes *class* as the means by which a person's political influence is determined). Meanwhile, radical-Liberalism can be seen in the American obsession with personal liberties and the value of the rational individual as a political actor. [NEWLINE] [NEWLINE] Thus, it sounds weird to call Romney a liberal, but according to conventions within the science of politics, he technically is a liberal ( although he is socially conservative). I don't know much about McCain's views, so I can't really label him. [NEWLINE] [NEWLINE] edited: a thing and then some other stuff</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Which bathroom should a transgender or transexual person use, the bathroom for their gender or the bathroom for their sex? [NEWLINE] [NEWLINE] The definitions of each of those words is useful in discussing the subject,<mask><mask> most people have little to no experience talking about sex and gender separately, they get kind of confused<mask> people don't use them interchangeably. Same thing for people coming from Tumblr.<mask> they use the words separately to talk about social issues, it's very confusing<mask> someone uses them interchangeably, especially<mask> the definitions are entirely different and always have been. [NEWLINE] [NEWLINE] I get that the difference is confusing<mask><mask> most people's gender lines up with their sex,<mask> for masculine females and feminine males, it's confusing<mask> people misuse terms colloquially without realizing they're doing<mask>. It seems like they don't actually understand<mask> the words mean. Once the majority of society starts calling the written definition garbage and the separation of gender and sex meaningless, it starts to seem like these people no longer belong or exist in society.<mask> of course they do, that's<mask> Tumblr's for right? Garbage? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Which bathroom should a transgender or transexual person use, the bathroom for their gender or the bathroom for their sex? [NEWLINE] [NEWLINE] The definitions of each of those words is useful in discussing the subject, but since most people have little to no experience talking about sex and gender separately, they get kind of confused when people don't use them interchangeably. Same thing for people coming from Tumblr. Since they use the words separately to talk about social issues, it's very confusing when someone uses them interchangeably, especially when the definitions are entirely different and always have been. [NEWLINE] [NEWLINE] I get that the difference is confusing given that most people's gender lines up with their sex, but for masculine females and feminine males, it's confusing why people misuse terms colloquially without realizing they're doing so. It seems like they don't actually understand what the words mean. Once the majority of society starts calling the written definition garbage and the separation of gender and sex meaningless, it starts to seem like these people no longer belong or exist in society. But of course they do, that's what Tumblr's for right? Garbage? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Reddit has made itself to be a place<mask> people of all colors, creeds and beliefs can come and speak freely [ENDQ] [NEWLINE] problem is, the racist and/or sexist users often deliberately drown out the voices of others or purposefully attempt to steer discussions into a certain direction (copy pasted stormfront 'data', throwaway racist/sexist meme spam accounts etc.). by allowing redpill/greatapes types to congregate on reddit, you're driving away non-bigoted people who'd like to have civil discussions. you're actually getting fewer different opinions and less of a lively debate,<mask> all remotely political threads just devolve into either circlejerks or complete shitstorms. [NEWLINE] [NEWLINE] essentially, by allowing bigotry to flourish, the admins are indirectly limiting the free speech of others. they say too many cooks spoil the broth. well, imagine<mask> some of the cooks insisted on deliberately ruining the broth: you'll always end up with a sub-par result and eventually, you will only have the assholes left. this is<mask>'s going on in numerous default subs and is<mask>'s driving the quality down.</s>
Label encoding: <s> [STARTQ] Reddit has made itself to be a place where people of all colors, creeds and beliefs can come and speak freely [ENDQ] [NEWLINE] problem is, the racist and/or sexist users often deliberately drown out the voices of others or purposefully attempt to steer discussions into a certain direction (copy pasted stormfront 'data', throwaway racist/sexist meme spam accounts etc.). by allowing redpill/greatapes types to congregate on reddit, you're driving away non-bigoted people who'd like to have civil discussions. you're actually getting fewer different opinions and less of a lively debate, as all remotely political threads just devolve into either circlejerks or complete shitstorms. [NEWLINE] [NEWLINE] essentially, by allowing bigotry to flourish, the admins are indirectly limiting the free speech of others. they say too many cooks spoil the broth. well, imagine if some of the cooks insisted on deliberately ruining the broth: you'll always end up with a sub-par result and eventually, you will only have the assholes left. this is what's going on in numerous default subs and is what's driving the quality down.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you head over to /r/apple now, you'll see a lot of philosophical hand waving about<mask> the bigger screens are actually really great and<mask> the watch is going to change the world. [ENDQ] [NEWLINE] the thing is that these are the same people who dissed Android for having bigger screens *earlier*. The key word "earlier" leads us to another angle - Apple's hype machine can make things (smart watch, bigger screens, NFC, OIS) appear<mask> groundbreaking<mask> in *fact* the technology has already long existed - usually on Android. [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] going back to OP... [NEWLINE] [NEWLINE] [STARTQ] <mask> after this initial wave of people, a second wave of often bigoted teenagers (and adults, more often than not) espouse statements like "Apple fanboy alert" or "ipoop mainstream and overpriced", etc, you know<mask> I mean [ENDQ] [NEWLINE] there are several people on a certain side, some more articulate, some more informed, some less open-minded. this broad spectrum can make the debate appear intellectual or circlejerky depending on which arguments get your attention.</s>
Label encoding: <s> [STARTQ] If you head over to /r/apple now, you'll see a lot of philosophical hand waving about why the bigger screens are actually really great and why the watch is going to change the world. [ENDQ] [NEWLINE] the thing is that these are the same people who dissed Android for having bigger screens *earlier*. The key word "earlier" leads us to another angle - Apple's hype machine can make things (smart watch, bigger screens, NFC, OIS) appear as groundbreaking when in *fact* the technology has already long existed - usually on Android. [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] going back to OP... [NEWLINE] [NEWLINE] [STARTQ] But after this initial wave of people, a second wave of often bigoted teenagers (and adults, more often than not) espouse statements like "Apple fanboy alert" or "ipoop mainstream and overpriced", etc, you know what I mean [ENDQ] [NEWLINE] there are several people on a certain side, some more articulate, some more informed, some less open-minded. this broad spectrum can make the debate appear intellectual or circlejerky depending on which arguments get your attention.</s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I feel then that OP is way too literally minded. There is nothing wrong with being literally minded, and is probably the only way to answer the question posed. [NEWLINE] [NEWLINE] <mask><mask> only literal things can be understood in this literal way, and there are no explanations that would fit.<mask> you perceive is literal.<mask> your senses can be inaccurate.<mask> you cannot absolutely be certain of anything. I get<mask> he is asking. [NEWLINE] [NEWLINE] I'm thinking the question is silly. Worth asking and discussing and nothing else. Human academic disciplines are models of reality that change based on<mask> we increasingly understand reality. Or, simply, become ever less wrong and more certain. Maybe 100% certainty is impossible. The right question is not '<mask> can I be certain, ever'<mask> '<mask> to become more certain'. That gives far more answers along the way, which is<mask> OP wants. He wants to know things. We can be certain enough to know more things based on<mask> we currently know. We must assume we know something. It might be wrong,<mask> the goal is to become more right,<mask> that should be resolved.</s>
Label encoding: <s>I feel then that OP is way too literally minded. There is nothing wrong with being literally minded, and is probably the only way to answer the question posed. [NEWLINE] [NEWLINE] I think only literal things can be understood in this literal way, and there are no explanations that would fit. What you perceive is literal. Yet your senses can be inaccurate. So you cannot absolutely be certain of anything. I get what he is asking. [NEWLINE] [NEWLINE] I'm thinking the question is silly. Worth asking and discussing and nothing else. Human academic disciplines are models of reality that change based on how we increasingly understand reality. Or, simply, become ever less wrong and more certain. Maybe 100% certainty is impossible. The right question is not'how can I be certain, ever' but'how to become more certain'. That gives far more answers along the way, which is what OP wants. He wants to know things. We can be certain enough to know more things based on what we currently know. We must assume we know something. It might be wrong, yet the goal is to become more right, so that should be resolved.</s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>These scenarios rarely come up. Most people don't even know that arian is a term from India, and<mask> much the Nazis stole from hindu culture (and twisted and corrupted it to their own sick views). [NEWLINE] [NEWLINE] <mask> I don't think it is that important. I'm sure once it was explained to your nieces teachers they understood. I doubt they think the little Indian girl is a Nazi. And once she is older she will understand the western significance.<mask> that doesn't mean it is a necessity to teach everyone. [NEWLINE] [NEWLINE] essentially the western swatzika and the Indian aum are simply two different words that sound similar. I don't think that we need to teach all americans that<mask> Germans say "dick" they mean thick? It is offensive in english to say dick,<mask> it is the german word for thick. [NEWLINE] [NEWLINE] The Indian aum is simply a foreign word that is mistranslated a lot, and<mask> in a western area or speaking a western language you must be careful to use this word.<mask> I don't think it is a pressing issue.</s>
Label encoding: <s>These scenarios rarely come up. Most people don't even know that arian is a term from India, and how much the Nazis stole from hindu culture (and twisted and corrupted it to their own sick views). [NEWLINE] [NEWLINE] But I don't think it is that important. I'm sure once it was explained to your nieces teachers they understood. I doubt they think the little Indian girl is a Nazi. And once she is older she will understand the western significance. But that doesn't mean it is a necessity to teach everyone. [NEWLINE] [NEWLINE] essentially the western swatzika and the Indian aum are simply two different words that sound similar. I don't think that we need to teach all americans that when Germans say "dick" they mean thick? It is offensive in english to say dick, but it is the german word for thick. [NEWLINE] [NEWLINE] The Indian aum is simply a foreign word that is mistranslated a lot, and when in a western area or speaking a western language you must be careful to use this word. But I don't think it is a pressing issue.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Women<mask> are involved in editing and publishing of ridiculous women's magazines like Cosmopolitan and that perpetuate harmful, unachievable beauty ideals, which give millions of girls crippling insecurities about their bodies and lead many of them to eating disorders, depression, and suicide. Women are involved in the production of those media, yes,<mask> that media is still harmful to women. [NEWLINE] [NEWLINE] The same goes for men, by the way, I don't want you to get the impression I'm being completely one-sided. Men were primarily responsible for the production of the "Marlboro Man" advertisements of yesteryear that promoted the idea that in order to be truly masculine, one needs to smoke cigarettes. This idea contributed to thousands or millions of men getting lung cancer over the course of the next two generations. Men promoted that vision of masculinity,<mask> it<mask> was ultimately problematic for men who bought into it. [NEWLINE] [NEWLINE] Both sexes are complicit in the gender-based oppression of both sexes, and there's space for both men and women in the greater feminist (or gender egalitarian<mask> you prefer that label) movement.</s>
Label encoding: <s>Women also are involved in editing and publishing of ridiculous women's magazines like Cosmopolitan and that perpetuate harmful, unachievable beauty ideals, which give millions of girls crippling insecurities about their bodies and lead many of them to eating disorders, depression, and suicide. Women are involved in the production of those media, yes, but that media is still harmful to women. [NEWLINE] [NEWLINE] The same goes for men, by the way, I don't want you to get the impression I'm being completely one-sided. Men were primarily responsible for the production of the "Marlboro Man" advertisements of yesteryear that promoted the idea that in order to be truly masculine, one needs to smoke cigarettes. This idea contributed to thousands or millions of men getting lung cancer over the course of the next two generations. Men promoted that vision of masculinity, but it also was ultimately problematic for men who bought into it. [NEWLINE] [NEWLINE] Both sexes are complicit in the gender-based oppression of both sexes, and there's space for both men and women in the greater feminist (or gender egalitarian if you prefer that label) movement.</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually, they are being mocked<mask> of their reactions to rejection and their own advances. I<mask> sometimes see posts on that subreddit about girls who have those reactions, too. I feel like the reason there aren't<mask> many female posts is simply<mask> females have those adverse reactions much less than males do. [NEWLINE] [NEWLINE] <mask><mask>, perhaps sometimes the sub goes too far and is mean to somebody who may not deserve it,<mask><mask><mask> they are mainly focusing on people who possibly don't take rejection well or blame others for their mistakes instead of owning up to their own mistakes. [NEWLINE] [NEWLINE] For example, many of the posts on that sub have men complaining that they were put in the friendzone. The real reason, perhaps, that they were not rewarded for their advances is<mask> they are not particularly attractive to the person they were pursuing, and that person has every right to turn them down. Everyone is allowed to have their own standards for who they want to be with. They then, instead of thinking about<mask> they could possibly better themselves, blame the girl and the "friendzone" for their misfortunes. </s>
Label encoding: <s>Actually, they are being mocked because of their reactions to rejection and their own advances. I also sometimes see posts on that subreddit about girls who have those reactions, too. I feel like the reason there aren't as many female posts is simply because females have those adverse reactions much less than males do. [NEWLINE] [NEWLINE] I agree, perhaps sometimes the sub goes too far and is mean to somebody who may not deserve it, but I think they are mainly focusing on people who possibly don't take rejection well or blame others for their mistakes instead of owning up to their own mistakes. [NEWLINE] [NEWLINE] For example, many of the posts on that sub have men complaining that they were put in the friendzone. The real reason, perhaps, that they were not rewarded for their advances is because they are not particularly attractive to the person they were pursuing, and that person has every right to turn them down. Everyone is allowed to have their own standards for who they want to be with. They then, instead of thinking about how they could possibly better themselves, blame the girl and the "friendzone" for their misfortunes. </s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think "queer" makes clear that hate-speach can, contrary to /u/stumblepretty, can be reappropriated.  Now, that may not be exactly<mask> OP was talking about (I'm not sure<mask><mask> with OP's point fully), I just think it is clear that we can reappropriate hate-speech, that a word isn't forever hate-speech simply<mask> it has been in the past. [NEWLINE] [NEWLINE] That said, I'm not sure<mask><mask> with you on the history of<mask> the word changed.  It went from being a thing that was yelled at people<mask> they were getting bashed, to a thing that was being chanted at Pride marches and a thing that was yelled at people<mask> they were getting bashed.  It reatined the neagitvie uses, it's just that the reclaimed descriptive use has overtaken it.  (It,<mask><mask>, retains the negative uses to this day, it's still used in hate-speech, just not exclusively<mask>, and<mask><mask><mask> it has lost some of its power) [NEWLINE] </s>
Label encoding: <s>I think "queer" makes clear that hate-speach can, contrary to /u/stumblepretty, can be reappropriated.  Now, that may not be exactly what OP was talking about (I'm not sure I agree with OP's point fully), I just think it is clear that we can reappropriate hate-speech, that a word isn't forever hate-speech simply because it has been in the past. [NEWLINE] [NEWLINE] That said, I'm not sure I agree with you on the history of how the word changed.  It went from being a thing that was yelled at people while they were getting bashed, to a thing that was being chanted at Pride marches and a thing that was yelled at people while they were getting bashed.  It reatined the neagitvie uses, it's just that the reclaimed descriptive use has overtaken it.  (It, in fact, retains the negative uses to this day, it's still used in hate-speech, just not exclusively so, and as a result it has lost some of its power) [NEWLINE] </s>
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not interested in trying to explain that the existence of a state doesn't change the fact that *social organization does and has occurred without the implicit threat of violence*. From the view of many people at Occupy, the police *were* the same thing<mask> Merrill Lynch paid thugs-- they are defending a system which we were calling into question. [NEWLINE] [NEWLINE] The rule of law is a fiction-- just like debt, just like the state, just like property relations. It only "exists"<mask><mask> there is a class of enforcers willing to carry out orders from those who insist they speak from a position of "rule of law". [NEWLINE] [NEWLINE] [STARTQ] I sincerely believe that the notion that human beings are inherently or naturally peaceful flies in the face of all evidence, experience and the historical record. [ENDQ] [NEWLINE] And I agreed with you-- naturalistic appeals aren't very good for making arguments in the first place.<mask> my point was that they are a social organization formed without hierarchy or the implicit threat of an authority, and your point about them not having 'ever existed' is based on playing word games.</s>
Label encoding: <s>I'm not interested in trying to explain that the existence of a state doesn't change the fact that *social organization does and has occurred without the implicit threat of violence*. From the view of many people at Occupy, the police *were* the same thing as Merrill Lynch paid thugs-- they are defending a system which we were calling into question. [NEWLINE] [NEWLINE] The rule of law is a fiction-- just like debt, just like the state, just like property relations. It only "exists" insofar as there is a class of enforcers willing to carry out orders from those who insist they speak from a position of "rule of law". [NEWLINE] [NEWLINE] [STARTQ] I sincerely believe that the notion that human beings are inherently or naturally peaceful flies in the face of all evidence, experience and the historical record. [ENDQ] [NEWLINE] And I agreed with you-- naturalistic appeals aren't very good for making arguments in the first place. But my point was that they are a social organization formed without hierarchy or the implicit threat of an authority, and your point about them not having 'ever existed' is based on playing word games.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>|lots of small businesses don't thrive<mask> much<mask> scrape by [NEWLINE] This I don't quite understand in a capitalist system. It seems to me like we do<mask> much to coddle small businesses due to potential political costs<mask> the vast majority of them bring little new to the table and show little to no chances for growth. I realize that it is stressful and damaging to have businesses go under and to decrease diversity in any market,<mask> shouldn't we let some of these business die and let more profitable ones take their place<mask> we pretend to be a primarily capitalist society? [NEWLINE] [NEWLINE] |an artificial cost [NEWLINE] <mask><mask> with your characterization of this<mask> artificial. I believe that it is a fundamental necessity to capitalism to make the next generation more fair, and that the system cannot survive for long without it. The fact that it would be new to lower the exemption does not mean that the stresses that go along with it should force us to abandon the idea. [NEWLINE] [NEWLINE] <mask><mask> that liquidation is probably something best avoided in most cases, and kitchen nightmares is a bad example of well run inheritances. [NEWLINE] </s>
Label encoding: <s>|lots of small businesses don't thrive so much as scrape by [NEWLINE] This I don't quite understand in a capitalist system. It seems to me like we do so much to coddle small businesses due to potential political costs when the vast majority of them bring little new to the table and show little to no chances for growth. I realize that it is stressful and damaging to have businesses go under and to decrease diversity in any market, but shouldn't we let some of these business die and let more profitable ones take their place if we pretend to be a primarily capitalist society? [NEWLINE] [NEWLINE] |an artificial cost [NEWLINE] I disagree with your characterization of this as artificial. I believe that it is a fundamental necessity to capitalism to make the next generation more fair, and that the system cannot survive for long without it. The fact that it would be new to lower the exemption does not mean that the stresses that go along with it should force us to abandon the idea. [NEWLINE] [NEWLINE] I agree that liquidation is probably something best avoided in most cases, and kitchen nightmares is a bad example of well run inheritances. [NEWLINE] </s>
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Anyone who gives anything is being generous.  Period. [NEWLINE] [NEWLINE] I am in a leadership role for a very successful up and coming charity program called Extra-Life, which raises money for sick kids.  Every donation... ANY donation, is greatly appreciated.  Even being in the role that I am, I can honestly say that I don't believe that the unfortunate have the right to expect help from others, rich or otherwise.  Anyone who gives anything is doing it out of the kindness of their heart (and sometimes for tax breaks) and making the claim that anyone who is giving should be giving more is frankly appalling. [NEWLINE] [NEWLINE] There is nothing about generousity that claims that those people helping others need to make some sort of sacrifice to meet some sort of arbitrary expectation you've given them. <mask> people were forced to sacrifice something in order to give to others, barely anyone would give anything. [NEWLINE] [NEWLINE] <mask><mask> it's great that you obviously have a passion to help the less fortunate,<mask> you're a bit misguided<mask> it comes to those who choose (or choose not) to give.</s>
Label encoding: <s>Anyone who gives anything is being generous.  Period. [NEWLINE] [NEWLINE] I am in a leadership role for a very successful up and coming charity program called Extra-Life, which raises money for sick kids.  Every donation... ANY donation, is greatly appreciated.  Even being in the role that I am, I can honestly say that I don't believe that the unfortunate have the right to expect help from others, rich or otherwise.  Anyone who gives anything is doing it out of the kindness of their heart (and sometimes for tax breaks) and making the claim that anyone who is giving should be giving more is frankly appalling. [NEWLINE] [NEWLINE] There is nothing about generousity that claims that those people helping others need to make some sort of sacrifice to meet some sort of arbitrary expectation you've given them.  If people were forced to sacrifice something in order to give to others, barely anyone would give anything. [NEWLINE] [NEWLINE] I think it's great that you obviously have a passion to help the less fortunate, but you're a bit misguided when it comes to those who choose (or choose not) to give.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Domestic Sheep exist for our use, they are no longer wild creatures and would not exist<mask> we did not use them, we create them and destroy them.  To not use them is to end them completely.  We have evolved into this symbiotic relationship. [NEWLINE] [NEWLINE] All Sheep must die eventually by our hand or otherwise,<mask> we can make their living better in respect to their status<mask> living creatures. [NEWLINE] [NEWLINE] Part of that moral concept is not using them for pleasure<mask> they are alive be it bestiality or torture for sport.  A surprise bullet to the head causes no pain,<mask><mask> it ends the life prematurely, it otherwise respects the emotional state of the animal to be itself. [NEWLINE] [NEWLINE] There is a moral difference between harvesting the dead sheep<mask> a resource, and using the live sheep<mask> a plaything,<mask> there is no pain in death, only a violation of the right to live until nature ends you, which in this framework, we do not grant to sheep<mask> a part of the pact we have to allowing them to exist and eat the food we provide them.</s>
Label encoding: <s>Domestic Sheep exist for our use, they are no longer wild creatures and would not exist if we did not use them, we create them and destroy them.  To not use them is to end them completely.  We have evolved into this symbiotic relationship. [NEWLINE] [NEWLINE] All Sheep must die eventually by our hand or otherwise, but we can make their living better in respect to their status as living creatures. [NEWLINE] [NEWLINE] Part of that moral concept is not using them for pleasure while they are alive be it bestiality or torture for sport.  A surprise bullet to the head causes no pain, so while it ends the life prematurely, it otherwise respects the emotional state of the animal to be itself. [NEWLINE] [NEWLINE] There is a moral difference between harvesting the dead sheep as a resource, and using the live sheep as a plaything, because there is no pain in death, only a violation of the right to live until nature ends you, which in this framework, we do not grant to sheep as a part of the pact we have to allowing them to exist and eat the food we provide them.</s>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're operating under the premise that democracy isn't an extreme and violent ideology,<mask> that's exactly<mask> it becomes<mask> the minority refuses to accept the rule of the majority.  Let's explore a few scenarios: [NEWLINE] [NEWLINE] 1. An individual rejects restrictions on his speech and continues to project his message.<mask> will your system do? Put him in a cage? Kill him? Rough him up? That sounds quite violent and extreme to me. [NEWLINE] [NEWLINE] 2. An individual rejects restrictions on the drugs he can use.<mask> will the system do? Put him in a cage? Kill him? Rough him up? Again, violent and extreme. [NEWLINE] [NEWLINE] 3. Let's say he no longer wants any of your rules<mask> he thinks they're violent, extreme, and insane. He peacefully rejects any of your system's commands and lives peacefully under his own.<mask> will your system do? [NEWLINE] [NEWLINE] <mask> you can convince me that your system operates in a nonviolent and non-extreme manner, you might have a point. Until then, you have the exact same violent and extreme views you despise.</s>
Label encoding: <s>You're operating under the premise that democracy isn't an extreme and violent ideology, yet that's exactly what it becomes when the minority refuses to accept the rule of the majority.  Let's explore a few scenarios: [NEWLINE] [NEWLINE] 1. An individual rejects restrictions on his speech and continues to project his message. What will your system do? Put him in a cage? Kill him? Rough him up? That sounds quite violent and extreme to me. [NEWLINE] [NEWLINE] 2. An individual rejects restrictions on the drugs he can use. What will the system do? Put him in a cage? Kill him? Rough him up? Again, violent and extreme. [NEWLINE] [NEWLINE] 3. Let's say he no longer wants any of your rules because he thinks they're violent, extreme, and insane. He peacefully rejects any of your system's commands and lives peacefully under his own. What will your system do? [NEWLINE] [NEWLINE] If you can convince me that your system operates in a nonviolent and non-extreme manner, you might have a point. Until then, you have the exact same violent and extreme views you despise.</s>
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not arguing that people who are in dire economic situations don't (or shouldn't) turn to circumventing the law<mask> it's a matter of life and death, I'm saying that there are millions of poor people who survive and do *not* poach, just<mask> there are those who do not murder, rape, or kidnap. [NEWLINE] [NEWLINE] <mask><mask> that it comes down to a question of the level of morality you're willing to drop yourself to. Poverty can (to a certain extent) explain someone else's unethical behavior,<mask> it cannot legitimize it, in Africa or in America. [NEWLINE] [NEWLINE] Furthermore, it seems strange that one cannot be simultaneously outraged over human trafficking, poaching, *and* factory farming. All of them seem morally dubious at best and reprehensible at the worst, and they all lead to a poorer state of human affairs. [NEWLINE] [NEWLINE] <mask> you're referring to the Reddit community's blindsighted attitude towards certain issues<mask> it happens to affect their personal lifestyle (such<mask> eating factory farmed meat), well, yeah, that's pretty well known. </s>
Label encoding: <s>I'm not arguing that people who are in dire economic situations don't (or shouldn't) turn to circumventing the law if it's a matter of life and death, I'm saying that there are millions of poor people who survive and do *not* poach, just as there are those who do not murder, rape, or kidnap. [NEWLINE] [NEWLINE] I think that it comes down to a question of the level of morality you're willing to drop yourself to. Poverty can (to a certain extent) explain someone else's unethical behavior, but it cannot legitimize it, in Africa or in America. [NEWLINE] [NEWLINE] Furthermore, it seems strange that one cannot be simultaneously outraged over human trafficking, poaching, *and* factory farming. All of them seem morally dubious at best and reprehensible at the worst, and they all lead to a poorer state of human affairs. [NEWLINE] [NEWLINE] If you're referring to the Reddit community's blindsighted attitude towards certain issues because it happens to affect their personal lifestyle (such as eating factory farmed meat), well, yeah, that's pretty well known. </s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We were all children once. [NEWLINE] [NEWLINE] Some of us were raised in wasteful ways that encouraged us to grow up living wastefully<mask> well. Many of us did<mask>. [NEWLINE] [NEWLINE] Other children were raised by parents who *modeled* "self-improvement through learning and improving society and humanity." Many of these children grew up to do likewise. [NEWLINE] [NEWLINE] It sounds like you are not against children, deep down,<mask> rather are against living wastefully.<mask> the solution is: [NEWLINE] [NEWLINE] (1) Self-improvement through learning and improving society and humanity (<mask> you already say); and [NEWLINE] (2) *Teach your children to do likewise.* [NEWLINE] [NEWLINE] <mask> the mature people are the ones not having kids, and the wasteful people are having all the kids, then the future generation will all have been raised to be wasteful. [NEWLINE] [NEWLINE] You want to change the world? Improve yourself (<mask> you already plan to), and raise your future kids to be better than the culture around you. Let them see there's a better way; let them learn the better way from you.</s>
Label encoding: <s>We were all children once. [NEWLINE] [NEWLINE] Some of us were raised in wasteful ways that encouraged us to grow up living wastefully as well. Many of us did so. [NEWLINE] [NEWLINE] Other children were raised by parents who *modeled* "self-improvement through learning and improving society and humanity." Many of these children grew up to do likewise. [NEWLINE] [NEWLINE] It sounds like you are not against children, deep down, but rather are against living wastefully. So the solution is: [NEWLINE] [NEWLINE] (1) Self-improvement through learning and improving society and humanity ( as you already say); and [NEWLINE] (2) *Teach your children to do likewise.* [NEWLINE] [NEWLINE] If the mature people are the ones not having kids, and the wasteful people are having all the kids, then the future generation will all have been raised to be wasteful. [NEWLINE] [NEWLINE] You want to change the world? Improve yourself ( as you already plan to), and raise your future kids to be better than the culture around you. Let them see there's a better way; let them learn the better way from you.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're thinking about this the wrong way. <mask> we start off in our captalism society with 1k people making $10k a year, our total economy is worth $10mil.  Now, a couple of years down the line, one of the guys who's making $10k a year now makes $10mil a year. <mask>, that doesn't mean everyone else is making 0 a year.  Rather, some people make $10k, some people make $20k, and<mask> on.  The total value of our economy has actually grown to $100mil, 10 times more than we started with. [NEWLINE] [NEWLINE] [NEWLINE] This is the basis of capitalism, that<mask> humans are allowed to be greedy and try to make<mask> much money<mask> they can, wealth is literally created.  I would say that your importance to the economy is measured by<mask> much wealth you are creating relative to<mask> much money you make a year,<mask> just<mask> important<mask><mask> much you can grow the economy today is<mask> much you can grow the economy tomorrow.</s>
Label encoding: <s>I think you're thinking about this the wrong way.  If we start off in our captalism society with 1k people making $10k a year, our total economy is worth $10mil.  Now, a couple of years down the line, one of the guys who's making $10k a year now makes $10mil a year.  But, that doesn't mean everyone else is making 0 a year.  Rather, some people make $10k, some people make $20k, and so on.  The total value of our economy has actually grown to $100mil, 10 times more than we started with. [NEWLINE] [NEWLINE] [NEWLINE] This is the basis of capitalism, that when humans are allowed to be greedy and try to make as much money as they can, wealth is literally created.  I would say that your importance to the economy is measured by how much wealth you are creating relative to how much money you make a year, because just as important as how much you can grow the economy today is how much you can grow the economy tomorrow.</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Again: "No" would've been a perfectly acceptable answer. They could've said "No"<mask> I asked<mask> they were still open. They could've said "No"<mask> I told there it would be a large party.<mask> I told them we didn't want a limited menu, and said we were leaving then, they could've said "Have a nice night". They chose to, not only say "Yes",<mask> to do<mask> over and over again. [ENDQ] [NEWLINE] It sounds like the waitstaff wanted you gone,<mask> were required by the manager to seat and serve you.<mask> again, you didn't break any rules,<mask> basic empathy would generally alert you to the fact that the waitstaff didn't want to be there, and would probably be fired<mask> they asked you to leave.<mask> you're putting people in a position that requires them to choose between being inconvenienced or being fired, then you're being a dick. You just are. You're entitled to be a dick,<mask> that doesn't mean there's nothing wrong with it. </s>
Label encoding: <s> [STARTQ] Again: "No" would've been a perfectly acceptable answer. They could've said "No" when I asked if they were still open. They could've said "No" when I told there it would be a large party. When I told them we didn't want a limited menu, and said we were leaving then, they could've said "Have a nice night". They chose to, not only say "Yes", but to do so over and over again. [ENDQ] [NEWLINE] It sounds like the waitstaff wanted you gone, but were required by the manager to seat and serve you. So again, you didn't break any rules, but basic empathy would generally alert you to the fact that the waitstaff didn't want to be there, and would probably be fired if they asked you to leave. If you're putting people in a position that requires them to choose between being inconvenienced or being fired, then you're being a dick. You just are. You're entitled to be a dick, but that doesn't mean there's nothing wrong with it. </s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just of out curiosity, do you watch<mask> your girlfriend eats all day? I know a couple of people who "only eat once a day"<mask> are constantly grazing and snacking. [NEWLINE] [NEWLINE] Of course your girlfriend may have a legitimate medical condition,<mask> those are generally pretty rare. [NEWLINE] [NEWLINE] And eating<mask> she eats doesn't mean that you will gain/lose the same weight.<mask> you're 7% body fat you are possibly partitioning<mask> you eat better than she does, meaning that your body processes the food into muscle rather than fat. I could be slightly off on that<mask><mask> I'm not a nutritionist. Maybe you<mask> fidget more than she does. Or walk more at work. People are going to have different rates of metabolic expenditure and guys are generally going to burn more calories than women. [NEWLINE] [NEWLINE] EDIT: Actually, I'm confused. Are you saying she was 190<mask> you met her? Or she was 130-140 then after some time gained weight up to 190?<mask> the former, then<mask> does that have to do with anything?</s>
Label encoding: <s>Just of out curiosity, do you watch what your girlfriend eats all day? I know a couple of people who "only eat once a day" but are constantly grazing and snacking. [NEWLINE] [NEWLINE] Of course your girlfriend may have a legitimate medical condition, but those are generally pretty rare. [NEWLINE] [NEWLINE] And eating what she eats doesn't mean that you will gain/lose the same weight. If you're 7% body fat you are possibly partitioning what you eat better than she does, meaning that your body processes the food into muscle rather than fat. I could be slightly off on that though as I'm not a nutritionist. Maybe you also fidget more than she does. Or walk more at work. People are going to have different rates of metabolic expenditure and guys are generally going to burn more calories than women. [NEWLINE] [NEWLINE] EDIT: Actually, I'm confused. Are you saying she was 190 when you met her? Or she was 130-140 then after some time gained weight up to 190? If the former, then what does that have to do with anything?</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] it is very damaging to the reputation<mask> a whole [ENDQ] [NEWLINE] It is only damaging<mask> you are bigoted and assign guilt by association. [NEWLINE] [NEWLINE] "Black civil rights leaders like Rev. Al Sharpton should stop advocating for civil rights<mask> people like Louis Farrakhan adopt extremist positions." [NEWLINE] [NEWLINE] This is the conservative frame for all of those who advocate for greater political equality for minority populations. Conservatives smear blacks, gays, women or frankly any minority with extreme examples<mask> they wrongly associate those extremists with the mainline advocates. [NEWLINE] [NEWLINE] That is just<mask> bigotry **is.** One is a bigot<mask> one holds all members of a minority position with it's most extreme members. It's called prejudice and is based on the sweeping generalization fallacy. [NEWLINE] [NEWLINE] Louis Farrakhan is an A. [NEWLINE] [NEWLINE] Louis Farrakhan holds an extreme position. [NEWLINE] [NEWLINE] Al Sharpton is an A [NEWLINE] [NEWLINE] <mask> Al Sharpton must repudiate Louis Farrakhan's extremism. [NEWLINE] [NEWLINE] The argument is invalid. OPs argument is based on this logical fallacy and should<mask> be rejected.</s>
Label encoding: <s> [STARTQ] it is very damaging to the reputation as a whole [ENDQ] [NEWLINE] It is only damaging if you are bigoted and assign guilt by association. [NEWLINE] [NEWLINE] "Black civil rights leaders like Rev. Al Sharpton should stop advocating for civil rights because people like Louis Farrakhan adopt extremist positions." [NEWLINE] [NEWLINE] This is the conservative frame for all of those who advocate for greater political equality for minority populations. Conservatives smear blacks, gays, women or frankly any minority with extreme examples because they wrongly associate those extremists with the mainline advocates. [NEWLINE] [NEWLINE] That is just what bigotry **is.** One is a bigot if one holds all members of a minority position with it's most extreme members. It's called prejudice and is based on the sweeping generalization fallacy. [NEWLINE] [NEWLINE] Louis Farrakhan is an A. [NEWLINE] [NEWLINE] Louis Farrakhan holds an extreme position. [NEWLINE] [NEWLINE] Al Sharpton is an A [NEWLINE] [NEWLINE] Therefore Al Sharpton must repudiate Louis Farrakhan's extremism. [NEWLINE] [NEWLINE] The argument is invalid. OPs argument is based on this logical fallacy and should therefore be rejected.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be fair<mask><mask> all religions should be equally up for jabs. [NEWLINE] [NEWLINE] <mask> a drawing of someone who you don't even know<mask> they look like<mask> just label it his name offends you,  I got no time to give a shit,  minority or not. [NEWLINE] [NEWLINE] <mask> someone takes those jesus is with you always drawings and captions them,  fine by me.   It does not matter.  Being a minority means ABSOLUTELY NOTHING to the protections  you should receive from society (chastising / protest / whatever) UNLESS the protections are protections against UNEQUAL treatment. [NEWLINE] [NEWLINE] Muslims are trashed just like any other and giving them a leg up just<mask> they're a minority<mask> people should be given shit<mask> they make fun of them,  no,  not cool. [NEWLINE] [NEWLINE] We similarly see such inversion with Judaism,  which this Charlie Hebdo illustrator (before he was fired for drawing pictures poking fun at jewish people,  which is a bit of unfortunate irony) [URL] </s>
Label encoding: <s>To be fair I think all religions should be equally up for jabs. [NEWLINE] [NEWLINE] If a drawing of someone who you don't even know what they look like but just label it his name offends you,  I got no time to give a shit,  minority or not. [NEWLINE] [NEWLINE] If someone takes those jesus is with you always drawings and captions them,  fine by me.   It does not matter.  Being a minority means ABSOLUTELY NOTHING to the protections  you should receive from society (chastising / protest / whatever) UNLESS the protections are protections against UNEQUAL treatment. [NEWLINE] [NEWLINE] Muslims are trashed just like any other and giving them a leg up just because they're a minority so people should be given shit if they make fun of them,  no,  not cool. [NEWLINE] [NEWLINE] We similarly see such inversion with Judaism,  which this Charlie Hebdo illustrator (before he was fired for drawing pictures poking fun at jewish people,  which is a bit of unfortunate irony) [URL] </s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you've ever circled a woman's areola with your fingertip, only to return by gently pinching her nipple, then you know unequivocally that a breast *is a sexual part* of the female anatomy.  Fetishes aside, you're not going to find anywhere the near the same arousal in one's shoulder, elbow or shin. <mask> such, the burqa isn't representative of a rational approach to addressing sexuality<mask> living in a social society.  Instead, it is excessive measure that ignores rationality, in favor of an absurd notion that *all* images of skin could be sexualized. [NEWLINE] [NEWLINE] <mask>, to go from that extreme to the other, is equally absurd.  In the middle, we find a rationalization that strikes a balance between pragmatism, vanity and our biological urges to procreate.  That balance moves and shifts, between cultures and eras. <mask> that doesn't mean *all* rationality should summarily be thrown out the window in favor of thoughtless extremes like the burqa or all-nudity.</s>
Label encoding: <s>If you've ever circled a woman's areola with your fingertip, only to return by gently pinching her nipple, then you know unequivocally that a breast *is a sexual part* of the female anatomy.  Fetishes aside, you're not going to find anywhere the near the same arousal in one's shoulder, elbow or shin.  As such, the burqa isn't representative of a rational approach to addressing sexuality while living in a social society.  Instead, it is excessive measure that ignores rationality, in favor of an absurd notion that *all* images of skin could be sexualized. [NEWLINE] [NEWLINE] However, to go from that extreme to the other, is equally absurd.  In the middle, we find a rationalization that strikes a balance between pragmatism, vanity and our biological urges to procreate.  That balance moves and shifts, between cultures and eras.  But that doesn't mean *all* rationality should summarily be thrown out the window in favor of thoughtless extremes like the burqa or all-nudity.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You want to get a job anywhere in the world? Want to easily make a $100,000 a year? Want to have lots of different and interesting specialties to choose from? Be a nurse. [ENDQ] [NEWLINE] Honestly this is exactly the kind of attitude that leads to<mask> much failure and debt in the college system. Children are encouraged to pursue an extremely limited set of professions for mostly financial reasons. Those that manage to take on all the debt and get through the programs for those professions quickly find out they don't actually have aptitude for it, or enjoy it, or can't find a real demand for it. A successful career starts with finding something you're decent at and can stand to do for a long time, not aiming for a pay bracket or type of benefits. [NEWLINE] [NEWLINE] I'm generally in favor of giving children more opportunities to actually discover their interests and aptitudes before going into debt for a degree, and to that end I somewhat agree with your overall idea about more service options,<mask><mask><mask> we both support similar action for entirely different reasons.</s>
Label encoding: <s> [STARTQ] You want to get a job anywhere in the world? Want to easily make a $100,000 a year? Want to have lots of different and interesting specialties to choose from? Be a nurse. [ENDQ] [NEWLINE] Honestly this is exactly the kind of attitude that leads to so much failure and debt in the college system. Children are encouraged to pursue an extremely limited set of professions for mostly financial reasons. Those that manage to take on all the debt and get through the programs for those professions quickly find out they don't actually have aptitude for it, or enjoy it, or can't find a real demand for it. A successful career starts with finding something you're decent at and can stand to do for a long time, not aiming for a pay bracket or type of benefits. [NEWLINE] [NEWLINE] I'm generally in favor of giving children more opportunities to actually discover their interests and aptitudes before going into debt for a degree, and to that end I somewhat agree with your overall idea about more service options, but I think we both support similar action for entirely different reasons.</s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Partisan: There are issues<mask> Congress is sharply divided along party lines. For example, on anything related to rich versus poor (taxes, the minimum wage, social welfare programs like food stamps and social security) the Republicans are unanimously in favor of things that help the rich and against things that help the poor. The Democrats tend to line up exactly opposite to this. Other highly partisan issues are regulation (Republicans oppose it claiming it hurts the economy, Democrats favor it<mask> it protects the environment and keeps the playing field level), economic policies (Republicans favor the Austrian school of Economics,<mask> Democrats favor Kensyianism), the debt (Republicans give lip service to reducing it, except<mask> they are in charge or<mask> it comes to cutting taxes; Democrats are more pragmatic). [ENDQ] [NEWLINE] Bullshit, the Dems talk a good game<mask><mask> it comes down to it they aren't particularly different. [NEWLINE] [NEWLINE] You don't see them supporting the $15 minimum wage campaign do you? [NEWLINE] [NEWLINE] You don't see them challenging the neo-liberal agenda at all. </s>
Label encoding: <s> [STARTQ] Partisan: There are issues where Congress is sharply divided along party lines. For example, on anything related to rich versus poor (taxes, the minimum wage, social welfare programs like food stamps and social security) the Republicans are unanimously in favor of things that help the rich and against things that help the poor. The Democrats tend to line up exactly opposite to this. Other highly partisan issues are regulation (Republicans oppose it claiming it hurts the economy, Democrats favor it because it protects the environment and keeps the playing field level), economic policies (Republicans favor the Austrian school of Economics, while Democrats favor Kensyianism), the debt (Republicans give lip service to reducing it, except when they are in charge or when it comes to cutting taxes; Democrats are more pragmatic). [ENDQ] [NEWLINE] Bullshit, the Dems talk a good game but when it comes down to it they aren't particularly different. [NEWLINE] [NEWLINE] You don't see them supporting the $15 minimum wage campaign do you? [NEWLINE] [NEWLINE] You don't see them challenging the neo-liberal agenda at all. </s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It seems like you think that the manager bending over backwards to accommodate you means that you're behavior is justified. [ENDQ] [NEWLINE] <mask><mask> it all comes down to exactly<mask> the conversations went. [NEWLINE] [NEWLINE] Maybe the party was polite at all times, calmly explaining that they would only eat<mask> the full meny were available,<mask> that they wouldn't be taking offense<mask> they had to leave<mask> of only some quick food being available. <mask> they were offered a full menu and service after that sort of conversation, then I don't see anything morally wrong at all with them staying and enjoying a regular meal. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>,<mask> they were brash and demanding, causing a scene that the manager felt she had to respond to, then they were absolutely in the wrong.  In that case, they were essentially bullying the restaurant into letting them have their own way. [NEWLINE] [NEWLINE] We just can't know who was actually acting like an arse in this situation.  Only the people there know<mask> happened, and they're all going to be biased.</s>
Label encoding: <s> [STARTQ] It seems like you think that the manager bending over backwards to accommodate you means that you're behavior is justified. [ENDQ] [NEWLINE] I think it all comes down to exactly how the conversations went. [NEWLINE] [NEWLINE] Maybe the party was polite at all times, calmly explaining that they would only eat if the full meny were available, but that they wouldn't be taking offense if they had to leave because of only some quick food being available.  If they were offered a full menu and service after that sort of conversation, then I don't see anything morally wrong at all with them staying and enjoying a regular meal. [NEWLINE] [NEWLINE] On the other hand, if they were brash and demanding, causing a scene that the manager felt she had to respond to, then they were absolutely in the wrong.  In that case, they were essentially bullying the restaurant into letting them have their own way. [NEWLINE] [NEWLINE] We just can't know who was actually acting like an arse in this situation.  Only the people there know what happened, and they're all going to be biased.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The fact is that we do decide<mask> is good and<mask> is bad<mask> best we can and make them into laws. People get their ethical and moral views from many sources,<mask> do we only allow those ethical and moral views to trump the law<mask> they come from religion? Even forgetting that religious ethics are generally inferior to modern secular ethics due to their dogmatic and supernatural basis, it is clearly discriminating against ideas based solely on their (dubious) source. [NEWLINE] [NEWLINE] <mask> it stands a PhD in philosophy who spends his life studying ethics, who comes to a moral position based on evidence and reason has no recourse to the same exemptions on moral grounds..... unless he<mask> had a religion that held the same morals. Even the more obviously made up religions like Scientology would have to be granted exemptions<mask><mask> RFRA. Hell<mask> I started my own Church of NuclearFirecracker and claimed a religious exemption is the government in the business of telling me that my religion is not valid?<mask><mask> them I am being given a pretty strong incentive to pretend to be religious.</s>
Label encoding: <s>The fact is that we do decide what is good and what is bad as best we can and make them into laws. People get their ethical and moral views from many sources, why do we only allow those ethical and moral views to trump the law when they come from religion? Even forgetting that religious ethics are generally inferior to modern secular ethics due to their dogmatic and supernatural basis, it is clearly discriminating against ideas based solely on their (dubious) source. [NEWLINE] [NEWLINE] As it stands a PhD in philosophy who spends his life studying ethics, who comes to a moral position based on evidence and reason has no recourse to the same exemptions on moral grounds..... unless he ALSO had a religion that held the same morals. Even the more obviously made up religions like Scientology would have to be granted exemptions according to RFRA. Hell if I started my own Church of NuclearFirecracker and claimed a religious exemption is the government in the business of telling me that my religion is not valid? If so them I am being given a pretty strong incentive to pretend to be religious.</s>
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I hate this sort of reply. Society was born *out of* some of our deepest instincts. Empathy. Need for safety and progress. Humans have many, different, conflicting needs. [NEWLINE] [NEWLINE] To temporarily uphold one need (ie sexual attraction) and then in other instances another (ie empathy) isn't a lie in any way. [NEWLINE] [NEWLINE] The thing wrong with OP's phrasing here is not acknowledging that we have many different needs other than sexual attraction, which he is apparently boiling down to "truth" in this example, whatever that may mean. [NEWLINE] [NEWLINE] [STARTQ] For society to be outraged at James Franco for wanting, and trying, to have sex with this girl is absurd. It's based on a fundamental lie [ENDQ] [NEWLINE] <mask> we take the example to the extreme of 12 years young, there arises a situation<mask> one side of the party can easily exploit the other, which we<mask> bystanders find unacceptable due to our base instinct of empathy.<mask> society condemning it isn't any more a "lie" than any other kind of urge us humans might have.</s>
Label encoding: <s>I hate this sort of reply. Society was born *out of* some of our deepest instincts. Empathy. Need for safety and progress. Humans have many, different, conflicting needs. [NEWLINE] [NEWLINE] To temporarily uphold one need (ie sexual attraction) and then in other instances another (ie empathy) isn't a lie in any way. [NEWLINE] [NEWLINE] The thing wrong with OP's phrasing here is not acknowledging that we have many different needs other than sexual attraction, which he is apparently boiling down to "truth" in this example, whatever that may mean. [NEWLINE] [NEWLINE] [STARTQ] For society to be outraged at James Franco for wanting, and trying, to have sex with this girl is absurd. It's based on a fundamental lie [ENDQ] [NEWLINE] If we take the example to the extreme of 12 years young, there arises a situation where one side of the party can easily exploit the other, which we as bystanders find unacceptable due to our base instinct of empathy. Therefore society condemning it isn't any more a "lie" than any other kind of urge us humans might have.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok -<mask><mask> many new friendships with men are you meeting in this same way? Are you approaching women with the intention of seeing<mask> you can get a date with them, or with the intention of meeting them the same<mask> you would a guy? [NEWLINE] [NEWLINE] Part of the problem may be that people you are meeting are not in the market for a relationship, and<mask> that is the only thing you are interested in (compared with just getting to know them<mask> a person) it is a huge turn off. [NEWLINE] [NEWLINE] I believe this is one of the reasons that online dating is<mask> powerful - you know that everyone on the site is interested in dating,<mask> you have eliminated the #1 reason<mask> most people would not want to pursue anything further.<mask>, many of the sites set up relative matches<mask> not only do they fit within<mask> you are looking for,<mask> you<mask> fit within<mask> they are looking for. This takes a lot of the work out of meeting someone, and you can instead focus on finding someone who you have good chemistry with. </s>
Label encoding: <s>Ok - so how many new friendships with men are you meeting in this same way? Are you approaching women with the intention of seeing if you can get a date with them, or with the intention of meeting them the same as you would a guy? [NEWLINE] [NEWLINE] Part of the problem may be that people you are meeting are not in the market for a relationship, and if that is the only thing you are interested in (compared with just getting to know them as a person) it is a huge turn off. [NEWLINE] [NEWLINE] I believe this is one of the reasons that online dating is so powerful - you know that everyone on the site is interested in dating, so you have eliminated the #1 reason why most people would not want to pursue anything further. Additionally, many of the sites set up relative matches where not only do they fit within what you are looking for, but you also fit within what they are looking for. This takes a lot of the work out of meeting someone, and you can instead focus on finding someone who you have good chemistry with. </s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Rather than just theories and assumptions [ENDQ] [NEWLINE] Are you really telling me that you only know that Australia's spy agency was actually spying on people is<mask> of Snowden? [NEWLINE] [NEWLINE] You know there is information on their website right? The only thing the hide is who they are specifically looking at [NEWLINE] [NEWLINE] [STARTQ] Reveal their secrets, protect our own [ENDQ] [NEWLINE] This is the [slogan of the ASD]( [URL] ) (Australian Signals Directive) [NEWLINE] [NEWLINE] [STARTQ] ASIS's primary goal is to obtain and distribute secret intelligence about the capabilities, intentions and activities of individuals or organisations outside Australia [ENDQ] [NEWLINE] And this is from the 'About Us' page from [ASIS's website]( [URL] ). [NEWLINE] [NEWLINE] Yep, it's totally super secret and not at all bloody obvious. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Nope, he let journalists do the filtering<mask> he himself knew he couldn't be in charge of deciding<mask> needs to be revealed. [ENDQ] [NEWLINE] That's not really much of an improvement, in some respects giving it to uncovered eyes is even more retarded. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Rather than just theories and assumptions [ENDQ] [NEWLINE] Are you really telling me that you only know that Australia's spy agency was actually spying on people is because of Snowden? [NEWLINE] [NEWLINE] You know there is information on their website right? The only thing the hide is who they are specifically looking at [NEWLINE] [NEWLINE] [STARTQ] Reveal their secrets, protect our own [ENDQ] [NEWLINE] This is the [slogan of the ASD]( [URL] ) (Australian Signals Directive) [NEWLINE] [NEWLINE] [STARTQ] ASIS's primary goal is to obtain and distribute secret intelligence about the capabilities, intentions and activities of individuals or organisations outside Australia [ENDQ] [NEWLINE] And this is from the 'About Us' page from [ASIS's website]( [URL] ). [NEWLINE] [NEWLINE] Yep, it's totally super secret and not at all bloody obvious. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Nope, he let journalists do the filtering as he himself knew he couldn't be in charge of deciding what needs to be revealed. [ENDQ] [NEWLINE] That's not really much of an improvement, in some respects giving it to uncovered eyes is even more retarded. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are 2 things you should consider<mask> talking about whether or not marriages should be recognized by the Government [NEWLINE] [NEWLINE] - medical decisions [NEWLINE] [NEWLINE] - inheritance/gifting [NEWLINE] [NEWLINE] I can live with whoever I want, I can have children with any woman, theoretically I don't need to be married.<mask><mask> I'm in an accident and there are important decisions to be made - I want the mother of my children to make those decisions. Marriage helps with that. [NEWLINE] <mask><mask> I live<mask> you give a big financial gift to someone (over 3000usd<mask><mask>?) you have to pay a tax for it, unless you are closely related to the person. Without marriage you wouldn't be able to give money to your spouse without having to pay taxes for it. [NEWLINE] [NEWLINE] Those are purely practical issues. Of course<mask> the government made it possible to write contracts about these things without having to get married, I'd be ok with it.<mask><mask> it stands I feel like marriage in the legal sense is a "package" of practical improvements for your life.</s>
Label encoding: <s>There are 2 things you should consider when talking about whether or not marriages should be recognized by the Government [NEWLINE] [NEWLINE] - medical decisions [NEWLINE] [NEWLINE] - inheritance/gifting [NEWLINE] [NEWLINE] I can live with whoever I want, I can have children with any woman, theoretically I don't need to be married. But if I'm in an accident and there are important decisions to be made - I want the mother of my children to make those decisions. Marriage helps with that. [NEWLINE] Also where I live if you give a big financial gift to someone (over 3000usd I think?) you have to pay a tax for it, unless you are closely related to the person. Without marriage you wouldn't be able to give money to your spouse without having to pay taxes for it. [NEWLINE] [NEWLINE] Those are purely practical issues. Of course if the government made it possible to write contracts about these things without having to get married, I'd be ok with it. But as it stands I feel like marriage in the legal sense is a "package" of practical improvements for your life.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] <mask> of your comment I actually looked up the definition of'respect' and there are basically two definitions: 1) deep admiration for someone's achievements or abilities and 2) due regard for the feelings, wishes, or rights of others. [NEWLINE] [NEWLINE] [NEWLINE] English is not my native language and in Portuguese we don't really use respect in the sense of deep admiration (we just use admiration for that), we consider the 2nd definition for respect. [NEWLINE] [NEWLINE] I made on comment on this thread stating<mask><mask>, saying that respect should be standard and apply to everyone by default until there is a reason not to. I maintain this<mask> OP was considering definition 2,<mask><mask> OP meant it<mask> deep admiration (definition 1), then it's different, admiration is not a default, there must be a reason to admire someone. [NEWLINE] [NEWLINE] [NEWLINE] In any case, I learned something that could question my view on this thread.<mask><mask> one automatically owes'respect 2' to one's teacher,<mask> not'respect 1'.</s>
Label encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] Because of your comment I actually looked up the definition of'respect' and there are basically two definitions: 1) deep admiration for someone's achievements or abilities and 2) due regard for the feelings, wishes, or rights of others. [NEWLINE] [NEWLINE] [NEWLINE] English is not my native language and in Portuguese we don't really use respect in the sense of deep admiration (we just use admiration for that), we consider the 2nd definition for respect. [NEWLINE] [NEWLINE] I made on comment on this thread stating my opinion, saying that respect should be standard and apply to everyone by default until there is a reason not to. I maintain this if OP was considering definition 2, but if OP meant it as deep admiration (definition 1), then it's different, admiration is not a default, there must be a reason to admire someone. [NEWLINE] [NEWLINE] [NEWLINE] In any case, I learned something that could question my view on this thread. I think one automatically owes'respect 2' to one's teacher, but not'respect 1'.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While 'it' isn't a proper term for humans, the word 'it' is often used to describe transgendered/transexual people.  By attempting to make a non-offensive option it creates a workaround that everyone can agree upon, and no one needs to worry about incorrectly referring to someone or more importantly worry about not being recognized<mask> a person in general. <mask>, 'xe' etc. are useful<mask> you don't want to say 'he/she.' [NEWLINE] [NEWLINE] Someone being truly offended<mask> being referred to<mask> the wrong pronoun is a little excessive,<mask> being aware that they are actually very useful words is very important (<mask><mask> ). <mask> leading a discussion with a group of people,<mask> not start with "<mask> is everyone's name? And preferred pronoun?" it takes 2 extra seconds and creates a situation<mask> people can both receive and give real respect for other humans.  Even<mask> we aren't dealing with any gender ambiguous individuals, asking for preferred pronouns can help people who have naturally androgynous appearances.</s>
Label encoding: <s>While 'it' isn't a proper term for humans, the word 'it' is often used to describe transgendered/transexual people.  By attempting to make a non-offensive option it creates a workaround that everyone can agree upon, and no one needs to worry about incorrectly referring to someone or more importantly worry about not being recognized as a person in general.  Also, 'xe' etc. are useful when you don't want to say 'he/she.' [NEWLINE] [NEWLINE] Someone being truly offended when being referred to as the wrong pronoun is a little excessive, but being aware that they are actually very useful words is very important ( IMHO ).  When leading a discussion with a group of people, why not start with " What is everyone's name? And preferred pronoun?" it takes 2 extra seconds and creates a situation where people can both receive and give real respect for other humans.  Even if we aren't dealing with any gender ambiguous individuals, asking for preferred pronouns can help people who have naturally androgynous appearances.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So just to make my point clear. [NEWLINE] [NEWLINE] Yes I admit that Gandhi has become a propaganda tool for the state.<mask> that does not mean that he did not accomplish stuff which has a lot for us to learn. [NEWLINE] [NEWLINE] Imagine<mask> someone said "I believe Albert Einstein is overrated, CMV", and then talks about all these stories Christians tell each other("..and that boy's name was Albert Einstein")<mask> of his belief in god, then would he be right? These stupid stories do not discount<mask> Einstein did. [NEWLINE] [NEWLINE] Maybe at least you admit that Gandhi the man and Gandhi the propaganda are two different things, and Gandhi is great for different reasons. I would recommend reading him more. There's a lot of bullshit to go through(like his religiousness),<mask> its nothing like<mask> you read through<mask> child. [NEWLINE] [NEWLINE] I must warn you, you will face palm everything you'll hear Anna Hazare take his name on television. Anna Hazare has resurrected Gandhi's corpse,<mask> it does not have his soul in it.</s>
Label encoding: <s>So just to make my point clear. [NEWLINE] [NEWLINE] Yes I admit that Gandhi has become a propaganda tool for the state. But that does not mean that he did not accomplish stuff which has a lot for us to learn. [NEWLINE] [NEWLINE] Imagine if someone said "I believe Albert Einstein is overrated, CMV", and then talks about all these stories Christians tell each other("..and that boy's name was Albert Einstein") because of his belief in god, then would he be right? These stupid stories do not discount what Einstein did. [NEWLINE] [NEWLINE] Maybe at least you admit that Gandhi the man and Gandhi the propaganda are two different things, and Gandhi is great for different reasons. I would recommend reading him more. There's a lot of bullshit to go through(like his religiousness), but its nothing like what you read through as child. [NEWLINE] [NEWLINE] I must warn you, you will face palm everything you'll hear Anna Hazare take his name on television. Anna Hazare has resurrected Gandhi's corpse, but it does not have his soul in it.</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of people use "he/she"/"he-she"<mask> a pretty bigoted slur towards transgender people<mask> that's probably not a great idea. [NEWLINE] [NEWLINE] "It" is incredibly dehumanizing and is used intentionally that way<mask> referring to transgender people, including those on your board, from<mask> I've seen. [NEWLINE] [NEWLINE] Singular "they" is well established in the English language and its common usage, and doesn't single out transgender people. This is<mask><mask><mask> it's easier and better to use than "ze"/"zim"/whatever else internet social justice warriors use and the offensive stuff in your post. [NEWLINE] [NEWLINE] I feel there is a very strong chance you are just trying to create "drama" with people by posting this,<mask><mask> a heads up I'm not interested in arguing about SRS's modding policies or whatever other "internet drama" idiots can pull out of this. I gave a completely serious response to your post, and will likely ignore anyone that doesn't extend me the same courtesy.</s>
Label encoding: <s>A lot of people use "he/she"/"he-she" as a pretty bigoted slur towards transgender people so that's probably not a great idea. [NEWLINE] [NEWLINE] "It" is incredibly dehumanizing and is used intentionally that way when referring to transgender people, including those on your board, from what I've seen. [NEWLINE] [NEWLINE] Singular "they" is well established in the English language and its common usage, and doesn't single out transgender people. This is why I think it's easier and better to use than "ze"/"zim"/whatever else internet social justice warriors use and the offensive stuff in your post. [NEWLINE] [NEWLINE] I feel there is a very strong chance you are just trying to create "drama" with people by posting this, so as a heads up I'm not interested in arguing about SRS's modding policies or whatever other "internet drama" idiots can pull out of this. I gave a completely serious response to your post, and will likely ignore anyone that doesn't extend me the same courtesy.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How do you figure government people are held accountable? [NEWLINE] [NEWLINE] Police mutilate a baby, no charges.  Choke and smother a man to death in public, no crime.  Wars are declared, that kill thousands or million of non combatants,<mask> it's not murder.  Property seized without due process,<mask> that's not theft. Intern thousands of people,<mask> of their ancestry.  Imprison tens of millions of people, ruining lives, families and livelihoods over plants and pills. [NEWLINE] [NEWLINE] <mask> governments are "accountable," like after a lawsuit, it's the tax payer that pays the damages, not the offensive actors. [NEWLINE] [NEWLINE] **<mask> evidence do you have the governments, even democratic ones, are ever accountable?**  Maybe only<mask> they lose a war. [NEWLINE] [NEWLINE] Governments are for<mask> you want to protect people from being accountable.  Sovereign immunity, qualified immunity, diplomatic immunity. [NEWLINE] [NEWLINE] Maybe, a particularly corrupt or abusive politician can be voted out of office,<mask> that's not accountability either.</s>
Label encoding: <s>How do you figure government people are held accountable? [NEWLINE] [NEWLINE] Police mutilate a baby, no charges.  Choke and smother a man to death in public, no crime.  Wars are declared, that kill thousands or million of non combatants, but it's not murder.  Property seized without due process, but that's not theft. Intern thousands of people, because of their ancestry.  Imprison tens of millions of people, ruining lives, families and livelihoods over plants and pills. [NEWLINE] [NEWLINE] When governments are "accountable," like after a lawsuit, it's the tax payer that pays the damages, not the offensive actors. [NEWLINE] [NEWLINE] ** What evidence do you have the governments, even democratic ones, are ever accountable?**  Maybe only when they lose a war. [NEWLINE] [NEWLINE] Governments are for when you want to protect people from being accountable.  Sovereign immunity, qualified immunity, diplomatic immunity. [NEWLINE] [NEWLINE] Maybe, a particularly corrupt or abusive politician can be voted out of office, but that's not accountability either.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1) Could you explain an "intersectionality argument" a little more, I don't quite understand. <mask> from<mask> you wrote it doesn't address my point.  The habitat is already destroyed and unlikely to be rebuilt.  In point 4 I do advocate for preserving<mask> is left. [NEWLINE] 2 &amp; 3) My point was not that the panda is an evolutionary dead end in it's natural setting, its that their adaptations make them especially vulnerable in the current environmental setting.  This means that humans need to provide substantially more intervention to save them than other vulnerbable and more important species. [NEWLINE] 4) I could not agree more that biodiversity is the long term goal, it's my argument that preserving the panda works against this goal.  A disproportionate amount of valuable and finite resources are being used to save them,<mask> they themselves don't seem to add much to biodiversity.  There are more important species that could use those resources.  I'm not advocating abandoning the panda, just reallocating the resources. </s>
Label encoding: <s>1) Could you explain an "intersectionality argument" a little more, I don't quite understand.  But from what you wrote it doesn't address my point.  The habitat is already destroyed and unlikely to be rebuilt.  In point 4 I do advocate for preserving what is left. [NEWLINE] 2 &amp; 3) My point was not that the panda is an evolutionary dead end in it's natural setting, its that their adaptations make them especially vulnerable in the current environmental setting.  This means that humans need to provide substantially more intervention to save them than other vulnerbable and more important species. [NEWLINE] 4) I could not agree more that biodiversity is the long term goal, it's my argument that preserving the panda works against this goal.  A disproportionate amount of valuable and finite resources are being used to save them, when they themselves don't seem to add much to biodiversity.  There are more important species that could use those resources.  I'm not advocating abandoning the panda, just reallocating the resources. </s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was under the impression that he was a science *educator*, rather than a scientist. It was<mask><mask><mask> alone I thought it was acceptable that he "debate" Ken Ham last year,<mask> "real" scientists debating Ken Ham would be to disrespect the scientific community and give Ham credibility which he has not earned. [NEWLINE] [NEWLINE] That being said, there is no standard definition of<mask> a scientist is.  It's not like being a doctor or lawyer<mask> there is a degree involved<mask> well<mask> membership in some sort of regulatory group (ie, the medical board, or bar association.) Most broadly, anyone who does science, is a scientist. An 8-year-old who conducts plant experiments using tomato plants is a scientist in this sense. [NEWLINE] [NEWLINE] I would say in the broadest sense he is a scientist,<mask> in the every-day usage of the word (ie, people who actively are part of the scientific process) I would say he is not, and would better be described<mask> a science-educator.</s>
Label encoding: <s>I was under the impression that he was a science *educator*, rather than a scientist. It was for this reason alone I thought it was acceptable that he "debate" Ken Ham last year, as "real" scientists debating Ken Ham would be to disrespect the scientific community and give Ham credibility which he has not earned. [NEWLINE] [NEWLINE] That being said, there is no standard definition of what a scientist is.  It's not like being a doctor or lawyer where there is a degree involved as well as membership in some sort of regulatory group (ie, the medical board, or bar association.) Most broadly, anyone who does science, is a scientist. An 8-year-old who conducts plant experiments using tomato plants is a scientist in this sense. [NEWLINE] [NEWLINE] I would say in the broadest sense he is a scientist, but in the every-day usage of the word (ie, people who actively are part of the scientific process) I would say he is not, and would better be described as a science-educator.</s>
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Suppose I'm a performer and a song by Biggie really inspired me and caused me to make significant changes in my life for the better. Suppose I get the opportunity to perform on stage and want to do my interpretation of the song that's been really meaningful to my life? [NEWLINE] [NEWLINE] Do I just say "Well shuckie darns. I come from different color folks<mask> I guess I can't be inspired by that hip hop music! Oh well!" and ignore the fact that it's been instrumental to my life<mask> the opportunity to pay homage? [NEWLINE] [NEWLINE] Or do I white wash it<mask> that I don't say a "bad word" and destroy the context of someone else's art? [NEWLINE] [NEWLINE] I pointed out in the beginning this isn't about me wanting to have permission to use the word - I really don't think I'd ever use it in that context<mask> it's not my slang. It's about feeling like I'm not *allowed*<mask> of my skin color which is really silly from my perspective. </s>
Label encoding: <s>Suppose I'm a performer and a song by Biggie really inspired me and caused me to make significant changes in my life for the better. Suppose I get the opportunity to perform on stage and want to do my interpretation of the song that's been really meaningful to my life? [NEWLINE] [NEWLINE] Do I just say "Well shuckie darns. I come from different color folks so I guess I can't be inspired by that hip hop music! Oh well!" and ignore the fact that it's been instrumental to my life despite the opportunity to pay homage? [NEWLINE] [NEWLINE] Or do I white wash it so that I don't say a "bad word" and destroy the context of someone else's art? [NEWLINE] [NEWLINE] I pointed out in the beginning this isn't about me wanting to have permission to use the word - I really don't think I'd ever use it in that context because it's not my slang. It's about feeling like I'm not *allowed* because of my skin color which is really silly from my perspective. </s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Less profit is still not an excuse. Perhaps you see business<mask> amoral,<mask> there are reasons there are restrictions on business. It is more profitable to lie to customers or have a monopoly on an industry or to pay illegals less than minimum wage,<mask> it's still wrong. It<mask> doesn't apply here<mask> OP isn't facing a lack of qualified females- he's disqualifying females based on their sex alone. I don't think the burden of maternity leave should be on the employers,<mask> apparently wherever OP lives it is and that's just that. You shouldn't say "there are women who can make up for that loss in profit"<mask> now you've set an even higher standard- men can be profitable, women have to be at least<mask> profitable<mask> their lost time to maternity leave. Finally, some companies are probably going to have to take a cut in profit to further the interests of diversity and equality. Maybe you think those aren't worthwhile pursuits,<mask> at least we can still fine your discriminatory hiring practices. </s>
Label encoding: <s>Less profit is still not an excuse. Perhaps you see business as amoral, but there are reasons there are restrictions on business. It is more profitable to lie to customers or have a monopoly on an industry or to pay illegals less than minimum wage, but it's still wrong. It also doesn't apply here because OP isn't facing a lack of qualified females- he's disqualifying females based on their sex alone. I don't think the burden of maternity leave should be on the employers, but apparently wherever OP lives it is and that's just that. You shouldn't say "there are women who can make up for that loss in profit" because now you've set an even higher standard- men can be profitable, women have to be at least as profitable as their lost time to maternity leave. Finally, some companies are probably going to have to take a cut in profit to further the interests of diversity and equality. Maybe you think those aren't worthwhile pursuits, but at least we can still fine your discriminatory hiring practices. </s>
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, thank you for being a rational and compassionate human being. Far too many times I see a topic relating to cyclists brought up, and the majority opinion—even on usually left-wing Reddit—is that cyclists are a waste of space and cars shouldn't have to deal with cyclists on the road. It's an entirely selfish and arrogant viewpoint. I completely agree with you,<mask> my argument is tentatively devil's advocate,<mask> not entirely. [NEWLINE] [NEWLINE] My argument is that bike lanes are only marginally safer than not having them,<mask> many drivers tend to ignore them<mask> they're concentrating on other things like changing lanes or turning. Instead, a better option is proper bike paths that are completely separated from the road.<mask><mask><mask> pedestrians are aware of them, it's much safer than having cyclists ride on the road,<mask><mask> whether or not there is a bike lane. Now, it isn't feasible in all cases, and it's a lot more expensive,<mask> in cases<mask> it is feasible it's much safer.</s>
Label encoding: <s>First of all, thank you for being a rational and compassionate human being. Far too many times I see a topic relating to cyclists brought up, and the majority opinion—even on usually left-wing Reddit—is that cyclists are a waste of space and cars shouldn't have to deal with cyclists on the road. It's an entirely selfish and arrogant viewpoint. I completely agree with you, so my argument is tentatively devil's advocate, although not entirely. [NEWLINE] [NEWLINE] My argument is that bike lanes are only marginally safer than not having them, since many drivers tend to ignore them if they're concentrating on other things like changing lanes or turning. Instead, a better option is proper bike paths that are completely separated from the road. As long as pedestrians are aware of them, it's much safer than having cyclists ride on the road, regardless of whether or not there is a bike lane. Now, it isn't feasible in all cases, and it's a lot more expensive, but in cases where it is feasible it's much safer.</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You think cars are the only risk to outdoor cats? [ENDQ] [NEWLINE] No ive had one cat eaten coyote and people have stolen another. Many die of old age. A few have gotten ran over. [NEWLINE] [NEWLINE] Again, yes there are risks to being outside<mask> i believe those risks do not trump the quality of life gotten from being outside. [NEWLINE] [NEWLINE] [STARTQ] And your cat urinates on your stuff for other reasons, not<mask> it's mad at you for keeping it inside. [ENDQ] [NEWLINE] Im sorry<mask> you dont know my cat at all. it pees<mask> it has seperation anxiety and its a brat. It does it now a lot less than the first few months<mask> it still occasionally does it. It uses the litterbox plenty. Its been fixed. Its been to the vet to test for iputis and whatever. It just is a highly stressed cat and maybe a little bored. i never said it peed<mask> it wanted outside, i was pointing out that not all cats are able to behave inside.</s>
Label encoding: <s> [STARTQ] You think cars are the only risk to outdoor cats? [ENDQ] [NEWLINE] No ive had one cat eaten coyote and people have stolen another. Many die of old age. A few have gotten ran over. [NEWLINE] [NEWLINE] Again, yes there are risks to being outside but i believe those risks do not trump the quality of life gotten from being outside. [NEWLINE] [NEWLINE] [STARTQ] And your cat urinates on your stuff for other reasons, not because it's mad at you for keeping it inside. [ENDQ] [NEWLINE] Im sorry but you dont know my cat at all. it pees because it has seperation anxiety and its a brat. It does it now a lot less than the first few months but it still occasionally does it. It uses the litterbox plenty. Its been fixed. Its been to the vet to test for iputis and whatever. It just is a highly stressed cat and maybe a little bored. i never said it peed because it wanted outside, i was pointing out that not all cats are able to behave inside.</s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0592, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I find it easier to see right-wing views<mask> based in the beliefs that the world is only a few short steps to total collapse of society<mask> we know it. The resources the world has are finite, and a free market is the best way we've got to allocate resources<mask> they are needed most. [NEWLINE] [NEWLINE] [STARTQ] I have not been convinced about<mask> right-wing views would support disabled people who are completely unable to look after themselves (<mask> no amount of 'allocating' their own time and money will help). [ENDQ] [NEWLINE] <mask><mask> that right-wing people would<mask><mask> charity would support those people. There is *some* evidence that this would be the case<mask> charities like those existed before social security was a thing. It was a very imperfect system and I have no idea<mask> right-wing people account for this. [NEWLINE] [NEWLINE] It's worth keeping in mind that there are strands of right-wing politics that support a basic income, which would solve a good chunk of the problems associated with the rest of the philosophy.</s>
Label encoding: <s>I find it easier to see right-wing views as based in the beliefs that the world is only a few short steps to total collapse of society as we know it. The resources the world has are finite, and a free market is the best way we've got to allocate resources where they are needed most. [NEWLINE] [NEWLINE] [STARTQ] I have not been convinced about how right-wing views would support disabled people who are completely unable to look after themselves ( where no amount of 'allocating' their own time and money will help). [ENDQ] [NEWLINE] I think that right-wing people would argue that charity would support those people. There is *some* evidence that this would be the case because charities like those existed before social security was a thing. It was a very imperfect system and I have no idea how right-wing people account for this. [NEWLINE] [NEWLINE] It's worth keeping in mind that there are strands of right-wing politics that support a basic income, which would solve a good chunk of the problems associated with the rest of the philosophy.</s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0851, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know, I know, it was a big goofup for me, I knew previously<mask> it completely slipped my mind,<mask> it happens. Truly embarrassing for me, seriously. [NEWLINE] [NEWLINE] And<mask> you find the men morally reprehensible, then that's fine and all,<mask><mask> you have to choose between a guy who shoplifted a pack of cigs<mask> he was a teenager and a guy who currently eats babies, the choice should be clear (I know that is a gross exaggeration,<mask><mask><mask> the magnitude of the difference is mostly irrelevant). The point was that even<mask> a third party nominee truly deserves the presidency more, it doesn't matter<mask> there is just a negligible chance of them winning. Maybe<mask><mask> I'M the best choice for president,<mask> I sure<mask> hell am not going to write in my own name. [NEWLINE] [NEWLINE] <mask> I suppose this is irrelevant now,<mask> my view on the topic was already changed, just not for ideological reasons<mask> you proposed. Thanks for the input anyway<mask>!</s>
Label encoding: <s>I know, I know, it was a big goofup for me, I knew previously but it completely slipped my mind, but it happens. Truly embarrassing for me, seriously. [NEWLINE] [NEWLINE] And if you find the men morally reprehensible, then that's fine and all, but if you have to choose between a guy who shoplifted a pack of cigs when he was a teenager and a guy who currently eats babies, the choice should be clear (I know that is a gross exaggeration, but imo the magnitude of the difference is mostly irrelevant). The point was that even if a third party nominee truly deserves the presidency more, it doesn't matter if there is just a negligible chance of them winning. Maybe I think I'M the best choice for president, but I sure as hell am not going to write in my own name. [NEWLINE] [NEWLINE] But I suppose this is irrelevant now, as my view on the topic was already changed, just not for ideological reasons as you proposed. Thanks for the input anyway though!</s>
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Basically I feel there is way too large an emphasis on avoiding causing offence. There was an example recently<mask> children<mask> young<mask> 3 had to sign a contract promising they would not be racist, homophobic, transphobic etc. I feel that this is ludicrous for such a young age. Source: [URL] Then there is positive discrimination which I'm<mask> against I feel there should be a system of meritocracy. Instead of quotas for women and/or ethnic minorities. I feel the best person should get the job. Then SJW's<mask> profoundly irritate me, I feel they are very sanctimonious. [NEWLINE] Am I wrong, has political correctness not gone far enough or is it just fine the way it is going. I am always open to changing my view<mask><mask> strong my views may seem. [NEWLINE] Edit: I am not saying I do not support the idea of everyone being equal and being free from discrimination. I am just stating I believe it is being implemented poorly. [NEWLINE] Edit 2: Seems<mask><mask> is very unpopular.</s>
Label encoding: <s>Basically I feel there is way too large an emphasis on avoiding causing offence. There was an example recently where children as young as 3 had to sign a contract promising they would not be racist, homophobic, transphobic etc. I feel that this is ludicrous for such a young age. Source: [URL] Then there is positive discrimination which I'm also against I feel there should be a system of meritocracy. Instead of quotas for women and/or ethnic minorities. I feel the best person should get the job. Then SJW's also profoundly irritate me, I feel they are very sanctimonious. [NEWLINE] Am I wrong, has political correctness not gone far enough or is it just fine the way it is going. I am always open to changing my view despite how strong my views may seem. [NEWLINE] Edit: I am not saying I do not support the idea of everyone being equal and being free from discrimination. I am just stating I believe it is being implemented poorly. [NEWLINE] Edit 2: Seems my opinion is very unpopular.</s>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If I can simplify<mask> you're suggesting, you're saying that basically "poor people don't know<mask> to spend their money" and<mask> the government should be responsible for choosing<mask> to spend it on for them. I don't think many people would agree with that,<mask><mask> it's fundamentally different than<mask> you're saying, it'd be interesting to see<mask> you'd describe the difference. [NEWLINE] [NEWLINE] And the biggest reason that people would agree with that kind of program, at least in this discussion, is that UBI isn't a welfare program. It's an economic program that's intended to help everyone.<mask><mask> that vast majority of people who will benefit from it will be middle class families. [NEWLINE] [NEWLINE] It just<mask> happens that with UBI in place, it would reduce or eliminate the need for many welfare programs. This would effect a small percentage of the population,<mask> it would have the nice side effect that direct cash transfers seem to be much more efficient and effective than many other kinds of assistance programs. [NEWLINE] </s>
Label encoding: <s>If I can simplify what you're suggesting, you're saying that basically "poor people don't know how to spend their money" and so the government should be responsible for choosing what to spend it on for them. I don't think many people would agree with that, but if it's fundamentally different than what you're saying, it'd be interesting to see how you'd describe the difference. [NEWLINE] [NEWLINE] And the biggest reason that people would agree with that kind of program, at least in this discussion, is that UBI isn't a welfare program. It's an economic program that's intended to help everyone. In fact that vast majority of people who will benefit from it will be middle class families. [NEWLINE] [NEWLINE] It just also happens that with UBI in place, it would reduce or eliminate the need for many welfare programs. This would effect a small percentage of the population, but it would have the nice side effect that direct cash transfers seem to be much more efficient and effective than many other kinds of assistance programs. [NEWLINE] </s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a few problems with this,<mask> I'll address<mask> I believe to be the main problem with the idea. [NEWLINE] [NEWLINE] You suggested this be done every 100 years or<mask>. There's nothing inherently wrong with that amount of time,<mask> rather having a set date in general. [NEWLINE] [NEWLINE] Think about<mask>, in today's political climate, everyone knew that in, say, 15 years the Constitution would have to be redone. Most politics for the next 15 years would be totally centered around each political party trying to gain the upper hand for the next rewrite. It would be similar to a lame duck president, except it would have far more detrimental effects to the overall political process. [NEWLINE] [NEWLINE] The Constitution was designed to allow for amendments<mask> we<mask> a nation see that something needs to be changed or updated. And the Supreme Court exists to interpret the Constitution and its amendments. Adding this countdown timer to its revision would distract from the day-to-day policy making that is, in reality, more important to our government.</s>
Label encoding: <s>There are a few problems with this, but I'll address what I believe to be the main problem with the idea. [NEWLINE] [NEWLINE] You suggested this be done every 100 years or so. There's nothing inherently wrong with that amount of time, but rather having a set date in general. [NEWLINE] [NEWLINE] Think about if, in today's political climate, everyone knew that in, say, 15 years the Constitution would have to be redone. Most politics for the next 15 years would be totally centered around each political party trying to gain the upper hand for the next rewrite. It would be similar to a lame duck president, except it would have far more detrimental effects to the overall political process. [NEWLINE] [NEWLINE] The Constitution was designed to allow for amendments when we as a nation see that something needs to be changed or updated. And the Supreme Court exists to interpret the Constitution and its amendments. Adding this countdown timer to its revision would distract from the day-to-day policy making that is, in reality, more important to our government.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am well aware of the law. That being said,<mask><mask> that such terms of service are inherently unethical and should<mask> be ignored or outright violated. It's<mask> nonsensical to say that I didn't buy the data<mask> I have a copy of the data that I can copy and manipulate without the control or consent of the person selling me the data.<mask> whoever is selling me data wants a term of service consistent with the method in which the data is transmitted, then those who create the data should release it in a manner<mask> I don't have direct access to the data that I have permission to use. There is nothing unethical about copying data and redistributing it in any way,<mask><mask><mask> the law says. [NEWLINE] [NEWLINE] And yes, I primarily will only use software<mask> it is legal for me to view, modify, and redistribute the source code and binaries. Coincidentally, software released with those conditions is typically of higher quality than software without those conditions,<mask> I'm actually benefiting from that.</s>
Label encoding: <s>I am well aware of the law. That being said, I think that such terms of service are inherently unethical and should therefore be ignored or outright violated. It's also nonsensical to say that I didn't buy the data when I have a copy of the data that I can copy and manipulate without the control or consent of the person selling me the data. If whoever is selling me data wants a term of service consistent with the method in which the data is transmitted, then those who create the data should release it in a manner where I don't have direct access to the data that I have permission to use. There is nothing unethical about copying data and redistributing it in any way, regardless of what the law says. [NEWLINE] [NEWLINE] And yes, I primarily will only use software where it is legal for me to view, modify, and redistribute the source code and binaries. Coincidentally, software released with those conditions is typically of higher quality than software without those conditions, so I'm actually benefiting from that.</s>
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No one seems to be actually trying to change your view. [NEWLINE] [NEWLINE] Generally, I'd say great literature is like any piece of art, it is meant to provoke an emotional response. You mention that Catcher in the Rye was a book you enjoyed<mask> it resonated with your age and issues you faced. Not all teenagers will<mask> necessarily reflect your own preferences and life experiences. Some may respond to King Lear, others to To Kill a Mockingbird (this was the classic that really resonated with me in English class at school, it didn't really have much about teenagers,<mask> it did speak to me about issues of justice and racism that reflected my own experiences). [NEWLINE] [NEWLINE] Essentially, you can't predict<mask> a person responds to a piece of art. Cover a decent range of classics and you can ensure that you a) increase the chances of reaching a student with a great book that resonates with them and b) give them the skills to recognise great writing even<mask> it doesn't resonate with them.</s><pad>
Label encoding: <s>No one seems to be actually trying to change your view. [NEWLINE] [NEWLINE] Generally, I'd say great literature is like any piece of art, it is meant to provoke an emotional response. You mention that Catcher in the Rye was a book you enjoyed because it resonated with your age and issues you faced. Not all teenagers will however necessarily reflect your own preferences and life experiences. Some may respond to King Lear, others to To Kill a Mockingbird (this was the classic that really resonated with me in English class at school, it didn't really have much about teenagers, but it did speak to me about issues of justice and racism that reflected my own experiences). [NEWLINE] [NEWLINE] Essentially, you can't predict how a person responds to a piece of art. Cover a decent range of classics and you can ensure that you a) increase the chances of reaching a student with a great book that resonates with them and b) give them the skills to recognise great writing even if it doesn't resonate with them.</s><pad>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, I'm implying that being an asshole is ruining this country. You can disagree with someone's beliefs and still be respectful. For example, I'm not Muslim and I don't agree with a lot of Muslim beliefs,<mask> I won't tell Muslims that their religion is stupid or bigoted or outdated or whatever<mask> that's just a dick move and it benefits absolutely no one. And I won't make jokes at the expense of them or their religion<mask> doing<mask> would turn them into a caricature and dehumanize them. There's no argument to be won in this case<mask> you're trying to dissuade someone from their cultural identity and that won't work due to factors on either side of the aisle. [NEWLINE] [NEWLINE] [NEWLINE] Political correctness isn't inherently about trying not to offend anyone; it's about respecting people and their cultures. And no, anti-vaxxers don't benefit from PC culture<mask> that's an idea not a cultural identity. There *is* an argument to be had in this case.</s>
Label encoding: <s>No, I'm implying that being an asshole is ruining this country. You can disagree with someone's beliefs and still be respectful. For example, I'm not Muslim and I don't agree with a lot of Muslim beliefs, but I won't tell Muslims that their religion is stupid or bigoted or outdated or whatever because that's just a dick move and it benefits absolutely no one. And I won't make jokes at the expense of them or their religion because doing so would turn them into a caricature and dehumanize them. There's no argument to be won in this case because you're trying to dissuade someone from their cultural identity and that won't work due to factors on either side of the aisle. [NEWLINE] [NEWLINE] [NEWLINE] Political correctness isn't inherently about trying not to offend anyone; it's about respecting people and their cultures. And no, anti-vaxxers don't benefit from PC culture because that's an idea not a cultural identity. There *is* an argument to be had in this case.</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How long did it take you to know<mask> to tell a story through film?<mask> many of those storytelling skills would translate to a shorter format, like TV or a web show?<mask> many new ones would you need to learn to do those competently? [NEWLINE] [NEWLINE] I bet the answers to 2 and 3 are less than you'd think. [NEWLINE] [NEWLINE] <mask> for the audience laughter, applause, reaction... have you considered directing a play? It's definitely not<mask> glamorous<mask> it used to be,<mask> still fun<mask><mask>. [NEWLINE] [NEWLINE] I guess I'm projecting a bit: I dabble in a bit of everything; I draw and animate, play musical instruments and sing, act in short film, improv, and theater (<mask> this hasn't happened in a long<mask> ), and juggle/clown. Just to name the most enjoyable and fulfilling. And I guess it's hard for me to understand<mask> you'd want to limit yourself to just one type of performance. It's all just<mask> much fun.</s>
Label encoding: <s>How long did it take you to know how to tell a story through film? How many of those storytelling skills would translate to a shorter format, like TV or a web show? How many new ones would you need to learn to do those competently? [NEWLINE] [NEWLINE] I bet the answers to 2 and 3 are less than you'd think. [NEWLINE] [NEWLINE] As for the audience laughter, applause, reaction... have you considered directing a play? It's definitely not as glamorous as it used to be, but still fun IMO. [NEWLINE] [NEWLINE] I guess I'm projecting a bit: I dabble in a bit of everything; I draw and animate, play musical instruments and sing, act in short film, improv, and theater ( though this hasn't happened in a long while ), and juggle/clown. Just to name the most enjoyable and fulfilling. And I guess it's hard for me to understand why you'd want to limit yourself to just one type of performance. It's all just so much fun.</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's a serious concern that many Oregonians, myself in particular, have. [NEWLINE] [NEWLINE] The main reason, hell the *only* reason, I don't want Southern California to dry up and blow away is that many of those people will move up north. And we don't have the housing, the jobs or the money to support them. [NEWLINE] [NEWLINE] A mass exodus of Californians to Oregon would destroy Oregon economically, politically and culturally. Even<mask> 10% of Californians moved to Oregon they'd make up half of our population. That's not even counting the Californians that are already here. [NEWLINE] [NEWLINE] And the people that would be starting businesses and creating jobs in Oregon won't come here. They'll either stay in California<mask> they can afford it or move to Washington to take advantage of their lower taxes. [NEWLINE] [NEWLINE] Our unemployment is just too high to let in millions of water-seeking Californians. It would be disastrous. California needs to get it's shit together before it starts adversely affecting other states.</s>
Label encoding: <s>That's a serious concern that many Oregonians, myself in particular, have. [NEWLINE] [NEWLINE] The main reason, hell the *only* reason, I don't want Southern California to dry up and blow away is that many of those people will move up north. And we don't have the housing, the jobs or the money to support them. [NEWLINE] [NEWLINE] A mass exodus of Californians to Oregon would destroy Oregon economically, politically and culturally. Even if 10% of Californians moved to Oregon they'd make up half of our population. That's not even counting the Californians that are already here. [NEWLINE] [NEWLINE] And the people that would be starting businesses and creating jobs in Oregon won't come here. They'll either stay in California because they can afford it or move to Washington to take advantage of their lower taxes. [NEWLINE] [NEWLINE] Our unemployment is just too high to let in millions of water-seeking Californians. It would be disastrous. California needs to get it's shit together before it starts adversely affecting other states.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0633, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How do you know it's not one of the dominant issues<mask> not the top issue already? Everybody talks about it.<mask> one thing I wish we could focus on is real solutions like more efficient ways to produce and distribute food (Hint: the solution isn't shipping chicken from the U.S. to China for processing and then shipping it back to the U.S.), ways to produce energy that doesn't involve burning a lot of fossil fuels or use up a lot of land the way solar panel "farms" do, etc., etc. We know<mask> the problem is. We just need real world solutions to solving it that don't presume that every individual human is going to make drastic changes to their lifestyles just to cut down on their impact. At this point, climate change is likely to get worse before it gets better no matter<mask> we do. Now it's just a matter of making sure humanity can survive the experience even<mask> it means finding all the *right* ways to change a planet.</s><pad>
Label encoding: <s>How do you know it's not one of the dominant issues if not the top issue already? Everybody talks about it. But one thing I wish we could focus on is real solutions like more efficient ways to produce and distribute food (Hint: the solution isn't shipping chicken from the U.S. to China for processing and then shipping it back to the U.S.), ways to produce energy that doesn't involve burning a lot of fossil fuels or use up a lot of land the way solar panel "farms" do, etc., etc. We know what the problem is. We just need real world solutions to solving it that don't presume that every individual human is going to make drastic changes to their lifestyles just to cut down on their impact. At this point, climate change is likely to get worse before it gets better no matter what we do. Now it's just a matter of making sure humanity can survive the experience even if it means finding all the *right* ways to change a planet.</s><pad>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> you liked this person and genuinely wanted to be with them, something this stupid wouldn't bother you. [ENDQ] [NEWLINE] This simply isn't true.  These things can bug you AND you still want to be with them. <mask>, you compromise and either talk to them about it or just learn to ignore.  No, neither of you deserve the relationship or deserve the change from the other,<mask> that doesn't mean that you can't choose to make the change. [NEWLINE] [NEWLINE] <mask> i read your posts I see a dude with a mental check list that he runs down.  And<mask> the girl doesn't meet every criteria than he just leaves. No questions asked, no talking about anything, just leaves.  And<mask> that's okay and your decision, it doesn't mean it is good. [NEWLINE] [NEWLINE] There are lot of things that people are allowed to do/capable of doing. <mask> just becasue you can do something doesn't mean it is the best thing to do.</s>
Label encoding: <s> [STARTQ] Because if you liked this person and genuinely wanted to be with them, something this stupid wouldn't bother you. [ENDQ] [NEWLINE] This simply isn't true.  These things can bug you AND you still want to be with them.  So, you compromise and either talk to them about it or just learn to ignore.  No, neither of you deserve the relationship or deserve the change from the other, but that doesn't mean that you can't choose to make the change. [NEWLINE] [NEWLINE] When i read your posts I see a dude with a mental check list that he runs down.  And if the girl doesn't meet every criteria than he just leaves. No questions asked, no talking about anything, just leaves.  And while that's okay and your decision, it doesn't mean it is good. [NEWLINE] [NEWLINE] There are lot of things that people are allowed to do/capable of doing.  But just becasue you can do something doesn't mean it is the best thing to do.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A lock is to only keep an honest person honest. [ENDQ] [NEWLINE] Oh come the fuck on,<mask> you don't think a lock will deter someone you are fucking crazy.<mask> many people do you think walk by random cars and try the handles to see<mask> they can get inside? Tons, they aren't going to break the window just<mask> it's locked unless they really want to get into that specific car, otherwise they will PICK AN EASIER TARGET. It is the same with rape and sexual assault, they know that they can't just grab someone on the sidewalk in broad daylight. They know that<mask> they don't want to go to prison they have to do it in a way that they can get away with. Do you think rapists are just walking around foaming at the mouth and unable to control themselves everywhere they go? No, they pick their time and place carefully just like any other smart criminal, which is<mask> it is important to not make yourself an easy target.</s>
Label encoding: <s> [STARTQ] A lock is to only keep an honest person honest. [ENDQ] [NEWLINE] Oh come the fuck on, if you don't think a lock will deter someone you are fucking crazy. How many people do you think walk by random cars and try the handles to see if they can get inside? Tons, they aren't going to break the window just because it's locked unless they really want to get into that specific car, otherwise they will PICK AN EASIER TARGET. It is the same with rape and sexual assault, they know that they can't just grab someone on the sidewalk in broad daylight. They know that if they don't want to go to prison they have to do it in a way that they can get away with. Do you think rapists are just walking around foaming at the mouth and unable to control themselves everywhere they go? No, they pick their time and place carefully just like any other smart criminal, which is why it is important to not make yourself an easy target.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Being liberal originally meant to support individual rights, freedom and the right to believe<mask> you want to and think<mask> you want to, no matter<mask> offensive or wrong. [NEWLINE] [NEWLINE] Originally it was more of a lockean or John Stuart Mill thing. Today it's more of a Marxist or Leninist ideology. [NEWLINE] [NEWLINE] I'm still confused<mask> "freedom" can be confused with and changed to "the only ideology that fostered for millions dying under oppressive regimes". [NEWLINE] [NEWLINE] Freedom for the proletariat! No, not really. They only have the freedom to rely almost entirely on government handouts (with no incentive to attain any pride in their work<mask> free money), businesses lack the profits after taxes to take the risk in hiring less qualified applicants, cost of living rises<mask><mask><mask> of higher costs implemented by government and high amounts of inflation and<mask><mask> our freedom<mask> a country has dropped down to in the 36th freest nation in the world. [NEWLINE] [NEWLINE] <mask> are our liberals who support freedom? </s>
Label encoding: <s>Being liberal originally meant to support individual rights, freedom and the right to believe what you want to and think what you want to, no matter how offensive or wrong. [NEWLINE] [NEWLINE] Originally it was more of a lockean or John Stuart Mill thing. Today it's more of a Marxist or Leninist ideology. [NEWLINE] [NEWLINE] I'm still confused how "freedom" can be confused with and changed to "the only ideology that fostered for millions dying under oppressive regimes". [NEWLINE] [NEWLINE] Freedom for the proletariat! No, not really. They only have the freedom to rely almost entirely on government handouts (with no incentive to attain any pride in their work because free money), businesses lack the profits after taxes to take the risk in hiring less qualified applicants, cost of living rises as a result of higher costs implemented by government and high amounts of inflation and lastly our freedom as a country has dropped down to in the 36th freest nation in the world. [NEWLINE] [NEWLINE] Where are our liberals who support freedom? </s>
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 8.There is no bad reason to end a relationship. I keep hearing people say<mask> someone broke up with them over something dumb - think of it, the dumber the reason someone breaks up with you over, the more it means that the relationship wasn't that important to them in the first place. Which means the break up was necessary, even more<mask> than<mask> people in love break up over something big and understandable. [ENDQ] [NEWLINE] Let's just tackle #8 right here - of course there are bad reasons to break up with someone. Misunderstandings happen all the time, people may have prior issues that they think is a game-changer<mask> actually isn't, etc. Do you really think people are perfect all the time? [NEWLINE] [NEWLINE] Even<mask> a generic claim this is illogical - you'd have to discount all the possible reasons people break up, and judge them all<mask> legitimate. This is absurd, and I'm sure most can create an appropriate counterexample.</s><pad>
Label encoding: <s> [STARTQ] 8.There is no bad reason to end a relationship. I keep hearing people say how someone broke up with them over something dumb - think of it, the dumber the reason someone breaks up with you over, the more it means that the relationship wasn't that important to them in the first place. Which means the break up was necessary, even more so than when people in love break up over something big and understandable. [ENDQ] [NEWLINE] Let's just tackle #8 right here - of course there are bad reasons to break up with someone. Misunderstandings happen all the time, people may have prior issues that they think is a game-changer but actually isn't, etc. Do you really think people are perfect all the time? [NEWLINE] [NEWLINE] Even as a generic claim this is illogical - you'd have to discount all the possible reasons people break up, and judge them all as legitimate. This is absurd, and I'm sure most can create an appropriate counterexample.</s><pad>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0722, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0733, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it is more accurate to say that hip hop was the most important musical movement *fifteen years ago*. KRS-One has this great line, ["Big Brother watching over you is a lie, you see. Hip hop has built its own secret society"]( [URL] ). It is well beyond the phase<mask> it was something unique and groundbreaking. We are now into the phase<mask> it is something that will reliably sell. And<mask><mask> Kanye is a perfect representative of this bloated and established period. He has talent,<mask> he thinks he is more important than he is. He is wildly successful and still can't stop complaining about<mask> "the man" keeps him from pursuing his dreams<mask> living in mansions and selling white t-shirts for a small fortune. [NEWLINE] [NEWLINE] A little over fifteen years ago, people like [Pete Rock &amp; CL Smooth]( [URL] ) were doing something incredible. They did make a massive impact on the world of music and now...we have Kanye.</s>
Label encoding: <s>I think it is more accurate to say that hip hop was the most important musical movement *fifteen years ago*. KRS-One has this great line, ["Big Brother watching over you is a lie, you see. Hip hop has built its own secret society"]( [URL] ). It is well beyond the phase where it was something unique and groundbreaking. We are now into the phase where it is something that will reliably sell. And I think Kanye is a perfect representative of this bloated and established period. He has talent, but he thinks he is more important than he is. He is wildly successful and still can't stop complaining about how "the man" keeps him from pursuing his dreams while living in mansions and selling white t-shirts for a small fortune. [NEWLINE] [NEWLINE] A little over fifteen years ago, people like [Pete Rock &amp; CL Smooth]( [URL] ) were doing something incredible. They did make a massive impact on the world of music and now...we have Kanye.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0628, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Small factoid: Jobs is actually on several Apple patents. [NEWLINE] [NEWLINE] He brought<mask> other electronics makers didn't, an ocsessive-compulsive approach to making the entire experience perfect. The Apple Store format that MS and Samsung are copying wasn't by accident. They built mockups in a warehouse and ran people through them until people loved it. The retail pays attention to customer satisfaction first, not squeezing every last penny,<mask> people love to shop there. They don't just throw products in boxes. They conceive a whole unboxing experience and rigorously test it on people. Everything, all the way through, tested for a positive user experience. [NEWLINE] [NEWLINE] And one more bit of genius, making the industrial designer of the products a senior VP with influence over tech and manufacturing<mask> it all works together. Others had lower-level design positions that work basically alone with no influence over anything else.<mask><mask> the functional and aesthetic design is always right in people's faces, it needs priority.</s>
Label encoding: <s>Small factoid: Jobs is actually on several Apple patents. [NEWLINE] [NEWLINE] He brought what other electronics makers didn't, an ocsessive-compulsive approach to making the entire experience perfect. The Apple Store format that MS and Samsung are copying wasn't by accident. They built mockups in a warehouse and ran people through them until people loved it. The retail pays attention to customer satisfaction first, not squeezing every last penny, so people love to shop there. They don't just throw products in boxes. They conceive a whole unboxing experience and rigorously test it on people. Everything, all the way through, tested for a positive user experience. [NEWLINE] [NEWLINE] And one more bit of genius, making the industrial designer of the products a senior VP with influence over tech and manufacturing so it all works together. Others had lower-level design positions that work basically alone with no influence over anything else. Given that the functional and aesthetic design is always right in people's faces, it needs priority.</s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Until you get caught up in a paradoxical loop of Dunning-Kruger self awareness! Ie: [NEWLINE] [NEWLINE] * I recognize the limitations of my niceness. I'm cognizant of all my weaknesses and flaws and know that I'm not always a good person. [NEWLINE] *<mask> the Dunning-Kruger effect states that my humility in this matter actually indicates that I'm a champion of righteousness. [NEWLINE] *<mask> I understand and agree with the principle behind the Dunning-Kruger effect, I now believe I am a super nice, super good person. [NEWLINE] *<mask><mask> the Dunning-Kruger effect,<mask> I believe I'm a super nice, super good person, I'm probably not.  I'm probably vastly over estimating niceness and goodness. [NEWLINE] *<mask> I understand and agree with the principle behind the Dunning-Kruger effect, I recognize this and become more humble about my goodness... [NEWLINE] [NEWLINE]...on and on!</s>
Label encoding: <s>Until you get caught up in a paradoxical loop of Dunning-Kruger self awareness! Ie: [NEWLINE] [NEWLINE] * I recognize the limitations of my niceness. I'm cognizant of all my weaknesses and flaws and know that I'm not always a good person. [NEWLINE] * But the Dunning-Kruger effect states that my humility in this matter actually indicates that I'm a champion of righteousness. [NEWLINE] * Because I understand and agree with the principle behind the Dunning-Kruger effect, I now believe I am a super nice, super good person. [NEWLINE] * According to the Dunning-Kruger effect, because I believe I'm a super nice, super good person, I'm probably not.  I'm probably vastly over estimating niceness and goodness. [NEWLINE] * Because I understand and agree with the principle behind the Dunning-Kruger effect, I recognize this and become more humble about my goodness... [NEWLINE] [NEWLINE]...on and on!</s>
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the root of this problem is age. You keep talking about "Western Culture"<mask> I feel<mask><mask> you are being overly broad. The only groups I can think of that encourage *heavy* drinking are 18-22 year olds. College kids. Children. This isn't really an excuse,<mask> an explanation.<mask><mask> that it is a pretty stupid thing for people to base their existence and pride on,<mask> *children* are idiots. Society at large does endorse drinking, sure,<mask> not nearly to the extent that you seem to be suggesting. It is a social lubricant, to be used in moderation, and we *do* shame and reject those that drink too much. [NEWLINE] [NEWLINE] I do take contention with your blanket statement that alcohol necessarily harms those drinking it and those around it. Plenty of people are capable of drinking a little, and stopping at appropriate times. I feel like maybe again you are looking at a very specific age group and applying it too broadly.</s>
Label encoding: <s>I think the root of this problem is age. You keep talking about "Western Culture" but I feel as though you are being overly broad. The only groups I can think of that encourage *heavy* drinking are 18-22 year olds. College kids. Children. This isn't really an excuse, but an explanation. I agree that it is a pretty stupid thing for people to base their existence and pride on, but *children* are idiots. Society at large does endorse drinking, sure, but not nearly to the extent that you seem to be suggesting. It is a social lubricant, to be used in moderation, and we *do* shame and reject those that drink too much. [NEWLINE] [NEWLINE] I do take contention with your blanket statement that alcohol necessarily harms those drinking it and those around it. Plenty of people are capable of drinking a little, and stopping at appropriate times. I feel like maybe again you are looking at a very specific age group and applying it too broadly.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That has been proven to be inherent to human psychology, not a subjective thing. [ENDQ] [NEWLINE] These are common generalizations across all of human genetics,<mask> there are certainly some outstanding and subjective ways to look at beauty. Some people like big breasts whereas some like small, some like being dominant in a relationship or otherwise submissive, some people like blondes and others brunettes... personally I'm not a huge fan of gauges<mask> I like "emos" and blue hair and the dark eyeliner and clothes, I'm not picky about bust size...<mask> that's subjective. There's plenty of people who don't like a more gothic style and prefer traditional things....<mask><mask> there is a level objective generalizations that can be made such<mask> someone being covered in puss and wounds is disturbing<mask> your instincts tell you to fear getting ill and potentially dying yourself, beyond those baser instincts to seek someone fruitful for reproduction is a wide range of personal tastes. </s>
Label encoding: <s> [STARTQ] That has been proven to be inherent to human psychology, not a subjective thing. [ENDQ] [NEWLINE] These are common generalizations across all of human genetics, but there are certainly some outstanding and subjective ways to look at beauty. Some people like big breasts whereas some like small, some like being dominant in a relationship or otherwise submissive, some people like blondes and others brunettes... personally I'm not a huge fan of gauges but I like "emos" and blue hair and the dark eyeliner and clothes, I'm not picky about bust size... but that's subjective. There's plenty of people who don't like a more gothic style and prefer traditional things.... so while there is a level objective generalizations that can be made such as someone being covered in puss and wounds is disturbing because your instincts tell you to fear getting ill and potentially dying yourself, beyond those baser instincts to seek someone fruitful for reproduction is a wide range of personal tastes. </s>
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That comment is like the mother of all Gish Gallops.  There is no way anyone could reasonably be expected to engage with that much content in a meaningful or thorough manner in this medium<mask> each video could probably be its own thread.  You say you're not a full-blown "conspiritard"<mask> that comment looks like every conversation I've ever had with one. [NEWLINE] [NEWLINE] Bottom line on this is that the evidence should be able to speak for itself and the process of pouring over dozens of youtube videos and feeling convinced is not good evidence.  Especially<mask> most of those videos usually amount to little more than anomaly hunting.  I doubt<mask> I'm saying will change anyone's view on the matter<mask> for the love of all that's good,<mask> you expect people to engage with you to this extent, at least have the courtesy to spend the time to make your own argument.  You haven't even spent the time to make your own youtube link compilation </s>
Label encoding: <s>That comment is like the mother of all Gish Gallops.  There is no way anyone could reasonably be expected to engage with that much content in a meaningful or thorough manner in this medium when each video could probably be its own thread.  You say you're not a full-blown "conspiritard" but that comment looks like every conversation I've ever had with one. [NEWLINE] [NEWLINE] Bottom line on this is that the evidence should be able to speak for itself and the process of pouring over dozens of youtube videos and feeling convinced is not good evidence.  Especially when most of those videos usually amount to little more than anomaly hunting.  I doubt what I'm saying will change anyone's view on the matter but for the love of all that's good, if you expect people to engage with you to this extent, at least have the courtesy to spend the time to make your own argument.  You haven't even spent the time to make your own youtube link compilation </s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ah, the good old no true feminist fallacy. [ENDQ] [NEWLINE] Way to misuse the "No true Scotsman" fallacy.  That's not<mask> it work. <mask> I was using that fallacy, I would have said these women aren't real feminists<mask> reasons A, B, and C.  Instead, I simply suggested that one group is not necessarily indicative of the whole.  Learn your fallacies. [NEWLINE] [NEWLINE] [STARTQ] <mask> the point is to draw attention, then they're failing spectacularly, apparently. They get boob pics in the papers, sure,<mask> it doesn't help activism in Russia. [ENDQ] [NEWLINE] Really?<mask> I just spent a whole day talking about the issue, its been on all the major news sites, people who actually read the article and didnt just dismiss it are now informed.  I would say it worked well. [NEWLINE] [NEWLINE] [STARTQ] <mask> their own demands may be entirely trivial. [ENDQ] [NEWLINE] Having your country invaded and annexed is trivial, is it?</s>
Label encoding: <s> [STARTQ] Ah, the good old no true feminist fallacy. [ENDQ] [NEWLINE] Way to misuse the "No true Scotsman" fallacy.  That's not how it work.  If I was using that fallacy, I would have said these women aren't real feminists because reasons A, B, and C.  Instead, I simply suggested that one group is not necessarily indicative of the whole.  Learn your fallacies. [NEWLINE] [NEWLINE] [STARTQ] If the point is to draw attention, then they're failing spectacularly, apparently. They get boob pics in the papers, sure, but it doesn't help activism in Russia. [ENDQ] [NEWLINE] Really? Because I just spent a whole day talking about the issue, its been on all the major news sites, people who actually read the article and didnt just dismiss it are now informed.  I would say it worked well. [NEWLINE] [NEWLINE] [STARTQ] While their own demands may be entirely trivial. [ENDQ] [NEWLINE] Having your country invaded and annexed is trivial, is it?</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0977, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Banning homosexuality means that half of the human species cannot be your sexual partner. [ENDQ] [NEWLINE] Uh no,<mask> half of the species are not gay. [NEWLINE] [NEWLINE] [STARTQ] Banning incests only bans you from, a couple dozen people at the most and really less than that. [ENDQ] [NEWLINE] <mask><mask> there are only two homosexuals left on earth, using your logic, gay marriage should be illegal? You are just inventing arbitrary limits to push your agenda. [NEWLINE] [NEWLINE] [STARTQ] Incest actually DOES screw with the family<mask> an institution. [ENDQ] [NEWLINE] <mask> does gay marriage. The family<mask> an institution is a FATHER and a MOTHER and children. [NEWLINE] [NEWLINE] [STARTQ] The obvious problems of consent. [ENDQ] [NEWLINE] We are talking about ADULTS. Once again you just invent an issue to push your agenda. Are you saying two consenting adults should be allowed to share their love?<mask> you say no, then gay marriage should be illegal<mask> there is no other reason for gay marriage. </s>
Label encoding: <s> [STARTQ] Banning homosexuality means that half of the human species cannot be your sexual partner. [ENDQ] [NEWLINE] Uh no, because half of the species are not gay. [NEWLINE] [NEWLINE] [STARTQ] Banning incests only bans you from, a couple dozen people at the most and really less than that. [ENDQ] [NEWLINE] So if there are only two homosexuals left on earth, using your logic, gay marriage should be illegal? You are just inventing arbitrary limits to push your agenda. [NEWLINE] [NEWLINE] [STARTQ] Incest actually DOES screw with the family as an institution. [ENDQ] [NEWLINE] So does gay marriage. The family as an institution is a FATHER and a MOTHER and children. [NEWLINE] [NEWLINE] [STARTQ] The obvious problems of consent. [ENDQ] [NEWLINE] We are talking about ADULTS. Once again you just invent an issue to push your agenda. Are you saying two consenting adults should be allowed to share their love? If you say no, then gay marriage should be illegal as there is no other reason for gay marriage. </s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This all feels pretty'scammy' to me,<mask><mask> there was no malice involved. There are many instances like this<mask> the school just seems more focused on getting more money from me than they already have. [ENDQ] [NEWLINE] <mask><mask> that's a straigh-up scam... by the book publishers. That has little to nothing to do with the professors. [NEWLINE] [NEWLINE] [STARTQ] <mask> can't I take a placement exam and skip these classes? [ENDQ] [NEWLINE] Don't know... have you asked?<mask> even<mask> you can't, the classes will be a breeze anyway right? Really, you can't expect most 100 level classes to be that great<mask> they're made for the lowest common denominator. The good stuff is usually going to be 200 300 and above. All stuff you can easily verify by looking at your school schedule and talking to students who are in those classes.<mask><mask> talking to other students is a great way to know which classes and profs to avoid.</s>
Label encoding: <s> [STARTQ] This all feels pretty'scammy' to me, even though there was no malice involved. There are many instances like this where the school just seems more focused on getting more money from me than they already have. [ENDQ] [NEWLINE] I agree that's a straigh-up scam... by the book publishers. That has little to nothing to do with the professors. [NEWLINE] [NEWLINE] [STARTQ] why can't I take a placement exam and skip these classes? [ENDQ] [NEWLINE] Don't know... have you asked? But even if you can't, the classes will be a breeze anyway right? Really, you can't expect most 100 level classes to be that great when they're made for the lowest common denominator. The good stuff is usually going to be 200 300 and above. All stuff you can easily verify by looking at your school schedule and talking to students who are in those classes. In fact talking to other students is a great way to know which classes and profs to avoid.</s>
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Examples: [NEWLINE] [NEWLINE] * We have 7 cutting boards and 5 colanders of differing sizes.  The largest of each of these basically requires its own shelf in the dishwasher.   I generally use one each per meal, and not the largest one, by re-using it at different stages of cooking.  He generally uses *at least* the largest of each, and sometimes some additional ones. [NEWLINE] [NEWLINE] * I generally take care to avoid splatters on the stovetop (e.g. by covering pots), he does not.  About half the time I make it through cooking without any splatters, he makes it through about 10% of the time. [NEWLINE] [NEWLINE] * We have a blender and 2 food processors of different sizes.  I will generally use only one of these, re-using it for different stages<mask> necessary.  He will often use 2 or all 3 for a single preparation<mask> he doesn't have to clean out the previous stage.</s><pad>
Label encoding: <s>Examples: [NEWLINE] [NEWLINE] * We have 7 cutting boards and 5 colanders of differing sizes.  The largest of each of these basically requires its own shelf in the dishwasher.   I generally use one each per meal, and not the largest one, by re-using it at different stages of cooking.  He generally uses *at least* the largest of each, and sometimes some additional ones. [NEWLINE] [NEWLINE] * I generally take care to avoid splatters on the stovetop (e.g. by covering pots), he does not.  About half the time I make it through cooking without any splatters, he makes it through about 10% of the time. [NEWLINE] [NEWLINE] * We have a blender and 2 food processors of different sizes.  I will generally use only one of these, re-using it for different stages as necessary.  He will often use 2 or all 3 for a single preparation so he doesn't have to clean out the previous stage.</s><pad>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0660, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not sure I believe your short answer.  No market (literally) without the hft means that somewhere, someone is making something that no one wants to buy. For a small fee, the hft will hide that little inefficiency from the rest of us.  Huh? [NEWLINE] [NEWLINE] And you keep descrbing a situation<mask> the hft twists the real market and then deny that the real market is twisted.<mask> I want to sell something for $10 and someone out there only wants to pay $8 for it, who is it that benefits<mask> the two of us aren't able to decide<mask> the real price should be?  I mean, I guess we know the exchange does.  That doesn't do much to make me happy. I know, you will make some comment about all the people trading which will force the hft to play honest or risk losing a ton. <mask> that would sort of contradict the short answer, wouldn't it?</s><pad>
Label encoding: <s>Not sure I believe your short answer.  No market (literally) without the hft means that somewhere, someone is making something that no one wants to buy. For a small fee, the hft will hide that little inefficiency from the rest of us.  Huh? [NEWLINE] [NEWLINE] And you keep descrbing a situation where the hft twists the real market and then deny that the real market is twisted. If I want to sell something for $10 and someone out there only wants to pay $8 for it, who is it that benefits if the two of us aren't able to decide what the real price should be?  I mean, I guess we know the exchange does.  That doesn't do much to make me happy. I know, you will make some comment about all the people trading which will force the hft to play honest or risk losing a ton.  But that would sort of contradict the short answer, wouldn't it?</s><pad>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We hold people of that profession to high standards.<mask> do cops get a pass currently? You're only defense<mask> far has been that is hard to do and it would dive some people out of the field or make it harder for new police to be hired. Both those problems are fine<mask> you have police who conduct themselves across the board with high standards. [NEWLINE] [NEWLINE] Just<mask> it is expensive or difficult to do<mask> isn't a valid reason to just let things slide. Your rejections of evidence on body cameras is odd<mask> well. They have been proven to reduce complaints. you can't just dismiss that out of hand<mask> it a hole in your argument. [NEWLINE] [NEWLINE] It is seems like you're making a lot of stretches. You main objections can be solved by simple economics or aren't really supported by facts. At this point, I fail to see the basis for you two main arguments,<mask> I fail to see<mask> ground you're standing on. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>We hold people of that profession to high standards. Why do cops get a pass currently? You're only defense so far has been that is hard to do and it would dive some people out of the field or make it harder for new police to be hired. Both those problems are fine if you have police who conduct themselves across the board with high standards. [NEWLINE] [NEWLINE] Just because it is expensive or difficult to do so isn't a valid reason to just let things slide. Your rejections of evidence on body cameras is odd as well. They have been proven to reduce complaints. you can't just dismiss that out of hand because it a hole in your argument. [NEWLINE] [NEWLINE] It is seems like you're making a lot of stretches. You main objections can be solved by simple economics or aren't really supported by facts. At this point, I fail to see the basis for you two main arguments, so I fail to see what ground you're standing on. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I can grant you that the intent may not be malicious from the performers' point of view, it doesn't stop it from being damaging by<mask> others interpret it<mask>.  A lot of gender performance plays into the idea that it is absurd to see someone who is assigned a certain gender at birth behave in a way that is associated with another gender.  Like "This wouldn't be funny<mask> it weren't for trans people". It's not limited to just trans issues either- other stereotypes are played for laughs too, without explicitly saying<mask> they're "funny". [NEWLINE] [NEWLINE] I could be reading you wrong,<mask> I feel like in your third paragraph you're drawing from the idea that trans people are gay, which is not necessarily true. [NEWLINE] [NEWLINE] I do take most of your points well<mask>, and you're close to changing my view especially with the last bit about being dirty stepping outside gender roles.  Can you clarify your position a little more?</s>
Label encoding: <s>While I can grant you that the intent may not be malicious from the performers' point of view, it doesn't stop it from being damaging by what others interpret it as.  A lot of gender performance plays into the idea that it is absurd to see someone who is assigned a certain gender at birth behave in a way that is associated with another gender.  Like "This wouldn't be funny if it weren't for trans people". It's not limited to just trans issues either- other stereotypes are played for laughs too, without explicitly saying why they're "funny". [NEWLINE] [NEWLINE] I could be reading you wrong, but I feel like in your third paragraph you're drawing from the idea that trans people are gay, which is not necessarily true. [NEWLINE] [NEWLINE] I do take most of your points well though, and you're close to changing my view especially with the last bit about being dirty stepping outside gender roles.  Can you clarify your position a little more?</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm not convinced this is the case, or even<mask> there is a clear distinction. [ENDQ] [NEWLINE] <mask><mask> that it's a grey area. [NEWLINE] [NEWLINE] [STARTQ] MW definition is far too broad. [ENDQ] [NEWLINE] Well, it's better than the Google def: "the use of violence and intimidation in the pursuit of political aims." [NEWLINE] [NEWLINE] <mask> not<mask> good<mask> the Wikipedia comment: [NEWLINE] [STARTQ] Common definitions of terrorism refer only to those violent acts that are intended to create fear (terror); are perpetrated for a religious, political, or ideological goal; and deliberately target or disregard the safety of non-combatants (e.g., neutral military personnel or civilians). [ENDQ] [NEWLINE] <mask>,<mask><mask> with you in theory -<mask> dictionaries generally look for simple one-liners that get the general point across,<mask> don't go much into the nuance.  (And<mask> the Wikipedia article points out, there isn't a fixed definition, which makes it harder).</s>
Label encoding: <s> [STARTQ] I'm not convinced this is the case, or even if there is a clear distinction. [ENDQ] [NEWLINE] I agree that it's a grey area. [NEWLINE] [NEWLINE] [STARTQ] MW definition is far too broad. [ENDQ] [NEWLINE] Well, it's better than the Google def: "the use of violence and intimidation in the pursuit of political aims." [NEWLINE] [NEWLINE] But not as good as the Wikipedia comment: [NEWLINE] [STARTQ] Common definitions of terrorism refer only to those violent acts that are intended to create fear (terror); are perpetrated for a religious, political, or ideological goal; and deliberately target or disregard the safety of non-combatants (e.g., neutral military personnel or civilians). [ENDQ] [NEWLINE] So, I agree with you in theory - but dictionaries generally look for simple one-liners that get the general point across, but don't go much into the nuance.  (And as the Wikipedia article points out, there isn't a fixed definition, which makes it harder).</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you familiar with the concept of the [Uncanny Valley?]( [URL].svg) Essentially, something that appears close to human,<mask> slightly off in some way will appear<mask> increasingly unnerving to people. ([See here for a demonstration.]( [URL] )) [NEWLINE] [NEWLINE] Now, think about clowns with this in mind. They look generally human,<mask> there are things that are just...not quite right, by the standards of a normal person. They can have false eyebrow markings over there real ones, which makes the clown's true emotive state difficult to read. They have a big, garish smile drawn over their real mouth, which again may or may not correspond with their true expression. [NEWLINE] [NEWLINE] Clowns by definition try to display an exaggerated sense of comical happiness,<mask> this can come off<mask> unnerving<mask> it is something artificial, and not necessarily demonstrative of the emotional state of the human being under the makeup. </s>
Label encoding: <s>Are you familiar with the concept of the [Uncanny Valley?]( [URL].svg) Essentially, something that appears close to human, but slightly off in some way will appear as increasingly unnerving to people. ([See here for a demonstration.]( [URL] )) [NEWLINE] [NEWLINE] Now, think about clowns with this in mind. They look generally human, but there are things that are just...not quite right, by the standards of a normal person. They can have false eyebrow markings over there real ones, which makes the clown's true emotive state difficult to read. They have a big, garish smile drawn over their real mouth, which again may or may not correspond with their true expression. [NEWLINE] [NEWLINE] Clowns by definition try to display an exaggerated sense of comical happiness, but this can come off as unnerving because it is something artificial, and not necessarily demonstrative of the emotional state of the human being under the makeup. </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll say this<mask> a envious European; the American Government has amazing fiscal policy. I mean, remember the international opinion around 2008? The credit crisis was supposed to crush you, either the banks would go down, or the dollar was going to massively inflate, causing foreign holders to swap out the dollars and decrease it's value even more. Neither have happened, and government bond interest rates on the dollar are still lower than inflation, which means people are effectively paying your government just have dollars. [NEWLINE] [NEWLINE] <mask>, in response to the credit crisis, the ECB didn't "print" any money,<mask> the Germans are more inflation averse than you are. This means that the bad loans from the banks which have been bailed out, have to be written off not through the sovereign power of the state (<mask> the EU isn't a sovereign)<mask> the money has had to come out of the real economy, which is stagnating for five years now.</s>
Label encoding: <s>I'll say this as a envious European; the American Government has amazing fiscal policy. I mean, remember the international opinion around 2008? The credit crisis was supposed to crush you, either the banks would go down, or the dollar was going to massively inflate, causing foreign holders to swap out the dollars and decrease it's value even more. Neither have happened, and government bond interest rates on the dollar are still lower than inflation, which means people are effectively paying your government just have dollars. [NEWLINE] [NEWLINE] Also, in response to the credit crisis, the ECB didn't "print" any money, because the Germans are more inflation averse than you are. This means that the bad loans from the banks which have been bailed out, have to be written off not through the sovereign power of the state ( because the EU isn't a sovereign) but the money has had to come out of the real economy, which is stagnating for five years now.</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All good points,<mask><mask><mask> those relate more directly to the human _body_, they don't necessarily relate more directly to the human experience. It's easier for me to think of water boiling<mask> a concrete, measurable thing than it is for me to think about<mask> temperatures I can survive at. I have no idea<mask> temperatures I can survive at.<mask><mask>,<mask> you said, it's a vague number that is not 0F or 100F. Even<mask> we invented a new measurement in which 100F was the magic amount that was just enough to kill anyone, it would be much less universally understandable than the boiling point of water. Boiling water is a thing that everyone has experience with. Dying is not. [NEWLINE] [NEWLINE] (After all that, I'd like to provide the disclaimer that I currently prefer F to C for the terribly subjective, personal reason that I'm used to it. I'm arguing with your rationale, not anything else.)</s>
Label encoding: <s>All good points, but even though those relate more directly to the human _body_, they don't necessarily relate more directly to the human experience. It's easier for me to think of water boiling as a concrete, measurable thing than it is for me to think about what temperatures I can survive at. I have no idea what temperatures I can survive at. In fact, as you said, it's a vague number that is not 0F or 100F. Even if we invented a new measurement in which 100F was the magic amount that was just enough to kill anyone, it would be much less universally understandable than the boiling point of water. Boiling water is a thing that everyone has experience with. Dying is not. [NEWLINE] [NEWLINE] (After all that, I'd like to provide the disclaimer that I currently prefer F to C for the terribly subjective, personal reason that I'm used to it. I'm arguing with your rationale, not anything else.)</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off I have to admit that I have a huge bias. I am far better at tests than I am at projects. [NEWLINE] [NEWLINE] <mask> the reason for this is that I have the exact opposite reaction to these types of things. I am never stressed in a test,<mask> often get extremely anxious and extremely stressed over projects. In these projects I often don't preform to my top measure<mask><mask><mask> stress,<mask> during a test I do preform my best. [NEWLINE] [NEWLINE] Now I'd<mask><mask><mask> most jobs are more test based. Other than engineering you usually have to do a lot of work on the spot. Being able to think quickly and memorizing material is important for<mask> you are interacting with clients or customers. [NEWLINE] [NEWLINE] Personally<mask><mask> both are important<mask> most people seem to be good at one or the other.<mask><mask><mask><mask><mask> that it would be best to give students to option to either do a project or a test.</s>
Label encoding: <s>First off I have to admit that I have a huge bias. I am far better at tests than I am at projects. [NEWLINE] [NEWLINE] But the reason for this is that I have the exact opposite reaction to these types of things. I am never stressed in a test, but often get extremely anxious and extremely stressed over projects. In these projects I often don't preform to my top measure because of this stress, but during a test I do preform my best. [NEWLINE] [NEWLINE] Now I'd also argue that most jobs are more test based. Other than engineering you usually have to do a lot of work on the spot. Being able to think quickly and memorizing material is important for when you are interacting with clients or customers. [NEWLINE] [NEWLINE] Personally I think both are important but most people seem to be good at one or the other. Because of this I think that it would be best to give students to option to either do a project or a test.</s>
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is old,<mask> I have a shorter interpretation. For the sake of argument, screaming is bad and tasteless. Its like farting into the microphone. [NEWLINE] [NEWLINE] Some people are going to be really good at farting into the microphone. Maybe I have extreme range and do deep lows,<mask> still being able to do really nice highs. Maybe I can fart really legible and fast, almost like rap except farts. Ive been farting into microphones for a decade and have honed my skill to a level most cant, even<mask> to you it just sounds like more farts. There's still a skill to this farting method, and even<mask> its a dumb fad, there still is an art form to it. [NEWLINE] [NEWLINE] Now, take away the farts and pretend its distortion. Just like guitar, its like normal vocals<mask> a little heavier. Most screaming has some distortion done in post<mask> thats a different story. </s>
Label encoding: <s>This is old, but I have a shorter interpretation. For the sake of argument, screaming is bad and tasteless. Its like farting into the microphone. [NEWLINE] [NEWLINE] Some people are going to be really good at farting into the microphone. Maybe I have extreme range and do deep lows, while still being able to do really nice highs. Maybe I can fart really legible and fast, almost like rap except farts. Ive been farting into microphones for a decade and have honed my skill to a level most cant, even if to you it just sounds like more farts. There's still a skill to this farting method, and even if its a dumb fad, there still is an art form to it. [NEWLINE] [NEWLINE] Now, take away the farts and pretend its distortion. Just like guitar, its like normal vocals but a little heavier. Most screaming has some distortion done in post but thats a different story. </s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry for your lost. [NEWLINE] [NEWLINE] I'd say that we all<mask> people aren't completely conscious of our actions. Sure you may have thought that you showed no emotion<mask> in reality your body language showed that you were grieving. Some indications would be showing [microexpression]( [URL] ) of sadness<mask> you talked. Your posture may have shown vulnerability depending on<mask> you felt. You could<mask> look at your relatives around you who are more emotional and easier to read<mask> being able to come to a conclusion. [NEWLINE] [NEWLINE] Overall, someone who is good at reading body language would have been able to come to the conclusion that<mask> this 15 year old boy is showing multiple signs of sadness and is dressed very nicely, he very well may be going to a funeral. [NEWLINE] [NEWLINE] Being able to tell<mask> it was your mother who passed away would be another story. We would<mask> be able to infer that whoever did pass away was very close to you.</s>
Label encoding: <s>I'm sorry for your lost. [NEWLINE] [NEWLINE] I'd say that we all as people aren't completely conscious of our actions. Sure you may have thought that you showed no emotion when in reality your body language showed that you were grieving. Some indications would be showing [microexpression]( [URL] ) of sadness when you talked. Your posture may have shown vulnerability depending on how you felt. You could also look at your relatives around you who are more emotional and easier to read thus being able to come to a conclusion. [NEWLINE] [NEWLINE] Overall, someone who is good at reading body language would have been able to come to the conclusion that because this 15 year old boy is showing multiple signs of sadness and is dressed very nicely, he very well may be going to a funeral. [NEWLINE] [NEWLINE] Being able to tell if it was your mother who passed away would be another story. We would although be able to infer that whoever did pass away was very close to you.</s>
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Well-done red meat [linked to aggressive prostate cancer]( [URL] /)" [NEWLINE] [NEWLINE] [STARTQ] "This is another piece of evidence for the notion that red meat, particularly grilled meat, contains carcinogens that may relate to prostate cancer," says Ronald D. Ennis, M.D., director of radiation oncology at St. Luke's -- Roosevelt Hospital Center, in New York City, who was not involved in the study. [ENDQ] [NEWLINE] [STARTQ] A new study has found that men have a higher risk of developing aggressive prostate cancer<mask> they consume a lot of ground beef and other red meat -- especially<mask> the meat is grilled or well-done [ENDQ] [NEWLINE] Raw meat, [is very healthy]( [URL] ).<mask>, cooked meat is always bad for your health. Well done meat is even worse for you. [NEWLINE] [NEWLINE] &gt; Inherently, there's not a reason that eating raw meat in and of itself would automatically create any problems.</s>
Label encoding: <s>"Well-done red meat [linked to aggressive prostate cancer]( [URL] /)" [NEWLINE] [NEWLINE] [STARTQ] "This is another piece of evidence for the notion that red meat, particularly grilled meat, contains carcinogens that may relate to prostate cancer," says Ronald D. Ennis, M.D., director of radiation oncology at St. Luke's -- Roosevelt Hospital Center, in New York City, who was not involved in the study. [ENDQ] [NEWLINE] [STARTQ] A new study has found that men have a higher risk of developing aggressive prostate cancer if they consume a lot of ground beef and other red meat -- especially if the meat is grilled or well-done [ENDQ] [NEWLINE] Raw meat, [is very healthy]( [URL] ). However, cooked meat is always bad for your health. Well done meat is even worse for you. [NEWLINE] [NEWLINE] &gt; Inherently, there's not a reason that eating raw meat in and of itself would automatically create any problems.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You still need to edit it. Get it into a format usable by plenty of ereaders. Host that content. Make it available for customers to buy. Update the format/hosting market availability<mask> technology changes. Care for those customers. Have access to lawyers. And all kinds of other things. Ebooks are cheaper to print/distribute,<mask> probably just<mask> complicated<mask> physical books. [NEWLINE] [NEWLINE] All of this is stuff that the author needs to have happen<mask> maybe can't do himself, and almost certainly doesn't want to put time/energy into himself.<mask>, the need for publishers. [NEWLINE] [NEWLINE] <mask> for marketing in particular: of course that is<mask> a lot of it consists of.<mask><mask> has marketing for *anything* been largely based on making reasonable arguments and discussing philosophical merits of the product? Marketing sucks, and is a drain.<mask> unfortunately, the author isn't going to sell many books without it. </s><pad>
Label encoding: <s>You still need to edit it. Get it into a format usable by plenty of ereaders. Host that content. Make it available for customers to buy. Update the format/hosting market availability as technology changes. Care for those customers. Have access to lawyers. And all kinds of other things. Ebooks are cheaper to print/distribute, but probably just as complicated as physical books. [NEWLINE] [NEWLINE] All of this is stuff that the author needs to have happen but maybe can't do himself, and almost certainly doesn't want to put time/energy into himself. Hence, the need for publishers. [NEWLINE] [NEWLINE] As for marketing in particular: of course that is what a lot of it consists of. Since when has marketing for *anything* been largely based on making reasonable arguments and discussing philosophical merits of the product? Marketing sucks, and is a drain. But unfortunately, the author isn't going to sell many books without it. </s><pad>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's<mask> I differ with you on things, I suppose.<mask><mask> the racism is more of a function of classism than just pure racial hatred. I feel like without a unified lower class, all the racial awareness and progress still hits a wall<mask> such a small minority tries to fight on it's own.<mask>, the poor are the majority these days,<mask> severely fragmented in to racial divides that only serve the rich. [NEWLINE] [NEWLINE] I'm not going to say I'm more right than anyone else, it's just<mask><mask>,<mask> I feel like the poor are dividing themselves to be conquered, no work even needs to be done by those who benefit from it, we're doing it for them by focusing on our individual grievances... (I say we<mask> I'm a poverty class white man). [NEWLINE] [NEWLINE] I still say I have more in common with a poor black man than a poor black man does with a rich black man. </s><pad>
Label encoding: <s>That's where I differ with you on things, I suppose. I think the racism is more of a function of classism than just pure racial hatred. I feel like without a unified lower class, all the racial awareness and progress still hits a wall when such a small minority tries to fight on it's own. Meanwhile, the poor are the majority these days, but severely fragmented in to racial divides that only serve the rich. [NEWLINE] [NEWLINE] I'm not going to say I'm more right than anyone else, it's just my opinion, but I feel like the poor are dividing themselves to be conquered, no work even needs to be done by those who benefit from it, we're doing it for them by focusing on our individual grievances... (I say we because I'm a poverty class white man). [NEWLINE] [NEWLINE] I still say I have more in common with a poor black man than a poor black man does with a rich black man. </s><pad>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I wonder to<mask> degree this has to do with past, not present inequality. [ENDQ] [NEWLINE] That could be very true,<mask> the past does help shape the present.<mask> we still see more men at the top, we are conditioned to think of high-level positions<mask> a man's place. [NEWLINE] [NEWLINE] It is going to take a<mask> to fix, I certainly won't deny that. I just have a problem with the argument that some people here are making that simply<mask> there are now individual protections in place, the problem ceases to exist. Without true, structural change, we really haven't done anything. [NEWLINE] [NEWLINE] <mask>, I'm absolutely not blaming "Men." For the vast majority of men, they have no say whatsoever in<mask> a woman makes--they are just<mask> dis-empowered<mask> anyone else.<mask><mask> the key distinction is that not all men have power,<mask> most of the people who do are men.</s>
Label encoding: <s> [STARTQ] I wonder to what degree this has to do with past, not present inequality. [ENDQ] [NEWLINE] That could be very true, but the past does help shape the present. If we still see more men at the top, we are conditioned to think of high-level positions as a man's place. [NEWLINE] [NEWLINE] It is going to take a while to fix, I certainly won't deny that. I just have a problem with the argument that some people here are making that simply because there are now individual protections in place, the problem ceases to exist. Without true, structural change, we really haven't done anything. [NEWLINE] [NEWLINE] Also, I'm absolutely not blaming "Men." For the vast majority of men, they have no say whatsoever in what a woman makes--they are just as dis-empowered as anyone else. I think the key distinction is that not all men have power, but most of the people who do are men.</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wow I know<mask> you mean. I first tried it<mask> I was 19 and smoked very regularly, even all-day for months on end, until I was 20. Then I moved to my dad's house, got caught, and had to seriously cut down to avoid getting kicked out. I still sneak it every now and then (only<mask> out of the house) and<mask> my tolerance is way lower. And recently, I was at a party and shared a blunt at the beginning. Well, I got way too high and suddenly felt really uncomfortable<mask> all the people around were strangers. And I was really afraid to talk to the girl I liked (whose birthday it was). I sobered up, and drank a little bit and smoked a few cigarettes and that really helped,<mask> I realized for once that maybe getting high isn't a great idea<mask> I'm trying to be social (beer is a much better choice in that situation!).</s>
Label encoding: <s>Wow I know what you mean. I first tried it when I was 19 and smoked very regularly, even all-day for months on end, until I was 20. Then I moved to my dad's house, got caught, and had to seriously cut down to avoid getting kicked out. I still sneak it every now and then (only when out of the house) and so my tolerance is way lower. And recently, I was at a party and shared a blunt at the beginning. Well, I got way too high and suddenly felt really uncomfortable since all the people around were strangers. And I was really afraid to talk to the girl I liked (whose birthday it was). I sobered up, and drank a little bit and smoked a few cigarettes and that really helped, but I realized for once that maybe getting high isn't a great idea if I'm trying to be social (beer is a much better choice in that situation!).</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0650, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You seem to be operating under the belief that admissions criteria are based on some attempt to provide merit-based education. This isn't really the case. [NEWLINE] [NEWLINE] Colleges want better students<mask> those students raise the average grade of their graduating classes. This makes the college look more desirable, allowing it to have higher tuition and still have full enrollment. [NEWLINE] [NEWLINE] Historically, Colleges have<mask> benefited from donations by alumni who have been very successful. This<mask> contributes to admissions requirements. [NEWLINE] [NEWLINE] Finally, certain Colleges must meet certain pass rates on third party exams (like the Bar exam) in order to remain accredited. [NEWLINE] [NEWLINE] The athlete contributes more to both prestige and the potential for grants than the average student who meets admissions requirements.<mask> the final point works against having a large pool of waived athletes, it doesn't really preclude their admission. [NEWLINE] [NEWLINE] Aside from these elements, is there really a compelling argument for having admissions requirements at all?</s>
Label encoding: <s>You seem to be operating under the belief that admissions criteria are based on some attempt to provide merit-based education. This isn't really the case. [NEWLINE] [NEWLINE] Colleges want better students because those students raise the average grade of their graduating classes. This makes the college look more desirable, allowing it to have higher tuition and still have full enrollment. [NEWLINE] [NEWLINE] Historically, Colleges have also benefited from donations by alumni who have been very successful. This also contributes to admissions requirements. [NEWLINE] [NEWLINE] Finally, certain Colleges must meet certain pass rates on third party exams (like the Bar exam) in order to remain accredited. [NEWLINE] [NEWLINE] The athlete contributes more to both prestige and the potential for grants than the average student who meets admissions requirements. While the final point works against having a large pool of waived athletes, it doesn't really preclude their admission. [NEWLINE] [NEWLINE] Aside from these elements, is there really a compelling argument for having admissions requirements at all?</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh good grief.<mask> a ridiculous comparison. [NEWLINE] [NEWLINE] Should a woman be essentially forced to do something with her body she may think abhorrent/is scared of/doesn't want to do that has physical and often mental side-effects or to give the child her body has been preparing her to care for and love away after she's carried it for 9 months and given birth simply<mask> a man has decided he doesn't want to pay for a baby? [NEWLINE] [NEWLINE] No. [NEWLINE] [NEWLINE] Is it fair on the man? The depends. It is her body that is affected. It is her that has to go through with any of the available options. It is her that has to raise a baby<mask> she keeps it in this situation. There is very little about this situation that is fair on either party.<mask><mask> it is reasonable the decision is hers and<mask><mask> it is reasonable to expect a father to provide for a child. </s>
Label encoding: <s>Oh good grief. What a ridiculous comparison. [NEWLINE] [NEWLINE] Should a woman be essentially forced to do something with her body she may think abhorrent/is scared of/doesn't want to do that has physical and often mental side-effects or to give the child her body has been preparing her to care for and love away after she's carried it for 9 months and given birth simply because a man has decided he doesn't want to pay for a baby? [NEWLINE] [NEWLINE] No. [NEWLINE] [NEWLINE] Is it fair on the man? The depends. It is her body that is affected. It is her that has to go through with any of the available options. It is her that has to raise a baby if she keeps it in this situation. There is very little about this situation that is fair on either party. I think it is reasonable the decision is hers and I think it is reasonable to expect a father to provide for a child. </s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Almost every student is eligible, and making all lunches free would let them do away with their administration of the free lunch program. [ENDQ] [NEWLINE] <mask> good<mask> this sounds, I believe it fails to take into account federal funding for free lunch programs. [NEWLINE] [NEWLINE] The cost of lunches (and breakfasts) received through the free lunch program is not absorbed by the individual school system,<mask> rather the school system receives federal funding for each child in the program to absorb the costs. Funding,<mask>, is based on registered participants in the program, and not the number of students in the school district who WOULD qualify<mask> they applied. [NEWLINE] [NEWLINE] <mask> it is true that your school district would SAVE money by providing free lunches to all students (and forgoing federal funds by not requiring them to sign up for the program) it honestly sounds like there is a MAJOR source of waste in their administration of the program.</s>
Label encoding: <s> [STARTQ] Almost every student is eligible, and making all lunches free would let them do away with their administration of the free lunch program. [ENDQ] [NEWLINE] As good as this sounds, I believe it fails to take into account federal funding for free lunch programs. [NEWLINE] [NEWLINE] The cost of lunches (and breakfasts) received through the free lunch program is not absorbed by the individual school system, but rather the school system receives federal funding for each child in the program to absorb the costs. Funding, however, is based on registered participants in the program, and not the number of students in the school district who WOULD qualify if they applied. [NEWLINE] [NEWLINE] If it is true that your school district would SAVE money by providing free lunches to all students (and forgoing federal funds by not requiring them to sign up for the program) it honestly sounds like there is a MAJOR source of waste in their administration of the program.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, perhaps no one should have children ever then,<mask> outside of the 1%, in a country<mask> insurance is tied to your job and you could lose it at any time, you can't always have enough in savings to pay for unexpected events. Which, of course, is<mask> we have insurance. [NEWLINE] [NEWLINE] Planning is having insurance that covers these things. And no, at the time I didn't have 30k in accessible (non-penalty) savings, which is<mask> my son's birth ended up costing. By that metric, its unlikely that most people have enough money for children. [NEWLINE] [NEWLINE] That's like saying don't buy a house unless you have enough money to pay for a replacement<mask> it burns down. The only difference is you can't lose your home insurance merely through losing your job. I had enough money to pay for private insurance,<mask> there wasn't' private insurance that would cover me. </s>
Label encoding: <s>Well, perhaps no one should have children ever then, because outside of the 1%, in a country where insurance is tied to your job and you could lose it at any time, you can't always have enough in savings to pay for unexpected events. Which, of course, is why we have insurance. [NEWLINE] [NEWLINE] Planning is having insurance that covers these things. And no, at the time I didn't have 30k in accessible (non-penalty) savings, which is what my son's birth ended up costing. By that metric, its unlikely that most people have enough money for children. [NEWLINE] [NEWLINE] That's like saying don't buy a house unless you have enough money to pay for a replacement if it burns down. The only difference is you can't lose your home insurance merely through losing your job. I had enough money to pay for private insurance, but there wasn't' private insurance that would cover me. </s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyday you drive by 100s of motorists who climbed in their cars knowing that a vehicle of that weight and speed can easily kill. Sure they aren't made to kill<mask><mask> machines don't have agency, that's hardly relevant. You don't know who among those drivers is not prone to use their car to kill and who's a moment away from jerking the wheel. Still, you don't walk around trembling at this idea. [NEWLINE] [NEWLINE] Furthermore, gun owners who carry don't plan to kill<mask>, should they NEED to use it to defend their lives or the lives of others, they've chosen the tool that is less likely to fail in that task (compared to a taser). I don't carry<mask> it's a pain in the ass and very unlikely to be useful.<mask> my firearm was tiny and light I'd carry<mask> I'd be crazy not to<mask> there was no discomfort or inconvenience. </s>
Label encoding: <s>Everyday you drive by 100s of motorists who climbed in their cars knowing that a vehicle of that weight and speed can easily kill. Sure they aren't made to kill but since machines don't have agency, that's hardly relevant. You don't know who among those drivers is not prone to use their car to kill and who's a moment away from jerking the wheel. Still, you don't walk around trembling at this idea. [NEWLINE] [NEWLINE] Furthermore, gun owners who carry don't plan to kill but, should they NEED to use it to defend their lives or the lives of others, they've chosen the tool that is less likely to fail in that task (compared to a taser). I don't carry because it's a pain in the ass and very unlikely to be useful. If my firearm was tiny and light I'd carry because I'd be crazy not to if there was no discomfort or inconvenience. </s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a better idea: [NEWLINE] [NEWLINE] You take the job with the construction company, work very hard and advance<mask> high<mask> you can go.  You then get into contact with the staff to whatever politicians are in charge of creating the regulations you have experience helping companies be in compliance with (could be local, state, federal).  You then offer your services<mask> an advisor to the mayor's office, or offer to testify at state legislative hearings on prospective environmental regulations.  "Yes Mr. Chairman, I have extensive experience with regulatory compliance and can say companies could meet the new standards at minimal cost." [NEWLINE] [NEWLINE] And in the best case scenario, actually get yourself hired on<mask> a regulator, and make the regulations not pathetic. [NEWLINE] [NEWLINE] Never underestimate the power of a small group of committed people flowing through the revolving door of corporate and regulatory bodies to change the world.<mask><mask>, it is the only thing that ever has.</s>
Label encoding: <s>I have a better idea: [NEWLINE] [NEWLINE] You take the job with the construction company, work very hard and advance as high as you can go.  You then get into contact with the staff to whatever politicians are in charge of creating the regulations you have experience helping companies be in compliance with (could be local, state, federal).  You then offer your services as an advisor to the mayor's office, or offer to testify at state legislative hearings on prospective environmental regulations.  "Yes Mr. Chairman, I have extensive experience with regulatory compliance and can say companies could meet the new standards at minimal cost." [NEWLINE] [NEWLINE] And in the best case scenario, actually get yourself hired on as a regulator, and make the regulations not pathetic. [NEWLINE] [NEWLINE] Never underestimate the power of a small group of committed people flowing through the revolving door of corporate and regulatory bodies to change the world. In fact, it is the only thing that ever has.</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Decides to have hot monkey sex and then can't remember it the next day [ENDQ] [NEWLINE] That's her problem. It was her choice. And honestly,<mask> it's left at that she can't even press charges, she doesn't know.<mask> the person she had monkey sex with then keeps talking about it, she would press charges for sexual harassment. [NEWLINE] [NEWLINE] [STARTQ] Gives consent,<mask> is too drunk to consider the ramifications [ENDQ] [NEWLINE] Her problem. [NEWLINE] [NEWLINE] [STARTQ] Is too drunk to know<mask>'s going on,<mask> just lays there<mask> the guy has his way with her [ENDQ] [NEWLINE] She didn't give consent. I should probably include in my main post that she<mask> has to say "yes" or something like that, which is already considered to be included in consent. [NEWLINE] [NEWLINE] [STARTQ] Is trying to fight the guy off,<mask> is too drunk to do<mask> [ENDQ] [NEWLINE] <mask> didn't give consent. Still rape.</s>
Label encoding: <s> [STARTQ] Decides to have hot monkey sex and then can't remember it the next day [ENDQ] [NEWLINE] That's her problem. It was her choice. And honestly, if it's left at that she can't even press charges, she doesn't know. If the person she had monkey sex with then keeps talking about it, she would press charges for sexual harassment. [NEWLINE] [NEWLINE] [STARTQ] Gives consent, but is too drunk to consider the ramifications [ENDQ] [NEWLINE] Her problem. [NEWLINE] [NEWLINE] [STARTQ] Is too drunk to know what's going on, so just lays there while the guy has his way with her [ENDQ] [NEWLINE] She didn't give consent. I should probably include in my main post that she also has to say "yes" or something like that, which is already considered to be included in consent. [NEWLINE] [NEWLINE] [STARTQ] Is trying to fight the guy off, but is too drunk to do so [ENDQ] [NEWLINE] Also didn't give consent. Still rape.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The only time I have ever seen that stuff be a problem is with reeeeally bitchy customers.<mask> you act nice about it and don't act like we are ruining your life by giving you the wrong food, you will be fine. [NEWLINE] [NEWLINE] <mask> your a server then you have to go back and tell the cook<mask> is wrong, generally this will pass without insident, its<mask> you come in with a horror story about someone being a bitch that everyone starts joking about spitting in it, or giving lower quality ingrediants, or even just taking your sweet time. [NEWLINE] [NEWLINE] <mask> for the manager thing, you can do that<mask> now your server thinks you're a dick,<mask> the server had the power to fix your food<mask> instead of asking them to do their job you go over their head and assume they will fuck it up. This<mask> makes the manager question<mask> the server acted rudly to you.</s>
Label encoding: <s>The only time I have ever seen that stuff be a problem is with reeeeally bitchy customers. If you act nice about it and don't act like we are ruining your life by giving you the wrong food, you will be fine. [NEWLINE] [NEWLINE] If your a server then you have to go back and tell the cook what is wrong, generally this will pass without insident, its when you come in with a horror story about someone being a bitch that everyone starts joking about spitting in it, or giving lower quality ingrediants, or even just taking your sweet time. [NEWLINE] [NEWLINE] As for the manager thing, you can do that but now your server thinks you're a dick, because the server had the power to fix your food but instead of asking them to do their job you go over their head and assume they will fuck it up. This also makes the manager question if the server acted rudly to you.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll just say to spare you and OP going back and forth... that's really, really strange. [NEWLINE] [NEWLINE] Like, not saying it's *wrong*. It's just really bizarre. [NEWLINE] [NEWLINE] There are people at my work whose relationships I know less about than others.<mask> I don't know anyone who just refuses to acknowledge that they have them<mask> "it's not relevant". [NEWLINE] [NEWLINE] I mostly just get the sense that you're trying to make yourself feel ok that you're uncomfortable with gay people. No idea<mask> that's true. Just<mask> you come across. [NEWLINE] [NEWLINE] Even<mask> you individually are very secretive about your personal life, just know that in general, people are talking about their boyfriends, girlfriends, husbands and wives all the time. And I can only imagine<mask> uncomfortable it would feel to have to bow out of every one of those conversations for risk of shame or reprimand. </s><pad><pad>
Label encoding: <s>I'll just say to spare you and OP going back and forth... that's really, really strange. [NEWLINE] [NEWLINE] Like, not saying it's *wrong*. It's just really bizarre. [NEWLINE] [NEWLINE] There are people at my work whose relationships I know less about than others. But I don't know anyone who just refuses to acknowledge that they have them because "it's not relevant". [NEWLINE] [NEWLINE] I mostly just get the sense that you're trying to make yourself feel ok that you're uncomfortable with gay people. No idea if that's true. Just how you come across. [NEWLINE] [NEWLINE] Even if you individually are very secretive about your personal life, just know that in general, people are talking about their boyfriends, girlfriends, husbands and wives all the time. And I can only imagine how uncomfortable it would feel to have to bow out of every one of those conversations for risk of shame or reprimand. </s><pad><pad>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because saying a joke is not morally wrong. I would scarcely call it an attack at all. Murder and poking fun cannot be compared, much less equated. Poking fun at someone is not 'violating peace'. The other person is free to disregard it or complain loudly<mask> he really wants,<mask>, to use an old adage, his right to swing his fist ends at another's face. The joke did absolutely no harm at all to another in any tangible way. [NEWLINE] [NEWLINE] And by your definition, isn't calling the police just another attack? I mean, we want peace right? We should just let him live his life and not bug him anymore. That's the MOST peaceful way. Except that won't stop him, and that's not justice. [NEWLINE] [NEWLINE] Now I'm not claiming drawing the prophet is justice. It is simply a demonstration that their acts are futile. A peaceful demonstration. </s>
Label encoding: <s>Because saying a joke is not morally wrong. I would scarcely call it an attack at all. Murder and poking fun cannot be compared, much less equated. Poking fun at someone is not 'violating peace'. The other person is free to disregard it or complain loudly if he really wants, but, to use an old adage, his right to swing his fist ends at another's face. The joke did absolutely no harm at all to another in any tangible way. [NEWLINE] [NEWLINE] And by your definition, isn't calling the police just another attack? I mean, we want peace right? We should just let him live his life and not bug him anymore. That's the MOST peaceful way. Except that won't stop him, and that's not justice. [NEWLINE] [NEWLINE] Now I'm not claiming drawing the prophet is justice. It is simply a demonstration that their acts are futile. A peaceful demonstration. </s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's true that in the presidential election an individual Californian's vote these days is underrepresented... ditto with you in the senate.  It's on some level the trade off of being a part of a large and wealthy state. [NEWLINE] [NEWLINE] <mask> of California's massive economic output, it's state government can have significant effects at the national level in ways that others simply cannot. [NEWLINE] [NEWLINE] Your vote matters at the local level.  Your US congress vote is no more or no less than anyone elses.  Your CA state votes matter, indirectly, more than votes for smaller states.  The US is a two party system, which means the *real* voting is done at the primaries... which shockingly few people participate in - and primary laws vary state to state. [NEWLINE] [NEWLINE] It's really intellectually lazy to just whine about a presidential election every four years.  The real battles are long before that.</s>
Label encoding: <s>It's true that in the presidential election an individual Californian's vote these days is underrepresented... ditto with you in the senate.  It's on some level the trade off of being a part of a large and wealthy state. [NEWLINE] [NEWLINE] Because of California's massive economic output, it's state government can have significant effects at the national level in ways that others simply cannot. [NEWLINE] [NEWLINE] Your vote matters at the local level.  Your US congress vote is no more or no less than anyone elses.  Your CA state votes matter, indirectly, more than votes for smaller states.  The US is a two party system, which means the *real* voting is done at the primaries... which shockingly few people participate in - and primary laws vary state to state. [NEWLINE] [NEWLINE] It's really intellectually lazy to just whine about a presidential election every four years.  The real battles are long before that.</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I doubt it, I personally know people who started with nothing and now are making 60k plus a year. My first job was 53k a year with no experience. I am 21 and just switches to a field<mask> I have no prior experience and making close to 100k. There are people who look for opportunities and put themselves out there and those who don't.<mask> some guy starts a software company or makes a website and becomes a billionaire suddenly you want to mark them<mask> evil. Someone starts a restaurant and it goes nationwide you consider them thieves. Capitalism may not seem fair<mask> you have to actually work for your money<mask> that doesn't bother those who strive for success.<mask> you make poor decisions like having a bunch of kids and dropping out of high school yes it will be hard to succeed. Guess<mask>? Our government will still send you checks, offer public housing and give you food stamps. </s>
Label encoding: <s>I doubt it, I personally know people who started with nothing and now are making 60k plus a year. My first job was 53k a year with no experience. I am 21 and just switches to a field where I have no prior experience and making close to 100k. There are people who look for opportunities and put themselves out there and those who don't. If some guy starts a software company or makes a website and becomes a billionaire suddenly you want to mark them as evil. Someone starts a restaurant and it goes nationwide you consider them thieves. Capitalism may not seem fair because you have to actually work for your money but that doesn't bother those who strive for success. If you make poor decisions like having a bunch of kids and dropping out of high school yes it will be hard to succeed. Guess what? Our government will still send you checks, offer public housing and give you food stamps. </s>
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is just plain greed. [ENDQ] [NEWLINE] Of course it's greed. That's<mask> humans operate. People do things<mask> they believe it will leave them better off<mask> they do. That includes performing tasks in exchange for money. [NEWLINE] [NEWLINE] [STARTQ] <mask> could a sane and empathetic person collect more than they could spend in a year [ENDQ] [NEWLINE] This makes no sense. You can spend an infinite amount of money in a year. [NEWLINE] [NEWLINE] [STARTQ] not want to give something back to those who had earned it for them? [ENDQ] [NEWLINE] They earned the money.<mask> someone says "I will give you $x amount of money<mask> you perform task y", then the person earns the money<mask> they perform task y. That goes for everyone, from entry-level employees all the way up to the CEO. [NEWLINE] [NEWLINE] Unless you are suggesting that the CEO did not perform the duties they agreed to in their employment contract?</s><pad><pad>
Label encoding: <s> [STARTQ] This is just plain greed. [ENDQ] [NEWLINE] Of course it's greed. That's how humans operate. People do things because they believe it will leave them better off if they do. That includes performing tasks in exchange for money. [NEWLINE] [NEWLINE] [STARTQ] How could a sane and empathetic person collect more than they could spend in a year [ENDQ] [NEWLINE] This makes no sense. You can spend an infinite amount of money in a year. [NEWLINE] [NEWLINE] [STARTQ] not want to give something back to those who had earned it for them? [ENDQ] [NEWLINE] They earned the money. When someone says "I will give you $x amount of money if you perform task y", then the person earns the money if they perform task y. That goes for everyone, from entry-level employees all the way up to the CEO. [NEWLINE] [NEWLINE] Unless you are suggesting that the CEO did not perform the duties they agreed to in their employment contract?</s><pad><pad>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] rape and wallets [ENDQ] [NEWLINE] Well no, not at all. You are not your wallet, your wallet is an object.<mask> you forgot to pick your wallet up, that isn't the same<mask> forgetting to pick yourself up. [NEWLINE] [NEWLINE] The main problem here<mask> isn't that you can't accidentally leave yourself<mask> an object on a seat, the main problem is the difference between a deterrent and blame. [NEWLINE] [NEWLINE] We blame a thief or rapist<mask> they chose to commit a crime. Without their choice, there is no crime. [NEWLINE] Victim's have 0% responsibility for being victimized,<mask> deterrents aren't the same thing<mask> blame.<mask><mask> someone already changed your view about not wearing sexy clothes actually being a deterrent, you should<mask> know that even in the case of a world<mask> not wearing sexy clothes is a perfect deterrent, we still don't say victim's have any blame for being victimized.</s>
Label encoding: <s> [STARTQ] rape and wallets [ENDQ] [NEWLINE] Well no, not at all. You are not your wallet, your wallet is an object. If you forgot to pick your wallet up, that isn't the same as forgetting to pick yourself up. [NEWLINE] [NEWLINE] The main problem here though isn't that you can't accidentally leave yourself as an object on a seat, the main problem is the difference between a deterrent and blame. [NEWLINE] [NEWLINE] We blame a thief or rapist because they chose to commit a crime. Without their choice, there is no crime. [NEWLINE] Victim's have 0% responsibility for being victimized, but deterrents aren't the same thing as blame. So while someone already changed your view about not wearing sexy clothes actually being a deterrent, you should also know that even in the case of a world where not wearing sexy clothes is a perfect deterrent, we still don't say victim's have any blame for being victimized.</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We don't have any system to ban people from owning and opperating motor vehicles, we can just take their license and make it illegal for them to drive on public roads. [NEWLINE] [NEWLINE] The problem is that alcohol affects judgement, and some people make very poor decisions<mask> they are drunk, such<mask> my friend who repeatedly drives drunk<mask><mask> he swears<mask> he is sober that he will never do it again. He gets drunk and his ability to make rational choices goes out the window, and he chooses to drive drunk<mask><mask> he never would have made that choice sober. [NEWLINE] [NEWLINE] Even<mask> his license is suspended he can still own a car and is likely to make that same poor decision<mask> he gets drunk.  He really shouldn't be allowed to buy, possess, or consume alcohol at all.<mask><mask><mask> he has lost that privilege by repeatedly putting innocent people in danger by his drunk driving.</s>
Label encoding: <s>We don't have any system to ban people from owning and opperating motor vehicles, we can just take their license and make it illegal for them to drive on public roads. [NEWLINE] [NEWLINE] The problem is that alcohol affects judgement, and some people make very poor decisions when they are drunk, such as my friend who repeatedly drives drunk even though he swears when he is sober that he will never do it again. He gets drunk and his ability to make rational choices goes out the window, and he chooses to drive drunk even though he never would have made that choice sober. [NEWLINE] [NEWLINE] Even when his license is suspended he can still own a car and is likely to make that same poor decision if he gets drunk.  He really shouldn't be allowed to buy, possess, or consume alcohol at all. In my opinion he has lost that privilege by repeatedly putting innocent people in danger by his drunk driving.</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>By "you," I man your conscious experience. I don't think that saying "I am my brain" is correct, for reasons that Daniel Dennet outlines [here]( [URL] /~mhb0/Dennett-WhereAmI.pdf). Note: Dennet proposes a compatibilist form of free will, which I don't agree with. [NEWLINE] [NEWLINE] Even<mask> the brain is in control, the brain isn't controlling<mask> the interactions of neurochemistry work, nor does it set up any initial conditions. Everything we think and decide is caused by the physical mechanisms in our brain. Unless it makes sense to say that other physical systems control over their output,<mask> opposed to the laws governing those systems and the conditions that those systems arose from, I don't think it is sensible to say that we/our brains are in control of<mask> we/our brains do.</s>
Label encoding: <s>By "you," I man your conscious experience. I don't think that saying "I am my brain" is correct, for reasons that Daniel Dennet outlines [here]( [URL] /~mhb0/Dennett-WhereAmI.pdf). Note: Dennet proposes a compatibilist form of free will, which I don't agree with. [NEWLINE] [NEWLINE] Even if the brain is in control, the brain isn't controlling how the interactions of neurochemistry work, nor does it set up any initial conditions. Everything we think and decide is caused by the physical mechanisms in our brain. Unless it makes sense to say that other physical systems control over their output, as opposed to the laws governing those systems and the conditions that those systems arose from, I don't think it is sensible to say that we/our brains are in control of what we/our brains do.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I didn't choose ridicule from among a list of other methods. It was already there. I wasn't logical from the very beginning. Otherwise I would have been much more specific and probably would have suggested other methods.<mask><mask><mask> ridicule has uses, you are right in this case. Especially your last bit about taking "some more effort" to make my point.<mask> I went to any trouble at all to make a my original point it would have been unrecognisable and not offensive or fatuous, or it would have deeply dishonest<mask> I still wanted to try to defend<mask> I started with. In the end I wouldn't have posted<mask> I did in the first place. [NEWLINE] ∆ "constructive criticism" is something I really really ought to always remember that,<mask><mask> I did I wouldn't have had this view,<mask> the delta. My criticism wasn't constructive it was just mean. </s>
Label encoding: <s>I didn't choose ridicule from among a list of other methods. It was already there. I wasn't logical from the very beginning. Otherwise I would have been much more specific and probably would have suggested other methods. While I think ridicule has uses, you are right in this case. Especially your last bit about taking "some more effort" to make my point. If I went to any trouble at all to make a my original point it would have been unrecognisable and not offensive or fatuous, or it would have deeply dishonest if I still wanted to try to defend what I started with. In the end I wouldn't have posted what I did in the first place. [NEWLINE] ∆ "constructive criticism" is something I really really ought to always remember that, because if I did I wouldn't have had this view, hence the delta. My criticism wasn't constructive it was just mean. </s>
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Great works of literature such<mask> To Kill a Mockingbird, 1984, Catcher in the Rye, etc, are all intended to persuade readers to one side of a moral issue. [ENDQ] [NEWLINE] <mask> I'm not misunderstanding you, you're saying that great literature is effectively moral propaganda on the part of the author.<mask><mask> an author writes a story with moral ambiguity and is a fence sitter about the actual ethics of the situation? Or<mask> they write about issues that they personally *wouldn't* find morally good,<mask> in a dramatic and sensationalized way for the sake of entertaining the reader? Or adventure stories that are primarily concerned with excitement and wonder, rather than moral consequences? [NEWLINE] [NEWLINE] We can't deny that literature is influenced by the moral biases of the author(s).<mask> it is *very* strong claim to say that the purpose of literature is to *define* morality.</s>
Label encoding: <s> [STARTQ] Great works of literature such as To Kill a Mockingbird, 1984, Catcher in the Rye, etc, are all intended to persuade readers to one side of a moral issue. [ENDQ] [NEWLINE] If I'm not misunderstanding you, you're saying that great literature is effectively moral propaganda on the part of the author. What if an author writes a story with moral ambiguity and is a fence sitter about the actual ethics of the situation? Or if they write about issues that they personally *wouldn't* find morally good, but in a dramatic and sensationalized way for the sake of entertaining the reader? Or adventure stories that are primarily concerned with excitement and wonder, rather than moral consequences? [NEWLINE] [NEWLINE] We can't deny that literature is influenced by the moral biases of the author(s). But it is *very* strong claim to say that the purpose of literature is to *define* morality.</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I currently works at a company that offers very affordable car leases to its employees. In that car lease they include everything, and for example insurance can be really expensive in Sweden<mask> I live for a young male that havent had a car in their name before. And they<mask> include the tax for driving through gates, which would be over half the price of leasing the car<mask> we assume you have to drive through those to get to work/home. And you got free mileage with service, which is obviously worth a lot of you drive a lot. [NEWLINE] [NEWLINE] <mask> those two conditions apply leasing a car are extremely cheap, I am currently not doing it<mask><mask> I would move outside the city gates<mask> I would have to pay that tax I would lease a car<mask> fast<mask> possible. [NEWLINE] [NEWLINE] <mask> in general<mask><mask> with you, there is no way I would lease a car under normal situations. </s>
Label encoding: <s>I currently works at a company that offers very affordable car leases to its employees. In that car lease they include everything, and for example insurance can be really expensive in Sweden where I live for a young male that havent had a car in their name before. And they also include the tax for driving through gates, which would be over half the price of leasing the car if we assume you have to drive through those to get to work/home. And you got free mileage with service, which is obviously worth a lot of you drive a lot. [NEWLINE] [NEWLINE] If those two conditions apply leasing a car are extremely cheap, I am currently not doing it but if I would move outside the city gates so I would have to pay that tax I would lease a car as fast as possible. [NEWLINE] [NEWLINE] But in general I agree with you, there is no way I would lease a car under normal situations. </s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's a hard one and you make a good point. The class could start with a lot of basic things like hygiene and helping your baby stay healthy. It could have Q&amp;A sessions<mask> instructors bust parenting myths and questions about the child's needs. Perhaps it could teach<mask> we know about the development of the brain over the years and<mask> certain things like physical abuse have (<mask> shown by studies) caused psychological issues down the road. This is just an example<mask><mask> we inform them of things less subjective and based more on scientific observation then they might be more informed<mask> they consider different ways to discipline and teach their child. Basically my feeling is<mask> we give them the data, they might make the better choices on their own.<mask> I would argue putting them in a room with other parents and instructors who have a lot of knowledge breeds collaboration, and that alone is better than nothing.</s>
Label encoding: <s>That's a hard one and you make a good point. The class could start with a lot of basic things like hygiene and helping your baby stay healthy. It could have Q&amp;A sessions where instructors bust parenting myths and questions about the child's needs. Perhaps it could teach what we know about the development of the brain over the years and how certain things like physical abuse have ( as shown by studies) caused psychological issues down the road. This is just an example but if we inform them of things less subjective and based more on scientific observation then they might be more informed as they consider different ways to discipline and teach their child. Basically my feeling is If we give them the data, they might make the better choices on their own. Also I would argue putting them in a room with other parents and instructors who have a lot of knowledge breeds collaboration, and that alone is better than nothing.</s>
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's a quote from another branch of this: [NEWLINE] [STARTQ] Walmart has 2 million employees. They can raise their CEO's salary by 2 million by decreasing all employees' annual salaries by a dollar. It is far more effective for Walmart to cut the pay of their cashiers than their executives. Large corporations can easily afford an executive's salary in the millions simply<mask> it doesn't matter<mask> they are paid. [ENDQ] [NEWLINE] Yes, CEO's are obviously much more skilled than the average manager.<mask> shareholders know that increasing a person's salary by a few million won't significantly change the value of the company.<mask> a CEO asks for a raise of a million dollars, shareholders don't have a huge reason not to give an extra million. In ideal business economics, shareholders may go for Joe,<mask> in real life a tiny change in the value of their company won't really matter to owners. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Here's a quote from another branch of this: [NEWLINE] [STARTQ] Walmart has 2 million employees. They can raise their CEO's salary by 2 million by decreasing all employees' annual salaries by a dollar. It is far more effective for Walmart to cut the pay of their cashiers than their executives. Large corporations can easily afford an executive's salary in the millions simply because it doesn't matter what they are paid. [ENDQ] [NEWLINE] Yes, CEO's are obviously much more skilled than the average manager. But shareholders know that increasing a person's salary by a few million won't significantly change the value of the company. When a CEO asks for a raise of a million dollars, shareholders don't have a huge reason not to give an extra million. In ideal business economics, shareholders may go for Joe, but in real life a tiny change in the value of their company won't really matter to owners. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask> I am aware it is not illegal to view revenge porn, only to post and share it. That would be the same<mask> it became legal to view child porn. I'm not saying that it shouldn't be illegal to produce or distribute the images<mask> it wouldn't be a good idea to give commercial pornographers licence to film children having sex to meet the market's demand. The idea would simply be let pedophiles view it,<mask> that they don't go on to abuse children. It's would still be a criminal offence to produce or distribute the images, like it is for revenge porn. [ENDQ] [NEWLINE] Just one thing: your title does state that you're specifically talking about viewing,<mask> you make want to make this clearer in your OP. I suspect you're going to get a lot of arguments about selling and distribution otherwise. Particularly<mask> the article you linked talks about it.</s>
Label encoding: <s> [STARTQ] As far as I am aware it is not illegal to view revenge porn, only to post and share it. That would be the same if it became legal to view child porn. I'm not saying that it shouldn't be illegal to produce or distribute the images because it wouldn't be a good idea to give commercial pornographers licence to film children having sex to meet the market's demand. The idea would simply be let pedophiles view it, so that they don't go on to abuse children. It's would still be a criminal offence to produce or distribute the images, like it is for revenge porn. [ENDQ] [NEWLINE] Just one thing: your title does state that you're specifically talking about viewing, but you make want to make this clearer in your OP. I suspect you're going to get a lot of arguments about selling and distribution otherwise. Particularly since the article you linked talks about it.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the use of it medically outweights any possible stigma. [ENDQ] [NEWLINE] Really?<mask> do you determine this weight? Do you have any research or statistics, or are you just claiming it<mask> objectively true? [NEWLINE] [NEWLINE] <mask><mask> the [study by the National Center for Transgender Equality and the National Gay and Lesbian Task Force]( [URL].pdf) (pdf), 19% of surveyed trans people have been refused medical care (pg 73), 28% have been verbally harassed in a medical setting (pg 74), and 2% report being physically attacked in a doctor's office (pg 74). [NEWLINE] [NEWLINE] Do you have evidence that more than 2% of trans people are ever in a medical situation<mask> their health or life are endangered by their doctor not knowing their assigned sex?<mask> not, I don't think you get to just *declare*<mask> fact that "any possible stigma" is outweighed.</s>
Label encoding: <s> [STARTQ] the use of it medically outweights any possible stigma. [ENDQ] [NEWLINE] Really? How do you determine this weight? Do you have any research or statistics, or are you just claiming it as objectively true? [NEWLINE] [NEWLINE] According to the [study by the National Center for Transgender Equality and the National Gay and Lesbian Task Force]( [URL].pdf) (pdf), 19% of surveyed trans people have been refused medical care (pg 73), 28% have been verbally harassed in a medical setting (pg 74), and 2% report being physically attacked in a doctor's office (pg 74). [NEWLINE] [NEWLINE] Do you have evidence that more than 2% of trans people are ever in a medical situation where their health or life are endangered by their doctor not knowing their assigned sex? If not, I don't think you get to just *declare* as fact that "any possible stigma" is outweighed.</s>
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is only<mask> you believe the lives of the mother (who has lived, has experience of the world, has family, potentially has other children in need of her care, has knowledge, intelligence, aspirations, dreams, goals, feelings) and the fetus (most abortions are done well before 12 weeks which means the fetus is little more than a blob of cells, can't feel pain, has no cognition that could be considered thought, has no impact on the world) are worth the exact same. I do not accept this at all. [NEWLINE] [NEWLINE] <mask>, the mother could go on to have plenty of other babies (for example; a teenager having an abortion could go on from this to finish her education, get a good job and have three more kids which she can provide for -<mask> she hadn't aborted, she'd only have had one and struggled. Wouldn't that be +2?)</s>
Label encoding: <s>This is only if you believe the lives of the mother (who has lived, has experience of the world, has family, potentially has other children in need of her care, has knowledge, intelligence, aspirations, dreams, goals, feelings) and the fetus (most abortions are done well before 12 weeks which means the fetus is little more than a blob of cells, can't feel pain, has no cognition that could be considered thought, has no impact on the world) are worth the exact same. I do not accept this at all. [NEWLINE] [NEWLINE] Also, the mother could go on to have plenty of other babies (for example; a teenager having an abortion could go on from this to finish her education, get a good job and have three more kids which she can provide for - if she hadn't aborted, she'd only have had one and struggled. Wouldn't that be +2?)</s>
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That seems to be driven by the assumption that people are listening to metal primarily for the vocals, like<mask> often happens with mainstream pop listeners -<mask> metal is a heavily instrumentally based genre, and always has been.  Part of the reason that death metal, black metal, etc. vocals are often amelodic is to remind listeners that there's all this other stuff going on<mask> well, and the band isn't just playing barebones music that's useless for anything other than for supporting the vocalist.  The other reason is that they tend to fit the harsher, rhythm-driven riffing and atypical melodies of the instruments more than a pop crooner or opera singer would - after all, I can't see anyone thinking that pop singing would fit in with a band like Cannibal Corpse at all.  Their vocals are made for the violent atmosphere that the music conveys.</s>
Label encoding: <s>That seems to be driven by the assumption that people are listening to metal primarily for the vocals, like what often happens with mainstream pop listeners - but metal is a heavily instrumentally based genre, and always has been.  Part of the reason that death metal, black metal, etc. vocals are often amelodic is to remind listeners that there's all this other stuff going on as well, and the band isn't just playing barebones music that's useless for anything other than for supporting the vocalist.  The other reason is that they tend to fit the harsher, rhythm-driven riffing and atypical melodies of the instruments more than a pop crooner or opera singer would - after all, I can't see anyone thinking that pop singing would fit in with a band like Cannibal Corpse at all.  Their vocals are made for the violent atmosphere that the music conveys.</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It might be best to instead use an analogy to a mental asylum than a prison. Prisoners are meant to have moral agency, whiles animals are not;<mask> it is not fair to 'punish' animals. Those with severe mental illness with violent tendencies often have their freedoms removed and are placed in care (asylums) which does not seek to punish<mask> simply to separate them from the outside world,<mask> they would cause harm, and allow them to have<mask> pleasant<mask> possible existence otherwise. [NEWLINE] [NEWLINE] <mask> yes,<mask> it were practical/possible/without-other-more-devastating-consequences, heavily modifying the natural world to reduce the amount of pain inflicted by animals on other animals could only be a good thing. The model for this would be the mental asylum,<mask> animals could have<mask> pleasant<mask> possible existence without causing harm to other animals.</s>
Label encoding: <s>It might be best to instead use an analogy to a mental asylum than a prison. Prisoners are meant to have moral agency, whiles animals are not; therefore it is not fair to 'punish' animals. Those with severe mental illness with violent tendencies often have their freedoms removed and are placed in care (asylums) which does not seek to punish but simply to separate them from the outside world, where they would cause harm, and allow them to have as pleasant as possible existence otherwise. [NEWLINE] [NEWLINE] But yes, if it were practical/possible/without-other-more-devastating-consequences, heavily modifying the natural world to reduce the amount of pain inflicted by animals on other animals could only be a good thing. The model for this would be the mental asylum, where animals could have as pleasant as possible existence without causing harm to other animals.</s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would only<mask><mask> "special care" should be taken<mask> dealing with *anyone*, introvert or not.  Part of social interaction is reading the person with whom you're talking, and being sensitive to their personality.  You try to avoid rude jokes around people you know aren't going to like it, you try to not yell around people who get freaked out by that kind of thing, and you don't force awkwardness onto someone who doesn't like it.  I don't think it's "kid gloves"<mask> much<mask> just being aware of the situation and acting<mask>. [NEWLINE] [NEWLINE] Extroverts have the advantage here<mask> we're<mask> used to engaging with people.  It's very easy for me to "adapt" my personality to fit a given conversation, simply<mask> I've got<mask> much practice talking to<mask> many different types of people.  </s>
Label encoding: <s>I would only argue that "special care" should be taken when dealing with *anyone*, introvert or not.  Part of social interaction is reading the person with whom you're talking, and being sensitive to their personality.  You try to avoid rude jokes around people you know aren't going to like it, you try to not yell around people who get freaked out by that kind of thing, and you don't force awkwardness onto someone who doesn't like it.  I don't think it's "kid gloves" so much as just being aware of the situation and acting accordingly. [NEWLINE] [NEWLINE] Extroverts have the advantage here because we're so used to engaging with people.  It's very easy for me to "adapt" my personality to fit a given conversation, simply because I've got so much practice talking to so many different types of people.  </s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask>, I'm kind of flamboyant. I'd like to know<mask> you define the word. [NEWLINE] [NEWLINE] For me, it means that I consider my face a canvas, and won't hesitate to paint it. It means I've used seduction<mask> a weapon, to avoid sexual harassment. (You'd be amazed<mask> many people knock it off,<mask> you intimidate them.) It means I go out of my way to entertain. Every sincere laugh is a jewel I can't hope to ever buy. [NEWLINE] [NEWLINE] It doesn't mean I hit on people who haven't hit on me, first. It doesn't mean I make my problems everyone else's problems. [NEWLINE] [NEWLINE] So, my question would be - are you opposed to flamboyant behavior, or just the toxic way in which some flamboyant people demand attention for themselves, rather than earning it? </s>
Label encoding: <s>So, I'm kind of flamboyant. I'd like to know how you define the word. [NEWLINE] [NEWLINE] For me, it means that I consider my face a canvas, and won't hesitate to paint it. It means I've used seduction as a weapon, to avoid sexual harassment. (You'd be amazed how many people knock it off, if you intimidate them.) It means I go out of my way to entertain. Every sincere laugh is a jewel I can't hope to ever buy. [NEWLINE] [NEWLINE] It doesn't mean I hit on people who haven't hit on me, first. It doesn't mean I make my problems everyone else's problems. [NEWLINE] [NEWLINE] So, my question would be - are you opposed to flamboyant behavior, or just the toxic way in which some flamboyant people demand attention for themselves, rather than earning it? </s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] And typing up an assignment is a great idea and all,<mask><mask> about tests? Arguably those are worth much more to their grades than homework, and<mask> you dock for illegible work on tests even<mask> they've tried their best that's not exactly fair. [ENDQ] [NEWLINE] OP isn't talking about tests.<mask> a fellow TA, I will work to decipher anything on a test... I'm not going to do the same thing for homework; I simply don't have that much time. I have 100 assignments to grade each week; I have no choice<mask> to prioritize quantity over quality, particularly<mask> I have to have a 2 or 3-day turnaround. For tests, I have a week or more for turnaround, and<mask> I can spend the time to understand<mask> people screwed up, work the problem through with their incorrect numbers, and decipher bad handwriting<mask> well. </s>
Label encoding: <s> [STARTQ] And typing up an assignment is a great idea and all, but what about tests? Arguably those are worth much more to their grades than homework, and if you dock for illegible work on tests even when they've tried their best that's not exactly fair. [ENDQ] [NEWLINE] OP isn't talking about tests. As a fellow TA, I will work to decipher anything on a test... I'm not going to do the same thing for homework; I simply don't have that much time. I have 100 assignments to grade each week; I have no choice but to prioritize quantity over quality, particularly because I have to have a 2 or 3-day turnaround. For tests, I have a week or more for turnaround, and so I can spend the time to understand where people screwed up, work the problem through with their incorrect numbers, and decipher bad handwriting as well. </s>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are wrong,  it won't be the same.  Not even close.  Of course I would still have participated in the whole Santa Claus thing<mask> a kid even<mask> I did know he wasn't real. <mask> it wouldn't even be close to the same. Believing and playing along are a huge difference<mask> your a kid.  You seem hell bent on not participating in the whole Santa thing with your kids.  Can I ask<mask>?<mask> could you or your kids possibly gain by them not believing in Santa and knowing the truth.  I'm<mask><mask>,  all you would achieve by doing that would be to rob your children of<mask> awesome it is to believe in Santa Claus for the few years that their brains will believe it.  Until they become more skeptical and mature and realize the world's not such a nice place all the time.  </s>
Label encoding: <s>You are wrong,  it won't be the same.  Not even close.  Of course I would still have participated in the whole Santa Claus thing as a kid even if I did know he wasn't real.  But it wouldn't even be close to the same. Believing and playing along are a huge difference when your a kid.  You seem hell bent on not participating in the whole Santa thing with your kids.  Can I ask why? What could you or your kids possibly gain by them not believing in Santa and knowing the truth.  I'm my opinion,  all you would achieve by doing that would be to rob your children of how awesome it is to believe in Santa Claus for the few years that their brains will believe it.  Until they become more skeptical and mature and realize the world's not such a nice place all the time.  </s>
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Think about it like this: Legally speaking, using a gun is considered deadly force in ANY scenario. The only way you're allowed to use deadly force is<mask> you feel that you are in imminent danger of great bodily harm/death, and<mask> someone is making you feel that way, the best/quickest way to end that threat is by firing into center mass until the threat has stopped. [NEWLINE] [NEWLINE] <mask> you have time for a warning shot/wounding shot, you weren't in imminent enough danger to truly use deadly force, and you're not legally protected. [NEWLINE] [NEWLINE] <mask>, aiming for extremities isn't taught anywhere<mask> it's the most likely to result in an unfavorable outcome.  Very few people are stopped by being shot only once, and again,<mask> you're in that situation<mask> you're using deadly force, do you want to take that chance?</s>
Label encoding: <s>Think about it like this: Legally speaking, using a gun is considered deadly force in ANY scenario. The only way you're allowed to use deadly force is if you feel that you are in imminent danger of great bodily harm/death, and if someone is making you feel that way, the best/quickest way to end that threat is by firing into center mass until the threat has stopped. [NEWLINE] [NEWLINE] If you have time for a warning shot/wounding shot, you weren't in imminent enough danger to truly use deadly force, and you're not legally protected. [NEWLINE] [NEWLINE] Additionally, aiming for extremities isn't taught anywhere because it's the most likely to result in an unfavorable outcome.  Very few people are stopped by being shot only once, and again, if you're in that situation where you're using deadly force, do you want to take that chance?</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>After the cops had then barricaded for 50. Years [NEWLINE] [NEWLINE] [NEWLINE] These kids didn't elect shit btw. They just die at the hand of the cops who think bombing their apartment complex which they and their parents  cannot leave  is a reasonable way to respond to this gunman who only gets more support exactly<mask><mask><mask> behavior From the police [NEWLINE] [NEWLINE] [NEWLINE] Some guy in the building is bad. These kids are dying and they and their parents have zero opportunity of leaving [NEWLINE] [NEWLINE] [NEWLINE] That zero opportunity is enforced by the cops.  Not the crazy man [NEWLINE] [NEWLINE] [NEWLINE] Then the cops blame the crazy man saying he has hostages.  These hostages were forced to remain hostages by the same police who now kill them and blame the crazy man [NEWLINE] [NEWLINE] [NEWLINE] Fuxking  bullshit man [NEWLINE] [NEWLINE] [NEWLINE] Weak ass bullshit [NEWLINE] [NEWLINE] [NEWLINE] On mobile grammar dies like Palestinian kids on mobile </s>
Label encoding: <s>After the cops had then barricaded for 50. Years [NEWLINE] [NEWLINE] [NEWLINE] These kids didn't elect shit btw. They just die at the hand of the cops who think bombing their apartment complex which they and their parents  cannot leave  is a reasonable way to respond to this gunman who only gets more support exactly because of this behavior From the police [NEWLINE] [NEWLINE] [NEWLINE] Some guy in the building is bad. These kids are dying and they and their parents have zero opportunity of leaving [NEWLINE] [NEWLINE] [NEWLINE] That zero opportunity is enforced by the cops.  Not the crazy man [NEWLINE] [NEWLINE] [NEWLINE] Then the cops blame the crazy man saying he has hostages.  These hostages were forced to remain hostages by the same police who now kill them and blame the crazy man [NEWLINE] [NEWLINE] [NEWLINE] Fuxking  bullshit man [NEWLINE] [NEWLINE] [NEWLINE] Weak ass bullshit [NEWLINE] [NEWLINE] [NEWLINE] On mobile grammar dies like Palestinian kids on mobile </s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Couple thoughts: [NEWLINE] [NEWLINE] 1) people cheating devalues the program.<mask> cheaters get out into the workforce, and employers realize they don't have the skills/education that they claim to have, it reflects on the school and degree program. "Don't hire people from X. They never know<mask> they're doing" [NEWLINE] [NEWLINE] 2) people cheating are really cheating themselves. They're spending considerable time and money on an education, and not getting that education. [NEWLINE] [NEWLINE] 3) a small number of students attempting to cheat is going to happen. A majority of students cheating may be an indication that there are problems with the curriculum itself. The cheating is masking this potential problem from faculty.<mask> the curriculum is broken, reporting the cheating my help them fix that, which is something you all could benefit from. Or at least those that come after you will. </s>
Label encoding: <s>Couple thoughts: [NEWLINE] [NEWLINE] 1) people cheating devalues the program. When cheaters get out into the workforce, and employers realize they don't have the skills/education that they claim to have, it reflects on the school and degree program. "Don't hire people from X. They never know what they're doing" [NEWLINE] [NEWLINE] 2) people cheating are really cheating themselves. They're spending considerable time and money on an education, and not getting that education. [NEWLINE] [NEWLINE] 3) a small number of students attempting to cheat is going to happen. A majority of students cheating may be an indication that there are problems with the curriculum itself. The cheating is masking this potential problem from faculty. If the curriculum is broken, reporting the cheating my help them fix that, which is something you all could benefit from. Or at least those that come after you will. </s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So by applying your argument to yourself, just<mask> you have no evidence to back up your claims of people's motivations (arguing against the majority just to feel superior) just<mask> I have no evidence to back up my claim that you're doing the same thing, wouldn't it make sense that<mask> antiproton is saying is more realistic. That these things are an emergent part of the human condition<mask> you bring a large number of people together. [NEWLINE] [NEWLINE] <mask> every argument you come up with defeats itself<mask> it can be applied to yourself, it seems more realistic that the argument is flawed in the first place unless you are trying to point out this self-defeating property (which is<mask> antiproton is doing by claiming that these phenomena you describe are explained by some overarching aspect of the human condition, i.e. everyone argues<mask> they have differing opinions)</s>
Label encoding: <s>So by applying your argument to yourself, just as you have no evidence to back up your claims of people's motivations (arguing against the majority just to feel superior) just as I have no evidence to back up my claim that you're doing the same thing, wouldn't it make sense that what antiproton is saying is more realistic. That these things are an emergent part of the human condition when you bring a large number of people together. [NEWLINE] [NEWLINE] Because every argument you come up with defeats itself because it can be applied to yourself, it seems more realistic that the argument is flawed in the first place unless you are trying to point out this self-defeating property (which is what antiproton is doing by claiming that these phenomena you describe are explained by some overarching aspect of the human condition, i.e. everyone argues when they have differing opinions)</s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It would be trivial to add a reproductive component to recognition of marriages; to require, within two years, that a couple submit a birth certificate for their child or have the marriage invalidated. [NEWLINE] [NEWLINE] I'm not aware of jurisdictions that have such a requirement. Marriage could be about<mask> many other things: forming stable pairings to give people larger support networks, promotion of mental health benefits, acknowledgement of the cultural reality that people enter into dyadic relationships and change<mask> they approach division of labor...in the US at least (not sure about Israel) spouses have rights and privileges that have nothing to do with children. Here, the only time marriage is about children is<mask> gay marriage is being debated. [NEWLINE] [NEWLINE] <mask>, to establish your point, you need to demonstrate either that there's strong evidence that marriage really is about procreation, or that it ought to be. </s>
Label encoding: <s>It would be trivial to add a reproductive component to recognition of marriages; to require, within two years, that a couple submit a birth certificate for their child or have the marriage invalidated. [NEWLINE] [NEWLINE] I'm not aware of jurisdictions that have such a requirement. Marriage could be about so many other things: forming stable pairings to give people larger support networks, promotion of mental health benefits, acknowledgement of the cultural reality that people enter into dyadic relationships and change how they approach division of labor...in the US at least (not sure about Israel) spouses have rights and privileges that have nothing to do with children. Here, the only time marriage is about children is when gay marriage is being debated. [NEWLINE] [NEWLINE] So, to establish your point, you need to demonstrate either that there's strong evidence that marriage really is about procreation, or that it ought to be. </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Realistically, with the amount of technology we have available, a 1:20 ratio in a classroom<mask> the majority of lecture is given through online video would be plenty enough (even 1:30 or something crazy like 3:100 should work). [NEWLINE] [NEWLINE] <mask> you want to see some cool stuff going on with modern education, look up the khan academy and the cool stuff that Sal is doing. [NEWLINE] [NEWLINE] The bigger problem right now seems to be the policy and structure of school.  Teachers are not the badguys right now.  And not having enough of them isn't the issue either.  We just seem to be stuck in this 1900s way of doing school<mask> we should be pushing the envelope, individualizing courses, and letting our kids pursue<mask> interests them most<mask> they are most receptive to that type of learning. [NEWLINE] [NEWLINE] Start here: [NEWLINE] [URL] </s>
Label encoding: <s>Realistically, with the amount of technology we have available, a 1:20 ratio in a classroom where the majority of lecture is given through online video would be plenty enough (even 1:30 or something crazy like 3:100 should work). [NEWLINE] [NEWLINE] If you want to see some cool stuff going on with modern education, look up the khan academy and the cool stuff that Sal is doing. [NEWLINE] [NEWLINE] The bigger problem right now seems to be the policy and structure of school.  Teachers are not the badguys right now.  And not having enough of them isn't the issue either.  We just seem to be stuck in this 1900s way of doing school when we should be pushing the envelope, individualizing courses, and letting our kids pursue what interests them most when they are most receptive to that type of learning. [NEWLINE] [NEWLINE] Start here: [NEWLINE] [URL] </s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0757, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] there are many black people in the US who put the blame of racism and slavery on ALL white people [ENDQ] [NEWLINE] Can you give me an example of this?  I don't think I have ever heard anybody say that all white people are to blame for slavery.  Many certainly believe that the majority of white people are ignorantly or willfully blind (or indifferent) to the ongoing effects of past discrimination, and<mask> complicit in the perpetuation of those effects -<mask> I don't think they say all white people are to blame for *slavery*. [NEWLINE] [NEWLINE] *Edit* Saying that this country should be doing more to remedy past discrimination is not the same thing<mask> saying white people alive today are to blame for past discrimination.  Just that they have a duty to either remedy those effects or stop pretending the racial divide in this country is meritocratic in any meaningful way.</s>
Label encoding: <s> [STARTQ] there are many black people in the US who put the blame of racism and slavery on ALL white people [ENDQ] [NEWLINE] Can you give me an example of this?  I don't think I have ever heard anybody say that all white people are to blame for slavery.  Many certainly believe that the majority of white people are ignorantly or willfully blind (or indifferent) to the ongoing effects of past discrimination, and therefore complicit in the perpetuation of those effects - but I don't think they say all white people are to blame for *slavery*. [NEWLINE] [NEWLINE] *Edit* Saying that this country should be doing more to remedy past discrimination is not the same thing as saying white people alive today are to blame for past discrimination.  Just that they have a duty to either remedy those effects or stop pretending the racial divide in this country is meritocratic in any meaningful way.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Situation you describe is only possible for a few decades, in developed enough countries, and among people that do well enough for themselves. In any other circumstances, family is pretty much instrumental for survival (mostly<mask> kids are involved of course). In most cases, family only works well with monogamy.<mask>, other ways don't really work in majority of cases (and did not work at all until recently), and<mask> can't be the norm. [NEWLINE] [NEWLINE] For the people who are "applicable" to this, it's of course just a matter of their mutual agreement, not something dictated by economical needs. You don't want society to control them in the direction of being monogamous,<mask> controlling in the opposite direction is really the same thing, just opposite.<mask> more accurate, the "norm" should be understanding and honoring the agreement you enter in. [NEWLINE] </s>
Label encoding: <s>Situation you describe is only possible for a few decades, in developed enough countries, and among people that do well enough for themselves. In any other circumstances, family is pretty much instrumental for survival (mostly when kids are involved of course). In most cases, family only works well with monogamy. So, other ways don't really work in majority of cases (and did not work at all until recently), and therefore can't be the norm. [NEWLINE] [NEWLINE] For the people who are "applicable" to this, it's of course just a matter of their mutual agreement, not something dictated by economical needs. You don't want society to control them in the direction of being monogamous, but controlling in the opposite direction is really the same thing, just opposite. So more accurate, the "norm" should be understanding and honoring the agreement you enter in. [NEWLINE] </s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am an organ donor and an athiest. I<mask> have a very close family friend who is on the donor list<mask> her kidneys are failing. Still, I dont believe it should be mandatory. I like the opt out option. I<mask> like living in a country<mask> people are free to practice any religion they see fit<mask><mask><mask> it isn't causing anyone else danger and I do not believe opting out of donating organs for religious reasons is putting anyone's life in danger. We certainly need a sort of campaign to make people more aware of<mask> immense good donating your organs can do<mask> I can't see making people who truly believe they need all their organs or body parts to pass into the next life give them up. Some people believe that every bit of their body is sacred and that is up to them to decide not the government. [NEWLINE] [NEWLINE] edit: caps</s>
Label encoding: <s>I am an organ donor and an athiest. I also have a very close family friend who is on the donor list because her kidneys are failing. Still, I dont believe it should be mandatory. I like the opt out option. I also like living in a country where people are free to practice any religion they see fit so long as it isn't causing anyone else danger and I do not believe opting out of donating organs for religious reasons is putting anyone's life in danger. We certainly need a sort of campaign to make people more aware of what immense good donating your organs can do but I can't see making people who truly believe they need all their organs or body parts to pass into the next life give them up. Some people believe that every bit of their body is sacred and that is up to them to decide not the government. [NEWLINE] [NEWLINE] edit: caps</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> you say men [ENDQ] [NEWLINE] no i don't say men, i say people. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> it hadn't then women wouldn't be voted<mask> leaders [ENDQ] [NEWLINE] ignoring that there are<mask> many more male leaders than female leaders,<mask> do you think that women can't be voted<mask> "leaders"(true leaders only exist in dictatorships and kingdoms) even<mask> a male equivalent would be taken more serious? Maybe there is no male rival that offers the same benefits<mask> this female leader? I don't<mask><mask> women can't lead anything in todays world<mask> rather that their opinions are subconsciously valued less.<mask><mask> my view is difficult to accept<mask> i don't think i can prove or anyone could disprove it,<mask> i would understand<mask> someone wouldn't want to have such a view or accept such a view<mask> vaild. [NEWLINE] </s>
Label encoding: <s> [STARTQ] since as you say men [ENDQ] [NEWLINE] no i don't say men, i say people. [NEWLINE] [NEWLINE] [STARTQ] because if it hadn't then women wouldn't be voted as leaders [ENDQ] [NEWLINE] ignoring that there are so many more male leaders than female leaders, why do you think that women can't be voted as "leaders"(true leaders only exist in dictatorships and kingdoms) even if a male equivalent would be taken more serious? Maybe there is no male rival that offers the same benefits as this female leader? I don't argue that women can't lead anything in todays world but rather that their opinions are subconsciously valued less. I think my view is difficult to accept because i don't think i can prove or anyone could disprove it, so i would understand why someone wouldn't want to have such a view or accept such a view as vaild. [NEWLINE] </s>
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry DefendWaifuWithRaifu, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=DefendWaifuWithRaifu+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad><pad>
Label encoding: <s>Sorry DefendWaifuWithRaifu, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=DefendWaifuWithRaifu+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad><pad>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course it's a biased question -- it assumes the answer is yes<mask> most reasonable people would find it very difficult to answer "no." [NEWLINE] [NEWLINE] [STARTQ] 1) technology doesn't exist to propel the missile that far [ENDQ] [NEWLINE] <mask> the issue here is atomic weapons, generally, and the point of comparison for trustworthiness purposes is the USA, hypo should assume that Afghanistan possessed weapons with capabilities similar to to the U.S. arsenal. [NEWLINE] [NEWLINE] This<mask> moots your (2). And even<mask> it didn't, the argument that Afghanistan is more trustworthy than the USA<mask> the genocidal criminals to whom it might leak nukes are merely incompetent...is not a very good argument. [NEWLINE] [NEWLINE] Edit:  lol at the assumption that I'm the one downvoting you.  I have not downvoted anyone ITT, including OP. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Of course it's a biased question -- it assumes the answer is yes because most reasonable people would find it very difficult to answer "no." [NEWLINE] [NEWLINE] [STARTQ] 1) technology doesn't exist to propel the missile that far [ENDQ] [NEWLINE] Since the issue here is atomic weapons, generally, and the point of comparison for trustworthiness purposes is the USA, hypo should assume that Afghanistan possessed weapons with capabilities similar to to the U.S. arsenal. [NEWLINE] [NEWLINE] This also moots your (2). And even if it didn't, the argument that Afghanistan is more trustworthy than the USA because the genocidal criminals to whom it might leak nukes are merely incompetent...is not a very good argument. [NEWLINE] [NEWLINE] Edit:  lol at the assumption that I'm the one downvoting you.  I have not downvoted anyone ITT, including OP. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0732, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Every country has something that is unique. Is it a douche move<mask> someone from Europe traveled to the united states to see the grand canyon? [ENDQ] [NEWLINE] Yes. Its a ditch. Tell them to look at the sky it doesnt even end its way more exciting. [NEWLINE] [NEWLINE] This isnt about the merits or lack thereof, of the alright, kinda decent canyon. [NEWLINE] [NEWLINE] Obviously every country is unique<mask> some are a lot more "unique" than others, at least in comparison to your home country.<mask> youre scared of other cultures youre a douche. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> that was something that person had wanted to do for years, is he a douche for fulfilling his dream instead of going on a safari in Africa? [ENDQ] [NEWLINE] <mask> its "my dream" to throw acid on children it still makes me a douche. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Every country has something that is unique. Is it a douche move if someone from Europe traveled to the united states to see the grand canyon? [ENDQ] [NEWLINE] Yes. Its a ditch. Tell them to look at the sky it doesnt even end its way more exciting. [NEWLINE] [NEWLINE] This isnt about the merits or lack thereof, of the alright, kinda decent canyon. [NEWLINE] [NEWLINE] Obviously every country is unique but some are a lot more "unique" than others, at least in comparison to your home country. If youre scared of other cultures youre a douche. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] What if that was something that person had wanted to do for years, is he a douche for fulfilling his dream instead of going on a safari in Africa? [ENDQ] [NEWLINE] If its "my dream" to throw acid on children it still makes me a douche. [NEWLINE] </s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of the time a mental illness is not only defined by the symptomes<mask><mask><mask> they cause the person experiencing them distress / cause big problems. [NEWLINE] [NEWLINE] People who feel no romantic attraction to anybody do not need to be in any distress about it, they can function well in society and the only problems they might encounter are others who wonder<mask> they never are in a relationship. [NEWLINE] [NEWLINE] <mask>, being aromantic would probably not qualify<mask> a mental illness and an aromantic person would probably not need therapy ( for being aromantic). [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] <mask> do you think love, romantic love, is THAT important? I somebody is happy just living their life, having friends, pursuing a career and hobbies,<mask> do you think they NEED to be able to fall in love?<mask> makes romantic love more important than love for your friends or family?</s><pad>
Label encoding: <s>A lot of the time a mental illness is not only defined by the symptomes but also if they cause the person experiencing them distress / cause big problems. [NEWLINE] [NEWLINE] People who feel no romantic attraction to anybody do not need to be in any distress about it, they can function well in society and the only problems they might encounter are others who wonder why they never are in a relationship. [NEWLINE] [NEWLINE] Thus, being aromantic would probably not qualify as a mental illness and an aromantic person would probably not need therapy ( for being aromantic). [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] Why do you think love, romantic love, is THAT important? I somebody is happy just living their life, having friends, pursuing a career and hobbies, why do you think they NEED to be able to fall in love? What makes romantic love more important than love for your friends or family?</s><pad>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [NEWLINE] Sorry 1leggeddog, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=1leggeddog+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \)) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s> [NEWLINE] [NEWLINE] Sorry 1leggeddog, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=1leggeddog+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \)) [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But<mask> come the gay population of the planet is predominantly western? I mean it goes back to Greeks and was continued until 1700s by men like Oscar Wilde and all.<mask> countries in Eastern Mediterranean have almost comparatively no gay population. I mean in the USA with the uprise and this neo-Renaissance of the homosexual culture ( like their own porn, movies, text book children education on homosexual family structures, pride parade, rainbow flags, specific music or stores, specific clothing a and even specific hand gestures or postures, etc etc ) even children  who are 'genetically' non-gay (<mask> in the people from eastern ancestry with no gay-universe culture or presence ) seem to be more increasingly gay.<mask> can we not deduce through these blatantly visible evidence that culture and environment is equally or more responsible for one to become gay? </s><pad>
Label encoding: <s>But how come the gay population of the planet is predominantly western? I mean it goes back to Greeks and was continued until 1700s by men like Oscar Wilde and all. But countries in Eastern Mediterranean have almost comparatively no gay population. I mean in the USA with the uprise and this neo-Renaissance of the homosexual culture ( like their own porn, movies, text book children education on homosexual family structures, pride parade, rainbow flags, specific music or stores, specific clothing a and even specific hand gestures or postures, etc etc ) even children  who are 'genetically' non-gay ( as in the people from eastern ancestry with no gay-universe culture or presence ) seem to be more increasingly gay. So can we not deduce through these blatantly visible evidence that culture and environment is equally or more responsible for one to become gay? </s><pad>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, sure, this idea is at least not ruled out by our understanding of physics.  It is<mask> not particularly supported by anything about string theory that I know of. [NEWLINE] [NEWLINE] Anyway, all that aside, my personal hypothesis about the underlying nature of reality is that 1) anything that isn't literally impossible does happen, spontaneously,<mask> nothing is ultimately stopping it and 2) all possible sets of rules that aren't logically contradictory exist, including the laws of physics in our universe (and infinite other universes) [NEWLINE] [NEWLINE] <mask> I guess my hypothesis is that the ultimate, most basic laws of physics are simply those of logic.  I base this idea on nothing,<mask> I really can't say I'm more rigorous about<mask> I believe than anyone else... I<mask> freely admit this idea is supported by zero evidence, it just seems plausible to me. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Well, sure, this idea is at least not ruled out by our understanding of physics.  It is also not particularly supported by anything about string theory that I know of. [NEWLINE] [NEWLINE] Anyway, all that aside, my personal hypothesis about the underlying nature of reality is that 1) anything that isn't literally impossible does happen, spontaneously, because nothing is ultimately stopping it and 2) all possible sets of rules that aren't logically contradictory exist, including the laws of physics in our universe (and infinite other universes) [NEWLINE] [NEWLINE] So I guess my hypothesis is that the ultimate, most basic laws of physics are simply those of logic.  I base this idea on nothing, so I really can't say I'm more rigorous about what I believe than anyone else... I also freely admit this idea is supported by zero evidence, it just seems plausible to me. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was specifically talking about [NEWLINE] [NEWLINE] [STARTQ] along with the millions of other women who have had to do the same [ENDQ] [NEWLINE] <mask> there are millions of men who have to do the same thing, partly<mask> women refuse to pay child support.<mask> your whole post is funny. The idea that your anecdotal evidence is<mask> overwhelming that it gives you facts about millions of people,<mask> you wouldn't believe a study even<mask> they had one<mask> men lie (for some reason) about wanting a child,<mask><mask> they have no reason to do<mask>. [NEWLINE] [NEWLINE] [STARTQ] <mask> did I say she didn't want to raise me? [ENDQ] [NEWLINE] Your dad didn't.<mask><mask> didn't she get an abortion<mask> she didn't want to raise you on her own? It was her choice, and she chose single motherhood. She has no one to blame<mask> herself. </s><pad>
Label encoding: <s>I was specifically talking about [NEWLINE] [NEWLINE] [STARTQ] along with the millions of other women who have had to do the same [ENDQ] [NEWLINE] Because there are millions of men who have to do the same thing, partly because women refuse to pay child support. But your whole post is funny. The idea that your anecdotal evidence is so overwhelming that it gives you facts about millions of people, but you wouldn't believe a study even if they had one because men lie (for some reason) about wanting a child, even though they have no reason to do so. [NEWLINE] [NEWLINE] [STARTQ] When did I say she didn't want to raise me? [ENDQ] [NEWLINE] Your dad didn't. So why didn't she get an abortion if she didn't want to raise you on her own? It was her choice, and she chose single motherhood. She has no one to blame but herself. </s><pad>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You should do a blind taste test, see<mask> your preferences are really<mask> clear cut<mask> you think. [NEWLINE] [NEWLINE] Here's the trick, to make it challenging, use more than two glasses, I'd suggest at least six.  Have someone else fill three with Coke, and three with Pepsi.  Label them 1-6, and make sure they keep a list somewhere out of your view.  Then, have them leave the room. [NEWLINE] [NEWLINE] Next, you enter, and taste each of the six glasses, and sort them into two piles.  Label each one<mask> Coke or Pepsi, and see<mask> you are right. [NEWLINE] [NEWLINE] <mask> you want an even greater challenge, use six glasses, remove the restriction that there will be three of each.  Or use more glasses. [NEWLINE] [NEWLINE] See<mask> you can still sort them with 100% accuracy.</s>
Label encoding: <s>You should do a blind taste test, see if your preferences are really as clear cut as you think. [NEWLINE] [NEWLINE] Here's the trick, to make it challenging, use more than two glasses, I'd suggest at least six.  Have someone else fill three with Coke, and three with Pepsi.  Label them 1-6, and make sure they keep a list somewhere out of your view.  Then, have them leave the room. [NEWLINE] [NEWLINE] Next, you enter, and taste each of the six glasses, and sort them into two piles.  Label each one as Coke or Pepsi, and see if you are right. [NEWLINE] [NEWLINE] If you want an even greater challenge, use six glasses, remove the restriction that there will be three of each.  Or use more glasses. [NEWLINE] [NEWLINE] See if you can still sort them with 100% accuracy.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0761, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0690, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The only secondhand ring I would give my<mask> is one from... a lifetime of bliss and happiness. [ENDQ] [NEWLINE] Exactly. Secondhand rings have,<mask> not in reality, then at least in perception, a massive stigma on them,<mask> the majority don't have that associated with them. [NEWLINE] [NEWLINE] <mask>, and this is based on personal research, most of the diamonds on rings you find at pawnshops are utter crap- poorly-cut I1/I2 diamonds with J/K color; sacrifices the original buyer made to get<mask> large a diamond<mask> possible for<mask> little money<mask> possible.<mask> I had been willing to compromise my standards and go with something like that, I could have bought a full-carat or larger diamond for my wife.<mask> it was, I went for a much higher-grade.71ct stone.</s>
Label encoding: <s> [STARTQ] The only secondhand ring I would give my SO is one from... a lifetime of bliss and happiness. [ENDQ] [NEWLINE] Exactly. Secondhand rings have, if not in reality, then at least in perception, a massive stigma on them, because the majority don't have that associated with them. [NEWLINE] [NEWLINE] Also, and this is based on personal research, most of the diamonds on rings you find at pawnshops are utter crap- poorly-cut I1/I2 diamonds with J/K color; sacrifices the original buyer made to get as large a diamond as possible for as little money as possible. If I had been willing to compromise my standards and go with something like that, I could have bought a full-carat or larger diamond for my wife. As it was, I went for a much higher-grade.71ct stone.</s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> about world class, award winning BBQ? [ENDQ] [NEWLINE] The south firmly disagrees.  We respect that it's good meat,<mask><mask> someone who gets to get a slice of authentic NY pizza from a real pizzeria any time you like, surely you can appreciate that there is more to it than the food.  There's an atmosphere, a soul, that you can only get by going straight to the source. [NEWLINE] [NEWLINE] We have good pizza down here, even made by NYC transplants,<mask> I know that I'm getting a good slice,<mask> I refuse to say that I've had true NY pizza until I get it from a greasy-ass stand in whatever part of NYC one gets pizza (I don't even know<mask> ).  This is<mask> we feel down here<mask> we hear about "great BBQ" anywhere north of Richmond.</s>
Label encoding: <s> [STARTQ] How about world class, award winning BBQ? [ENDQ] [NEWLINE] The south firmly disagrees.  We respect that it's good meat, but as someone who gets to get a slice of authentic NY pizza from a real pizzeria any time you like, surely you can appreciate that there is more to it than the food.  There's an atmosphere, a soul, that you can only get by going straight to the source. [NEWLINE] [NEWLINE] We have good pizza down here, even made by NYC transplants, so I know that I'm getting a good slice, but I refuse to say that I've had true NY pizza until I get it from a greasy-ass stand in whatever part of NYC one gets pizza (I don't even know yet ).  This is how we feel down here when we hear about "great BBQ" anywhere north of Richmond.</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The industry has lots of virus-filled, sketchy websites. [ENDQ] [NEWLINE] <mask> does that make pornography a corrupting enterprise?<mask>, it is important to remember that the ads are the virus filled nonsense.<mask> people are tech literate enough not to click on sketchy ads you'll be just fine.<mask><mask> a a company that sells computer security products tested the old "porn sites are filled with viruses" hypothesis in 2010 and discovered that<mask><mask> [for every infected porn site there are 70 non-porn sites infected with viruses]( [URL].toptenreviews.com/<mask> -are-some-types-of-websites-filled-with-viruses-.html). [NEWLINE] [NEWLINE] [STARTQ] It<mask> can be oppressive to its 'employees.' [ENDQ] [NEWLINE] <mask><mask>? Can you demonstrate that this is the case? </s>
Label encoding: <s> [STARTQ] The industry has lots of virus-filled, sketchy websites. [ENDQ] [NEWLINE] How does that make pornography a corrupting enterprise? Also, it is important to remember that the ads are the virus filled nonsense. If people are tech literate enough not to click on sketchy ads you'll be just fine. In fact a a company that sells computer security products tested the old "porn sites are filled with viruses" hypothesis in 2010 and discovered that in fact [for every infected porn site there are 70 non-porn sites infected with viruses]( [URL].toptenreviews.com/ why -are-some-types-of-websites-filled-with-viruses-.html). [NEWLINE] [NEWLINE] [STARTQ] It also can be oppressive to its 'employees.' [ENDQ] [NEWLINE] How so? Can you demonstrate that this is the case? </s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As for Senior Citizens, it is a fine line. On one hand, you expect them to have lots of money after working all their life or at least have kids supporting them financially along with all their other benefits...<mask> there are Senior Citizens who do not have kids (or at least are not in a good relationship with them) or do not have kids that are financially capable. There are<mask> those who do not have hefty retirement funds and have not been able to save well. They are<mask> going to be hard-pressed to find a job in their age. It is a form of repaying this generation that has gone off to make the U.S. we have now.<mask> a Senior Citizen is screwed over... they have no real way to get back up. They are not strong enough to work. And never will be again. </s>
Label encoding: <s>As for Senior Citizens, it is a fine line. On one hand, you expect them to have lots of money after working all their life or at least have kids supporting them financially along with all their other benefits... But there are Senior Citizens who do not have kids (or at least are not in a good relationship with them) or do not have kids that are financially capable. There are also those who do not have hefty retirement funds and have not been able to save well. They are also going to be hard-pressed to find a job in their age. It is a form of repaying this generation that has gone off to make the U.S. we have now. If a Senior Citizen is screwed over... they have no real way to get back up. They are not strong enough to work. And never will be again. </s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, you torrented a whole series, that probably cost more than 7 dollars had you paid for it, and afterwards bought a book for 7 dollars<mask><mask><mask> of this? And<mask> did you not just torrent the book? That's possible you know.<mask> spend money on something you can get for free? [NEWLINE] [NEWLINE] I know moralizing won't help anyone, I just need to point out the blunder you're making<mask> you're saying<mask> of piracy you have helped the industry,<mask> your piracy led to purchase.<mask> not be consistent all the way, you're still a pirate, right? To make a discussion that's relevant to this discussion,<mask> did you pay 7 dollars for that book you could've gotten for free?<mask> piracy is A-OK, then<mask> at all pay for anything, even<mask> it's available?</s>
Label encoding: <s>So, you torrented a whole series, that probably cost more than 7 dollars had you paid for it, and afterwards bought a book for 7 dollars as a result of this? And why did you not just torrent the book? That's possible you know. Why spend money on something you can get for free? [NEWLINE] [NEWLINE] I know moralizing won't help anyone, I just need to point out the blunder you're making while you're saying because of piracy you have helped the industry, because your piracy led to purchase. Why not be consistent all the way, you're still a pirate, right? To make a discussion that's relevant to this discussion, why did you pay 7 dollars for that book you could've gotten for free? If piracy is A-OK, then why at all pay for anything, even if it's available?</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>3D can be used to convey emotion and meaning to a film. It can spent depth and clarify position: I see a good use for it in war movies<mask> tactics are important to storytelling. [NEWLINE] [NEWLINE] One great example is in the Tintin movie,<mask> the control of the focus of the viewer is used in one scene to draw the eyes deeper into the scene<mask> Tintin wanders through an old room, only to realise... The shadow *in the foreground* is significant. It blew me away with<mask> clever a use of 3D, and effective in making the surprise of a shadowy-figure more visceral, not<mask> it was a jump-scare/throw at the audience moment,<mask> by revealing that the audience had been *looking past* the potential attacker all along. It was brilliant and emotional use of 3D.</s>
Label encoding: <s>3D can be used to convey emotion and meaning to a film. It can spent depth and clarify position: I see a good use for it in war movies where tactics are important to storytelling. [NEWLINE] [NEWLINE] One great example is in the Tintin movie, where the control of the focus of the viewer is used in one scene to draw the eyes deeper into the scene as Tintin wanders through an old room, only to realise... The shadow *in the foreground* is significant. It blew me away with how clever a use of 3D, and effective in making the surprise of a shadowy-figure more visceral, not because it was a jump-scare/throw at the audience moment, but by revealing that the audience had been *looking past* the potential attacker all along. It was brilliant and emotional use of 3D.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But we don't mandate tissue donation legally. We<mask> give parents legal routes to get out of parenting (adoption). You can give birth, place your child for adoption, and have no further obligations of any sort to that child. There is no analogous situation for pregnancy--no one else can carry the fetus for you. [NEWLINE] [NEWLINE] Adoption is not a particularly common outcome for infants born in developed countries today,<mask> it is at least a possibility that the parents have. And that's the point here; we give people the legal right, whether or not they use it. [NEWLINE] [NEWLINE] Regardless,<mask> this argument just doesn't do it for you, that's okay.<mask><mask> arguments about birth being the start of life are totally at odds with<mask> we know about fetal development and learning. We can consider the evidence and come to different conclusions.</s>
Label encoding: <s>But we don't mandate tissue donation legally. We also give parents legal routes to get out of parenting (adoption). You can give birth, place your child for adoption, and have no further obligations of any sort to that child. There is no analogous situation for pregnancy--no one else can carry the fetus for you. [NEWLINE] [NEWLINE] Adoption is not a particularly common outcome for infants born in developed countries today, but it is at least a possibility that the parents have. And that's the point here; we give people the legal right, whether or not they use it. [NEWLINE] [NEWLINE] Regardless, if this argument just doesn't do it for you, that's okay. I think arguments about birth being the start of life are totally at odds with what we know about fetal development and learning. We can consider the evidence and come to different conclusions.</s>
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This will sound a little flippant...<mask><mask> we are going to segregate people based on the expectation of something uncomfortable, then I want young people only gym hours<mask> I can go and expect not to have 75 year old testicles wiggled in my face. [NEWLINE] [NEWLINE] I'd say that not having restrictions in the first place is the answer. In life we have to accept that at times there will be things that make us feel uncomfortable. Banning others from things<mask> of that discomfort is, in my view, morally wrong.<mask> someone is uncomfortable to the point<mask> they will not confront the situation, then perhaps they should seek alternatives. There are women's only gyms,<mask> well<mask> home gym equipment.<mask> to exclude people<mask> a few are not comfortable<mask> thousands of others tolerate the uncomfortableness is wrong.</s>
Label encoding: <s>This will sound a little flippant... but if we are going to segregate people based on the expectation of something uncomfortable, then I want young people only gym hours so I can go and expect not to have 75 year old testicles wiggled in my face. [NEWLINE] [NEWLINE] I'd say that not having restrictions in the first place is the answer. In life we have to accept that at times there will be things that make us feel uncomfortable. Banning others from things because of that discomfort is, in my view, morally wrong. If someone is uncomfortable to the point where they will not confront the situation, then perhaps they should seek alternatives. There are women's only gyms, as well as home gym equipment. But to exclude people because a few are not comfortable when thousands of others tolerate the uncomfortableness is wrong.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this partially depends on the parents. I was about 9<mask> my parents first sent me for guitar lessons, and I really wasn't that interested in it,<mask> I was forced to stick with it.<mask> I reached my teens<mask> I was pleased to have had that push and started to appreciate the wonderful gift that my parents had given me and really began to appreciate music. [NEWLINE] [NEWLINE] At this age it was completely my choice whether to carry on or not (I did), and I do think trying to force older children to do something that they have honestly tried and not enjoyed is a waste of time and money.<mask>, by letting me start younger<mask> I had basically no pressure on time and the lessons were cheaper gave me a great start and something that now gives me hours of enjoyment and relief<mask> a working adult in my late twenties.</s>
Label encoding: <s>I think this partially depends on the parents. I was about 9 when my parents first sent me for guitar lessons, and I really wasn't that interested in it, however I was forced to stick with it. When I reached my teens however I was pleased to have had that push and started to appreciate the wonderful gift that my parents had given me and really began to appreciate music. [NEWLINE] [NEWLINE] At this age it was completely my choice whether to carry on or not (I did), and I do think trying to force older children to do something that they have honestly tried and not enjoyed is a waste of time and money. However, by letting me start younger when I had basically no pressure on time and the lessons were cheaper gave me a great start and something that now gives me hours of enjoyment and relief as a working adult in my late twenties.</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This might be getting over my head.<mask> it is true that Descartes pointed out that you cannot doubt the fact that you are thinking.<mask> yes "the fact that you are thinking" is the only thing you can think of without faith. [NEWLINE] [NEWLINE] <mask> at the same time this is the only thing you can be sure of.<mask> this means that you can't be sure that God exists and you can't be sure that 1+1 exists. Maybe there is some bad or sadistic god that puts these wrong thoughts in our head and make it all make sense. [NEWLINE] [NEWLINE] (This is all very theoretical) [I<mask> found this<mask> didn't have time to read it]( [URL].blogspot.be/2011/04/criticisms-to-descartes-cogito-ie-i.html)</s>
Label encoding: <s>This might be getting over my head. But it is true that Descartes pointed out that you cannot doubt the fact that you are thinking. So yes "the fact that you are thinking" is the only thing you can think of without faith. [NEWLINE] [NEWLINE] But at the same time this is the only thing you can be sure of. So this means that you can't be sure that God exists and you can't be sure that 1+1 exists. Maybe there is some bad or sadistic god that puts these wrong thoughts in our head and make it all make sense. [NEWLINE] [NEWLINE] (This is all very theoretical) [I also found this but didn't have time to read it]( [URL].blogspot.be/2011/04/criticisms-to-descartes-cogito-ie-i.html)</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>your position rests on the assumption that most people, given ideal education and social environments, have the ability to do just about any job<mask> well<mask> anyone else. this assumption runs counter to the available data regarding the heritability of psychological traits and limits of environmental intervention. [NEWLINE] [NEWLINE] the reason this is significant is that there is a solid chance a great number of people will not have the interest or ability to transfer into these new areas of the economy no matter<mask> programs are in place to assist them. even in the wildest post scarcity economic dream, the idea that each person has the ability, creativity, and will to provide valued service seems pretty far fetched. [NEWLINE] [NEWLINE] <mask><mask> do we do to feed these people? ever more programs for this and that, further bloating our bureaucracy and keeping money from the people that need it to live?</s>
Label encoding: <s>your position rests on the assumption that most people, given ideal education and social environments, have the ability to do just about any job as well as anyone else. this assumption runs counter to the available data regarding the heritability of psychological traits and limits of environmental intervention. [NEWLINE] [NEWLINE] the reason this is significant is that there is a solid chance a great number of people will not have the interest or ability to transfer into these new areas of the economy no matter what programs are in place to assist them. even in the wildest post scarcity economic dream, the idea that each person has the ability, creativity, and will to provide valued service seems pretty far fetched. [NEWLINE] [NEWLINE] so what do we do to feed these people? ever more programs for this and that, further bloating our bureaucracy and keeping money from the people that need it to live?</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me make sure I understand you right: Humanity doesn't need God,<mask> someday we might be able to build a society that's better than heaven and resurrect everyone that ever lived. Is that right? [NEWLINE] [NEWLINE] There are a handful of issues with your reasoning. [NEWLINE] [NEWLINE] 1. Your idea is WAY out there. We aren't even close to the kind of tech required to pull of your ideas, not to mention it's probably just impossible. [NEWLINE] [NEWLINE] 2. [STARTQ] All you need are the data, and you can reconstruct anything. [ENDQ] [NEWLINE] We're missing data.<mask> you or I die, society probably isn't going to save enough of our data to resurrect us. [NEWLINE] [NEWLINE] 3.<mask> God does exist, then some sort of soul does<mask> well, which would make truly resurrecting someone impossible. [NEWLINE] [NEWLINE] Edit: formatting</s>
Label encoding: <s>Let me make sure I understand you right: Humanity doesn't need God, because someday we might be able to build a society that's better than heaven and resurrect everyone that ever lived. Is that right? [NEWLINE] [NEWLINE] There are a handful of issues with your reasoning. [NEWLINE] [NEWLINE] 1. Your idea is WAY out there. We aren't even close to the kind of tech required to pull of your ideas, not to mention it's probably just impossible. [NEWLINE] [NEWLINE] 2. [STARTQ] All you need are the data, and you can reconstruct anything. [ENDQ] [NEWLINE] We're missing data. If you or I die, society probably isn't going to save enough of our data to resurrect us. [NEWLINE] [NEWLINE] 3. If God does exist, then some sort of soul does as well, which would make truly resurrecting someone impossible. [NEWLINE] [NEWLINE] Edit: formatting</s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As I said on that post, you don't need to use the alienating language of privilege to quickly explain<mask> you don't agree with someone. You can just cite your different experiences and move on: "You think this isn't offensive,<mask> I've experienced this more than you have and think it is." [NEWLINE] [NEWLINE] Privilege is one-sided; by definition only white people can experience racial privilege. Do you see<mask> white people, or men, or straight people, might interpret "check your privilege"<mask> "we don't care<mask> you have to say?" There isn't a way to continue discussion after that. [NEWLINE] [NEWLINE] Is there a good way to prove that one has sufficiently checked their privilege? Without one, I don't see<mask> someone who has a good-faith disagreement can voice it unless they are a minority.</s>
Label encoding: <s>As I said on that post, you don't need to use the alienating language of privilege to quickly explain why you don't agree with someone. You can just cite your different experiences and move on: "You think this isn't offensive, but I've experienced this more than you have and think it is." [NEWLINE] [NEWLINE] Privilege is one-sided; by definition only white people can experience racial privilege. Do you see why white people, or men, or straight people, might interpret "check your privilege" as "we don't care what you have to say?" There isn't a way to continue discussion after that. [NEWLINE] [NEWLINE] Is there a good way to prove that one has sufficiently checked their privilege? Without one, I don't see how someone who has a good-faith disagreement can voice it unless they are a minority.</s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're arguing for a fourth possibility: "Declare poor people to be unfit, task CPS with taking away their kids." [NEWLINE] [NEWLINE] Personally,<mask><mask> that's worse than our current system, for any number of reasons. <mask> this is a preference issue.  You could be working from different ethics. [NEWLINE] [NEWLINE] <mask>, even<mask> your preferences differ, the "order matters" rule still holds.  Absent abuse, CPS would allow children to live in shocking poverty.  (Check out the medicaid limits to see<mask> we allow.  People live on *less* than that). [NEWLINE] [NEWLINE] Simply removing child-support would move you to the 'worst case'.  To avoid that, you'd need to replace child support with your new system.  This creates a position more nuanced than 'dismantle the current system'.</s>
Label encoding: <s>You're arguing for a fourth possibility: "Declare poor people to be unfit, task CPS with taking away their kids." [NEWLINE] [NEWLINE] Personally, I think that's worse than our current system, for any number of reasons.  But this is a preference issue.  You could be working from different ethics. [NEWLINE] [NEWLINE] However, even if your preferences differ, the "order matters" rule still holds.  Absent abuse, CPS would allow children to live in shocking poverty.  (Check out the medicaid limits to see what we allow.  People live on *less* than that). [NEWLINE] [NEWLINE] Simply removing child-support would move you to the 'worst case'.  To avoid that, you'd need to replace child support with your new system.  This creates a position more nuanced than 'dismantle the current system'.</s>
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Penalties for drug dealers have been ineffective partially<mask> there is<mask> much profit to be made by people with often few other options to make a good living, and partially<mask> dealers can see many other dealers who face no penalty and<mask> believe it very probably that they can escape consequences. [NEWLINE] [NEWLINE] <mask> much<mask> incarceration is of limited effectiveness, at the end of the day, people do evaluate their actions on a risk reward basis, even<mask> they don't think about it too critically.<mask> something has high risk and little reward, you'll reduce it (barring addiction and other complicating factors). [NEWLINE] [NEWLINE] Swatting doesn't have nearly the upside of drug dealing, and it's relatively rare,<mask> harsh sentences can easily create the image of high risk. It's actually a far better candidate than drugs for long sentences to be effective.</s>
Label encoding: <s>Penalties for drug dealers have been ineffective partially because there is so much profit to be made by people with often few other options to make a good living, and partially because dealers can see many other dealers who face no penalty and thus believe it very probably that they can escape consequences. [NEWLINE] [NEWLINE] As much as incarceration is of limited effectiveness, at the end of the day, people do evaluate their actions on a risk reward basis, even when they don't think about it too critically. If something has high risk and little reward, you'll reduce it (barring addiction and other complicating factors). [NEWLINE] [NEWLINE] Swatting doesn't have nearly the upside of drug dealing, and it's relatively rare, so harsh sentences can easily create the image of high risk. It's actually a far better candidate than drugs for long sentences to be effective.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's considered excessively cruel for use in warfare or policing activity, due to the excessive damage it causes,<mask> my understanding is correct.  It's kinda the absurdity of drawing the line of a "humane" way of punching a hole in someone's liver,<mask> whatever. [NEWLINE] [NEWLINE] See my other comment<mask> to<mask> private citizens don't NEED to stand up to predator drones.  TL;DR would be that unless the government is interested in wholesale slaughter of it's citizens, all they need to do it make sure they stay withing explosive range of the general populace, and the drone's effectiveness is reduced significantly.  Once you've done that, forget the drone, shoot the operator.  He's a citizen, just like everyone else, which means he has to go home, or shopping, or elsewhere eventually.  </s>
Label encoding: <s>It's considered excessively cruel for use in warfare or policing activity, due to the excessive damage it causes, if my understanding is correct.  It's kinda the absurdity of drawing the line of a "humane" way of punching a hole in someone's liver, but whatever. [NEWLINE] [NEWLINE] See my other comment as to why private citizens don't NEED to stand up to predator drones.  TL;DR would be that unless the government is interested in wholesale slaughter of it's citizens, all they need to do it make sure they stay withing explosive range of the general populace, and the drone's effectiveness is reduced significantly.  Once you've done that, forget the drone, shoot the operator.  He's a citizen, just like everyone else, which means he has to go home, or shopping, or elsewhere eventually.  </s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0666, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is evidence to suggest that breasts are [not innately sexual.]( [URL] %E2%80%99t-hard-wired-to-find-breasts-attractive/) [NEWLINE] [NEWLINE] Breasts might be an erogenous zone<mask> other erogenous zones include: the mouth, the neck, the abdomen, arms, feet... Are those innately sexual?<mask> do some cultures require the upper arm to be covered?<mask> are people in countries<mask> nudity is more accepted desensitized to it? [NEWLINE] [NEWLINE] <mask>, breasts's *primary function* is breastfeeding. The hormones that make them sensitive (oxytocin) are the same hormones that facilitate breastfeeding and maternal bonding. The clitoris,<mask><mask><mask><mask>, is *only* ever used during sex,<mask> your analogy doesn't make sense. </s>
Label encoding: <s>There is evidence to suggest that breasts are [not innately sexual.]( [URL] %E2%80%99t-hard-wired-to-find-breasts-attractive/) [NEWLINE] [NEWLINE] Breasts might be an erogenous zone but other erogenous zones include: the mouth, the neck, the abdomen, arms, feet... Are those innately sexual? Why do some cultures require the upper arm to be covered? Why are people in countries where nudity is more accepted desensitized to it? [NEWLINE] [NEWLINE] Lastly, breasts's *primary function* is breastfeeding. The hormones that make them sensitive (oxytocin) are the same hormones that facilitate breastfeeding and maternal bonding. The clitoris, on the other hand, is *only* ever used during sex, so your analogy doesn't make sense. </s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] In either case, I feel like it would be beneficial to cause a clear divide between feminists and feminazis in order to garner more support and to have feminist ideas to be taken more seriously. [ENDQ] [NEWLINE] [NEWLINE] This is an impossibility.  Those labels are self imposed.  Most people  - namely those fighting for a cause - won't elect pejorative labels for themselves.  You want to distinguish those who want equality (feminists) by calling them Egalitarian.  Femenazis (who currently go by feminist) would then do<mask>?  Continue to call themselves by an abandoned name<mask> it was the equal opposite of macho?  Probably not.  Those who are perceived to be Femenazis, probably perceive themselves<mask> Feminists. <mask> the title changes, everyone will switch over.</s>
Label encoding: <s> [STARTQ] In either case, I feel like it would be beneficial to cause a clear divide between feminists and feminazis in order to garner more support and to have feminist ideas to be taken more seriously. [ENDQ] [NEWLINE] [NEWLINE] This is an impossibility.  Those labels are self imposed.  Most people  - namely those fighting for a cause - won't elect pejorative labels for themselves.  You want to distinguish those who want equality (feminists) by calling them Egalitarian.  Femenazis (who currently go by feminist) would then do what?  Continue to call themselves by an abandoned name because it was the equal opposite of macho?  Probably not.  Those who are perceived to be Femenazis, probably perceive themselves as Feminists.  When the title changes, everyone will switch over.</s>
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the few men that are still in college [ENDQ] [NEWLINE] This is a ridiculous and unfounded exaggeration, and you know it. [NEWLINE] [NEWLINE] <mask>, the [Census Bureau]( [URL] %203.xlsx) reports that 14.2% of women have a Bachelor's,<mask> 14.7% of men have a Bachelor's, 1.1% of women have a professional degree<mask> 1.9% of men have one, and 1.2% of women have a doctorate<mask> 2.2% of men have one. Women had 21,715 Bachelor's degrees in 2013, whereas men had a total of 19,860. That means that women only held 52% of bachelor's degrees, which is not surprising considering [there are more women than men in the United States]( [URL].pdf). </s>
Label encoding: <s> [STARTQ] the few men that are still in college [ENDQ] [NEWLINE] This is a ridiculous and unfounded exaggeration, and you know it. [NEWLINE] [NEWLINE] Also, the [Census Bureau]( [URL] %203.xlsx) reports that 14.2% of women have a Bachelor's, but 14.7% of men have a Bachelor's, 1.1% of women have a professional degree but 1.9% of men have one, and 1.2% of women have a doctorate yet 2.2% of men have one. Women had 21,715 Bachelor's degrees in 2013, whereas men had a total of 19,860. That means that women only held 52% of bachelor's degrees, which is not surprising considering [there are more women than men in the United States]( [URL].pdf). </s>
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I dislike the way they make themselves look,<mask><mask> they're being innovative and changing the way we do everything for the better. [NEWLINE] Remember<mask> they said they simply could *not* take away internet connectivity<mask> "the console was built from the ground up around it",<mask> they managed to change it once everyone voiced their opinions? [NEWLINE] Microsoft with the xbox one is simply not worth it, the price is currently ~450$ for the console, plus a year of gold is 60$.<mask> you pay around 500$ to have the ability to watch netflix (<mask> you pay 8$ for netflix<mask> well) and browse the web. New AAA games are usually around 60-70$,<mask> you're almost at 600$<mask> you want to play a AAA title on xbone, and that's just one game.</s>
Label encoding: <s>I dislike the way they make themselves look, as if they're being innovative and changing the way we do everything for the better. [NEWLINE] Remember when they said they simply could *not* take away internet connectivity because "the console was built from the ground up around it", yet they managed to change it once everyone voiced their opinions? [NEWLINE] Microsoft with the xbox one is simply not worth it, the price is currently ~450$ for the console, plus a year of gold is 60$. So you pay around 500$ to have the ability to watch netflix ( if you pay 8$ for netflix as well) and browse the web. New AAA games are usually around 60-70$, so you're almost at 600$ if you want to play a AAA title on xbone, and that's just one game.</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Rights are a myth, and we need a better model for thinking about personal freedom. [ENDQ] [NEWLINE] I understand your point, and agree that terms like "God-given rights" are flowery language at best and misleading at worst. Obviously many people in the world, and numerous people in America don't have these rights. [NEWLINE] [NEWLINE] <mask>, I do think there's a difference between a myth and a social construction. We live in a society in which Representatives make the rules and, being elected by the people, those representatives have good reason to make those rules in keeping with public will.<mask>, in an imperfect, abstract sense, popular opinion controls<mask> rights we should all have and<mask> rights we should not have. And I would<mask><mask> those rights that society dictates are real, albeit impermanent and intangible. </s>
Label encoding: <s> [STARTQ] Rights are a myth, and we need a better model for thinking about personal freedom. [ENDQ] [NEWLINE] I understand your point, and agree that terms like "God-given rights" are flowery language at best and misleading at worst. Obviously many people in the world, and numerous people in America don't have these rights. [NEWLINE] [NEWLINE] However, I do think there's a difference between a myth and a social construction. We live in a society in which Representatives make the rules and, being elected by the people, those representatives have good reason to make those rules in keeping with public will. So, in an imperfect, abstract sense, popular opinion controls what rights we should all have and what rights we should not have. And I would argue that those rights that society dictates are real, albeit impermanent and intangible. </s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You haven't really participated in this thread,<mask> I'd like to point out federal vs. municipal difference<mask> it comes to location of the discount. [NEWLINE] [NEWLINE] A coffee shop in Seattle might very well decide to have a local firefighters discount, (I can name a few that do.)<mask> should they honor a vacationing firefighter from Omaha?<mask> can he prove he's a firefighter?<mask> should a Seattlite care that a man's employed by a city in Nebraska? [NEWLINE] [NEWLINE] <mask> states may have local branches of National Guard, active duty military members serve the nation<mask> a whole and carry one (1) form of identification that is accepted *everywhere.* [NEWLINE] [NEWLINE] This simplifies the discount process and may point to a succinct reason<mask> they tend to receive a blanket discount in most areas of the US. </s>
Label encoding: <s>You haven't really participated in this thread, but I'd like to point out federal vs. municipal difference when it comes to location of the discount. [NEWLINE] [NEWLINE] A coffee shop in Seattle might very well decide to have a local firefighters discount, (I can name a few that do.) But should they honor a vacationing firefighter from Omaha? How can he prove he's a firefighter? Why should a Seattlite care that a man's employed by a city in Nebraska? [NEWLINE] [NEWLINE] While states may have local branches of National Guard, active duty military members serve the nation as a whole and carry one (1) form of identification that is accepted *everywhere.* [NEWLINE] [NEWLINE] This simplifies the discount process and may point to a succinct reason why they tend to receive a blanket discount in most areas of the US. </s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I really love this. Maybe part of<mask> was unsettling was that I was picking up a little bit on this. For the most part, the crowd didn't seem spirited and fun, the energy was a little darker than that. Just there to numb out, rub on each other (I mean, two people attempting to have sex bumped into me at one point), and forget who they are for a night - not in just an escapist way. Hard to describe. [NEWLINE] [NEWLINE] I don't want to make it sound like I had no fun, I did some wild dancing with the people I came with and lots of laughter and sweating and silliness. I'm all kinds of fun.<mask> the atmosphere of this place (and other clubs I've been to over the years) is just depressing to me in general.</s>
Label encoding: <s>I really love this. Maybe part of what was unsettling was that I was picking up a little bit on this. For the most part, the crowd didn't seem spirited and fun, the energy was a little darker than that. Just there to numb out, rub on each other (I mean, two people attempting to have sex bumped into me at one point), and forget who they are for a night - not in just an escapist way. Hard to describe. [NEWLINE] [NEWLINE] I don't want to make it sound like I had no fun, I did some wild dancing with the people I came with and lots of laughter and sweating and silliness. I'm all kinds of fun. But the atmosphere of this place (and other clubs I've been to over the years) is just depressing to me in general.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you keep saying that "a welfare state is just a burden over the taxpayer" then yeah, you kind of are arguing against the existence of welfare programs. [NEWLINE] [NEWLINE] They provide people with the *food they need to eat*.<mask> successful is anyone going to be at getting themselves out of poverty<mask> they can't even feed themselves or their family? [NEWLINE] [NEWLINE] I suppose they could always turn to a life of crime,<mask> that sounds like it would burden the taxpayers even more. [NEWLINE] [NEWLINE] The point is that welfare isn't "the solution" to poverty, it is "the solution" to people starving in the streets. We need bigger solutions to poverty, of which education is a huge part,<mask> we're not going to get very far into a long term solution<mask> people can't survive in the short term.</s>
Label encoding: <s>If you keep saying that "a welfare state is just a burden over the taxpayer" then yeah, you kind of are arguing against the existence of welfare programs. [NEWLINE] [NEWLINE] They provide people with the *food they need to eat*. How successful is anyone going to be at getting themselves out of poverty if they can't even feed themselves or their family? [NEWLINE] [NEWLINE] I suppose they could always turn to a life of crime, but that sounds like it would burden the taxpayers even more. [NEWLINE] [NEWLINE] The point is that welfare isn't "the solution" to poverty, it is "the solution" to people starving in the streets. We need bigger solutions to poverty, of which education is a huge part, but we're not going to get very far into a long term solution if people can't survive in the short term.</s>
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The majority of your points seem predicated on the idea that a) Basic Income requires a flat tax or rejects progressive taxation, and b) that absolutely every social program must be scrapped for a BI to take effect. [NEWLINE] [NEWLINE] Neither of these are true. Of course there are many conservatives or right-libertarian who would prefer those outcomes,<mask> there are<mask> many liberals and left-libertarians who do not. [NEWLINE] [NEWLINE] The title of your CMV is worded incorrectly. You are arguing against not Basic Income,<mask> the questions of<mask> to pay for it and<mask> exactly would be "replaced." [NEWLINE] [NEWLINE] <mask> that doesn't change your view, I'm really not sure<mask> would.  The post is very well written and laid out, it's just a perfectly shot arrow aimed at the wrong target.</s><pad>
Label encoding: <s>The majority of your points seem predicated on the idea that a) Basic Income requires a flat tax or rejects progressive taxation, and b) that absolutely every social program must be scrapped for a BI to take effect. [NEWLINE] [NEWLINE] Neither of these are true. Of course there are many conservatives or right-libertarian who would prefer those outcomes, but there are also many liberals and left-libertarians who do not. [NEWLINE] [NEWLINE] The title of your CMV is worded incorrectly. You are arguing against not Basic Income, but the questions of how to pay for it and what exactly would be "replaced." [NEWLINE] [NEWLINE] If that doesn't change your view, I'm really not sure what would.  The post is very well written and laid out, it's just a perfectly shot arrow aimed at the wrong target.</s><pad>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Crazed22, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Crazed22+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry Crazed22, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Crazed22+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0565, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Who is the "cause" of being mauled in that scenario, the bear or my actions? [ENDQ] [NEWLINE] Both you and the bear. Your actions were viewed<mask> aggressive by the bear, and the bear responded.<mask>, bears aren't able to make moral decisions<mask> are people. [NEWLINE] [NEWLINE] [STARTQ] Similarly,<mask> a man walks into a street that is known to have petty criminals around, at night, alone, he increases his chances dramatically of being robbed. [ENDQ] [NEWLINE] Did the man walk into a pick-pocketing convention, or is it expected that people use the sidewalk for things other than crime? Did he dare the criminals to rob him? Bear pits aren't intended to hold people,<mask> sidewalks are intended for travel on foot.<mask>, criminals are thinking individuals (unlike bears). [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Who is the "cause" of being mauled in that scenario, the bear or my actions? [ENDQ] [NEWLINE] Both you and the bear. Your actions were viewed as aggressive by the bear, and the bear responded. However, bears aren't able to make moral decisions as are people. [NEWLINE] [NEWLINE] [STARTQ] Similarly, if a man walks into a street that is known to have petty criminals around, at night, alone, he increases his chances dramatically of being robbed. [ENDQ] [NEWLINE] Did the man walk into a pick-pocketing convention, or is it expected that people use the sidewalk for things other than crime? Did he dare the criminals to rob him? Bear pits aren't intended to hold people, but sidewalks are intended for travel on foot. Also, criminals are thinking individuals (unlike bears). [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry AhmadSahrab, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=AhmadSahrab+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad><pad>
Label encoding: <s>Sorry AhmadSahrab, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=AhmadSahrab+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad><pad>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You cannot use a principle of liberal social democracy, and then turn around [ENDQ] [NEWLINE] <mask> not?  Marx believed that capitalism is a vital stage in the development of communism.  Just<mask> capitalism was necessary at one point does not mean it is necessary at all points. [NEWLINE] [NEWLINE] Likewise, Lenin said "The bourgeoisie is many times stronger than we. To give it the weapon of freedom of the press is to ease the enemy’s cause, to help the class enemy. We do not desire to end in suicide,<mask> we will not do this." [NEWLINE] [NEWLINE] For most liberals, freedom of speech is a vital good.  For certain people who have a very clear vision of<mask> progress should be, freedom of speech is more like a bus ride:<mask> it arrives at your stop, you get off.</s>
Label encoding: <s> [STARTQ] You cannot use a principle of liberal social democracy, and then turn around [ENDQ] [NEWLINE] Why not?  Marx believed that capitalism is a vital stage in the development of communism.  Just because capitalism was necessary at one point does not mean it is necessary at all points. [NEWLINE] [NEWLINE] Likewise, Lenin said "The bourgeoisie is many times stronger than we. To give it the weapon of freedom of the press is to ease the enemy’s cause, to help the class enemy. We do not desire to end in suicide, so we will not do this." [NEWLINE] [NEWLINE] For most liberals, freedom of speech is a vital good.  For certain people who have a very clear vision of what progress should be, freedom of speech is more like a bus ride: when it arrives at your stop, you get off.</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Have you taught many kids? Counting in 5s is *much* easier for most of them than counting to 60 and then remembering to start over. [NEWLINE] [NEWLINE] <mask><mask> would they have to count in 5s anyway? Most clocks do have indicators in between<mask> there's no reason they can't count normally. The different is with an analog clock they can visualise the whole thing, it's all in front of their eyes<mask> they don't have to memorise, they can keep working it out each time until it becomes more natural. [NEWLINE] [NEWLINE] For the average 6 year old an hour could be anywhere between 5 minutes and a day depending on<mask> desperately they're looking forward to whatever's at the end of that hour. And no parents are technically accurate<mask> they tell their kids '5 more minutes'.</s>
Label encoding: <s>Have you taught many kids? Counting in 5s is *much* easier for most of them than counting to 60 and then remembering to start over. [NEWLINE] [NEWLINE] Also why would they have to count in 5s anyway? Most clocks do have indicators in between so there's no reason they can't count normally. The different is with an analog clock they can visualise the whole thing, it's all in front of their eyes so they don't have to memorise, they can keep working it out each time until it becomes more natural. [NEWLINE] [NEWLINE] For the average 6 year old an hour could be anywhere between 5 minutes and a day depending on how desperately they're looking forward to whatever's at the end of that hour. And no parents are technically accurate when they tell their kids '5 more minutes'.</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We pay people based on their worth and<mask> much potential money they can pull in.<mask> you are a good surgeon is everyone gonna follow you on instagram and buy anything that you advertise? Nope [NEWLINE] [NEWLINE] I would<mask><mask> being a Doctor or a engineer is more about being capable and meeting certain qualifications. In sports we reward people who are the best players on the planet, it is a much more competitive scene and<mask> I would argue they deserve to be payed<mask> much<mask> they do [NEWLINE] [NEWLINE] <mask><mask> I would say that<mask> comparing pro athletes you should only compare them to the top engineers and doctors. Unfortunately we don't create a ranking system for the best doctors and engineers on the planet, i'm sure some of them out there are getting payed<mask> much<mask> not more than the pro athletes </s>
Label encoding: <s>We pay people based on their worth and how much potential money they can pull in. If you are a good surgeon is everyone gonna follow you on instagram and buy anything that you advertise? Nope [NEWLINE] [NEWLINE] I would argue that being a Doctor or a engineer is more about being capable and meeting certain qualifications. In sports we reward people who are the best players on the planet, it is a much more competitive scene and therefore I would argue they deserve to be payed as much as they do [NEWLINE] [NEWLINE] In fact I would say that when comparing pro athletes you should only compare them to the top engineers and doctors. Unfortunately we don't create a ranking system for the best doctors and engineers on the planet, i'm sure some of them out there are getting payed as much if not more than the pro athletes </s>
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Harvard, off tuition alone, makes roughly $112,000,000 from each class [ENDQ] [NEWLINE] Did you factor in financial aid? Most Harvard students don't post full tuition. I get 77M from [URL].  Harvard has a 37B endowment;<mask> they make 1% interest off it, that's almost 5 times tuition. [NEWLINE] [NEWLINE] [STARTQ] College seems like a glorified high school to me. [ENDQ] [NEWLINE] College of much more that high school.  I don't know<mask> high school you go to our<mask> college you're going to go to,<mask> in general, you will have less busywork and more thought provoking assignments.  Many schools expect/encourage collaboration in a way that high schools don't.  Furthermore, the social aspects of college make for an altogether different experience</s>
Label encoding: <s> [STARTQ] Harvard, off tuition alone, makes roughly $112,000,000 from each class [ENDQ] [NEWLINE] Did you factor in financial aid? Most Harvard students don't post full tuition. I get 77M from [URL].  Harvard has a 37B endowment; if they make 1% interest off it, that's almost 5 times tuition. [NEWLINE] [NEWLINE] [STARTQ] College seems like a glorified high school to me. [ENDQ] [NEWLINE] College of much more that high school.  I don't know what high school you go to our what college you're going to go to, but in general, you will have less busywork and more thought provoking assignments.  Many schools expect/encourage collaboration in a way that high schools don't.  Furthermore, the social aspects of college make for an altogether different experience</s>
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People don't like it<mask> other people disagree with them, especially on matters<mask> "important"<mask> life after death, fate, determinism, free will, the existence of God, morality, politics, etc. These are subjects that arouse strong feelings.<mask><mask> that's<mask> you were being called a moron. People feel threatened<mask> others disagree on those subjects. [NEWLINE] [NEWLINE] The actual teachings of some religious communities are actually quite peaceful and friendly, e.g. "love thy neighbor<mask> thyself" and whatnot. The animosity comes from the fact that you disagree about the nature of reality...<mask> it's not really taught by the religion necessarily. [NEWLINE] [NEWLINE] On a side note, you speak absolutely fantastic English. Education in the Czech Republic is not failing you, that's for sure.</s>
Label encoding: <s>People don't like it when other people disagree with them, especially on matters as "important" as life after death, fate, determinism, free will, the existence of God, morality, politics, etc. These are subjects that arouse strong feelings. I think that's why you were being called a moron. People feel threatened when others disagree on those subjects. [NEWLINE] [NEWLINE] The actual teachings of some religious communities are actually quite peaceful and friendly, e.g. "love thy neighbor as thyself" and whatnot. The animosity comes from the fact that you disagree about the nature of reality... but it's not really taught by the religion necessarily. [NEWLINE] [NEWLINE] On a side note, you speak absolutely fantastic English. Education in the Czech Republic is not failing you, that's for sure.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] he's not sexist<mask> he insults men too [ENDQ] [NEWLINE] <mask> does this make him both a misogynist and a misandrist simultaneously? [NEWLINE] [NEWLINE] <mask>,<mask> for his caring only about women's looks:<mask><mask> that's a horrible exaggeration. Look up videos of him insulting women online. Look up<mask> he said about Hillary Clinton. He insulted her policies, her actions. [NEWLINE] [NEWLINE] And, no matter<mask> the media says, a half-dozen cherry-picked examples from the last several *years* do not constitute proof that he only values women's bodies. [NEWLINE] [NEWLINE] Edit: and to be clear, I do not like Trump. Not at all. I don't like his words, it don't like his policies.<mask> I<mask> don't like the media misrepresenting him.</s>
Label encoding: <s> [STARTQ] he's not sexist because he insults men too [ENDQ] [NEWLINE] So does this make him both a misogynist and a misandrist simultaneously? [NEWLINE] [NEWLINE] Also, as for his caring only about women's looks: I think that's a horrible exaggeration. Look up videos of him insulting women online. Look up what he said about Hillary Clinton. He insulted her policies, her actions. [NEWLINE] [NEWLINE] And, no matter what the media says, a half-dozen cherry-picked examples from the last several *years* do not constitute proof that he only values women's bodies. [NEWLINE] [NEWLINE] Edit: and to be clear, I do not like Trump. Not at all. I don't like his words, it don't like his policies. But I also don't like the media misrepresenting him.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because it often doesn't lift all dysphoria. Sometimes its just not enough. Just for your info, not all transgender people seek surgery. Many are happy with their genitals, and just hormone therapy.<mask> for those who are not, surgery is an often the final straw.<mask> draw the line somewhere. HRT changes the body, and its the thing people want out of HRT. Its the way HRT helps people.<mask> draw the line there?<mask> is it different to surgery? [NEWLINE] [NEWLINE] Edit: For a more personal experience. HRT did wonders for me, and i am statsified in the way it helped me.<mask> it didn't lift my genital dysphoria. And that still sucks.<mask> i will seek surgery in the future, to lift that problem<mask> well.</s>
Label encoding: <s>Because it often doesn't lift all dysphoria. Sometimes its just not enough. Just for your info, not all transgender people seek surgery. Many are happy with their genitals, and just hormone therapy. But for those who are not, surgery is an often the final straw. Why draw the line somewhere. HRT changes the body, and its the thing people want out of HRT. Its the way HRT helps people. Why draw the line there? Why is it different to surgery? [NEWLINE] [NEWLINE] Edit: For a more personal experience. HRT did wonders for me, and i am statsified in the way it helped me. But it didn't lift my genital dysphoria. And that still sucks. So i will seek surgery in the future, to lift that problem as well.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It might seem strange<mask> some people still see property<mask> something evil. [NEWLINE] [NEWLINE] <mask> look at your last statement: [NEWLINE] [NEWLINE] A right to live, freedom from bodily harm and freedom of association are easy to support,<mask> you see your body<mask> property of yourself. [NEWLINE] [NEWLINE] <mask> the latter two points are more difficult<mask> of that two: [NEWLINE] [NEWLINE] You may have a right to your property, meaning that other people aren't free to take it away from you,<mask> you aren't free to take away things from other people to protect your property. (This is<mask> the state fails,<mask> this is another topic) [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] <mask> you don't have to change your views about human rights,<mask> you just got a logically consistent explanation<mask> you hold onto such human rights.</s>
Label encoding: <s>It might seem strange because some people still see property as something evil. [NEWLINE] [NEWLINE] But look at your last statement: [NEWLINE] [NEWLINE] A right to live, freedom from bodily harm and freedom of association are easy to support, if you see your body as property of yourself. [NEWLINE] [NEWLINE] Also the latter two points are more difficult because of that two: [NEWLINE] [NEWLINE] You may have a right to your property, meaning that other people aren't free to take it away from you, but you aren't free to take away things from other people to protect your property. (This is where the state fails, but this is another topic) [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] So you don't have to change your views about human rights, but you just got a logically consistent explanation why you hold onto such human rights.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆  This is the best answer I have heard<mask> far<mask> it offers a clear solution to the problem of removing legal protections<mask> the government were to just stop being involved.  Under this system people are still protected, and there is no authority that can prevent anybody from full participation in a'marriage,' whatever that happens to mean to people at the time. <mask> some may still quibble over the definition of the word, this will still be the most inclusionary system,<mask><mask><mask> the government is willing to offer the same protections and benefits to all consenting adults to desire to take part in the partnership registration. [NEWLINE] [NEWLINE] I suspect that there are some who would still try to limit equality in this area,<mask> that will likely happen no matter<mask> solution is proposed. </s>
Label encoding: <s>∆  This is the best answer I have heard so far because it offers a clear solution to the problem of removing legal protections if the government were to just stop being involved.  Under this system people are still protected, and there is no authority that can prevent anybody from full participation in a'marriage,' whatever that happens to mean to people at the time.  While some may still quibble over the definition of the word, this will still be the most inclusionary system, as long as the government is willing to offer the same protections and benefits to all consenting adults to desire to take part in the partnership registration. [NEWLINE] [NEWLINE] I suspect that there are some who would still try to limit equality in this area, but that will likely happen no matter what solution is proposed. </s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You really think a 5'1" girl/woman/grandmother can protect herself with her hands against a 6'5" muscle bound man?  Maybe there are some professional fighters who could do<mask>,<mask> not the average person, without huge amounts of training.  And<mask><mask> you add a knife (or even a shiv) to the mix? [NEWLINE] [NEWLINE] A gun is the great equalizer.  It requires minimal training to use adequately and effectively.  It is just<mask> lethal in the hands of a novice<mask> a professional.  It gives a small and frail person the same defensive capabilities<mask> a hulking mass of a man. [NEWLINE] [NEWLINE] I am a fan of the quote: [NEWLINE] [NEWLINE] &gt;God made men.  Samuel Colt made men equal.</s>
Label encoding: <s>You really think a 5'1" girl/woman/grandmother can protect herself with her hands against a 6'5" muscle bound man?  Maybe there are some professional fighters who could do so, but not the average person, without huge amounts of training.  And what if you add a knife (or even a shiv) to the mix? [NEWLINE] [NEWLINE] A gun is the great equalizer.  It requires minimal training to use adequately and effectively.  It is just as lethal in the hands of a novice as a professional.  It gives a small and frail person the same defensive capabilities as a hulking mass of a man. [NEWLINE] [NEWLINE] I am a fan of the quote: [NEWLINE] [NEWLINE] &gt;God made men.  Samuel Colt made men equal.</s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sure most reddit users have noticed this bias. It will probably hit your frontpage, no matter<mask> you are subbed to. [NEWLINE] [NEWLINE] The question is:<mask> does this bias become *hate*. This,<mask> I see it, is the only subjective part.<mask> go ahead and try to change my view! [NEWLINE] [NEWLINE] [NEWLINE] <mask> it looks like I need to say more. Here are a few of my reasons: [NEWLINE] [NEWLINE] /r/Bad_Cop_No_Donut has four times<mask> many subscribers<mask> /r/Good_Cop_Free_Donut *and* /r/ProtectAndServe combined. [NEWLINE] [NEWLINE] After searching "cop" using reddits search function. The second result is an /r/AskReddit thread about moronic cops.</s>
Label encoding: <s>I'm sure most reddit users have noticed this bias. It will probably hit your frontpage, no matter what you are subbed to. [NEWLINE] [NEWLINE] The question is: when does this bias become *hate*. This, as I see it, is the only subjective part. So go ahead and try to change my view! [NEWLINE] [NEWLINE] [NEWLINE] So it looks like I need to say more. Here are a few of my reasons: [NEWLINE] [NEWLINE] /r/Bad_Cop_No_Donut has four times as many subscribers as /r/Good_Cop_Free_Donut *and* /r/ProtectAndServe combined. [NEWLINE] [NEWLINE] After searching "cop" using reddits search function. The second result is an /r/AskReddit thread about moronic cops.</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you've taken the first and most important step in your search for understanding. <mask> has been mentioned already, those who take the time to question religion are the ones who find meaningful answers in the end.  Unfortunately,<mask> many can guide you toward the right answers, ultimately it is your journey and experiences in life that will help you in your ability to grow closer to Christ. [NEWLINE] To start, I'd suggest looking into Christian debaters and reading historical accounts (including the Bible of course) to build a foundation of credible "evidence"<mask> you will. [NEWLINE] <mask>, know that you and I and everyone around you have many of the same questions.  These really are the most important questions to ask. <mask><mask> the popularity of this post is testament enough for that.</s>
Label encoding: <s>I think you've taken the first and most important step in your search for understanding.  As has been mentioned already, those who take the time to question religion are the ones who find meaningful answers in the end.  Unfortunately, although many can guide you toward the right answers, ultimately it is your journey and experiences in life that will help you in your ability to grow closer to Christ. [NEWLINE] To start, I'd suggest looking into Christian debaters and reading historical accounts (including the Bible of course) to build a foundation of credible "evidence" if you will. [NEWLINE] Also, know that you and I and everyone around you have many of the same questions.  These really are the most important questions to ask.  I think the popularity of this post is testament enough for that.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, I have the feeling that such self-censorship can go too far. A very similar example is something I remember to be a quite frequent occurrence in elementary school. A kid comes to the teacher and says "<mask> and<mask> called me a faggot" and the teacher's first reaction is to reprimand the victim for repeating the "bad word". [NEWLINE] [NEWLINE] Now, it can be argued that at such a young age this is a deterrent from making any such words habitual (and I do mean that it can be argued<mask> making it 'forbidden' can give a kind of appeal to young kids and actually explaining *<mask> * it is bad would be much better<mask><mask> ).<mask> we are adults and should have passed the "don't use the bad word!" phase.</s><pad>
Label encoding: <s>Well, I have the feeling that such self-censorship can go too far. A very similar example is something I remember to be a quite frequent occurrence in elementary school. A kid comes to the teacher and says " so and so called me a faggot" and the teacher's first reaction is to reprimand the victim for repeating the "bad word". [NEWLINE] [NEWLINE] Now, it can be argued that at such a young age this is a deterrent from making any such words habitual (and I do mean that it can be argued because making it 'forbidden' can give a kind of appeal to young kids and actually explaining * why * it is bad would be much better IMHO ). But we are adults and should have passed the "don't use the bad word!" phase.</s><pad>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0674, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think<mask> you describe gives doctors a monopoly on information and expresses a very patriarchal view of medicine (I.e., "just do<mask> the doctor tells you"). Doctors are human and suffer from the same limitations anyone else does. They don't know everything and they make mistakes. A patient should understand that, generally speaking, her doctor knows more than her<mask> it comes to medicine,<mask> the best healthcare outcomes are achieved<mask> the patient is actively involved in her own care, instead of blindly following directions. In my view, anything that helps a patient gather information about her own health<mask> that she can make informed decisions is a plus.<mask><mask><mask> it's a fine line and care should be taken to make sure that the drug companies don't oversell their medications in misleading ads. </s>
Label encoding: <s>I think what you describe gives doctors a monopoly on information and expresses a very patriarchal view of medicine (I.e., "just do what the doctor tells you"). Doctors are human and suffer from the same limitations anyone else does. They don't know everything and they make mistakes. A patient should understand that, generally speaking, her doctor knows more than her when it comes to medicine, but the best healthcare outcomes are achieved when the patient is actively involved in her own care, instead of blindly following directions. In my view, anything that helps a patient gather information about her own health so that she can make informed decisions is a plus. However I think it's a fine line and care should be taken to make sure that the drug companies don't oversell their medications in misleading ads. </s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> after a person dies, their spirit continues. [ENDQ] [NEWLINE] I hope this is true. Especially the part about still having a voice and being able to speak even after death. In some ways,<mask><mask> being a spirit or ghost might be even more comforting and peaceful than going to a "heaven"<mask> I would be able to stay here in the universe I've come to love. [NEWLINE] [NEWLINE] I'm not<mask> convinced it's possible,<mask>,<mask> I haven't had enough time to watch the movie. I'm not actually sure I'll get the chance anytime soon either,<mask> I apologize for that.<mask> one day I watch it and it changes my belief, I'll have to track down the post and give you a delta. [NEWLINE] [NEWLINE] Thank you. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] But after a person dies, their spirit continues. [ENDQ] [NEWLINE] I hope this is true. Especially the part about still having a voice and being able to speak even after death. In some ways, I think being a spirit or ghost might be even more comforting and peaceful than going to a "heaven" because I would be able to stay here in the universe I've come to love. [NEWLINE] [NEWLINE] I'm not yet convinced it's possible, though, but I haven't had enough time to watch the movie. I'm not actually sure I'll get the chance anytime soon either, so I apologize for that. If one day I watch it and it changes my belief, I'll have to track down the post and give you a delta. [NEWLINE] [NEWLINE] Thank you. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As much<mask> it may seem like it now; depression is not a terminal diagnosis. <mask> is such a final solution really necessary for a temporary problem?  Don't you want to find out<mask> you can get better? <mask> you weren't<mask> depressed anymore, you'd likely enjoy being alive.  Most people who have failed a suicide attempt say they are glad they failed... eventually. You've not reached that eventually<mask><mask> you're still struggling.  This is a real battle,<mask> it's one you can persevere through, just<mask> others have before you.  Don't deny yourself the chance at a future just<mask> the present sucks royally: fight for future you.  Learn to love yourself<mask> you do others.  Please don't give up.</s>
Label encoding: <s>As much as it may seem like it now; depression is not a terminal diagnosis.  So is such a final solution really necessary for a temporary problem?  Don't you want to find out if you can get better?  If you weren't so depressed anymore, you'd likely enjoy being alive.  Most people who have failed a suicide attempt say they are glad they failed... eventually. You've not reached that eventually yet since you're still struggling.  This is a real battle, but it's one you can persevere through, just as others have before you.  Don't deny yourself the chance at a future just because the present sucks royally: fight for future you.  Learn to love yourself as you do others.  Please don't give up.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A great video covers this topic, the main justification for the bombings was that it was the 'lesser' of two evils, whether we would bomb 2 Japanese cities, or invade Japan. [NEWLINE] [NEWLINE] The latter of which would require a lot more effort on our part. We (assuming we are both from the US) would lose *a lot* of resources, including people, from invading. Japan would potentially lose a lot more people, whether or not they were soldiers or civilians. It would overall change the stability of the country for a long ass time too, the nation may not even be the same today<mask> we invaded them. Japan recovered fairly fast from the bomb -<mask> a full out invasion would have taken a lot more time, and a lot more people potentially. [NEWLINE] </s>
Label encoding: <s>A great video covers this topic, the main justification for the bombings was that it was the 'lesser' of two evils, whether we would bomb 2 Japanese cities, or invade Japan. [NEWLINE] [NEWLINE] The latter of which would require a lot more effort on our part. We (assuming we are both from the US) would lose *a lot* of resources, including people, from invading. Japan would potentially lose a lot more people, whether or not they were soldiers or civilians. It would overall change the stability of the country for a long ass time too, the nation may not even be the same today if we invaded them. Japan recovered fairly fast from the bomb - but a full out invasion would have taken a lot more time, and a lot more people potentially. [NEWLINE] </s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this view doesn't really warrant a particularly compelling counterargument<mask> no one in their right mind is going to<mask><mask> absolutely everyone on a desktop must run a Linux distribution. Like, of course not -<mask> about older people who can barely work a computer and have enough trouble with OSX?<mask> about hardcore gamers whose library would shrink<mask> they were forced off of Windows? Nobody would want to force a user onto Linux, especially given<mask> complicated it is at first and<mask> many people have serious fears of computers. [NEWLINE] [NEWLINE] Have you actually run into anyone who's argued that absolutely everyone should use Linux? I personally can't imagine that person exists outside of a frequenter of a hardcore Ubuntu forum who assumes that everyone's<mask> used to using the command line<mask> he is.</s>
Label encoding: <s>I think this view doesn't really warrant a particularly compelling counterargument because no one in their right mind is going to argue that absolutely everyone on a desktop must run a Linux distribution. Like, of course not - what about older people who can barely work a computer and have enough trouble with OSX? What about hardcore gamers whose library would shrink if they were forced off of Windows? Nobody would want to force a user onto Linux, especially given how complicated it is at first and how many people have serious fears of computers. [NEWLINE] [NEWLINE] Have you actually run into anyone who's argued that absolutely everyone should use Linux? I personally can't imagine that person exists outside of a frequenter of a hardcore Ubuntu forum who assumes that everyone's as used to using the command line as he is.</s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The police officer didn't know about the robbery or that the person shot was the suspect<mask><mask> the police chief. [ENDQ] [NEWLINE] That may not be true. [NEWLINE] [NEWLINE] Reports are coming out that<mask> Wilson didn't know about the robbery during the initial confrontation, he did get a radio transmission describing the robbery and the suspects which prompted him to confront Wilson and Johnson a second time. It does explain eyewitness reports that Wilson backed up his car after initially driving off. [NEWLINE] [NEWLINE] From [URL] / [NEWLINE] [NEWLINE] "In Josie's version, Wilson may have heard a call about a strong-arm robbery and saw the young men carrying something that might have been stolen cigars." [NEWLINE] [NEWLINE] Still need more info. Allegedly there are over a dozen eyewitnesses who back up the police report.</s>
Label encoding: <s> [STARTQ] The police officer didn't know about the robbery or that the person shot was the suspect according to the police chief. [ENDQ] [NEWLINE] That may not be true. [NEWLINE] [NEWLINE] Reports are coming out that while Wilson didn't know about the robbery during the initial confrontation, he did get a radio transmission describing the robbery and the suspects which prompted him to confront Wilson and Johnson a second time. It does explain eyewitness reports that Wilson backed up his car after initially driving off. [NEWLINE] [NEWLINE] From [URL] / [NEWLINE] [NEWLINE] "In Josie's version, Wilson may have heard a call about a strong-arm robbery and saw the young men carrying something that might have been stolen cigars." [NEWLINE] [NEWLINE] Still need more info. Allegedly there are over a dozen eyewitnesses who back up the police report.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> you would rather die than stay the rest of your life in prison doesn't mean that other people would. I someone wants to kill himself they can do it in prison<mask> well,<mask> it should be their choice. [NEWLINE] [NEWLINE] Other than that, one might find peace in strange places like prison. He should have that right<mask><mask><mask> he doesn't hurt anyone. He would have a lot of time for introspection. In a sense the person in prison for life is more free than us, given he has a lot of time to do nothing.<mask> we try to do nothing for too long we get ridiculed, homeless, get diseases and<mask> on. I know some people cannot stand to do nothing<mask> there are other people who search this in life like monks.</s>
Label encoding: <s>Just because you would rather die than stay the rest of your life in prison doesn't mean that other people would. I someone wants to kill himself they can do it in prison as well, but it should be their choice. [NEWLINE] [NEWLINE] Other than that, one might find peace in strange places like prison. He should have that right as long as he doesn't hurt anyone. He would have a lot of time for introspection. In a sense the person in prison for life is more free than us, given he has a lot of time to do nothing. If we try to do nothing for too long we get ridiculed, homeless, get diseases and so on. I know some people cannot stand to do nothing but there are other people who search this in life like monks.</s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Different than the person you originally replied to. [NEWLINE] [NEWLINE] [STARTQ] It sounded like you were claiming that six-year-olds would never be exposed to the word "botanist" in a context<mask> the meaning wasn't clear. [ENDQ] [NEWLINE] Perhaps they would or wouldn't. I don't know that it would matter whether a 6 year old would know<mask> a botanist was on first sight or not. By the time it *did* matter that they know<mask> a botanist is, they'd find out. [NEWLINE] [NEWLINE] <mask><mask> the point ButtaBeButtaFree was trying to make is that<mask> your examples may *seem* "easier" they are largely irrelevant<mask> English speakers do not actually have a hard time figuring out<mask> words mean.</s><pad>
Label encoding: <s>Different than the person you originally replied to. [NEWLINE] [NEWLINE] [STARTQ] It sounded like you were claiming that six-year-olds would never be exposed to the word "botanist" in a context where the meaning wasn't clear. [ENDQ] [NEWLINE] Perhaps they would or wouldn't. I don't know that it would matter whether a 6 year old would know what a botanist was on first sight or not. By the time it *did* matter that they know what a botanist is, they'd find out. [NEWLINE] [NEWLINE] I think the point ButtaBeButtaFree was trying to make is that while your examples may *seem* "easier" they are largely irrelevant as English speakers do not actually have a hard time figuring out what words mean.</s><pad>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's pretty difficult to display a medieval, male-dominated era and pretend that females have control over their lives. Daughters are bartered<mask> wives, wives are treated like slaves, etc. It's impossible to pretend that at some point in our history women weren't entirely subservient to men. Relatively speaking, women have only had rights and true agency for an incredibly tiny fraction of the history of mankind. [NEWLINE] [NEWLINE] <mask>, I don't really think it was "ignoring" the character. It's an interesting point,<mask><mask><mask> it was actually to take a step back from the scene. It was graphic enough<mask> it was,<mask><mask> literally the entire scene focused solely on her for the duration that it went, it would be even more traumatic.</s>
Label encoding: <s>It's pretty difficult to display a medieval, male-dominated era and pretend that females have control over their lives. Daughters are bartered as wives, wives are treated like slaves, etc. It's impossible to pretend that at some point in our history women weren't entirely subservient to men. Relatively speaking, women have only had rights and true agency for an incredibly tiny fraction of the history of mankind. [NEWLINE] [NEWLINE] Also, I don't really think it was "ignoring" the character. It's an interesting point, but I think it was actually to take a step back from the scene. It was graphic enough as it was, but if literally the entire scene focused solely on her for the duration that it went, it would be even more traumatic.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your analogy doesn't quite work -- birth control is not a form of treatment (in the context of this cmv),<mask> a way to avoid the risks that come with sex. Here's a more apt analogy: would we force employers to provide their smoking employees with electronic cigarettes? This would allow them to get their nicotine fix without the health risks of Tabacco, just like birth control allows employees to have sex with none of the associated risk. The idea of forcing employees to provide electronic cigarettes seems kind of ridiculous, doesn't it? [NEWLINE] [NEWLINE] I do think that the economic argument has merit,<mask> that raises the question of whether the right to act in accordance with your belief (<mask><mask><mask> the rights of others aren't infringed) trumps the economic benefit.</s>
Label encoding: <s>Your analogy doesn't quite work -- birth control is not a form of treatment (in the context of this cmv), but a way to avoid the risks that come with sex. Here's a more apt analogy: would we force employers to provide their smoking employees with electronic cigarettes? This would allow them to get their nicotine fix without the health risks of Tabacco, just like birth control allows employees to have sex with none of the associated risk. The idea of forcing employees to provide electronic cigarettes seems kind of ridiculous, doesn't it? [NEWLINE] [NEWLINE] I do think that the economic argument has merit, but that raises the question of whether the right to act in accordance with your belief ( so long as the rights of others aren't infringed) trumps the economic benefit.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A positive right obliges others to do something for you - it's an entitlement. You have the right to a trial by jury in some cases, this obliges the state to conduct such a trial on your behalf in those cases. Similarly, you have the right to a K-12 education, and this obliges the state to maintain schools for you to attend. [NEWLINE] [NEWLINE] A negative right obliges others not to do something to you. You have the right to freedom of expression, this obliges others not to restrict you from expressing yourself<mask> you can (<mask> it does not oblige anyone to facilitate your expression). You<mask> have the right to life, this obliges others not to kill you (<mask> it does not oblige anyone to save your life).</s>
Label encoding: <s>A positive right obliges others to do something for you - it's an entitlement. You have the right to a trial by jury in some cases, this obliges the state to conduct such a trial on your behalf in those cases. Similarly, you have the right to a K-12 education, and this obliges the state to maintain schools for you to attend. [NEWLINE] [NEWLINE] A negative right obliges others not to do something to you. You have the right to freedom of expression, this obliges others not to restrict you from expressing yourself how you can ( though it does not oblige anyone to facilitate your expression). You also have the right to life, this obliges others not to kill you ( though it does not oblige anyone to save your life).</s>
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0747, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've got some cells in my body that have separate DNA. I've got a ton of cells that have nearly 0% genetic similarity to me. I've got a mess of cells in my abdomen that have at most a bit over 50% genetic similarity to cells next to them, and I've got some mutant cell lines hanging around in various places. I'm not a giant mass of cells with identical nuclei. With the right medical procedures, I could have cells in my body from a different *species* or that originated in a different person. [NEWLINE] [NEWLINE] <mask> unique DNA doesn't mean that a cell's not mine. Can it be physically separated from me? Is it dependent on me, and only me, for nutrients? Then it's mine. [NEWLINE] </s>
Label encoding: <s>I've got some cells in my body that have separate DNA. I've got a ton of cells that have nearly 0% genetic similarity to me. I've got a mess of cells in my abdomen that have at most a bit over 50% genetic similarity to cells next to them, and I've got some mutant cell lines hanging around in various places. I'm not a giant mass of cells with identical nuclei. With the right medical procedures, I could have cells in my body from a different *species* or that originated in a different person. [NEWLINE] [NEWLINE] So unique DNA doesn't mean that a cell's not mine. Can it be physically separated from me? Is it dependent on me, and only me, for nutrients? Then it's mine. [NEWLINE] </s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0686, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>yeah,<mask> I'm not a true progressive. I'm liberal on some issues.. conservative on some and downright libertarian on some.. I'm<mask> 45 and I've been hearing that "hold your nose and vote for<mask> and<mask>,<mask> of the Supreme Court" all my life...Screw that.. I'm not living my day to life under a terrible President all<mask> maybe she'll nominate a Supreme Court justice some time in the next four years... In the meantime she'll launch another war in the Middle East and 100,000 people have to die<mask> I was worried about the Supreme Court. BTW, I'm opposed to gay marriage.. Like [Hillary]( [URL] /) (er wait.. did she "evolve"<mask>? I still haven't)</s>
Label encoding: <s>yeah, but I'm not a true progressive. I'm liberal on some issues.. conservative on some and downright libertarian on some.. I'm also 45 and I've been hearing that "hold your nose and vote for so and so, because of the Supreme Court" all my life...Screw that.. I'm not living my day to life under a terrible President all because maybe she'll nominate a Supreme Court justice some time in the next four years... In the meantime she'll launch another war in the Middle East and 100,000 people have to die because I was worried about the Supreme Court. BTW, I'm opposed to gay marriage.. Like [Hillary]( [URL] /) (er wait.. did she "evolve" yet? I still haven't)</s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How do we judge<mask> is defensible? Logically I know you would argue it isn't<mask> morally defensible you can argue it is.<mask> you say defensible on any grounds then do you only mean logically defensible?<mask> the positives of the belief outweigh the negatives and the belief is not harming others then morally the belief is certainly morally defensible. [NEWLINE] [NEWLINE] <mask> to logically defensible is everything we value in society logical? Love<mask> an emotion is illogical and<mask> we are told it is  the highest aim in our lives to experience it. Humans<mask> a species are not robots, we are emotional creatures and<mask> many of our emotions are logical (eg pain) love, excitement, fear can all be illogical and<mask> benefit us. </s>
Label encoding: <s>How do we judge what is defensible? Logically I know you would argue it isn't but morally defensible you can argue it is. When you say defensible on any grounds then do you only mean logically defensible? If the positives of the belief outweigh the negatives and the belief is not harming others then morally the belief is certainly morally defensible. [NEWLINE] [NEWLINE] As to logically defensible is everything we value in society logical? Love as an emotion is illogical and yet we are told it is  the highest aim in our lives to experience it. Humans as a species are not robots, we are emotional creatures and while many of our emotions are logical (eg pain) love, excitement, fear can all be illogical and yet benefit us. </s>
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah I edited my original post to show I changed my view, I won't be trying to convince anyone of GMO harm to small farmers per se,<mask> related issues of big ag are still very troubling to me. I'm surprised you don't seen to find anything wrong with anti-competitive behavior,<mask>. I can't imagine that a similar situation would be allowed to happen in the US.<mask> should it not be a concern in developing countries? [NEWLINE] [NEWLINE] <mask><mask> only a handful of corporations control livestock processing in the US, at least there is still competition.<mask>,<mask> WH Group in China bought one of the largest pork processors, Smithfield, there were talks about national food security and intervention<mask> such international acquisitions of our food economy were to increase.</s>
Label encoding: <s>Yeah I edited my original post to show I changed my view, I won't be trying to convince anyone of GMO harm to small farmers per se, but related issues of big ag are still very troubling to me. I'm surprised you don't seen to find anything wrong with anti-competitive behavior, though. I can't imagine that a similar situation would be allowed to happen in the US. Why should it not be a concern in developing countries? [NEWLINE] [NEWLINE] Even though only a handful of corporations control livestock processing in the US, at least there is still competition. However, when WH Group in China bought one of the largest pork processors, Smithfield, there were talks about national food security and intervention if such international acquisitions of our food economy were to increase.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks, these are great points I'm taking on board.<mask><mask> particularly the impracticalities of punishing people for driving without a licence,<mask> is the punishment for that by the way? Can you go to jail<mask> you are *repeatedly* caught driving without a licence? ∆ [NEWLINE] [NEWLINE] Out of curiousity,<mask> happened to "reform" you from a street racer to a cop?<mask>,<mask> kind of thing might have stopped you street racing? There's a law in my state<mask> the police can impound the cars of people street/drag racing etc. I imagine something like that would only make you angry and cause you to hate the police,<mask> I kinda like the idea just to stick it to those darned kids.</s>
Label encoding: <s>Thanks, these are great points I'm taking on board. I think particularly the impracticalities of punishing people for driving without a licence, what is the punishment for that by the way? Can you go to jail if you are *repeatedly* caught driving without a licence? ∆ [NEWLINE] [NEWLINE] Out of curiousity, what happened to "reform" you from a street racer to a cop? Also, what kind of thing might have stopped you street racing? There's a law in my state where the police can impound the cars of people street/drag racing etc. I imagine something like that would only make you angry and cause you to hate the police, but I kinda like the idea just to stick it to those darned kids.</s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That's the thing<mask>,<mask> I'm already going 10+ over, I'm not really going slow [ENDQ] [NEWLINE] The thing is that the left lane (in the US at least) is the *passing lane*. You seem to be under the impression that the left lane is the go-over-the-speed-limit lane,<mask> that is not always true.<mask> you are not passing, you need to be in the right (<mask> in direction) lane. [NEWLINE] [NEWLINE] The reason for this is that forcing people to pass you in the lane designated for slower traffic increases the speed disparity (and<mask> danger) between cars. In other words, you are forcing someone going faster into the same lane<mask> the slowest cars on the road.</s>
Label encoding: <s> [STARTQ] That's the thing though, if I'm already going 10+ over, I'm not really going slow [ENDQ] [NEWLINE] The thing is that the left lane (in the US at least) is the *passing lane*. You seem to be under the impression that the left lane is the go-over-the-speed-limit lane, but that is not always true. If you are not passing, you need to be in the right ( as in direction) lane. [NEWLINE] [NEWLINE] The reason for this is that forcing people to pass you in the lane designated for slower traffic increases the speed disparity (and therefore danger) between cars. In other words, you are forcing someone going faster into the same lane as the slowest cars on the road.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, XXY is male, XO is female. In some cases, the SRY gene can be moved to an X chromosome. Such a case results in an XX male[2]. [ENDQ] [NEWLINE] Not quite<mask> simple.<mask><mask> you have an XXY with CAIS? They have the SRY their body just doesn't respond to androgens and is entirely phenotypically female. They have testes not ovaries,<mask> they have a vagina, breasts, etc. Would you say this person is male<mask> of that SRY gene? Or would you say they are female due to the presence of all of the female secondary sex characteristics? [NEWLINE] [NEWLINE] Or would you say that they should decide<mask> they identify, present, etc?</s><pad>
Label encoding: <s> [STARTQ] Thus, XXY is male, XO is female. In some cases, the SRY gene can be moved to an X chromosome. Such a case results in an XX male[2]. [ENDQ] [NEWLINE] Not quite so simple. What if you have an XXY with CAIS? They have the SRY their body just doesn't respond to androgens and is entirely phenotypically female. They have testes not ovaries, but they have a vagina, breasts, etc. Would you say this person is male because of that SRY gene? Or would you say they are female due to the presence of all of the female secondary sex characteristics? [NEWLINE] [NEWLINE] Or would you say that they should decide how they identify, present, etc?</s><pad>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OK, here's a report of a study which concluded that [women are better able to remember things than men in some ways]( [URL].cfm) [NEWLINE] [NEWLINE] <mask> you need more examples, I could find more,<mask> it could descend into a'shifting the goalposts' situation,<mask> could you at least concede that men are not better than women at ''everything'' and just amend that view to ''better than women at most things''? [NEWLINE] [NEWLINE] And<mask><mask> that<mask> a woman is given a job or a scholarship just<mask> she is female and they have a quota to fill, then it implies that she could not have earned that place on her own merit, which is harmful to the view that women are capable of earning their own place</s>
Label encoding: <s>OK, here's a report of a study which concluded that [women are better able to remember things than men in some ways]( [URL].cfm) [NEWLINE] [NEWLINE] If you need more examples, I could find more, but it could descend into a'shifting the goalposts' situation, so could you at least concede that men are not better than women at ''everything'' and just amend that view to ''better than women at most things''? [NEWLINE] [NEWLINE] And I agree that if a woman is given a job or a scholarship just because she is female and they have a quota to fill, then it implies that she could not have earned that place on her own merit, which is harmful to the view that women are capable of earning their own place</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They actually are composed of real people who make mistakes. [ENDQ] [NEWLINE] And those mistakes are far better than the alternative.  Nobody's saying that it's a perfect system.  No such thing exists. [NEWLINE] [NEWLINE] [STARTQ] These people have the right to kill me.  And I don't have the right to kill them - even<mask> I am right and they are wrong and I was just defending myself. [ENDQ] [NEWLINE] Barring the obvious extenuating circumstances, no, they don't. <mask> they kill you and it is unjust, then punishment is totally in order.  One could<mask><mask> that punishment is rarely carried out in our current society,<mask> that's a problem of enforcement, not the giving away of our rights in the first place.</s>
Label encoding: <s> [STARTQ] They actually are composed of real people who make mistakes. [ENDQ] [NEWLINE] And those mistakes are far better than the alternative.  Nobody's saying that it's a perfect system.  No such thing exists. [NEWLINE] [NEWLINE] [STARTQ] These people have the right to kill me.  And I don't have the right to kill them - even if I am right and they are wrong and I was just defending myself. [ENDQ] [NEWLINE] Barring the obvious extenuating circumstances, no, they don't.  If they kill you and it is unjust, then punishment is totally in order.  One could argue that that punishment is rarely carried out in our current society, but that's a problem of enforcement, not the giving away of our rights in the first place.</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do you now see<mask> your objections to moral realism are absurd [ENDQ] [NEWLINE] Not really.  An Example: [NEWLINE] [NEWLINE] [STARTQ] &gt; Human perception of morality varies wildly based on time, place, situation, and who you ask. [ENDQ] [NEWLINE] [STARTQ] Human perception of the size, location and shape of the Earth [ENDQ] [NEWLINE] The size or shape of the earth is a fact we have access to by several methods of measuring it.  There are no wars fought<mask> people feel passionately that those that think it's flat deserve to die. [NEWLINE] There is almost global consensus worldwide of the size, shape and location of the earth, and<mask> time goes by the accuracy and consensus regarding this only increases. [NEWLINE] [NEWLINE] This does not apply to morality very well. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Do you now see how your objections to moral realism are absurd [ENDQ] [NEWLINE] Not really.  An Example: [NEWLINE] [NEWLINE] [STARTQ] &gt; Human perception of morality varies wildly based on time, place, situation, and who you ask. [ENDQ] [NEWLINE] [STARTQ] Human perception of the size, location and shape of the Earth [ENDQ] [NEWLINE] The size or shape of the earth is a fact we have access to by several methods of measuring it.  There are no wars fought because people feel passionately that those that think it's flat deserve to die. [NEWLINE] There is almost global consensus worldwide of the size, shape and location of the earth, and as time goes by the accuracy and consensus regarding this only increases. [NEWLINE] [NEWLINE] This does not apply to morality very well. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But it does add to the utility for users,<mask> it's another option. People who use the "just friends" feature appreciate it. They'd say that it adds end user utility to the site. [NEWLINE] [NEWLINE] <mask> you're basically saying that it's a chore to have to go through all these "just friends" folks<mask> all you're looking for is actual dating, and<mask>, it's less user-friendly for you, I'd say 1, there's probably an option to filter those out,<mask>, problem solved, and 2,<mask> about the ugly people? You have to filter those out<mask> well,<mask> I don't think you're gonna suggest that dating sites shouldn't allow ugly people just<mask> it'll make your search easier.</s>
Label encoding: <s>But it does add to the utility for users, because it's another option. People who use the "just friends" feature appreciate it. They'd say that it adds end user utility to the site. [NEWLINE] [NEWLINE] If you're basically saying that it's a chore to have to go through all these "just friends" folks when all you're looking for is actual dating, and thus, it's less user-friendly for you, I'd say 1, there's probably an option to filter those out, so, problem solved, and 2, what about the ugly people? You have to filter those out as well, but I don't think you're gonna suggest that dating sites shouldn't allow ugly people just so it'll make your search easier.</s>
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"You haven't even mentioned woman on male sexual assault, which a lot of people even claim is impossible" [NEWLINE] [STARTQ] I didn't really mention it<mask> I don't really understand<mask> it could work that way [ENDQ] [NEWLINE] aaaaaaand that's another big problem MRA's are working to overcome. Rape by definition is unlawful intercourse (not: not JUST being penetrated), which means a man can be raped<mask> he has sex against his will. You just said you don't understand<mask> it could work that way. A man is not always willing to have sex.<mask> the general way people see it at the moment,<mask> a young man has sex he's always 'done well' and is congratulated. Even<mask> he didn't want it? Hmm.</s>
Label encoding: <s>"You haven't even mentioned woman on male sexual assault, which a lot of people even claim is impossible" [NEWLINE] [STARTQ] I didn't really mention it because I don't really understand how it could work that way [ENDQ] [NEWLINE] aaaaaaand that's another big problem MRA's are working to overcome. Rape by definition is unlawful intercourse (not: not JUST being penetrated), which means a man can be raped if he has sex against his will. You just said you don't understand how it could work that way. A man is not always willing to have sex. Although the general way people see it at the moment, if a young man has sex he's always 'done well' and is congratulated. Even if he didn't want it? Hmm.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How can you say that society is a fictional entity? That is similar to the cells of your body saying that the body is a fictional entity. All the cells experience are their close neighbors without awareness of the entire body. [NEWLINE] [NEWLINE] Most people would die<mask> they were to leave society<mask> very few people have the skills needed to survive on ther own. [NEWLINE] [NEWLINE] Bodies have evolved macrophages to kill sick and old cells.<mask> organisms have evolved these mechanisms to enhance their survival we should set up similar mechanisms in society to enhance our survival. [NEWLINE] [NEWLINE] From [URL] [NEWLINE] [NEWLINE] [STARTQ] Macrophages are highly specialized in removal of dying or dead cells and cellular debris. [ENDQ] [NEWLINE] Natural mechanisms created by evolution have inspired a variety of useful technologies, [URL] </s>
Label encoding: <s>How can you say that society is a fictional entity? That is similar to the cells of your body saying that the body is a fictional entity. All the cells experience are their close neighbors without awareness of the entire body. [NEWLINE] [NEWLINE] Most people would die if they were to leave society as very few people have the skills needed to survive on ther own. [NEWLINE] [NEWLINE] Bodies have evolved macrophages to kill sick and old cells. Since organisms have evolved these mechanisms to enhance their survival we should set up similar mechanisms in society to enhance our survival. [NEWLINE] [NEWLINE] From [URL] [NEWLINE] [NEWLINE] [STARTQ] Macrophages are highly specialized in removal of dying or dead cells and cellular debris. [ENDQ] [NEWLINE] Natural mechanisms created by evolution have inspired a variety of useful technologies, [URL] </s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This means a lot to me! I'm actually going to school to become a public school teacher. I'm only going into my second year and haven't really hit any of the formal University classes,<mask> I've read a lot, been fairly active in peer to peer  teaching (in high school), and like to thing that I'm fairly passionate about facilitating education. [NEWLINE] [NEWLINE] Would you suggest any authors or books? To be honest, most of my reading has all dealt with Critical Literature/Pedagogy such<mask> multiple writings from Paulo Friere (<mask> is obvious throughout this entire thread) and few of the authors the he influenced, and not a lot on actual accounts that take place directly in our school system or the classroom. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>This means a lot to me! I'm actually going to school to become a public school teacher. I'm only going into my second year and haven't really hit any of the formal University classes, but I've read a lot, been fairly active in peer to peer  teaching (in high school), and like to thing that I'm fairly passionate about facilitating education. [NEWLINE] [NEWLINE] Would you suggest any authors or books? To be honest, most of my reading has all dealt with Critical Literature/Pedagogy such as multiple writings from Paulo Friere ( as is obvious throughout this entire thread) and few of the authors the he influenced, and not a lot on actual accounts that take place directly in our school system or the classroom. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, I guess I usually think of it in terms of small scale conflicts. Waco is exactly the example I and in mind<mask> I started thinking about this. [NEWLINE] [NEWLINE] <mask>,<mask> folks are arguing in favor of the need for personal firearms<mask> of the threat of tyranny are they really imagining a large scale revolution? I guess my problem with that is that a large scale revolutionary movement would first require slow base-building, and the U.S. government has historically been very good at infiltrating and disrupting movements that it viewed<mask> potentially threatening its authority. (For example Cointelpro during the 60's and early 70s.) I wonder<mask> dissent would ever realistically be able to reach the point of revolution without being violently quashed.</s>
Label encoding: <s>Yeah, I guess I usually think of it in terms of small scale conflicts. Waco is exactly the example I and in mind when I started thinking about this. [NEWLINE] [NEWLINE] So, when folks are arguing in favor of the need for personal firearms because of the threat of tyranny are they really imagining a large scale revolution? I guess my problem with that is that a large scale revolutionary movement would first require slow base-building, and the U.S. government has historically been very good at infiltrating and disrupting movements that it viewed as potentially threatening its authority. (For example Cointelpro during the 60's and early 70s.) I wonder if dissent would ever realistically be able to reach the point of revolution without being violently quashed.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If the dictionary has 20,000 words and each string was between 1-8 words in combination, with there being three forms of each word (all lowercase, all uppercase, first letter capitalized-- making the dictionary 60,000 words long)...With repetition allowed, that's easily the 27th order. (Ex. NoYesPonyAtFunTheDogCat) [NEWLINE] [NEWLINE] <mask><mask> (this website)[ [URL] ], that's on the 33rd order.  Now imagine<mask> you had all lowercase, one first-letter capitalized, one all upper-case, etc... (Ex. WhiteTHUSradioRadiopony)... I don't bother to calculate the order,<mask> it's just hard. </s>
Label encoding: <s>If the dictionary has 20,000 words and each string was between 1-8 words in combination, with there being three forms of each word (all lowercase, all uppercase, first letter capitalized-- making the dictionary 60,000 words long)...With repetition allowed, that's easily the 27th order. (Ex. NoYesPonyAtFunTheDogCat) [NEWLINE] [NEWLINE] According to (this website)[ [URL] ], that's on the 33rd order.  Now imagine if you had all lowercase, one first-letter capitalized, one all upper-case, etc... (Ex. WhiteTHUSradioRadiopony)... I don't bother to calculate the order, but it's just hard. </s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Welp,<mask> we're at it, every time you get caught speeding, everytime you do a rolling stop, everytime you don't put your blinker on before you turn, everytime you run a red light, you should not be allowed to drive again. Don't say it's not the same, you are still putting people at risk including yourself. I'm sorry for your loss, don't get me wrong,<mask><mask> you wanna start revoking privileges for one offense (<mask><mask> unsafe it is) you gotta do it for all of them. And I'm sure you wouldn't be driving for very long<mask> law enforcement started cracking down on all the laws. [NEWLINE] [NEWLINE] Sorry<mask> my grammar/spelling sucks</s>
Label encoding: <s>Welp, while we're at it, every time you get caught speeding, everytime you do a rolling stop, everytime you don't put your blinker on before you turn, everytime you run a red light, you should not be allowed to drive again. Don't say it's not the same, you are still putting people at risk including yourself. I'm sorry for your loss, don't get me wrong, but if you wanna start revoking privileges for one offense ( despite how unsafe it is) you gotta do it for all of them. And I'm sure you wouldn't be driving for very long if law enforcement started cracking down on all the laws. [NEWLINE] [NEWLINE] Sorry if my grammar/spelling sucks</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree, the comparison you made with dolphins hits the nail right on the head. The problem is, is the comment I'm replying to says "they don't, they simply respond to your anger." The dog understands that<mask> its doing is wrong. It might not have a firm grasp on the pragmatics of right and wrong that humans have,<mask> it has a functional ability to understand that there are things it should do and things it shouldn't do. It's not a mindless animal. [NEWLINE] [NEWLINE] To compare to your dolphin analogy, I would say that it would be like arguing humans can't swim. We can swim, dolphins swim better than us,<mask> we can get from point A to point B in the water.</s>
Label encoding: <s>I agree, the comparison you made with dolphins hits the nail right on the head. The problem is, is the comment I'm replying to says "they don't, they simply respond to your anger." The dog understands that what its doing is wrong. It might not have a firm grasp on the pragmatics of right and wrong that humans have, but it has a functional ability to understand that there are things it should do and things it shouldn't do. It's not a mindless animal. [NEWLINE] [NEWLINE] To compare to your dolphin analogy, I would say that it would be like arguing humans can't swim. We can swim, dolphins swim better than us, but we can get from point A to point B in the water.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you have a very poor understanding of alcoholism. You can drink responsibly, have fun, and not feel a need to consume alcohol. [NEWLINE] [NEWLINE] You<mask> seem to be basing most of this on parties<mask> people get way too drunk and go crazy. Getting a few friends together, drinking some and playing games is a great experience. It makes simple, otherwise boring things seem fun. It's<mask> a great bonding experience. I really think you should just try it<mask> you have some experience to have a basis for your views.<mask>'s the worst that could happen? You get drunk, don't like it, fall asleep, then go about your life. [NEWLINE] [NEWLINE] The problem isn't the alcohol, it's the people. </s>
Label encoding: <s>I think you have a very poor understanding of alcoholism. You can drink responsibly, have fun, and not feel a need to consume alcohol. [NEWLINE] [NEWLINE] You also seem to be basing most of this on parties where people get way too drunk and go crazy. Getting a few friends together, drinking some and playing games is a great experience. It makes simple, otherwise boring things seem fun. It's also a great bonding experience. I really think you should just try it so you have some experience to have a basis for your views. What's the worst that could happen? You get drunk, don't like it, fall asleep, then go about your life. [NEWLINE] [NEWLINE] The problem isn't the alcohol, it's the people. </s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> we don't (and can't) know anything about the nature of God. [ENDQ] [NEWLINE] <mask> that is true, then<mask> can you say: [NEWLINE] [NEWLINE] [STARTQ] God is all powerful and infinite [ENDQ] [NEWLINE] That's a claim about God. Which you have no proof for. [NEWLINE] [NEWLINE] Saying God brought the universe in to existence is more along the lines of saying a flying pink squirrel created the universe. There's no reason to believe that it is the case. [NEWLINE] [NEWLINE] Saying "I don't know" is the more honest answer<mask> it is not making a claim which needs to be proven or disproven.<mask><mask> saying that god created the universe is a claim that  **could** be proven or disproven. </s>
Label encoding: <s> [STARTQ] because we don't (and can't) know anything about the nature of God. [ENDQ] [NEWLINE] If that is true, then how can you say: [NEWLINE] [NEWLINE] [STARTQ] God is all powerful and infinite [ENDQ] [NEWLINE] That's a claim about God. Which you have no proof for. [NEWLINE] [NEWLINE] Saying God brought the universe in to existence is more along the lines of saying a flying pink squirrel created the universe. There's no reason to believe that it is the case. [NEWLINE] [NEWLINE] Saying "I don't know" is the more honest answer because it is not making a claim which needs to be proven or disproven. Where as saying that god created the universe is a claim that  **could** be proven or disproven. </s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Average pie [STARTQ] Average cake. [ENDQ] Great pie &lt; Great cake. [NEWLINE] [NEWLINE] The simple fact is that pie is incredibly easy to make compared to real life cake. <mask> much<mask> that there are only a couple of bakeries that still make actual cake.  Most of<mask> people call cake today is an insult to the food.  It is like classifying McDonalds<mask> French cuisine<mask> of french fries. [NEWLINE] [NEWLINE] I won't go into details for cake<mask> it's<mask> complex,<mask> do yourself a favor and make this recipe for Italian Meringue Buttercream.  It will show you a hint of<mask> you're missing. <mask> you taste that and still think pie is better... you won't. [URL] </s>
Label encoding: <s>Average pie [STARTQ] Average cake. [ENDQ] Great pie &lt; Great cake. [NEWLINE] [NEWLINE] The simple fact is that pie is incredibly easy to make compared to real life cake.  So much so that there are only a couple of bakeries that still make actual cake.  Most of what people call cake today is an insult to the food.  It is like classifying McDonalds as French cuisine because of french fries. [NEWLINE] [NEWLINE] I won't go into details for cake because it's so complex, but do yourself a favor and make this recipe for Italian Meringue Buttercream.  It will show you a hint of what you're missing.  If you taste that and still think pie is better... you won't. [URL] </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I started thinking about this after seeing my brother's reaction to playing things like call of duty. Utter, maddening frustration at being interrupted or killed. Really horrible behaviour and we have a very nice loving family, he doesn't have any issues,<mask> anything he's a very sweet, caring boy.<mask> then there are outside factors too, which<mask><mask> combined with obesessing over games just brought the worse in him. [NEWLINE] [NEWLINE] My conclusion, based on a kinda psychology 101 and observation (<mask> not very scientific) is that is a bit like weed being a risk factor. Neither is going to make you crack by itself,<mask> they can both be the straw that breaks the camel's back<mask> paired with other things.</s>
Label encoding: <s>I started thinking about this after seeing my brother's reaction to playing things like call of duty. Utter, maddening frustration at being interrupted or killed. Really horrible behaviour and we have a very nice loving family, he doesn't have any issues, if anything he's a very sweet, caring boy. But then there are outside factors too, which I think combined with obesessing over games just brought the worse in him. [NEWLINE] [NEWLINE] My conclusion, based on a kinda psychology 101 and observation ( so not very scientific) is that is a bit like weed being a risk factor. Neither is going to make you crack by itself, but they can both be the straw that breaks the camel's back when paired with other things.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Bastiat's [That Which is Seen, and That Which is Not Seen]( [URL] ) is relevant to your comment. Sure, the space race was a big leap ahead in technology.<mask> it was ultra-super expensive (relatively speaking for the time in question).<mask> the cost-benefit analysis might not pan out<mask> well<mask> you think. It is impossible to know<mask> would have happened instead,<mask> we can look at the efficiency of<mask> was done. Certainly there were a great many things that came out of it,<mask> at its core, a crap ton of money was spent sending a few humans to an inhabitable environment for the sake of doing it, all<mask> a bunch more humans at home suffered in poverty.</s>
Label encoding: <s>Bastiat's [That Which is Seen, and That Which is Not Seen]( [URL] ) is relevant to your comment. Sure, the space race was a big leap ahead in technology. But it was ultra-super expensive (relatively speaking for the time in question). So the cost-benefit analysis might not pan out as well as you think. It is impossible to know what would have happened instead, but we can look at the efficiency of what was done. Certainly there were a great many things that came out of it, but at its core, a crap ton of money was spent sending a few humans to an inhabitable environment for the sake of doing it, all while a bunch more humans at home suffered in poverty.</s>
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ Possibly<mask> for me it does not have the same effect.  You have changed<mask><mask><mask><mask><mask> I am suggesting will not apply to everyone<mask> footage of refugees who are hungry does not have the same effect<mask> it does not show<mask> they are running from, only the environment they are going to.  For me it is the horrors that they are trying to escape from that really disturb me. [NEWLINE] [NEWLINE] <mask> a side not I never used to donate <mask> now I feel guilty for not<mask><mask> my inactions are actively endangering people may otherwise be in safe hands. <mask><mask> about viewing the videos is that it may help others realise the severity of the situation and prevent some people forgetting about it without acting.</s>
Label encoding: <s>∆ Possibly but for me it does not have the same effect.  You have changed my opinion insofar as what I am suggesting will not apply to everyone however footage of refugees who are hungry does not have the same effect because it does not show what they are running from, only the environment they are going to.  For me it is the horrors that they are trying to escape from that really disturb me. [NEWLINE] [NEWLINE] As a side not I never used to donate  but now I feel guilty for not as if my inactions are actively endangering people may otherwise be in safe hands.  My opinion about viewing the videos is that it may help others realise the severity of the situation and prevent some people forgetting about it without acting.</s>
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If both parties are intoxicated, it is not rape.<mask> one party is intoxicated<mask> willing and the other is<mask> willing<mask> not intoxicated, it is not rape. [NEWLINE] [NEWLINE] It becomes rape<mask> the intoxication of one individual is done with the express intent to create a situation in which the other person might be able to have sex with them<mask><mask> they would not normally consent. [NEWLINE] [NEWLINE] <mask>,<mask> both people are intoxicated, then it is not rape.<mask> for this situation to be rape, it requires a victim to be selected, intoxicated for the express purpose of gaining sexual intercourse, the second party in question must not be willing<mask> intoxicated to a degree that they cannot legally make their own decisions. THEN it is rape. </s>
Label encoding: <s>If both parties are intoxicated, it is not rape. If one party is intoxicated but willing and the other is also willing but not intoxicated, it is not rape. [NEWLINE] [NEWLINE] It becomes rape when the intoxication of one individual is done with the express intent to create a situation in which the other person might be able to have sex with them where as they would not normally consent. [NEWLINE] [NEWLINE] However, if both people are intoxicated, then it is not rape. So for this situation to be rape, it requires a victim to be selected, intoxicated for the express purpose of gaining sexual intercourse, the second party in question must not be willing but intoxicated to a degree that they cannot legally make their own decisions. THEN it is rape. </s>
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0591, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it only looks that way in retrospect and rock only looks culturally irrelevant compared to an embellished past. Look back at the heyday of punk or prog rock or thrash metal and you'll see it was largely some version of pop that dominated the charts. Pop-punk, grunge and nu metal were largely short-lived trends (from a mainstream exposure standpoint.) Rock has always had one foot firmly in the underground, which is<mask> its descent into irrelevance has been proclaimed in every decade. Most rock bands reach their commercial peak late in their career. Bands like Led Zeppelin, Deep Purple, Black Sabbath, Rush and Iron Maiden weren't critical darlings at the time of their most classic albums.</s>
Label encoding: <s>I think it only looks that way in retrospect and rock only looks culturally irrelevant compared to an embellished past. Look back at the heyday of punk or prog rock or thrash metal and you'll see it was largely some version of pop that dominated the charts. Pop-punk, grunge and nu metal were largely short-lived trends (from a mainstream exposure standpoint.) Rock has always had one foot firmly in the underground, which is why its descent into irrelevance has been proclaimed in every decade. Most rock bands reach their commercial peak late in their career. Bands like Led Zeppelin, Deep Purple, Black Sabbath, Rush and Iron Maiden weren't critical darlings at the time of their most classic albums.</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But surely this belief extends to other children<mask> that OP's child? More to the point, my solution to that problem is that we should provide better supports for families of children with disabilities, included fully-funded respite care and in-home nursing,<mask> needed. Most parents don't want their children dead even<mask> they are causing horrific strain on the family; even<mask> they do, most would prefer to have a version of life in which their child was alive,<mask> no longer impacting them<mask> negatively. Right? I mean, this seems pretty basic.<mask> do you feel that euthanasia, which would undoubtedly be very traumatic for most parents, is a better answer than providing support?<mask> reason is there,<mask> money?</s>
Label encoding: <s>But surely this belief extends to other children besides that OP's child? More to the point, my solution to that problem is that we should provide better supports for families of children with disabilities, included fully-funded respite care and in-home nursing, if needed. Most parents don't want their children dead even if they are causing horrific strain on the family; even if they do, most would prefer to have a version of life in which their child was alive, but no longer impacting them so negatively. Right? I mean, this seems pretty basic. Why do you feel that euthanasia, which would undoubtedly be very traumatic for most parents, is a better answer than providing support? What reason is there, besides money?</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the main thing should be distinguishing between responsible smokers and stoner. Such<mask> the difference between responsible drinkers and drunks. I go out and drink quite a bit<mask> still slightly look down upon the people that get rowdy and far out of control<mask> under the influence. Such<mask> I smoke quite regularly<mask> get annoyed by the stoners who make that their entire life, get a oil rig setup, and are never sober. Before I started doing volunteer work, I would smoke every day<mask> I would know the time and place for it and wasn't about to go to work completely blazed and get myself fired. In summary, I'd say the best way to put it is moderation and balance is key.</s>
Label encoding: <s>I think the main thing should be distinguishing between responsible smokers and stoner. Such as the difference between responsible drinkers and drunks. I go out and drink quite a bit but still slightly look down upon the people that get rowdy and far out of control when under the influence. Such as I smoke quite regularly but get annoyed by the stoners who make that their entire life, get a oil rig setup, and are never sober. Before I started doing volunteer work, I would smoke every day but I would know the time and place for it and wasn't about to go to work completely blazed and get myself fired. In summary, I'd say the best way to put it is moderation and balance is key.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>He's speaking about college majors, and logic is usually (or always?) part of a philosophy department. [NEWLINE] [NEWLINE] <mask><mask><mask> it's a different question whether or not it *should* be part of the philosophy department. I would counter by saying that metaphysics can be entirely independent of humans, and there is no other area of learning that would be able to deal with metaphysics. Metaphysics can't be entirely dealt with through scientific means, nor mathematical means. [NEWLINE] [NEWLINE] <mask><mask><mask> you take logic and metaphysics out of the purview of philosophy, you'll just end up with a new scholarly area that is simply "Logic and Metaphysics,"<mask> there is nowhere else to put them. </s><pad>
Label encoding: <s>He's speaking about college majors, and logic is usually (or always?) part of a philosophy department. [NEWLINE] [NEWLINE] So I think it's a different question whether or not it *should* be part of the philosophy department. I would counter by saying that metaphysics can be entirely independent of humans, and there is no other area of learning that would be able to deal with metaphysics. Metaphysics can't be entirely dealt with through scientific means, nor mathematical means. [NEWLINE] [NEWLINE] I think if you take logic and metaphysics out of the purview of philosophy, you'll just end up with a new scholarly area that is simply "Logic and Metaphysics," because there is nowhere else to put them. </s><pad>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] OccamsChainsaw [ENDQ] [NEWLINE] Interesting username. I'm guessing you're big on atheism and a nihilistic worldview you associate with science. [NEWLINE] [NEWLINE] [STARTQ] I don't see<mask><mask> the want comes from would matter, or<mask> chasing that outcome necessarily means I have a moral obligation to<mask> do. [ENDQ] [NEWLINE] My point is that obedience to an objective standard is implicit in many of a nihilist's actions.<mask> someone claims to have abandoned Catholicism<mask> continues going to mass, praying, and doing everything else a Catholic does, then they haven't really abandoned Catholicism except perhaps in an uninteresting intellectual sense. I am saying that, similarly, the nihilist has not really abandoned the belief that morality is objective.</s>
Label encoding: <s> [STARTQ] OccamsChainsaw [ENDQ] [NEWLINE] Interesting username. I'm guessing you're big on atheism and a nihilistic worldview you associate with science. [NEWLINE] [NEWLINE] [STARTQ] I don't see how where the want comes from would matter, or why chasing that outcome necessarily means I have a moral obligation to so do. [ENDQ] [NEWLINE] My point is that obedience to an objective standard is implicit in many of a nihilist's actions. If someone claims to have abandoned Catholicism but continues going to mass, praying, and doing everything else a Catholic does, then they haven't really abandoned Catholicism except perhaps in an uninteresting intellectual sense. I am saying that, similarly, the nihilist has not really abandoned the belief that morality is objective.</s>
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What you may want to focus on, is Yield. [NEWLINE] [NEWLINE] <mask> a company is seeking a 10% ROI.  "Investing" 60,000 in an employee that results in 5,000 profit, put's yield at 8.3%,<mask> they maybe prefer to not hire. [NEWLINE] [NEWLINE] Really,<mask>, total cost of an employee, is on average about 30% higher (more and less depending on the jobs),<mask> that's a rough national average. <mask> the 60,000 should not just be the salary,<mask> the salary, benefits, and additional insurances, regulatory requirements, and taxes.  From the payroll perspective, the employer pays 100% of FICA contributions for example.</s>
Label encoding: <s>What you may want to focus on, is Yield. [NEWLINE] [NEWLINE] If a company is seeking a 10% ROI.  "Investing" 60,000 in an employee that results in 5,000 profit, put's yield at 8.3%, so they maybe prefer to not hire. [NEWLINE] [NEWLINE] Really, however, total cost of an employee, is on average about 30% higher (more and less depending on the jobs), but that's a rough national average.  So the 60,000 should not just be the salary, but the salary, benefits, and additional insurances, regulatory requirements, and taxes.  From the payroll perspective, the employer pays 100% of FICA contributions for example.</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> a lot of scientific breakthroughs come through the military. You do like the internet right? [ENDQ] [NEWLINE] Yeah,<mask> scientists go 'Hell, a central research system is paying for my work... better not work<mask> fast. Only<mask> the military fund it will I be truly effective!' [NEWLINE] [NEWLINE] The military fund shitloads of research that spins off into effective things.<mask> any other place... hell<mask> you gave the military's research budget to a *kindergarten* to set up a few hundred research labs then that research program would spin off hundreds of successful, useful and societal-positive programs. [NEWLINE] [NEWLINE] There is nothing special about the military that allows their success except having the money in the first place.</s>
Label encoding: <s> [STARTQ] Also a lot of scientific breakthroughs come through the military. You do like the internet right? [ENDQ] [NEWLINE] Yeah, because scientists go 'Hell, a central research system is paying for my work... better not work as fast. Only if the military fund it will I be truly effective!' [NEWLINE] [NEWLINE] The military fund shitloads of research that spins off into effective things. If any other place... hell if you gave the military's research budget to a *kindergarten* to set up a few hundred research labs then that research program would spin off hundreds of successful, useful and societal-positive programs. [NEWLINE] [NEWLINE] There is nothing special about the military that allows their success except having the money in the first place.</s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0798, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do rules of evidence have to do with due process? Due process means that he doesn't go to jail without a trial [ENDQ] [NEWLINE] <mask><mask><mask>, you never have a trial before going to jail. Prison follows a trial. Second of all, it's not just about having a trial,<mask> having a fair trial. Would you try to<mask><mask> due process exists in China? [NEWLINE] [NEWLINE] [STARTQ] The exception is<mask> the defendant has a history of sexual assault. [ENDQ] [NEWLINE] With a complementary exception to the accuser's history, which is normally admissable. [URL] [NEWLINE] [NEWLINE] [STARTQ] Oversimplification of complex matters is bad, m'kay? [ENDQ] [NEWLINE] Not researching<mask> you're talking about is worse.</s>
Label encoding: <s> [STARTQ] What do rules of evidence have to do with due process? Due process means that he doesn't go to jail without a trial [ENDQ] [NEWLINE] First of all, you never have a trial before going to jail. Prison follows a trial. Second of all, it's not just about having a trial, but having a fair trial. Would you try to argue that due process exists in China? [NEWLINE] [NEWLINE] [STARTQ] The exception is if the defendant has a history of sexual assault. [ENDQ] [NEWLINE] With a complementary exception to the accuser's history, which is normally admissable. [URL] [NEWLINE] [NEWLINE] [STARTQ] Oversimplification of complex matters is bad, m'kay? [ENDQ] [NEWLINE] Not researching what you're talking about is worse.</s>
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's often pointed out that a vegetarian diet has its own death toll of the creatures who lost their lives through the process of farming plants. [NEWLINE] [NEWLINE] The rebuttal, which I wholly agree with is that the point is to lessen animal suffering, and that lessening is a positive step even<mask> it isn't perfect. [NEWLINE] [NEWLINE] <mask><mask> the same applies here. Given two possible worlds, both of which contain factory farming,<mask> one of which has those animals dying more quickly and painlessly and the other including the additional suffering of throat slitting, which possible world contains less animal suffering? The answer is clear. To reject this change<mask> too little or hypocritical is letting the perfect be the enemy of the good.</s>
Label encoding: <s>It's often pointed out that a vegetarian diet has its own death toll of the creatures who lost their lives through the process of farming plants. [NEWLINE] [NEWLINE] The rebuttal, which I wholly agree with is that the point is to lessen animal suffering, and that lessening is a positive step even if it isn't perfect. [NEWLINE] [NEWLINE] I think the same applies here. Given two possible worlds, both of which contain factory farming, but one of which has those animals dying more quickly and painlessly and the other including the additional suffering of throat slitting, which possible world contains less animal suffering? The answer is clear. To reject this change as too little or hypocritical is letting the perfect be the enemy of the good.</s>
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry KingMinish, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=KingMinish+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry KingMinish, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=KingMinish+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can't simply declare concepts to be things you dislike and then cry no true scotsman<mask> it's pointed out that the thing you're criticizing doesn't actually fall under the definition you would prefer. It's a completely ignorant misappropriation of the fallacy, which follows this model: [NEWLINE] [NEWLINE] -All Scotsmen eat lamb. [NEWLINE] [NEWLINE] -My father is Scottish and he doesn't eat lamb. [NEWLINE] [NEWLINE] -He's not a true Scotsman. [NEWLINE] [NEWLINE] You are using it in this way: [NEWLINE] [NEWLINE] -All Scotsmen eat alfalfa. [NEWLINE] [NEWLINE] -No they don't,<mask> those aren't Scotsmen, they're horses. [NEWLINE] [NEWLINE] -NO TRUE SCOTSMAN [NEWLINE] </s>
Label encoding: <s>You can't simply declare concepts to be things you dislike and then cry no true scotsman when it's pointed out that the thing you're criticizing doesn't actually fall under the definition you would prefer. It's a completely ignorant misappropriation of the fallacy, which follows this model: [NEWLINE] [NEWLINE] -All Scotsmen eat lamb. [NEWLINE] [NEWLINE] -My father is Scottish and he doesn't eat lamb. [NEWLINE] [NEWLINE] -He's not a true Scotsman. [NEWLINE] [NEWLINE] You are using it in this way: [NEWLINE] [NEWLINE] -All Scotsmen eat alfalfa. [NEWLINE] [NEWLINE] -No they don't, also those aren't Scotsmen, they're horses. [NEWLINE] [NEWLINE] -NO TRUE SCOTSMAN [NEWLINE] </s>
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The purpose of the physical exams is to determine endurance. Each test is to determine the endurance of the various muscles. It isn't gender specific becaue it can't be. [NEWLINE] [NEWLINE] Let's say you, the man in charge of who get's into the army, decide that you need to be able to carry this box to this location in this time.<mask> many men can do<mask>, almost no women can.<mask>,<mask> you've determined<mask> part of your job,<mask> they can't get that box there in that time, they can't effectively perform combat duties. Do you lower the requirements for the women, or do you only pass the women who can pass the test already in place?</s>
Label encoding: <s>The purpose of the physical exams is to determine endurance. Each test is to determine the endurance of the various muscles. It isn't gender specific becaue it can't be. [NEWLINE] [NEWLINE] Let's say you, the man in charge of who get's into the army, decide that you need to be able to carry this box to this location in this time. While many men can do so, almost no women can. However, as you've determined as part of your job, if they can't get that box there in that time, they can't effectively perform combat duties. Do you lower the requirements for the women, or do you only pass the women who can pass the test already in place?</s>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry TheNewColor, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=TheNewColor+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry TheNewColor, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=TheNewColor+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My incoherently worded theme was meant to be more focused on the powerlessness humans have to chance. Not<mask> much about<mask> those chances stand, i.e. the scientific community assuring that chances are low,<mask> you hinted at. Yes my views are ridiculous and not backed by viable sources of information,<mask> the chances are still there no matter<mask> minimal or<mask> strange. Other's have already convinced that humans have or will have the capability, know<mask>, technology, etc. to repair or prevent major disaster. [NEWLINE] [NEWLINE] <mask> thank you<mask> your words have opened my eyes on a different note. I am aware of my own thoughts and do often stop and think! They are definitely ridiculous.</s>
Label encoding: <s>My incoherently worded theme was meant to be more focused on the powerlessness humans have to chance. Not so much about how those chances stand, i.e. the scientific community assuring that chances are low, as you hinted at. Yes my views are ridiculous and not backed by viable sources of information, but the chances are still there no matter how minimal or how strange. Other's have already convinced that humans have or will have the capability, know how, technology, etc. to repair or prevent major disaster. [NEWLINE] [NEWLINE] But thank you as your words have opened my eyes on a different note. I am aware of my own thoughts and do often stop and think! They are definitely ridiculous.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Does pregnancy count<mask> reasonable fear of serious death or injury in every case? I get that in some cases there's a very good chance that the mother will die.<mask> in an average pregnancy I don't think that it meets that standard. [NEWLINE] [NEWLINE] This argument does actually help explain<mask> it may not be a contradiction to some degree,<mask> is that really the reason we allow abortions?<mask> it's self defense? That seems like a pretty weak self defense case in most circumstances.<mask> a mother had the exact same odds of dying<mask> she would in her average pregnancy, due to her 1 year old being alive I don't think we'd call it self defense<mask> she killed the 1 year old. </s>
Label encoding: <s>Does pregnancy count as reasonable fear of serious death or injury in every case? I get that in some cases there's a very good chance that the mother will die. But in an average pregnancy I don't think that it meets that standard. [NEWLINE] [NEWLINE] This argument does actually help explain why it may not be a contradiction to some degree, but is that really the reason we allow abortions? Because it's self defense? That seems like a pretty weak self defense case in most circumstances. If a mother had the exact same odds of dying as she would in her average pregnancy, due to her 1 year old being alive I don't think we'd call it self defense if she killed the 1 year old. </s>
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the kid has thrown himself on the floor in the middle of a restaurant and the waiters are going to trip over him and the parents are refusing to do anything about it [ENDQ] [NEWLINE] <mask> does that justify a violent reaction? [NEWLINE] [NEWLINE] [STARTQ] Just a few instances of getting knocked on his butt by a complete stranger should be enough to get it through the kid's head that this is not something you do in public. [ENDQ] [NEWLINE] Are you going to help the kid with his homework<mask> the parents don't? [NEWLINE] [NEWLINE] Are you going to pay for the kid's college education<mask> the parents don't? [NEWLINE] [NEWLINE] <mask><mask> do you think its your responsibility to give this kid lessons<mask> the parents don't?</s>
Label encoding: <s> [STARTQ] the kid has thrown himself on the floor in the middle of a restaurant and the waiters are going to trip over him and the parents are refusing to do anything about it [ENDQ] [NEWLINE] Why does that justify a violent reaction? [NEWLINE] [NEWLINE] [STARTQ] Just a few instances of getting knocked on his butt by a complete stranger should be enough to get it through the kid's head that this is not something you do in public. [ENDQ] [NEWLINE] Are you going to help the kid with his homework when the parents don't? [NEWLINE] [NEWLINE] Are you going to pay for the kid's college education when the parents don't? [NEWLINE] [NEWLINE] So why do you think its your responsibility to give this kid lessons when the parents don't?</s>
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>'Perspective' would seem to miss out on a key part of the (sociological) concept of privilege - it's not<mask> X group of people think, or<mask> they see things, it's *<mask> has led them to think that way*. In the case of 'white privilege', the contingent fact of their being born with white skin has resulted in their experience of racism being radically different to that of racism<mask> felt by a person with black skin. [NEWLINE] [NEWLINE] The emphasis in 'perspective' is on a way of thinking. 'Privilege' is concerned with the systematic exclusion of certain types of experience from certain groups of people, and the way this affects their thinking.</s>
Label encoding: <s>'Perspective' would seem to miss out on a key part of the (sociological) concept of privilege - it's not what X group of people think, or how they see things, it's * what has led them to think that way*. In the case of 'white privilege', the contingent fact of their being born with white skin has resulted in their experience of racism being radically different to that of racism as felt by a person with black skin. [NEWLINE] [NEWLINE] The emphasis in 'perspective' is on a way of thinking. 'Privilege' is concerned with the systematic exclusion of certain types of experience from certain groups of people, and the way this affects their thinking.</s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0612, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It depends on who you are,<mask> you're middle class in a first-world country then yes, you have it pretty good compared to last few thousand years. [NEWLINE] [NEWLINE] <mask>, slavery and exploitation is workers is terrible in many parts of the world,<mask> [there are ways to make sure you don't support it, which can be difficult.]( [URL] /) [NEWLINE] [NEWLINE] Before there were factories in places like Thailand and Bangladesh the people still managed to live without working in terrible sweatshops. It's just sad and appalling<mask><mask> much of the products made in the world are through worker exploitation. [NEWLINE] [NEWLINE] I'd rather be a hunter/gatherer or farmer than work in a sweatshop. </s>
Label encoding: <s>It depends on who you are, if you're middle class in a first-world country then yes, you have it pretty good compared to last few thousand years. [NEWLINE] [NEWLINE] However, slavery and exploitation is workers is terrible in many parts of the world, but [there are ways to make sure you don't support it, which can be difficult.]( [URL] /) [NEWLINE] [NEWLINE] Before there were factories in places like Thailand and Bangladesh the people still managed to live without working in terrible sweatshops. It's just sad and appalling how so much of the products made in the world are through worker exploitation. [NEWLINE] [NEWLINE] I'd rather be a hunter/gatherer or farmer than work in a sweatshop. </s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I asked myself, 'would I want The Godfather in 3d?'. The answer, of course, is yes. <mask>?  Be cause 3d creates depth.  A scene can look spectacular in 3d<mask> it pops. [NEWLINE] [NEWLINE] Think of it like going to see a play or a musical.  Part of the experience is live acting,<mask> another is that you experience the depth of the scene.  3d brings that depth to the screen.  It's immersive. <mask> you let yourself go, you almost feel<mask><mask> you are experiencing<mask> is being shown.  Of course,<mask> you go into it<mask> a skeptic, you'll never understand.</s>
Label encoding: <s>I asked myself, 'would I want The Godfather in 3d?'. The answer, of course, is yes.  Why?  Be cause 3d creates depth.  A scene can look spectacular in 3d because it pops. [NEWLINE] [NEWLINE] Think of it like going to see a play or a musical.  Part of the experience is live acting, but another is that you experience the depth of the scene.  3d brings that depth to the screen.  It's immersive.  If you let yourself go, you almost feel as if you are experiencing what is being shown.  Of course, if you go into it as a skeptic, you'll never understand.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyone on here is articulating their answers better than I could,<mask> I will say that I<mask> used to hate Kanye and then one day on a 8 hour car ride my friend forced me to listen to a few of his albums (this one included) in their entirety and it completely changed my mind. [NEWLINE] [NEWLINE] <mask>, OP says his personality isn't factoring into their opinion of him<mask><mask><mask> it probably is on a subconscious level. It's just human nature. A lot of people hate him<mask> a person (being an obnoxious jerk, marrying Kim Kardashian, etc, etc), and it's very hard (<mask> not impossible) to separate someone<mask> a person from themselves<mask> a professional</s>
Label encoding: <s>Everyone on here is articulating their answers better than I could, but I will say that I also used to hate Kanye and then one day on a 8 hour car ride my friend forced me to listen to a few of his albums (this one included) in their entirety and it completely changed my mind. [NEWLINE] [NEWLINE] Also, OP says his personality isn't factoring into their opinion of him but I think it probably is on a subconscious level. It's just human nature. A lot of people hate him as a person (being an obnoxious jerk, marrying Kim Kardashian, etc, etc), and it's very hard ( if not impossible) to separate someone as a person from themselves as a professional</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Seriously, I've never understood the mindset of It Doesn't Matter. [ENDQ] [NEWLINE] There was actually a discussion about this on /r/AskReddit the other day.  People who use that argument (your vote doesn't matter, both parties are the same, etc.) are subtly reinforcing the right-wing narrative that "government doesn't work".  Some do that on purpose to further their own interests,<mask> others have simply bought into it, oblivious to the fact that they are doing<mask>.  A person that hears this sort of rhetoric might think it makes some sort of intuitive sense, without realizing that it directly benefits the political right. [NEWLINE] [NEWLINE] [Relevant bit here]( [URL] ).</s>
Label encoding: <s> [STARTQ] Seriously, I've never understood the mindset of It Doesn't Matter. [ENDQ] [NEWLINE] There was actually a discussion about this on /r/AskReddit the other day.  People who use that argument (your vote doesn't matter, both parties are the same, etc.) are subtly reinforcing the right-wing narrative that "government doesn't work".  Some do that on purpose to further their own interests, while others have simply bought into it, oblivious to the fact that they are doing so.  A person that hears this sort of rhetoric might think it makes some sort of intuitive sense, without realizing that it directly benefits the political right. [NEWLINE] [NEWLINE] [Relevant bit here]( [URL] ).</s>
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry notacrackheadofficer, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=notacrackheadofficer+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry notacrackheadofficer, your post has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=notacrackheadofficer+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0763, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0806, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Prism was signed during the Bush Administration and we never went against<mask> they were doing [ENDQ] [NEWLINE] Reminder: *Snowden is a whistleblower*. [NEWLINE] [NEWLINE] The American people *didn't know*<mask> the NSA were doing. Obama evidently wanted that to remain a secret. Can't have the democratic process getting in the way of some good domestic spying, after all. [NEWLINE] [NEWLINE] I'm sure you're right that the whole thing should have been shot-down years ago,<mask> still. [NEWLINE] [NEWLINE] <mask><mask>, Obama is not siding with Snowden. He is<mask> anti-whistleblower<mask> they come,<mask><mask> his pre-presidency rhetoric might have had one believe.</s>
Label encoding: <s> [STARTQ] Prism was signed during the Bush Administration and we never went against what they were doing [ENDQ] [NEWLINE] Reminder: *Snowden is a whistleblower*. [NEWLINE] [NEWLINE] The American people *didn't know* what the NSA were doing. Obama evidently wanted that to remain a secret. Can't have the democratic process getting in the way of some good domestic spying, after all. [NEWLINE] [NEWLINE] I'm sure you're right that the whole thing should have been shot-down years ago, but still. [NEWLINE] [NEWLINE] In addition, Obama is not siding with Snowden. He is as anti-whistleblower as they come, despite what his pre-presidency rhetoric might have had one believe.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd like to point out a few misconceptions you have. Most rapes aren't committed by strangers, nor are they commited in back alleys. Most rapes happen in the home, and most rapes are committed by someone the victim knows. Someone like a family member, significant other, friend etc. Furthermore, most rapes aren't committed by psychopaths and people with antisocial personality disorder (the untreatable psychological conditions that have lack of empathy<mask> a defining characteristic). This means that there are plenty of rapes being commited by average joes and janes.<mask> rape rates in North America happened at 1/10 the rate that they happen now, I would agree with you. </s>
Label encoding: <s>I'd like to point out a few misconceptions you have. Most rapes aren't committed by strangers, nor are they commited in back alleys. Most rapes happen in the home, and most rapes are committed by someone the victim knows. Someone like a family member, significant other, friend etc. Furthermore, most rapes aren't committed by psychopaths and people with antisocial personality disorder (the untreatable psychological conditions that have lack of empathy as a defining characteristic). This means that there are plenty of rapes being commited by average joes and janes. If rape rates in North America happened at 1/10 the rate that they happen now, I would agree with you. </s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0931, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, more people are identifying<mask> non-religious,<mask> that doesn't mean that religion is going to go away. <mask><mask><mask>, there have always been athiests,<mask> often times these people were just "closet athiests" out of fear of being killed. [NEWLINE] [NEWLINE] <mask> it comes to religion existing, religions don't have to coexist.  Wars have taken place over thousands of years in the name of religion. <mask> you are referring to religious people losing the battle on having proof, they don't have proof, never had proof, never will have proof,<mask> that has never stopped any given religion from existing.    </s>
Label encoding: <s>Yes, more people are identifying as non-religious, but that doesn't mean that religion is going to go away.  First of all, there have always been athiests, but often times these people were just "closet athiests" out of fear of being killed. [NEWLINE] [NEWLINE] When it comes to religion existing, religions don't have to coexist.  Wars have taken place over thousands of years in the name of religion.  If you are referring to religious people losing the battle on having proof, they don't have proof, never had proof, never will have proof, but that has never stopped any given religion from existing.    </s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0800, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But then you are going through a list of popular games that weren't intended for VR to judge its validity.<mask><mask> you are excluding the possibility that VR can be<mask> cool and wanted that it will have more games (and other things) specifically developed for it, rather than trying to make it adapt to things that weren't made for it. It would be like<mask> mobile games have become popular,<mask><mask> all of those games you mentioned would not be good on mobile devices. [NEWLINE] [NEWLINE] <mask><mask> this time the processing power and technology is really there to create something new. We<mask> a society have been dreaming about VR entertainment for a long time, and now it's here.</s><pad><pad>
Label encoding: <s>But then you are going through a list of popular games that weren't intended for VR to judge its validity. I think you are excluding the possibility that VR can be so cool and wanted that it will have more games (and other things) specifically developed for it, rather than trying to make it adapt to things that weren't made for it. It would be like how mobile games have become popular, even though all of those games you mentioned would not be good on mobile devices. [NEWLINE] [NEWLINE] I think this time the processing power and technology is really there to create something new. We as a society have been dreaming about VR entertainment for a long time, and now it's here.</s><pad><pad>
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry DeliberateConfusion, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=DeliberateConfusion+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry DeliberateConfusion, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=DeliberateConfusion+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you were the first to touch on probability which should be worth a ∆ for presenting a perspective I hadn't fully considered. [NEWLINE] [NEWLINE] The only real problem I have with this explanation is that it could be oversimplified. It would only hold<mask> we assume that the bisexual will have significantly less sex partners of the opposite sex than the heterosexual and some other variables that may or may not be relevant anyway. [NEWLINE] [NEWLINE] I would need to see some kind of analysis to be convinced to change my belief that we do not have enough information claim heterosexuality is advantageous<mask> before I wasn't even fully aware of this route<mask> I count it<mask> a "Change of Perspective". </s>
Label encoding: <s>I think you were the first to touch on probability which should be worth a ∆ for presenting a perspective I hadn't fully considered. [NEWLINE] [NEWLINE] The only real problem I have with this explanation is that it could be oversimplified. It would only hold if we assume that the bisexual will have significantly less sex partners of the opposite sex than the heterosexual and some other variables that may or may not be relevant anyway. [NEWLINE] [NEWLINE] I would need to see some kind of analysis to be convinced to change my belief that we do not have enough information claim heterosexuality is advantageous but before I wasn't even fully aware of this route so I count it as a "Change of Perspective". </s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Now, in an ideal world, we would scrap ALL nukes from ALL countries.<mask><mask><mask><mask> some potentially dangerous countries have nukes, it's good for some more sane countries to have nukes too to keep them in check. [ENDQ] [NEWLINE] This actually doesn't follow. Consider the same scenario in the case<mask> neither country has nuclear weapons.<mask> B has stronger (or even comparable) military forces there's a good chance they will attack, and even<mask> they lose the war will kill many thousands or even millions of people. Compared to scenario 1<mask> there's no war at all it seems to me that we're much better off<mask> everyone has nukes.</s>
Label encoding: <s> [STARTQ] Now, in an ideal world, we would scrap ALL nukes from ALL countries. But as long as some potentially dangerous countries have nukes, it's good for some more sane countries to have nukes too to keep them in check. [ENDQ] [NEWLINE] This actually doesn't follow. Consider the same scenario in the case where neither country has nuclear weapons. If B has stronger (or even comparable) military forces there's a good chance they will attack, and even if they lose the war will kill many thousands or even millions of people. Compared to scenario 1 where there's no war at all it seems to me that we're much better off when everyone has nukes.</s>
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When you say "shit happens" do you mean they lose their job or that a family just like appears?  Cause losing a good job, means they have skills, and sooner or later they can get a good job back, and we have welfare programs (unemployment payments) to help them out until then.  And<mask> you're saying that they just "shit happened" into a family, that doesn't happen.  (Unless you're a male that gets raped and the woman decides to keep the kid and then you could be stuck with a kid of some woman in prison.  That's the only way you can have a family without consenting to it.)</s><pad>
Label encoding: <s>When you say "shit happens" do you mean they lose their job or that a family just like appears?  Cause losing a good job, means they have skills, and sooner or later they can get a good job back, and we have welfare programs (unemployment payments) to help them out until then.  And if you're saying that they just "shit happened" into a family, that doesn't happen.  (Unless you're a male that gets raped and the woman decides to keep the kid and then you could be stuck with a kid of some woman in prison.  That's the only way you can have a family without consenting to it.)</s><pad>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Any word like "typical", "normal" and<mask> on implies a value judgement. You could use "common" and "uncommon"<mask> the most value-neutral word for that concept I can think of,<mask> that doesn't really mean the same thing. [NEWLINE] [NEWLINE] Heteronormativity is the social pressure in society to only acknowledge straight people and straight relationships. It's<mask> true that straight relationships really are more common than queer ones<mask> that's not directly related in any way. [NEWLINE] [NEWLINE] Long story short<mask> you call straight relationships "typical", "normal" or any other word that implies "expected" that is itself heteronormative. </s>
Label encoding: <s>Any word like "typical", "normal" and so on implies a value judgement. You could use "common" and "uncommon" as the most value-neutral word for that concept I can think of, but that doesn't really mean the same thing. [NEWLINE] [NEWLINE] Heteronormativity is the social pressure in society to only acknowledge straight people and straight relationships. It's also true that straight relationships really are more common than queer ones but that's not directly related in any way. [NEWLINE] [NEWLINE] Long story short when you call straight relationships "typical", "normal" or any other word that implies "expected" that is itself heteronormative. </s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You might find the alternative more terrifying: not being able to die. [NEWLINE] [NEWLINE] In such a world, experiencing the physical aspect of dying might even become an escape from the boredom of eternal existence. Such a world is described in [The Metamorphosis of Prime Intellect]( [URL] /),<mask> dying has become a sport known<mask> "death jockeying". [NEWLINE] [NEWLINE] Aside from the physical, there are only your emotional attachments (the state of death itself being nothing at all).<mask><mask> the only way to be free of the fear of death, short of joining a religion, is to extricate oneself from existing attachments and actively cultivate non-attachment. [NEWLINE] </s>
Label encoding: <s>You might find the alternative more terrifying: not being able to die. [NEWLINE] [NEWLINE] In such a world, experiencing the physical aspect of dying might even become an escape from the boredom of eternal existence. Such a world is described in [The Metamorphosis of Prime Intellect]( [URL] /), where dying has become a sport known as "death jockeying". [NEWLINE] [NEWLINE] Aside from the physical, there are only your emotional attachments (the state of death itself being nothing at all). I think the only way to be free of the fear of death, short of joining a religion, is to extricate oneself from existing attachments and actively cultivate non-attachment. [NEWLINE] </s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think it's reducible to this kind of simple model.<mask> there are plenty of other variables and other actors in the model: On the top of my head, mainly law enforcement, which always have guns even<mask> "pop has no guns". It seems that people have grown to distrust or not rely on law enforcement in the US,<mask> maybe the problems come from there,<mask> law enforcement was better there would be no need for discussion about guns in regards to self-defense.<mask> you count law enforcement in your "pop" category, then even in countries with no guns in the general population, then "pop has guns" still apply in a sense?</s>
Label encoding: <s>I don't think it's reducible to this kind of simple model. Because there are plenty of other variables and other actors in the model: On the top of my head, mainly law enforcement, which always have guns even when "pop has no guns". It seems that people have grown to distrust or not rely on law enforcement in the US, but maybe the problems come from there, if law enforcement was better there would be no need for discussion about guns in regards to self-defense. If you count law enforcement in your "pop" category, then even in countries with no guns in the general population, then "pop has guns" still apply in a sense?</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you were planning on selling the car anyway, then you should be able to sell it after I replace the vehicle. After the car is replaced and you sell it, then you may do whatever you'd like with those funds. [NEWLINE] [NEWLINE] I believe that it's unacceptable<mask> you actually haven't been inconvenienced, and have instead actually been helped, by the totaling of the car. I am not dictating<mask> the money is used for, the law technically would (<mask> that is<mask> I am trying to change). You would be compensated fully for your vehicle, or even have a vehicle (the same model, year, etc.) in place of your old one. </s><pad>
Label encoding: <s>If you were planning on selling the car anyway, then you should be able to sell it after I replace the vehicle. After the car is replaced and you sell it, then you may do whatever you'd like with those funds. [NEWLINE] [NEWLINE] I believe that it's unacceptable because you actually haven't been inconvenienced, and have instead actually been helped, by the totaling of the car. I am not dictating what the money is used for, the law technically would ( as that is what I am trying to change). You would be compensated fully for your vehicle, or even have a vehicle (the same model, year, etc.) in place of your old one. </s><pad>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don think that. The woman would be the one crashing the car,<mask> she cant be forced to give her bodily resources to the baby( guy who needs kidney).<mask> she would have to pay the dudes medical bills.<mask> did i say the man was at fault? Im saying,<mask> both parents are at fault, you cant force either to use their bodily resources for the baby<mask> both can be forced to support it financially. The fact that woman are the only one who can give their bodily resources to the baby is unfortunate and unfair to both parties. Once technology advances to do this without the womans need to be involved then that can make it more fair.</s>
Label encoding: <s>I don think that. The woman would be the one crashing the car, but she cant be forced to give her bodily resources to the baby( guy who needs kidney). But she would have to pay the dudes medical bills. Where did i say the man was at fault? Im saying, while both parents are at fault, you cant force either to use their bodily resources for the baby but both can be forced to support it financially. The fact that woman are the only one who can give their bodily resources to the baby is unfortunate and unfair to both parties. Once technology advances to do this without the womans need to be involved then that can make it more fair.</s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It certainly would mean higher taxes (assuming you don't cut elsewhere like military spending);<mask> it *could* mean lower healthcare *costs*.  Right now lots of uninsured people do two things that are very expensive for everyone.  They will ignore problems until they become unignorable; which not only makes their health outcome statistically worse,<mask><mask> makes their treatment cost more.  The other thing the uninsured do is go to the emergency room for non-emergencies,<mask> it's the only access to healthcare they have. <mask> we had universal coverage, then many emergency room visits could be handled during scheduled doctors appointments at a much lower cost to everyone.</s>
Label encoding: <s>It certainly would mean higher taxes (assuming you don't cut elsewhere like military spending); but it *could* mean lower healthcare *costs*.  Right now lots of uninsured people do two things that are very expensive for everyone.  They will ignore problems until they become unignorable; which not only makes their health outcome statistically worse, but also makes their treatment cost more.  The other thing the uninsured do is go to the emergency room for non-emergencies, because it's the only access to healthcare they have.  If we had universal coverage, then many emergency room visits could be handled during scheduled doctors appointments at a much lower cost to everyone.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Certainly. In modern context, they are completely interchangeable terms with regard to the US government. The term "representative democracy" is all over middle and high school text books too by the way. I can specifically remember having to define a representative democracy versus a direct democracy. [NEWLINE] [NEWLINE] <mask> someone randomly came up to me and asked me to name the form of government practiced by the US, I would say a democracy. Not<mask> saying republic is wrong or<mask> saying democracy is more correct. Its just the term I more commonly use to describe this type of government and I know it won't be confused with a direct democracy<mask> they no longer exist anywhere on a national level.</s>
Label encoding: <s>Certainly. In modern context, they are completely interchangeable terms with regard to the US government. The term "representative democracy" is all over middle and high school text books too by the way. I can specifically remember having to define a representative democracy versus a direct democracy. [NEWLINE] [NEWLINE] If someone randomly came up to me and asked me to name the form of government practiced by the US, I would say a democracy. Not because saying republic is wrong or because saying democracy is more correct. Its just the term I more commonly use to describe this type of government and I know it won't be confused with a direct democracy because they no longer exist anywhere on a national level.</s>
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But ISIL is *not* and entirely new sounding entity.  The US government is regularly using that name to refer to<mask> most of the rest of the media are calling "ISIS".  The media reports on the government's statements, often with direct quotes,<mask> we often get "ISIL" and "ISIS" used in the same sentence (or at least paragraph) to refer to the same group. [NEWLINE] [NEWLINE] The confusion (<mask> there is any) is due to two names being used concurrently.  The most elegant solution is for one of the two names to be adopted universally. <mask><mask> the better name for that adoption would be "ISIL".</s>
Label encoding: <s>But ISIL is *not* and entirely new sounding entity.  The US government is regularly using that name to refer to what most of the rest of the media are calling "ISIS".  The media reports on the government's statements, often with direct quotes, so we often get "ISIL" and "ISIS" used in the same sentence (or at least paragraph) to refer to the same group. [NEWLINE] [NEWLINE] The confusion ( if there is any) is due to two names being used concurrently.  The most elegant solution is for one of the two names to be adopted universally.  I think the better name for that adoption would be "ISIL".</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Bitcoin is hybrid in nature. <mask> it works<mask> a currency, it<mask> works<mask> a money transferring service and people speculate that the network will find other uses over time.  Currently, merchants using Bitpay or Coinbase can set up an account<mask> they can recieve payments in Bitcoin without ever touching them or experiencing fluctuation. <mask> a bitcoin is just a representation of value, applications can be designed to take an input of any currency on the sending end and spit them back out in a different currency on the receiving end.  It doesn't take much imagination to think of it's other potential uses. It's<mask> people are<mask> fascinated by it. [NEWLINE] </s>
Label encoding: <s>Bitcoin is hybrid in nature.  While it works as a currency, it also works as a money transferring service and people speculate that the network will find other uses over time.  Currently, merchants using Bitpay or Coinbase can set up an account where they can recieve payments in Bitcoin without ever touching them or experiencing fluctuation.  Since a bitcoin is just a representation of value, applications can be designed to take an input of any currency on the sending end and spit them back out in a different currency on the receiving end.  It doesn't take much imagination to think of it's other potential uses. It's why people are so fascinated by it. [NEWLINE] </s>
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> basically you're the guy who thinks rape isn't<mask> bad of a crime<mask> people make it out to be... [ENDQ] [NEWLINE] Never said that. The specific words I used are "I don't understand<mask> ". There's a big difference. [NEWLINE] [NEWLINE] <mask> for armed robbery devaluing the victim<mask> a person, yes. This is the case for any crime<mask> a criminal disregards the other victim's personhood for their own personal gains, is it not? [NEWLINE] [NEWLINE] EDIT: For further explanation, in the case of armed robbery, the criminal disregards their victim's desires and personal character for the sole purpose of taking something from that victim, right?</s>
Label encoding: <s> [STARTQ] So basically you're the guy who thinks rape isn't as bad of a crime as people make it out to be... [ENDQ] [NEWLINE] Never said that. The specific words I used are "I don't understand why ". There's a big difference. [NEWLINE] [NEWLINE] As for armed robbery devaluing the victim as a person, yes. This is the case for any crime where a criminal disregards the other victim's personhood for their own personal gains, is it not? [NEWLINE] [NEWLINE] EDIT: For further explanation, in the case of armed robbery, the criminal disregards their victim's desires and personal character for the sole purpose of taking something from that victim, right?</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Down's Syndrome is caused by a physical failure in the reproductive process, it's a result of genetic material,<mask> it's not like some inherited trait and would not be passed down in that sense.  There is a rare form of Down's Syndrome that *is* geneticly inherited,<mask> we<mask> know<mask> to check for that, and it only accounts for a small percent of people with DS. [NEWLINE] [NEWLINE] Wait<mask>,<mask> about the carriers of this genetic trait for whom it is not expressed?  They are fully functional,<mask> would have the same probability of passing down DS<mask> the people who had the trait expressed.  Sterlize them too?</s>
Label encoding: <s>Down's Syndrome is caused by a physical failure in the reproductive process, it's a result of genetic material, but it's not like some inherited trait and would not be passed down in that sense.  There is a rare form of Down's Syndrome that *is* geneticly inherited, but we also know how to check for that, and it only accounts for a small percent of people with DS. [NEWLINE] [NEWLINE] Wait though, what about the carriers of this genetic trait for whom it is not expressed?  They are fully functional, but would have the same probability of passing down DS as the people who had the trait expressed.  Sterlize them too?</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Counting calories isn't eccentric or obsessive.  Caloric intake vs. caloric demand is a pretty simple concept, and people would probably do well to understand it<mask> one piece of the metabolic puzzle.  Understanding the calorie<mask> a unit of energy and<mask> that affects your metabolism is probably a step in the right direction for most people. [NEWLINE] [NEWLINE] It's not a ridiculous burden for restaurants either.  Calorie counts can easily be expressed<mask> a reasonable range, and most restaurants, even ones that aren't "premade and standardized",<mask> you say, cook with a degree of consistency that would allow them to ballpark an entree to within 50-100 calories.</s>
Label encoding: <s>Counting calories isn't eccentric or obsessive.  Caloric intake vs. caloric demand is a pretty simple concept, and people would probably do well to understand it as one piece of the metabolic puzzle.  Understanding the calorie as a unit of energy and how that affects your metabolism is probably a step in the right direction for most people. [NEWLINE] [NEWLINE] It's not a ridiculous burden for restaurants either.  Calorie counts can easily be expressed as a reasonable range, and most restaurants, even ones that aren't "premade and standardized", as you say, cook with a degree of consistency that would allow them to ballpark an entree to within 50-100 calories.</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] To you it really doesn't matter whether you push the button or not.<mask> there's no more reason not to push it than there is to push it. For you personally, it comes down to whether or not you want revenge. Is it petty to want revenge for your murder? I don't think<mask>. Plenty of people go to court seeking justice for the murder of a loved one, and that's not petty. [ENDQ] [NEWLINE] Another way to see it:<mask> you know you're dead, then your decision shouldn't be based on<mask>'s right for you (getting your revenge),<mask> based on<mask>'s best for those who will still be alive.</s>
Label encoding: <s> [STARTQ] To you it really doesn't matter whether you push the button or not. So there's no more reason not to push it than there is to push it. For you personally, it comes down to whether or not you want revenge. Is it petty to want revenge for your murder? I don't think so. Plenty of people go to court seeking justice for the murder of a loved one, and that's not petty. [ENDQ] [NEWLINE] Another way to see it: if you know you're dead, then your decision shouldn't be based on what's right for you (getting your revenge), but based on what's best for those who will still be alive.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it forces many anti-vaxxers into even more of an echo chamber. One of the must successful ways to spread information is to have anti-vaxxers have a friend or a neighbor who can discuss these things with them. [ENDQ] [NEWLINE] This is an outcome I hadn't thought of. I'll award a ∆ for a partial view change. This could cause more people to choose to isolate themselves socially or geographically around vaccine-belief lines, which is just bad for everyone in the long run. I still think in the short term it is appropriate to mandate vaccination,<mask> I hadn't thought<mask> it could affect perceptions over time.</s>
Label encoding: <s> [STARTQ] So it forces many anti-vaxxers into even more of an echo chamber. One of the must successful ways to spread information is to have anti-vaxxers have a friend or a neighbor who can discuss these things with them. [ENDQ] [NEWLINE] This is an outcome I hadn't thought of. I'll award a ∆ for a partial view change. This could cause more people to choose to isolate themselves socially or geographically around vaccine-belief lines, which is just bad for everyone in the long run. I still think in the short term it is appropriate to mandate vaccination, but I hadn't thought how it could affect perceptions over time.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Fast-forward a thousand Earth years. [ENDQ] [NEWLINE] <mask> the Church teaches that Heaven is outside of time.  Trying to apply the concept of time, with one event happening "after" another is meaningless<mask> discussing the Christian conception of Heaven.  Remember that someone who enters Heaven before or after other people (whether by dying at a later date, or by Purgatory<mask> you believe in it) has the identical reward.  It is not "Heaven for a thousand fewer years", it is "Heaven."  Being outside of time, the ways in which you are conceiving of boredom, events, and years are totally inapplicable.</s>
Label encoding: <s> [STARTQ] Fast-forward a thousand Earth years. [ENDQ] [NEWLINE] But the Church teaches that Heaven is outside of time.  Trying to apply the concept of time, with one event happening "after" another is meaningless when discussing the Christian conception of Heaven.  Remember that someone who enters Heaven before or after other people (whether by dying at a later date, or by Purgatory if you believe in it) has the identical reward.  It is not "Heaven for a thousand fewer years", it is "Heaven."  Being outside of time, the ways in which you are conceiving of boredom, events, and years are totally inapplicable.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> in that sense, no one has a right to life. [ENDQ] [NEWLINE] <mask> does that follow? [NEWLINE] [NEWLINE] [STARTQ] We still have to draw a line<mask> a human society on<mask> is acceptable, and to some atheists or libertarians, that line could be conception. [ENDQ] [NEWLINE] <mask><mask> drawing the line there is absurd *in general*,<mask> I haven't claimed that it's incompatible with either libertarianism or atheism (<mask> it's in tension with the naturalism usually esposed by atheists<mask> well<mask> other moral intuitions). [NEWLINE] [NEWLINE] <mask>, denying bodily autonomy to a woman seems like a very hard,<mask> not impossible, pill to swallow for a libertarian.</s>
Label encoding: <s> [STARTQ] But in that sense, no one has a right to life. [ENDQ] [NEWLINE] How does that follow? [NEWLINE] [NEWLINE] [STARTQ] We still have to draw a line as a human society on what is acceptable, and to some atheists or libertarians, that line could be conception. [ENDQ] [NEWLINE] I think drawing the line there is absurd *in general*, so I haven't claimed that it's incompatible with either libertarianism or atheism ( although it's in tension with the naturalism usually esposed by atheists as well as other moral intuitions). [NEWLINE] [NEWLINE] However, denying bodily autonomy to a woman seems like a very hard, if not impossible, pill to swallow for a libertarian.</s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do we draw the line? [ENDQ] [NEWLINE] <mask> should we? /r/basicincome [NEWLINE] [NEWLINE] That said, orphanages/foster care is a state system (or should be).<mask> life support and late pregnancy were funded and compensated for by the state, and this were justified<mask> essentially pre-birth welfare in the public interest, then<mask><mask> that might provide some degree of justification for an anti-abortion policy. [NEWLINE] [NEWLINE] The public interest does sometimes override property and ownership rights, and sometimes even medical authority (example: parents who insist on 'prayer healthcare'). I'm not sure<mask> this is a practical use for this principle,<mask>.</s>
Label encoding: <s> [STARTQ] Where do we draw the line? [ENDQ] [NEWLINE] Why should we? /r/basicincome [NEWLINE] [NEWLINE] That said, orphanages/foster care is a state system (or should be). If life support and late pregnancy were funded and compensated for by the state, and this were justified as essentially pre-birth welfare in the public interest, then I think that might provide some degree of justification for an anti-abortion policy. [NEWLINE] [NEWLINE] The public interest does sometimes override property and ownership rights, and sometimes even medical authority (example: parents who insist on 'prayer healthcare'). I'm not sure if this is a practical use for this principle, though.</s>
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Multiple, private, competing agencies who are authorized to use force to enforce contracts by their paying customers. Judges, lawyers, and police are all different companies, removing the crazy conflict of interest that currently exists<mask> they are all on the same (government) payroll. Competition and reputation is used<mask> a mechanism to ensure the most level playing field possible.<mask> you can withdraw your funding from a company who doesn't do<mask> you like, companies will cater to provide you with service you agree with, unlike the government. [NEWLINE] [NEWLINE] The concept is called [Polycentric Law]( [URL] ) [NEWLINE] [NEWLINE] For an intro - you can check out [this video]( [URL] ). </s>
Label encoding: <s>Multiple, private, competing agencies who are authorized to use force to enforce contracts by their paying customers. Judges, lawyers, and police are all different companies, removing the crazy conflict of interest that currently exists when they are all on the same (government) payroll. Competition and reputation is used as a mechanism to ensure the most level playing field possible. When you can withdraw your funding from a company who doesn't do what you like, companies will cater to provide you with service you agree with, unlike the government. [NEWLINE] [NEWLINE] The concept is called [Polycentric Law]( [URL] ) [NEWLINE] [NEWLINE] For an intro - you can check out [this video]( [URL] ). </s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Mm, I don't really agree that it's unfair to tax the richer and poorer people in the same bracket at the same rate. We don't tax the rich and the poor at the same rate<mask> it's literally impossible to do<mask><mask> collecting the needed revenue and not killing all the poor people. It doesn't really matter<mask> someone making $250,000 and someone making $300,000 both get taxed at the same rate<mask><mask> at the end of the day they'll both still have a ton of money.<mask> neither of them will really miss the money I don't see anything unfair about giving them the same relative tax burden. [NEWLINE] </s>
Label encoding: <s>Mm, I don't really agree that it's unfair to tax the richer and poorer people in the same bracket at the same rate. We don't tax the rich and the poor at the same rate because it's literally impossible to do so while collecting the needed revenue and not killing all the poor people. It doesn't really matter if someone making $250,000 and someone making $300,000 both get taxed at the same rate though because at the end of the day they'll both still have a ton of money. Since neither of them will really miss the money I don't see anything unfair about giving them the same relative tax burden. [NEWLINE] </s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It depends on<mask> your goal is with such a philosophy. [NEWLINE] [NEWLINE] A: Is the goal to improve the lives of each individual to the maximum degree? [NEWLINE] [NEWLINE] B: Is the goal to improve YOUR life to the maximum degree? ^(And possibly the lives of a few people you care most about ) [NEWLINE] [NEWLINE] C: Is the goal to improve humanity<mask> a whole? [NEWLINE] [NEWLINE] D: Is the goal to do whatever is morally just and right? [NEWLINE] [NEWLINE] E: Some other goal entirely. [NEWLINE] [NEWLINE] I can't tell you whether this idea would work out in favor of your goals<mask> I don't know<mask> you are trying to accomplish.</s>
Label encoding: <s>It depends on what your goal is with such a philosophy. [NEWLINE] [NEWLINE] A: Is the goal to improve the lives of each individual to the maximum degree? [NEWLINE] [NEWLINE] B: Is the goal to improve YOUR life to the maximum degree? ^(And possibly the lives of a few people you care most about ) [NEWLINE] [NEWLINE] C: Is the goal to improve humanity as a whole? [NEWLINE] [NEWLINE] D: Is the goal to do whatever is morally just and right? [NEWLINE] [NEWLINE] E: Some other goal entirely. [NEWLINE] [NEWLINE] I can't tell you whether this idea would work out in favor of your goals if I don't know what you are trying to accomplish.</s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with this.  I personally don't like the ambiguity of religious organizations<mask> the current code applies.  The accounting and activities that a charitable non-profit demonstrates to acquire exemption are fairly easy to identify,<mask> more esoteric religions may be difficult to establish<mask> such.  Scientology is largely based on works of science fiction, not that I question the validity of L. Ron Hubbard or of Christ,<mask> then neither should the state.  The Flying Spaghetti Monster is no different in observable substance,<mask><mask> Pastafarians were to engage in charitable activities per the tax code, I would feel much more inclined to approve of their exemption status. </s>
Label encoding: <s>I agree with this.  I personally don't like the ambiguity of religious organizations as the current code applies.  The accounting and activities that a charitable non-profit demonstrates to acquire exemption are fairly easy to identify, but more esoteric religions may be difficult to establish as such.  Scientology is largely based on works of science fiction, not that I question the validity of L. Ron Hubbard or of Christ, but then neither should the state.  The Flying Spaghetti Monster is no different in observable substance, but if Pastafarians were to engage in charitable activities per the tax code, I would feel much more inclined to approve of their exemption status. </s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point is that the bible is not a mind blowing book<mask> you don't have faith to begin with.  The people who debate it start with the premise that it's the literal or figurative word of god, save those who<mask><mask> it isn't. [NEWLINE] [NEWLINE] <mask> I don't read with 100% comprehension,<mask><mask> I understood<mask> I was reading for the most part.  Without the faith that it's 100% accurate (<mask> some believe) or divinely inspired (<mask> all christians believe) it's just not all that great.  I doubt that many people would be inspired to become christians simply by reading the bible.</s>
Label encoding: <s>The point is that the bible is not a mind blowing book if you don't have faith to begin with.  The people who debate it start with the premise that it's the literal or figurative word of god, save those who argue that it isn't. [NEWLINE] [NEWLINE] While I don't read with 100% comprehension, I think I understood what I was reading for the most part.  Without the faith that it's 100% accurate ( as some believe) or divinely inspired ( as all christians believe) it's just not all that great.  I doubt that many people would be inspired to become christians simply by reading the bible.</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have not read Ayn Rand, nor do I ever intend to,<mask> I consider her philosophy to be frankly immoral. Moral people should look out for their fellow humans. [ENDQ] [NEWLINE] <mask> does your position differ substantially from ms rand's? [NEWLINE] [NEWLINE] [STARTQ] betterment of others who will just attribute my good deed to their god anyway most likely [ENDQ] [NEWLINE] <mask> is the significance of god? [NEWLINE] [NEWLINE] on a somewhat personal note<mask> long<mask> your view on this changed and (presumably in general terms<mask> i can only presume you wont want to go into too much detail)<mask> happened to change your view from the advocacy of others? [NEWLINE] EDIT: formatting</s>
Label encoding: <s> [STARTQ] I have not read Ayn Rand, nor do I ever intend to, as I consider her philosophy to be frankly immoral. Moral people should look out for their fellow humans. [ENDQ] [NEWLINE] how does your position differ substantially from ms rand's? [NEWLINE] [NEWLINE] [STARTQ] betterment of others who will just attribute my good deed to their god anyway most likely [ENDQ] [NEWLINE] what is the significance of god? [NEWLINE] [NEWLINE] on a somewhat personal note how long since your view on this changed and (presumably in general terms as i can only presume you wont want to go into too much detail) what happened to change your view from the advocacy of others? [NEWLINE] EDIT: formatting</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0708, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh, I knew about that. I just figured "toll-kill" and "ka-lao-keh" would've been close enough<mask> much easier to not mess up for people who aren't familiar with Japanese sounds at all. [NEWLINE] [NEWLINE] <mask><mask> a lot of people would see "kyo" and still say "key-yo".<mask>,<mask><mask> the "r" sound,<mask> in-between, is actually closer to the American l than the American r,<mask> it has a "flicking" action with the tongue,<mask> faint it may be. [NEWLINE] [NEWLINE] Thanks for responding after<mask> many days, lol.</s>
Label encoding: <s>Oh, I knew about that. I just figured "toll-kill" and "ka-lao-keh" would've been close enough but much easier to not mess up for people who aren't familiar with Japanese sounds at all. [NEWLINE] [NEWLINE] I think a lot of people would see "kyo" and still say "key-yo". Also, I think the "r" sound, while in-between, is actually closer to the American l than the American r, since it has a "flicking" action with the tongue, however faint it may be. [NEWLINE] [NEWLINE] Thanks for responding after so many days, lol.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0714, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I'm saying is, there are several different parts of the brain that are sexually dimporphic. Some of these parts of the brain are aplastic, and no amount of hormones of either type will change whether they are male shaped or female shaped. Some parts are plastic, and will become male shaped in the presence of Testosterone, and female shaped in the presence of Estrogen.<mask> the parts of the brain that are Aplastic can sometimes be halfway between male and female shaped, they might not be saying male *or* female very strongly, and<mask> leave more control in the hands of the plastic parts of the brain.</s>
Label encoding: <s>What I'm saying is, there are several different parts of the brain that are sexually dimporphic. Some of these parts of the brain are aplastic, and no amount of hormones of either type will change whether they are male shaped or female shaped. Some parts are plastic, and will become male shaped in the presence of Testosterone, and female shaped in the presence of Estrogen. Since the parts of the brain that are Aplastic can sometimes be halfway between male and female shaped, they might not be saying male *or* female very strongly, and thus leave more control in the hands of the plastic parts of the brain.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Other posters have done a great job of covering the major arguments<mask> here's a more subtle one. [NEWLINE] [NEWLINE] Let's say that you pirated a 2-hour movie and then watched it.<mask> piracy were somehow (magically) impossible,<mask> would you have done with those 2 hours instead? [NEWLINE] [NEWLINE] I'd<mask><mask><mask> you pirate a product you never would have bought, you deprive competing products of money too. Imagine<mask> much smaller your library of entertainment would be<mask> you couldn't pirate anything. Then you might need something to fill your free time and products that you currently think are slightly too expensive would move into your price range.</s>
Label encoding: <s>Other posters have done a great job of covering the major arguments so here's a more subtle one. [NEWLINE] [NEWLINE] Let's say that you pirated a 2-hour movie and then watched it. If piracy were somehow (magically) impossible, what would you have done with those 2 hours instead? [NEWLINE] [NEWLINE] I'd argue that when you pirate a product you never would have bought, you deprive competing products of money too. Imagine how much smaller your library of entertainment would be if you couldn't pirate anything. Then you might need something to fill your free time and products that you currently think are slightly too expensive would move into your price range.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah I know<mask> you mean. I just think there's a link you're having making between "I don't personally understad/like this system" and "This isn't a valid system" in both the art and evaluation of art.<mask> you don't like eggs you don't say "That's not food. Eggs aren't worth any money." You can see that the world clearly doesn't value things on utility alone, and that's certainly an arguable point,<mask> it exists and you can't eject it from your world view.<mask> you do you won't be able to understand things like modern art selling for millions. </s>
Label encoding: <s>Yeah I know what you mean. I just think there's a link you're having making between "I don't personally understad/like this system" and "This isn't a valid system" in both the art and evaluation of art. If you don't like eggs you don't say "That's not food. Eggs aren't worth any money." You can see that the world clearly doesn't value things on utility alone, and that's certainly an arguable point, but it exists and you can't eject it from your world view. If you do you won't be able to understand things like modern art selling for millions. </s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What I was getting at with my questions: [NEWLINE] <mask> you have a "bad" teacher, you fail to learn and<mask> perform poorly on tests. Those underperforming teacher should<mask> be removed from teaching. [NEWLINE] [NEWLINE] <mask> you have a "good" teacher, you successfully learn the material, and sometimes in doing<mask>. [NEWLINE] [NEWLINE] <mask> you had a good teacher, you perform well on tests. I've never known anyone who honestly felt they knew the material being taught,<mask> still failed the test. [NEWLINE] [NEWLINE] This is essentially the system we have now,<mask> I feel we need to do more to encourage good teaching, and punish poor teaching.</s>
Label encoding: <s>What I was getting at with my questions: [NEWLINE] When you have a "bad" teacher, you fail to learn and thus perform poorly on tests. Those underperforming teacher should thus be removed from teaching. [NEWLINE] [NEWLINE] When you have a "good" teacher, you successfully learn the material, and sometimes in doing so. [NEWLINE] [NEWLINE] Because you had a good teacher, you perform well on tests. I've never known anyone who honestly felt they knew the material being taught, but still failed the test. [NEWLINE] [NEWLINE] This is essentially the system we have now, but I feel we need to do more to encourage good teaching, and punish poor teaching.</s>
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Nitpick: Steven Sotloff was Jewish and was killed by ISIS. [NEWLINE] [NEWLINE] More importantly, look at attacks like the Charlie Hebdo shooting that followed up with an attack on a Jewish grocery.  Or the Mumbai attacks that had to include the local Chabad.  Or the Kansas City Jewish community shootings.  Jews are not generally targets of random violence,<mask><mask> it comes to racially motivated mass murder they are clearly heavily targeted. [NEWLINE] [NEWLINE] Not to mention all the semicovert claims that the bankers and neocons and Israel lobby control the US.  Even Bernie Sanders was accused of being an Israeli citizen by Diane Rehm... [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Nitpick: Steven Sotloff was Jewish and was killed by ISIS. [NEWLINE] [NEWLINE] More importantly, look at attacks like the Charlie Hebdo shooting that followed up with an attack on a Jewish grocery.  Or the Mumbai attacks that had to include the local Chabad.  Or the Kansas City Jewish community shootings.  Jews are not generally targets of random violence, but when it comes to racially motivated mass murder they are clearly heavily targeted. [NEWLINE] [NEWLINE] Not to mention all the semicovert claims that the bankers and neocons and Israel lobby control the US.  Even Bernie Sanders was accused of being an Israeli citizen by Diane Rehm... [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't see<mask> it matters. We aren't designed for any specific purpose.<mask> people choose monogamy, whatever their brains may be craving deep down, they're doing monogamy.<mask><mask><mask> it doesn't kill us it's a net positive. Or at least neutral. I don't care<mask> people choose to be monogamous or not. We are not designed to be polyamorous any more than designed to be monogamous. We aren't designed to be either. We exist and make choices and those choices determine our survival and procreation. That which survives isn't necessarily right or wrong it's just<mask> survives. </s>
Label encoding: <s>I don't see how it matters. We aren't designed for any specific purpose. If people choose monogamy, whatever their brains may be craving deep down, they're doing monogamy. As long as it doesn't kill us it's a net positive. Or at least neutral. I don't care if people choose to be monogamous or not. We are not designed to be polyamorous any more than designed to be monogamous. We aren't designed to be either. We exist and make choices and those choices determine our survival and procreation. That which survives isn't necessarily right or wrong it's just what survives. </s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes I watched the hospice people try to talk the spouse into signing a no-intubation order (the spouse didn't understand this was essentially the same<mask> a DNR until later), watched them push morphine early, saw<mask> the oncologist elected to stop chemo (it was a drug interaction they should have caught). A parade of horribles, this one case,<mask> I have a feeling this kind of thing goes on quite a bit more often than most of us realize. Especially for people with subpar insurance, indigents, people without family members willing/able to look over caregivers' shoulders, etc.</s>
Label encoding: <s>Yes I watched the hospice people try to talk the spouse into signing a no-intubation order (the spouse didn't understand this was essentially the same as a DNR until later), watched them push morphine early, saw why the oncologist elected to stop chemo (it was a drug interaction they should have caught). A parade of horribles, this one case, but I have a feeling this kind of thing goes on quite a bit more often than most of us realize. Especially for people with subpar insurance, indigents, people without family members willing/able to look over caregivers' shoulders, etc.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll give you the ∆ on the waste of time point, which, frankly, was not<mask> well worded on my part. It is fair to<mask><mask> even<mask> OP does not change his mind, some people in the comments will.<mask><mask><mask> with your first paragraph: just<mask> deltas are awarded does not mean anything.<mask> I said, controversial issues exist<mask> both points are valid exist, and these posts are likely to dispense deltas liberally. The fact that deltas are awarded alone does not justify whether the time spent convincing a stupid person of a single point was time well spent.</s><pad>
Label encoding: <s>I'll give you the ∆ on the waste of time point, which, frankly, was not so well worded on my part. It is fair to argue that even if OP does not change his mind, some people in the comments will. But I disagree with your first paragraph: just because deltas are awarded does not mean anything. As I said, controversial issues exist where both points are valid exist, and these posts are likely to dispense deltas liberally. The fact that deltas are awarded alone does not justify whether the time spent convincing a stupid person of a single point was time well spent.</s><pad>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't disagree that it's okay,<mask><mask> that it's pragmatic.<mask><mask> people will act a certain way<mask> of **stereotypes** is hardly practical. [ENDQ] [NEWLINE] <mask> someone has had a lot of casual sex partners in the past, it's not stereotyping them to assume that this will continue in the future.  People's behavior demonstrates their values. <mask> the majority of their previous partners were in long term relationships, they value long term relationships. <mask> the majority of their previous partners were one night stands, they value casual sex.  It's always better to follow actions instead of words. </s>
Label encoding: <s> [STARTQ] I don't disagree that it's okay, I disagree that it's pragmatic. Assuming that people will act a certain way because of **stereotypes** is hardly practical. [ENDQ] [NEWLINE] If someone has had a lot of casual sex partners in the past, it's not stereotyping them to assume that this will continue in the future.  People's behavior demonstrates their values.  If the majority of their previous partners were in long term relationships, they value long term relationships.  If the majority of their previous partners were one night stands, they value casual sex.  It's always better to follow actions instead of words. </s>
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can't speak for the cops<mask> I was in the military stationed in Korea<mask> 9/11 happened. We were told to guard the gates with unloaded M16s and we asked the same kind of thing, "<mask> are we carrying weapons around<mask> we aren't prepared to use them?" [NEWLINE] [NEWLINE] That was very much a situation<mask> the show of force was more important than the use of force. [NEWLINE] [NEWLINE] I don't think<mask><mask> with<mask> the cop did in this situation<mask> I can see<mask> sometimes a show of force would be a good idea to deter any further escalation even<mask> you don't intend to fire it. [NEWLINE] </s>
Label encoding: <s>Can't speak for the cops but I was in the military stationed in Korea when 9/11 happened. We were told to guard the gates with unloaded M16s and we asked the same kind of thing, " Why are we carrying weapons around if we aren't prepared to use them?" [NEWLINE] [NEWLINE] That was very much a situation where the show of force was more important than the use of force. [NEWLINE] [NEWLINE] I don't think I agree with what the cop did in this situation but I can see how sometimes a show of force would be a good idea to deter any further escalation even when you don't intend to fire it. [NEWLINE] </s>
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Without a government to back up court verdicts with violence, suing is worthless [ENDQ] [NEWLINE] Nope. Let's say A owes B some money. The court rules in B's favor. Now A would have their economic reputation (credit score, business reputation, etc) tarnished and B now has the right to take part of A's property [NEWLINE] [NEWLINE] [STARTQ] And it wouldn't have to be personal, that could just be libertarian-CEO guy's standard way of getting employees to do<mask> they're told. [ENDQ] [NEWLINE] <mask> that were company policy, nobody would want to work for that company and it would soon go bankrupt. </s>
Label encoding: <s> [STARTQ] Without a government to back up court verdicts with violence, suing is worthless [ENDQ] [NEWLINE] Nope. Let's say A owes B some money. The court rules in B's favor. Now A would have their economic reputation (credit score, business reputation, etc) tarnished and B now has the right to take part of A's property [NEWLINE] [NEWLINE] [STARTQ] And it wouldn't have to be personal, that could just be libertarian-CEO guy's standard way of getting employees to do what they're told. [ENDQ] [NEWLINE] If that were company policy, nobody would want to work for that company and it would soon go bankrupt. </s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> whenever I've come close or had an offer I get...anxious. [ENDQ] [NEWLINE] Op you are like me, the idea of casual sex with someone whom i'm not close with seems awkward and uncomfortable. To me sex gets better<mask> you really know the person, the first time is the worst time and I don't like people touching me that much unless I trust them. [NEWLINE] [NEWLINE] <mask> I know that some people are more adventurous and relaxed with this type of thing, its no big deal to them<mask> they don't have to face any mental hurdles to engage in casual sex.<mask><mask> for those people it is fine</s>
Label encoding: <s> [STARTQ] But whenever I've come close or had an offer I get...anxious. [ENDQ] [NEWLINE] Op you are like me, the idea of casual sex with someone whom i'm not close with seems awkward and uncomfortable. To me sex gets better when you really know the person, the first time is the worst time and I don't like people touching me that much unless I trust them. [NEWLINE] [NEWLINE] But I know that some people are more adventurous and relaxed with this type of thing, its no big deal to them because they don't have to face any mental hurdles to engage in casual sex. I think for those people it is fine</s>
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's become a cliche at this point,<mask> I believe that I can be absolutely 100% sure without an inkling of a doubt that **I exist**. I'm not saying I know *<mask> * I am, *who* I am,<mask> it is that it means to *exist*.<mask> whatever it is, I do it, and I'm doing it right now. Whatever I am, I am that. I am existing right now doing the things that I am doing and thinking the things I am thinking and I can be absolutely certain of that. No one can ever possibly prove it otherwise to me.</s>
Label encoding: <s>It's become a cliche at this point, but I believe that I can be absolutely 100% sure without an inkling of a doubt that **I exist**. I'm not saying I know * what * I am, *who* I am, what it is that it means to *exist*. But whatever it is, I do it, and I'm doing it right now. Whatever I am, I am that. I am existing right now doing the things that I am doing and thinking the things I am thinking and I can be absolutely certain of that. No one can ever possibly prove it otherwise to me.</s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To me a scamer who tells you the truth is better than a preacher who says this and that then has sex with the choirboy. [NEWLINE] [NEWLINE] [STARTQ] legitimacy...  questionable. [ENDQ] [NEWLINE] Well, they do have to keep themselves alive, after all they are non-profit organizations. Better the believers paying than the taxpayer... [NEWLINE] [NEWLINE] [STARTQ] going to worship and receive communion. [ENDQ] [NEWLINE] Not in every case.<mask> you are a Mormon and don't tithe you can not attend the temple.... [NEWLINE] [NEWLINE] [STARTQ] try to convert them is different than "overtaking the world", [ENDQ] [NEWLINE] History begs to differ....</s>
Label encoding: <s>To me a scamer who tells you the truth is better than a preacher who says this and that then has sex with the choirboy. [NEWLINE] [NEWLINE] [STARTQ] legitimacy...  questionable. [ENDQ] [NEWLINE] Well, they do have to keep themselves alive, after all they are non-profit organizations. Better the believers paying than the taxpayer... [NEWLINE] [NEWLINE] [STARTQ] going to worship and receive communion. [ENDQ] [NEWLINE] Not in every case. If you are a Mormon and don't tithe you can not attend the temple.... [NEWLINE] [NEWLINE] [STARTQ] try to convert them is different than "overtaking the world", [ENDQ] [NEWLINE] History begs to differ....</s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Id have to agree with this. Posers are chameleons. They embrace a type of culture in order to fit in. This will either end<mask> they realise they dont belong or decide they belong anyway and cease being posers<mask> they just are. [NEWLINE] [NEWLINE] Hipsters adapt other cultural styles and ideas<mask> a form of plumage in order to stand out amd attract attention. This<mask> course all culminates in the hipster cultural conundrum, - "<mask> everyone's special, no one is." Which is<mask><mask> hipsters spend too much time together you get an arms race of increasingly outlandish garments and habits.</s>
Label encoding: <s>Id have to agree with this. Posers are chameleons. They embrace a type of culture in order to fit in. This will either end when they realise they dont belong or decide they belong anyway and cease being posers as they just are. [NEWLINE] [NEWLINE] Hipsters adapt other cultural styles and ideas as a form of plumage in order to stand out amd attract attention. This if course all culminates in the hipster cultural conundrum, - " when everyone's special, no one is." Which is why when hipsters spend too much time together you get an arms race of increasingly outlandish garments and habits.</s>
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Interesting theory,<mask> i don't think we meet many of these points. The way the elections are run makes these points difficult to acheive. [NEWLINE] [NEWLINE] There are large differences in voter participation based on age and income. Individual votes are equally valued,<mask> individual votes don't matter, the votes of the groups you agree with are<mask> changes things.<mask><mask> we often stop felons from voting. These problems limit voting equality and potential to participation. The combination of a media which almost any political ideology has a problem with and the system which pushes for two well defined parties makes our control of the agenda very indirect.</s>
Label encoding: <s>Interesting theory, But i don't think we meet many of these points. The way the elections are run makes these points difficult to acheive. [NEWLINE] [NEWLINE] There are large differences in voter participation based on age and income. Individual votes are equally valued, but individual votes don't matter, the votes of the groups you agree with are what changes things. In addition we often stop felons from voting. These problems limit voting equality and potential to participation. The combination of a media which almost any political ideology has a problem with and the system which pushes for two well defined parties makes our control of the agenda very indirect.</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [STARTQ] See, there's no concept of "being gay" in that time period,<mask><mask> there's no Latin or Greek word from that time which corresponds to the modern meaning of "homosexuality". No one was "out" and all same-sex sexual occurences were between obstensibly *heterosexual people* and was in coercive situations. [ENDQ] [NEWLINE] <mask><mask> with most of your argument,<mask> it's important to point out that consensual, exclusive homosexuality did not just spring up in the 20th century. [Wikipedia]( [URL] ) is the best I can do<mask> I'm supposed to be working. [NEWLINE] </s>
Label encoding: <s> [NEWLINE] [STARTQ] See, there's no concept of "being gay" in that time period, in fact there's no Latin or Greek word from that time which corresponds to the modern meaning of "homosexuality". No one was "out" and all same-sex sexual occurences were between obstensibly *heterosexual people* and was in coercive situations. [ENDQ] [NEWLINE] I agree with most of your argument, but it's important to point out that consensual, exclusive homosexuality did not just spring up in the 20th century. [Wikipedia]( [URL] ) is the best I can do while I'm supposed to be working. [NEWLINE] </s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Half of your reply seems to refer to another post. [ENDQ] [NEWLINE] You're right.  My bad.  I deleted the part that was quoting another post, and reposted it<mask> it belongs.  Hope you don't mind. [NEWLINE] [NEWLINE] EDIT:  And I totally agree about the technology.  My biggest gripe with the scenario is that the capitol clearly have very advanced genetic engineering capabilities at their disposal. <mask> they use them to make mockingjays and stupid amusements in the arena, rather than just making GE miner creatures (for instance) which would remove the need for a slave workforce.</s>
Label encoding: <s> [STARTQ] Half of your reply seems to refer to another post. [ENDQ] [NEWLINE] You're right.  My bad.  I deleted the part that was quoting another post, and reposted it where it belongs.  Hope you don't mind. [NEWLINE] [NEWLINE] EDIT:  And I totally agree about the technology.  My biggest gripe with the scenario is that the capitol clearly have very advanced genetic engineering capabilities at their disposal.  But they use them to make mockingjays and stupid amusements in the arena, rather than just making GE miner creatures (for instance) which would remove the need for a slave workforce.</s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0721, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> there's a reason<mask> the New York Times is considered a newspaper of record<mask>, say, the National Inquirer is not. [ENDQ] [NEWLINE] <mask> the New York Times has a record of printing well-researched articles<mask> the National Enquirer prints anything that'll sell papers. [NEWLINE] [NEWLINE] The Mises Institute is likely to report selectively and only talk about research that supports its views,<mask> it still has to use actual research or else it would lose all credibility. All claims made in the article are themselves sourced,<mask> the Mises Institute is actually not even being used<mask> a source here.</s>
Label encoding: <s> [STARTQ] but there's a reason why the New York Times is considered a newspaper of record while, say, the National Inquirer is not. [ENDQ] [NEWLINE] Because the New York Times has a record of printing well-researched articles while the National Enquirer prints anything that'll sell papers. [NEWLINE] [NEWLINE] The Mises Institute is likely to report selectively and only talk about research that supports its views, but it still has to use actual research or else it would lose all credibility. All claims made in the article are themselves sourced, so the Mises Institute is actually not even being used as a source here.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0809, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Strike Suit Zero [ENDQ] [NEWLINE] Funnily enough I actually preferred KB&amp;M for Strike Suit Zero. [NEWLINE] [NEWLINE] [STARTQ] I wish there were more space combat sims on the market today [ENDQ] [NEWLINE] <mask> setphazerstopun said, Elite: Dangerous may be right up your alley. We've<mask> got Star Citizen to look forward to and Eve: Valkyrie. I loved X-Wing &amp; TIE Fighter back in the day and it's looking to be a great time to be a space sim fan. Add in the Oculus Rift or HTC Vive to the mix and it's going to be amazing.</s>
Label encoding: <s> [STARTQ] Strike Suit Zero [ENDQ] [NEWLINE] Funnily enough I actually preferred KB&amp;M for Strike Suit Zero. [NEWLINE] [NEWLINE] [STARTQ] I wish there were more space combat sims on the market today [ENDQ] [NEWLINE] As setphazerstopun said, Elite: Dangerous may be right up your alley. We've also got Star Citizen to look forward to and Eve: Valkyrie. I loved X-Wing &amp; TIE Fighter back in the day and it's looking to be a great time to be a space sim fan. Add in the Oculus Rift or HTC Vive to the mix and it's going to be amazing.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They are things. Biological things. You can change your gender, you can even change your *legal* sex,<mask> without future advancements in genetic therapy, you can't change your biological sex. OP is right in that much, at least. The only exceptions are rare genetic diseases<mask> someone may have a defective chromosome, additional chromosome, etc... that influences them to be a "hermaphrodite" genetically. Unfortunately, there's no such thing<mask> a true hermaphrodite,<mask> most end up<mask> a mutated version of one gender with secondary characteristics (even some primary) from the alternate.</s>
Label encoding: <s>They are things. Biological things. You can change your gender, you can even change your *legal* sex, but without future advancements in genetic therapy, you can't change your biological sex. OP is right in that much, at least. The only exceptions are rare genetic diseases where someone may have a defective chromosome, additional chromosome, etc... that influences them to be a "hermaphrodite" genetically. Unfortunately, there's no such thing as a true hermaphrodite, so most end up as a mutated version of one gender with secondary characteristics (even some primary) from the alternate.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They have been putting up with a lot of crap for a long time, [ENDQ] [NEWLINE] Ah, yes all their anguish over website trumps one person being compared to hitler on /r/all for days on end, sent death threats, and bombarded with tons of racism and sexist comments. I don't care<mask> any grief they have after the shit that was put on Reddit after the FPH ban and last weekends Victoria firing. A very vocal minority have shown they can really be assholes. [NEWLINE] [NEWLINE] Assholes shouldn't even get the time of day after the stunts that were pulled like that. </s>
Label encoding: <s> [STARTQ] They have been putting up with a lot of crap for a long time, [ENDQ] [NEWLINE] Ah, yes all their anguish over website trumps one person being compared to hitler on /r/all for days on end, sent death threats, and bombarded with tons of racism and sexist comments. I don't care what any grief they have after the shit that was put on Reddit after the FPH ban and last weekends Victoria firing. A very vocal minority have shown they can really be assholes. [NEWLINE] [NEWLINE] Assholes shouldn't even get the time of day after the stunts that were pulled like that. </s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do you judge the worth of a donation by the impact it had on the life of the donator, or on their reasoning? Surely the most logical way to judge the worth of a donation is<mask> much good it does,<mask> with that being the intention of the donation. [NEWLINE] [NEWLINE] By that metric, a billionaire donating 1% of their wealth is far better than a normal person,<mask> they do far more good in the world. Who cares<mask> impact it has on their life? Something making someone's life worse is a bad thing<mask> it makes them less happy,<mask> should we value it? </s>
Label encoding: <s>Why do you judge the worth of a donation by the impact it had on the life of the donator, or on their reasoning? Surely the most logical way to judge the worth of a donation is how much good it does, what with that being the intention of the donation. [NEWLINE] [NEWLINE] By that metric, a billionaire donating 1% of their wealth is far better than a normal person, because they do far more good in the world. Who cares what impact it has on their life? Something making someone's life worse is a bad thing as it makes them less happy, why should we value it? </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>is this a problem with first person shooters overall or just ones killing people?  from her reasonings<mask> can't you just get a game<mask> you kill robots, aliens or some other "non human, definitely evil enemy" <mask> that you can play an fps that isn't "killing people" [NEWLINE] [NEWLINE] <mask> you were arguing "i should be able to play call of duty and battlefield at 17" i could come up with some arguments to change your view (it is killing people, desensitizes and whatnot)<mask> just any fps? there are fps' made for toddlers to learn language.</s>
Label encoding: <s>is this a problem with first person shooters overall or just ones killing people?  from her reasonings why can't you just get a game where you kill robots, aliens or some other "non human, definitely evil enemy"  so that you can play an fps that isn't "killing people" [NEWLINE] [NEWLINE] if you were arguing "i should be able to play call of duty and battlefield at 17" i could come up with some arguments to change your view (it is killing people, desensitizes and whatnot) but just any fps? there are fps' made for toddlers to learn language.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Willing to question<mask> is claimed<mask> established fact is fine and an important part of skepticism. Doggedly persisting with those questions in the face of evidence,<mask>, is no longer skepticism and rapidly veers into cherry-picking and begging the question to support preconceived or politically motivated ideas. Case in point: citing James Randi, who is a magician, not a climatologist,<mask> someone who's ideas on AGW might be a valid challenge to the 97% of climate scientists who say otherwise. Or maintaining that a 97% consensus is achievable with a lack of "reliable evidence".</s>
Label encoding: <s>Willing to question what is claimed as established fact is fine and an important part of skepticism. Doggedly persisting with those questions in the face of evidence, however, is no longer skepticism and rapidly veers into cherry-picking and begging the question to support preconceived or politically motivated ideas. Case in point: citing James Randi, who is a magician, not a climatologist, as someone who's ideas on AGW might be a valid challenge to the 97% of climate scientists who say otherwise. Or maintaining that a 97% consensus is achievable with a lack of "reliable evidence".</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean eventually the universe will collapse in on itself. All energy gets used up by the ever expanding universe, and eventually it all collapses. Bringing an end to everything. Everything. [NEWLINE] [NEWLINE] Everything we do<mask> humans ultimately affects only our small ecosystem, our small speck of dust in space, Earth. [NEWLINE] [NEWLINE] <mask> even<mask> we save up every iota of information and personal data,<mask> the Earth gets destroyed, by us, an asteroid, the sun going supernova, whatever, it'll all disappear, along with everything we ever were. [NEWLINE] [NEWLINE] Kind of bleak I guess,<mask> inevitable. </s><pad>
Label encoding: <s>I mean eventually the universe will collapse in on itself. All energy gets used up by the ever expanding universe, and eventually it all collapses. Bringing an end to everything. Everything. [NEWLINE] [NEWLINE] Everything we do as humans ultimately affects only our small ecosystem, our small speck of dust in space, Earth. [NEWLINE] [NEWLINE] But even if we save up every iota of information and personal data, when the Earth gets destroyed, by us, an asteroid, the sun going supernova, whatever, it'll all disappear, along with everything we ever were. [NEWLINE] [NEWLINE] Kind of bleak I guess, but inevitable. </s><pad>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Going to a big party and getting really drunk is, objectively, a very dangerous thing to do. It opens you up to being a victim of a lot of crimes, not just rape,<mask><mask> theft and assault. [ENDQ] [NEWLINE] The amount of things stolen at house parties is ridiculous. I have seen someone walk out a house party with a snow board. [NEWLINE] [NEWLINE] A great piece of advice is that is you are having a large house party with any amount of strangers in your house you should secure you valuables.<mask><mask> you are the host, you shouldn't get black out drunk. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Going to a big party and getting really drunk is, objectively, a very dangerous thing to do. It opens you up to being a victim of a lot of crimes, not just rape, but also theft and assault. [ENDQ] [NEWLINE] The amount of things stolen at house parties is ridiculous. I have seen someone walk out a house party with a snow board. [NEWLINE] [NEWLINE] A great piece of advice is that is you are having a large house party with any amount of strangers in your house you should secure you valuables. Also if you are the host, you shouldn't get black out drunk. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Qing Dynasty China was very advanced in the 1700s,<mask> unlike most European nations it was relatively disinterested in expansion or foreign trade. <mask>, Europeans were *very* interested in doing business in China. <mask> the British East India Tea Company did a pretty good job of intentionally fucking up the Chinese economy via illegal opium dealing in order to secure better conditions for foreign trade.  This lead to war and some ridiculously uneven trade agreements.  The [Century of Humiliation]( [URL] ) caused by Western imperialism lead to the weakening and collapse of the Qing Dynasty. [NEWLINE] [NEWLINE] <mask> that's one.</s><pad>
Label encoding: <s>Qing Dynasty China was very advanced in the 1700s, but unlike most European nations it was relatively disinterested in expansion or foreign trade.  However, Europeans were *very* interested in doing business in China.  So the British East India Tea Company did a pretty good job of intentionally fucking up the Chinese economy via illegal opium dealing in order to secure better conditions for foreign trade.  This lead to war and some ridiculously uneven trade agreements.  The [Century of Humiliation]( [URL] ) caused by Western imperialism lead to the weakening and collapse of the Qing Dynasty. [NEWLINE] [NEWLINE] So that's one.</s><pad>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But you KNOW<mask> the circumstances are.  You KNOW the size of the seat.  You have the option to upgrade. [NEWLINE] [NEWLINE] <mask> you choose to buy a Ferrari should they be required to retrofit it free of charge<mask> you don't fit in it? <mask> you buy a bed that's too short, should they give you a larger mattress for free?  Should "extra long" pants which require extra material and stitching be required to be the same price<mask> shorter ones? [NEWLINE] [NEWLINE] <mask> can giving away for free something that people are willing to pay for not be bad for business? </s>
Label encoding: <s>But you KNOW what the circumstances are.  You KNOW the size of the seat.  You have the option to upgrade. [NEWLINE] [NEWLINE] If you choose to buy a Ferrari should they be required to retrofit it free of charge because you don't fit in it?  If you buy a bed that's too short, should they give you a larger mattress for free?  Should "extra long" pants which require extra material and stitching be required to be the same price as shorter ones? [NEWLINE] [NEWLINE] How can giving away for free something that people are willing to pay for not be bad for business? </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0645, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is quite unfortunate that philosophers and physicists don't communicate<mask> much<mask> they should; I've witnessed attempts at this first hand and they are not pretty (I'm a grad student of physics at a university with a strong philosphy department with philosophy of science experts). There are,<mask>, plenty of physicists interested in the philosophical implications of physical theories. Here's a blog post by one of the foremost theoretical physicists today, who specialises in quantum field theory and cosmology among other topics: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] He details an argument similar to mine, albeit fleshed out and far more convincing.</s><pad>
Label encoding: <s>It is quite unfortunate that philosophers and physicists don't communicate as much as they should; I've witnessed attempts at this first hand and they are not pretty (I'm a grad student of physics at a university with a strong philosphy department with philosophy of science experts). There are, however, plenty of physicists interested in the philosophical implications of physical theories. Here's a blog post by one of the foremost theoretical physicists today, who specialises in quantum field theory and cosmology among other topics: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] He details an argument similar to mine, albeit fleshed out and far more convincing.</s><pad>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0603, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I generally don't want to complain about food<mask> most mistakes are minor and it's not worth the trouble to the servers to fix minor mistakes.&gt; [ENDQ] [NEWLINE] I 100 BILLION PERCENT DISAGREE with you COMPLETELY!  Everyone should get<mask> *THEY* want for their *TIP* money, which means altering the server about issues, especially<mask> it's something that is their fault like they forgot your side dish or they forgot a side of ranch, etc. [NEWLINE] [NEWLINE] NO food issues are truly minor really in general.  They are ALL important.</s>
Label encoding: <s> [STARTQ] I generally don't want to complain about food because most mistakes are minor and it's not worth the trouble to the servers to fix minor mistakes.&gt; [ENDQ] [NEWLINE] I 100 BILLION PERCENT DISAGREE with you COMPLETELY!  Everyone should get what *THEY* want for their *TIP* money, which means altering the server about issues, especially if it's something that is their fault like they forgot your side dish or they forgot a side of ranch, etc. [NEWLINE] [NEWLINE] NO food issues are truly minor really in general.  They are ALL important.</s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So are you telling me that<mask><mask> Mac's have less viruses,<mask> their market share is less (profits are less) they are less effective in protecting consumers against viruses.<mask> compared to PC's that have many viruses developed for them<mask> have a large market share (profits are more) and are  more likely to protect users against viruses. [NEWLINE] [NEWLINE] <mask> this is<mask> you are telling me, then I guess you are saying it is easier to get a virus on a Mac then a PC. I would like some evidence to this claim,<mask> I've never heard this claim made before. </s>
Label encoding: <s>So are you telling me that even though Mac's have less viruses, since their market share is less (profits are less) they are less effective in protecting consumers against viruses. As compared to PC's that have many viruses developed for them but have a large market share (profits are more) and are  more likely to protect users against viruses. [NEWLINE] [NEWLINE] If this is what you are telling me, then I guess you are saying it is easier to get a virus on a Mac then a PC. I would like some evidence to this claim, because I've never heard this claim made before. </s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Couldn't see your op status cuz of my phone app lol. well, please help me understand<mask> my post doesn't relate, and help me understand<mask> personal preference is<mask> strong<mask> strong a reason to opt out<mask> accountability to a higher power. Keep in mind that exceptions for religious people are made under the assumptions that their religion could very well be true (<mask> unlikely), and that the truth of that religion would undoubtedly supercede the obligation of an employer to pay for every last form of birth control,  or the requirement of a religious pacifist to enter the draft etc.</s><pad>
Label encoding: <s>Couldn't see your op status cuz of my phone app lol. well, please help me understand how my post doesn't relate, and help me understand how personal preference is as strong as strong a reason to opt out as accountability to a higher power. Keep in mind that exceptions for religious people are made under the assumptions that their religion could very well be true ( however unlikely), and that the truth of that religion would undoubtedly supercede the obligation of an employer to pay for every last form of birth control,  or the requirement of a religious pacifist to enter the draft etc.</s><pad>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>to answer your questions, even<mask> milk productions slows or stops, or teeth come in, or you want to transition or supplement, consultants can help you with these minor problems. having seen my wife come upon and surpass all of these challenges with ease, it's not<mask> difficult<mask> our culture has led you to believe. [NEWLINE] [NEWLINE] full disclosure: i am only *almost* certain of all of these things<mask> of the consults we've had. i wouldn't put it past my wife to keep feeding until she was more than a year old,<mask><mask> we've already started weening her.</s>
Label encoding: <s>to answer your questions, even if milk productions slows or stops, or teeth come in, or you want to transition or supplement, consultants can help you with these minor problems. having seen my wife come upon and surpass all of these challenges with ease, it's not as difficult as our culture has led you to believe. [NEWLINE] [NEWLINE] full disclosure: i am only *almost* certain of all of these things because of the consults we've had. i wouldn't put it past my wife to keep feeding until she was more than a year old, even though we've already started weening her.</s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] False. Even the most strident of anti-gun researchers (in this case, David Hemenway) estimates<mask> many<mask> 80,000 defensive gun uses per year,<mask> some estimates run<mask> high<mask> 3.6 million, compared to ~30,000 gun deaths. [ENDQ] [NEWLINE] And<mask><mask> ALL the derailing below, to stay on message...  it's still an 8-3 ratio of Defensive Gun Use to suicides + homicides. [NEWLINE] [NEWLINE] (homicides which<mask> include justifiable homicides btw...  like the boston bomber who was shot and killed by police...)</s>
Label encoding: <s> [STARTQ] False. Even the most strident of anti-gun researchers (in this case, David Hemenway) estimates as many as 80,000 defensive gun uses per year, while some estimates run as high as 3.6 million, compared to ~30,000 gun deaths. [ENDQ] [NEWLINE] And regardless of ALL the derailing below, to stay on message...  it's still an 8-3 ratio of Defensive Gun Use to suicides + homicides. [NEWLINE] [NEWLINE] (homicides which also include justifiable homicides btw...  like the boston bomber who was shot and killed by police...)</s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why not the taxpayer? Both the father and the taxpayer were not responsible for the child's appearance (that was solely the responsibility of the woman,<mask> the last say determines responsibility),<mask> in this aspect they are equal.<mask> forcing the taxpayer to do it causes less harm (<mask> the burden of assisting the child is spread on more people than one).<mask> forcing the taxpayer to do it seems like the least bad choice. [NEWLINE] [NEWLINE] I would gladly pay slightly higher taxes<mask> it would mean that I would have the option of financial abortion in the case of unwanted pregnancy. Think of it<mask> insurance.</s>
Label encoding: <s>Why not the taxpayer? Both the father and the taxpayer were not responsible for the child's appearance (that was solely the responsibility of the woman, since the last say determines responsibility), so in this aspect they are equal. But forcing the taxpayer to do it causes less harm ( since the burden of assisting the child is spread on more people than one). So forcing the taxpayer to do it seems like the least bad choice. [NEWLINE] [NEWLINE] I would gladly pay slightly higher taxes if it would mean that I would have the option of financial abortion in the case of unwanted pregnancy. Think of it as insurance.</s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The violent rape you are talking about is seldom about sexual urges. It's more about power and dominance. People don't rape<mask> they lack sexual outlets, they do it for other reasons. [NEWLINE] [NEWLINE] <mask>, I wouldn't be surprised<mask> rape accusations increase.<mask> you are caught in an affair and the consequences are high enough, you might just claim rape to get out of it.<mask><mask>, that might even be optimal. Claim rape<mask> lacking any strong evidence and choose not to press charges. You get away with the affair in a legal sense and your partner never goes to court.</s>
Label encoding: <s>The violent rape you are talking about is seldom about sexual urges. It's more about power and dominance. People don't rape because they lack sexual outlets, they do it for other reasons. [NEWLINE] [NEWLINE] However, I wouldn't be surprised if rape accusations increase. If you are caught in an affair and the consequences are high enough, you might just claim rape to get out of it. In fact, that might even be optimal. Claim rape while lacking any strong evidence and choose not to press charges. You get away with the affair in a legal sense and your partner never goes to court.</s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0758, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The father is often compelled to pay child support (whether he wants it or not),<mask> the mother is very seldom compelled to grant the father parental rights, and even<mask> it happens, she can effectively stall and make paternal visits minimal/impossible, without significant legal repercussions. [ENDQ] [NEWLINE] <mask> that's the exact inequality being addressed by allowing fathers to forfeit their parental rights and responsibilities. The fact that the mother doesn't stand to gain anything from leveling the playing field only shows a flaw in the status quo. [NEWLINE] [NEWLINE] [STARTQ] Would be an extremely hard sell politically. [ENDQ] [NEWLINE] <mask><mask>.</s>
Label encoding: <s> [STARTQ] The father is often compelled to pay child support (whether he wants it or not), but the mother is very seldom compelled to grant the father parental rights, and even if it happens, she can effectively stall and make paternal visits minimal/impossible, without significant legal repercussions. [ENDQ] [NEWLINE] But that's the exact inequality being addressed by allowing fathers to forfeit their parental rights and responsibilities. The fact that the mother doesn't stand to gain anything from leveling the playing field only shows a flaw in the status quo. [NEWLINE] [NEWLINE] [STARTQ] Would be an extremely hard sell politically. [ENDQ] [NEWLINE] I agree.</s>
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> we wish to<mask><mask> means we are justified in killing 1+ foreign persons for every US person's life we save [ENDQ] [NEWLINE] I do not believe this, and do not argue for it.  Our moral high ground comes from the fact that the US does not target foreign civilians, that our military takes steps and precautions to avoid killing foreign civilians, and that we are (I optimistically hope) in the process of rethinking the role of the CIA.  The CIA's drones are reportedly not taking appropriate efforts to avoid killing foreign civilians, which should certainly merit pointed investigation.</s>
Label encoding: <s> [STARTQ] if we wish to argue that means we are justified in killing 1+ foreign persons for every US person's life we save [ENDQ] [NEWLINE] I do not believe this, and do not argue for it.  Our moral high ground comes from the fact that the US does not target foreign civilians, that our military takes steps and precautions to avoid killing foreign civilians, and that we are (I optimistically hope) in the process of rethinking the role of the CIA.  The CIA's drones are reportedly not taking appropriate efforts to avoid killing foreign civilians, which should certainly merit pointed investigation.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A slave does not get to choose<mask> job they have,<mask> hours they work,<mask> hard they work,<mask> they take breaks,<mask> they take breaks,<mask> they live, or with whom they live. Slaves may be traded, bought, and sold. I could definitely go on. [ENDQ] [NEWLINE] [NEWLINE] <mask> I was not talking about total slavery,<mask> only contractual temporary slavery, like for example parenthood, employment or child support / aliments (just<mask> we are not talking about total loss of bodily rights,<mask> only contractual and temporary loss of bodily rights - pregnancy).</s>
Label encoding: <s> [STARTQ] A slave does not get to choose what job they have, what hours they work, how hard they work, when they take breaks, if they take breaks, where they live, or with whom they live. Slaves may be traded, bought, and sold. I could definitely go on. [ENDQ] [NEWLINE] [NEWLINE] But I was not talking about total slavery, but only contractual temporary slavery, like for example parenthood, employment or child support / aliments (just as we are not talking about total loss of bodily rights, but only contractual and temporary loss of bodily rights - pregnancy).</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Don't tell me<mask> my argument is<mask> you misunderstand it. [NEWLINE] [NEWLINE] [STARTQ] You are saying that I should intentionally not tell my children the reasons I most strongly feel something is wrong,<mask> instead I should only tell them reasons that have been...<mask>?... pre-approved by some government committee on<mask> valid a-religious reasoning for all common childhood questions are? [ENDQ] [NEWLINE] You never answered my very first question. Are you ONLY not racist<mask> of your religion? Do you not see ANY valid reason to raise your child to not be racist other than religion? </s><pad><pad><pad><pad>
Label encoding: <s>Don't tell me what my argument is when you misunderstand it. [NEWLINE] [NEWLINE] [STARTQ] You are saying that I should intentionally not tell my children the reasons I most strongly feel something is wrong, but instead I should only tell them reasons that have been... what?... pre-approved by some government committee on what valid a-religious reasoning for all common childhood questions are? [ENDQ] [NEWLINE] You never answered my very first question. Are you ONLY not racist because of your religion? Do you not see ANY valid reason to raise your child to not be racist other than religion? </s><pad><pad><pad><pad>
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I wouldn't classify asking<mask><mask> disrespect. The child has a right to ask<mask> it is being told to do something,<mask><mask><mask> it does<mask> respectfully. That can even help; sometimes, the child is aware of something the parent isn't. [NEWLINE] [NEWLINE] <mask>, that respect comes before the question itself.<mask> the parent finally insists on having the child do something, it would be very disrespectful for the child to say "not until you explain it to me." A parent doesn't *owe* the child anything, everything that is given is a gift. That includes explanations.</s>
Label encoding: <s>I wouldn't classify asking why as disrespect. The child has a right to ask why it is being told to do something, so long as it does so respectfully. That can even help; sometimes, the child is aware of something the parent isn't. [NEWLINE] [NEWLINE] However, that respect comes before the question itself. If the parent finally insists on having the child do something, it would be very disrespectful for the child to say "not until you explain it to me." A parent doesn't *owe* the child anything, everything that is given is a gift. That includes explanations.</s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But time is not a comparable metric<mask> attempting to compare<mask> two people value. [NEWLINE] [NEWLINE] Here's a simple example: a person with a terminal illness who is likely to die in the next year has much less time to live than a 22 year old graduating college.  I would<mask><mask> the former's time is much more valuable<mask> he has less of it left.  In this regard, someone who has very little time and decides to devote it to working at a soup kitchen is much more "generous" than someone who has 1 billion dollars and gives 100.  </s>
Label encoding: <s>But time is not a comparable metric when attempting to compare what two people value. [NEWLINE] [NEWLINE] Here's a simple example: a person with a terminal illness who is likely to die in the next year has much less time to live than a 22 year old graduating college.  I would argue that the former's time is much more valuable since he has less of it left.  In this regard, someone who has very little time and decides to devote it to working at a soup kitchen is much more "generous" than someone who has 1 billion dollars and gives 100.  </s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe it was Socrates who said during his Defense Speech that no one should fear death<mask> no one alive can know anything about it, many people fear death simply<mask> they are ignorant of it. There is just<mask> strong of a possibility that death is even greater than life<mask><mask> no one can prove it everyone assumes the worst. He did not have a strong sense of religion like we do now a days,<mask> he is right in saying that fear of the unknown is an irrational fear<mask> in the end there might not be anything to have been afraid of in the first place.</s>
Label encoding: <s>I believe it was Socrates who said during his Defense Speech that no one should fear death because no one alive can know anything about it, many people fear death simply because they are ignorant of it. There is just as strong of a possibility that death is even greater than life but since no one can prove it everyone assumes the worst. He did not have a strong sense of religion like we do now a days, but he is right in saying that fear of the unknown is an irrational fear because in the end there might not be anything to have been afraid of in the first place.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, have you looked at the [wikipedia page]( [URL] ) for heteronormativity? [NEWLINE] [NEWLINE] "Normativity" is an evaluative concept and should not be confused with typical/atypical or common/uncommon, which are purely descriptive. "Heteronormative" behavior implies the view that heterosexual behavior is the only or the "correct" norm to follow, and the equivalence of sex and gender. In other words to claim that someone is behaving in a heteronormative or otherwise way is to claim something about their beliefs about sexuality.</s>
Label encoding: <s>Well, have you looked at the [wikipedia page]( [URL] ) for heteronormativity? [NEWLINE] [NEWLINE] "Normativity" is an evaluative concept and should not be confused with typical/atypical or common/uncommon, which are purely descriptive. "Heteronormative" behavior implies the view that heterosexual behavior is the only or the "correct" norm to follow, and the equivalence of sex and gender. In other words to claim that someone is behaving in a heteronormative or otherwise way is to claim something about their beliefs about sexuality.</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You make a few faulty assumptions. [NEWLINE] [NEWLINE] 1. That the extra time would be used for in-class instruction rather than being used for physical activity or "homework" time [NEWLINE] [NEWLINE] 2.  That the "main" teacher would need to be involved in the above activities. [NEWLINE] [NEWLINE] Imagine kids coming in the morning and engaging in organized physical activity for a few hours followed by breakfast.  Teachers could easily lesson plan during this time (no longer would they need to stay after school) AND kids would be far more alert<mask> in class after engaging in such activities. [NEWLINE] </s>
Label encoding: <s>You make a few faulty assumptions. [NEWLINE] [NEWLINE] 1. That the extra time would be used for in-class instruction rather than being used for physical activity or "homework" time [NEWLINE] [NEWLINE] 2.  That the "main" teacher would need to be involved in the above activities. [NEWLINE] [NEWLINE] Imagine kids coming in the morning and engaging in organized physical activity for a few hours followed by breakfast.  Teachers could easily lesson plan during this time (no longer would they need to stay after school) AND kids would be far more alert while in class after engaging in such activities. [NEWLINE] </s>
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It was a real wake-up call for me<mask> I realized that the people most commonly telling me "<mask> you look fine!" were roughly the same size I was. Telling me that I didn't need to go to the gym or watch<mask> I ate was them telling *themselves* that they didn't need to. One of my coworkers became openly hostile toward my efforts, "joking" frequently that I was going to get ugly and mannish<mask> I kept lifting. It was frustrating and bizarre to see just<mask> much she disdained my self improvement.</s>
Label encoding: <s>It was a real wake-up call for me when I realized that the people most commonly telling me " But you look fine!" were roughly the same size I was. Telling me that I didn't need to go to the gym or watch what I ate was them telling *themselves* that they didn't need to. One of my coworkers became openly hostile toward my efforts, "joking" frequently that I was going to get ugly and mannish if I kept lifting. It was frustrating and bizarre to see just how much she disdained my self improvement.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I thought the opposition came from employers being required to offer a plan that covered it.<mask>, for example, Catholic hospitals who employed women are being forced to have options available to women who want BCP covered,<mask> the women themselves are not required to buy that plan. [NEWLINE] [NEWLINE] [STARTQ] and<mask> there is any immorality in this equation, it is using force to induce purchases that people do not want. [ENDQ] [NEWLINE] <mask> we could use that to argue against anything then. ED meds, health care plans that aren't custom-made to your needs, car insurance...</s>
Label encoding: <s>I thought the opposition came from employers being required to offer a plan that covered it. So, for example, Catholic hospitals who employed women are being forced to have options available to women who want BCP covered, but the women themselves are not required to buy that plan. [NEWLINE] [NEWLINE] [STARTQ] and if there is any immorality in this equation, it is using force to induce purchases that people do not want. [ENDQ] [NEWLINE] But we could use that to argue against anything then. ED meds, health care plans that aren't custom-made to your needs, car insurance...</s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In a platonic relationship you're thinking about yourself first and<mask> the other person compliments it. In a purely sexual relationship you're thinking about your own needs and<mask> the other person serves them. You can combine these two into a fuck buddy relationship - and it would still be a selfish experience.  A romantic relationship is giving yourself over to someone,  and they to you. Often you'll have to do things you wouldn't do for a platonic friend or a fuck buddy,  just for the good of your relationship.  That's the difference<mask><mask>. </s>
Label encoding: <s>In a platonic relationship you're thinking about yourself first and how the other person compliments it. In a purely sexual relationship you're thinking about your own needs and how the other person serves them. You can combine these two into a fuck buddy relationship - and it would still be a selfish experience.  A romantic relationship is giving yourself over to someone,  and they to you. Often you'll have to do things you wouldn't do for a platonic friend or a fuck buddy,  just for the good of your relationship.  That's the difference IMO. </s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not saying your opinion is not worthy of expression, simply that it possesses an unrealistic set of criteria which are not constructive at all. I do dare you to write/find me a story which meets your standards<mask> I applied them. It would be impossible<mask> 1) anything could set me off, and 2) I could find some other way for the story to express<mask> happened in a less 'triggering' way. Your standards have a right to be expressed,<mask> they don't have a right to be taken seriously or be considered logical<mask> they aren't.</s>
Label encoding: <s>I'm not saying your opinion is not worthy of expression, simply that it possesses an unrealistic set of criteria which are not constructive at all. I do dare you to write/find me a story which meets your standards if I applied them. It would be impossible because 1) anything could set me off, and 2) I could find some other way for the story to express what happened in a less 'triggering' way. Your standards have a right to be expressed, but they don't have a right to be taken seriously or be considered logical when they aren't.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>IMO, This is the inevitable lifestyle of a product like this. [NEWLINE] [NEWLINE] <mask> others mentioned, chrome pioneered the current look and feel of browser's that everyone else is emulating. [NEWLINE] [NEWLINE] You want chrome to innovate again,<mask> doing<mask> would likely result in a loss of share, or at best no improvement (<mask> a huge cost to develop).<mask>? Current Chrome users like it<mask> it is now. Those who want something different will slowly migrate to other browsers. [NEWLINE] [NEWLINE] Then<mask> Chrome's user base begins to shrink THEN they innovate again and wow everyone back.</s>
Label encoding: <s>IMO, This is the inevitable lifestyle of a product like this. [NEWLINE] [NEWLINE] As others mentioned, chrome pioneered the current look and feel of browser's that everyone else is emulating. [NEWLINE] [NEWLINE] You want chrome to innovate again, but doing so would likely result in a loss of share, or at best no improvement ( but a huge cost to develop). Why? Current Chrome users like it how it is now. Those who want something different will slowly migrate to other browsers. [NEWLINE] [NEWLINE] Then when Chrome's user base begins to shrink THEN they innovate again and wow everyone back.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> the electoral college system sucks balls. [ENDQ] [NEWLINE] We aren't voting on a president today,<mask> I'm not sure<mask> electoral college matters much.<mask><mask>, your local election likely involves a lot of races that will effect you in a more direct way than you think. My town is electing a commissioner, a sheriff and filling two council seats. We are<mask> deciding whether to borrow to build a recreation center and raise sales tax by two cents over the next twenty years. A lot of these decisions are likely going to come down to a few votes. </s>
Label encoding: <s> [STARTQ] I think the electoral college system sucks balls. [ENDQ] [NEWLINE] We aren't voting on a president today, so I'm not sure why electoral college matters much. In fact, your local election likely involves a lot of races that will effect you in a more direct way than you think. My town is electing a commissioner, a sheriff and filling two council seats. We are also deciding whether to borrow to build a recreation center and raise sales tax by two cents over the next twenty years. A lot of these decisions are likely going to come down to a few votes. </s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You might want to clarify this in your posts. For an avid comic reader, the concept of a "comic book" includes much more than the mainstream Marvel/DC offerings you seem to be mainly thinking about. [ENDQ] [NEWLINE] Thank you for pointing this out.  I honestly didn't know to do that.  I'm mostly trying to sort out my own misgivings vis-a-vis my colleague.  I admit that most of<mask> I know about the genre comes from him, and I<mask> I overhear him talking about to his friends.</s>
Label encoding: <s> [STARTQ] You might want to clarify this in your posts. For an avid comic reader, the concept of a "comic book" includes much more than the mainstream Marvel/DC offerings you seem to be mainly thinking about. [ENDQ] [NEWLINE] Thank you for pointing this out.  I honestly didn't know to do that.  I'm mostly trying to sort out my own misgivings vis-a-vis my colleague.  I admit that most of what I know about the genre comes from him, and I what I overhear him talking about to his friends.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It seems that having friends and "doing life" is something reserved for the young...then people have families/get hitched and loose interest to go out. [ENDQ] [NEWLINE] <mask> this is true that means you're in your prime right now to "do life".<mask> would you waste that?<mask> you've stated, there's plenty of time for being a hermit<mask> you're older. Right now you're in college, there is no easier place to make connections that last for a lifetime. You might not get a chance to make those connections later in life.</s>
Label encoding: <s> [STARTQ] It seems that having friends and "doing life" is something reserved for the young...then people have families/get hitched and loose interest to go out. [ENDQ] [NEWLINE] If this is true that means you're in your prime right now to "do life". Why would you waste that? As you've stated, there's plenty of time for being a hermit when you're older. Right now you're in college, there is no easier place to make connections that last for a lifetime. You might not get a chance to make those connections later in life.</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Furthermore,<mask> you're a 13-year-old girl, you have no reason to believe that your role in sex is anything more than ensuring your partner achieves a climax. [ENDQ] [NEWLINE] Except that 13 year olds get different views of sex from every part of our media driven culture. Music videos, sitcoms, reality tv shows etc.. Pornography isn't the sole source that impressionable people get their ideas about sex from. [NEWLINE] Watch a romantic comedy and a kid will get an entirely alternate view of<mask> sex is, can be, and should be. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Furthermore, if you're a 13-year-old girl, you have no reason to believe that your role in sex is anything more than ensuring your partner achieves a climax. [ENDQ] [NEWLINE] Except that 13 year olds get different views of sex from every part of our media driven culture. Music videos, sitcoms, reality tv shows etc.. Pornography isn't the sole source that impressionable people get their ideas about sex from. [NEWLINE] Watch a romantic comedy and a kid will get an entirely alternate view of what sex is, can be, and should be. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No they don't. People often respect traditions, even<mask> they shouldn't. [NEWLINE] [NEWLINE] <mask><mask> you're<mask> ignoring two important differences between religion and any other diet. The first is implication:<mask> god exists that has a way bigger implication than<mask> gluten is bad for you. It<mask> warrants a higher change to your lifestyle at a lower probability. The second is probability that the belief is true.<mask> we assume both beliefs are false, we can at least agree on the fact that more people believe in God, which in turn means tha their error is more understandable.</s>
Label encoding: <s>No they don't. People often respect traditions, even if they shouldn't. [NEWLINE] [NEWLINE] I think you're also ignoring two important differences between religion and any other diet. The first is implication: If god exists that has a way bigger implication than if gluten is bad for you. It therefore warrants a higher change to your lifestyle at a lower probability. The second is probability that the belief is true. If we assume both beliefs are false, we can at least agree on the fact that more people believe in God, which in turn means tha their error is more understandable.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Morality at its most basic is<mask> is necessary to resolve conflicting interests among moral agents. [NEWLINE] [NEWLINE] * Every moral agent has interests [NEWLINE] * Whatever one's interests are, there are situations<mask> you will be in conflict with others to fulfill your own interests. This means that it's impossible to fulfill every interest of everyone. [NEWLINE] * For everyone in a group to get the maximum of their interests fulfilled, it's in their best interest to cooperate in some way. [NEWLINE] * There are multiple frameworks that ensure cooperation,<mask> I'm not going to argue a specific one.</s>
Label encoding: <s>Morality at its most basic is what is necessary to resolve conflicting interests among moral agents. [NEWLINE] [NEWLINE] * Every moral agent has interests [NEWLINE] * Whatever one's interests are, there are situations where you will be in conflict with others to fulfill your own interests. This means that it's impossible to fulfill every interest of everyone. [NEWLINE] * For everyone in a group to get the maximum of their interests fulfilled, it's in their best interest to cooperate in some way. [NEWLINE] * There are multiple frameworks that ensure cooperation, so I'm not going to argue a specific one.</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In my view<mask> all sides have equivalent or equal requirements or conflicting views or<mask> have you, that's fine.<mask> the moment it becomes war, either one party or all were unjust. Justice would settle without war. [NEWLINE] [NEWLINE] It is<mask> possible<mask> that one side is unjust and oppresses, and the only recourse to freedom and justice is violence.<mask>, in order for violence to be justified by a party, the other needs to be unjust.<mask> all parties are just, there can be no war. [NEWLINE] [NEWLINE] That's *my* view anyway.</s>
Label encoding: <s>In my view if all sides have equivalent or equal requirements or conflicting views or what have you, that's fine. But the moment it becomes war, either one party or all were unjust. Justice would settle without war. [NEWLINE] [NEWLINE] It is however possible though that one side is unjust and oppresses, and the only recourse to freedom and justice is violence. So, in order for violence to be justified by a party, the other needs to be unjust. If all parties are just, there can be no war. [NEWLINE] [NEWLINE] That's *my* view anyway.</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Reply to Paragraph 1: We could set different policies for birth control pills being sought for medical purposes and birth control pills being sought for solely contraceptive purposes. [NEWLINE] [NEWLINE] [STARTQ] Making birth control<mask> easy to get and ideally free would [be good.] [ENDQ] [NEWLINE] Yes,<mask>, again, having insurance companies cover birth control pills might have the opposite effect. I probably should have made this a more central theme of my OP.<mask><mask> its a fallacy to think "covered by health insurance = more widely available." That's certainly not the case with check-ups...</s>
Label encoding: <s>Reply to Paragraph 1: We could set different policies for birth control pills being sought for medical purposes and birth control pills being sought for solely contraceptive purposes. [NEWLINE] [NEWLINE] [STARTQ] Making birth control as easy to get and ideally free would [be good.] [ENDQ] [NEWLINE] Yes, but, again, having insurance companies cover birth control pills might have the opposite effect. I probably should have made this a more central theme of my OP. I think its a fallacy to think "covered by health insurance = more widely available." That's certainly not the case with check-ups...</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] a system that is actually fit for a scientific or engineering purpose [ENDQ] [NEWLINE] <mask><mask> the imperial system is fit for these purposes. I am a scientist,<mask> that helps make<mask><mask> count for something. [NEWLINE] [NEWLINE] <mask><mask> you're right,<mask> the reason you're right is not<mask> metric is objectively better,<mask> rather<mask> metric is dominant.<mask> imperial were dominant, the argument that learning a second system adds cost and the possibility for error would apply to metric or any other system,<mask> that argument doesn't really help show that imperial is objectively better.</s>
Label encoding: <s> [STARTQ] a system that is actually fit for a scientific or engineering purpose [ENDQ] [NEWLINE] I think the imperial system is fit for these purposes. I am a scientist, if that helps make my opinion count for something. [NEWLINE] [NEWLINE] I think you're right, but the reason you're right is not because metric is objectively better, but rather because metric is dominant. If imperial were dominant, the argument that learning a second system adds cost and the possibility for error would apply to metric or any other system, so that argument doesn't really help show that imperial is objectively better.</s>
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The problem I have with the current system is that wealthier families send their kids to private schools and many poorer families are forced to send their kids to public schools. [ENDQ] [NEWLINE] <mask><mask> this is a good thing. Public schools have very limited funds and resources and should not have to expend them on students who would rather attend a private institution.<mask> wealthier students attend private schools, then the quality of public school education rises significantly. It lowers the student to teacher ratio, lowers budgetary strain, and allows the school to develop strategies specifically tailored to less affluent students. </s>
Label encoding: <s> [STARTQ] The problem I have with the current system is that wealthier families send their kids to private schools and many poorer families are forced to send their kids to public schools. [ENDQ] [NEWLINE] I think this is a good thing. Public schools have very limited funds and resources and should not have to expend them on students who would rather attend a private institution. If wealthier students attend private schools, then the quality of public school education rises significantly. It lowers the student to teacher ratio, lowers budgetary strain, and allows the school to develop strategies specifically tailored to less affluent students. </s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know about Croatia,<mask> in the UK candidates have to pay a deposit to stand, which they lose<mask> they don't achieve a certain percentage of the vote. This is one reason that I feel even a vote for a loser has some positive effect<mask> they don't lose their deposit. [NEWLINE] [NEWLINE] <mask> really, people over the world live without democracy. People have died for the chance to vote and<mask> the system here in the UK isn't perfect, it could be far worse. You should support your right to vote by using it. </s>
Label encoding: <s>I don't know about Croatia, but in the UK candidates have to pay a deposit to stand, which they lose if they don't achieve a certain percentage of the vote. This is one reason that I feel even a vote for a loser has some positive effect if they don't lose their deposit. [NEWLINE] [NEWLINE] But really, people over the world live without democracy. People have died for the chance to vote and although the system here in the UK isn't perfect, it could be far worse. You should support your right to vote by using it. </s>
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Advertising is good. Advertising is powerful. The freedom to advertise medicine is the freedom to tell people<mask> works. That's<mask> medicine is. Treatments for diseases that work more times than they don't, and with known and accepted side effects. Baring the advertisement of prescription drugs without<mask> banning the advertisement of natural and alternative health supplements places science at a disadvantage in the market by their inability to communicate with consumers. Rather than restricting companies ability to advertise we should tax and educate consumers to make good choices and think critically about their health care options. </s>
Label encoding: <s>Advertising is good. Advertising is powerful. The freedom to advertise medicine is the freedom to tell people what works. That's what medicine is. Treatments for diseases that work more times than they don't, and with known and accepted side effects. Baring the advertisement of prescription drugs without also banning the advertisement of natural and alternative health supplements places science at a disadvantage in the market by their inability to communicate with consumers. Rather than restricting companies ability to advertise we should tax and educate consumers to make good choices and think critically about their health care options. </s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Now, unless you are planning for some type of zombie apocalypse or nuclear war [ENDQ] [NEWLINE] Nop. [NEWLINE] [NEWLINE] More like events that result in destruction of stock market. [NEWLINE] [NEWLINE] Wars, revolutions, communists governments nationalizing everything. [NEWLINE] [NEWLINE] Think Germany closing down Polish stock market in 1939. Or Russian Civil war. [NEWLINE] [NEWLINE] [STARTQ] Unallocated Storage [ENDQ] [NEWLINE] [STARTQ] Digital IOU [ENDQ] [NEWLINE] [STARTQ] Gold ETF [ENDQ] [NEWLINE] Useless for the type of event I am talking about. I might still buy these things<mask> only<mask><mask><mask> to physical stash.</s><pad><pad><pad>
Label encoding: <s> [STARTQ] Now, unless you are planning for some type of zombie apocalypse or nuclear war [ENDQ] [NEWLINE] Nop. [NEWLINE] [NEWLINE] More like events that result in destruction of stock market. [NEWLINE] [NEWLINE] Wars, revolutions, communists governments nationalizing everything. [NEWLINE] [NEWLINE] Think Germany closing down Polish stock market in 1939. Or Russian Civil war. [NEWLINE] [NEWLINE] [STARTQ] Unallocated Storage [ENDQ] [NEWLINE] [STARTQ] Digital IOU [ENDQ] [NEWLINE] [STARTQ] Gold ETF [ENDQ] [NEWLINE] Useless for the type of event I am talking about. I might still buy these things but only in ADDITION to physical stash.</s><pad><pad><pad>
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You seem to be forgetting that it's not just North vs. South Korea. It's<mask> US/NATO vs. China. I *really* don't think you want to deal with a war between China and the US. And on top of the fact, North Korea is still a nuclear power, and<mask> they would probably never use their nukes for anything other than political leverage normally, directly invading them would absolutely cause them to use their nukes, which<mask> short range, could still hit some major areas in Korea or even Japan.</s>
Label encoding: <s>You seem to be forgetting that it's not just North vs. South Korea. It's also US/NATO vs. China. I *really* don't think you want to deal with a war between China and the US. And on top of the fact, North Korea is still a nuclear power, and although they would probably never use their nukes for anything other than political leverage normally, directly invading them would absolutely cause them to use their nukes, which although short range, could still hit some major areas in Korea or even Japan.</s>
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We don't weigh other people's misfortunes and compare them to ours and decide which one is greater and go on the path of least misfortune in our society!<mask> someone is speeding<mask> they are en route to the hospital to deliver a baby, they are still breaking the law,<mask><mask> their reasoning. They should get a ticket, whether or not their getting to the hospital is more or less important than other drivers not having the safest driving experience<mask> of them. [NEWLINE] We are not responsible for other people's misfortunes.</s>
Label encoding: <s>We don't weigh other people's misfortunes and compare them to ours and decide which one is greater and go on the path of least misfortune in our society! If someone is speeding because they are en route to the hospital to deliver a baby, they are still breaking the law, regardless of their reasoning. They should get a ticket, whether or not their getting to the hospital is more or less important than other drivers not having the safest driving experience because of them. [NEWLINE] We are not responsible for other people's misfortunes.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me show you<mask> that argument doesn't follow: [NEWLINE] [NEWLINE] It's certainly true that genes determine (say)<mask> many arms you have. And<mask> many arms humans have has massively shaped all sorts of everyday objects around you. [NEWLINE] [NEWLINE] <mask> it doesn't follow from that that you could alter the shape of, say, a keyboard by breeding humans to have more arms. Even<mask> you could breed humans to have more arms (which you can't) selective breeding certainly wouldn't be the best way to go about altering<mask> keyboards are shaped.</s>
Label encoding: <s>Let me show you why that argument doesn't follow: [NEWLINE] [NEWLINE] It's certainly true that genes determine (say) how many arms you have. And how many arms humans have has massively shaped all sorts of everyday objects around you. [NEWLINE] [NEWLINE] But it doesn't follow from that that you could alter the shape of, say, a keyboard by breeding humans to have more arms. Even if you could breed humans to have more arms (which you can't) selective breeding certainly wouldn't be the best way to go about altering how keyboards are shaped.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't think cat owners have the RIGHT to subject an animal that depends on them to danger. [ENDQ] [NEWLINE] There's always danger. The question is to<mask> extent. The danger to outdoor cats is low enough<mask> to be a worthwhile risk/benefit. [NEWLINE] [NEWLINE] Cats do not depend on humans to the extent you seem to think. Cats are not domesticated animals, they evolved to live outdoors, alongside humans in human settlements. That's their natural habitat. They're not helpless domestic animals who aren't adapted to being let outside.</s><pad><pad>
Label encoding: <s> [STARTQ] I don't think cat owners have the RIGHT to subject an animal that depends on them to danger. [ENDQ] [NEWLINE] There's always danger. The question is to what extent. The danger to outdoor cats is low enough as to be a worthwhile risk/benefit. [NEWLINE] [NEWLINE] Cats do not depend on humans to the extent you seem to think. Cats are not domesticated animals, they evolved to live outdoors, alongside humans in human settlements. That's their natural habitat. They're not helpless domestic animals who aren't adapted to being let outside.</s><pad><pad>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ugh, these social uses of the words "man" and "male" are derailing the conversation. [ENDQ] [NEWLINE] <mask> they lack any coherent, socially-sanctioned meaning?  :) [NEWLINE] [NEWLINE] <mask> ok, I see your point: [NEWLINE] [NEWLINE] [STARTQ] A man is a person who identifies<mask> a person with a penis,<mask><mask> his actual biology. [ENDQ] [NEWLINE] PC would say this is impermissibly binarist and reductive.  You can have no desire to get rid of your penis and still be considered a woman. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Ugh, these social uses of the words "man" and "male" are derailing the conversation. [ENDQ] [NEWLINE] Because they lack any coherent, socially-sanctioned meaning?  :) [NEWLINE] [NEWLINE] But ok, I see your point: [NEWLINE] [NEWLINE] [STARTQ] A man is a person who identifies as a person with a penis, regardless of his actual biology. [ENDQ] [NEWLINE] PC would say this is impermissibly binarist and reductive.  You can have no desire to get rid of your penis and still be considered a woman. [NEWLINE] </s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not my rule<mask> OP's rule,<mask> I'll argue for it. [NEWLINE] [NEWLINE] In your hypothetical situation<mask> his tattoos are offensive (e.g. Nazi swastika), the same thing applies. Wearing long-sleeved shirts is still a mere "discomfort". No one's forcing him to destroy his pin-up-shirt or remove his hypothetically offensive tattoos. He's free to show them off in private or even in public<mask> not in public<mask> representing both the best of an industry and the government.</s>
Label encoding: <s>It's not my rule but OP's rule, but I'll argue for it. [NEWLINE] [NEWLINE] In your hypothetical situation where his tattoos are offensive (e.g. Nazi swastika), the same thing applies. Wearing long-sleeved shirts is still a mere "discomfort". No one's forcing him to destroy his pin-up-shirt or remove his hypothetically offensive tattoos. He's free to show them off in private or even in public but not in public while representing both the best of an industry and the government.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That logic could be applied to anything expensive. Buying cars. Vacations. Homes. Going out to eat on a regular basis. Adopting a dog. [NEWLINE] [NEWLINE] Fiscally responsible adults budget appropriately and have enough left over that they can handle unexpected financial issues. [NEWLINE] [NEWLINE] <mask> said people want to have a fancy party<mask> they get married<mask> it's *fun* and you don't get a lot of excuses to do stuff like that, I don't know<mask> the purpose of shitting on that party would be. </s>
Label encoding: <s>That logic could be applied to anything expensive. Buying cars. Vacations. Homes. Going out to eat on a regular basis. Adopting a dog. [NEWLINE] [NEWLINE] Fiscally responsible adults budget appropriately and have enough left over that they can handle unexpected financial issues. [NEWLINE] [NEWLINE] If said people want to have a fancy party when they get married because it's *fun* and you don't get a lot of excuses to do stuff like that, I don't know what the purpose of shitting on that party would be. </s>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This needs more upvotes. [NEWLINE] [NEWLINE] <mask> you are in the lower 80% of males, then you will think women don't really "need" sex, and that their sex drives are lower. [NEWLINE] [NEWLINE] <mask><mask><mask>...<mask> you are in the top 20% of males... then you will think women are<mask> horny/sexual<mask> men,<mask> they are constantly propositioning you and making passes at you. [NEWLINE] [NEWLINE] It's really all about whether you see the sexual side of women or not that forms your view on their sexuality.</s>
Label encoding: <s>This needs more upvotes. [NEWLINE] [NEWLINE] If you are in the lower 80% of males, then you will think women don't really "need" sex, and that their sex drives are lower. [NEWLINE] [NEWLINE] HOWEVER... if you are in the top 20% of males... then you will think women are as horny/sexual as men, because they are constantly propositioning you and making passes at you. [NEWLINE] [NEWLINE] It's really all about whether you see the sexual side of women or not that forms your view on their sexuality.</s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The only arguments I’ve read for heterosexuality being default is the biological urge to have children, which I believe is neither universal nor exclusive to heterosexuality and the (<mask><mask> fallacious) “most people are straight<mask> there must be a biological mechanism that supports this phenomenon” path of reasoning. [ENDQ] [NEWLINE] <mask> do you view these arguments<mask> fallacious? [NEWLINE] [NEWLINE] Occam's Razor would suggest that a statistically significant portion of a population holding an orientation which pairs with a species-wide biological imperative would be "normal".</s>
Label encoding: <s> [STARTQ] The only arguments I’ve read for heterosexuality being default is the biological urge to have children, which I believe is neither universal nor exclusive to heterosexuality and the ( imo fallacious) “most people are straight so there must be a biological mechanism that supports this phenomenon” path of reasoning. [ENDQ] [NEWLINE] Why do you view these arguments as fallacious? [NEWLINE] [NEWLINE] Occam's Razor would suggest that a statistically significant portion of a population holding an orientation which pairs with a species-wide biological imperative would be "normal".</s>
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would really only agree with your point<mask> people socialised exclusively through gay only events. A lot of the time these events are used to help create networks for gay positive social or political actions, and sometimes their used for parties<mask> who wants to talk politics all the time? [NEWLINE] [NEWLINE] Segregation isn't the best way to bring about social change,<mask> it is a good start in finding out<mask> other people are bringing about social change in their own communities and<mask> challenges they face before being able to deal with it head on. </s>
Label encoding: <s>I would really only agree with your point if people socialised exclusively through gay only events. A lot of the time these events are used to help create networks for gay positive social or political actions, and sometimes their used for parties because who wants to talk politics all the time? [NEWLINE] [NEWLINE] Segregation isn't the best way to bring about social change, but it is a good start in finding out how other people are bringing about social change in their own communities and what challenges they face before being able to deal with it head on. </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I<mask> share the same view<mask> you. <mask> reading through these comments make me think that you and i are not<mask>'susceptible'to peer pressure, addiction, or whatever its called. Ive smoked cigarettes before, and i got cravjngs weeks after<mask> i stopped.<mask> i didnt feed that craving, i ignored it. Then again I'm<mask> subscribed to fph and honestly have no sympathy for fat people just<mask> i have no sympathy for people that get off drugs or alcohol. </s>
Label encoding: <s>I also share the same view as you.  But reading through these comments make me think that you and i are not as'susceptible'to peer pressure, addiction, or whatever its called. Ive smoked cigarettes before, and i got cravjngs weeks after when i stopped. But i didnt feed that craving, i ignored it. Then again I'm also subscribed to fph and honestly have no sympathy for fat people just how i have no sympathy for people that get off drugs or alcohol. </s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> you have never seen a teaching survey that questions comprehension doesn't mean it's not being considered. The administration is well aware of<mask> each professor sounds like - they did interviews, and they sit in meetings and such. They are not ignorant of the handicap of an accent, and have taken that into consideration at the time of hiring the professor. [NEWLINE] [NEWLINE] <mask>, there is not a limitless supply of experts worthy of being professors. Get rid of all the professors with accents, and who will you replace them with?</s>
Label encoding: <s>Just because you have never seen a teaching survey that questions comprehension doesn't mean it's not being considered. The administration is well aware of what each professor sounds like - they did interviews, and they sit in meetings and such. They are not ignorant of the handicap of an accent, and have taken that into consideration at the time of hiring the professor. [NEWLINE] [NEWLINE] Also, there is not a limitless supply of experts worthy of being professors. Get rid of all the professors with accents, and who will you replace them with?</s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you give tacit consent any time you claim to use anything granted by the collective will [ENDQ] [NEWLINE] This is the part that he disagrees with, and you're merely claiming it's self-evident, which it most certainly is not. [NEWLINE] [NEWLINE] [STARTQ] <mask> you decide to exercise your right to private property, which solely exist by the will of the collective, you are giving tacit consent. [ENDQ] [NEWLINE] There is no "collective" entity, only individuals.  At no magical point does a group of people stop being individuals.</s>
Label encoding: <s> [STARTQ] you give tacit consent any time you claim to use anything granted by the collective will [ENDQ] [NEWLINE] This is the part that he disagrees with, and you're merely claiming it's self-evident, which it most certainly is not. [NEWLINE] [NEWLINE] [STARTQ] When you decide to exercise your right to private property, which solely exist by the will of the collective, you are giving tacit consent. [ENDQ] [NEWLINE] There is no "collective" entity, only individuals.  At no magical point does a group of people stop being individuals.</s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The best way to make the process faster is to not have carry on bags that go into the overhead bins. Then to speed it up more you just have people line up via seat number from back of plane to front with window seats going before seats in the middle or aisle. [NEWLINE] [NEWLINE] <mask> soon<mask> you're dealing with people trying to play overhead tetris you're going to be boarding quite slowly no matter<mask> people are lined up. And<mask> you want to make sure there's space for your luggage you line up sooner.</s>
Label encoding: <s>The best way to make the process faster is to not have carry on bags that go into the overhead bins. Then to speed it up more you just have people line up via seat number from back of plane to front with window seats going before seats in the middle or aisle. [NEWLINE] [NEWLINE] As soon as you're dealing with people trying to play overhead tetris you're going to be boarding quite slowly no matter how people are lined up. And since you want to make sure there's space for your luggage you line up sooner.</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If women don't want to be in STEM,<mask> are you trying to shoehorn 50/50 splits? That's complete bullshit. Again, in the most egalitarian countries (the nordic states), women are REJECTING science/engineering.  Basically, you're saying, you don't care about women's actual preferences in an egalitarian society, you want them to do jobs they are unhappy doing.  Is it any surprise that after all the gains feminism has made, women are more unhappy than ever? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>If women don't want to be in STEM, why are you trying to shoehorn 50/50 splits? That's complete bullshit. Again, in the most egalitarian countries (the nordic states), women are REJECTING science/engineering.  Basically, you're saying, you don't care about women's actual preferences in an egalitarian society, you want them to do jobs they are unhappy doing.  Is it any surprise that after all the gains feminism has made, women are more unhappy than ever? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In terms of<mask> things "should" be,<mask><mask> it's primarily a matter of evidence rather than the degree of harm. Compared to emotional harm, physical violence leaves behind unambiguous evidence and tends to occur in distinct and individually significant events. It's<mask> easier to draw consistent lines between levels of physical violence than it is with emotional harm. [NEWLINE] [NEWLINE] That said,<mask><mask> emotional harm is not given enough weight in a moral sense, even<mask> it is not practical to consider it more heavily in a legal sense.</s>
Label encoding: <s>In terms of how things "should" be, IMO it's primarily a matter of evidence rather than the degree of harm. Compared to emotional harm, physical violence leaves behind unambiguous evidence and tends to occur in distinct and individually significant events. It's also easier to draw consistent lines between levels of physical violence than it is with emotional harm. [NEWLINE] [NEWLINE] That said, I think emotional harm is not given enough weight in a moral sense, even if it is not practical to consider it more heavily in a legal sense.</s>
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Permanent unemployment is going to become more and more of a thing. Automation is going to happen. There simply are going to be less and less jobs, and we need to figure out a way to deal with that, and letting people starve isn't a good plan. [ENDQ] [NEWLINE] I keep hearing rumblings of this,<mask> hasn't automation been happening for a long, long time? Don't we have loads of technology that has replaced jobs already?<mask> should we assume this time it will be different?</s>
Label encoding: <s> [STARTQ] Permanent unemployment is going to become more and more of a thing. Automation is going to happen. There simply are going to be less and less jobs, and we need to figure out a way to deal with that, and letting people starve isn't a good plan. [ENDQ] [NEWLINE] I keep hearing rumblings of this, but hasn't automation been happening for a long, long time? Don't we have loads of technology that has replaced jobs already? Why should we assume this time it will be different?</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I grew up in Nova Scotia, Canada,<mask> not only alcohol sales were banned on sundays<mask> general shopping<mask> was (and has just recently changed for shopping). [NEWLINE] [NEWLINE] <mask> a young adult, I remember one benefit of this was that everyone had sundays off work, which made for very exciting night life on saturday nights. [NEWLINE] [NEWLINE] <mask> I now live in a province<mask> alcohol sales are on all days of the week, I do sometimes miss the feeling that saturday night is THE night of the week.</s>
Label encoding: <s>I grew up in Nova Scotia, Canada, where not only alcohol sales were banned on sundays but general shopping also was (and has just recently changed for shopping). [NEWLINE] [NEWLINE] As a young adult, I remember one benefit of this was that everyone had sundays off work, which made for very exciting night life on saturday nights. [NEWLINE] [NEWLINE] While I now live in a province where alcohol sales are on all days of the week, I do sometimes miss the feeling that saturday night is THE night of the week.</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] smoking cigarettes isn't very pleasant the first time. The effect from smoking cigarettes comes from a long time use of the drug.<mask> it's not really a "want" feeling. It's a "need" feeling. You don't want to smoke cigarettes<mask> it's fun. You just need to. [ENDQ] [NEWLINE] You're talking about an entirely subjective experience. I smoke around 3 cigarettes a week. I am in no way addicted to nicotine.  The first time I smoked a cigarette, it was enjoyable. [NEWLINE] </s>
Label encoding: <s> [STARTQ] smoking cigarettes isn't very pleasant the first time. The effect from smoking cigarettes comes from a long time use of the drug. But it's not really a "want" feeling. It's a "need" feeling. You don't want to smoke cigarettes because it's fun. You just need to. [ENDQ] [NEWLINE] You're talking about an entirely subjective experience. I smoke around 3 cigarettes a week. I am in no way addicted to nicotine.  The first time I smoked a cigarette, it was enjoyable. [NEWLINE] </s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For about the bajillionth time, rapes that happen outside of this circumstance are not relevant here. Rapes that happen to men or to women who were were not getting themselves into dangerous situations are unfortunate and should not be tolerated,<mask> all rapes should not be.<mask>, such women who do not deserve sympathy. [NEWLINE] [NEWLINE] My gender gives me more perspective into this, I feel. I simply cannot fathom<mask> someone would go about doing this knowing the consequences.<mask>, no sympathy will be given from me.</s>
Label encoding: <s>For about the bajillionth time, rapes that happen outside of this circumstance are not relevant here. Rapes that happen to men or to women who were were not getting themselves into dangerous situations are unfortunate and should not be tolerated, as all rapes should not be. However, such women who do not deserve sympathy. [NEWLINE] [NEWLINE] My gender gives me more perspective into this, I feel. I simply cannot fathom why someone would go about doing this knowing the consequences. Hence, no sympathy will be given from me.</s>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Taking on faith that an entire religion is 100% factually true without question requires more faith than simply saying "1+1=2". [ENDQ] [NEWLINE] You're trying to redefine (or rather nullify)<mask> faith means<mask> you don't agree with it. [NEWLINE] [NEWLINE] I'm not sure<mask> it came from<mask> I enjoy the quote "faith isn't a virtue, it's the glorification of voluntary ignorance." It's one of the few things from /r/atheism that I found rather entertaining.</s>
Label encoding: <s> [STARTQ] Taking on faith that an entire religion is 100% factually true without question requires more faith than simply saying "1+1=2". [ENDQ] [NEWLINE] You're trying to redefine (or rather nullify) what faith means because you don't agree with it. [NEWLINE] [NEWLINE] I'm not sure where it came from but I enjoy the quote "faith isn't a virtue, it's the glorification of voluntary ignorance." It's one of the few things from /r/atheism that I found rather entertaining.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The difference is that "physical sciences" study a static subject,<mask> sciences preoccupied by social matters and human behaviors need to deal with ever changing variables. Humans change, the way we interact informs following interaction. We're kinda late to the party and we have a huge mess to untangle. [NEWLINE] [NEWLINE] This does lead to a lot of "hit and miss", true,<mask> it<mask> provides valuable knowledge. Statistical work, for example, enables us to know a lot of things about people and their lives. [NEWLINE] </s>
Label encoding: <s>The difference is that "physical sciences" study a static subject, while sciences preoccupied by social matters and human behaviors need to deal with ever changing variables. Humans change, the way we interact informs following interaction. We're kinda late to the party and we have a huge mess to untangle. [NEWLINE] [NEWLINE] This does lead to a lot of "hit and miss", true, but it also provides valuable knowledge. Statistical work, for example, enables us to know a lot of things about people and their lives. [NEWLINE] </s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0681, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0664, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Engineers aren't scientists. Scientists do experiments to answer scientific questions. Engineers don't (necessarily).  That doesn't mean that an engineer can't be a scientist or vice versa,<mask> being an engineer doesn't make you a scientist. [NEWLINE] [NEWLINE] Contrary to<mask> other people are suggesting, the name of the degree doesn't matter. My wife is a music teacher,<mask> she has a B.S. in Education. My father majored in marketing and got a B.S. in business. </s>
Label encoding: <s>Engineers aren't scientists. Scientists do experiments to answer scientific questions. Engineers don't (necessarily).  That doesn't mean that an engineer can't be a scientist or vice versa, but being an engineer doesn't make you a scientist. [NEWLINE] [NEWLINE] Contrary to what other people are suggesting, the name of the degree doesn't matter. My wife is a music teacher, but she has a B.S. in Education. My father majored in marketing and got a B.S. in business. </s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If 1000 terrorists bombed America every year and killed 30 people each, it still wouldn't be<mask> many deaths<mask> we have<mask> of vehicle accidents each year. I don't believe soldiers recently accomplished anything to do with America. The immense waste we've seen in war could've eliminated poverty and removed focus from America simply through our lack of intervention. The ***tragic*** extent to which our actions are misguided would be similar to helping astronauts get food on a lunar base by shipping supplies out into the ocean and sinking them.</s>
Label encoding: <s>If 1000 terrorists bombed America every year and killed 30 people each, it still wouldn't be as many deaths as we have because of vehicle accidents each year. I don't believe soldiers recently accomplished anything to do with America. The immense waste we've seen in war could've eliminated poverty and removed focus from America simply through our lack of intervention. The ***tragic*** extent to which our actions are misguided would be similar to helping astronauts get food on a lunar base by shipping supplies out into the ocean and sinking them.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't mean incapability in any sort of metaphysical sense.<mask> most people have neither the education nor the inclination to make fine logical distinctions. I teach test prep, and among other things, I work with students on critical reasoning problems. Even well-educated students pursuing graduate degrees - and most of my students do go on to get them eventually - frequently have trouble with even basic logical problems. It's just not taught -<mask><mask> this is a bad thing,<mask><mask><mask> it is a true thing.</s>
Label encoding: <s>I don't mean incapability in any sort of metaphysical sense. But most people have neither the education nor the inclination to make fine logical distinctions. I teach test prep, and among other things, I work with students on critical reasoning problems. Even well-educated students pursuing graduate degrees - and most of my students do go on to get them eventually - frequently have trouble with even basic logical problems. It's just not taught - I think this is a bad thing, but I think it is a true thing.</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> you say "<mask> shouldn't we be able to shame fat people"<mask> followed with "not necessarily to their face, that's jst dickish"<mask> are you trying to say? [NEWLINE] [NEWLINE] <mask> you're saying that we shouldn't encourage people for being fat,<mask><mask>. [NEWLINE] <mask> you're saying that by being skinnier makes you a better person, I completely disagree. [NEWLINE] [NEWLINE] <mask> in any case, I prefer to leave them alone unless their obesity is a serious health concern [NEWLINE] </s>
Label encoding: <s>So when you say " Why shouldn't we be able to shame fat people" but followed with "not necessarily to their face, that's jst dickish" what are you trying to say? [NEWLINE] [NEWLINE] If you're saying that we shouldn't encourage people for being fat, I agree. [NEWLINE] If you're saying that by being skinnier makes you a better person, I completely disagree. [NEWLINE] [NEWLINE] But in any case, I prefer to leave them alone unless their obesity is a serious health concern [NEWLINE] </s>
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Studies show that equally competent resumes (and sometimes, identical resumes) are perceived<mask> showing less competence and ability<mask> the name is female. [ENDQ] [NEWLINE] This is much too strong a statement for the research you are providing.  The article you provided only links to one study with 126 participants who were self-selected.  That isn't enough to draw any conclusions about the population<mask> a whole.  It may raise some interesting questions,<mask> it isn't sound enough to make the kind of statement you did.</s>
Label encoding: <s> [STARTQ] Studies show that equally competent resumes (and sometimes, identical resumes) are perceived as showing less competence and ability if the name is female. [ENDQ] [NEWLINE] This is much too strong a statement for the research you are providing.  The article you provided only links to one study with 126 participants who were self-selected.  That isn't enough to draw any conclusions about the population as a whole.  It may raise some interesting questions, but it isn't sound enough to make the kind of statement you did.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See this is<mask> I know you don't read comics, much less Spider-Man. [NEWLINE] [NEWLINE] It's not just a masked mute swinging around, it's Peter Parker.  A person with hopes dreams goals relationships and his own personality. [NEWLINE] [NEWLINE] <mask> the only thing that changed is his skin color it is racist and arbitrary.<mask> his character is changed, it is offensive to fans. [NEWLINE] [NEWLINE] From an outsider perspective it is just a human fighting crime,<mask> its much more involved than that. </s>
Label encoding: <s>See this is how I know you don't read comics, much less Spider-Man. [NEWLINE] [NEWLINE] It's not just a masked mute swinging around, it's Peter Parker.  A person with hopes dreams goals relationships and his own personality. [NEWLINE] [NEWLINE] If the only thing that changed is his skin color it is racist and arbitrary. If his character is changed, it is offensive to fans. [NEWLINE] [NEWLINE] From an outsider perspective it is just a human fighting crime, but its much more involved than that. </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I'm currently in High School, and I hate reading. My reasoning? They throw<mask> much at you<mask> fast. Last year I had to read the Odyssey in 2 weeks (read and annotate it) all leading up to a 10 paragraph essay. There were<mask> other assignments that came along with it. Most of which were internet articles ranging from 2 to 5 pages that we would have to read and annotate. This was the point that I determined<mask> I hates reading<mask> much now. </s>
Label encoding: <s>Well I'm currently in High School, and I hate reading. My reasoning? They throw so much at you so fast. Last year I had to read the Odyssey in 2 weeks (read and annotate it) all leading up to a 10 paragraph essay. There were also other assignments that came along with it. Most of which were internet articles ranging from 2 to 5 pages that we would have to read and annotate. This was the point that I determined why I hates reading so much now. </s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks, this is a really good point for effects of sudden dip in the population. This quite aptly describes the small part of age old social experiment of population curve (the demographic transition model).<mask> really makes me sad is, that this demographic transition is being repeated in underdeveloped countries, even after learning it from developed countries. [NEWLINE] Its like a person who has been a victim of road accident, is watching another road accident and teaching the effects of it, instead of doing anything about it.  </s>
Label encoding: <s>Thanks, this is a really good point for effects of sudden dip in the population. This quite aptly describes the small part of age old social experiment of population curve (the demographic transition model). What really makes me sad is, that this demographic transition is being repeated in underdeveloped countries, even after learning it from developed countries. [NEWLINE] Its like a person who has been a victim of road accident, is watching another road accident and teaching the effects of it, instead of doing anything about it.  </s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Either way, it'll run out<mask> losing out 150 dollars per session is a pretty expensive campaign in the long run. [NEWLINE] [NEWLINE] The longer she does this, the more established it becomes<mask> a pattern of behavior and not a business promotion. This would prove personal reasons rather than business reasons (promoting word of mouth and increasing the clientele list.) [NEWLINE] [NEWLINE] Keep in mind that I am totally not an lawyer and<mask> a layman, I assume that is a paramount requirement in a discrimination case.</s>
Label encoding: <s>Either way, it'll run out since losing out 150 dollars per session is a pretty expensive campaign in the long run. [NEWLINE] [NEWLINE] The longer she does this, the more established it becomes as a pattern of behavior and not a business promotion. This would prove personal reasons rather than business reasons (promoting word of mouth and increasing the clientele list.) [NEWLINE] [NEWLINE] Keep in mind that I am totally not an lawyer and as a layman, I assume that is a paramount requirement in a discrimination case.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't agree with that, that is<mask> the upvoted system is for. [ENDQ] [NEWLINE] I don't see<mask> attracting minorities conflicts with the upvote system, or<mask> reddit is currently changing that. [NEWLINE] [NEWLINE] [STARTQ] <mask> the majority of teddit thinks that black people are thugs yeah that sucks<mask> we shouldn't the say oh now we must change that [ENDQ] [NEWLINE] <mask> it sucks, then<mask> should't we change that? Not by censoring anyone,<mask> just by providing alternative viewpoints.</s>
Label encoding: <s> [STARTQ] I don't agree with that, that is what the upvoted system is for. [ENDQ] [NEWLINE] I don't see where attracting minorities conflicts with the upvote system, or how reddit is currently changing that. [NEWLINE] [NEWLINE] [STARTQ] If the majority of teddit thinks that black people are thugs yeah that sucks but we shouldn't the say oh now we must change that [ENDQ] [NEWLINE] If it sucks, then why should't we change that? Not by censoring anyone, but just by providing alternative viewpoints.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>None of your points addressed the "virtue" of hard work. You only talked about<mask> hard work would personally get you. Here's the definition of virtue: [NEWLINE] [NEWLINE] [STARTQ] noun. behavior showing high moral standards. [ENDQ] [NEWLINE] You're only looking at hard work from your perspective. Generosity is a virtue<mask><mask> it doesn't benefit the person who has it. Hard work would benefit yourself and others around you. Your company benefits from hard workers. Society benefits from hard workers. </s>
Label encoding: <s>None of your points addressed the "virtue" of hard work. You only talked about what hard work would personally get you. Here's the definition of virtue: [NEWLINE] [NEWLINE] [STARTQ] noun. behavior showing high moral standards. [ENDQ] [NEWLINE] You're only looking at hard work from your perspective. Generosity is a virtue even though it doesn't benefit the person who has it. Hard work would benefit yourself and others around you. Your company benefits from hard workers. Society benefits from hard workers. </s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the idea of eco-systems like that is bad in the long run. At first, it might seem awesome,<mask> sooner or later it will become very limiting.<mask> a non-Apple alternative appears that has some amazing new features that would be useful for you, you'd need to either break the eco-system by getting one device that's incompatible with the rest, or keep using the Apple device without these new features and hope that maybe one day Apple will create something similar.</s>
Label encoding: <s>I think the idea of eco-systems like that is bad in the long run. At first, it might seem awesome, but sooner or later it will become very limiting. If a non-Apple alternative appears that has some amazing new features that would be useful for you, you'd need to either break the eco-system by getting one device that's incompatible with the rest, or keep using the Apple device without these new features and hope that maybe one day Apple will create something similar.</s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not a belief. It's a lack of belief.<mask> it is a label that defines your position on the question "does God exist?" This label is meaningful in a country like the US<mask> over 90% of people are religious and say yes to that claim.  I find it odd that people have such a problem with anyone identifying<mask> atheist and use your logic to idk, shame them out of considering themself an atheist? I'm not sure<mask> the point is. </s>
Label encoding: <s>It's not a belief. It's a lack of belief. But it is a label that defines your position on the question "does God exist?" This label is meaningful in a country like the US where over 90% of people are religious and say yes to that claim.  I find it odd that people have such a problem with anyone identifying as atheist and use your logic to idk, shame them out of considering themself an atheist? I'm not sure what the point is. </s>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll admit, social constructs such<mask> clothing don't have to be 100% traceable to biological instincts. Some distortion and cultural adaptation over the millennia is reasonable - I can't imagine our hominid ancestors parading around in frills and neckties. [NEWLINE] [NEWLINE] <mask>, I would maintain that there's at least a distant biological reason for the gender roles in dress we have today. For one, short skirts may draw attention to a female's bottom half and can clearly signal sexual availability.</s>
Label encoding: <s>I'll admit, social constructs such as clothing don't have to be 100% traceable to biological instincts. Some distortion and cultural adaptation over the millennia is reasonable - I can't imagine our hominid ancestors parading around in frills and neckties. [NEWLINE] [NEWLINE] However, I would maintain that there's at least a distant biological reason for the gender roles in dress we have today. For one, short skirts may draw attention to a female's bottom half and can clearly signal sexual availability.</s>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>haha I would try to change your view<mask> you didn't give me much to go on. you can check comment history of my profile, this is my first post,<mask> ive been commenting for a<mask>. Comments will verify Im from arkansas, not that that would be any proof of validity<mask> Im not sure<mask> would be. I would hope the discussion is worth more than the deltas, eliminating the need for covert delta-gaining operations that you are implying.</s>
Label encoding: <s>haha I would try to change your view but you didn't give me much to go on. you can check comment history of my profile, this is my first post, but ive been commenting for a while. Comments will verify Im from arkansas, not that that would be any proof of validity but Im not sure what would be. I would hope the discussion is worth more than the deltas, eliminating the need for covert delta-gaining operations that you are implying.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0657, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Watching a sport is<mask> engaging<mask> you are invested in it. [NEWLINE] [NEWLINE] <mask> you and many generations of your family watch soccer then you will be interested in a team. The same is true of NFL or hockey. [NEWLINE] [NEWLINE] Having played the game yourself will<mask> make it better for you. [NEWLINE] [NEWLINE] Or makes no sense to say one is more enjoyable than the other<mask> enjoyment is entirely subjective. You can absolutely say you prefer one<mask> you cannot demands someone else not like the other.</s>
Label encoding: <s>Watching a sport is as engaging as you are invested in it. [NEWLINE] [NEWLINE] If you and many generations of your family watch soccer then you will be interested in a team. The same is true of NFL or hockey. [NEWLINE] [NEWLINE] Having played the game yourself will also make it better for you. [NEWLINE] [NEWLINE] Or makes no sense to say one is more enjoyable than the other because enjoyment is entirely subjective. You can absolutely say you prefer one but you cannot demands someone else not like the other.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0675, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ You've made some really compelling arguments that have nearly change my view in it's entirety. Unfortunately I'm still really bugged by the chance of a meteor impact. Yes they happen once in hundreds of millions of years,<mask> from my understanding this is just a modeled average. Two meteors could strike back to back within, lets say 100 years, just like someone could get struck by lighting six times in one day, it may be incredibly unlikely<mask> it *can* happen.</s>
Label encoding: <s>∆ You've made some really compelling arguments that have nearly change my view in it's entirety. Unfortunately I'm still really bugged by the chance of a meteor impact. Yes they happen once in hundreds of millions of years, but from my understanding this is just a modeled average. Two meteors could strike back to back within, lets say 100 years, just like someone could get struck by lighting six times in one day, it may be incredibly unlikely but it *can* happen.</s>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can't exactly shop around, looking for someone to bring your reports of extra-judicial NSA activity to. Talk to the wrong person, and you find yourself in solitary confinement for the rest of your life. [NEWLINE] [NEWLINE] The government required that Snowden act the way he did by their barbaric treatment of other whistleblowers. Kennedy said "<mask> you make peaceful change impossible, violent revolution is inevitable."  The same principles apply here. The government closed any route for reform<mask> the one Snowden took. </s><pad>
Label encoding: <s>You can't exactly shop around, looking for someone to bring your reports of extra-judicial NSA activity to. Talk to the wrong person, and you find yourself in solitary confinement for the rest of your life. [NEWLINE] [NEWLINE] The government required that Snowden act the way he did by their barbaric treatment of other whistleblowers. Kennedy said " If you make peaceful change impossible, violent revolution is inevitable."  The same principles apply here. The government closed any route for reform but the one Snowden took. </s><pad>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree there's probably not a pure empirical answer to this.  51% of variance in post-college salary would be a decent metric.  Mostly, I'm looking for evidence that many/most employers care about substantive material taught to most college grads they hire. [NEWLINE] [NEWLINE] It's a fuzzy thing.  The 100%/51% thing was me saying that a single counterexample won't do it, not that I'm really looking for a precise number.</s>
Label encoding: <s>I agree there's probably not a pure empirical answer to this.  51% of variance in post-college salary would be a decent metric.  Mostly, I'm looking for evidence that many/most employers care about substantive material taught to most college grads they hire. [NEWLINE] [NEWLINE] It's a fuzzy thing.  The 100%/51% thing was me saying that a single counterexample won't do it, not that I'm really looking for a precise number.</s>
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a tangential challenge,<mask><mask><mask> considering the case of proper nouns is worth<mask>. [NEWLINE] [NEWLINE] Let's talk about Nikon.  Most of the people reading this will be familiar with Nike-on, the manufacturer of cameras. <mask>, most of the people working for the company know it<mask> Nee-kon. [NEWLINE] [NEWLINE] Are you suggesting that<mask> a Japanese person with excellent English was speaking, they should adjust the pronunciation of their company depending on audience?</s>
Label encoding: <s>This is a tangential challenge, but I think considering the case of proper nouns is worth while. [NEWLINE] [NEWLINE] Let's talk about Nikon.  Most of the people reading this will be familiar with Nike-on, the manufacturer of cameras.  However, most of the people working for the company know it as Nee-kon. [NEWLINE] [NEWLINE] Are you suggesting that if a Japanese person with excellent English was speaking, they should adjust the pronunciation of their company depending on audience?</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This is exactly the instance of "overrated" that I am speaking out against,<mask> we are in agreement. [ENDQ] [NEWLINE] I understand,<mask><mask><mask> even in this case, the word has the implication that "sometime in the future everyone will realize that this is not<mask> good relative to these other things."  This is different than just saying, I don't like it. [NEWLINE] [NEWLINE] Of course, someone should have to *defend*<mask> something is overrated.</s>
Label encoding: <s> [STARTQ] This is exactly the instance of "overrated" that I am speaking out against, so we are in agreement. [ENDQ] [NEWLINE] I understand, but I think even in this case, the word has the implication that "sometime in the future everyone will realize that this is not as good relative to these other things."  This is different than just saying, I don't like it. [NEWLINE] [NEWLINE] Of course, someone should have to *defend* why something is overrated.</s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right<mask><mask> with that. Morals vary with cultures and time. 300 years ago we had few qualms with slavery. Before that, we were trying and executing people we suspected to be witches based on discrimination and little proof. [NEWLINE] [NEWLINE] Locke's theory of natural rights advanced morality and we now recognize these rights to be true and valid, at least in most developed nations. I hope that<mask> time progresses, other nations and cultures become more tolerant and accepting<mask> well. </s>
Label encoding: <s>Right I agree with that. Morals vary with cultures and time. 300 years ago we had few qualms with slavery. Before that, we were trying and executing people we suspected to be witches based on discrimination and little proof. [NEWLINE] [NEWLINE] Locke's theory of natural rights advanced morality and we now recognize these rights to be true and valid, at least in most developed nations. I hope that as time progresses, other nations and cultures become more tolerant and accepting as well. </s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Young men are more at risk to reckless [ENDQ] [NEWLINE] <mask> aren't blacks in the USA more at risk to be criminal? Of course that is not for an inherent reason, of course it's not in their DNA, it's just the way things are right now. [NEWLINE] [NEWLINE] It is not necessary to know about testosterone in order to profile young men. Even<mask> you couldn't explain their statistical dominance it would still be profitable to charge them more, simply<mask> of the statistics.</s>
Label encoding: <s> [STARTQ] Young men are more at risk to reckless [ENDQ] [NEWLINE] But aren't blacks in the USA more at risk to be criminal? Of course that is not for an inherent reason, of course it's not in their DNA, it's just the way things are right now. [NEWLINE] [NEWLINE] It is not necessary to know about testosterone in order to profile young men. Even if you couldn't explain their statistical dominance it would still be profitable to charge them more, simply because of the statistics.</s>
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I should probably have clarified a bit more,<mask> I will do<mask> here: I talk only of people who get drunk on a regular basis and/or have drinking problems. Glass of wine? Not really a problem. Even a beer or two? Not a problem. [NEWLINE] [NEWLINE] With regard to drugs and masturbation, and comparing them to video games and TV, I will have to think a little more on that and will get back to you on it. Thanks for replying</s>
Label encoding: <s>I should probably have clarified a bit more, so I will do so here: I talk only of people who get drunk on a regular basis and/or have drinking problems. Glass of wine? Not really a problem. Even a beer or two? Not a problem. [NEWLINE] [NEWLINE] With regard to drugs and masturbation, and comparing them to video games and TV, I will have to think a little more on that and will get back to you on it. Thanks for replying</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The idea that people are against is<mask> Politicians change their views just to gain voters or to be elected into office. People who before election appeal to a voter base for legalizing MJ and then in office are staunchly against it. [NEWLINE] [NEWLINE] Generally speaking, I do agree with you. Politicians like everyone should keep their mind open to new information and be willing to adapt their view based on the best evidence available to them. Essentially people don't want to feel gamed.</s>
Label encoding: <s>The idea that people are against is when Politicians change their views just to gain voters or to be elected into office. People who before election appeal to a voter base for legalizing MJ and then in office are staunchly against it. [NEWLINE] [NEWLINE] Generally speaking, I do agree with you. Politicians like everyone should keep their mind open to new information and be willing to adapt their view based on the best evidence available to them. Essentially people don't want to feel gamed.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> no, I'm not afraid of dying in a terrorist attack. [ENDQ] [NEWLINE] then<mask> do you disagree with me. I dont deny anything you said I just dont see<mask> you're trying to say. [NEWLINE] [NEWLINE] its like you're trying to say you're not afraid of a terrorist attack<mask>... you still think its a serious enough threat to deserve all the attention it gets?<mask> shouldnt you be afraid<mask> its a serious enough threat to deserve the attention?</s>
Label encoding: <s> [STARTQ] So no, I'm not afraid of dying in a terrorist attack. [ENDQ] [NEWLINE] then why do you disagree with me. I dont deny anything you said I just dont see what you're trying to say. [NEWLINE] [NEWLINE] its like you're trying to say you're not afraid of a terrorist attack but... you still think its a serious enough threat to deserve all the attention it gets? but shouldnt you be afraid if its a serious enough threat to deserve the attention?</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0663, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, first someone (in this case, the owners of Hobby Lobby) have to believe that their rights are being infringed, and lodge a legal challenge. Then, it gets escalated through the courts,<mask> the justices use previously established case law to decide whether or not religious freedom applies. Finally,<mask> it gets challenged all the way to the supreme court, all of those justices will vote on it, and decide. [NEWLINE] [NEWLINE] Kind of like<mask> actually happened. </s>
Label encoding: <s>Well, first someone (in this case, the owners of Hobby Lobby) have to believe that their rights are being infringed, and lodge a legal challenge. Then, it gets escalated through the courts, where the justices use previously established case law to decide whether or not religious freedom applies. Finally, if it gets challenged all the way to the supreme court, all of those justices will vote on it, and decide. [NEWLINE] [NEWLINE] Kind of like what actually happened. </s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0695, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your view of those executives is distorted. Most of them *don't* live like that or view it like that. The amount of money is a lot,<mask> the people who are skilled enough to manage a company<mask> they can make that much money are few and far between. The pay is competitive. [NEWLINE] [NEWLINE] You should remember that most of these executives aren't making millions of dollars every year. They're<mask> the ones footing the bill for everything. </s>
Label encoding: <s>I think your view of those executives is distorted. Most of them *don't* live like that or view it like that. The amount of money is a lot, but the people who are skilled enough to manage a company where they can make that much money are few and far between. The pay is competitive. [NEWLINE] [NEWLINE] You should remember that most of these executives aren't making millions of dollars every year. They're also the ones footing the bill for everything. </s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Group a knows variables x &amp; y, group b knows only variable x. This question involves understanding x &amp; y to debate strongly unfortunately. [NEWLINE] [NEWLINE] You can postulate y,<mask> people who know x + y will have to spend time explaining the variable and<mask> some form or logic makes sense. [NEWLINE] [NEWLINE] It's like someone who has never had sex before debating<mask> it's not important. Their opinion should be made from a point of actual understanding.</s>
Label encoding: <s>Group a knows variables x &amp; y, group b knows only variable x. This question involves understanding x &amp; y to debate strongly unfortunately. [NEWLINE] [NEWLINE] You can postulate y, but people who know x + y will have to spend time explaining the variable and why some form or logic makes sense. [NEWLINE] [NEWLINE] It's like someone who has never had sex before debating why it's not important. Their opinion should be made from a point of actual understanding.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>'s the last time you saw a biker in hi-vis clothing? Never, right? They wear black leather. [ENDQ] [NEWLINE] Actually, synthetic jackets (often with reflective patches) are becoming more and more common. [NEWLINE] [NEWLINE] Regardless, color of someone's jacket is only a small issue. Leather became a staple not just<mask> of its looks -- it serves vital protection. Lights and loud noises are much more attention-grabbing than the color of your jacket.</s>
Label encoding: <s> [STARTQ] When's the last time you saw a biker in hi-vis clothing? Never, right? They wear black leather. [ENDQ] [NEWLINE] Actually, synthetic jackets (often with reflective patches) are becoming more and more common. [NEWLINE] [NEWLINE] Regardless, color of someone's jacket is only a small issue. Leather became a staple not just because of its looks -- it serves vital protection. Lights and loud noises are much more attention-grabbing than the color of your jacket.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People already take out huge loans to go to school, and the dropout rate is unacceptable. Free education would result in overutilization, without the attendant increase in quality necessary to support that. The result? Dilution and debt. K-12 should be free and high-quality. I would<mask><mask> college should not be unless there's a push for credits earned through MOOCs. [NEWLINE] [NEWLINE] **<mask> you increase pressure, you must have release valves.**</s>
Label encoding: <s>People already take out huge loans to go to school, and the dropout rate is unacceptable. Free education would result in overutilization, without the attendant increase in quality necessary to support that. The result? Dilution and debt. K-12 should be free and high-quality. I would argue that college should not be unless there's a push for credits earned through MOOCs. [NEWLINE] [NEWLINE] ** If you increase pressure, you must have release valves.**</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except it's not better in every day. [NEWLINE] [NEWLINE] Fahrenheit is based literally on body temperature and gives a more "human friendly" scale. [NEWLINE] [NEWLINE] Add to that the fact that 0°F is<mask> salt water freezes on the roads, which is helpful to know. [NEWLINE] [NEWLINE] Celsius is absolutely better in terms of scientific measurements (Kelvin is best,<mask> whatever). [NEWLINE] [NEWLINE] <mask> in terms of direct human impact, F is better than C </s>
Label encoding: <s>Except it's not better in every day. [NEWLINE] [NEWLINE] Fahrenheit is based literally on body temperature and gives a more "human friendly" scale. [NEWLINE] [NEWLINE] Add to that the fact that 0°F is when salt water freezes on the roads, which is helpful to know. [NEWLINE] [NEWLINE] Celsius is absolutely better in terms of scientific measurements (Kelvin is best, but whatever). [NEWLINE] [NEWLINE] But in terms of direct human impact, F is better than C </s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you are obedient, your changes are<mask> close to zero<mask> non-zero mathematically allows. [ENDQ] [NEWLINE] And<mask><mask> I'm not obedient? Does that mean I brought it on myself? The fact that I can be completely compliant,<mask> be shot to death<mask> I went to get my license [like this guy]( [URL] /) is crazy.  Thankfully, that officer was fired and charged,<mask> many aren't<mask> lucky,<mask> we've recently seen.</s>
Label encoding: <s> [STARTQ] If you are obedient, your changes are as close to zero as non-zero mathematically allows. [ENDQ] [NEWLINE] And what if I'm not obedient? Does that mean I brought it on myself? The fact that I can be completely compliant, but be shot to death because I went to get my license [like this guy]( [URL] /) is crazy.  Thankfully, that officer was fired and charged, but many aren't so lucky, as we've recently seen.</s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oops, I hadn't read the comment above the one I was responding to.  You're right, the Equal Pay Act does not directly address promotions, only hiring and wages. [NEWLINE] [NEWLINE] Regardless, people have successfully sued numerous times against discriminatory promotion practices.  I can't find the law behind it,<mask> it is illegal in the US.  Walmart v. Dukes, Velez v. Novartis, and the EEOC all address this.</s>
Label encoding: <s>Oops, I hadn't read the comment above the one I was responding to.  You're right, the Equal Pay Act does not directly address promotions, only hiring and wages. [NEWLINE] [NEWLINE] Regardless, people have successfully sued numerous times against discriminatory promotion practices.  I can't find the law behind it, but it is illegal in the US.  Walmart v. Dukes, Velez v. Novartis, and the EEOC all address this.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To lead you need a mandate and they failed to secure that. Rebooting is the better option. Such a move is not about the individual; it's an indication that it's not about their own individual power<mask> concern for<mask> is better moving forward.<mask>, there is no better time to regroup and reform than at an early point after an election. Even the winners reconsider<mask> to make the best of<mask> they have now.. resetting the cabinet etc.</s>
Label encoding: <s>To lead you need a mandate and they failed to secure that. Rebooting is the better option. Such a move is not about the individual; it's an indication that it's not about their own individual power but concern for what is better moving forward. Also, there is no better time to regroup and reform than at an early point after an election. Even the winners reconsider how to make the best of what they have now.. resetting the cabinet etc.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> do you think the military should have to do with the equipment? [NEWLINE] [NEWLINE] I would<mask> say that having local police forces better armed against the federal government outweighs the jackass cops and scary raids.<mask> the SWAT raids result in many more deaths or citizens being deathly afraid of the police then I will agree with you,<mask> I just don't think that's the case.  Ferguson still burned down their city even with guns pointed at them.</s>
Label encoding: <s>So what do you think the military should have to do with the equipment? [NEWLINE] [NEWLINE] I would also say that having local police forces better armed against the federal government outweighs the jackass cops and scary raids. If the SWAT raids result in many more deaths or citizens being deathly afraid of the police then I will agree with you, but I just don't think that's the case.  Ferguson still burned down their city even with guns pointed at them.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0987, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Writing about this clearly is hard,<mask> luckily it has been done already by people that are better at this than me,<mask> I'd say read [this]( [URL] /). [NEWLINE] [NEWLINE] The short version is that you are part of the system. Your mind is one of the inputs to the equations that determine<mask> happens: that's<mask> you exercise free will. [NEWLINE] [NEWLINE] The even longer than the link version I guess is [this]( [URL] (solution)).</s>
Label encoding: <s>Writing about this clearly is hard, but luckily it has been done already by people that are better at this than me, so I'd say read [this]( [URL] /). [NEWLINE] [NEWLINE] The short version is that you are part of the system. Your mind is one of the inputs to the equations that determine what happens: that's how you exercise free will. [NEWLINE] [NEWLINE] The even longer than the link version I guess is [this]( [URL] (solution)).</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fair enough. [NEWLINE] [NEWLINE] <mask> even<mask> we assume high school English isn't actively killing the desire to read for pleasure, you'd have a tough time making the case that it's doing anything to cultivate it. [NEWLINE] [NEWLINE] And shouldn't the first order goal of high school English be to instill a love of reading? After all,<mask> someone doesn't ever read, then there's little point in teaching them the reading skills the class is trying to impart.</s>
Label encoding: <s>Fair enough. [NEWLINE] [NEWLINE] But even if we assume high school English isn't actively killing the desire to read for pleasure, you'd have a tough time making the case that it's doing anything to cultivate it. [NEWLINE] [NEWLINE] And shouldn't the first order goal of high school English be to instill a love of reading? After all, if someone doesn't ever read, then there's little point in teaching them the reading skills the class is trying to impart.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why would the government kill thousands of people for money<mask> they are the ones who print the money? [NEWLINE] [NEWLINE] <mask>, different people are going to have different takes on<mask> happened that day. Look at all of the interviews you posted. Think about<mask> many different people all have their own say, and<mask> it's really  no different than all of the comments you see in this thread. Everyone has their own version of events, especially in something<mask> complex.</s>
Label encoding: <s>Why would the government kill thousands of people for money when they are the ones who print the money? [NEWLINE] [NEWLINE] Also, different people are going to have different takes on what happened that day. Look at all of the interviews you posted. Think about how many different people all have their own say, and how it's really  no different than all of the comments you see in this thread. Everyone has their own version of events, especially in something so complex.</s>
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Isn't this a bit of a cop-out via a self-fulfilling prophecy? [NEWLINE] [NEWLINE] Once you stop believing that you have any control over your own future, you certainly lose any control you had.<mask>, at that point it isn't<mask> of some cruel fate of birth,<mask><mask> you've willingly given up your agency. [NEWLINE] [NEWLINE] Looking at your past submissions, it might be more effective to redefine your definition of success and value.</s>
Label encoding: <s>Isn't this a bit of a cop-out via a self-fulfilling prophecy? [NEWLINE] [NEWLINE] Once you stop believing that you have any control over your own future, you certainly lose any control you had. However, at that point it isn't because of some cruel fate of birth, but because you've willingly given up your agency. [NEWLINE] [NEWLINE] Looking at your past submissions, it might be more effective to redefine your definition of success and value.</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ah,<mask> is it democratic to enforce the Constitution upon today's populace?<mask> not one person alive voted for it? [NEWLINE] [NEWLINE] Even<mask> you<mask><mask> it could be amended, the (exceedingly difficult) processes by which to do<mask> _are dictated by the constitution._ [NEWLINE] [NEWLINE] The constitution itself is fairly undemocratic - and (at least in my view) the safe guards it provides are the same counter-democratic methods I mentioned above.</s>
Label encoding: <s>Ah, but is it democratic to enforce the Constitution upon today's populace? When not one person alive voted for it? [NEWLINE] [NEWLINE] Even if you argue that it could be amended, the (exceedingly difficult) processes by which to do so _are dictated by the constitution._ [NEWLINE] [NEWLINE] The constitution itself is fairly undemocratic - and (at least in my view) the safe guards it provides are the same counter-democratic methods I mentioned above.</s>
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>if someone came to me and said that 1+1=3, i would want to know<mask><mask> it would probably be crazy and interesting to listen to.<mask> you are right i wouldnt really think they had any substantial argument.<mask> they did<mask> i might reconsider my views. the point i am trying to make is that you should never say, "i am unwilling to reconsider my views even<mask> you provide all evidence necessary to do<mask>." </s>
Label encoding: <s>if someone came to me and said that 1+1=3, i would want to know why because it would probably be crazy and interesting to listen to. but you are right i wouldnt really think they had any substantial argument. if they did though i might reconsider my views. the point i am trying to make is that you should never say, "i am unwilling to reconsider my views even if you provide all evidence necessary to do so." </s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a difference between accepting that something came about naturally and believing that it shouldn't have a negative connotation,<mask>. (I realize you weren't giving your opinion,<mask> I guess this is more of a hypothetical debate) Just<mask> we cannot change the fact that some people are pedophiles, doesn't mean we should accept it. This could cause pedophiles to believe that their thoughts are fine, and<mask> more willing to act on them. </s>
Label encoding: <s>There's a difference between accepting that something came about naturally and believing that it shouldn't have a negative connotation, though. (I realize you weren't giving your opinion, so I guess this is more of a hypothetical debate) Just because we cannot change the fact that some people are pedophiles, doesn't mean we should accept it. This could cause pedophiles to believe that their thoughts are fine, and therefore more willing to act on them. </s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not saying that I bypass it,<mask> I will admit that I tend to be a "buffet Catholic." I take the good and leave the bad. I have never been strict in my religion<mask> I will say that prayer has been relaxing and encouraging to me. It does feel good to put the burden of death upon someone higher than myself. You're right on my view of morality,<mask>. Whatever causes people to live better is a win. </s>
Label encoding: <s>Not saying that I bypass it, but I will admit that I tend to be a "buffet Catholic." I take the good and leave the bad. I have never been strict in my religion but I will say that prayer has been relaxing and encouraging to me. It does feel good to put the burden of death upon someone higher than myself. You're right on my view of morality, however. Whatever causes people to live better is a win. </s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Say<mask> you want about the Sex Pistols,<mask> I can't hear<mask> you're talking about with the Ramones. Sure, the actual songs are extremely simple,<mask>, in their recordings they sound *very* together to me. Could you mention particular songs, and the parts of those songs that sound lazy to you? [NEWLINE] [NEWLINE] I guess in their live recordings its often clear that they're playing faster and probably drunker than they should be.</s>
Label encoding: <s>Say what you want about the Sex Pistols, but I can't hear what you're talking about with the Ramones. Sure, the actual songs are extremely simple, however, in their recordings they sound *very* together to me. Could you mention particular songs, and the parts of those songs that sound lazy to you? [NEWLINE] [NEWLINE] I guess in their live recordings its often clear that they're playing faster and probably drunker than they should be.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0805, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While people are put on sex offenders lists for public urination (and that's appalling) it is not a widespread issue and can only legally occur in a dozen or<mask> states. [NEWLINE] [NEWLINE] <mask>, the 'need' for public urination in the US is substantially different. Public lavatories are pretty much everywhere in urban and suburban areas<mask> in Italy I've sometimes had to wait an hour or two<mask> I don't have proper change, haha.</s>
Label encoding: <s>While people are put on sex offenders lists for public urination (and that's appalling) it is not a widespread issue and can only legally occur in a dozen or so states. [NEWLINE] [NEWLINE] Also, the 'need' for public urination in the US is substantially different. Public lavatories are pretty much everywhere in urban and suburban areas where in Italy I've sometimes had to wait an hour or two if I don't have proper change, haha.</s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0945, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yep. That's<mask> I am screaming. I feel like open relationships are for Ancient China with concubines or the 70's before STDs were lethal. [NEWLINE] <mask> that's a whole other topic in and of itself. I was really hoping someone could come on here and talk about<mask> they once thought like me and they tried XYZ and saw some other light.<mask><mask><mask> you are right - it's just not for some people.</s>
Label encoding: <s>Yep. That's what I am screaming. I feel like open relationships are for Ancient China with concubines or the 70's before STDs were lethal. [NEWLINE] But that's a whole other topic in and of itself. I was really hoping someone could come on here and talk about how they once thought like me and they tried XYZ and saw some other light. But I think you are right - it's just not for some people.</s>
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't agree with the liar/slut thing. I'm sure you could find sensationalist news stories about this happening n<mask> that doesn't mean it "essentially" happens. I realize that sometimes the entire case can work against the victim,<mask> in my experience,<mask> someone is accused, people accept it. Don't misunderstand me, I get that sometimes this happens,<mask> it isn't even close to most of the time.</s>
Label encoding: <s>I don't agree with the liar/slut thing. I'm sure you could find sensationalist news stories about this happening n but that doesn't mean it "essentially" happens. I realize that sometimes the entire case can work against the victim, but in my experience, when someone is accused, people accept it. Don't misunderstand me, I get that sometimes this happens, but it isn't even close to most of the time.</s>
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> should we define human rights, and<mask> do people lose these rights,<mask> it than becomes moral to take them. Such<mask> imprisonment or the death penalty. [ENDQ] [NEWLINE] We define human rights by<mask> we can agree on are things that are human rights. That's really it.<mask> morality is subjective, "it becomes moral to take away rights" in the scenarios that we agree it is moral to take away rights<mask> a society.</s>
Label encoding: <s> [STARTQ] How should we define human rights, and when do people lose these rights, where it than becomes moral to take them. Such as imprisonment or the death penalty. [ENDQ] [NEWLINE] We define human rights by what we can agree on are things that are human rights. That's really it. as morality is subjective, "it becomes moral to take away rights" in the scenarios that we agree it is moral to take away rights as a society.</s>
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0920, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyone partakes in some form of escapism.  Reading a story, listening to music (usually), watching a movie or tv show, or even engaging in a debate about drugs are all ways to forget the other stuff in your life and focus on something interesting or different.<mask> most of the stuff we do in life is kind of a defense mechanism/escape mechanism,<mask> doing drugs is just another part of life in that regard.</s>
Label encoding: <s>Everyone partakes in some form of escapism.  Reading a story, listening to music (usually), watching a movie or tv show, or even engaging in a debate about drugs are all ways to forget the other stuff in your life and focus on something interesting or different. Also most of the stuff we do in life is kind of a defense mechanism/escape mechanism, so doing drugs is just another part of life in that regard.</s>
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The the "buzz" you get from nicotine is actually acts<mask> a treatment for the symptoms of ADHD.  This is<mask> you see more than 40% of the adult ADHD population smoking compared with 26% in the general population.  Nicotine is no substitute for methylphenidate,<mask> many un-diagnosed adults with ADHD turn to it<mask> it provides a temporary relief. [NEWLINE] [NEWLINE] &gt; [source]( [URL].pdf)</s>
Label encoding: <s>The the "buzz" you get from nicotine is actually acts as a treatment for the symptoms of ADHD.  This is why you see more than 40% of the adult ADHD population smoking compared with 26% in the general population.  Nicotine is no substitute for methylphenidate, but many un-diagnosed adults with ADHD turn to it because it provides a temporary relief. [NEWLINE] [NEWLINE] &gt; [source]( [URL].pdf)</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be fair, dogs do have a lot of roles in society, and many breeds were bred for hunting and guarding,<mask> I'm having a hard time imagining the cow equivalent of a Labrador, being bred almost solely for companionship. Am I saying that all dogs are sacred companions that are designed solely to be loyal to us? No.<mask><mask> I am saying is that there is definitely a biological difference that goes deeper than culture alone.</s>
Label encoding: <s>To be fair, dogs do have a lot of roles in society, and many breeds were bred for hunting and guarding, but I'm having a hard time imagining the cow equivalent of a Labrador, being bred almost solely for companionship. Am I saying that all dogs are sacred companions that are designed solely to be loyal to us? No. But what I am saying is that there is definitely a biological difference that goes deeper than culture alone.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not justifiable, you total up actual costs in the $200 billion range, then say a very calculable number is incalculable (<mask> the inclusion of this point is entirely debatable, the dot com bubble had already burst and the economy was hemorrhaging before 9/11 happened), to cap it off you then throw in the cost of a poorly calculated response to the attack<mask> a cost of the attack.  </s>
Label encoding: <s>That's not justifiable, you total up actual costs in the $200 billion range, then say a very calculable number is incalculable ( also the inclusion of this point is entirely debatable, the dot com bubble had already burst and the economy was hemorrhaging before 9/11 happened), to cap it off you then throw in the cost of a poorly calculated response to the attack as a cost of the attack.  </s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I work for a big engineering firm; and I would beg to differ about your understanding of<mask> it works. [NEWLINE] [NEWLINE] <mask> the principles behind engineering are rock solid, I've<mask> to see one thing that worked precisely<mask> it was designed the first time around. <mask> you must challenge<mask> you *thought* your object would work, find the failure point, and refine your design. [NEWLINE] [NEWLINE] Not unlike<mask> sunny views his spiritual journey.</s><pad>
Label encoding: <s>I work for a big engineering firm; and I would beg to differ about your understanding of how it works. [NEWLINE] [NEWLINE] While the principles behind engineering are rock solid, I've yet to see one thing that worked precisely as it was designed the first time around.  So you must challenge how you *thought* your object would work, find the failure point, and refine your design. [NEWLINE] [NEWLINE] Not unlike how sunny views his spiritual journey.</s><pad>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0669, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh, I don't know<mask> that's everyone who believes that. Aren't people who call themselves moral universalists in the same vein? Again, I don't know all the philosophical lingo -- maybe consequentialists are kind of<mask> I'm getting at. [NEWLINE] [NEWLINE] All I know is people like Sam Harris believe the science can, one day, distinguish between moral and immoral.<mask>, whatever he is, and he's not religious.</s>
Label encoding: <s>Oh, I don't know if that's everyone who believes that. Aren't people who call themselves moral universalists in the same vein? Again, I don't know all the philosophical lingo -- maybe consequentialists are kind of what I'm getting at. [NEWLINE] [NEWLINE] All I know is people like Sam Harris believe the science can, one day, distinguish between moral and immoral. So, whatever he is, and he's not religious.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] Lots of really good responses in this thread, and viewpoints that I never considered,<mask> this is the one that did it for me. [NEWLINE] [NEWLINE] <mask><mask> I may not personally have or see an issue with unequal pay in my own life doesn't mean it doesn't exist for others, and just<mask> it's not an issue for ME doesn't mean it's not an issue. Thanks :)</s>
Label encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] Lots of really good responses in this thread, and viewpoints that I never considered, but this is the one that did it for me. [NEWLINE] [NEWLINE] Even though I may not personally have or see an issue with unequal pay in my own life doesn't mean it doesn't exist for others, and just because it's not an issue for ME doesn't mean it's not an issue. Thanks :)</s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, they are. About 345,000 in annual income or 1.5 million in savings puts you in the 1% of either category, quite doable with married couples with shared finances.<mask> you want to know one of the many reasons OWS failed, it was<mask> a lot of initially sympathetic professionals and business owners did the math and realized they were part of the 1% whose blood and assets were being called for. </s>
Label encoding: <s>Yes, they are. About 345,000 in annual income or 1.5 million in savings puts you in the 1% of either category, quite doable with married couples with shared finances. If you want to know one of the many reasons OWS failed, it was because a lot of initially sympathetic professionals and business owners did the math and realized they were part of the 1% whose blood and assets were being called for. </s>
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Without any sort of meaningful regulation, "market forces",<mask> you euphemize them, would push average monthly rent into five figure territory, hastening NYC's decline into a Singaporean playground for super-rich part time residents who see housing<mask> an investment instead of a necessity, pricing out anybody who isn't a millionaire. It would<mask> turn the rental market into a speculative bubble and we all know<mask> those tend to work out. </s>
Label encoding: <s>Without any sort of meaningful regulation, "market forces", as you euphemize them, would push average monthly rent into five figure territory, hastening NYC's decline into a Singaporean playground for super-rich part time residents who see housing as an investment instead of a necessity, pricing out anybody who isn't a millionaire. It would also turn the rental market into a speculative bubble and we all know how those tend to work out. </s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks, I'll do that. [NEWLINE] [NEWLINE] Its worth pointing out it was your first post that got me to start questioning my view. [NEWLINE] [NEWLINE] I'm not there<mask>, I'll admit that the thought does worry and scare me,<mask> I'm going to put myself out there and offer everyone a welcoming hand no-matter who they are or who their partners are. And hopefully we'll all come off better people for it.</s>
Label encoding: <s>Thanks, I'll do that. [NEWLINE] [NEWLINE] Its worth pointing out it was your first post that got me to start questioning my view. [NEWLINE] [NEWLINE] I'm not there yet, I'll admit that the thought does worry and scare me, but I'm going to put myself out there and offer everyone a welcoming hand no-matter who they are or who their partners are. And hopefully we'll all come off better people for it.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay, I want to try this one again<mask> I've thought of another analogy. [NEWLINE] [NEWLINE] Two people design and arm a bomb that will detonate automatically in an hour in a building full of people.  One of them is holding the remote with a "cancellation" button.  He doesn't press it and the bomb explodes<mask> it was designed to do. [NEWLINE] [NEWLINE] Who is responsible for the explosion?</s>
Label encoding: <s>Okay, I want to try this one again since I've thought of another analogy. [NEWLINE] [NEWLINE] Two people design and arm a bomb that will detonate automatically in an hour in a building full of people.  One of them is holding the remote with a "cancellation" button.  He doesn't press it and the bomb explodes as it was designed to do. [NEWLINE] [NEWLINE] Who is responsible for the explosion?</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From my experience, very much yes.<mask> I was dialing in on<mask> medicating worked, I had a problem in determining<mask> much I should smoke and<mask> in the day for the best effect. This was<mask> I was switching strains every month<mask> I didn't have a stable dealer. The big split is between Indica and Sativa, the first better for pain, the second for focus and appetite. </s>
Label encoding: <s>From my experience, very much yes. When I was dialing in on how medicating worked, I had a problem in determining how much I should smoke and when in the day for the best effect. This was because I was switching strains every month because I didn't have a stable dealer. The big split is between Indica and Sativa, the first better for pain, the second for focus and appetite. </s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In Latin, *res publica* simply meant the state. Even<mask> the Senate became a sham, Rome still called itself a *res publica.* In terms of political science, a republic like the US is contrasted with a direct democracy. The US certainly is a democratic republic,<mask> it is a republic first and foremost. Keep in mind, we didn't even elect senators directly until just over 100 years ago.</s>
Label encoding: <s>In Latin, *res publica* simply meant the state. Even when the Senate became a sham, Rome still called itself a *res publica.* In terms of political science, a republic like the US is contrasted with a direct democracy. The US certainly is a democratic republic, but it is a republic first and foremost. Keep in mind, we didn't even elect senators directly until just over 100 years ago.</s>
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Came here to say this. My wife uses your strategy,<mask> I am more of your sister's style. After solving a puzzle, my wife really, really knows the painting (we do puzzles of paintings),<mask> my share perception and spacial cognitive skills are better than hers. [NEWLINE] Before I met my wife, it hadn't even crossed my mind to look at the picture in order to solve the puzzle! </s>
Label encoding: <s>Came here to say this. My wife uses your strategy, while I am more of your sister's style. After solving a puzzle, my wife really, really knows the painting (we do puzzles of paintings), while my share perception and spacial cognitive skills are better than hers. [NEWLINE] Before I met my wife, it hadn't even crossed my mind to look at the picture in order to solve the puzzle! </s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're ignoring the fact that there is already a standard.  The law is the standard. The law says "the person opening the door is supposed to check."  Anybody assuming different is the one acting differently.  In the narrative you've created, the person opening the door is effectively thinking "it doesn't matter that the average person would think that I would look, I'm in the right here."</s>
Label encoding: <s>You're ignoring the fact that there is already a standard.  The law is the standard. The law says "the person opening the door is supposed to check."  Anybody assuming different is the one acting differently.  In the narrative you've created, the person opening the door is effectively thinking "it doesn't matter that the average person would think that I would look, I'm in the right here."</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is absolutely terribly advice. Someone breaking into your home at night, in the United States, is presumed to be a lethal threat until otherwise proven. [NEWLINE] [NEWLINE] You don't ever meet a lethal threat with anything other than lethal force. To do<mask> is to invite injury or death upon yourself and your family. [NEWLINE] [NEWLINE] Not to mention, pepper spray will not incapacitate a person who intends to do you harm.</s>
Label encoding: <s>This is absolutely terribly advice. Someone breaking into your home at night, in the United States, is presumed to be a lethal threat until otherwise proven. [NEWLINE] [NEWLINE] You don't ever meet a lethal threat with anything other than lethal force. To do so is to invite injury or death upon yourself and your family. [NEWLINE] [NEWLINE] Not to mention, pepper spray will not incapacitate a person who intends to do you harm.</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're 100% correct<mask> for me it's about the ideal world view. I guess it would've taken the idea of being homosexual not being any different to being to straight to make people never feel the need to have to have a specific gay community, to me the idea of needing a community is all about safety in numbers homosexual people feel they need a specific place to actually be able to express themselves.</s>
Label encoding: <s>I think you're 100% correct however for me it's about the ideal world view. I guess it would've taken the idea of being homosexual not being any different to being to straight to make people never feel the need to have to have a specific gay community, to me the idea of needing a community is all about safety in numbers homosexual people feel they need a specific place to actually be able to express themselves.</s>
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're trying to force labels.<mask> someone has brown, black, or white skin, they are objectively brown, black, or white. All I'm saying is it's okay to describe people by skin color. That black person who's family has been in America for generations isn't african american, he's a black skinned american. That isn't a label. That is fact. </s>
Label encoding: <s>You're trying to force labels. If someone has brown, black, or white skin, they are objectively brown, black, or white. All I'm saying is it's okay to describe people by skin color. That black person who's family has been in America for generations isn't african american, he's a black skinned american. That isn't a label. That is fact. </s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And then that buddy gets reviewed and punished. Then the buddy who protects him. Then the buddy who protects him. Like I said: a great way to get all of the abusive cops out quickly<mask> it goes the way you think. [NEWLINE] [NEWLINE] You are essentially saying "the police force is broken and we can never fix it." Which is bullshit. Plenty of places have professional, appropriate, accountable police forces.</s>
Label encoding: <s>And then that buddy gets reviewed and punished. Then the buddy who protects him. Then the buddy who protects him. Like I said: a great way to get all of the abusive cops out quickly if it goes the way you think. [NEWLINE] [NEWLINE] You are essentially saying "the police force is broken and we can never fix it." Which is bullshit. Plenty of places have professional, appropriate, accountable police forces.</s>
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] God makes the universe. He doesn't need to describe it to himself.<mask><mask> he's explaining it to someone else, he has to use words that they will understand.<mask> he used the words "day and night,"<mask><mask> at the time of Creation, that wouldn't have made any sense until the sun was created. [ENDQ] [NEWLINE] <mask> you believe the creation story in genesis is an allegory?</s>
Label encoding: <s> [STARTQ] God makes the universe. He doesn't need to describe it to himself. But when he's explaining it to someone else, he has to use words that they will understand. So he used the words "day and night," even though at the time of Creation, that wouldn't have made any sense until the sun was created. [ENDQ] [NEWLINE] So you believe the creation story in genesis is an allegory?</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're not arguing from any principle. You're simply arguing logistics, and once again, difficulty of implementation is not a good argument to deny people rights.<mask><mask> we would both agree on that. [NEWLINE] [NEWLINE] <mask>,<mask> you are<mask> worried about logistics, then we'll hire the best legal minds in the country and go from there. A new legal framework would then naturally be built up over time. [NEWLINE] </s>
Label encoding: <s>You're not arguing from any principle. You're simply arguing logistics, and once again, difficulty of implementation is not a good argument to deny people rights. I think we would both agree on that. [NEWLINE] [NEWLINE] However, if you are so worried about logistics, then we'll hire the best legal minds in the country and go from there. A new legal framework would then naturally be built up over time. [NEWLINE] </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That isn't a homosexual specific experience. [ENDQ] [NEWLINE] No,<mask> homosexuals experience an *order of magnitude more* people being assholes to them.  Source:  I have gay friends and hear them talk about this shit all the time. [NEWLINE] [NEWLINE] Quit drawing a false equivalence and acting like just<mask> straight people have to deal with assholes that somehow we're in the same boat<mask> them.</s>
Label encoding: <s> [STARTQ] That isn't a homosexual specific experience. [ENDQ] [NEWLINE] No, but homosexuals experience an *order of magnitude more* people being assholes to them.  Source:  I have gay friends and hear them talk about this shit all the time. [NEWLINE] [NEWLINE] Quit drawing a false equivalence and acting like just because straight people have to deal with assholes that somehow we're in the same boat as them.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>True, their brains may be simple,<mask> of their reflexes are quick, then you're still screwed. [NEWLINE] [NEWLINE] Flies are simple. It's easy to outwit them,<mask> they can process that you're trying to hit them and flee before your hand smacks the table. [NEWLINE] [NEWLINE] Let's say a dinosaur armed with sharp teeth and claws has very quick reaction times. Little more complicated.</s>
Label encoding: <s>True, their brains may be simple, but of their reflexes are quick, then you're still screwed. [NEWLINE] [NEWLINE] Flies are simple. It's easy to outwit them, but they can process that you're trying to hit them and flee before your hand smacks the table. [NEWLINE] [NEWLINE] Let's say a dinosaur armed with sharp teeth and claws has very quick reaction times. Little more complicated.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No offense taken. I was<mask> brought up by pretty religious parents and family. I am mostly Deistic now,<mask> I mean the lifestyle stays, ya know? [NEWLINE] I've never said I didn't indulge in these habits.<mask><mask> I stated in OP that I have. Even with the perspective gained from it I still never understood it. I guess that's just<mask> it isn't for me.</s>
Label encoding: <s>No offense taken. I was indeed brought up by pretty religious parents and family. I am mostly Deistic now, but I mean the lifestyle stays, ya know? [NEWLINE] I've never said I didn't indulge in these habits. In fact I stated in OP that I have. Even with the perspective gained from it I still never understood it. I guess that's just because it isn't for me.</s>
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0662, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Honestly, I usually don't even try with 9/11 Truthers simply<mask> they demonstrate poor scientific method.** [NEWLINE] [NEWLINE] <mask> you have a quick second could you explain for 12 tonnes of titanium (engines) didn't leave a scratch on the pentagon?<mask> at all possible could you demonstrate the proper scientific method on this one... you know... just<mask> everyone is clear. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>**Honestly, I usually don't even try with 9/11 Truthers simply because they demonstrate poor scientific method.** [NEWLINE] [NEWLINE] If you have a quick second could you explain for 12 tonnes of titanium (engines) didn't leave a scratch on the pentagon? If at all possible could you demonstrate the proper scientific method on this one... you know... just so everyone is clear. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yell at the black people to shut up. I am a white person and<mask> I yell at someone to shut up who happens to be black, they will call me racist. Only black people can yell at other black people, and please do<mask>. I cannot stand it that lower class black people can be<mask> rude<mask> hell,<mask><mask> I white person yells at them they are called racist.</s><pad>
Label encoding: <s>Yell at the black people to shut up. I am a white person and if I yell at someone to shut up who happens to be black, they will call me racist. Only black people can yell at other black people, and please do so. I cannot stand it that lower class black people can be as rude as hell, but if I white person yells at them they are called racist.</s><pad>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not everyone is allowed to choose to serve - it was only recently that don't ask don't tell was repealed, and people with disabilities are are not always eligible to join the military. [NEWLINE] [NEWLINE] People born with disabilities do not choose to be born with them, and<mask> a consequence they are unable to receive this discount. In that sense, veterans discounts<mask> hold an implicit bias against the disabled community.</s>
Label encoding: <s>Not everyone is allowed to choose to serve - it was only recently that don't ask don't tell was repealed, and people with disabilities are are not always eligible to join the military. [NEWLINE] [NEWLINE] People born with disabilities do not choose to be born with them, and as a consequence they are unable to receive this discount. In that sense, veterans discounts also hold an implicit bias against the disabled community.</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just a quick question to clear up your view:<mask> are you attracted to ISIS and European nationalists? In my experience, European nationalists still have materialistic goals, they just feel that those goals are being threatened by foreigners/the EU. And ISIS might not have materialistic goals,<mask> it's not the case that any organisation that has non-materialistic goals will do for you, is it?</s>
Label encoding: <s>Just a quick question to clear up your view: Why are you attracted to ISIS and European nationalists? In my experience, European nationalists still have materialistic goals, they just feel that those goals are being threatened by foreigners/the EU. And ISIS might not have materialistic goals, but it's not the case that any organisation that has non-materialistic goals will do for you, is it?</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why? [NEWLINE] [NEWLINE] <mask>, put yourself in their shoes: they just won a defensive war against surrounding Arab nations, and now a population is trying to enter this state. The enemy population and the entering population are *virtually indistinguishable*,<mask> that it is impossible on an individual basis to tell them apart.<mask> can the new state know who to let in? That's a horrific security risk.</s>
Label encoding: <s>Why? [NEWLINE] [NEWLINE] Also, put yourself in their shoes: they just won a defensive war against surrounding Arab nations, and now a population is trying to enter this state. The enemy population and the entering population are *virtually indistinguishable*, so that it is impossible on an individual basis to tell them apart. How can the new state know who to let in? That's a horrific security risk.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0762, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People are just noticing the general trend.  20 years is too short for China to surpass the US in terms of military power, and China's current economic growth needs to perform a soft landing to avoid a very painful economic crash.  They<mask> have a severe demographic crisis. <mask>,<mask> they can overcome their many issues, they have great potential within the next 20-50 years. </s>
Label encoding: <s>People are just noticing the general trend.  20 years is too short for China to surpass the US in terms of military power, and China's current economic growth needs to perform a soft landing to avoid a very painful economic crash.  They also have a severe demographic crisis.  However, if they can overcome their many issues, they have great potential within the next 20-50 years. </s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's the definition I go by and<mask> I've seen<mask> the basis for most PC ideas. Surely there will always be people who identify<mask> being PC<mask> say nasty things about other cultures under the guise of being progressive,<mask> more often than not, these people are speaking out of anger and frustration. They aren't great examples,<mask> every philosophy has its not-<mask> -star students.</s>
Label encoding: <s>It's the definition I go by and what I've seen as the basis for most PC ideas. Surely there will always be people who identify as being PC but say nasty things about other cultures under the guise of being progressive, but more often than not, these people are speaking out of anger and frustration. They aren't great examples, but every philosophy has its not- so -star students.</s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> people pirate the media rather than buy it, [ENDQ] [NEWLINE] Not everybody who pirates would have bought the product without the ability to pirate it. They would have just gone without,<mask> the creators wouldn't have gotten their money either way. [NEWLINE] [NEWLINE] <mask> somebody pirates (who would have otherwise not consumed the media at all) they can tell others about the media who then might buy it. </s>
Label encoding: <s> [STARTQ] When people pirate the media rather than buy it, [ENDQ] [NEWLINE] Not everybody who pirates would have bought the product without the ability to pirate it. They would have just gone without, therefore the creators wouldn't have gotten their money either way. [NEWLINE] [NEWLINE] When somebody pirates (who would have otherwise not consumed the media at all) they can tell others about the media who then might buy it. </s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My doctor expressed some concern about this and was thinking about adjusting my HRT. [ENDQ] [NEWLINE] Did your doctor not know you were trans? [NEWLINE] [NEWLINE] <mask>, ∆  for this part: [NEWLINE] &gt;I'm a trans woman,<mask> my medical records have me<mask> male.<mask> I get blood analyzed, the results are returned with comparisons for the male-typical range.</s>
Label encoding: <s> [STARTQ] My doctor expressed some concern about this and was thinking about adjusting my HRT. [ENDQ] [NEWLINE] Did your doctor not know you were trans? [NEWLINE] [NEWLINE] Also, ∆  for this part: [NEWLINE] &gt;I'm a trans woman, but my medical records have me as male. When I get blood analyzed, the results are returned with comparisons for the male-typical range.</s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I suppose<mask> you need to be exact, imperial units have global standards too,<mask> my point was that is<mask> you're speaking **casually**, it seems to me that imperial whole units fit their subject matter better.  Like, "last night we got a foot of snow."  Not precisely,<mask> it just rolls off the tongue and that's all you need sometimes.</s>
Label encoding: <s>Well I suppose if you need to be exact, imperial units have global standards too, but my point was that is if you're speaking **casually**, it seems to me that imperial whole units fit their subject matter better.  Like, "last night we got a foot of snow."  Not precisely, but it just rolls off the tongue and that's all you need sometimes.</s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>These are my thoughts exactly. We<mask> a culture determined that breasts are sexual and<mask> made them illegal. This argument can be used in favor of the burqa, and it can<mask> be used against same sex marriages and other civil rights issues. Just<mask> we determined that something isn't okay based on our moral standards, doesn't mean we can impose these moral standards over everyone else.</s>
Label encoding: <s>These are my thoughts exactly. We as a culture determined that breasts are sexual and therefore made them illegal. This argument can be used in favor of the burqa, and it can also be used against same sex marriages and other civil rights issues. Just because we determined that something isn't okay based on our moral standards, doesn't mean we can impose these moral standards over everyone else.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Each people have a different experience in life. I know many environmental engineers, some are corporate drones working to skirt regulations and other work in NGO's next to other environment studies majors, biologists, naturalists, and many other people who have a passion for for making a difference. You are entitled to your opinion,<mask> don't think all of reality reflects your limited experience.   </s>
Label encoding: <s>Each people have a different experience in life. I know many environmental engineers, some are corporate drones working to skirt regulations and other work in NGO's next to other environment studies majors, biologists, naturalists, and many other people who have a passion for for making a difference. You are entitled to your opinion, but don't think all of reality reflects your limited experience.   </s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] it seems you want to change the definition of bigot and prejudice. [ENDQ] [NEWLINE] I have already cited the wikipedia page. There is not one definition of "bigot."<mask> generally, yes, I get to use words<mask> I wish to use them. That is<mask> language works. Attempts to ground language in absolute definitions failed long ago. Words acquire meaning by their use.</s>
Label encoding: <s> [STARTQ] it seems you want to change the definition of bigot and prejudice. [ENDQ] [NEWLINE] I have already cited the wikipedia page. There is not one definition of "bigot." But generally, yes, I get to use words however I wish to use them. That is how language works. Attempts to ground language in absolute definitions failed long ago. Words acquire meaning by their use.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>More to the point, political correctitude presupposes that it's possible to catalogue all offensive expression and bottle it up. [NEWLINE] [NEWLINE] It isn't, and it isn't. It's like trying to rake leaves on a windy day. A lot of fuss and bother, and no little frustration, for no constructive end. A well-intended goal prosecuted by hopeless tactics. [NEWLINE] </s>
Label encoding: <s>More to the point, political correctitude presupposes that it's possible to catalogue all offensive expression and bottle it up. [NEWLINE] [NEWLINE] It isn't, and it isn't. It's like trying to rake leaves on a windy day. A lot of fuss and bother, and no little frustration, for no constructive end. A well-intended goal prosecuted by hopeless tactics. [NEWLINE] </s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I recognize most of<mask> you mentioned<mask> problematic<mask> I don't agree that it has gotten worse over time. In most of those areas we have had those problems many many years.<mask> anything has changed it has been the transparency of media organizations, governments, and corporations. [NEWLINE] [NEWLINE] Things were still terribly biased and corrupt in the 60's<mask> it was way easier to hide it.</s>
Label encoding: <s>I recognize most of what you mentioned as problematic but I don't agree that it has gotten worse over time. In most of those areas we have had those problems many many years. If anything has changed it has been the transparency of media organizations, governments, and corporations. [NEWLINE] [NEWLINE] Things were still terribly biased and corrupt in the 60's but it was way easier to hide it.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Polyamorous people can just<mask> well marry people they love.<mask> only one at a time. [ENDQ] [NEWLINE] <mask>, one week I marry the first guy I love, the second week I marry the second guy....and just rotate my whole life?  Or marry guy A for the first half of my life...divorce..and marry Guy B the rest of it?</s>
Label encoding: <s> [STARTQ] Polyamorous people can just as well marry people they love. But only one at a time. [ENDQ] [NEWLINE] So, one week I marry the first guy I love, the second week I marry the second guy....and just rotate my whole life?  Or marry guy A for the first half of my life...divorce..and marry Guy B the rest of it?</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0789, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Eeek, that's... That's not<mask> it was intended to be consumed. [NEWLINE] [NEWLINE] I mean, to a certain extent, art (food included) is subjective,<mask><mask> you go look at the mona lisa through inverted Neon-Yellow tinted glasses, Leonardo Da Vinci's gonna be like "that's not<mask> you're meant to experience it!"</s>
Label encoding: <s>Eeek, that's... That's not how it was intended to be consumed. [NEWLINE] [NEWLINE] I mean, to a certain extent, art (food included) is subjective, but if you go look at the mona lisa through inverted Neon-Yellow tinted glasses, Leonardo Da Vinci's gonna be like "that's not how you're meant to experience it!"</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Knowing that "sch" = "sh" is very basic, and come on, roth = rot = red, and schild = shield... It's trivial. [ENDQ] [NEWLINE] I could cherry-pick a phrase/surname from every language on the planet and make the same statement. Do you know every language on earth? Can you analyse their surnames?</s>
Label encoding: <s> [STARTQ] Knowing that "sch" = "sh" is very basic, and come on, roth = rot = red, and schild = shield... It's trivial. [ENDQ] [NEWLINE] I could cherry-pick a phrase/surname from every language on the planet and make the same statement. Do you know every language on earth? Can you analyse their surnames?</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The feminist cause isn't to manufacture rape culture or "prove" that some women are harassed/some men harass women. Therefor posing<mask> a man to harass a woman doesn't help the feminist cause for equality. It would create a media stir,<mask><mask> the goal of feminism isn't to create media stirs, that wouldn't advance feminism's goals. </s>
Label encoding: <s>The feminist cause isn't to manufacture rape culture or "prove" that some women are harassed/some men harass women. Therefor posing as a man to harass a woman doesn't help the feminist cause for equality. It would create a media stir, but since the goal of feminism isn't to create media stirs, that wouldn't advance feminism's goals. </s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many cities have people going 40mph+ through the downtown areas,<mask> that's<mask> we need lights and such. Some cities are better, like in DC, you'll rarely go over 40mph, it's more like 10-20mph in the city. I like<mask> in Asia/Europe, they don't have all these lights, people just make it work.</s>
Label encoding: <s>Many cities have people going 40mph+ through the downtown areas, so that's why we need lights and such. Some cities are better, like in DC, you'll rarely go over 40mph, it's more like 10-20mph in the city. I like how in Asia/Europe, they don't have all these lights, people just make it work.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ahh, except after you save the people dying of thirst, you have to add at least some of them to the people dying of hunger! Better to not bother helping anyone,<mask><mask>. It's all that "Give a mouse a cookie" nonsense... [NEWLINE] [NEWLINE] "<mask> you give someone clean water, then they'll be able to tell you they're hungry..."</s>
Label encoding: <s>Ahh, except after you save the people dying of thirst, you have to add at least some of them to the people dying of hunger! Better to not bother helping anyone, I think. It's all that "Give a mouse a cookie" nonsense... [NEWLINE] [NEWLINE] " If you give someone clean water, then they'll be able to tell you they're hungry..."</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You ignore the possibility of the emigration of blacks to America, which was obviously stunted by slavery.  Who's to say<mask> more or less blacks would be in America and<mask> the quality of life would differ in Africa and America had slavery not occurred?  We simply don't know,<mask> saying that slavery benefitted the slaves' descendants is preposterous.</s><pad><pad>
Label encoding: <s>You ignore the possibility of the emigration of blacks to America, which was obviously stunted by slavery.  Who's to say if more or less blacks would be in America and how the quality of life would differ in Africa and America had slavery not occurred?  We simply don't know, so saying that slavery benefitted the slaves' descendants is preposterous.</s><pad><pad>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wife asks me to pass the TV remote control that is just 4 feet away from her.  I tell her to get it herself.  She says "I'm too lazy to get up... pleeezee" [NEWLINE] [NEWLINE] <mask> is that<mask> laziness?  And<mask> can you judge her actions<mask> not being lazy<mask> she explicitly says she is lazy?</s>
Label encoding: <s>Wife asks me to pass the TV remote control that is just 4 feet away from her.  I tell her to get it herself.  She says "I'm too lazy to get up... pleeezee" [NEWLINE] [NEWLINE] What is that besides laziness?  And how can you judge her actions as not being lazy when she explicitly says she is lazy?</s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Reddit should be free and open and there should be *less* regulation.<mask> it is you can get banned for damn near anything<mask> of the SJW mod cabal that runs the whole site.<mask> you want to live in a sheltered fantasy world<mask> no one disagrees with you then shut off the internet, cancel your phone service, break your TV and stay inside.</s>
Label encoding: <s>Reddit should be free and open and there should be *less* regulation. As it is you can get banned for damn near anything because of the SJW mod cabal that runs the whole site. If you want to live in a sheltered fantasy world where no one disagrees with you then shut off the internet, cancel your phone service, break your TV and stay inside.</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.8555e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[Here is 2004]( [URL].0.html) - John Kerry won the black vote with 88%. [NEWLINE] [NEWLINE] I'm having trouble finding stats from before this,<mask> really this started with the [southern strategy]( [URL] )<mask> the republican party made a calculated decision to forgo the smaller black vote in hopes of winning a much larger southern white vote. [NEWLINE] </s>
Label encoding: <s>[Here is 2004]( [URL].0.html) - John Kerry won the black vote with 88%. [NEWLINE] [NEWLINE] I'm having trouble finding stats from before this, but really this started with the [southern strategy]( [URL] ) where the republican party made a calculated decision to forgo the smaller black vote in hopes of winning a much larger southern white vote. [NEWLINE] </s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0626, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Kolomgorov complexity is the "perfect" way to measure entropy,<mask> I believe it's not computable.  There are ways to do a reasonable job,<mask>. [NEWLINE] [NEWLINE] One way is to compress a dictionary followed by the password; the number of additional bits required to store the password is a good approximation of the entropy contained in the password.</s><pad>
Label encoding: <s>Kolomgorov complexity is the "perfect" way to measure entropy, but I believe it's not computable.  There are ways to do a reasonable job, however. [NEWLINE] [NEWLINE] One way is to compress a dictionary followed by the password; the number of additional bits required to store the password is a good approximation of the entropy contained in the password.</s><pad>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0656, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>but that is the same now for any type of voting. [NEWLINE] [NEWLINE] The thing is, people wouldn't need to vote on every issue. They could just vote on issues that they cared more about or that they were more involved with. [NEWLINE] [NEWLINE] <mask>, campaigners would be bound to have tv campaigns on issues, or poster campaigns just like they do now. </s>
Label encoding: <s>but that is the same now for any type of voting. [NEWLINE] [NEWLINE] The thing is, people wouldn't need to vote on every issue. They could just vote on issues that they cared more about or that they were more involved with. [NEWLINE] [NEWLINE] Also, campaigners would be bound to have tv campaigns on issues, or poster campaigns just like they do now. </s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0730, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And preventing people who could be productive taxpayers from killing themselves allows those people to continue contributing to society in the form of productivity and taxes. They're similar on the fiscal side in that legalized suicide leads to reduced revenue,<mask> a lack of seatbelt laws lads to increased expenditures. They're similar on the moral side in that both prevent people from killing themselves.</s>
Label encoding: <s>And preventing people who could be productive taxpayers from killing themselves allows those people to continue contributing to society in the form of productivity and taxes. They're similar on the fiscal side in that legalized suicide leads to reduced revenue, while a lack of seatbelt laws lads to increased expenditures. They're similar on the moral side in that both prevent people from killing themselves.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Look at<mask> well it works for sex offenders. Or even convicts in general. There are websites that will just publish your face and background (to ruin your reputation,<mask> someone googles you like say for a job) and demand money to take it down. [NEWLINE] [NEWLINE] Imagine same with aids. An already pariah disease will be<mask> much worse.</s>
Label encoding: <s>Look at how well it works for sex offenders. Or even convicts in general. There are websites that will just publish your face and background (to ruin your reputation, if someone googles you like say for a job) and demand money to take it down. [NEWLINE] [NEWLINE] Imagine same with aids. An already pariah disease will be so much worse.</s>
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0699, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Also inequality is a natural product of growth. Here is a question on inequality: Would you rather real GDP double for every single person on earth?(would greatly increase inequality) Or would you rather that real GDP today is equally divided out between every person on earth? [NEWLINE] [NEWLINE] I'm just curious to know your answer<mask> you think inequality is that important. </s><pad>
Label encoding: <s>Also inequality is a natural product of growth. Here is a question on inequality: Would you rather real GDP double for every single person on earth?(would greatly increase inequality) Or would you rather that real GDP today is equally divided out between every person on earth? [NEWLINE] [NEWLINE] I'm just curious to know your answer if you think inequality is that important. </s><pad>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>At the age of 17, yeah.<mask><mask> you're going to give $100k away, you should let a teenager make that decision. [NEWLINE] [NEWLINE] I accept that I made the choice, I just wish I didn't have a choice to make. I didn't even need a co-signer, they just said, "Yeah, sure". </s>
Label encoding: <s>At the age of 17, yeah. Because if you're going to give $100k away, you should let a teenager make that decision. [NEWLINE] [NEWLINE] I accept that I made the choice, I just wish I didn't have a choice to make. I didn't even need a co-signer, they just said, "Yeah, sure". </s>
Loss: tensor(0.1089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If he removes our humanness, do we not more or less become demigods? Or<mask> we become like him<mask> seperates us from him? Just a bunch of omnipotent all knowing intelligences, working in unison,<mask> is the difference between that and a single god acting alone omnipotently, omnipresently, and all knowingly?</s><pad>
Label encoding: <s>If he removes our humanness, do we not more or less become demigods? Or if we become like him what seperates us from him? Just a bunch of omnipotent all knowing intelligences, working in unison, what is the difference between that and a single god acting alone omnipotently, omnipresently, and all knowingly?</s><pad>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why not flip that and apply it to western products? [NEWLINE] [NEWLINE] The US is involved in shady operations all over the world. Would you support american products bearing the label "Made by the CIA-funding, brown people killers of World Police USA" [NEWLINE] [NEWLINE] <mask> that's exactly<mask> you are going to get by pushing this kind of ridiculous name calling.</s>
Label encoding: <s>Why not flip that and apply it to western products? [NEWLINE] [NEWLINE] The US is involved in shady operations all over the world. Would you support american products bearing the label "Made by the CIA-funding, brown people killers of World Police USA" [NEWLINE] [NEWLINE] Because that's exactly what you are going to get by pushing this kind of ridiculous name calling.</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would add to this that even<mask> Russia and China were to give up their nukes too, the lack of any western nuclear deterrence would still make them a lot more likely to take aggressive actions with conventional military forces. Even<mask> they were defeated, the costs and casualties of such a war make our present state of pax atomica look pretty good.</s>
Label encoding: <s>I would add to this that even if Russia and China were to give up their nukes too, the lack of any western nuclear deterrence would still make them a lot more likely to take aggressive actions with conventional military forces. Even if they were defeated, the costs and casualties of such a war make our present state of pax atomica look pretty good.</s>
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But that doesn't stop (smart) veganism from being healthy.<mask> your worried about regulating the vitamin industry, I'll back that up,<mask> it doesn't follow to say a diet is unhealthy<mask> it requires supplements. Supplements are part of the diet, and<mask> the diet satisfies all nutritional needs and avoids the bad stuff, it's healthy.</s>
Label encoding: <s>But that doesn't stop (smart) veganism from being healthy. If your worried about regulating the vitamin industry, I'll back that up, but it doesn't follow to say a diet is unhealthy if it requires supplements. Supplements are part of the diet, and if the diet satisfies all nutritional needs and avoids the bad stuff, it's healthy.</s>
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Free speech is not absolute. Political speech is considered more important to protect than other types of speech. [NEWLINE] [NEWLINE] The original justification for censoring broadcasts comes from this court case: [URL]._Pacifica_Foundation [NEWLINE] [NEWLINE] An example of political speech being considered more improtant than other types: [URL] #Revocation [NEWLINE] [NEWLINE] Edit: an overview: [URL] </s>
Label encoding: <s>Free speech is not absolute. Political speech is considered more important to protect than other types of speech. [NEWLINE] [NEWLINE] The original justification for censoring broadcasts comes from this court case: [URL]._Pacifica_Foundation [NEWLINE] [NEWLINE] An example of political speech being considered more improtant than other types: [URL] #Revocation [NEWLINE] [NEWLINE] Edit: an overview: [URL] </s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I told my wife<mask> I met her "I'll put down the seat for you,<mask> you pick it up for me." [NEWLINE] [NEWLINE] She never does,<mask> I never do. It stays<mask> it stays.  We're both fine with this. [NEWLINE] [NEWLINE] That said, I put down the seat at other people's houses out of respect. </s>
Label encoding: <s>I told my wife when I met her "I'll put down the seat for you, if you pick it up for me." [NEWLINE] [NEWLINE] She never does, so I never do. It stays where it stays.  We're both fine with this. [NEWLINE] [NEWLINE] That said, I put down the seat at other people's houses out of respect. </s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree<mask> that couldn't be more irrelevant to the discussion. Remember, we're talking about a young woman having sex with an idol of hers,<mask> she's going to let herself get giddy over it and then let herself get crushed by the reality she already knew, which is that her idol wants nothing to do with her long-term.</s>
Label encoding: <s>I agree although that couldn't be more irrelevant to the discussion. Remember, we're talking about a young woman having sex with an idol of hers, so she's going to let herself get giddy over it and then let herself get crushed by the reality she already knew, which is that her idol wants nothing to do with her long-term.</s>
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So you think it is equally<mask> valid to say "you shouldn't do that<mask> the old man in the sky will get mad and send you to a flaming pit deep under ground"<mask> it is to say "you shouldn't do that<mask> it will hurt other people?"<mask><mask> one of those phenomena can be observed and the other can not?</s>
Label encoding: <s>So you think it is equally as valid to say "you shouldn't do that because the old man in the sky will get mad and send you to a flaming pit deep under ground" as it is to say "you shouldn't do that because it will hurt other people?" Even though one of those phenomena can be observed and the other can not?</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.9081e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.4618e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.4795e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/looklistencreate. [NEWLINE] [NEWLINE] [^looklistencreate's ^delta ^history](/r/ChangeMyView/wiki/user/looklistencreate) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/looklistencreate. [NEWLINE] [NEWLINE] [^looklistencreate's ^delta ^history](/r/ChangeMyView/wiki/user/looklistencreate) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, exactly. [NEWLINE] [NEWLINE] Just for fun, one could<mask><mask> a game being popular makes it better to<mask>. A popular game has a larger community and is more likely to be supported in the future by the manufacturer.<mask> I am just making that argument for fun. In reality I just wanted to explain Occams argument a bit better.</s>
Label encoding: <s>Yes, exactly. [NEWLINE] [NEWLINE] Just for fun, one could argue that a game being popular makes it better to though. A popular game has a larger community and is more likely to be supported in the future by the manufacturer. But I am just making that argument for fun. In reality I just wanted to explain Occams argument a bit better.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sex is very emotionally charged for most people. Just sleeping with someone isn't usually enough to make them want to get married<mask> having sex regularly will form a bond that they think will last the rest of their lives. It's not the only factor<mask> it's a large one in relationships and<mask> can't be discounted<mask> a reason to get married</s>
Label encoding: <s>Sex is very emotionally charged for most people. Just sleeping with someone isn't usually enough to make them want to get married but having sex regularly will form a bond that they think will last the rest of their lives. It's not the only factor but it's a large one in relationships and therefore can't be discounted as a reason to get married</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, yes, I believe everyone is entitled to their own beliefs. That being said, his actions were radical and possibly inhumane,<mask> they still had a cause; the greater good for an Aryan race. It wasn't merciless bloodshed for selfish ideals like many other tyrants.<mask> for your condescending question, I'm 19. </s>
Label encoding: <s>Well, yes, I believe everyone is entitled to their own beliefs. That being said, his actions were radical and possibly inhumane, but they still had a cause; the greater good for an Aryan race. It wasn't merciless bloodshed for selfish ideals like many other tyrants. As for your condescending question, I'm 19. </s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> exactly is stopping from simply expanding English, or the language chosen to be THE language, to hold more words and,<mask> such, more perspectives? [ENDQ] [NEWLINE] Well for one there are more native speakers of Mandarin and Spanish in the world. Would you feel the same way<mask> one of those two were proposed<mask> the official world language?</s>
Label encoding: <s> [STARTQ] So what exactly is stopping from simply expanding English, or the language chosen to be THE language, to hold more words and, as such, more perspectives? [ENDQ] [NEWLINE] Well for one there are more native speakers of Mandarin and Spanish in the world. Would you feel the same way if one of those two were proposed as the official world language?</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.6373e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well it is a birth right<mask> by lineage not by<mask> you were born.  You could be 6th generation born there<mask><mask> your parents aren't citizens then neither are you.  You still have to naturalized. [NEWLINE] [NEWLINE] The most interesting citizenship is probably the Vatican<mask> your citizenship is purely at the discretion of the standing Pope. </s>
Label encoding: <s>Well it is a birth right but by lineage not by where you were born.  You could be 6th generation born there but if your parents aren't citizens then neither are you.  You still have to naturalized. [NEWLINE] [NEWLINE] The most interesting citizenship is probably the Vatican where your citizenship is purely at the discretion of the standing Pope. </s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think I am good at my job, and I always strive to get better at it. [NEWLINE] [NEWLINE] <mask><mask> much better? There is no upper limit. I could get much better much faster<mask> I devoted 100% of my life to work.<mask> I choose to draw a line<mask><mask><mask> there is more to life than work.</s><pad>
Label encoding: <s>I think I am good at my job, and I always strive to get better at it. [NEWLINE] [NEWLINE] But how much better? There is no upper limit. I could get much better much faster if I devoted 100% of my life to work. But I choose to draw a line because I think there is more to life than work.</s><pad>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>NASAs budget gets payloads to Mars, Jupiter, and Pluto.  A similar budget can get nuclear weapons to targets beyond low earth orbit and expand things that NASA already does. The agency I propose would mostly be concerned with planning and allocating already existing resources, not developing new technology.  We have DARPA and NASA for that.</s>
Label encoding: <s>NASAs budget gets payloads to Mars, Jupiter, and Pluto.  A similar budget can get nuclear weapons to targets beyond low earth orbit and expand things that NASA already does. The agency I propose would mostly be concerned with planning and allocating already existing resources, not developing new technology.  We have DARPA and NASA for that.</s>
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Like my reply a second ago, it would be on the artist to have felt it was produced from a point of artistic intention.  One Direction was formed and produces from a point of financial gain first and artistic expression AT BEST second, more likely third or fourth.  Which<mask><mask> causes the group<mask> a whole to lack artistic integrity.</s>
Label encoding: <s>Like my reply a second ago, it would be on the artist to have felt it was produced from a point of artistic intention.  One Direction was formed and produces from a point of financial gain first and artistic expression AT BEST second, more likely third or fourth.  Which IMO causes the group as a whole to lack artistic integrity.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to say that no, you do not have equal rights.<mask><mask><mask> in the 21st century love is the main factor of marriage not to reproduce.<mask> many non-same sex couples don't have children<mask><mask> do you believe that same-sex couples should not have this choice, by marrying who they want? [NEWLINE] </s>
Label encoding: <s>I'm going to say that no, you do not have equal rights. Because I think in the 21st century love is the main factor of marriage not to reproduce. Also many non-same sex couples don't have children so why do you believe that same-sex couples should not have this choice, by marrying who they want? [NEWLINE] </s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I guess 1984 was a supremely terrible example for that. It's just the most recent thing I've read<mask> it's fresh in my mind. I suppose with the symbology of 1984 it's unique.<mask><mask><mask> many books use sex<mask> just another plot point, it dilutes the importance of<mask>'s done skillfully.</s>
Label encoding: <s>I guess 1984 was a supremely terrible example for that. It's just the most recent thing I've read so it's fresh in my mind. I suppose with the symbology of 1984 it's unique. But when so many books use sex as just another plot point, it dilutes the importance of what's done skillfully.</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fair enough,<mask> the way I see it<mask> the woman chooses the irresponsible path she may view child support<mask> a reason or an incentive to keep the baby.<mask> a lack of incentive causes less children to be born into suffering<mask> opposed to more children having their suffering lessened somewhat, would it not be considered a better outcome? </s>
Label encoding: <s>Fair enough, but the way I see it if the woman chooses the irresponsible path she may view child support as a reason or an incentive to keep the baby. If a lack of incentive causes less children to be born into suffering as opposed to more children having their suffering lessened somewhat, would it not be considered a better outcome? </s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/arbaard. [NEWLINE] [NEWLINE] [^arbaard's ^delta ^history](/r/ChangeMyView/wiki/user/arbaard) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/arbaard. [NEWLINE] [NEWLINE] [^arbaard's ^delta ^history](/r/ChangeMyView/wiki/user/arbaard) ^| [^delta ^system ^explained](/r/ChangeMyView/wiki/DeltaBot)</s>
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Chrome and Chromium should be nearly functionally identical, one is built by Google (Chrome), the other is built by a third party from the opensource codebase (Chromium). [NEWLINE] [NEWLINE] You can get binary blobs for Linux<mask> you want true Chrome. [NEWLINE] [NEWLINE] The future channels are canary, and dev.</s>
Label encoding: <s>Chrome and Chromium should be nearly functionally identical, one is built by Google (Chrome), the other is built by a third party from the opensource codebase (Chromium). [NEWLINE] [NEWLINE] You can get binary blobs for Linux if you want true Chrome. [NEWLINE] [NEWLINE] The future channels are canary, and dev.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Additionally, I want to express that I really and truly do want someone to CMV on this.  My views are creating friction in my life<mask> they really piss my wife off.  Everywhere I turn, I see nothing<mask> full support for this shit and it really feels like I may be in the wrong here.  </s>
Label encoding: <s>Additionally, I want to express that I really and truly do want someone to CMV on this.  My views are creating friction in my life as they really piss my wife off.  Everywhere I turn, I see nothing but full support for this shit and it really feels like I may be in the wrong here.  </s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I tend to consider it in terms of "sexual compatibility". [NEWLINE] [NEWLINE] That being said, I don't think clingy-ness is an aspect of sex at all and is entirely a personality trait and<mask> such is subject to different forces aside from sexual promiscuity. It seems shortsighted to link the two together<mask> directly.</s>
Label encoding: <s>I tend to consider it in terms of "sexual compatibility". [NEWLINE] [NEWLINE] That being said, I don't think clingy-ness is an aspect of sex at all and is entirely a personality trait and as such is subject to different forces aside from sexual promiscuity. It seems shortsighted to link the two together so directly.</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Youself being a virgin holds a intengient value to you that you feel a partner needs to value for you to be equals in lifes journey. It's not even about the sex itself<mask> the ideals you hold which the moment you 'lose', with whoever significant other it is, becomes quiet honestly a non-topic.</s>
Label encoding: <s>Youself being a virgin holds a intengient value to you that you feel a partner needs to value for you to be equals in lifes journey. It's not even about the sex itself but the ideals you hold which the moment you 'lose', with whoever significant other it is, becomes quiet honestly a non-topic.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is catholic,<mask> protestant values have basically been integrated in Europe culture<mask><mask> declared religion (talking work and finance ethics here). Protestantism and Catholicism have been interacting<mask> much<mask> Luther and Calvin that I don't see a porblem in grouping them, something that for instance wouldn't be<mask> easy with Orthodoxy.</s>
Label encoding: <s>It is catholic, but protestant values have basically been integrated in Europe culture regardless of declared religion (talking work and finance ethics here). Protestantism and Catholicism have been interacting so much since Luther and Calvin that I don't see a porblem in grouping them, something that for instance wouldn't be as easy with Orthodoxy.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>An airburst actually doesn't generate appreciable fallout - the nuclear residues on their own are too finely dispersed and flung too high into the atmosphere to come down right away. Fallout happens<mask> there is a lot of dirt and debris sucked into the mushroom cloud<mask> well - these heavier particles collect fission products and fall out quickly.</s>
Label encoding: <s>An airburst actually doesn't generate appreciable fallout - the nuclear residues on their own are too finely dispersed and flung too high into the atmosphere to come down right away. Fallout happens when there is a lot of dirt and debris sucked into the mushroom cloud as well - these heavier particles collect fission products and fall out quickly.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. Homosexuals CAN reproduce.  Just not with their partners. [NEWLINE] 2. The world is overpopulated and in the event that our populations would be threatened I guarantee that for most homosexuals, logic would supersede sexuality and they'd take one for the team and have heterosexual intercourse in order to reproduce.  </s>
Label encoding: <s>1. Homosexuals CAN reproduce.  Just not with their partners. [NEWLINE] 2. The world is overpopulated and in the event that our populations would be threatened I guarantee that for most homosexuals, logic would supersede sexuality and they'd take one for the team and have heterosexual intercourse in order to reproduce.  </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What's wrong with rappers? [NEWLINE] [NEWLINE] Just kidding.  I still don't think that's a "good reason" to kill someone.  I guess we just disagree? [NEWLINE] [NEWLINE] Edit:  Rappers and Rapists are not the same thing.  One is a joke, and the other is no laughing matter.</s>
Label encoding: <s>What's wrong with rappers? [NEWLINE] [NEWLINE] Just kidding.  I still don't think that's a "good reason" to kill someone.  I guess we just disagree? [NEWLINE] [NEWLINE] Edit:  Rappers and Rapists are not the same thing.  One is a joke, and the other is no laughing matter.</s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That phrase is a caricature--I've never heard anyone say it seriously with the exception of try hard radicals on the internet and I hang around a lot of very left political types. The mere existence of that phrase does not imply that "kill yourself for being born in America" is an "extremely common" political sentiment.</s>
Label encoding: <s>That phrase is a caricature--I've never heard anyone say it seriously with the exception of try hard radicals on the internet and I hang around a lot of very left political types. The mere existence of that phrase does not imply that "kill yourself for being born in America" is an "extremely common" political sentiment.</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0794, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Everyone are different. I've seen many cases<mask> pressure only got the person more stubborn, angry and frustrated. I was one of those cases, albeit about another issue altogether,<mask> still. People are mostly protective of their little shell. The only way to get to many of those is to encourage them to open up.</s>
Label encoding: <s>Everyone are different. I've seen many cases where pressure only got the person more stubborn, angry and frustrated. I was one of those cases, albeit about another issue altogether, but still. People are mostly protective of their little shell. The only way to get to many of those is to encourage them to open up.</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And we're back to theory vs. practice. Sure... theoretically that could happen.<mask> you say, it's not very likely. Pragmatically,<mask> that approach results in laws we like better, and it's not technically dishonest or unscientific,<mask> possible objection could their be to taking that approach?</s>
Label encoding: <s>And we're back to theory vs. practice. Sure... theoretically that could happen. As you say, it's not very likely. Pragmatically, if that approach results in laws we like better, and it's not technically dishonest or unscientific, what possible objection could their be to taking that approach?</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Could the effects of chemicals like DMT or LSD be considered<mask> a possible evidence for 'God' (not necessarily the Christian God,<mask> even a pantheistic god)? Not the drugs themselves,<mask> the spiritual experience, which has been a recurring experience in the history of man? Just a thought. (:</s>
Label encoding: <s>Could the effects of chemicals like DMT or LSD be considered as a possible evidence for 'God' (not necessarily the Christian God, but even a pantheistic god)? Not the drugs themselves, but the spiritual experience, which has been a recurring experience in the history of man? Just a thought. (:</s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You fail to see my points and force me to act<mask> tactful<mask> you<mask> we can converse.<mask> my point, you want me to come to your level of discussion the same<mask> you wanting me to be tactful. You want the world to kneel before you<mask> you can feel special. </s>
Label encoding: <s>You fail to see my points and force me to act as tactful as you so we can converse. Thus my point, you want me to come to your level of discussion the same as you wanting me to be tactful. You want the world to kneel before you so you can feel special. </s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, most people do not realize that [9 out 100 on the pill still get pregnant]( [URL] /). [Condoms—<mask> used correctly and consistently—prevent pregnancy about 98 percent of the time. The typical effectiveness rate—<mask> mistakes are made or condoms break—is about 82 percent.]( [URL] /)</s>
Label encoding: <s>Yes, most people do not realize that [9 out 100 on the pill still get pregnant]( [URL] /). [Condoms— when used correctly and consistently—prevent pregnancy about 98 percent of the time. The typical effectiveness rate— where mistakes are made or condoms break—is about 82 percent.]( [URL] /)</s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I guess you are correct in saying my argument is on the "don't feed the troll" side. I promote everyone being nice to each other, and<mask><mask> that<mask> people start receiving a lot of attention for pointing out<mask> the "troll" makes noise, it starts to derail the issues a bit.</s>
Label encoding: <s>I guess you are correct in saying my argument is on the "don't feed the troll" side. I promote everyone being nice to each other, and I think that when people start receiving a lot of attention for pointing out when the "troll" makes noise, it starts to derail the issues a bit.</s>
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That is an untenable standard of historical evidence.  Jesus is far from the only figure for whom we lack contemporary sources.  We<mask> lack surviving contemporary sources for Hannibal and Alexander. <mask> even men of that stature can reach us without such sources, it's absurd to expect them for a Jewish apocalyptic prophet.</s>
Label encoding: <s>That is an untenable standard of historical evidence.  Jesus is far from the only figure for whom we lack contemporary sources.  We also lack surviving contemporary sources for Hannibal and Alexander.  If even men of that stature can reach us without such sources, it's absurd to expect them for a Jewish apocalyptic prophet.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When they are not campaigning yes. They spend their time studying or being briefed by their assistants on things. In both situations they are informed far more than the average citizen who would be voting in a direct democracy. Tyranny of the masses is a legitimate concern and<mask> the founding fathers chose a representative government. </s>
Label encoding: <s>When they are not campaigning yes. They spend their time studying or being briefed by their assistants on things. In both situations they are informed far more than the average citizen who would be voting in a direct democracy. Tyranny of the masses is a legitimate concern and why the founding fathers chose a representative government. </s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sites like reddit aren't meant to be a morally upright website. Reddit is meant to be a medium for people to use.<mask><mask> it was only<mask> activity in reddit was actually becoming illegal that anyone took any action. [NEWLINE] [NEWLINE] Reddit is a vehicle, not a managed entity<mask><mask><mask> content goes.</s>
Label encoding: <s>Sites like reddit aren't meant to be a morally upright website. Reddit is meant to be a medium for people to use. I think it was only when activity in reddit was actually becoming illegal that anyone took any action. [NEWLINE] [NEWLINE] Reddit is a vehicle, not a managed entity as far as content goes.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I highly doubt that you will find too many scientists with an average(100) IQ, The vast majority have higher than average IQ and most of the best have very high IQ. [NEWLINE] [NEWLINE] <mask> in a world filled with average intelligence,<mask> highly stable people,<mask> would our scientific advancement come from? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I highly doubt that you will find too many scientists with an average(100) IQ, The vast majority have higher than average IQ and most of the best have very high IQ. [NEWLINE] [NEWLINE] So in a world filled with average intelligence, but highly stable people, where would our scientific advancement come from? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm doing computer science at a university and I prefer to work outside the computer lab<mask> most of our work has to be done on their computers<mask> SSH is the only solution (<mask> I generally use a normal text editor instead<mask> the connection to the dorms is fast enough to do X forwarding practically). </s>
Label encoding: <s>I'm doing computer science at a university and I prefer to work outside the computer lab but most of our work has to be done on their computers so SSH is the only solution ( though I generally use a normal text editor instead since the connection to the dorms is fast enough to do X forwarding practically). </s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is a real choice. The woman has to think about whether she wants to raise the kid alone or not at all.<mask> it is not true that they are equally responsible. He/she who wants the child brought to term is the one responsible. Which in this particular scenario is only the woman.</s>
Label encoding: <s>It is a real choice. The woman has to think about whether she wants to raise the kid alone or not at all. Because it is not true that they are equally responsible. He/she who wants the child brought to term is the one responsible. Which in this particular scenario is only the woman.</s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So I would have to ask you, legality set aside<mask> I am really focused on the ethical implications  of the issue. Is there any moral difference between the government acting on this maxim and a father whose son was murdered? Are the limits of just punishment relative to who is the one dealing it out?</s>
Label encoding: <s>So I would have to ask you, legality set aside since I am really focused on the ethical implications  of the issue. Is there any moral difference between the government acting on this maxim and a father whose son was murdered? Are the limits of just punishment relative to who is the one dealing it out?</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not arbitrary. There are numerous distinctions in the law between closely held corporations and other publicly-owned corporations. It doesn't strike me<mask> weird or odd that a judge would say that a company held by a family can more clearly assert a religious belief than a company owned by thousands of stockholders.</s>
Label encoding: <s>It's not arbitrary. There are numerous distinctions in the law between closely held corporations and other publicly-owned corporations. It doesn't strike me as weird or odd that a judge would say that a company held by a family can more clearly assert a religious belief than a company owned by thousands of stockholders.</s>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But you are choosing to play a sport every time you go play. It's within your control every day. Heroin is completely out of your control, addicts will hurt family and friends to maintain said addiction.<mask> sports usually do the opposite, by bringing family and friends together for a fun bonding experience.</s>
Label encoding: <s>But you are choosing to play a sport every time you go play. It's within your control every day. Heroin is completely out of your control, addicts will hurt family and friends to maintain said addiction. While sports usually do the opposite, by bringing family and friends together for a fun bonding experience.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm from Germany.  Few people smoke here. It used to be different.<mask>, Europe is bigger than three countries and things change. There used to be many more smokers. Myself included and some of my friends who<mask> quit. [NEWLINE] [NEWLINE] It is true that smoking is allowed outside generally.</s>
Label encoding: <s>I'm from Germany.  Few people smoke here. It used to be different. So, Europe is bigger than three countries and things change. There used to be many more smokers. Myself included and some of my friends who also quit. [NEWLINE] [NEWLINE] It is true that smoking is allowed outside generally.</s>
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You broke it down pretty well. I'd be inclined to rescind my above comment,<mask> this muffin is pretty good and I have *at least* another 25 minutes of lunch break here. [NEWLINE] [NEWLINE] EDIT:<mask> I'm saying is, you made valid points and I should rethink my reasoning.</s><pad>
Label encoding: <s>You broke it down pretty well. I'd be inclined to rescind my above comment, but this muffin is pretty good and I have *at least* another 25 minutes of lunch break here. [NEWLINE] [NEWLINE] EDIT: what I'm saying is, you made valid points and I should rethink my reasoning.</s><pad>
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0731, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/franklin_wi. ^[[History](/r/changemyview/wiki/user/franklin_wi)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/franklin_wi. ^[[History](/r/changemyview/wiki/user/franklin_wi)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you implying that a company cannot operate in multiple countries?<mask> you create an engineering firm in Canada, you may want to outsource partly to US<mask> there's a much larger talent pool. You are not going to move headquarters just to get an extra 3 teams of friggin engineers. </s><pad>
Label encoding: <s>Are you implying that a company cannot operate in multiple countries? If you create an engineering firm in Canada, you may want to outsource partly to US where there's a much larger talent pool. You are not going to move headquarters just to get an extra 3 teams of friggin engineers. </s><pad>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>His discography is overrated<mask><mask> of the way he died, he is a legent. After he died, an entire generation grew up sort of worshipping him. Other than maybe hail mary, his songs are pretty forgettable and he has done some pretty terrible movies<mask> well.</s>
Label encoding: <s>His discography is overrated but because of the way he died, he is a legent. After he died, an entire generation grew up sort of worshipping him. Other than maybe hail mary, his songs are pretty forgettable and he has done some pretty terrible movies as well.</s>
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Don't forget delusional<mask> well! Ive known people who thought they were awesome, kind, and very attractive and wouldn't settle for anything less in a partner.<mask>... They were none of they things they thought they were. They could never have a relationship and couldn't figure out<mask>!</s>
Label encoding: <s>Don't forget delusional as well! Ive known people who thought they were awesome, kind, and very attractive and wouldn't settle for anything less in a partner. But... They were none of they things they thought they were. They could never have a relationship and couldn't figure out why!</s>
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So at<mask> point does it become an addiction?<mask>'s the threshold for healthy usage and addictive usage?<mask> did it take for you to self-diagnose (<mask> that's<mask> happened)?<mask> changes did you notice in your life that lead you to that conclusion and the need to change?</s>
Label encoding: <s>So at what point does it become an addiction? Where's the threshold for healthy usage and addictive usage? What did it take for you to self-diagnose ( if that's what happened)? What changes did you notice in your life that lead you to that conclusion and the need to change?</s>
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] in practise it will always be people from the same political dynasties that gets elected [ENDQ] [NEWLINE] In practice it's the people with the most money who get elected. [NEWLINE] [NEWLINE] Name recognition and begin able to lean on the expertise and connections of your family makes it easier to make money. </s>
Label encoding: <s> [STARTQ] in practise it will always be people from the same political dynasties that gets elected [ENDQ] [NEWLINE] In practice it's the people with the most money who get elected. [NEWLINE] [NEWLINE] Name recognition and begin able to lean on the expertise and connections of your family makes it easier to make money. </s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/ItIsOnlyRain. ^[[History](/r/changemyview/wiki/user/ItIsOnlyRain)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/ItIsOnlyRain. ^[[History](/r/changemyview/wiki/user/ItIsOnlyRain)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you cannot state objectively that we've actually practiced Keynesian economics in this country, (for example paying off debt with surplus in times of boom among other things). [ENDQ] [NEWLINE] We did for a<mask>, after WWII. Stagflation in the 1970s put an end to that.</s>
Label encoding: <s> [STARTQ] But you cannot state objectively that we've actually practiced Keynesian economics in this country, (for example paying off debt with surplus in times of boom among other things). [ENDQ] [NEWLINE] We did for a while, after WWII. Stagflation in the 1970s put an end to that.</s>
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason solar and wind aren't viable right now is<mask> we don't have adequate energy storage. Storing the energy produced by using batteries, producing hydrogen or hydrocarbon fuels, or even pumping water uphill, would allow use to use solar and wind and not have to worry about reliability issues</s>
Label encoding: <s>The reason solar and wind aren't viable right now is because we don't have adequate energy storage. Storing the energy produced by using batteries, producing hydrogen or hydrocarbon fuels, or even pumping water uphill, would allow use to use solar and wind and not have to worry about reliability issues</s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, in a combined bathrooms you would have the same number of stalls and urinals<mask> in both bathrooms combined,<mask> I don't really see your argument there. [NEWLINE] [NEWLINE] <mask>,<mask> would gyms or pools be especially uncomfortable for people? I wasn't talking about changing rooms. </s>
Label encoding: <s>Well, in a combined bathrooms you would have the same number of stalls and urinals as in both bathrooms combined, so I don't really see your argument there. [NEWLINE] [NEWLINE] Also, why would gyms or pools be especially uncomfortable for people? I wasn't talking about changing rooms. </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Yorubaland. ^[[History](/r/changemyview/wiki/user/Yorubaland)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Yorubaland. ^[[History](/r/changemyview/wiki/user/Yorubaland)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From a utilitarian perspective, it made a lot of people involuntarily sad and angry - doesn't that qualify?<mask><mask> I understand your point,<mask>. People were emotional, and lacked rational perspective on the actual value of a particular lion compared with the many many others that are hunted regularly.</s>
Label encoding: <s>From a utilitarian perspective, it made a lot of people involuntarily sad and angry - doesn't that qualify? I think I understand your point, though. People were emotional, and lacked rational perspective on the actual value of a particular lion compared with the many many others that are hunted regularly.</s>
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0637, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> much about that.<mask> I do feel that clear, neat writing helps the reader understand the ideas behind the words. Perhaps MORE<mask> than the ability to construct proper sentences,<mask> still important to the goal of getting the reader to understand the thought/idea.</s>
Label encoding: <s>I don't know so much about that. But I do feel that clear, neat writing helps the reader understand the ideas behind the words. Perhaps MORE so than the ability to construct proper sentences, but still important to the goal of getting the reader to understand the thought/idea.</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/redraven937. ([History](/r/changemyview/wiki/redraven937)) [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/redraven937. ([History](/r/changemyview/wiki/redraven937)) [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Dude, I don't like Tina Fey's personality, and don't think she's very funny,<mask> I still think she's attractive. She's not *incredibly* physically attractive,<mask> that's different from saying she's not physically attractive, period, end of sentence.</s>
Label encoding: <s>Dude, I don't like Tina Fey's personality, and don't think she's very funny, but I still think she's attractive. She's not *incredibly* physically attractive, but that's different from saying she's not physically attractive, period, end of sentence.</s>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your statement about keeping businesses from leaving shows a complete misunderstanding of business. You can't say we can't let them leave then<mask> say we need to make the u.s. an attractive place for business. This is substantively the same argument put forward by collectivists.</s>
Label encoding: <s>Your statement about keeping businesses from leaving shows a complete misunderstanding of business. You can't say we can't let them leave then also say we need to make the u.s. an attractive place for business. This is substantively the same argument put forward by collectivists.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Cavemonster. ^[[History](/r/changemyview/wiki/user/Cavemonster)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Cavemonster. ^[[History](/r/changemyview/wiki/user/Cavemonster)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0575, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Someone has to be mentally capable of consent in order to get married in the first place, and once they are married their interests are protected in other ways. [NEWLINE] [NEWLINE] The better comparison would be "can a 17 year old enter marriage", and in several places the answer is no.</s>
Label encoding: <s>Someone has to be mentally capable of consent in order to get married in the first place, and once they are married their interests are protected in other ways. [NEWLINE] [NEWLINE] The better comparison would be "can a 17 year old enter marriage", and in several places the answer is no.</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If it is necessary for survival then it becomes similar to a situation of being stranded on a life raft and having to eat a companion for survival. I don't think the OP is saying that you would have to respect other animals over humans in order to be an animal lover. </s>
Label encoding: <s>If it is necessary for survival then it becomes similar to a situation of being stranded on a life raft and having to eat a companion for survival. I don't think the OP is saying that you would have to respect other animals over humans in order to be an animal lover. </s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes<mask><mask> is it necessary,<mask> progressive taxation has shown through many real world examples (Putin's implementation in 2001 of a flat tax of 13% creating massive economic growth) to be an inhibitor of economic growth then<mask> is its purpose even<mask> the rich can afford it? </s>
Label encoding: <s>Yes but why is it necessary, if progressive taxation has shown through many real world examples (Putin's implementation in 2001 of a flat tax of 13% creating massive economic growth) to be an inhibitor of economic growth then what is its purpose even if the rich can afford it? </s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Suppose he *did* create an ecommerce website that functioned poorly.<mask><mask> he couldn't do better, he would still be justified in criticizing him. He didn't pay a professional for a semi-functional ecommerce website. He paid for a fully functional one.</s>
Label encoding: <s>Suppose he *did* create an ecommerce website that functioned poorly. Even though he couldn't do better, he would still be justified in criticizing him. He didn't pay a professional for a semi-functional ecommerce website. He paid for a fully functional one.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The hypocrisy would be<mask> we *think* "Someone has harmed and wronged us, and we should take action to stop it,"<mask> we *say* we are fighting for perceived American values such<mask> justice, freedom, democracy, equality, right to privacy, etc. </s>
Label encoding: <s>The hypocrisy would be when we *think* "Someone has harmed and wronged us, and we should take action to stop it," but we *say* we are fighting for perceived American values such as justice, freedom, democracy, equality, right to privacy, etc. </s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This video was made due to a specific event a few months ago,<mask> it's relevant and the guy who made it (TotalBiscuit) shares his thoughts about video game violence and any links between it and violent upbringings. [Have a look.]( [URL] )</s>
Label encoding: <s>This video was made due to a specific event a few months ago, but it's relevant and the guy who made it (TotalBiscuit) shares his thoughts about video game violence and any links between it and violent upbringings. [Have a look.]( [URL] )</s>
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I had never thought about it like that before -<mask> I guess the original view could be similar to someone saying "I dont think we should have seperate museums for jazz and rock - they should be mixed together", and you are pointing out that they are not the same thing</s>
Label encoding: <s>I had never thought about it like that before - so I guess the original view could be similar to someone saying "I dont think we should have seperate museums for jazz and rock - they should be mixed together", and you are pointing out that they are not the same thing</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You're the one who put it out there [ENDQ] [NEWLINE] <mask> you're just gunna ignore<mask> I wrote after that? [NEWLINE] [NEWLINE] The social values are all good,<mask> we learn that in the 12 years of mandatory schooling already.<mask>, it still sounds very statist.</s>
Label encoding: <s> [STARTQ] You're the one who put it out there [ENDQ] [NEWLINE] So you're just gunna ignore what I wrote after that? [NEWLINE] [NEWLINE] The social values are all good, but we learn that in the 12 years of mandatory schooling already. Also, it still sounds very statist.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason demo teams don't start fires and just let buildings collapse is<mask> they have virtually no control over the situation in that case and it is way more dangerous. It has nothing to do with the building falling differently. [NEWLINE] [NEWLINE] <mask> specifically about my post is incorrect? </s>
Label encoding: <s>The reason demo teams don't start fires and just let buildings collapse is because they have virtually no control over the situation in that case and it is way more dangerous. It has nothing to do with the building falling differently. [NEWLINE] [NEWLINE] What specifically about my post is incorrect? </s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A super majority of the elected representatives and states. There is no mechanism for revoking citizenship in the constitution, it would have to be amended. [NEWLINE] [NEWLINE] The chances of it happening are low, and they really have bigger fish to fry,<mask> the question was asked..... </s>
Label encoding: <s>A super majority of the elected representatives and states. There is no mechanism for revoking citizenship in the constitution, it would have to be amended. [NEWLINE] [NEWLINE] The chances of it happening are low, and they really have bigger fish to fry, but the question was asked..... </s>
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>that's<mask> court is for.<mask><mask><mask> the false confessions. they are extremely rare these days<mask> not almost non existent. false convictions happen more often<mask> a false confession with conviction... not<mask> much. like it our not its not going to change. </s>
Label encoding: <s>that's what court is for. as far as the false confessions. they are extremely rare these days if not almost non existent. false convictions happen more often but a false confession with conviction... not so much. like it our not its not going to change. </s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/iserane. ^[[History](/r/changemyview/wiki/user/iserane)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/iserane. ^[[History](/r/changemyview/wiki/user/iserane)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know a lot of addicts in recovery. one thing they say is that they cannot control their addiction,<mask> they CAN control their recovery. Active addicts are not at fault for HAVING an addiction,<mask> their choice not to seek help *is* their fault.</s>
Label encoding: <s>I know a lot of addicts in recovery. one thing they say is that they cannot control their addiction, but they CAN control their recovery. Active addicts are not at fault for HAVING an addiction, but their choice not to seek help *is* their fault.</s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That seems strange,<mask> these things aren't selling<mask> don't they just stock less of them? [NEWLINE] [NEWLINE] Of course a pants tax could be used to subsidize something beneficial to public health without raising the price of high calorie foods,<mask> making the hungry hungrier.</s>
Label encoding: <s>That seems strange, if these things aren't selling why don't they just stock less of them? [NEWLINE] [NEWLINE] Of course a pants tax could be used to subsidize something beneficial to public health without raising the price of high calorie foods, thus making the hungry hungrier.</s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not doing x would do y. [NEWLINE] [NEWLINE] The problem is not doing tons of different x's would do y.<mask><mask> we want to agree we need more y, you have to explain<mask> one x is better than any other. No one has done this.</s>
Label encoding: <s>Not doing x would do y. [NEWLINE] [NEWLINE] The problem is not doing tons of different x's would do y. So if we want to agree we need more y, you have to explain why one x is better than any other. No one has done this.</s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Men under-report due to the fact that they don't feel like they will get any help. Many stats exclude made to penetrate from rape statistics, saying that these are not legitimate rape,<mask> sexual assault. The number is likely much higher for male victims. </s>
Label encoding: <s>Men under-report due to the fact that they don't feel like they will get any help. Many stats exclude made to penetrate from rape statistics, saying that these are not legitimate rape, but sexual assault. The number is likely much higher for male victims. </s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>These people dont care about facts or true history, they just like to push this narrative<mask> they heard Cenk Uygur say it once on TYT and it made them feel good. [NEWLINE] [NEWLINE] <mask> matters is<mask> they feel, not<mask> is true.</s>
Label encoding: <s>These people dont care about facts or true history, they just like to push this narrative because they heard Cenk Uygur say it once on TYT and it made them feel good. [NEWLINE] [NEWLINE] What matters is how they feel, not what is true.</s>
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because literature is the best way to explore<mask> language can be used to communicate. Joyce's Ulysses, with its differing styles and voices, explores issues with a complexity that can't be conveyed simply through treatises on politics, law, and economy. </s>
Label encoding: <s>Because literature is the best way to explore how language can be used to communicate. Joyce's Ulysses, with its differing styles and voices, explores issues with a complexity that can't be conveyed simply through treatises on politics, law, and economy. </s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While<mask><mask> that no one has used it in that context before (that I've seen). I don't think it falls under straw man only<mask> you used the word "impossible". They are simply demonstrating that it IS possible by giving an example. </s>
Label encoding: <s>While I agree that no one has used it in that context before (that I've seen). I don't think it falls under straw man only because you used the word "impossible". They are simply demonstrating that it IS possible by giving an example. </s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Women don't have either of those rights.  Men have total autonomy and the ability to decide both things for themselves; don't want to be responsible for a child?  Don't have sex.  Pretty simple stuff, and it's 100% effective.</s>
Label encoding: <s>Women don't have either of those rights.  Men have total autonomy and the ability to decide both things for themselves; don't want to be responsible for a child?  Don't have sex.  Pretty simple stuff, and it's 100% effective.</s>
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Also, there's a big focus on the instrumentation in metal, especially with the more extreme stuff.  The harsh vocals were - in a lot of ways - a reminder that hey, there's things<mask> singing in music that you can focus on too!</s>
Label encoding: <s>Also, there's a big focus on the instrumentation in metal, especially with the more extreme stuff.  The harsh vocals were - in a lot of ways - a reminder that hey, there's things besides singing in music that you can focus on too!</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All of the ones<mask><mask> of are ones to poke holes in something that is otherwise correct. [NEWLINE] [NEWLINE] I'd be open to being proven wrong<mask> it would show holes in my process. That's<mask> I'd want out of a CMV. </s>
Label encoding: <s>All of the ones I think of are ones to poke holes in something that is otherwise correct. [NEWLINE] [NEWLINE] I'd be open to being proven wrong as it would show holes in my process. That's what I'd want out of a CMV. </s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The world is full of selfish, foolish, ill-mannered, self-centered, narcissistic, hypocritical people.  That is our greatest weakness<mask> a species, and one which we will perhaps never be able to overcome.  Depressing<mask> true.</s>
Label encoding: <s>The world is full of selfish, foolish, ill-mannered, self-centered, narcissistic, hypocritical people.  That is our greatest weakness as a species, and one which we will perhaps never be able to overcome.  Depressing but true.</s>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your point is that women should be okay with paternity testing<mask> sometimes men end up claiming paternity on children that aren't theres and shouldn't be willing to break up with them for it. [ENDQ] [NEWLINE] No. please go back and read my title.</s>
Label encoding: <s> [STARTQ] Your point is that women should be okay with paternity testing because sometimes men end up claiming paternity on children that aren't theres and shouldn't be willing to break up with them for it. [ENDQ] [NEWLINE] No. please go back and read my title.</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's a place for learning.  Not for debate.  You go there<mask> you have an open mind about the subject and want to learn more, not to try and change other people's opinions. [NEWLINE] [NEWLINE] That's<mask> this subreddit is for.</s><pad>
Label encoding: <s>It's a place for learning.  Not for debate.  You go there if you have an open mind about the subject and want to learn more, not to try and change other people's opinions. [NEWLINE] [NEWLINE] That's what this subreddit is for.</s><pad>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks and you're welcome. [NEWLINE] [NEWLINE] Now,<mask> my brain continues down this train of thought, I'm wondering<mask> it's ethical to breed with someone who has a known history of addiction or obesity. Kind of like intentionally having an autistic child. </s>
Label encoding: <s>Thanks and you're welcome. [NEWLINE] [NEWLINE] Now, as my brain continues down this train of thought, I'm wondering if it's ethical to breed with someone who has a known history of addiction or obesity. Kind of like intentionally having an autistic child. </s>
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When I use a bathroom, I sometimes make it smelly.<mask> I was succeeded by a person of opposite gender, I would feel much more embarrassed then I do now. [NEWLINE] [NEWLINE] The same goes with sounds that sometimes occur<mask> using a bathroom.</s>
Label encoding: <s>When I use a bathroom, I sometimes make it smelly. If I was succeeded by a person of opposite gender, I would feel much more embarrassed then I do now. [NEWLINE] [NEWLINE] The same goes with sounds that sometimes occur while using a bathroom.</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If many poor people spending all of their money on neccessities didn't create jobs, walmart wouldn't be the biggest employer in the country. The problem is that the jobs being created are only in industries poor people can afford to utilize. </s>
Label encoding: <s>If many poor people spending all of their money on neccessities didn't create jobs, walmart wouldn't be the biggest employer in the country. The problem is that the jobs being created are only in industries poor people can afford to utilize. </s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Relevant: [Everything is a Remix Part 3 - Apple]( [URL] ) [NEWLINE] [NEWLINE] Edit: the Apple part begins at [3:00]( [URL] ;v=wq5D43qAsVg&amp;t=180) Min</s>
Label encoding: <s>Relevant: [Everything is a Remix Part 3 - Apple]( [URL] ) [NEWLINE] [NEWLINE] Edit: the Apple part begins at [3:00]( [URL] ;v=wq5D43qAsVg&amp;t=180) Min</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0615, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I guess I was taking more of an issue with<mask> your premise was worded.<mask> you took a middle position, there isn't much to be changed. Try something like "CMV: The current level of police militarization is not problematic."</s>
Label encoding: <s>I guess I was taking more of an issue with how your premise was worded. Because you took a middle position, there isn't much to be changed. Try something like "CMV: The current level of police militarization is not problematic."</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You don't know<mask> atheist means.  There are many definitions,<mask> none are<mask> you're implying. Which is to say that one does not have to be certain of anything to be an atheist.  It's just not in the definition.</s>
Label encoding: <s>You don't know what atheist means.  There are many definitions, but none are what you're implying. Which is to say that one does not have to be certain of anything to be an atheist.  It's just not in the definition.</s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There isn't really that much range in terms of male representation either. Most protagonists are voice-less entities with little to no personality. Virtually any game<mask> you can customize a character the way you want<mask> has the option to change genders.</s>
Label encoding: <s>There isn't really that much range in terms of male representation either. Most protagonists are voice-less entities with little to no personality. Virtually any game where you can customize a character the way you want also has the option to change genders.</s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>lol do they? I guess gun manufactures do nothing for the economy.<mask> deaths by automobiles are acceptable losses? You threw out a statistic and i threw out a statistic. Fact automobiles harm people much more the firearms. Again<mask> are firearms impractical?</s>
Label encoding: <s>lol do they? I guess gun manufactures do nothing for the economy. So deaths by automobiles are acceptable losses? You threw out a statistic and i threw out a statistic. Fact automobiles harm people much more the firearms. Again how are firearms impractical?</s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A corporation is a group of people working towards a common purpose of some kind, be it charitable, political, economic, etc.  Are you asserting that a group of people has or should have fewer rights than an individual person? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>A corporation is a group of people working towards a common purpose of some kind, be it charitable, political, economic, etc.  Are you asserting that a group of people has or should have fewer rights than an individual person? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>True,<mask> couldn't the argument be made that the reason the likelihood is low is<mask> of the social convention that the seat remain down? And even<mask> it wasn't the likelihood would be higher and<mask><mask> would the culpability? [NEWLINE] </s>
Label encoding: <s>True, however couldn't the argument be made that the reason the likelihood is low is because of the social convention that the seat remain down? And even if it wasn't the likelihood would be higher and therefore so would the culpability? [NEWLINE] </s>
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most minority groups are okay<mask> we make the jokes to each other. I will make an occasional Jewish joke with my Jewish friends<mask> it's fine<mask> we are all Jewish. I don't go around making black jokes or Asian jokes. </s>
Label encoding: <s>Most minority groups are okay when we make the jokes to each other. I will make an occasional Jewish joke with my Jewish friends but it's fine because we are all Jewish. I don't go around making black jokes or Asian jokes. </s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1.) Earth doesn't have Zombies made out of ice [NEWLINE] 2.) earth doesn't have mile high ice wall older than 100,000 years [NEWLINE] 3.) Earth doesn't have climate change of multi-century winters and decade long summers. [NEWLINE] </s>
Label encoding: <s>1.) Earth doesn't have Zombies made out of ice [NEWLINE] 2.) earth doesn't have mile high ice wall older than 100,000 years [NEWLINE] 3.) Earth doesn't have climate change of multi-century winters and decade long summers. [NEWLINE] </s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. it takes 2 people to make a baby, the baby is<mask> much the mans responsibility<mask> the women's [NEWLINE] 2. <mask> we want men and women to be treated equally we can't make laws that discriminate between genders.</s>
Label encoding: <s>1. it takes 2 people to make a baby, the baby is as much the mans responsibility as the women's [NEWLINE] 2.  if we want men and women to be treated equally we can't make laws that discriminate between genders.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] At least<mask> I am (Boston area) [ENDQ] [NEWLINE] I went to Boston for the first time a few weeks ago and driving through there was insanity. Things obviously don't mean the same thing in Boston<mask> they do in VA...</s>
Label encoding: <s> [STARTQ] At least where I am (Boston area) [ENDQ] [NEWLINE] I went to Boston for the first time a few weeks ago and driving through there was insanity. Things obviously don't mean the same thing in Boston as they do in VA...</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0672, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/silvy96. [NEWLINE] [NEWLINE] ^^[ [^^Awardee's ^^History](/r/changemyview/wiki/user/silvy96) ^^]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/silvy96. [NEWLINE] [NEWLINE] ^^[ [^^Awardee's ^^History](/r/changemyview/wiki/user/silvy96) ^^]</s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It sounds like they're getting photos of them *<mask> a couple*.<mask> yes, you could probably get a 'gay couple' photo with another person of your sex,<mask> it wouldn't be much of a couple's photo.</s>
Label encoding: <s>It sounds like they're getting photos of them * as a couple*. So yes, you could probably get a 'gay couple' photo with another person of your sex, but it wouldn't be much of a couple's photo.</s>
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The US is the most powerful country in the world. Its decisions directly affect people, not just in Britain,<mask> around the world. [NEWLINE] [NEWLINE] It is perfectly reasonable to be angry with anyone who makes decisions that negatively impact you.</s>
Label encoding: <s>The US is the most powerful country in the world. Its decisions directly affect people, not just in Britain, but around the world. [NEWLINE] [NEWLINE] It is perfectly reasonable to be angry with anyone who makes decisions that negatively impact you.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Racing games? [NEWLINE] [NEWLINE] Using WASD for steering and having no way to apply variable acceleration really sucks. [NEWLINE] [NEWLINE] Fighting games are similar. [NEWLINE] [NEWLINE] <mask><mask> do wheels/hotas/fighting sticks figure into your view?</s>
Label encoding: <s>Racing games? [NEWLINE] [NEWLINE] Using WASD for steering and having no way to apply variable acceleration really sucks. [NEWLINE] [NEWLINE] Fighting games are similar. [NEWLINE] [NEWLINE] Also how do wheels/hotas/fighting sticks figure into your view?</s>
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A CMV. [NEWLINE] [NEWLINE] The heart of my position is "By doing<mask>, they run a grave risk of alienating friends, family, or coworkers. There is no commeasurate benefit" to balance this risk.</s>
Label encoding: <s>A CMV. [NEWLINE] [NEWLINE] The heart of my position is "By doing so, they run a grave risk of alienating friends, family, or coworkers. There is no commeasurate benefit" to balance this risk.</s>
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are actually not very many people who draw the line at birth or conception. Legally in the united states, the line is drawn at either viability (24 weeks) or response to pain and stimuli (20 weeks).</s>
Label encoding: <s>There are actually not very many people who draw the line at birth or conception. Legally in the united states, the line is drawn at either viability (24 weeks) or response to pain and stimuli (20 weeks).</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree. I didn't mean to say that embargoes are *good*, just that they are effective at causing poverty. By extension, I meant to show that the opposite, free trade, would create wealth.</s>
Label encoding: <s>I agree. I didn't mean to say that embargoes are *good*, just that they are effective at causing poverty. By extension, I meant to show that the opposite, free trade, would create wealth.</s>
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Being overweight is debilitating in ways remarkably similar to drunkenness; reduced coordination, inability to perform many tasks requiring manual control, etc. it doesn't fog your mind,<mask> your original point was about dependence, not judgment</s><pad>
Label encoding: <s>Being overweight is debilitating in ways remarkably similar to drunkenness; reduced coordination, inability to perform many tasks requiring manual control, etc. it doesn't fog your mind, but your original point was about dependence, not judgment</s><pad>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Being a 16 year old, I can tell you that he wouldn't get teased,<mask> all of his friends would want to mess around impersonate cops, even<mask> it's just to freak out a friend. [NEWLINE] </s>
Label encoding: <s>Being a 16 year old, I can tell you that he wouldn't get teased, but all of his friends would want to mess around impersonate cops, even if it's just to freak out a friend. [NEWLINE] </s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, the UN is just a conglomeration of bodies that claim use of violent aggression is legitimate.  It's the furthest away from a civilisation<mask> no organisation who claims that legitimacy is even permitted to exist.</s>
Label encoding: <s>No, the UN is just a conglomeration of bodies that claim use of violent aggression is legitimate.  It's the furthest away from a civilisation where no organisation who claims that legitimacy is even permitted to exist.</s>
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wouldnt i<mask> be responsible for the ones i shoot too? Or does your mind get hazy about that part<mask> you arent above using the same force to advance your agenda for<mask> the world should work?</s>
Label encoding: <s>Wouldnt i also be responsible for the ones i shoot too? Or does your mind get hazy about that part because you arent above using the same force to advance your agenda for how the world should work?</s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0908, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I wonder<mask> the British spy agency was thinking. [NEWLINE] [NEWLINE] "maybe<mask> we make it sound delicious no one will say anything." [NEWLINE] [NEWLINE] "<mask> about tempora? " [NEWLINE] [NEWLINE] " hell yea! "</s>
Label encoding: <s>I wonder what the British spy agency was thinking. [NEWLINE] [NEWLINE] "maybe if we make it sound delicious no one will say anything." [NEWLINE] [NEWLINE] " how about tempora? " [NEWLINE] [NEWLINE] " hell yea! "</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0641, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think I understand your question about dishonesty. I honestly have no problem with the adoption process. It's pretty cost prohibitive for the majority of people<mask>. Foster care system is horrible. </s>
Label encoding: <s>I don't think I understand your question about dishonesty. I honestly have no problem with the adoption process. It's pretty cost prohibitive for the majority of people though. Foster care system is horrible. </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can you provide some justification for that? I've been reading some Kuhn lately and I haven't seen any indication of that,<mask> it's possible that I haven't gotten deep enough in<mask>. </s>
Label encoding: <s>Can you provide some justification for that? I've been reading some Kuhn lately and I haven't seen any indication of that, but it's possible that I haven't gotten deep enough in yet. </s>
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the word "deserves" throws this off. Nobody deserves anything--not even five bucks. That someone GETS five bucks is nice,<mask> the word "deserves" is bullshit. </s>
Label encoding: <s>I think the word "deserves" throws this off. Nobody deserves anything--not even five bucks. That someone GETS five bucks is nice, but the word "deserves" is bullshit. </s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Spain performed it's last execution in 1975, which is relatively recent.  Furthermore,<mask><mask><mask> that capital punishment should be abolished, my entire view of a justice system does not hinge on this point.</s>
Label encoding: <s>Spain performed it's last execution in 1975, which is relatively recent.  Furthermore, while I agree that capital punishment should be abolished, my entire view of a justice system does not hinge on this point.</s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your explanation<mask> Scar did it.  He did it to his own bloodline, just for power, and<mask> you said achieved exactly that (albeit for not too long). Other villains did not. </s>
Label encoding: <s>Your explanation why Scar did it.  He did it to his own bloodline, just for power, and as you said achieved exactly that (albeit for not too long). Other villains did not. </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>that's still technically atheism,<mask> an agnostic does not actually believe in god. Agnostic is often used in combination with the other two terms, an agnostic atheist or agnostic theist.</s>
Label encoding: <s>that's still technically atheism, as an agnostic does not actually believe in god. Agnostic is often used in combination with the other two terms, an agnostic atheist or agnostic theist.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What's even worse are the english teachers who, after learning that the author intended no such symbols, go on to claim that authorial intent is irrelevant and that the interpretation being taught is obviously correct.</s><pad>
Label encoding: <s>What's even worse are the english teachers who, after learning that the author intended no such symbols, go on to claim that authorial intent is irrelevant and that the interpretation being taught is obviously correct.</s><pad>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Every person on this earth that can speak is capable of rapping. [ENDQ] [NEWLINE] There is no way that every person on earth can deliver (or write) this Grammy winning performance: [NEWLINE] [NEWLINE] [URL] </s>
Label encoding: <s> [STARTQ] Every person on this earth that can speak is capable of rapping. [ENDQ] [NEWLINE] There is no way that every person on earth can deliver (or write) this Grammy winning performance: [NEWLINE] [NEWLINE] [URL] </s>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>IF you had read my OP you'd know that I'm talking about a colloquial use of the word "rape," not a legal one. That is the entire point of this discussion.</s>
Label encoding: <s>IF you had read my OP you'd know that I'm talking about a colloquial use of the word "rape," not a legal one. That is the entire point of this discussion.</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes we agree there, I would pursue the perpetrator(s) normally. [NEWLINE] [NEWLINE] <mask><mask> a voice in my head says "serves him right" I can't help feelng guilty.</s>
Label encoding: <s>Yes we agree there, I would pursue the perpetrator(s) normally. [NEWLINE] [NEWLINE] But if a voice in my head says "serves him right" I can't help feelng guilty.</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well yeah I was discussing it within the context of someone else's breakdown. I'd still be pissed about shitting 6 grand, not to mention costs in efficiency for HR to post and interview people</s>
Label encoding: <s>Well yeah I was discussing it within the context of someone else's breakdown. I'd still be pissed about shitting 6 grand, not to mention costs in efficiency for HR to post and interview people</s>
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's correlation vs causality. The fact that wages were lower<mask> we didn't have a minimum wage doesn't show that wages were lower _<mask> _ we didn't have a minimum wage.</s>
Label encoding: <s>It's correlation vs causality. The fact that wages were lower when we didn't have a minimum wage doesn't show that wages were lower _ because _ we didn't have a minimum wage.</s>
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You might want to read rule B on the submission rules. [NEWLINE] [NEWLINE] You aren't supposed to be the devil's advocate, you should personally hold the view and be willing to have it changed.</s>
Label encoding: <s>You might want to read rule B on the submission rules. [NEWLINE] [NEWLINE] You aren't supposed to be the devil's advocate, you should personally hold the view and be willing to have it changed.</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay, and<mask> are women supposed to do to avoid rape on top of the hundreds of other things they do already?<mask> is "giving advice" after they've been victimized any help?</s>
Label encoding: <s>Okay, and what are women supposed to do to avoid rape on top of the hundreds of other things they do already? How is "giving advice" after they've been victimized any help?</s>
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[Here]( [URL] /) is a blog post by a linguist who has lived and worked in the USA and the UK. He disagrees with your idea that our Englishes have much difference.</s>
Label encoding: <s>[Here]( [URL] /) is a blog post by a linguist who has lived and worked in the USA and the UK. He disagrees with your idea that our Englishes have much difference.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Love and relationships are not a competition. You don't have to compare yourself to other people to be happy.<mask><mask> science shows that comparing yourself to others hurts your ability to be happy.</s>
Label encoding: <s>Love and relationships are not a competition. You don't have to compare yourself to other people to be happy. In fact science shows that comparing yourself to others hurts your ability to be happy.</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Nah, I'm referring to shit like banking and insurance that a society without money wouldn't need. This is a pretty good video: [URL] -BZ3eWxM </s>
Label encoding: <s>Nah, I'm referring to shit like banking and insurance that a society without money wouldn't need. This is a pretty good video: [URL] -BZ3eWxM </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] ~~teeth-hurtingly~~ deliciously sweet.  Especially<mask> you get a ~~dreaded~~ choice corner piece or flower. [ENDQ] [NEWLINE] Fixed that for you</s>
Label encoding: <s> [STARTQ] ~~teeth-hurtingly~~ deliciously sweet.  Especially if you get a ~~dreaded~~ choice corner piece or flower. [ENDQ] [NEWLINE] Fixed that for you</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not an intelligent show, and it isn't trying to be. Just<mask> the main characters are supposed to be intelligent doesn't mean the dialog or humor has to be. </s>
Label encoding: <s>It's not an intelligent show, and it isn't trying to be. Just because the main characters are supposed to be intelligent doesn't mean the dialog or humor has to be. </s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's a good thing I repeatedly stipulated I was talking about fearing one's own death in itself. [NEWLINE] [NEWLINE] <mask> it doesn't seem to have helped people's reading comprehension.</s><pad>
Label encoding: <s>It's a good thing I repeatedly stipulated I was talking about fearing one's own death in itself. [NEWLINE] [NEWLINE] Though it doesn't seem to have helped people's reading comprehension.</s><pad>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Utilitarianism is the most common moral theory,<mask><mask> some call it 'common sense morality'. Most people agree that an action yielding the greatest amount of happiness is good. </s>
Label encoding: <s>Utilitarianism is the most common moral theory, hence why some call it 'common sense morality'. Most people agree that an action yielding the greatest amount of happiness is good. </s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So forced oral sex is not rape? Foreign objects are not rape? Seems quite surprising, even assuming laws are made from the viewpoint<mask> "men rape women and that's it."</s>
Label encoding: <s>So forced oral sex is not rape? Foreign objects are not rape? Seems quite surprising, even assuming laws are made from the viewpoint if "men rape women and that's it."</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh, I thought you meant "working on my car"<mask> in polishing it and keeping it in super condition<mask> a hobby, not<mask> in "occasional repairs".</s>
Label encoding: <s>Oh, I thought you meant "working on my car" as in polishing it and keeping it in super condition as a hobby, not as in "occasional repairs".</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's pretty deep for a guy who calls himself gnarly butt-queef. [NEWLINE] [NEWLINE] I know this will be deleted,<mask> I just couldn't hold back.</s>
Label encoding: <s>That's pretty deep for a guy who calls himself gnarly butt-queef. [NEWLINE] [NEWLINE] I know this will be deleted, but I just couldn't hold back.</s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem of world hunger isn't one of redistribution of wealth...especially on that scale. Even with money there are hugely limiting factors on getting some of these people food. </s>
Label encoding: <s>The problem of world hunger isn't one of redistribution of wealth...especially on that scale. Even with money there are hugely limiting factors on getting some of these people food. </s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0624, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0609, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's<mask> I mentioned punishing anyone or any group who shows a bias in who they pull over to remove any incentives there are to pull over one group more than another.</s>
Label encoding: <s>That's why I mentioned punishing anyone or any group who shows a bias in who they pull over to remove any incentives there are to pull over one group more than another.</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[I'm curious]( [URL] )<mask> to<mask> you define 'intelligence'. [NEWLINE] [NEWLINE] <mask> is 'intelligence' for you?<mask> is an 'intelligent response'?</s><pad>
Label encoding: <s>[I'm curious]( [URL] ) as to how you define 'intelligence'. [NEWLINE] [NEWLINE] What is 'intelligence' for you? What is an 'intelligent response'?</s><pad>
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So this is just the issue of "all squares are rectangles,<mask> not all rectangles are squares"? Is the added confusion worth the slight differences in terminology? </s>
Label encoding: <s>So this is just the issue of "all squares are rectangles, but not all rectangles are squares"? Is the added confusion worth the slight differences in terminology? </s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right,<mask> the KKK is a political group.  Are we really going to give the government the power to decide<mask> political groups can exist based on rhetoric alone? </s>
Label encoding: <s>Right, but the KKK is a political group.  Are we really going to give the government the power to decide what political groups can exist based on rhetoric alone? </s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You wouldn't add those values together.  The risk of a walker being hit by a car is not equal to the risk of a driver hitting a pedestrian.</s>
Label encoding: <s>You wouldn't add those values together.  The risk of a walker being hit by a car is not equal to the risk of a driver hitting a pedestrian.</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How far does this go?<mask> we catch illegals on a public road (paid by tax) should we detain them?<mask> would that even be enforced? </s><pad>
Label encoding: <s>How far does this go? If we catch illegals on a public road (paid by tax) should we detain them? How would that even be enforced? </s><pad>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've a video that is relevant to your CMV :) [NEWLINE] [Here you go.]( [URL] ) [NEWLINE] <mask><mask> the answer is somewhere in the middle =]</s>
Label encoding: <s>I've a video that is relevant to your CMV :) [NEWLINE] [Here you go.]( [URL] ) [NEWLINE] I think the answer is somewhere in the middle =]</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not saying that my idea is perfect :).<mask> there was a non-hormonal, reversible method available, I would much prefer to use that :).</s>
Label encoding: <s>I'm not saying that my idea is perfect :). If there was a non-hormonal, reversible method available, I would much prefer to use that :).</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe I should have clarified a little more,<mask> essentially<mask> I mean by "typical" is whatever the average work hours are for a given nation. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Maybe I should have clarified a little more, but essentially what I mean by "typical" is whatever the average work hours are for a given nation. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Might I suggest Matthew 7? [URL] [NEWLINE] [NEWLINE] i.e. it's not on you to determine whether others' activities are sinful.  Judge not...</s>
Label encoding: <s>Might I suggest Matthew 7? [URL] [NEWLINE] [NEWLINE] i.e. it's not on you to determine whether others' activities are sinful.  Judge not...</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Adderall is treatment for AD(H)D, not a cure. [NEWLINE] <mask><mask> with<mask> you're saying I'm just making a clarification.</s>
Label encoding: <s>Adderall is treatment for AD(H)D, not a cure. [NEWLINE] I agree with what you're saying I'm just making a clarification.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Society merely sets the standard of certain things;<mask> you choose to follow a standard that can get you injured, yes, it does become your fault.</s>
Label encoding: <s>Society merely sets the standard of certain things; if you choose to follow a standard that can get you injured, yes, it does become your fault.</s>
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Haha, that's exactly<mask> I was thinking,<mask> I didn't want to name names in case OP was a huge Micheal Bay fan too lol</s>
Label encoding: <s>Haha, that's exactly what I was thinking, but I didn't want to name names in case OP was a huge Micheal Bay fan too lol</s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ Gave a clear view of the feelings going through his head at the time, needless to say those weren't enjoyable. Changed my view.</s>
Label encoding: <s>∆ Gave a clear view of the feelings going through his head at the time, needless to say those weren't enjoyable. Changed my view.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I set up online backup for my wife's photos and important data. It took 45 days of 24/7 uploading to finish the initial upload.</s>
Label encoding: <s>I set up online backup for my wife's photos and important data. It took 45 days of 24/7 uploading to finish the initial upload.</s>
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right, of course,<mask> until those police policies change, I don't think we can fault victims for not wanting to come forward.</s>
Label encoding: <s>You're right, of course, but until those police policies change, I don't think we can fault victims for not wanting to come forward.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you kind of are. You may not claim to be an activist in it,<mask> you are certainly assigning yourself along an ideological movement.</s>
Label encoding: <s>I think you kind of are. You may not claim to be an activist in it, but you are certainly assigning yourself along an ideological movement.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Forrest Gump is an excellent example of CGI done well. Over 20 years old and people will still not be able to spot it all.</s>
Label encoding: <s>Forrest Gump is an excellent example of CGI done well. Over 20 years old and people will still not be able to spot it all.</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0949, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Whats the average career length for a professional dancer? I was under the impression that it ends<mask> you are quite young similar to gymnastics</s>
Label encoding: <s>Whats the average career length for a professional dancer? I was under the impression that it ends what you are quite young similar to gymnastics</s>
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Or whatever causes a dinoasaur to become a fossil pressed between two pieces of rock. That was my point, not the boulder.</s>
Label encoding: <s>Or whatever causes a dinoasaur to become a fossil pressed between two pieces of rock. That was my point, not the boulder.</s>
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to disagree that students in first or second grace learn any mathematics, they learn<mask> to calculate easy things, that's a difference.</s>
Label encoding: <s>I have to disagree that students in first or second grace learn any mathematics, they learn how to calculate easy things, that's a difference.</s>
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0541, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm impacted more by my fellow men feeling the need to keep their legs spread than I ever have been by overweight people taking up space.</s>
Label encoding: <s>I'm impacted more by my fellow men feeling the need to keep their legs spread than I ever have been by overweight people taking up space.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, there was not a clear path to success. That is you applying the path taken<mask> being clear after the fact. </s><pad>
Label encoding: <s>No, there was not a clear path to success. That is you applying the path taken as being clear after the fact. </s><pad>
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Petty theft doesn't require the same kind of attention or discussion<mask> rape, its several levels of magnitude less important then rape.</s>
Label encoding: <s>Petty theft doesn't require the same kind of attention or discussion as rape, its several levels of magnitude less important then rape.</s>
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not just wait staff, it's tipping in general. Those waiting my table<mask> an example of my view. </s>
Label encoding: <s>It's not just wait staff, it's tipping in general. Those waiting my table where an example of my view. </s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0772, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The sexual assault frequency points are just<mask> valid in Europe,<mask><mask> whether the population would "mind" integrated bathrooms.</s>
Label encoding: <s>The sexual assault frequency points are just as valid in Europe, regardless of whether the population would "mind" integrated bathrooms.</s>
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think they consider the deal to be weak and ineffective.  A stronger deal wouldn't have drawn<mask> much criticism.</s>
Label encoding: <s>I think they consider the deal to be weak and ineffective.  A stronger deal wouldn't have drawn as much criticism.</s>
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, it was a serious event which deserves consideration.<mask> that does not mean he need be emotionally affected by it</s>
Label encoding: <s>Yes, it was a serious event which deserves consideration. But that does not mean he need be emotionally affected by it</s>
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Good point,<mask> now I can be 100% sure that I am not 100% sure about<mask> I know.</s>
Label encoding: <s>Good point, but now I can be 100% sure that I am not 100% sure about what I know.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Don't forget that men get ogled and harassed at<mask> well,<mask> women getting special treatment is stupid.</s>
Label encoding: <s>Don't forget that men get ogled and harassed at as well, so women getting special treatment is stupid.</s>
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But it makes it almost impossible to fire a bad teacher, which costs way more than a few thousand dollars.</s>
Label encoding: <s>But it makes it almost impossible to fire a bad teacher, which costs way more than a few thousand dollars.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Believe it or not, ladies' nights are actually banned in many places. [NEWLINE] [NEWLINE] [URL] '_night</s>
Label encoding: <s>Believe it or not, ladies' nights are actually banned in many places. [NEWLINE] [NEWLINE] [URL] '_night</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Active voice is "I picked up the ball". Passive voice is "The ball was picked up by me".</s>
Label encoding: <s>Active voice is "I picked up the ball". Passive voice is "The ball was picked up by me".</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you, apologies, I was on my phone and neglected to look before posting. Much appreciate the links.</s><pad>
Label encoding: <s>Thank you, apologies, I was on my phone and neglected to look before posting. Much appreciate the links.</s><pad>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>TiA is a little more self-aware, and that makes a lot more of a difference.</s>
Label encoding: <s>TiA is a little more self-aware, and that makes a lot more of a difference.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you. Twas not my intention. Which part in particular,<mask> I could ask? </s>
Label encoding: <s>Thank you. Twas not my intention. Which part in particular, if I could ask? </s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] capitalize on the work of those that did [ENDQ] [NEWLINE] which in turn changes the genre...</s>
Label encoding: <s> [STARTQ] capitalize on the work of those that did [ENDQ] [NEWLINE] which in turn changes the genre...</s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe they should have a right to garnish my wages a reasonable amount, certainly.</s>
Label encoding: <s>I believe they should have a right to garnish my wages a reasonable amount, certainly.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know, I thought you were suggesting that the army was entirely draft based, nm.</s>
Label encoding: <s>I know, I thought you were suggesting that the army was entirely draft based, nm.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you<mask> think protection of religion under the law or in policy is discrimination? </s>
Label encoding: <s>Do you also think protection of religion under the law or in policy is discrimination? </s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok valid point. It would allow you to check notifications during a conversation.</s>
Label encoding: <s>Ok valid point. It would allow you to check notifications during a conversation.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed - 1 delta awarded to /u/jennerality</s>
Label encoding: <s>Confirmed - 1 delta awarded to /u/jennerality</s>
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Plasma pistol + battle rifle = best newb combo ever</s>
Label encoding: <s>Plasma pistol + battle rifle = best newb combo ever</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah...THAT'S<mask> I'm saying...</s>
Label encoding: <s>Yeah...THAT'S what I'm saying...</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I thought I made that pretty clear. </s>
Label encoding: <s>I thought I made that pretty clear. </s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Cooking fish in a dorm room?</s>
Label encoding: <s>Cooking fish in a dorm room?</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No problem, friend.</s><pad>
Label encoding: <s>No problem, friend.</s><pad>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.6628e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.4504e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8336e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.9884e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.1153e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.4105e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.7623e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.9904e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.8234e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.0609e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.0444e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.9351e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.8857e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7578e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.9001e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.1823e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6238e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.0486e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.9476e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.2189e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7122e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.1961e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.2754e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7267e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7647e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6583e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.1031e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.4259e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6791e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7853e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.5493e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.4503e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6853e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.9374e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.2263e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(3.8427e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.9536e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8749e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.0130e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.6766e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6150e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4019e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.8865e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.8547e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8056e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.8837e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.3405e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.5872e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.1825e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.3037e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6308e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.1754e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.2016e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.6361e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6127e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.4019e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7501e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3301e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.3113e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7786e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.1513e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.3454e-05, device='cuda:0', grad_fn=<NllLossBackward>)
--------------EPOCH 4-------------
Test Accuracy: tensor(0.7838, device='cuda:0')
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Saving model at iteration: 0
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To begin my argument, I need to make sure we are using a common set of definitions.<mask> for clarity in this thread, I would like to use the following definitions: [NEWLINE] [NEWLINE] Ethnicity: A *socially-defined* category of people who identify with each other based on common ancestral, social, *cultural* or national experience. [NEWLINE] [NEWLINE] Culture: The attitudes and behavior characteristic of a particular *social* group. [NEWLINE] [NEWLINE] Race: Major divisions of humankind, having distinct physical characteristics (i.e., defined primarily by *physical* differences). [NEWLINE] [NEWLINE] 1. It would appear to me that it is clearly wrong to judge a person in advance based on the physical traits given to them simply by virtue of being born. To my mind this is<mask> would constitute racism proper (race being defined<mask> above). Racism<mask> such, I hold<mask> categorically immoral. [NEWLINE] [NEWLINE] 2. Culture (<mask> defined above) consists of attitudes and behaviors associated with social groups. *[edit: wording]* Barring genetic explanations or explanations from psychiatric disorders, it seems like talk about behavior and attitudes in individual people are generally explained from the perspective of the ideas people hold. It seems to stand to reason that<mask> explanations of behavior and attitudes in individuals are explained by ideas held, then “attitudes and behavior characteristic of a particular social group” would most easily be explained by a commonly held set of ideas. [NEWLINE] [NEWLINE] 3. Ideas and behaviors, per se, can and should always be looked at with a critical eye and always open to scrutiny, satire, debate, and criticism.<mask> culture is understood to be a social group’s set of common ideas and behaviors, they should be open to the same. I hold this<mask> categorical, and<mask> you want to CMV, this is really the heart of the matter. [NEWLINE] [NEWLINE] 4. One of the linguistic rat’s nests that frequently arise in discussions about these topics is the conflation of race and culture (and<mask> ideas) under the umbrella term “ethnicity.”<mask> to criticize the culture (i.e. ideas) common to an ethnic group it is implied that you are criticizing the race<mask> well. It seems like this is a rather cheap way to insulate ideas from criticism. Race is inborn, culture is an idea construct. Ideas and behaviors can and should be open to criticism. [NEWLINE] [NEWLINE] 5. John Stewart is not believed to be an asshole<mask> he criticized the ideas and behaviors of Ferguson and New York
Label encoding: <s>To begin my argument, I need to make sure we are using a common set of definitions. So for clarity in this thread, I would like to use the following definitions: [NEWLINE] [NEWLINE] Ethnicity: A *socially-defined* category of people who identify with each other based on common ancestral, social, *cultural* or national experience. [NEWLINE] [NEWLINE] Culture: The attitudes and behavior characteristic of a particular *social* group. [NEWLINE] [NEWLINE] Race: Major divisions of humankind, having distinct physical characteristics (i.e., defined primarily by *physical* differences). [NEWLINE] [NEWLINE] 1. It would appear to me that it is clearly wrong to judge a person in advance based on the physical traits given to them simply by virtue of being born. To my mind this is what would constitute racism proper (race being defined as above). Racism as such, I hold as categorically immoral. [NEWLINE] [NEWLINE] 2. Culture ( as defined above) consists of attitudes and behaviors associated with social groups. *[edit: wording]* Barring genetic explanations or explanations from psychiatric disorders, it seems like talk about behavior and attitudes in individual people are generally explained from the perspective of the ideas people hold. It seems to stand to reason that if explanations of behavior and attitudes in individuals are explained by ideas held, then “attitudes and behavior characteristic of a particular social group” would most easily be explained by a commonly held set of ideas. [NEWLINE] [NEWLINE] 3. Ideas and behaviors, per se, can and should always be looked at with a critical eye and always open to scrutiny, satire, debate, and criticism. If culture is understood to be a social group’s set of common ideas and behaviors, they should be open to the same. I hold this as categorical, and if you want to CMV, this is really the heart of the matter. [NEWLINE] [NEWLINE] 4. One of the linguistic rat’s nests that frequently arise in discussions about these topics is the conflation of race and culture (and therefore ideas) under the umbrella term “ethnicity.” Therefore to criticize the culture (i.e. ideas) common to an ethnic group it is implied that you are criticizing the race as well. It seems like this is a rather cheap way to insulate ideas from criticism. Race is inborn, culture is an idea construct. Ideas and behaviors can and should be open to criticism. [NEWLINE] [NEWLINE] 5. John Stewart is not believed to be an asshole when he criticized the ideas and behaviors of Ferguson and New York
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I want to take a crack at this. Before I get into the subject matter, I want to build my ethos a bit (to show I'm not just making all this up). I do a lot of side work for Scott Horton, the host of Antiwar Radio and The Scott Horton Show. He's a longtime editor of Antiwar.com, and most of<mask> I'm about to say, I got directly from him (just now). [NEWLINE] [NEWLINE] The first thing that is really important to note is that non-interventionism will not prevent every crisis. Not even Ron Paul would say that. It will<mask>, prevent escalation of various crises. You say that you're abandoning non-interventionism,<mask> a point that I'll keep reiterating is that non-interventionism is a more realistic and *better* foreign policy than interventionism.<mask> you respond, I challenge you to provide specific actions that should have been taken to resolve all of the issues you address: ISIS, Israel, China, and Russia. [NEWLINE] [NEWLINE] **I'll take on the Islamic State first.** The premise of your argument is that non-interventionists supported the withdrawal from Iraq in 2010-2011, which was a bad move<mask> that led to the rise of ISIS. There's a lot to say here. [NEWLINE] [NEWLINE] 1. <mask> the US had stayed in Iraq longer, the ISIS would have grown sooner.<mask><mask><mask> that these jihadists thrive off of violence. They recruit by painting the West<mask> the enemy.<mask> the US bombs them, it proves them true.<mask> the US first started bombing them this summer,[CNN reported an uptick in recruitment]( [URL] /). Likewise,<mask> the US had stayed, it would have given the ISIS another enemy to fight.<mask> US troops really are bad for them,<mask> are they practically begging the US to send more troops? It's telling that most of their atrocities (and all of their journalist beheadings) have been propagated by *them.* This is empirically proven by<mask> happened in 2005 (a particularly bad year in the war). Search "the El Salvador Option" and "James Steele."<mask> happened is that the US began training Shia militias to fight the Sunni insurgency. A big partner of the US was the Badr Brigade. This ended up exacerbating sectarian civil war. Sunnis and Shias weren't always killing each other: it was the US that made it<mask>. It was<mask> al-Zarqawi's big year. [NEWLINE] [NEWLINE] 2.<mask>
Label encoding: <s>I want to take a crack at this. Before I get into the subject matter, I want to build my ethos a bit (to show I'm not just making all this up). I do a lot of side work for Scott Horton, the host of Antiwar Radio and The Scott Horton Show. He's a longtime editor of Antiwar.com, and most of what I'm about to say, I got directly from him (just now). [NEWLINE] [NEWLINE] The first thing that is really important to note is that non-interventionism will not prevent every crisis. Not even Ron Paul would say that. It will however, prevent escalation of various crises. You say that you're abandoning non-interventionism, so a point that I'll keep reiterating is that non-interventionism is a more realistic and *better* foreign policy than interventionism. If you respond, I challenge you to provide specific actions that should have been taken to resolve all of the issues you address: ISIS, Israel, China, and Russia. [NEWLINE] [NEWLINE] **I'll take on the Islamic State first.** The premise of your argument is that non-interventionists supported the withdrawal from Iraq in 2010-2011, which was a bad move because that led to the rise of ISIS. There's a lot to say here. [NEWLINE] [NEWLINE] 1.  If the US had stayed in Iraq longer, the ISIS would have grown sooner. The reason is that these jihadists thrive off of violence. They recruit by painting the West as the enemy. When the US bombs them, it proves them true. When the US first started bombing them this summer,[CNN reported an uptick in recruitment]( [URL] /). Likewise, if the US had stayed, it would have given the ISIS another enemy to fight. If US troops really are bad for them, why are they practically begging the US to send more troops? It's telling that most of their atrocities (and all of their journalist beheadings) have been propagated by *them.* This is empirically proven by what happened in 2005 (a particularly bad year in the war). Search "the El Salvador Option" and "James Steele." What happened is that the US began training Shia militias to fight the Sunni insurgency. A big partner of the US was the Badr Brigade. This ended up exacerbating sectarian civil war. Sunnis and Shias weren't always killing each other: it was the US that made it so. It was also al-Zarqawi's big year. [NEWLINE] [NEWLINE] 2. If
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you don't mind, im going to try and change your view by looking at some of the policies you mention (which,<mask> my disagreement, are legitimate policies) and by trying to explain some of the more structural issues with the party. [NEWLINE] [NEWLINE] <mask> Nigel Farage is obviously a well worded and rather charismatic fella. His facial expressions and some of his behaviour make him look like a low-brow village idiot that lives in a pub<mask> the reallity is that he's much more savy than that. One major problem is that<mask> he might be described<mask> an astute politician, the rest of his party is riddled with incompetent gaffe masters who have contributed heavily to the perception that the party is all about "bashing queers, expelling foreigners and keeping england for the english just like old eunoch would have wanted" [NEWLINE] Granted these party members and their gaffes (which are later condemned) have done a great deal to help UKIP get popularist support<mask> those that aren't<mask> keen on the more extreme manifestations of the party are, rightfully, terrified that Nigel has no chance of keeping them in check once they get into parliament. [NEWLINE] [NEWLINE] Now the policies: [NEWLINE] [NEWLINE] * Getting out of the EU<mask> it has a lot of regulations which put a hold on the UK [NEWLINE] [NEWLINE] The EU does have a lot of regulations which the UK is held too,<mask> all the countries in the EU are held to them too. Furthermore the UK most definitely does have a say  in these regulations, it's sort of<mask> MEPs are there for. The UK pushed very hard to get into Europe in the past and it knew perfectly well<mask> it was getting into<mask><mask> it joined the european community it has<mask> had  a say in EU regulations.<mask> this is a facile argument<mask> the EU does impose regulations on the UK,<mask> the UK contributes to forming these regulations. [NEWLINE] [NEWLINE] * London is very dirty people come from other countries and don't bother to learn english and'sharia police'. [NEWLINE] [NEWLINE] London is not all that dirty mate, trust me, i've lived in naples, rome, berlin and paris. All those cities are a lot dirtier than london. People do bother to learn english, some don't learn<mask> well<mask> others<mask> most learn a little bit. A lot of people speak perfect english<mask> still frequent people with whom they have another common language and with them they speak that other language (yes even on public transport).<mask><mask>
Label encoding: <s>If you don't mind, im going to try and change your view by looking at some of the policies you mention (which, despite my disagreement, are legitimate policies) and by trying to explain some of the more structural issues with the party. [NEWLINE] [NEWLINE] So Nigel Farage is obviously a well worded and rather charismatic fella. His facial expressions and some of his behaviour make him look like a low-brow village idiot that lives in a pub but the reallity is that he's much more savy than that. One major problem is that while he might be described as an astute politician, the rest of his party is riddled with incompetent gaffe masters who have contributed heavily to the perception that the party is all about "bashing queers, expelling foreigners and keeping england for the english just like old eunoch would have wanted" [NEWLINE] Granted these party members and their gaffes (which are later condemned) have done a great deal to help UKIP get popularist support but those that aren't so keen on the more extreme manifestations of the party are, rightfully, terrified that Nigel has no chance of keeping them in check once they get into parliament. [NEWLINE] [NEWLINE] Now the policies: [NEWLINE] [NEWLINE] * Getting out of the EU because it has a lot of regulations which put a hold on the UK [NEWLINE] [NEWLINE] The EU does have a lot of regulations which the UK is held too, but all the countries in the EU are held to them too. Furthermore the UK most definitely does have a say  in these regulations, it's sort of what MEPs are there for. The UK pushed very hard to get into Europe in the past and it knew perfectly well what it was getting into because since it joined the european community it has also had  a say in EU regulations. so this is a facile argument because the EU does impose regulations on the UK, but the UK contributes to forming these regulations. [NEWLINE] [NEWLINE] * London is very dirty people come from other countries and don't bother to learn english and'sharia police'. [NEWLINE] [NEWLINE] London is not all that dirty mate, trust me, i've lived in naples, rome, berlin and paris. All those cities are a lot dirtier than london. People do bother to learn english, some don't learn as well as others but most learn a little bit. A lot of people speak perfect english but still frequent people with whom they have another common language and with them they speak that other language (yes even on public transport). As far
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't particularly hold this view dearly, I just expect there might be some budding historians on Reddit with more information! It was influenced mostly by [this article]( [URL] /?no-ist), which is fascinating<mask> you just fancy some historical reading. Basically,<mask> Napoleon has won at Waterloo, and completed his conquest of Europe (bar Russia, of course, and maybe England), the continent would have flourished economically, socially and politically, more<mask> than now. I will highlight the main points below. [NEWLINE] [NEWLINE] [NEWLINE] **1. Revolutionary domestic reforms which all of Europe would have continued to benefit from.** [NEWLINE] [NEWLINE] *"<mask> he said he would be remembered not for his military victories,<mask> for his domestic reforms, especially the Code Napoleon, that brilliant distillation of 42 competing and often contradictory legal codes into a single, easily comprehensible body of French law.<mask><mask>, Napoleon’s years<mask> first consul, from 1799 to 1804, were extraordinarily peaceful and productive. He<mask> created the educational system based on lycées and grandes écoles and the Sorbonne, which put France at the forefront of European educational achievement. He consolidated the administrative system based on departments and prefects. He initiated the Council of State, which still vets the laws of France, and the Court of Audit, which oversees its public accounts. He organized the Banque de France and the Légion d’Honneur, which thrive today. He<mask> built or renovated much of the Parisian architecture that we still enjoy, both the useful—the quays along the Seine and four bridges over it, the sewers and reservoirs—and the beautiful, such<mask> the Arc de Triomphe, the Rue de Rivoli and the Vendôme column."* [NEWLINE] [NEWLINE] Napoleon was an extraordinary governor. Many of his reforms are<mask> the rest of Europe based their institutions on, only much later. [NEWLINE] [NEWLINE] [NEWLINE] **2. Napoleon was a lover, not a fighter.** [NEWLINE] [NEWLINE] *In September 1805, Austria invaded Napoleon’s ally Bavaria, and Russia declared war on France<mask> well. Napoleon swiftly won the ensuing War of the Third Coalition with his finest victory, at Austerlitz in 1805. The next year the Prussians<mask> declared war on him,<mask> they were soundly defeated at Jena; Napoleon’s peace treaty of Tilsit with Russia and Prussia followed. The Austrians declared war on France once more in 1809,<mask>
Label encoding: <s>I don't particularly hold this view dearly, I just expect there might be some budding historians on Reddit with more information! It was influenced mostly by [this article]( [URL] /?no-ist), which is fascinating if you just fancy some historical reading. Basically, if Napoleon has won at Waterloo, and completed his conquest of Europe (bar Russia, of course, and maybe England), the continent would have flourished economically, socially and politically, more so than now. I will highlight the main points below. [NEWLINE] [NEWLINE] [NEWLINE] **1. Revolutionary domestic reforms which all of Europe would have continued to benefit from.** [NEWLINE] [NEWLINE] *" Yet he said he would be remembered not for his military victories, but for his domestic reforms, especially the Code Napoleon, that brilliant distillation of 42 competing and often contradictory legal codes into a single, easily comprehensible body of French law. In fact, Napoleon’s years as first consul, from 1799 to 1804, were extraordinarily peaceful and productive. He also created the educational system based on lycées and grandes écoles and the Sorbonne, which put France at the forefront of European educational achievement. He consolidated the administrative system based on departments and prefects. He initiated the Council of State, which still vets the laws of France, and the Court of Audit, which oversees its public accounts. He organized the Banque de France and the Légion d’Honneur, which thrive today. He also built or renovated much of the Parisian architecture that we still enjoy, both the useful—the quays along the Seine and four bridges over it, the sewers and reservoirs—and the beautiful, such as the Arc de Triomphe, the Rue de Rivoli and the Vendôme column."* [NEWLINE] [NEWLINE] Napoleon was an extraordinary governor. Many of his reforms are what the rest of Europe based their institutions on, only much later. [NEWLINE] [NEWLINE] [NEWLINE] **2. Napoleon was a lover, not a fighter.** [NEWLINE] [NEWLINE] *In September 1805, Austria invaded Napoleon’s ally Bavaria, and Russia declared war on France as well. Napoleon swiftly won the ensuing War of the Third Coalition with his finest victory, at Austerlitz in 1805. The next year the Prussians also declared war on him, but they were soundly defeated at Jena; Napoleon’s peace treaty of Tilsit with Russia and Prussia followed. The Austrians declared war on France once more in 1809, but
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What does coontown's existence really matter<mask> to not be exposed to it all you have to do is not visit coontown?<mask> they brigade someone then the brigadiers ought to be punished, banned, etc.<mask> I understand it spez, Reddit's CEO, doesn't believe there's enough evidence of brigading, or consistent brigading, on the part of coontown's user base to have banned it for that reason like fatpeoplehate. Even<mask> they were brigading the consequences of their actions amount to the loss of useless internet points at most. A target could simply make a new account in a matter of seconds. This doesn't excuse brigading,<mask><mask><mask> its effects should be put into perspective. [NEWLINE] [NEWLINE] <mask>... I like that I can look at communities like coontown, theredpill, and others. These are people who actually exist. They're our neighbors and co-workers. They're out there being spooky.<mask><mask> all of us are interested in them. That's<mask> we're interested in documentaries like [Welcome to Leith]( [URL] ) and [Jesus Camp]( [URL] ). Unlike those documentaries,<mask>, these subreddits exist in real time, responding to the world, debating, philosophizing, moralizing. These subreddits offer windows into communities in ways that are practically unprecedented.<mask><mask> stormfront.com or.net or whatever it was/is had a forum, or does,<mask> the sheer diversity of subreddits (and ease of access to them) like coontown and theredpill is unlike anything I've experienced prior to reddit. [NEWLINE] [NEWLINE] Thirdly, I don't see<mask> coontown, theredpill, etc's continued existences will actually do any meaningful harm.<mask> they may possibly attract and 'convert' borderline racists they<mask> are exposed to direct criticism constantly. By the virtue of their existence they galvanize discussions about important issues that exist in our society.<mask> coontown didn't exist people would still be racist. Again, yes, their users can validate, embolden, and radicalize each other,<mask> they're<mask> exposed to (the aforementioned) criticism, judgement, and opposition. [NEWLINE] [NEWLINE] Fourthly,<mask><mask> it's interesting to consider, having brought up Jesus Camp, that benign appearing subreddits like, say, r/Christianity and r/Islam (which are probably mostly full of perfectly decent, sweet people) are more likely to advocate for less obviously hateful<mask> more actually negatively affecting legislation and real-world action.
Label encoding: <s>What does coontown's existence really matter when to not be exposed to it all you have to do is not visit coontown? If they brigade someone then the brigadiers ought to be punished, banned, etc. As I understand it spez, Reddit's CEO, doesn't believe there's enough evidence of brigading, or consistent brigading, on the part of coontown's user base to have banned it for that reason like fatpeoplehate. Even if they were brigading the consequences of their actions amount to the loss of useless internet points at most. A target could simply make a new account in a matter of seconds. This doesn't excuse brigading, but I think its effects should be put into perspective. [NEWLINE] [NEWLINE] Secondly... I like that I can look at communities like coontown, theredpill, and others. These are people who actually exist. They're our neighbors and co-workers. They're out there being spooky. I think all of us are interested in them. That's why we're interested in documentaries like [Welcome to Leith]( [URL] ) and [Jesus Camp]( [URL] ). Unlike those documentaries, however, these subreddits exist in real time, responding to the world, debating, philosophizing, moralizing. These subreddits offer windows into communities in ways that are practically unprecedented. I think stormfront.com or.net or whatever it was/is had a forum, or does, but the sheer diversity of subreddits (and ease of access to them) like coontown and theredpill is unlike anything I've experienced prior to reddit. [NEWLINE] [NEWLINE] Thirdly, I don't see how coontown, theredpill, etc's continued existences will actually do any meaningful harm. While they may possibly attract and 'convert' borderline racists they also are exposed to direct criticism constantly. By the virtue of their existence they galvanize discussions about important issues that exist in our society. If coontown didn't exist people would still be racist. Again, yes, their users can validate, embolden, and radicalize each other, but they're also exposed to (the aforementioned) criticism, judgement, and opposition. [NEWLINE] [NEWLINE] Fourthly, I think it's interesting to consider, having brought up Jesus Camp, that benign appearing subreddits like, say, r/Christianity and r/Islam (which are probably mostly full of perfectly decent, sweet people) are more likely to advocate for less obviously hateful but more actually negatively affecting legislation and real-world action.
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you could easily<mask><mask> learning the bare-bones of business breaks down certain assumptions, etc.<mask>, my impression of business school is not that one learns<mask> businesses work,<mask> rather<mask> to make businesses work. It's not a study<mask> much<mask> it is instruction. [ENDQ] [NEWLINE] Unfortunately you're impression is not the reality. [NEWLINE] [NEWLINE] Business school is<mask> businesses work. It's no different than studying any other field. [NEWLINE] [NEWLINE] There are a lot of different businesses, business strategies, management strategies, etc. And there are examples of excellently done strategies, and poorly done strategies. [NEWLINE] [NEWLINE] Case study was the bread and butter of my classes.  My senior management class had all of us break off into teams and play a simulator to see who could run a successful business. [NEWLINE] [NEWLINE] [STARTQ] I don't see<mask><mask> I claim that business isn't a legitimate discipline, I must<mask> accept that social sciences and humanities are not legitimate...<mask><mask> my point would be best understood<mask> saying that business school uses these disciplines only to promote generation of wealth, rather than contribute anything meaningful to them. We can point to mathematicians and economists<mask> responsible for major theories in risk analysis and market fluctuations - businessmen simply implement them. [ENDQ] [NEWLINE] <mask><mask>'s the difference between studying them and studying business? Really? [NEWLINE] [NEWLINE] <mask><mask> without implementation those disciplines would be sorely lacking in any concrete info, or data sets to build their models off of. [NEWLINE] [NEWLINE] Ideas are nice,<mask> practice is<mask> shit gets done. That thesis on observed student's and<mask> to elicit certain behaviors  may be nice;<mask><mask> could you develop the necessary data to prove it's a viable or true among adults at large? *You put it into practice*. [NEWLINE] [NEWLINE] Business, at a minimum, is the field of of putting things into practice. And studying business, is studying<mask> to put *anything* into practice, and ideally make a living off of it. [NEWLINE] [NEWLINE] [STARTQ] No, I'm simply saying that it is not and should not be treated<mask> an academic discipline. I'm asking someone to convince me that there is anything of non-monetary value derived from a business education. [ENDQ] [NEWLINE] Know<mask> mathematicians and economists were able to develop major theories in risk analysis and market fluctuations?<mask> of the relentless practice that goes on in Business.  It's an incredibly diverse field of study, not all that different than biology, or sociology.<mask> more importantly, it's all about putting your ideas into practice. [NEWLINE] [NEWLINE] <mask>'s the
Label encoding: <s> [STARTQ] you could easily argue that learning the bare-bones of business breaks down certain assumptions, etc. However, my impression of business school is not that one learns how businesses work, but rather how to make businesses work. It's not a study as much as it is instruction. [ENDQ] [NEWLINE] Unfortunately you're impression is not the reality. [NEWLINE] [NEWLINE] Business school is HOW businesses work. It's no different than studying any other field. [NEWLINE] [NEWLINE] There are a lot of different businesses, business strategies, management strategies, etc. And there are examples of excellently done strategies, and poorly done strategies. [NEWLINE] [NEWLINE] Case study was the bread and butter of my classes.  My senior management class had all of us break off into teams and play a simulator to see who could run a successful business. [NEWLINE] [NEWLINE] [STARTQ] I don't see how if I claim that business isn't a legitimate discipline, I must also accept that social sciences and humanities are not legitimate... I think my point would be best understood as saying that business school uses these disciplines only to promote generation of wealth, rather than contribute anything meaningful to them. We can point to mathematicians and economists as responsible for major theories in risk analysis and market fluctuations - businessmen simply implement them. [ENDQ] [NEWLINE] Because what's the difference between studying them and studying business? Really? [NEWLINE] [NEWLINE] I think without implementation those disciplines would be sorely lacking in any concrete info, or data sets to build their models off of. [NEWLINE] [NEWLINE] Ideas are nice, but practice is where shit gets done. That thesis on observed student's and how to elicit certain behaviors  may be nice; but how could you develop the necessary data to prove it's a viable or true among adults at large? *You put it into practice*. [NEWLINE] [NEWLINE] Business, at a minimum, is the field of of putting things into practice. And studying business, is studying how to put *anything* into practice, and ideally make a living off of it. [NEWLINE] [NEWLINE] [STARTQ] No, I'm simply saying that it is not and should not be treated as an academic discipline. I'm asking someone to convince me that there is anything of non-monetary value derived from a business education. [ENDQ] [NEWLINE] Know why mathematicians and economists were able to develop major theories in risk analysis and market fluctuations? Because of the relentless practice that goes on in Business.  It's an incredibly diverse field of study, not all that different than biology, or sociology. But more importantly, it's all about putting your ideas into practice. [NEWLINE] [NEWLINE] What's the
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A large part of<mask> makes things either frightening or funny is<mask> unexpected they are. Once we've been repeatedly exposed to certain jokes or frightening situations they start to lose their effect. In a sense, we become inoculated to them. This is well understood with humor through the 'Seinfeld's Not Funny' trope. Seinfeld was a pathfinder that influenced a generation of comedy. Younger audiences looking back at Seinfeld often don't get<mask> was<mask> special about it<mask> they've grew up on media that was hugely influenced by it. This has led them to become desensitized to its setups, punchlines and style of humor before even encountering it. [NEWLINE] [NEWLINE] The same can be said about the effectiveness of horror media. It was in a formulaic lull before Blair Witch Project showed up. There was very little innovation beyond than applying old formulas to the new'monster of the week'. Audiences were becoming<mask> used to the emotional mechanisms and patterns employed n horror cinema that they just weren't working anymore. The Blair Witch Project was a massive culture shock. [NEWLINE] [NEWLINE] The Blair Witch Project is a pretty terrible movie by conventional standards,<mask> it wasn't concerned with being a good film in the classical sense and that's<mask> made it special. It really only cared about scaring the shit out of audiences, and it went about achieving that goal in an extremely unconventional way. It was<mask> unconventional that it literally created a new sub-genera of horror. It was the first well known example of a 'found footage' horror film, and it appeared at a very special time in history. [NEWLINE] [NEWLINE] The internet was still incredibly young back in 1999, and the directors littered it with websites, articles, and forum posts to support the idea that the events documented in the film had really happened. These posts were often the first things people found<mask> they would 'Yahoo' this weird movie they were hearing about. There was a stunning lack of critical thinking, and plenty of people who would normally know better were caught off guard. We just weren't accustomed to questioning<mask> we read on the internet<mask>. The end result was that<mask> you were willing to suspend your disbelief even a little bit, everything about this movie felt real and convincing. On top of that it had almost no distribution in the beginning and its inaccessibility added to the mystery. The true believers started this weird buzz unlike anything I've ever seen, and there was a level of deception with the early audiences that can only be compared to the legendary 'War of the Worlds'
Label encoding: <s>A large part of what makes things either frightening or funny is how unexpected they are. Once we've been repeatedly exposed to certain jokes or frightening situations they start to lose their effect. In a sense, we become inoculated to them. This is well understood with humor through the 'Seinfeld's Not Funny' trope. Seinfeld was a pathfinder that influenced a generation of comedy. Younger audiences looking back at Seinfeld often don't get what was so special about it because they've grew up on media that was hugely influenced by it. This has led them to become desensitized to its setups, punchlines and style of humor before even encountering it. [NEWLINE] [NEWLINE] The same can be said about the effectiveness of horror media. It was in a formulaic lull before Blair Witch Project showed up. There was very little innovation beyond than applying old formulas to the new'monster of the week'. Audiences were becoming so used to the emotional mechanisms and patterns employed n horror cinema that they just weren't working anymore. The Blair Witch Project was a massive culture shock. [NEWLINE] [NEWLINE] The Blair Witch Project is a pretty terrible movie by conventional standards, but it wasn't concerned with being a good film in the classical sense and that's what made it special. It really only cared about scaring the shit out of audiences, and it went about achieving that goal in an extremely unconventional way. It was so unconventional that it literally created a new sub-genera of horror. It was the first well known example of a 'found footage' horror film, and it appeared at a very special time in history. [NEWLINE] [NEWLINE] The internet was still incredibly young back in 1999, and the directors littered it with websites, articles, and forum posts to support the idea that the events documented in the film had really happened. These posts were often the first things people found when they would 'Yahoo' this weird movie they were hearing about. There was a stunning lack of critical thinking, and plenty of people who would normally know better were caught off guard. We just weren't accustomed to questioning what we read on the internet yet. The end result was that if you were willing to suspend your disbelief even a little bit, everything about this movie felt real and convincing. On top of that it had almost no distribution in the beginning and its inaccessibility added to the mystery. The true believers started this weird buzz unlike anything I've ever seen, and there was a level of deception with the early audiences that can only be compared to the legendary 'War of the Worlds'
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] there is no connection to reality there [ENDQ] [NEWLINE] Before I even begin,<mask> you want a good persuasive argument, the last thing you should tell someone is that they have no connection to reality. Reading that line made me want to disregard everything you said in the rest of your post. I decided to push through it,<mask> it took me a few hours to even want to give you the time of day. [NEWLINE] [NEWLINE] Now to continue... [NEWLINE] [NEWLINE] [STARTQ] Feminism has been about the liberation of men from the very beginning. [ENDQ] [NEWLINE] Feminism is about equal pay, reproductive rights, maternity leave, domestic violence, woman's suffrage, and sexual harassment. Yes some of these include liberation from men. Absolutely.<mask> that's not the main goal. That's not<mask> it's completely about like you claim. These are all issues of inequality against women. The focus of feminism is on our inequalities and<mask> that may include the liberation from men, it<mask> includes laws that give us the right to equal pay, job security, the right to vote, the right to determine<mask> happens to our bodies, etc. that have nothing to do with liberation from oppressive men. [NEWLINE] [NEWLINE] [STARTQ] No, the cause of an imbalanced ratio of men vs women in prison is largely<mask> of socializing men to use violence to solve their problems. Men commit the vast majority of violent crimes, and this is something that is true in nearly every culture globally,<mask> with less violent societies, there is less social pressure for men to be violent. [ENDQ] [NEWLINE] <mask> it is true that men are committing more violent crimes, I stated that women are getting off the hook more easily for the *same* crimes which is causing the unbalanced ratio. I was not implying that the ratio would be 50/50<mask> women were properly convicted,<mask> that the ratio is not<mask> it would be<mask> they weren't. [NEWLINE] [NEWLINE] [STARTQ] From this, I find it clear that you have no idea<mask> feminism is. [ENDQ] [NEWLINE] Again with the demeaning arguments here.<mask> you believe that I don't know<mask> feminism is, enlighten me. Don't put me down. Teach me the things I don't know about feminism without telling me that I don't understand it.<mask> you use these arguments you aren't teaching. You are putting the person on a defense which causes them to take the focus off of<mask> you are saying and defending themselves. It makes your argument far weaker<mask> it makes the person not even want to listen to<mask> you have to say. [NEWLINE]
Label encoding: <s> [STARTQ] there is no connection to reality there [ENDQ] [NEWLINE] Before I even begin, if you want a good persuasive argument, the last thing you should tell someone is that they have no connection to reality. Reading that line made me want to disregard everything you said in the rest of your post. I decided to push through it, but it took me a few hours to even want to give you the time of day. [NEWLINE] [NEWLINE] Now to continue... [NEWLINE] [NEWLINE] [STARTQ] Feminism has been about the liberation of men from the very beginning. [ENDQ] [NEWLINE] Feminism is about equal pay, reproductive rights, maternity leave, domestic violence, woman's suffrage, and sexual harassment. Yes some of these include liberation from men. Absolutely. But that's not the main goal. That's not what it's completely about like you claim. These are all issues of inequality against women. The focus of feminism is on our inequalities and while that may include the liberation from men, it also includes laws that give us the right to equal pay, job security, the right to vote, the right to determine what happens to our bodies, etc. that have nothing to do with liberation from oppressive men. [NEWLINE] [NEWLINE] [STARTQ] No, the cause of an imbalanced ratio of men vs women in prison is largely because of socializing men to use violence to solve their problems. Men commit the vast majority of violent crimes, and this is something that is true in nearly every culture globally, but with less violent societies, there is less social pressure for men to be violent. [ENDQ] [NEWLINE] While it is true that men are committing more violent crimes, I stated that women are getting off the hook more easily for the *same* crimes which is causing the unbalanced ratio. I was not implying that the ratio would be 50/50 if women were properly convicted, but that the ratio is not what it would be if they weren't. [NEWLINE] [NEWLINE] [STARTQ] From this, I find it clear that you have no idea what feminism is. [ENDQ] [NEWLINE] Again with the demeaning arguments here. If you believe that I don't know what feminism is, enlighten me. Don't put me down. Teach me the things I don't know about feminism without telling me that I don't understand it. When you use these arguments you aren't teaching. You are putting the person on a defense which causes them to take the focus off of what you are saying and defending themselves. It makes your argument far weaker because it makes the person not even want to listen to what you have to say. [NEWLINE]
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You haven't got the point, America codifies these things in a constitution that it is practically impossible for a sitting government to over turn. No other country has that. [NEWLINE] [NEWLINE] <mask><mask> you want to go down the lists we can: [NEWLINE] [NEWLINE] Australia [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] does not have explicit freedom of speech in any constitutional or statutory declaration of rights... [ENDQ] [NEWLINE] Council of Europe [NEWLINE] [STARTQ] The exercise of these freedoms,<mask> it carries with it duties and responsibilities, may be subject to such formalities, conditions, restrictions or penalties<mask> are prescribed by law and are necessary in a democratic society, in the interests of national security, territorial integrity or public safety, for the prevention of disorder or crime, for the protection of health or morals, for the protection of the reputation or the rights of others, for preventing the disclosure of information received in confidence, or for maintaining the authority and impartiality of the judiciary. [ENDQ] [NEWLINE] (You'll notice<mask> subjective many of those exceptions are.) [NEWLINE] [NEWLINE] Czech Republic [NEWLINE] [STARTQ] The freedom of expression and the right to seek and disseminate information may be limited by law in the case of measures necessary in a democratic society for protecting the rights and freedoms of others, the security of the State, public security, public health, and morals. [ENDQ] [NEWLINE] (Again, note the subjectivity) [NEWLINE] [NEWLINE] Denmark [NEWLINE] [STARTQ] Hate speech is illegal<mask><mask> the Danish Penal Code § 266(b): [ENDQ] [NEWLINE] [NEWLINE] Finnland [NEWLINE] [NEWLINE] [STARTQ] Blasphemy and hate speech are forbidden. The blasphemy law applies to all religions. The hate speech law protects people of different sexual orientations, races, skin colors, places of birth, national or ethnic origins, religions or beliefs and disabled people.[80] The sentence for committing these crimes could theoretically be imprisonment,<mask> during the modern juridical history the sentence has always been a fine. [ENDQ] [NEWLINE] France [NEWLINE] [STARTQ] The Pleven Act of 1972 (after Justice Minister René Pleven) prohibits incitement to hatred, discrimination, slander and racial insults.[84][85] The Gayssot Act of 1990 prohibits any racist, anti-Semite, or xenophobic activities, including Holocaust denial.[85] The Law of 30 December 2004 prohibits hatred against people<mask> of their gender, sexual orientation, or disability. [ENDQ] [NEWLINE] [NEWLINE] Germany [NEWLINE] [STARTQ] The press is regulated by the law of Germany<mask> well<mask> all 16 States of Germany. [ENDQ] [NEWLINE] Greece [NEWLINE] [STARTQ] The 14th article of the Constitution of Greece makes it an offence for the press to insult the
Label encoding: <s>You haven't got the point, America codifies these things in a constitution that it is practically impossible for a sitting government to over turn. No other country has that. [NEWLINE] [NEWLINE] But if you want to go down the lists we can: [NEWLINE] [NEWLINE] Australia [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] does not have explicit freedom of speech in any constitutional or statutory declaration of rights... [ENDQ] [NEWLINE] Council of Europe [NEWLINE] [STARTQ] The exercise of these freedoms, since it carries with it duties and responsibilities, may be subject to such formalities, conditions, restrictions or penalties as are prescribed by law and are necessary in a democratic society, in the interests of national security, territorial integrity or public safety, for the prevention of disorder or crime, for the protection of health or morals, for the protection of the reputation or the rights of others, for preventing the disclosure of information received in confidence, or for maintaining the authority and impartiality of the judiciary. [ENDQ] [NEWLINE] (You'll notice how subjective many of those exceptions are.) [NEWLINE] [NEWLINE] Czech Republic [NEWLINE] [STARTQ] The freedom of expression and the right to seek and disseminate information may be limited by law in the case of measures necessary in a democratic society for protecting the rights and freedoms of others, the security of the State, public security, public health, and morals. [ENDQ] [NEWLINE] (Again, note the subjectivity) [NEWLINE] [NEWLINE] Denmark [NEWLINE] [STARTQ] Hate speech is illegal according to the Danish Penal Code § 266(b): [ENDQ] [NEWLINE] [NEWLINE] Finnland [NEWLINE] [NEWLINE] [STARTQ] Blasphemy and hate speech are forbidden. The blasphemy law applies to all religions. The hate speech law protects people of different sexual orientations, races, skin colors, places of birth, national or ethnic origins, religions or beliefs and disabled people.[80] The sentence for committing these crimes could theoretically be imprisonment, but during the modern juridical history the sentence has always been a fine. [ENDQ] [NEWLINE] France [NEWLINE] [STARTQ] The Pleven Act of 1972 (after Justice Minister René Pleven) prohibits incitement to hatred, discrimination, slander and racial insults.[84][85] The Gayssot Act of 1990 prohibits any racist, anti-Semite, or xenophobic activities, including Holocaust denial.[85] The Law of 30 December 2004 prohibits hatred against people because of their gender, sexual orientation, or disability. [ENDQ] [NEWLINE] [NEWLINE] Germany [NEWLINE] [STARTQ] The press is regulated by the law of Germany as well as all 16 States of Germany. [ENDQ] [NEWLINE] Greece [NEWLINE] [STARTQ] The 14th article of the Constitution of Greece makes it an offence for the press to insult the
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One thing I believe you are missing is that the assumption that there are objective rules and qualifications, values and feelings that you can use to measure<mask> difficult it is for someone to come out<mask> bisexual, gay or<mask> have you. Situations all differ and people may feel more at ease in certain environments rather than in others. Not to mention people react to themselves and their sexuality differently than others do. Certain individuals may feel more comfortable with themselves than others no matter<mask> they fit on the scale either<mask> of their upbringing, surroundings or simply<mask> of their personality. [NEWLINE] [NEWLINE] <mask> already it is hard to take this particular oppressed group has it harder than this other particular oppressed group at any more than a mass experience, not a personal one, in your case it would be more-<mask> my experience<mask> this oppressed group is such and such, whereas my experience in seeing<mask> this other oppressed group is such and such. In other words your experience bisexuals (or in reality you and the people you pay attention to) have it harder than gays (again the people you pay attention to.) In other environments you might say something different. [NEWLINE] [NEWLINE] It's easy to say in general people think such and such,<mask> in reality many different people you label believe this or that actually don't believe it in the black and white terms you say. They might only believe that certain personalities of bisexuals are "sluts", basically they judge someone based on preconceived personality traits and they are labeled "slut". And another may notice that the bisexuals they know are "sluts"<mask> not necessarily all. And<mask> on and<mask> on. My point really is that myths about oppressed groups are vague and ambiguous and differ greatly from social group to group. That's a problem with generalizations, you have to keep them vague and general.<mask> you define generalizations too specific you lose a sense of reality and are simply labeling in black and white terms. Which in itself is a major cause of oppressive ways of thinking. Using your examples<mask> an example (hehe!) instead of saying bisexuals are more likely to cheat, are sluts, female bisexuals are just using it for male attention and are just attracted to everyone and using that<mask> a definite generalization it's a bit more accurate to say that bisexuals are generally seen<mask> promiscuous. The other two can<mask> be further generalized into bisexuals don't know their sexuality, they are just confused. [NEWLINE] [NEWLINE] <mask> we have bisexuals are: [NEWLINE] [NEWLINE] [NEWLINE] 1.  promiscuous [NEWLINE] [NEWLINE]
Label encoding: <s>One thing I believe you are missing is that the assumption that there are objective rules and qualifications, values and feelings that you can use to measure how difficult it is for someone to come out as bisexual, gay or what have you. Situations all differ and people may feel more at ease in certain environments rather than in others. Not to mention people react to themselves and their sexuality differently than others do. Certain individuals may feel more comfortable with themselves than others no matter where they fit on the scale either because of their upbringing, surroundings or simply because of their personality. [NEWLINE] [NEWLINE] So already it is hard to take this particular oppressed group has it harder than this other particular oppressed group at any more than a mass experience, not a personal one, in your case it would be more- so my experience AS this oppressed group is such and such, whereas my experience in seeing how this other oppressed group is such and such. In other words your experience bisexuals (or in reality you and the people you pay attention to) have it harder than gays (again the people you pay attention to.) In other environments you might say something different. [NEWLINE] [NEWLINE] It's easy to say in general people think such and such, but in reality many different people you label believe this or that actually don't believe it in the black and white terms you say. They might only believe that certain personalities of bisexuals are "sluts", basically they judge someone based on preconceived personality traits and they are labeled "slut". And another may notice that the bisexuals they know are "sluts" but not necessarily all. And so on and so on. My point really is that myths about oppressed groups are vague and ambiguous and differ greatly from social group to group. That's a problem with generalizations, you have to keep them vague and general. If you define generalizations too specific you lose a sense of reality and are simply labeling in black and white terms. Which in itself is a major cause of oppressive ways of thinking. Using your examples as an example (hehe!) instead of saying bisexuals are more likely to cheat, are sluts, female bisexuals are just using it for male attention and are just attracted to everyone and using that as a definite generalization it's a bit more accurate to say that bisexuals are generally seen as promiscuous. The other two can also be further generalized into bisexuals don't know their sexuality, they are just confused. [NEWLINE] [NEWLINE] So we have bisexuals are: [NEWLINE] [NEWLINE] [NEWLINE] 1.  promiscuous [NEWLINE] [NEWLINE]
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A farmer instead has incentive to make tomatoes that last longer, and can be grown year round. Making more beneficial tomatoes for society is asking farmers to sacrifice profits. [ENDQ] [NEWLINE] Error #1, assumption of benefit. You assume that a "good tasting, nutritional" tomato is<mask> is of most social value. Tomatoes that last longer can reach more people, poison them less, and make more efficient use of food (rather than throw out a larger percentage). Having them grown year round means there are tomatoes available year round. Even *<mask> * some reduced nutrition, it is better nutrition than not eating them at all<mask> they aren't available. Mass production of tomatoes likewise cuts costs and makes tomatoes more affordable<mask> that the poor can eat them (and improve nutrition), and everybody else can afford to eat more of them. [NEWLINE] [NEWLINE] The rest of your points have similar problems,<mask> I'll refrain from addressing them one by one. It appears to me that you've just never thought through the tradeoffs and net social value. You see the item itself in a bubble and don't consider its cost, its value, access to it, or the industry on the whole. [NEWLINE] [NEWLINE] One variation that I will address is the "ethically reasonable" meat production comment. Yes, you are correct that flat out profit aims for pure efficiency and may lead to unethical behaviour.<mask>, that is not a problem with profit motivation or capitalism in general; it is a problem with the costs and benefits tied to the activity. The solution tends to take the form of either collective regulation (via our collective government), or otherwise tying costs to the problems themselves (known<mask> externalities in economics). For example, carbon tax adds the cost of cleaning the environment to each transaction that generates the carbon that results in the need to clean it up. [NEWLINE] [NEWLINE] Based on your title too, you imply that there are other systems that would do better to help things "reach full potential". To be clear, the trend towards maximizing efficiency in lieu of other things is an inherent mathematical property. Natural selection works on it even. Issues like the Prisoners Dilemma and Tragedy of the Commons are not products of a given economic system,<mask> are fundamental properties of systems with social transactions between multiple stakeholders. It matters not<mask> economic system you use, they will always exist. [NEWLINE] [NEWLINE] For example, imagine a system that encourages you to spend all day making one really good tomato,<mask> tasty and nutritious<mask> you can imagine. OK,<mask><mask> do you survive? You could eat it
Label encoding: <s> [STARTQ] A farmer instead has incentive to make tomatoes that last longer, and can be grown year round. Making more beneficial tomatoes for society is asking farmers to sacrifice profits. [ENDQ] [NEWLINE] Error #1, assumption of benefit. You assume that a "good tasting, nutritional" tomato is what is of most social value. Tomatoes that last longer can reach more people, poison them less, and make more efficient use of food (rather than throw out a larger percentage). Having them grown year round means there are tomatoes available year round. Even * if * some reduced nutrition, it is better nutrition than not eating them at all because they aren't available. Mass production of tomatoes likewise cuts costs and makes tomatoes more affordable so that the poor can eat them (and improve nutrition), and everybody else can afford to eat more of them. [NEWLINE] [NEWLINE] The rest of your points have similar problems, so I'll refrain from addressing them one by one. It appears to me that you've just never thought through the tradeoffs and net social value. You see the item itself in a bubble and don't consider its cost, its value, access to it, or the industry on the whole. [NEWLINE] [NEWLINE] One variation that I will address is the "ethically reasonable" meat production comment. Yes, you are correct that flat out profit aims for pure efficiency and may lead to unethical behaviour. However, that is not a problem with profit motivation or capitalism in general; it is a problem with the costs and benefits tied to the activity. The solution tends to take the form of either collective regulation (via our collective government), or otherwise tying costs to the problems themselves (known as externalities in economics). For example, carbon tax adds the cost of cleaning the environment to each transaction that generates the carbon that results in the need to clean it up. [NEWLINE] [NEWLINE] Based on your title too, you imply that there are other systems that would do better to help things "reach full potential". To be clear, the trend towards maximizing efficiency in lieu of other things is an inherent mathematical property. Natural selection works on it even. Issues like the Prisoners Dilemma and Tragedy of the Commons are not products of a given economic system, but are fundamental properties of systems with social transactions between multiple stakeholders. It matters not what economic system you use, they will always exist. [NEWLINE] [NEWLINE] For example, imagine a system that encourages you to spend all day making one really good tomato, as tasty and nutritious as you can imagine. OK, so how do you survive? You could eat it
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is not easy (for me at the moment at least) to find studies that look at<mask> representation in media influences<mask> we think about and treat certain groups or<mask> certain groups behave and think about themselves. [NEWLINE] For many people it is obvious that<mask> for example a little black kid sees that the hero in the movies he watches is always a white guy, and people that look like him are only side-characters that might influence (not determine) the way he thinks about himself, the things he considers himself capable of and the things he goes on to pursue. And that media<mask> influences<mask> other people think he is capable of. Same could be said for representation of gender. This goes not only for movies<mask> media in general. [NEWLINE] [NEWLINE] The expectation we have of certain groups (gender, race, orientation etc.) come largely from the society we life in and the things we perceive<mask> normal or given. The alternative would be that we are born or naturally develop gender roles or stereotypes even without any interaction, to me that seems highly unlikely. I don't think that we are able to form our subconscious biases only by real world interaction without even being influenced by 'imaginary' content, like for example video games. [NEWLINE] [NEWLINE] <mask>, you are right, that just<mask> something seems obvious that doesn't mean that we don't need to go searching for scientific proof. [NEWLINE] [NEWLINE] Here is one study that I could find; [NEWLINE] [STARTQ] To give one example of this effect, we know that (for reason not relevant here) women and minorities are underrepresented in media especially<mask> it comes to characters that are professionally successful. There was a study done "that finds evidence of a self-esteem boosting effect of television for white boys,<mask> self-esteem damaging effects for white girls, black girls, and black boys.  One primary reason?  White boys see lots of white boys and men in the shows they watch.  And, not just that,<mask> they regularly see these characters and actors in positive, powerful, and central roles." ([quote]( [URL] /), [source]( [URL].abstract)). [ENDQ] [NEWLINE] A quote from a [2000 paper]( [URL] %253A10.1023%252FA%253A1007046204478.pdf?originUrl=http%3A%2F%2Flink.springer.com%2Farticle%2F10.1023%2FA%3A1007046204478&amp;token2=exp=1436542114~
Label encoding: <s>It is not easy (for me at the moment at least) to find studies that look at how representation in media influences how we think about and treat certain groups or how certain groups behave and think about themselves. [NEWLINE] For many people it is obvious that if for example a little black kid sees that the hero in the movies he watches is always a white guy, and people that look like him are only side-characters that might influence (not determine) the way he thinks about himself, the things he considers himself capable of and the things he goes on to pursue. And that media also influences what other people think he is capable of. Same could be said for representation of gender. This goes not only for movies but media in general. [NEWLINE] [NEWLINE] The expectation we have of certain groups (gender, race, orientation etc.) come largely from the society we life in and the things we perceive as normal or given. The alternative would be that we are born or naturally develop gender roles or stereotypes even without any interaction, to me that seems highly unlikely. I don't think that we are able to form our subconscious biases only by real world interaction without even being influenced by 'imaginary' content, like for example video games. [NEWLINE] [NEWLINE] However, you are right, that just because something seems obvious that doesn't mean that we don't need to go searching for scientific proof. [NEWLINE] [NEWLINE] Here is one study that I could find; [NEWLINE] [STARTQ] To give one example of this effect, we know that (for reason not relevant here) women and minorities are underrepresented in media especially when it comes to characters that are professionally successful. There was a study done "that finds evidence of a self-esteem boosting effect of television for white boys, but self-esteem damaging effects for white girls, black girls, and black boys.  One primary reason?  White boys see lots of white boys and men in the shows they watch.  And, not just that, but they regularly see these characters and actors in positive, powerful, and central roles." ([quote]( [URL] /), [source]( [URL].abstract)). [ENDQ] [NEWLINE] A quote from a [2000 paper]( [URL] %253A10.1023%252FA%253A1007046204478.pdf?originUrl=http%3A%2F%2Flink.springer.com%2Farticle%2F10.1023%2FA%3A1007046204478&amp;token2=exp=1436542114~
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Probably not.<mask> in that context (let's say you're hanging out with your buddies and the topic of bad video games comes up), your reason for saying that was probably for humor, or a sense of fraternity in making fun of something that you think is terrible. [ENDQ] [NEWLINE] <mask><mask> I say it on a YouTube video's comment section? Random anonymous imageboard? I'm not sure<mask> you think a person can't express an opinion just for its own sake, and they always have some ulterior motive for it. [NEWLINE] [NEWLINE] [STARTQ] The creators will never hear it; you aren't intending to humiliate. It will not serve the function of intimidating anyone. [ENDQ] [NEWLINE] Couldn't you<mask> say that about the OP commenting on a YouTube video? Seems like any criticism that isn't face-to-face or person-to-person will<mask> qualify<mask> non-intimidatory,<mask> you aren't speaking directly to the people responsible. [NEWLINE] [NEWLINE] [STARTQ] That said, depending on circumstance, it could make sense.<mask> you're really insulting the game humorously, in a way you are making fun of it, then<mask> shouldn't we call that an attempt at superiority? [ENDQ] [NEWLINE] <mask> at no point did I claim to be superior at making games? You got 6 deltas for this revelation and I really don't see<mask>. You never explained<mask> someone is claiming their superiority, or exactly<mask> they are claiming to be superior. [NEWLINE] [NEWLINE] [STARTQ] <mask> the creator comes in and says "Well, could you do better?", it wouldn't necessarily mean that your criticisms were incorrect,<mask> rather that,<mask> they were personally insulted, it makes sense to diffuse that humiliation and social tension by pointing out that you are just<mask> bad. [ENDQ] [NEWLINE] <mask> you are saying [NEWLINE] [NEWLINE] - it is possible (quite likely in most cases<mask><mask> ) that the person's criticism is correct, [NEWLINE] - the "can you do better response" only makes sense<mask> coming from the subject of criticism themselves. [NEWLINE] [NEWLINE] Otherwise I really don't understand<mask> the "humiliation" and "social tension" is coming from. I guess your underlying assumption is that<mask> people hear negative criticism they disagree with, their internal response is [NEWLINE] [NEWLINE] [STARTQ] "Ohmygosh, this person just said something I like sucks!<mask> he is correct in his assessment, then I am a fool for liking this thing. I feel incredibly tense and humiliated from this guy voicing an opinion I don't like, and must provide my own to ensure others do not accept his view
Label encoding: <s> [STARTQ] Probably not. Because in that context (let's say you're hanging out with your buddies and the topic of bad video games comes up), your reason for saying that was probably for humor, or a sense of fraternity in making fun of something that you think is terrible. [ENDQ] [NEWLINE] What if I say it on a YouTube video's comment section? Random anonymous imageboard? I'm not sure why you think a person can't express an opinion just for its own sake, and they always have some ulterior motive for it. [NEWLINE] [NEWLINE] [STARTQ] The creators will never hear it; you aren't intending to humiliate. It will not serve the function of intimidating anyone. [ENDQ] [NEWLINE] Couldn't you also say that about the OP commenting on a YouTube video? Seems like any criticism that isn't face-to-face or person-to-person will also qualify as non-intimidatory, as you aren't speaking directly to the people responsible. [NEWLINE] [NEWLINE] [STARTQ] That said, depending on circumstance, it could make sense. If you're really insulting the game humorously, in a way you are making fun of it, then why shouldn't we call that an attempt at superiority? [ENDQ] [NEWLINE] Because at no point did I claim to be superior at making games? You got 6 deltas for this revelation and I really don't see why. You never explained HOW someone is claiming their superiority, or exactly how they are claiming to be superior. [NEWLINE] [NEWLINE] [STARTQ] If the creator comes in and says "Well, could you do better?", it wouldn't necessarily mean that your criticisms were incorrect, but rather that, if they were personally insulted, it makes sense to diffuse that humiliation and social tension by pointing out that you are just as bad. [ENDQ] [NEWLINE] So you are saying [NEWLINE] [NEWLINE] - it is possible (quite likely in most cases IMO ) that the person's criticism is correct, [NEWLINE] - the "can you do better response" only makes sense when coming from the subject of criticism themselves. [NEWLINE] [NEWLINE] Otherwise I really don't understand where the "humiliation" and "social tension" is coming from. I guess your underlying assumption is that when people hear negative criticism they disagree with, their internal response is [NEWLINE] [NEWLINE] [STARTQ] "Ohmygosh, this person just said something I like sucks! If he is correct in his assessment, then I am a fool for liking this thing. I feel incredibly tense and humiliated from this guy voicing an opinion I don't like, and must provide my own to ensure others do not accept his view
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] To pair Korra, a seemingly strong female character, with anyone at all, would not do her justice. Avatar Korra is a strong, independent Water Tribe woman who don't need no man (or, uh, girl). She broke up with Mako in the second season<mask> she couldn't be with him and be Avatar at the same time, which is a theme consistent with The Last Airbender; Aang had to let go his earthly tethers to enter the Avatar State [ENDQ] [NEWLINE] Aang had to let go his earthly connections to open the seventh chakra.<mask> this doesn't mean all Avatars must never have romantic connections, the waterbender Avatar before Korra had a love interests, and Roku had a wife! [NEWLINE] [NEWLINE] And to further expand on the whole Aang thing, he very clearly *didn't* let Katara go. They kissed at the end of the season and<mask> you pay attention real close in LoK you can even spot their offspring. Aang's conversation with Avatar Yangchen on the Lion Turtle pretty much summed it up: the Avatar's job is *balance* and<mask> of that they cannot completly sever their Earthly connection.<mask><mask> Guru Patik was wrong about this aspect of the Avatar and was trying to bring Aang back to his Airbender roots. [NEWLINE] [NEWLINE] All that said, pairing Korra with someone romantically doesn't make her not a strong and independant woman. Both her and Asami are independant women who fell in love. Their relationship is clearly not codependent or anything like that. Korra didn't write for two whole years! [NEWLINE] [NEWLINE] [STARTQ] To pair Korra with a relatively minor character who was only significant<mask> it was convenient to the team,<mask><mask> gender, is objectively bad fiction. [ENDQ] [NEWLINE] Did you even watch Book 3? Korra and Asami spent something like 85% of that season together. Asami was a pretty important character even back in book 1! And<mask> she was pushed to the sideline in book 2 she came back with a bang and is just<mask> part of the krew<mask> Bolin. [NEWLINE] [NEWLINE] <mask> really,<mask> is pairing up a main character with a minor one (not that Asami was, mind) bad fiction? Should main character romantic interests only involve other main characters or something? [NEWLINE] [NEWLINE] [STARTQ] There is no way this ending wasn't fanservice. [ENDQ] [NEWLINE] <mask> Bryan put it...which fans? Plenty were in your camp and didn't want to see Korrasami become cannon, plenty more wanted to see Makorra
Label encoding: <s> [STARTQ] To pair Korra, a seemingly strong female character, with anyone at all, would not do her justice. Avatar Korra is a strong, independent Water Tribe woman who don't need no man (or, uh, girl). She broke up with Mako in the second season because she couldn't be with him and be Avatar at the same time, which is a theme consistent with The Last Airbender; Aang had to let go his earthly tethers to enter the Avatar State [ENDQ] [NEWLINE] Aang had to let go his earthly connections to open the seventh chakra. But this doesn't mean all Avatars must never have romantic connections, the waterbender Avatar before Korra had a love interests, and Roku had a wife! [NEWLINE] [NEWLINE] And to further expand on the whole Aang thing, he very clearly *didn't* let Katara go. They kissed at the end of the season and if you pay attention real close in LoK you can even spot their offspring. Aang's conversation with Avatar Yangchen on the Lion Turtle pretty much summed it up: the Avatar's job is *balance* and because of that they cannot completly sever their Earthly connection. I think Guru Patik was wrong about this aspect of the Avatar and was trying to bring Aang back to his Airbender roots. [NEWLINE] [NEWLINE] All that said, pairing Korra with someone romantically doesn't make her not a strong and independant woman. Both her and Asami are independant women who fell in love. Their relationship is clearly not codependent or anything like that. Korra didn't write for two whole years! [NEWLINE] [NEWLINE] [STARTQ] To pair Korra with a relatively minor character who was only significant when it was convenient to the team, regardless of gender, is objectively bad fiction. [ENDQ] [NEWLINE] Did you even watch Book 3? Korra and Asami spent something like 85% of that season together. Asami was a pretty important character even back in book 1! And while she was pushed to the sideline in book 2 she came back with a bang and is just as part of the krew as Bolin. [NEWLINE] [NEWLINE] But really, why is pairing up a main character with a minor one (not that Asami was, mind) bad fiction? Should main character romantic interests only involve other main characters or something? [NEWLINE] [NEWLINE] [STARTQ] There is no way this ending wasn't fanservice. [ENDQ] [NEWLINE] As Bryan put it...which fans? Plenty were in your camp and didn't want to see Korrasami become cannon, plenty more wanted to see Makorra
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I feel like people don't deserve to have money in our society<mask> they don't put forth anything that makes our society prosper. [ENDQ] [NEWLINE] It's not a question of being given money. It's about<mask> that money is able to get someone. Money in and of itself is useless. It serves no function unless it is utilized to purchase a product or service. For everyone, money allows them to put food on the table, put gas in their car to get to work/drop kids off to school, purchase cloths, pay rent etc. We can agree that everyone should have access to food and shelter<mask> without money, this task is (nearly) impossible. Money<mask> impacts health, more<mask> in the US than say in Canada or Europe<mask> healthcare is free of charge FOR ALL citizens<mask><mask> finance. It's a proven fact that those in poverty are at higher risk for health problems down the road, problems that they won't even be able to afford to fix<mask> of a lack of financial resources. All this to say that money isn't<mask> people actually need.<mask> they need (or to use your work,<mask> they *deserve*) are the things that money is able to provide (i.e. shelter, food, health, etc). A society cannot prosper<mask> it's members are living in a state of destitution. This is<mask> societies like those in socialist countries (e.g. Norway, Denmark, Australia, Canada) prosper<mask> a whole<mask> the basic needs of the poorest are still taken into account. [NEWLINE] [NEWLINE] [STARTQ] which,<mask> they knew that they wouldn't make enough money to support one anyways, then they shouldn't have had kids [ENDQ] [NEWLINE] This is a strange statement. I take it you're from the US (<mask> I may be wrong). You're making the argument that people should not be having kids<mask> they can't afford it.<mask>, just recently in a ruling in the US, companies are able to withdraw birth control from employees.<mask>, the US still has a highly belligerent view regarding reproductive health. It's not taught in schools, it's not promoted in public discussions, condoms are not made available for free in health centres, women are discouraged from getting abortions and people don't know their options.<mask> often, women will have kids<mask> they have access to little or no information on their reproductive health. This is a much deeper issue than you are making out to be. [NEWLINE] [NEWLINE] [STARTQ] <mask> the U.S is in $17 trillion in debt [ENDQ] [NEWLINE] This debt
Label encoding: <s> [STARTQ] I feel like people don't deserve to have money in our society if they don't put forth anything that makes our society prosper. [ENDQ] [NEWLINE] It's not a question of being given money. It's about what that money is able to get someone. Money in and of itself is useless. It serves no function unless it is utilized to purchase a product or service. For everyone, money allows them to put food on the table, put gas in their car to get to work/drop kids off to school, purchase cloths, pay rent etc. We can agree that everyone should have access to food and shelter but without money, this task is (nearly) impossible. Money also impacts health, more so in the US than say in Canada or Europe where healthcare is free of charge FOR ALL citizens regardless of finance. It's a proven fact that those in poverty are at higher risk for health problems down the road, problems that they won't even be able to afford to fix because of a lack of financial resources. All this to say that money isn't what people actually need. What they need (or to use your work, what they *deserve*) are the things that money is able to provide (i.e. shelter, food, health, etc). A society cannot prosper if it's members are living in a state of destitution. This is why societies like those in socialist countries (e.g. Norway, Denmark, Australia, Canada) prosper as a whole because the basic needs of the poorest are still taken into account. [NEWLINE] [NEWLINE] [STARTQ] which, if they knew that they wouldn't make enough money to support one anyways, then they shouldn't have had kids [ENDQ] [NEWLINE] This is a strange statement. I take it you're from the US ( though I may be wrong). You're making the argument that people should not be having kids if they can't afford it. Yet, just recently in a ruling in the US, companies are able to withdraw birth control from employees. Moreover, the US still has a highly belligerent view regarding reproductive health. It's not taught in schools, it's not promoted in public discussions, condoms are not made available for free in health centres, women are discouraged from getting abortions and people don't know their options. So often, women will have kids because they have access to little or no information on their reproductive health. This is a much deeper issue than you are making out to be. [NEWLINE] [NEWLINE] [STARTQ] when the U.S is in $17 trillion in debt [ENDQ] [NEWLINE] This debt
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Interestingly enough, they did exactly this in the U.K. about thirty or<mask> years ago.  It all seems logical and practical and like it should work.   Then they decided that legally, ownership was satisfied<mask> you could own something at all. <mask> handguns were turned in in huge numbers and you were allowed to keep rifles and shotguns.  Then it was shotguns only.  Then you could only keep them<mask> you had a proven use for them like on a farm or similar. [NEWLINE] [NEWLINE] Thirty years later and the inventers of most of our modern arms and the people who helped us the most in WWII are completely castrated.  Firearms are essentially gone and crime is out of control. [NEWLINE] [NEWLINE] <mask><mask> was the critical difference? <mask> was pointed out, it's that your argument pre-supposes that firearms are a priviledge and not a right.  You go into the creation of the list with a view that logically will lead you down this path. <mask> let's get into each statement/idea. <mask> some are good, and some are not.  And this isn't really abot<mask> you set up the argument<mask> much<mask> the individual points. [NEWLINE] [NEWLINE] 1 - This sounds great. <mask> this is alrady<mask> happens with concealed carry permits.  A smarter approach would be to make a nationwide standard and database for CCW permits (it's a hodge-podge of conflicting laws currently and some states accept others, some do not), <mask>, make a concerted effort to promote such permit ownership to our youths (say, training leading up to a full permit at adulthood) and<mask> part of self-defense programs and<mask> on.  We need more firearms in the hands of good, trained citizens. [NEWLINE] [NEWLINE] 2 - Absolutely.  We do this in California already to buy any firearm. <mask> the issue again is every state has its own rules and laws.  There needs to be one database and one standard for all states.  I know I'm a big proponent of States rights,<mask> in many ways, the patchwork of laws creates more headache that it's worth<mask> we're trying to tackle a natiowide problem. [NEWLINE] [NEWLINE] 3 - Unfortunately this is completely unworkable. <mask><mask> you can reload your own ammo for pennies a round, and hundreds of millions of rounds are already in private hands, it's never going to happen. [NEWLINE] [NEWLINE] 4 - Correct.  People will just load their own or drive across town
Label encoding: <s>Interestingly enough, they did exactly this in the U.K. about thirty or so years ago.  It all seems logical and practical and like it should work.   Then they decided that legally, ownership was satisfied if you could own something at all.  So handguns were turned in in huge numbers and you were allowed to keep rifles and shotguns.  Then it was shotguns only.  Then you could only keep them if you had a proven use for them like on a farm or similar. [NEWLINE] [NEWLINE] Thirty years later and the inventers of most of our modern arms and the people who helped us the most in WWII are completely castrated.  Firearms are essentially gone and crime is out of control. [NEWLINE] [NEWLINE] So what was the critical difference?  As was pointed out, it's that your argument pre-supposes that firearms are a priviledge and not a right.  You go into the creation of the list with a view that logically will lead you down this path.  But let's get into each statement/idea.  Because some are good, and some are not.  And this isn't really abot how you set up the argument so much as the individual points. [NEWLINE] [NEWLINE] 1 - This sounds great.  But this is alrady what happens with concealed carry permits.  A smarter approach would be to make a nationwide standard and database for CCW permits (it's a hodge-podge of conflicting laws currently and some states accept others, some do not),  Also, make a concerted effort to promote such permit ownership to our youths (say, training leading up to a full permit at adulthood) and as part of self-defense programs and so on.  We need more firearms in the hands of good, trained citizens. [NEWLINE] [NEWLINE] 2 - Absolutely.  We do this in California already to buy any firearm.  But the issue again is every state has its own rules and laws.  There needs to be one database and one standard for all states.  I know I'm a big proponent of States rights, but in many ways, the patchwork of laws creates more headache that it's worth when we're trying to tackle a natiowide problem. [NEWLINE] [NEWLINE] 3 - Unfortunately this is completely unworkable.  Given that you can reload your own ammo for pennies a round, and hundreds of millions of rounds are already in private hands, it's never going to happen. [NEWLINE] [NEWLINE] 4 - Correct.  People will just load their own or drive across town
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As for the stereotypes, most groups get something similar,<mask> they may range from silly to offensive (worth mentioning here that being silly doesn't make a stereotype acceptable). Black people in the U.S., especially in the south, have some hateful stereotypes that need not be repeated,<mask><mask> some less offensive ones, e.g., they like watermelon and are great athletes. White people in the U.S. can't dance, jump, or play basketball. Some are labeled rednecks and stereotyped<mask> lazy and dumb inbreeders. English people have crooked, yellow teeth, spend all their time at the pub, and they all love football. Italians are loud, in your face, and use excessive hand gestures, and they love spicy meatballs. Germans get Nazi jokes,<mask> they have good taste in beer. Canadians are overly friendly and abuse maple syrup like drug addicts. Indians smell like curry and they all worship cows. The list could go on and on. My point is that all groups are stereotyped (we didn't even get into gender, sexuality, religion, age, etc.). Obviously stereotyping is not a fair thing,<mask><mask> others have said, there are some more serious issues out there. Being labeled<mask> "good at math, great with technology, and bad drivers" may be wrong,<mask>,<mask><mask><mask>, it takes a backseat to being labeled<mask> unfit for college, incapable of a job, likely a rapist, murderer, thief, or other type of criminal on the basis of one's skin color. This may be a bit of an aside,<mask><mask><mask> it's key to call people out<mask> you hear a stereotype. It could be<mask> simple<mask> "That's textbook racial stereotyping," or "That's a racial stereotype and I don't appreciate you labeling me [or others] like that."<mask><mask> we don't call people out on these things in our day-to-day then I don't think we can expect much to change. [NEWLINE] [NEWLINE] [STARTQ] I kinda get the feeling Asians tend to be higher achievers<mask> we're systematically forced to compete and fight against one another for pursuing similar fields which seems incredibly bogus to me. [ENDQ] [NEWLINE] This may well be the case,<mask> my Korean roommate, with whom I've had many discussions about all manner of things, including social injustices, attributes Asians' high achievement rates to Asian cultures. He says that he is expected to graduate college ASAP, get<mask> high a paying job<mask> possible, and start supporting his parents<mask>
Label encoding: <s>As for the stereotypes, most groups get something similar, although they may range from silly to offensive (worth mentioning here that being silly doesn't make a stereotype acceptable). Black people in the U.S., especially in the south, have some hateful stereotypes that need not be repeated, but also some less offensive ones, e.g., they like watermelon and are great athletes. White people in the U.S. can't dance, jump, or play basketball. Some are labeled rednecks and stereotyped as lazy and dumb inbreeders. English people have crooked, yellow teeth, spend all their time at the pub, and they all love football. Italians are loud, in your face, and use excessive hand gestures, and they love spicy meatballs. Germans get Nazi jokes, but they have good taste in beer. Canadians are overly friendly and abuse maple syrup like drug addicts. Indians smell like curry and they all worship cows. The list could go on and on. My point is that all groups are stereotyped (we didn't even get into gender, sexuality, religion, age, etc.). Obviously stereotyping is not a fair thing, but as others have said, there are some more serious issues out there. Being labeled as "good at math, great with technology, and bad drivers" may be wrong, but, in my opinion, it takes a backseat to being labeled as unfit for college, incapable of a job, likely a rapist, murderer, thief, or other type of criminal on the basis of one's skin color. This may be a bit of an aside, but I think it's key to call people out when you hear a stereotype. It could be as simple as "That's textbook racial stereotyping," or "That's a racial stereotype and I don't appreciate you labeling me [or others] like that." But if we don't call people out on these things in our day-to-day then I don't think we can expect much to change. [NEWLINE] [NEWLINE] [STARTQ] I kinda get the feeling Asians tend to be higher achievers because we're systematically forced to compete and fight against one another for pursuing similar fields which seems incredibly bogus to me. [ENDQ] [NEWLINE] This may well be the case, but my Korean roommate, with whom I've had many discussions about all manner of things, including social injustices, attributes Asians' high achievement rates to Asian cultures. He says that he is expected to graduate college ASAP, get as high a paying job as possible, and start supporting his parents as
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> this attitude mostly comes from two places: One, many redditors are products of helicopter parenting and feel that the world should be about them and they shouldn't have to tolerate "annoying little ankle-biters." [ENDQ] [NEWLINE] <mask>, in this sense, are parents implying that everyone must tolerate their decisions<mask><mask> they don't have a choice? [NEWLINE] [NEWLINE] This reminds me of someone telling me to be more "patient"<mask> they hold the line up to continue shopping.<mask> right does someone have to tell someone else they have to tolerate something<mask> that one person chose to do the thing others find annoying? [NEWLINE] [NEWLINE] [STARTQ] And two, it has become cool to be a loner, to be against increasing the surplus population, to be anti-breeder. [ENDQ] [NEWLINE] In my Buddhist practice, being able to not be a loner<mask> rather "alone" is actually a rather important aspect of ones practice.<mask> many have kids to actually fill a void in themselves and seem to think it's their sole purpose on this planet. [NEWLINE] [NEWLINE] Generalizing aside, was it cool to not be a loner, before, and cool to increase the population? [NEWLINE] [NEWLINE] <mask> someone who walks alone, my question is, who thinks I am cool? I am alone. "Cool" is for a younger demographic. I am only 33. This is important in my next point. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask>,<mask><mask> its origins, this popular negative opinion of children is unhealthy and will damage this generation's youth and<mask> our future<mask> left unchecked. [ENDQ] [NEWLINE] I don't really grasp this negativity<mask> I don't think it's fully extends into life outside reddit. [NEWLINE] [NEWLINE] <mask><mask> many, on reddit, are of the age demographic<mask> they actually are still kids. I would put this number at 25. I don't think one has a grasp of who they are by 25 and their perspective's will change. [NEWLINE] [NEWLINE] [STARTQ] A child growing up thinking he is less important than everyone else could be just<mask>,<mask> not more, damaging than a child growing up thinking he is more important than everyone else. [ENDQ] [NEWLINE] All I ever hear about is "think of the children." It falls into my comments, above, about someone telling me to be patient. I am dictated by parents of those children. [NEWLINE] [NEWLINE] In Canada, Health Canada just went against studies and lied about Cannabis<mask> the video they had tested well with "parent demographic." [NEWLINE] [NEWLINE] <mask>, I'm not sure<mask> a child grows up feeling unimportant<mask> all we ever
Label encoding: <s> [STARTQ] I think this attitude mostly comes from two places: One, many redditors are products of helicopter parenting and feel that the world should be about them and they shouldn't have to tolerate "annoying little ankle-biters." [ENDQ] [NEWLINE] So, in this sense, are parents implying that everyone must tolerate their decisions as if they don't have a choice? [NEWLINE] [NEWLINE] This reminds me of someone telling me to be more "patient" as they hold the line up to continue shopping. What right does someone have to tell someone else they have to tolerate something because that one person chose to do the thing others find annoying? [NEWLINE] [NEWLINE] [STARTQ] And two, it has become cool to be a loner, to be against increasing the surplus population, to be anti-breeder. [ENDQ] [NEWLINE] In my Buddhist practice, being able to not be a loner but rather "alone" is actually a rather important aspect of ones practice. So many have kids to actually fill a void in themselves and seem to think it's their sole purpose on this planet. [NEWLINE] [NEWLINE] Generalizing aside, was it cool to not be a loner, before, and cool to increase the population? [NEWLINE] [NEWLINE] As someone who walks alone, my question is, who thinks I am cool? I am alone. "Cool" is for a younger demographic. I am only 33. This is important in my next point. [NEWLINE] [NEWLINE] [STARTQ] I think, regardless of its origins, this popular negative opinion of children is unhealthy and will damage this generation's youth and thus our future if left unchecked. [ENDQ] [NEWLINE] I don't really grasp this negativity as I don't think it's fully extends into life outside reddit. [NEWLINE] [NEWLINE] I think many, on reddit, are of the age demographic where they actually are still kids. I would put this number at 25. I don't think one has a grasp of who they are by 25 and their perspective's will change. [NEWLINE] [NEWLINE] [STARTQ] A child growing up thinking he is less important than everyone else could be just as, if not more, damaging than a child growing up thinking he is more important than everyone else. [ENDQ] [NEWLINE] All I ever hear about is "think of the children." It falls into my comments, above, about someone telling me to be patient. I am dictated by parents of those children. [NEWLINE] [NEWLINE] In Canada, Health Canada just went against studies and lied about Cannabis because the video they had tested well with "parent demographic." [NEWLINE] [NEWLINE] So, I'm not sure why a child grows up feeling unimportant when all we ever
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] This decision for the Jewish people to start a Zionist quest going back to the 1880, 60 years before the end of WWII was foolish and based on religious ideologies that were not rooted in rationality, after WWII they should have emigrated out to the allied countries which saved them from persecution, and joined those secular nations who were more than happen to take them.<mask> they had done this, they would not be here and Jews would be safer and happier worldwide<mask><mask><mask>. [ENDQ] [NEWLINE] <mask><mask> one of the articles you linked [(on Old Yishuv)]( [URL] #Revival) Jewish emigration to the biblical Land of Israel<mask> an escape from religious persecution in Europe began<mask> early<mask> 1492.<mask>, the formulation of modern Zionism was a response to increasing levels of anti semitism across Europe. It really began with the [Biluim]( [URL] ) from Russia, with the first wave of Aliyah in 1882, who emigrated alongside those from Yemen, and began farming and wine producing. [NEWLINE] [NEWLINE] Your claim that the aliyah movements from 1880 were **"foolish and based on religious ideologies that were not rooted in rationality"** is patently untrue and unfair. Yes, the decision to move to [Eretz Yisrael]( [URL] ) was based on religious beliefs and traditions,<mask> clearly a rational one based on both an actual and a perceived idea that they had a historical community there - not to mention that it makes sense to respond to threats against you<mask> of your religion to move somewhere that people thought that they would be safe.<mask> else would you have had these communities go, prior to 1945? There was persecution of Jews<mask> a religious group in [Eastern Europe]( [URL] )<mask> well<mask> persecution of Jews<mask> a [racial and ethnic minority]( [URL] #Modern_and_the_racial_antisemitism) in Western Europe. [NEWLINE] [NEWLINE] This leads on to my final point, which is that it is disingenuous to say that **"they should have emigrated out to the allied countries which saved them from persecution, and joined those secular nations who were more than happen to take them. "** There are [significant examples of anti-semitism]( [URL] #After_1945) in post ww2 Europe, which proves the point that Jewish people did not have a right to feel that their security would be ensured in a secular (read: Christian) country. The attacks in France and Belgium on Jewish people and businesses (Jewish, not Israeli) just goes to show that Jewish people
Label encoding: <s> [STARTQ] This decision for the Jewish people to start a Zionist quest going back to the 1880, 60 years before the end of WWII was foolish and based on religious ideologies that were not rooted in rationality, after WWII they should have emigrated out to the allied countries which saved them from persecution, and joined those secular nations who were more than happen to take them. If they had done this, they would not be here and Jews would be safer and happier worldwide in my opinion. [ENDQ] [NEWLINE] According to one of the articles you linked [(on Old Yishuv)]( [URL] #Revival) Jewish emigration to the biblical Land of Israel as an escape from religious persecution in Europe began as early as 1492. However, the formulation of modern Zionism was a response to increasing levels of anti semitism across Europe. It really began with the [Biluim]( [URL] ) from Russia, with the first wave of Aliyah in 1882, who emigrated alongside those from Yemen, and began farming and wine producing. [NEWLINE] [NEWLINE] Your claim that the aliyah movements from 1880 were **"foolish and based on religious ideologies that were not rooted in rationality"** is patently untrue and unfair. Yes, the decision to move to [Eretz Yisrael]( [URL] ) was based on religious beliefs and traditions, but clearly a rational one based on both an actual and a perceived idea that they had a historical community there - not to mention that it makes sense to respond to threats against you because of your religion to move somewhere that people thought that they would be safe. Where else would you have had these communities go, prior to 1945? There was persecution of Jews as a religious group in [Eastern Europe]( [URL] ) as well as persecution of Jews as a [racial and ethnic minority]( [URL] #Modern_and_the_racial_antisemitism) in Western Europe. [NEWLINE] [NEWLINE] This leads on to my final point, which is that it is disingenuous to say that **"they should have emigrated out to the allied countries which saved them from persecution, and joined those secular nations who were more than happen to take them. "** There are [significant examples of anti-semitism]( [URL] #After_1945) in post ww2 Europe, which proves the point that Jewish people did not have a right to feel that their security would be ensured in a secular (read: Christian) country. The attacks in France and Belgium on Jewish people and businesses (Jewish, not Israeli) just goes to show that Jewish people
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A few quick thoughts: [NEWLINE] [NEWLINE] - It's really hard to make a super convincing argument on marriage<mask> it's very fact specific, depending on who you are,<mask> you are in life,<mask> you want to go and with whom.  The extent to which individual piece of my argument is going to be compelling is going to depend on all of these things and I can't really exhaustively anticipate everybody's life situation.  That said, OP's primary arguments boiled marriage down to legitimizing sex and living together, and I was hoping that, by pointing out all the things marital status touches on, OP would realize he grossly oversimplified the considerations and possibilities of marriage, even<mask> they didn't tip the scales for him this very minute. [NEWLINE] [NEWLINE] - I still think everybody is ignoring the extent to which long-term cohabitation creates many of the same problems<mask> marriage<mask> without a clearly defined body of law to go alongside it, and it's relevant<mask> OP treats cohabitation<mask> interchangeable with marriage<mask> with less drawbacks,<mask>, really, we're just talking about balancing different trade-offs.  It's very nice to sit there and think you're totally going to keep everything separate for the next several decades (giving the benefit of the doubt to the relationship, marriage or not)<mask>, realistically, commingling happens<mask> a matter of course.  Homes, apartments and rent, tangible personal property, etc. are all things that just sort of eventually meld together over time - even<mask> you do some things like maintain separate bank accounts - and I really think it just defies reality to place all of the cost of ending a relationship into a marital context<mask> we deal with tons of this in long-term relationships, period.  I<mask> feel obligated to point out that<mask><mask> we're going to see more cohabitating couples in long-term relationships being held to more marital-like standards down the road,<mask> that's speculative food to chew on more than anything. [NEWLINE] [NEWLINE] - You're right that everybody should engage in estate planning, tax planning, have a will, healthcare proxy,<mask> most people *don't* and many can't afford to consistently approach their attorney, accountant, etc. to create these schemes and update them.  In this respect, marriage is a very egalitarian institution that allows people to create some kind of default scheme at low cost<mask> they have something that recognizes their relationship to another person and the rights and obligations that follow.  It's<mask> a different conversation from having marriage and
Label encoding: <s>A few quick thoughts: [NEWLINE] [NEWLINE] - It's really hard to make a super convincing argument on marriage because it's very fact specific, depending on who you are, where you are in life, where you want to go and with whom.  The extent to which individual piece of my argument is going to be compelling is going to depend on all of these things and I can't really exhaustively anticipate everybody's life situation.  That said, OP's primary arguments boiled marriage down to legitimizing sex and living together, and I was hoping that, by pointing out all the things marital status touches on, OP would realize he grossly oversimplified the considerations and possibilities of marriage, even if they didn't tip the scales for him this very minute. [NEWLINE] [NEWLINE] - I still think everybody is ignoring the extent to which long-term cohabitation creates many of the same problems as marriage but without a clearly defined body of law to go alongside it, and it's relevant because OP treats cohabitation as interchangeable with marriage but with less drawbacks, when, really, we're just talking about balancing different trade-offs.  It's very nice to sit there and think you're totally going to keep everything separate for the next several decades (giving the benefit of the doubt to the relationship, marriage or not) but, realistically, commingling happens as a matter of course.  Homes, apartments and rent, tangible personal property, etc. are all things that just sort of eventually meld together over time - even if you do some things like maintain separate bank accounts - and I really think it just defies reality to place all of the cost of ending a relationship into a marital context when we deal with tons of this in long-term relationships, period.  I also feel obligated to point out that I think we're going to see more cohabitating couples in long-term relationships being held to more marital-like standards down the road, but that's speculative food to chew on more than anything. [NEWLINE] [NEWLINE] - You're right that everybody should engage in estate planning, tax planning, have a will, healthcare proxy, but most people *don't* and many can't afford to consistently approach their attorney, accountant, etc. to create these schemes and update them.  In this respect, marriage is a very egalitarian institution that allows people to create some kind of default scheme at low cost so they have something that recognizes their relationship to another person and the rights and obligations that follow.  It's also a different conversation from having marriage and
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I guess I don't see<mask> the fact that [WWI] was bad for everyone demonstrates non-self-interest,<mask><mask> it's possible for self-interest to leave everyone worse off. [ENDQ] [NEWLINE] It would be one thing<mask> self-interest led countries to make narrow, short-term decisions that they later came to regret,<mask> that wasn't really the case. Instead, Germany, Russia, France &amp; Britain all perceived at the time that their commitments and alliances were drawing them inexorably toward a larger out-of-control conflict that none of them wanted, could afford, or could likely profit from.<mask> their inability to free themselves from the prevailing doctrine of balance of power caused them to take actions they knew were against their own interests. [NEWLINE] [NEWLINE] [STARTQ] It doesn't take ideology to figure that it makes sense to increase our country's arms, given<mask> the other countries are doing. [ENDQ] [NEWLINE] Yes,<mask><mask> doing<mask> leads directly to increasing the likelihood of war,<mask> the opposite is<mask> you intend, then you can't call that self-interest. Compare to a related concept, [the tragedy of the commons]( [URL] ), whereby individuals acting in their individual self interest wind up depleting a common resource (air pollution, overfishing a fishery, deforestation, etc.). Tragedy of the commons *is* driven by self interest: individuals benefit<mask> the environment and other/future individuals suffer. The tragedy of the commons accords with your notion that everything is driven by self interest. The security dilemma<mask> does not. A nation acting to protect itself by ramping up military spending is trying to increase its own stability by increasing the security of its people and armed forces.<mask> doing<mask> in certain cases directly undermines that self interested aim by increasing the threats to its security from other nations. Foregoing an arms race and resisting the impulse to build up would leave the nation more secure and the government more stable. This is exactly the logic behind the SALT and START treaties between the US and Russia/USSR... the logic of nuclear war leads each nation towards an impulse towards an arms race with its rival.<mask> leaders in both nations could see that that impulse was inexorably making a nuclear conflict ever more likely,<mask> they intervened to reverse the process. [NEWLINE] [NEWLINE] [STARTQ] Did the Nazis believe that the Slavs, Communists, and Jews posed a danger to the Nazi state? [ENDQ] [NEWLINE] Yes,<mask><mask> they did.<mask> you have to draw a distinction between illusory self interest and actual self
Label encoding: <s> [STARTQ] I guess I don't see how the fact that [WWI] was bad for everyone demonstrates non-self-interest, given that it's possible for self-interest to leave everyone worse off. [ENDQ] [NEWLINE] It would be one thing if self-interest led countries to make narrow, short-term decisions that they later came to regret, but that wasn't really the case. Instead, Germany, Russia, France &amp; Britain all perceived at the time that their commitments and alliances were drawing them inexorably toward a larger out-of-control conflict that none of them wanted, could afford, or could likely profit from. Yet their inability to free themselves from the prevailing doctrine of balance of power caused them to take actions they knew were against their own interests. [NEWLINE] [NEWLINE] [STARTQ] It doesn't take ideology to figure that it makes sense to increase our country's arms, given what the other countries are doing. [ENDQ] [NEWLINE] Yes, but when doing so leads directly to increasing the likelihood of war, when the opposite is what you intend, then you can't call that self-interest. Compare to a related concept, [the tragedy of the commons]( [URL] ), whereby individuals acting in their individual self interest wind up depleting a common resource (air pollution, overfishing a fishery, deforestation, etc.). Tragedy of the commons *is* driven by self interest: individuals benefit but the environment and other/future individuals suffer. The tragedy of the commons accords with your notion that everything is driven by self interest. The security dilemma however does not. A nation acting to protect itself by ramping up military spending is trying to increase its own stability by increasing the security of its people and armed forces. But doing so in certain cases directly undermines that self interested aim by increasing the threats to its security from other nations. Foregoing an arms race and resisting the impulse to build up would leave the nation more secure and the government more stable. This is exactly the logic behind the SALT and START treaties between the US and Russia/USSR... the logic of nuclear war leads each nation towards an impulse towards an arms race with its rival. But leaders in both nations could see that that impulse was inexorably making a nuclear conflict ever more likely, so they intervened to reverse the process. [NEWLINE] [NEWLINE] [STARTQ] Did the Nazis believe that the Slavs, Communists, and Jews posed a danger to the Nazi state? [ENDQ] [NEWLINE] Yes, I think they did. But you have to draw a distinction between illusory self interest and actual self
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi Reddit. With all the news over the past couple weeks regarding Caitlyn Jenner’s transition and coming out, I feel now is a good time to discuss gender’s role in society and its position in<mask> people identify<mask> an individual. I would like to preface myself by saying that I identify<mask> a pansexual cis-male, and I certainly have zero issues with people that are transgendered. I do,<mask>, get very confused<mask> to *<mask> * some people are transgendered. [NEWLINE] [NEWLINE] [NEWLINE] Many people use the phrase, “I am a male trapped in a female’s body,” or, “my body is male<mask> my brain is female.”<mask> does this even mean? From a very young age, people are raised with ideas imposed upon us that dictate<mask> boys are and<mask> girls are, and these perceptions have nothing to do with the biological sex of someone and are typically based upon baseless stereotypical personality traits or<mask> that gender should like to do. We are taught boys should play with toy cars and tools<mask> girls are supposed to like barbies and dresses. These stereotypes are completely disparate from the reality that people are often multi-faceted, and might like many things across several disciplines and across several “gender” boundaries. I, personally, love to wrench my own car, build things, and in the same day I might decide to look totally different and wear makeup<mask> an artistic outlet or go out shopping with friends to look at heels. These interests have nothing to do with my biological sex, and shouldn’t have an affect on<mask> I identify. [NEWLINE] [NEWLINE] [NEWLINE] Society has dictated that girls do certain things and boys do certain things.<mask> does this mean<mask> someone who is anatomically male tends to like a lot of the things that the opposite gender stereotypically likes?<mask><mask> it’s unfair to oneself to say, “I like girl things<mask><mask> I’m a girl,” or, “I like looking like a girl<mask> I’m a girl.” This idea of outward presentation equating to one’s gender identity seems restrictive to me. I question<mask> many people would identify<mask> transgender<mask> these societal norms became antiquated and everyone was truly free to express all sides of themselves<mask><mask> physical sex. It should not be odd to be a very “feminine” person (makeup, hair, dress, etc) and be anatomically
Label encoding: <s>Hi Reddit. With all the news over the past couple weeks regarding Caitlyn Jenner’s transition and coming out, I feel now is a good time to discuss gender’s role in society and its position in how people identify as an individual. I would like to preface myself by saying that I identify as a pansexual cis-male, and I certainly have zero issues with people that are transgendered. I do, however, get very confused as to * why * some people are transgendered. [NEWLINE] [NEWLINE] [NEWLINE] Many people use the phrase, “I am a male trapped in a female’s body,” or, “my body is male but my brain is female.” What does this even mean? From a very young age, people are raised with ideas imposed upon us that dictate what boys are and what girls are, and these perceptions have nothing to do with the biological sex of someone and are typically based upon baseless stereotypical personality traits or what that gender should like to do. We are taught boys should play with toy cars and tools while girls are supposed to like barbies and dresses. These stereotypes are completely disparate from the reality that people are often multi-faceted, and might like many things across several disciplines and across several “gender” boundaries. I, personally, love to wrench my own car, build things, and in the same day I might decide to look totally different and wear makeup as an artistic outlet or go out shopping with friends to look at heels. These interests have nothing to do with my biological sex, and shouldn’t have an affect on how I identify. [NEWLINE] [NEWLINE] [NEWLINE] Society has dictated that girls do certain things and boys do certain things. What does this mean when someone who is anatomically male tends to like a lot of the things that the opposite gender stereotypically likes? I think it’s unfair to oneself to say, “I like girl things so therefore I’m a girl,” or, “I like looking like a girl so I’m a girl.” This idea of outward presentation equating to one’s gender identity seems restrictive to me. I question how many people would identify as transgender if these societal norms became antiquated and everyone was truly free to express all sides of themselves regardless of physical sex. It should not be odd to be a very “feminine” person (makeup, hair, dress, etc) and be anatomically
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 500 billion planets [ENDQ] [NEWLINE] And here may be<mask> your perspective differs from others'. There aren't 500 billion planets in the observable universe. Forgive the bluntness,<mask> that is a shockingly egregious underestimation. [NEWLINE] [NEWLINE] There are actually somewhere around<mask> many *galaxies* in the observable universe<mask> you think there are planets. The estimate for number of planets is many orders of magnitude higher. [NEWLINE] [NEWLINE] Your estimate for the *total* number of planets in the entire universe is about equal to the number of *habitable* planets in the Milky Way alone (100 billion). That's not just "planets,"<mask> the number that could specifically support our type of life. [NEWLINE] [NEWLINE] And that's just one galaxy. [NEWLINE] [NEWLINE] -------------- [NEWLINE] [NEWLINE] Given [this estimate]( [URL] /#safe=active&amp;q=number+of+planets+in+the+universe) of 5×10^22 habitable planets, and given your hypothetical probability-of-life of one in a trillion, the probability (assuming independence) of there not being any other life in the universe is: [NEWLINE] [NEWLINE] (probability of not having life)^(number of habitable planets in the universe) [NEWLINE] [NEWLINE] =(1 - probability of having life)^(number of habitable planets in the universe) [NEWLINE] [NEWLINE] =(999,999,999,999/1,000,000,000,000)^(5×10^22) = 1/(10^(10^10.33674470818953)) [NEWLINE] [NEWLINE] That's one over a number with 1 x 10^10 *decimal digits*. For any reasonable application, this probability is zero. [NEWLINE] [NEWLINE] [wolfram alpha calculation]( [URL] /?i=%28999%2C999%2C999%2C999%2F1%2C000%2C000%2C000%2C000%29%5E%285%C3%9710%5E22%29) [NEWLINE] [NEWLINE] <mask>, using your probability, we can say with exceptionally confident certainty that there is<mask><mask> other life in the universe. [NEWLINE] [NEWLINE] For fun, the expected number of inhabited planets in the universe would be: [NEWLINE] [NEWLINE] (number of habitable planets)*(probability of being inhabited) [NEWLINE] [NEWLINE] =(5 x 10^(22))*(1/one trillion) [NEWLINE] [NEWLINE] = **50 billion inhabited planets** [NEWLINE] [NEWLINE] [wolfram alpha calculation]( [URL] /?i=%285+x+10%5E
Label encoding: <s> [STARTQ] 500 billion planets [ENDQ] [NEWLINE] And here may be where your perspective differs from others'. There aren't 500 billion planets in the observable universe. Forgive the bluntness, but that is a shockingly egregious underestimation. [NEWLINE] [NEWLINE] There are actually somewhere around as many *galaxies* in the observable universe as you think there are planets. The estimate for number of planets is many orders of magnitude higher. [NEWLINE] [NEWLINE] Your estimate for the *total* number of planets in the entire universe is about equal to the number of *habitable* planets in the Milky Way alone (100 billion). That's not just "planets," but the number that could specifically support our type of life. [NEWLINE] [NEWLINE] And that's just one galaxy. [NEWLINE] [NEWLINE] -------------- [NEWLINE] [NEWLINE] Given [this estimate]( [URL] /#safe=active&amp;q=number+of+planets+in+the+universe) of 5×10^22 habitable planets, and given your hypothetical probability-of-life of one in a trillion, the probability (assuming independence) of there not being any other life in the universe is: [NEWLINE] [NEWLINE] (probability of not having life)^(number of habitable planets in the universe) [NEWLINE] [NEWLINE] =(1 - probability of having life)^(number of habitable planets in the universe) [NEWLINE] [NEWLINE] =(999,999,999,999/1,000,000,000,000)^(5×10^22) = 1/(10^(10^10.33674470818953)) [NEWLINE] [NEWLINE] That's one over a number with 1 x 10^10 *decimal digits*. For any reasonable application, this probability is zero. [NEWLINE] [NEWLINE] [wolfram alpha calculation]( [URL] /?i=%28999%2C999%2C999%2C999%2F1%2C000%2C000%2C000%2C000%29%5E%285%C3%9710%5E22%29) [NEWLINE] [NEWLINE] Thus, using your probability, we can say with exceptionally confident certainty that there is in fact other life in the universe. [NEWLINE] [NEWLINE] For fun, the expected number of inhabited planets in the universe would be: [NEWLINE] [NEWLINE] (number of habitable planets)*(probability of being inhabited) [NEWLINE] [NEWLINE] =(5 x 10^(22))*(1/one trillion) [NEWLINE] [NEWLINE] = **50 billion inhabited planets** [NEWLINE] [NEWLINE] [wolfram alpha calculation]( [URL] /?i=%285+x+10%5E
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A load to go over,<mask> I'll just clear the bullets then. [NEWLINE] [NEWLINE] A)  This idea appears to work from a perspective<mask> the nature of the porn is hardcore, bdsm, and like. <mask> it looks like you need some perspective,<mask> come along lassy on adventure! Lol [NEWLINE] <mask> young males discover porn, we're appeasing just this itty bitty obnoxious voice in the back of our skulls that have been nagging on about<mask> girls aren't<mask> bad<mask> we thought they were-- and at that point in life, even<mask> you had sex ed, you don't even watch porn for the sex really.  Honestly, masturbation isn't even a thing<mask>.  All you care for are breasted of rediculous preportions.  From<mask> I can remember, it was just the fact that they were there, and maybe<mask> we were told they were forbidden, right?  Then the whole nipple conspiracy begins, like "do they exist there," "are these bags of wonder just that or is there more?!" [NEWLINE] All said with a brother and no sisters mind you, and in primary schools, their sex ed course--I remember, 5th grade man. [NEWLINE] The whole level of shenanigans was just great.  Like they just put the gym teacher up at the front of the room telling us about our "special package" and that was it.  Like that's all we got out of it, training in penis induendos and the realization that we still had no idea<mask> laid on the other side of the spectrum. They never even brought it up in that cla-- well to be fair, we probably wouldn't have learned anything anyways. Lol<mask> anything, that side split anatomy diagram of a vagina would've just confused us to a limitless extent. [NEWLINE] <mask> yeah, at that point, that's all we knew. <mask>, at that age you don't even get off on porn, you just sit there and stare in awe<mask> you think the boy from A Christmas Story might. [NEWLINE] <mask>, you give it a few years then, and this moment comes<mask> you move on from gazing at pics of breasts for reasons you can't even define, to eventually coming across videos, which depending on<mask> time in your life you do<mask> could confuse the hell out of you, and maybe even scare you. xD [NEWLINE] <mask> once your hormones click into shape you watch it more for this mysterious action towards the back simply out of curiosity.  It's still weird<mask> your
Label encoding: <s>A load to go over, so I'll just clear the bullets then. [NEWLINE] [NEWLINE] A)  This idea appears to work from a perspective where the nature of the porn is hardcore, bdsm, and like.  But it looks like you need some perspective, so come along lassy on adventure! Lol [NEWLINE] As young males discover porn, we're appeasing just this itty bitty obnoxious voice in the back of our skulls that have been nagging on about how girls aren't as bad as we thought they were-- and at that point in life, even if you had sex ed, you don't even watch porn for the sex really.  Honestly, masturbation isn't even a thing yet.  All you care for are breasted of rediculous preportions.  From what I can remember, it was just the fact that they were there, and maybe how we were told they were forbidden, right?  Then the whole nipple conspiracy begins, like "do they exist there," "are these bags of wonder just that or is there more?!" [NEWLINE] All said with a brother and no sisters mind you, and in primary schools, their sex ed course--I remember, 5th grade man. [NEWLINE] The whole level of shenanigans was just great.  Like they just put the gym teacher up at the front of the room telling us about our "special package" and that was it.  Like that's all we got out of it, training in penis induendos and the realization that we still had no idea what laid on the other side of the spectrum. They never even brought it up in that cla-- well to be fair, we probably wouldn't have learned anything anyways. Lol if anything, that side split anatomy diagram of a vagina would've just confused us to a limitless extent. [NEWLINE] So yeah, at that point, that's all we knew.  So, at that age you don't even get off on porn, you just sit there and stare in awe how you think the boy from A Christmas Story might. [NEWLINE] So, you give it a few years then, and this moment comes where you move on from gazing at pics of breasts for reasons you can't even define, to eventually coming across videos, which depending on what time in your life you do so could confuse the hell out of you, and maybe even scare you. xD [NEWLINE] But once your hormones click into shape you watch it more for this mysterious action towards the back simply out of curiosity.  It's still weird but your
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*Edit: Thanks for all the info. I more or less quit responding<mask> a couple people effectively changed the way I was looking at the issue. To be honest<mask>, I don't think I'll come back here again. I wanted to have a discussion, and it was nice to talk with some of you,<mask> I don't have enough comment karma to spare for all the downvotes I got<mask> responding to comments. That seems to be a theme in most posts in this sub. [NEWLINE] [NEWLINE] Like I said, I appreciate all the responses, and a big thanks to those who took the time to respond to me and add sources and information to back up their reasoning.* [NEWLINE] [NEWLINE] I'm really struggling with this one, CMV. This is my first post here. Reading through threads concerning this issue here on Reddit has made me wonder<mask> I'm just a heartless jackass or not<mask> it appears most of Reddit goes against the grain with me on this issue here. [NEWLINE] [NEWLINE] I don't feel like it's wrong to drug test welfare recipients.<mask><mask> somewhat with the view that<mask> I am required to pass a drug test to even have a shot at a job, a welfare recipient should have to do the same to receive assistance which they're not required to work for. I've heard that a 30 day rehabilitation option is offered in some states for people who test positive, and benefits are cut off for those who refuse that option. I don't agree with punishing someone for addiction and not offering help,<mask><mask> the drug testing states implemented this policy, I don't think I'd see any problems with it at all. [NEWLINE] [NEWLINE] I feel this way<mask> I've been in a lot of situations<mask> this has been a relevant issue. Cashiering for a small town grocery store growing up, knowing the people using food stamps at the counter, and<mask> knowing that the majority<mask> them are using drugs (small town) helped me form an opinion early on. I didn't think it was fair that at 16, my tax money was being used to help someone pay for food<mask> they spent their money on drugs.<mask> they wanted food,<mask> not by it instead of drugs? Of course, now that I'm older I realize that not everything is<mask> black and white, and that this goes deeper than my 16 year old opinion. [NEWLINE] [NEWLINE] I'm a home health nurse that works with pediatric patients in their homes. My patients are almost always in very poor, run down neighborhoods. All of these families are on food stamps and
Label encoding: <s>*Edit: Thanks for all the info. I more or less quit responding because a couple people effectively changed the way I was looking at the issue. To be honest though, I don't think I'll come back here again. I wanted to have a discussion, and it was nice to talk with some of you, but I don't have enough comment karma to spare for all the downvotes I got when responding to comments. That seems to be a theme in most posts in this sub. [NEWLINE] [NEWLINE] Like I said, I appreciate all the responses, and a big thanks to those who took the time to respond to me and add sources and information to back up their reasoning.* [NEWLINE] [NEWLINE] I'm really struggling with this one, CMV. This is my first post here. Reading through threads concerning this issue here on Reddit has made me wonder if I'm just a heartless jackass or not since it appears most of Reddit goes against the grain with me on this issue here. [NEWLINE] [NEWLINE] I don't feel like it's wrong to drug test welfare recipients. I agree somewhat with the view that if I am required to pass a drug test to even have a shot at a job, a welfare recipient should have to do the same to receive assistance which they're not required to work for. I've heard that a 30 day rehabilitation option is offered in some states for people who test positive, and benefits are cut off for those who refuse that option. I don't agree with punishing someone for addiction and not offering help, but if the drug testing states implemented this policy, I don't think I'd see any problems with it at all. [NEWLINE] [NEWLINE] I feel this way because I've been in a lot of situations where this has been a relevant issue. Cashiering for a small town grocery store growing up, knowing the people using food stamps at the counter, and also knowing that the majority if them are using drugs (small town) helped me form an opinion early on. I didn't think it was fair that at 16, my tax money was being used to help someone pay for food when they spent their money on drugs. If they wanted food, why not by it instead of drugs? Of course, now that I'm older I realize that not everything is so black and white, and that this goes deeper than my 16 year old opinion. [NEWLINE] [NEWLINE] I'm a home health nurse that works with pediatric patients in their homes. My patients are almost always in very poor, run down neighborhoods. All of these families are on food stamps and
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I thought about posting this in Men's Rights (<mask> it's three times bigger than Feminism and seems a lot more active),<mask> I don't think I'd get the kind of open-minded discussion I'm looking for there. [NEWLINE] [NEWLINE] Whenever I'm reading through a thread linking to an article that depicts one sex in a negative light (rape, domestic abuse, false rape accusations, etc.), it seems inevitable that I come across some MRA/feminism discussion somewhere in the comments.  The first few times I saw this happening, I tried to follow the discussion for any compelling arguments,<mask> it always somehow spiraled into absurdity.   I started skipping over most of those sections whenever I came across them, writing them off<mask> a waste of time. [NEWLINE] [NEWLINE] Until one day<mask> I was reading one of these threads, I came across someone who said, "I really only read the Men's Rights sub<mask> I'm on reddit."  A quick look in his comment history proved his statement to be true.  Then I started doing that whenever I came across a sexism discussion on a thread I was reading.  I'd say 7 out of 10 times, the user's comment history had 80% of their comments in either /r/mensrights, /r/feminism, or /r/shitredditsays... (usually the former two subs,<mask> SRS comments are always downvoted to the bottom of most threads). [NEWLINE] [NEWLINE] [NEWLINE] I've looked through several threads on both subreddits, and aside from links to interesting articles, I guess I just don't see the good that they're doing.  The rational, intelligent conversations in those places are few and far in between... name-calling and berating seem to be commonplace, and this definitely spills out into the rest of reddit. [NEWLINE] [NEWLINE] [NEWLINE] Have I misinterpretted the point of these places?  I see them<mask> a little destructive and somewhat counter-productive to their cause.  Some of these users seem like impressionable people who have spent<mask> much time in these places that their views are completely shaped by them. [NEWLINE] [NEWLINE] [NEWLINE] <mask> a disclaimer:  I'm not gonna pretend I know<mask> it's like<mask> a man in this day and age, just<mask> I don't think it's possible for a man to know<mask> it's like to be a woman.  I acknowledge there are assholes, idiots, and all-around terrible examples of both sexes and that these people do not represent either sex<mask> a whole. [NEWLINE] [NEWLINE] I
Label encoding: <s>I thought about posting this in Men's Rights ( since it's three times bigger than Feminism and seems a lot more active), but I don't think I'd get the kind of open-minded discussion I'm looking for there. [NEWLINE] [NEWLINE] Whenever I'm reading through a thread linking to an article that depicts one sex in a negative light (rape, domestic abuse, false rape accusations, etc.), it seems inevitable that I come across some MRA/feminism discussion somewhere in the comments.  The first few times I saw this happening, I tried to follow the discussion for any compelling arguments, but it always somehow spiraled into absurdity.   I started skipping over most of those sections whenever I came across them, writing them off as a waste of time. [NEWLINE] [NEWLINE] Until one day when I was reading one of these threads, I came across someone who said, "I really only read the Men's Rights sub when I'm on reddit."  A quick look in his comment history proved his statement to be true.  Then I started doing that whenever I came across a sexism discussion on a thread I was reading.  I'd say 7 out of 10 times, the user's comment history had 80% of their comments in either /r/mensrights, /r/feminism, or /r/shitredditsays... (usually the former two subs, as SRS comments are always downvoted to the bottom of most threads). [NEWLINE] [NEWLINE] [NEWLINE] I've looked through several threads on both subreddits, and aside from links to interesting articles, I guess I just don't see the good that they're doing.  The rational, intelligent conversations in those places are few and far in between... name-calling and berating seem to be commonplace, and this definitely spills out into the rest of reddit. [NEWLINE] [NEWLINE] [NEWLINE] Have I misinterpretted the point of these places?  I see them as a little destructive and somewhat counter-productive to their cause.  Some of these users seem like impressionable people who have spent so much time in these places that their views are completely shaped by them. [NEWLINE] [NEWLINE] [NEWLINE] As a disclaimer:  I'm not gonna pretend I know what it's like as a man in this day and age, just as I don't think it's possible for a man to know what it's like to be a woman.  I acknowledge there are assholes, idiots, and all-around terrible examples of both sexes and that these people do not represent either sex as a whole. [NEWLINE] [NEWLINE] I
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's this weird quality that lies outside of the physical realm and cannot be justified by it (<mask> you can't easily leap from something existing to something being moral without something being taken<mask> axiomatic) and<mask> we are bound by it with threads made up of this<mask> strange material? [ENDQ] [NEWLINE] I don't think it's correct to think of morality<mask> something that *exists*. Rather, it's something that is *true*,<mask> an application of reason. In this way it is similar to math and logic. Those things are true and can have objectively right and wrong applications,<mask> we don't have to describe them<mask> something that "lies outside of the physical realm" (unless you're a platonist). [NEWLINE] [NEWLINE] Consider a person who slightly enjoys the sensation of falling, and very much dislikes pain. He jumps from a roof to enjoy the 2 seconds of falling, followed by breaking his leg and being in a large amount of pain for a long time. By his own preferences, he dislikes that pain much more than he likes the 2 seconds of falling. Before he jumps, you point out that this is a bad idea by his own preferences. Suppose he says: "Yes,<mask> my current-self will get<mask> he wants with no downside. It's only my future-self (in 2 seconds) that will be in pain. And the one making this decision is my current-self." Do you think there is something irrational about that way of thinking? [NEWLINE] [NEWLINE] <mask><mask><mask>. His reason for jumping is based on getting<mask> he wants.<mask> telling him it's irrational to jump is *<mask> * based on<mask> he wants, counting all the effects of that choice rather than ignoring some of them. The desires of his current and future self do not matter in some ultimate way to the universe,<mask> they matter *to him* both now and in the future. In the same way that jumping is a good decision for his present-self, jumping is an overall bad decision *all things considered*. [NEWLINE] [NEWLINE] In other words, people act for reasons, and reasons can be rational or irrational. This does not have to mean the rationality or irrationality of those reasons are some non-physical material. [NEWLINE] [NEWLINE] Now consider a similar situation to the guy who jumps,<mask> instead of getting something slightly good for his present-self in exchange for something very bad for his future-self, the same trade-off applies across 2 different people rather than across time for 1 person. For instance, maybe I
Label encoding: <s> [STARTQ] There's this weird quality that lies outside of the physical realm and cannot be justified by it ( because you can't easily leap from something existing to something being moral without something being taken as axiomatic) and yet we are bound by it with threads made up of this also strange material? [ENDQ] [NEWLINE] I don't think it's correct to think of morality as something that *exists*. Rather, it's something that is *true*, as an application of reason. In this way it is similar to math and logic. Those things are true and can have objectively right and wrong applications, yet we don't have to describe them as something that "lies outside of the physical realm" (unless you're a platonist). [NEWLINE] [NEWLINE] Consider a person who slightly enjoys the sensation of falling, and very much dislikes pain. He jumps from a roof to enjoy the 2 seconds of falling, followed by breaking his leg and being in a large amount of pain for a long time. By his own preferences, he dislikes that pain much more than he likes the 2 seconds of falling. Before he jumps, you point out that this is a bad idea by his own preferences. Suppose he says: "Yes, but my current-self will get what he wants with no downside. It's only my future-self (in 2 seconds) that will be in pain. And the one making this decision is my current-self." Do you think there is something irrational about that way of thinking? [NEWLINE] [NEWLINE] I think so. His reason for jumping is based on getting what he wants. But telling him it's irrational to jump is * also * based on what he wants, counting all the effects of that choice rather than ignoring some of them. The desires of his current and future self do not matter in some ultimate way to the universe, but they matter *to him* both now and in the future. In the same way that jumping is a good decision for his present-self, jumping is an overall bad decision *all things considered*. [NEWLINE] [NEWLINE] In other words, people act for reasons, and reasons can be rational or irrational. This does not have to mean the rationality or irrationality of those reasons are some non-physical material. [NEWLINE] [NEWLINE] Now consider a similar situation to the guy who jumps, but instead of getting something slightly good for his present-self in exchange for something very bad for his future-self, the same trade-off applies across 2 different people rather than across time for 1 person. For instance, maybe I
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] racism, bigotry, religious conservatism are not all that bad. [ENDQ] [NEWLINE] They aren't. A person can be racist, bigoted, and religiously conservative simultaneously and not be a bad person. Sure, these things can lead to violence,<mask><mask> can anything else. Violence is bad; simply holding a belief is not "bad" at all. For centuries, it was thought that whites were superior to blacks. Was every single person that thought this instantaneously the offspring of the devil? Morally bankrupt? [NEWLINE] [NEWLINE] I do concede that it is somewhat irrational to say that racism is not bad, I'm not an ignorant person.<mask> let me address your claim that religious conservatism. Here, you're setting up a political argument that liberals are somehow superior to conservatives. Do you really believe this? Would you tell that to the face of a religious conservative, that<mask> of their religious conservatism, they are a bad person and their *beliefs* should be "weeded out?" Ideological genocide is still Nazism. In particular, I get the feeling that you are attacking white, Southern Christians, which obviously are not a minority.<mask> makes it okay to attack the majority and advocate for their ideological elimination,<mask> wrong to attack the Jews (minority) and advocate for their ideological elimination? [NEWLINE] [NEWLINE] Just<mask> you aren't planning a mass extermination camp doesn't mean that you aren't a subscriber to Nazism. It's one thing to disagree,<mask> it's an entirely different thing to advocate for the extinction of an idea or belief. You wanted to have your view changed that views can be changed, which seems to me to be saying that<mask> you are willing to change your view, you are superior to others that will not change their view. Do you demand the elimination of "negative traits" like religious conservatism? It says in the Bible that homosexuality is bad (I'm not Christian and I'm not looking for a religious debate,<mask> this is not true then forgive me.):<mask> can you deny a person his religious views and force them to change? Gays are bad for all time,<mask><mask> Christianity, and this will not, and should not change just<mask> the majority now believes otherwise. [NEWLINE] [NEWLINE] [STARTQ] Once again, I'll state it - I am not advocating teaching the kids<mask> I believe to be correct. I just believe that<mask> we're to remove these kinds of traits from the society, [ENDQ] [NEWLINE] You seem to have missed the point. Thinkers of greater intellect than you have realized that it is nigh
Label encoding: <s> [STARTQ] racism, bigotry, religious conservatism are not all that bad. [ENDQ] [NEWLINE] They aren't. A person can be racist, bigoted, and religiously conservative simultaneously and not be a bad person. Sure, these things can lead to violence, but so can anything else. Violence is bad; simply holding a belief is not "bad" at all. For centuries, it was thought that whites were superior to blacks. Was every single person that thought this instantaneously the offspring of the devil? Morally bankrupt? [NEWLINE] [NEWLINE] I do concede that it is somewhat irrational to say that racism is not bad, I'm not an ignorant person. But let me address your claim that religious conservatism. Here, you're setting up a political argument that liberals are somehow superior to conservatives. Do you really believe this? Would you tell that to the face of a religious conservative, that because of their religious conservatism, they are a bad person and their *beliefs* should be "weeded out?" Ideological genocide is still Nazism. In particular, I get the feeling that you are attacking white, Southern Christians, which obviously are not a minority. What makes it okay to attack the majority and advocate for their ideological elimination, but wrong to attack the Jews (minority) and advocate for their ideological elimination? [NEWLINE] [NEWLINE] Just because you aren't planning a mass extermination camp doesn't mean that you aren't a subscriber to Nazism. It's one thing to disagree, but it's an entirely different thing to advocate for the extinction of an idea or belief. You wanted to have your view changed that views can be changed, which seems to me to be saying that because you are willing to change your view, you are superior to others that will not change their view. Do you demand the elimination of "negative traits" like religious conservatism? It says in the Bible that homosexuality is bad (I'm not Christian and I'm not looking for a religious debate, if this is not true then forgive me.): how can you deny a person his religious views and force them to change? Gays are bad for all time, according to Christianity, and this will not, and should not change just because the majority now believes otherwise. [NEWLINE] [NEWLINE] [STARTQ] Once again, I'll state it - I am not advocating teaching the kids what I believe to be correct. I just believe that if we're to remove these kinds of traits from the society, [ENDQ] [NEWLINE] You seem to have missed the point. Thinkers of greater intellect than you have realized that it is nigh
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I will try to explain myself without sounding like I'm just enticing the "PC Master Race" circlejerk. [NEWLINE] [NEWLINE] In the past, there were significant difference in hardware and software between consoles and home computers. Granted, older consoles used chips often found in PC's (the NES's CPU was based off the design of the one used in the Apple II and C64, Genesis used the Zilog Z80 And Motorola 68K, ect.)<mask> they produced results that greatly differed from one another in terms of processing, sound and graphics.<mask> with this current gen, both the Xbox One and PS4 use the same micro-architecture (x86) in their APU's that are manufactured by the same company, AMD, that is used in current PC's today. Many of both consoles' non-exclusives have<mask> been shown to underperform graphically<mask> pitted against desktops made for the same price<mask> them. [NEWLINE] [NEWLINE] [NEWLINE] The other point that I feel no longer separates consoles and PC's is their firmware and the need to constantly update it<mask> well<mask> patch games via internet. Again, with older consoles,<mask> I put a functioning copy of Super Metroid in any working SNES, it should play immediately, and that's it. I can understand with the 6th Generation's addition with a UI to adjust the clock, manage saves and other basic functions,<mask><mask> there were additions like internet browsers, multimedia players, Streaming Video, and other services that added nothing to one's gameplay experiences, it seemed like a waste of resources, and redundant<mask> my computer can do the same things. [NEWLINE] [NEWLINE] And on the subject of online patches, I'd like to give another example concerning cartridge-based games. Sure, there were rereleases of patched games, like Final Fantasy 6 fixing a bug with a character's common skill or one of the Shinobi games taking out the blatant copyright infringing characters or Ocarina of Time's removal of bugs and Islamic references.<mask> most times, the game you bought was the game you got. Now, we have to add storage devices to hold all the revisions of Skyrim, Borderlands, and Halo; a similar issue PC users have to deal with. [NEWLINE] [NEWLINE] [NEWLINE] Another small point is that of maintenance. Ignoring that older consoles last longer due to simpler internals and arguably more durable designs, 360's are imfamous for their Red Ring of Death errors. And<mask> it wasn't covered under warranty, and you don't have the specific parts to fix it,
Label encoding: <s>I will try to explain myself without sounding like I'm just enticing the "PC Master Race" circlejerk. [NEWLINE] [NEWLINE] In the past, there were significant difference in hardware and software between consoles and home computers. Granted, older consoles used chips often found in PC's (the NES's CPU was based off the design of the one used in the Apple II and C64, Genesis used the Zilog Z80 And Motorola 68K, ect.) but they produced results that greatly differed from one another in terms of processing, sound and graphics. But with this current gen, both the Xbox One and PS4 use the same micro-architecture (x86) in their APU's that are manufactured by the same company, AMD, that is used in current PC's today. Many of both consoles' non-exclusives have also been shown to underperform graphically when pitted against desktops made for the same price as them. [NEWLINE] [NEWLINE] [NEWLINE] The other point that I feel no longer separates consoles and PC's is their firmware and the need to constantly update it as well as patch games via internet. Again, with older consoles, if I put a functioning copy of Super Metroid in any working SNES, it should play immediately, and that's it. I can understand with the 6th Generation's addition with a UI to adjust the clock, manage saves and other basic functions, but when there were additions like internet browsers, multimedia players, Streaming Video, and other services that added nothing to one's gameplay experiences, it seemed like a waste of resources, and redundant when my computer can do the same things. [NEWLINE] [NEWLINE] And on the subject of online patches, I'd like to give another example concerning cartridge-based games. Sure, there were rereleases of patched games, like Final Fantasy 6 fixing a bug with a character's common skill or one of the Shinobi games taking out the blatant copyright infringing characters or Ocarina of Time's removal of bugs and Islamic references. But most times, the game you bought was the game you got. Now, we have to add storage devices to hold all the revisions of Skyrim, Borderlands, and Halo; a similar issue PC users have to deal with. [NEWLINE] [NEWLINE] [NEWLINE] Another small point is that of maintenance. Ignoring that older consoles last longer due to simpler internals and arguably more durable designs, 360's are imfamous for their Red Ring of Death errors. And if it wasn't covered under warranty, and you don't have the specific parts to fix it,
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you have to vigorously research the MSRP<mask><mask> to get<mask> close to the salesmens floor price<mask> possible. [ENDQ] [NEWLINE] See, this is the problem. You are under the impression that your custom is in some way desirable for the business to have, and<mask> you appear to have a desire to pay<mask> little<mask> humanly possible for<mask> much<mask> you can possibly get. [NEWLINE] [NEWLINE] From the point of view of a business owner, this makes you<mask> is called a "Class C" customer. I don't employ salespeople to waste their time with you. Your "research" will consume inordinate amounts of staff time, you will tyre-kick all day long<mask> you nitpickingly analyze all possible factors, and you will try to play my business off against our competition. In general you will demand resources be expended on you that are in no way justified by the margin that can be earned from you, and you will resent every second you spend interacting with us. [NEWLINE] [NEWLINE] One of my salespersons' core tasks is to identify you early, and redirect you off to the competition. You're just not worth dealing with. They would far rather deal with a normal person who just wants to buy a car suit, is prepared to pay a fair price for it, has some reasonable questions to ask, will accept reasonable answers, and then will *actually buy*. [NEWLINE] [NEWLINE] <mask> you're a nice person, we'll go to some lengths to show you<mask> to inform yourself, maybe spend some time doing some comparison,<mask> from your questions and answers in this thread, it's pretty clear that you're out to take us for whatever you can. [NEWLINE] [NEWLINE] And<mask> there's some problem with<mask> you bought,<mask> you get that suit home and it doesn't quite fit your car, then oh boy, will you ever be down here the next day at opening time jumping up and down demanding we deal with it straight away. This is<mask> you are a class "C" customer - the most price-conscious people are, almost without exception, the most troublesome. Everything is someone else's fault. [NEWLINE] [NEWLINE] My salespeople are employed to earn me money. Each salesperson has a turnover target of two to four times his or her salary depending on experience, and gets a bonus based on turnover above this and the gross margin at which they are able to sell car suits.<mask> they can sell at a higher price, they will get a share of that.<mask> they have 100% conversion rate, they need to sell at
Label encoding: <s> [STARTQ] but you have to vigorously research the MSRP so as to get as close to the salesmens floor price as possible. [ENDQ] [NEWLINE] See, this is the problem. You are under the impression that your custom is in some way desirable for the business to have, and yet you appear to have a desire to pay as little as humanly possible for as much as you can possibly get. [NEWLINE] [NEWLINE] From the point of view of a business owner, this makes you what is called a "Class C" customer. I don't employ salespeople to waste their time with you. Your "research" will consume inordinate amounts of staff time, you will tyre-kick all day long as you nitpickingly analyze all possible factors, and you will try to play my business off against our competition. In general you will demand resources be expended on you that are in no way justified by the margin that can be earned from you, and you will resent every second you spend interacting with us. [NEWLINE] [NEWLINE] One of my salespersons' core tasks is to identify you early, and redirect you off to the competition. You're just not worth dealing with. They would far rather deal with a normal person who just wants to buy a car suit, is prepared to pay a fair price for it, has some reasonable questions to ask, will accept reasonable answers, and then will *actually buy*. [NEWLINE] [NEWLINE] If you're a nice person, we'll go to some lengths to show you how to inform yourself, maybe spend some time doing some comparison, but from your questions and answers in this thread, it's pretty clear that you're out to take us for whatever you can. [NEWLINE] [NEWLINE] And if there's some problem with what you bought, if you get that suit home and it doesn't quite fit your car, then oh boy, will you ever be down here the next day at opening time jumping up and down demanding we deal with it straight away. This is why you are a class "C" customer - the most price-conscious people are, almost without exception, the most troublesome. Everything is someone else's fault. [NEWLINE] [NEWLINE] My salespeople are employed to earn me money. Each salesperson has a turnover target of two to four times his or her salary depending on experience, and gets a bonus based on turnover above this and the gross margin at which they are able to sell car suits. If they can sell at a higher price, they will get a share of that. If they have 100% conversion rate, they need to sell at
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yay!:) [NEWLINE] To continue the discussion: [NEWLINE] [NEWLINE] **Re: First Quote** [NEWLINE] [NEWLINE] *<mask> * political systems are corrupted, that's all the more reason<mask> the social sciences and humanities are needed. You fix politics by studying politics, not engineering. Lessons from science, etc. may be useful in informing the design of political systems,<mask> are never its primary source.<mask> we're trying to increase fair representation in politics - give me John Locke over Isaac Newton, John Rawls over Einstein. [NEWLINE] [NEWLINE] The second paragraph of my original response is important in explaining<mask> a fully technocratic solution to government is impossible. Uncertainty is a fact of life in governance and society<mask> the subjects are ever-changing humans in complex environments, not simple micro-organisms in the safety of a lab. Fun aside - the yearning for technocracy in politics is resonant with philosophical positivism, a view that no one really holds anymore,<mask> you'd need a good humanities education to soundly refute it (Which is<mask> I mean by all science/tech is informed by politics and philosophy, and vice versa) [NEWLINE] [NEWLINE] Finally, in<mask> far<mask> technocratic elements should be partially incorporated into politics, that's<mask> another reason<mask> the social sciences are necessary. Using data to study the effects of policy is the domain of economics, sociology, political science, etc. [NEWLINE] [NEWLINE] **Re:Second Quote** [NEWLINE] [NEWLINE] You've heard the parable about judging rabbits by their ability to swim and elephants by their ability to climb trees, yes? You can't separate the quality of the student from the subject they are studying. Say student X has +99 math points<mask> only +20 writing points.<mask> X becomes an engineering major they'd probably be considered an awesome student. Take that same awesome student and make them an English major and you'd think they're terrible. [NEWLINE] [NEWLINE] And I guess I have no idea<mask> this is such a paradigmatic case for generalization. Even<mask> it were, you've only asserted some vague personal experience (a very small *n*), which is not a substantive basis for a generalization. I kind of think you're just projecting your (percieved) failure and disappointment onto these fields writ large. [NEWLINE] [NEWLINE] And I'd hope your generalization would give way in the face of larger n data that complicates your simple picture. The following studies aren't perfect,<mask> they're a better basis for knowledge than your intuition,<mask><mask>. [This]( [URL] ) ranks majors by their average verbal
Label encoding: <s>Yay!:) [NEWLINE] To continue the discussion: [NEWLINE] [NEWLINE] **Re: First Quote** [NEWLINE] [NEWLINE] * If * political systems are corrupted, that's all the more reason why the social sciences and humanities are needed. You fix politics by studying politics, not engineering. Lessons from science, etc. may be useful in informing the design of political systems, but are never its primary source. If we're trying to increase fair representation in politics - give me John Locke over Isaac Newton, John Rawls over Einstein. [NEWLINE] [NEWLINE] The second paragraph of my original response is important in explaining why a fully technocratic solution to government is impossible. Uncertainty is a fact of life in governance and society because the subjects are ever-changing humans in complex environments, not simple micro-organisms in the safety of a lab. Fun aside - the yearning for technocracy in politics is resonant with philosophical positivism, a view that no one really holds anymore, though you'd need a good humanities education to soundly refute it (Which is what I mean by all science/tech is informed by politics and philosophy, and vice versa) [NEWLINE] [NEWLINE] Finally, in so far as technocratic elements should be partially incorporated into politics, that's yet another reason why the social sciences are necessary. Using data to study the effects of policy is the domain of economics, sociology, political science, etc. [NEWLINE] [NEWLINE] **Re:Second Quote** [NEWLINE] [NEWLINE] You've heard the parable about judging rabbits by their ability to swim and elephants by their ability to climb trees, yes? You can't separate the quality of the student from the subject they are studying. Say student X has +99 math points but only +20 writing points. If X becomes an engineering major they'd probably be considered an awesome student. Take that same awesome student and make them an English major and you'd think they're terrible. [NEWLINE] [NEWLINE] And I guess I have no idea why this is such a paradigmatic case for generalization. Even if it were, you've only asserted some vague personal experience (a very small *n*), which is not a substantive basis for a generalization. I kind of think you're just projecting your (percieved) failure and disappointment onto these fields writ large. [NEWLINE] [NEWLINE] And I'd hope your generalization would give way in the face of larger n data that complicates your simple picture. The following studies aren't perfect, but they're a better basis for knowledge than your intuition, I think. [This]( [URL] ) ranks majors by their average verbal
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it is the opinion of many people on the internet, and of reddit in particular that the companies trying to fight piracy should "Deal with it" -- that they need to realize that the internet is the future and trying to fight piracy is impossible (or prohibitively expensive) in a free web. [NEWLINE] [NEWLINE] A little bit about myself, and my habits -- I am 20 years old, a liberal raised on the internet and currently in school studying film. I have torrented rarely for probably 5 years,<mask> I would consider myself on the bottom end of piracy. I subscribe to Netflix, Amazon prime and Spotify, and usually buy my games on steam.<mask> of these services I never pirate music or games ever, and I usually do not pirate Television or Movies either. [NEWLINE] [NEWLINE] <mask> an aspiring filmmaker myself, it feels hypocritical to steal someone else's hard work, and<mask><mask> at all possible I do not do that. I am more than willing to pay 3-5 dollars to rent a movie I want to watch.<mask> this is not the internet we live in.<mask> some movies are available on youtube, itunes and amazon instant for rental, many are either only available to buy or not available at all. On top of that, these services tend to have a charge to rent the "HD" version of the film, a practice I believe to be archaic and stupid in an age<mask> everyone has HD monitors and/or TVs. [NEWLINE] [NEWLINE] In my mind,<mask> there were a service that made a huge selection of films available for a one-time viewing for $5 easily and at full quality -- with options for downloading<mask> need be -- I personally would never pirate films or TV. [NEWLINE] [NEWLINE] <mask>, this may not represent the consumer base<mask> a whole and may not be enough to save the companies afraid of the internet. [NEWLINE] [NEWLINE] (WARNING: I AM GOING OFF<mask> I HAVE READ IN THIS SECTION -- I DO NOT HAVE SOURCES FOR MY CLAIMS<mask><mask> ) [NEWLINE] [NEWLINE] <mask> Spotify does seem mitigate piracy, it<mask> does not give much money to artists<mask> a rule.<mask>, at the same time, most artists make their money from touring and not from album sales,<mask> the record label would end up taking most of the money from itunes/CD sales anyways.<mask> part of me thinks that in this open age that perhaps it is the companies that should fail,<mask> they cannot accept the internet's existence, freeing the artists to receive more money for sales directly from services such<mask> Spotify and
Label encoding: <s>I think it is the opinion of many people on the internet, and of reddit in particular that the companies trying to fight piracy should "Deal with it" -- that they need to realize that the internet is the future and trying to fight piracy is impossible (or prohibitively expensive) in a free web. [NEWLINE] [NEWLINE] A little bit about myself, and my habits -- I am 20 years old, a liberal raised on the internet and currently in school studying film. I have torrented rarely for probably 5 years, but I would consider myself on the bottom end of piracy. I subscribe to Netflix, Amazon prime and Spotify, and usually buy my games on steam. Because of these services I never pirate music or games ever, and I usually do not pirate Television or Movies either. [NEWLINE] [NEWLINE] As an aspiring filmmaker myself, it feels hypocritical to steal someone else's hard work, and so if at all possible I do not do that. I am more than willing to pay 3-5 dollars to rent a movie I want to watch. But this is not the internet we live in. While some movies are available on youtube, itunes and amazon instant for rental, many are either only available to buy or not available at all. On top of that, these services tend to have a charge to rent the "HD" version of the film, a practice I believe to be archaic and stupid in an age where everyone has HD monitors and/or TVs. [NEWLINE] [NEWLINE] In my mind, if there were a service that made a huge selection of films available for a one-time viewing for $5 easily and at full quality -- with options for downloading if need be -- I personally would never pirate films or TV. [NEWLINE] [NEWLINE] However, this may not represent the consumer base as a whole and may not be enough to save the companies afraid of the internet. [NEWLINE] [NEWLINE] (WARNING: I AM GOING OFF WHAT I HAVE READ IN THIS SECTION -- I DO NOT HAVE SOURCES FOR MY CLAIMS YET ) [NEWLINE] [NEWLINE] While Spotify does seem mitigate piracy, it also does not give much money to artists as a rule. But, at the same time, most artists make their money from touring and not from album sales, as the record label would end up taking most of the money from itunes/CD sales anyways. So part of me thinks that in this open age that perhaps it is the companies that should fail, if they cannot accept the internet's existence, freeing the artists to receive more money for sales directly from services such as Spotify and
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to assume we're talking about the US here,<mask> all countries have different cultures regarding things like race and gender. [NEWLINE] [NEWLINE] 1) I will readily agree that<mask> a white person, I have it way easier than a person of a minority race (especially black or latino) living in Western society.<mask><mask> it's awful that I have this privilege<mask> I know I do. [NEWLINE] [NEWLINE] 2) Ok, the gender thing is a sticky issue.<mask> much<mask> gender is compared to race, the two are not at all alike. Sex is a binary (gender is not,<mask> that's a separate issue) - you're either a man or a woman,<mask> once society decides that there are roles for men and roles for women, there are going to be pros and cons of living on one side of that binary, no matter which side you're on.<mask><mask> it's irrational to<mask><mask> one gender has it better than another<mask> there are areas<mask> women will benefit and areas<mask> men will due to that binary. [NEWLINE] Now that men and women have essentially equal political rights, they are closer to being equal on a social level,<mask> there's no denying that society still sees men<mask> dominating and women<mask> submissive.<mask> a man is made to be submissive (<mask> he gets raped,<mask> he chooses to dress like a woman), people will not be<mask> receptive to him<mask> they would be a to a woman in the same situation;<mask> a woman is dominant (gets a job<mask> a politician or CEO, chooses job over family), she will likely be more belittled and not taken<mask> seriously<mask> a man in that position. To say one gender "has it worse" is oversimplifying a multifaceted issue. [NEWLINE] [NEWLINE] 3) Attractiveness. The problem is, you can't just say "This gender has it easier, this race has it easier, this level of beauty has it easier,<mask><mask> I combine them all, that type of person will necessarily have it easier". Hamburgers are good, ice cream is good,<mask> hamburger ice cream would taste like shit,<mask> you get<mask> I mean. [NEWLINE] Being an attractive woman is different from being an attractive man. A big issue for women is sexual assault and harassment, which ties into the issue of objectification both by the media and in real life. I'm not going to<mask><mask> ugly women have it easy - they don't,<mask> that doesn't mean being beautiful is a cakewalk. [NEWLINE] You mention getting
Label encoding: <s>I'm going to assume we're talking about the US here, since all countries have different cultures regarding things like race and gender. [NEWLINE] [NEWLINE] 1) I will readily agree that as a white person, I have it way easier than a person of a minority race (especially black or latino) living in Western society. I think it's awful that I have this privilege but I know I do. [NEWLINE] [NEWLINE] 2) Ok, the gender thing is a sticky issue. As much as gender is compared to race, the two are not at all alike. Sex is a binary (gender is not, but that's a separate issue) - you're either a man or a woman, so once society decides that there are roles for men and roles for women, there are going to be pros and cons of living on one side of that binary, no matter which side you're on. I think it's irrational to argue that one gender has it better than another since there are areas where women will benefit and areas where men will due to that binary. [NEWLINE] Now that men and women have essentially equal political rights, they are closer to being equal on a social level, but there's no denying that society still sees men as dominating and women as submissive. When a man is made to be submissive ( if he gets raped, if he chooses to dress like a woman), people will not be as receptive to him as they would be a to a woman in the same situation; when a woman is dominant (gets a job as a politician or CEO, chooses job over family), she will likely be more belittled and not taken as seriously as a man in that position. To say one gender "has it worse" is oversimplifying a multifaceted issue. [NEWLINE] [NEWLINE] 3) Attractiveness. The problem is, you can't just say "This gender has it easier, this race has it easier, this level of beauty has it easier, so if I combine them all, that type of person will necessarily have it easier". Hamburgers are good, ice cream is good, but hamburger ice cream would taste like shit, if you get what I mean. [NEWLINE] Being an attractive woman is different from being an attractive man. A big issue for women is sexual assault and harassment, which ties into the issue of objectification both by the media and in real life. I'm not going to argue that ugly women have it easy - they don't, but that doesn't mean being beautiful is a cakewalk. [NEWLINE] You mention getting
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd have to disagree that the world was better off<mask> the Soviet Union existed and acted<mask> a counterweight to the US in geopolitics. [NEWLINE] [NEWLINE] <mask> the Iraq War (est. 150,000 to 600,000 dead) and Afghanistan/War on Terror (est. 50,000 in Afghanistan + 5-20,000 worldwide) are the major controversial subjects of the post-Cold War and post-9/11 world... [NEWLINE] [NEWLINE] They all simply pale in comparison to the battle over ideology between the West and East that manifested itself in the form of proxy wars, regime changes, and other controversies from the period of 1945 to 1991. [NEWLINE] [NEWLINE] Here's a sampling: [NEWLINE] **Proxy Wars** [NEWLINE] [NEWLINE] * Greek Civil War (1946 - 1949) - Greece (with US + UK support) vs. Communist insurgents - 150,000 killed [NEWLINE] * Arab Israeli Conflict (main phase 1948 - 1973) - Israel (with US support) vs. Arab nations (with USSR support) - around 110,000 killed [NEWLINE] * Korean War (1950 - 1953) - North Korea (with USSR + China support) vs. South Korea + United States + UN support - 3 million+ killed [NEWLINE] * Cuban Revolution (1953 - 1959) - Communist insurgents vs. Batista government (with US support) - 5,000 killed [NEWLINE] * Vietnam War (1955 - 1975) - North Vietnam (supported primarily by USSR and China) vs. South Vietnam (supported primarily by United States) - 3-4 million killed [NEWLINE] * Nicaraguan Revolution and Contra War (1960s to 1990) - Somonza government (with US aid) vs. FSLN (supported by Soviet Union and Cuba) - 40,000 killed [NEWLINE] * Angolan Civil War (1975 - 2002) - MPLA aided by the USSR, Vietnam, and Cuba vs. UNITA and FNLA (aided by the US, China, and South Africa) - 500,000 killed [NEWLINE] * Soviet War in Afghanistan (1979 - 1989) - Soviet Union vs. Afghan rebels (with Pakistan + United States + China support) - 1 million killed [NEWLINE] [NEWLINE] Not wars,<mask> still prominent changes aided either covertly or overtly: [NEWLINE] [NEWLINE] **Regime Changes / Interventions** [NEWLINE] [NEWLINE] * 1948 - Czechoslovakia coup - USSR installs communist government [NEWLINE] * 1953 - Iran's Mossadegh, Prime Minister appointed by the Shah, overthrown after reducing Shah's power, nationalizing British oil, and turning towards USSR - Shah's power restored [NEWLINE] * 1954
Label encoding: <s>I'd have to disagree that the world was better off when the Soviet Union existed and acted as a counterweight to the US in geopolitics. [NEWLINE] [NEWLINE] While the Iraq War (est. 150,000 to 600,000 dead) and Afghanistan/War on Terror (est. 50,000 in Afghanistan + 5-20,000 worldwide) are the major controversial subjects of the post-Cold War and post-9/11 world... [NEWLINE] [NEWLINE] They all simply pale in comparison to the battle over ideology between the West and East that manifested itself in the form of proxy wars, regime changes, and other controversies from the period of 1945 to 1991. [NEWLINE] [NEWLINE] Here's a sampling: [NEWLINE] **Proxy Wars** [NEWLINE] [NEWLINE] * Greek Civil War (1946 - 1949) - Greece (with US + UK support) vs. Communist insurgents - 150,000 killed [NEWLINE] * Arab Israeli Conflict (main phase 1948 - 1973) - Israel (with US support) vs. Arab nations (with USSR support) - around 110,000 killed [NEWLINE] * Korean War (1950 - 1953) - North Korea (with USSR + China support) vs. South Korea + United States + UN support - 3 million+ killed [NEWLINE] * Cuban Revolution (1953 - 1959) - Communist insurgents vs. Batista government (with US support) - 5,000 killed [NEWLINE] * Vietnam War (1955 - 1975) - North Vietnam (supported primarily by USSR and China) vs. South Vietnam (supported primarily by United States) - 3-4 million killed [NEWLINE] * Nicaraguan Revolution and Contra War (1960s to 1990) - Somonza government (with US aid) vs. FSLN (supported by Soviet Union and Cuba) - 40,000 killed [NEWLINE] * Angolan Civil War (1975 - 2002) - MPLA aided by the USSR, Vietnam, and Cuba vs. UNITA and FNLA (aided by the US, China, and South Africa) - 500,000 killed [NEWLINE] * Soviet War in Afghanistan (1979 - 1989) - Soviet Union vs. Afghan rebels (with Pakistan + United States + China support) - 1 million killed [NEWLINE] [NEWLINE] Not wars, but still prominent changes aided either covertly or overtly: [NEWLINE] [NEWLINE] **Regime Changes / Interventions** [NEWLINE] [NEWLINE] * 1948 - Czechoslovakia coup - USSR installs communist government [NEWLINE] * 1953 - Iran's Mossadegh, Prime Minister appointed by the Shah, overthrown after reducing Shah's power, nationalizing British oil, and turning towards USSR - Shah's power restored [NEWLINE] * 1954
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me start by saying I'm sorry this got<mask> long, I really didn't mean it, I just got carried away. [NEWLINE] [NEWLINE] [STARTQ] I suspect that the vast majority of people do not care at all about the quality of anything other than the game itself [ENDQ] [NEWLINE] I must disagree with this. Granted, this is only anecdotal evidence and<mask> not a valid source,<mask> I traded uses games through a website called Goozex for a couple of years. Goozex had multiple options for "package quality" on their trades. There was full package, which was the disc, the case, and any manuals that came with it. There was disc and manual, which is self explanatory, and there was disc only. Somewhere akin to 90% of the trades that occurred on Goozex were for full package only. I realize this is not conclusive in and of itself, I am just offering you a differing viewpoint. [NEWLINE] [NEWLINE] [STARTQ] a company like GAME or GameStop being able to sell you a game of equivalent quality for less money and having none of that profit going to the publisher (I count GAME purchasing a game from the publisher in order sell<mask> profit for the publisher) still seems wrong. [ENDQ] [NEWLINE] I can understand this point, sort of. I suppose it depends on whether you only count the game itself<mask> the "quality" part. A lot of used games I've seen from gamestop do not actually have the original case, and even more don't have the manuals. [NEWLINE] [NEWLINE] <mask> you are considering the game itself to be the valuable part, I would point you towards the housing market.<mask> I purchase a house for $200k from the company that built the house, and two years later it turns out that my house is $300k,<mask> of location or better economy, whatever, do I owe the building company that $100k to continue to live there?<mask><mask> I sold it? Should the company be entitled to a portion of the profits I made from selling that house? [NEWLINE] [NEWLINE] The first sale doctrine protects consumers by limiting the scope of control that a company has over it's products *after* they have been sold.<mask> I buy a baseball for $5, is it fair to me for the company to say I can only sell that baseball to other people for $10? Is it fair for the company to say that I can only use that baseball on company-approved baseball fields? Is it fair for the company to say I can only use the baseball between the hours of 6 am and 10
Label encoding: <s>Let me start by saying I'm sorry this got so long, I really didn't mean it, I just got carried away. [NEWLINE] [NEWLINE] [STARTQ] I suspect that the vast majority of people do not care at all about the quality of anything other than the game itself [ENDQ] [NEWLINE] I must disagree with this. Granted, this is only anecdotal evidence and so not a valid source, but I traded uses games through a website called Goozex for a couple of years. Goozex had multiple options for "package quality" on their trades. There was full package, which was the disc, the case, and any manuals that came with it. There was disc and manual, which is self explanatory, and there was disc only. Somewhere akin to 90% of the trades that occurred on Goozex were for full package only. I realize this is not conclusive in and of itself, I am just offering you a differing viewpoint. [NEWLINE] [NEWLINE] [STARTQ] a company like GAME or GameStop being able to sell you a game of equivalent quality for less money and having none of that profit going to the publisher (I count GAME purchasing a game from the publisher in order sell as profit for the publisher) still seems wrong. [ENDQ] [NEWLINE] I can understand this point, sort of. I suppose it depends on whether you only count the game itself as the "quality" part. A lot of used games I've seen from gamestop do not actually have the original case, and even more don't have the manuals. [NEWLINE] [NEWLINE] If you are considering the game itself to be the valuable part, I would point you towards the housing market. If I purchase a house for $200k from the company that built the house, and two years later it turns out that my house is $300k, because of location or better economy, whatever, do I owe the building company that $100k to continue to live there? What if I sold it? Should the company be entitled to a portion of the profits I made from selling that house? [NEWLINE] [NEWLINE] The first sale doctrine protects consumers by limiting the scope of control that a company has over it's products *after* they have been sold. If I buy a baseball for $5, is it fair to me for the company to say I can only sell that baseball to other people for $10? Is it fair for the company to say that I can only use that baseball on company-approved baseball fields? Is it fair for the company to say I can only use the baseball between the hours of 6 am and 10
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just found this subreddit, it's interesting. I'll take a stab sure. [NEWLINE] [NEWLINE] Ok,<mask><mask> you're saying that STEM majors are more difficult than non-STEM majors. You've seemed to define "difficult" by saying the amount of hours one puts in to complete coursework - and<mask> the amount of effort involved to complete the coursework. Hours =/= Effort,<mask> we'll say effort is a better thing to talk about for this. Yes it is more subjective than hours,<mask> hours are just one quantifiable part of a much larger part, which is the effort itself. [NEWLINE] [NEWLINE] The adage, "<mask> you want to get something done quick, find a lazy person to do it,<mask> they'll do it the fastest" rings true to me here. Effort is not just the hours involved<mask> it is the critical application of thought and logic (and many, many other factors) to overcome the challenge. For example: I could spend 3 hours trying to solve a maths problem by myself, unaware of the correct method and learning<mask> I go - or I could could simply google the question - understand the method to solve it and use that knowledge to complete the question in 1 hour. In both cases, I've expended effort - just applied in different ways (brute force vs. more creative thinking) My point here is that not all effort - gives the same result. [NEWLINE] [NEWLINE] <mask> we've talked about<mask> effort is a strange and hard to define thing in relation to the ways people expend it towards solving problems.<mask> lets talk about the way people solve problems. People have an individual natural inclination to solve a problem in a certain way. Some people are better than others at logical reasoning - and others are better at spacial reasoning. These are differences in information metabolism. i.e. -<mask> you take in, process, and output information. Different majors are suited to different ways of inputing, processing and outputting information -<mask> the people in these courses are obviously not on a level playing field from the get go. People will have advantages and disadvantages - and thats<mask><mask><mask> you have - a natural disadvantage to processing information in a way that suits a STEM major. Perhaps<mask> you were in an interior design course - you might flourish. Or<mask> you did nursing - you might be great at it naturally -<mask> you've chosen a STEM major and are now finding it difficult. You haven't spoken about other classmates -<mask> I have no idea of their opinion of the perceived difficulty of the
Label encoding: <s>Just found this subreddit, it's interesting. I'll take a stab sure. [NEWLINE] [NEWLINE] Ok, firstly you're saying that STEM majors are more difficult than non-STEM majors. You've seemed to define "difficult" by saying the amount of hours one puts in to complete coursework - and secondly the amount of effort involved to complete the coursework. Hours =/= Effort, so we'll say effort is a better thing to talk about for this. Yes it is more subjective than hours, but hours are just one quantifiable part of a much larger part, which is the effort itself. [NEWLINE] [NEWLINE] The adage, " if you want to get something done quick, find a lazy person to do it, as they'll do it the fastest" rings true to me here. Effort is not just the hours involved but it is the critical application of thought and logic (and many, many other factors) to overcome the challenge. For example: I could spend 3 hours trying to solve a maths problem by myself, unaware of the correct method and learning as I go - or I could could simply google the question - understand the method to solve it and use that knowledge to complete the question in 1 hour. In both cases, I've expended effort - just applied in different ways (brute force vs. more creative thinking) My point here is that not all effort - gives the same result. [NEWLINE] [NEWLINE] So we've talked about how effort is a strange and hard to define thing in relation to the ways people expend it towards solving problems. So lets talk about the way people solve problems. People have an individual natural inclination to solve a problem in a certain way. Some people are better than others at logical reasoning - and others are better at spacial reasoning. These are differences in information metabolism. i.e. - how you take in, process, and output information. Different majors are suited to different ways of inputing, processing and outputting information - so the people in these courses are obviously not on a level playing field from the get go. People will have advantages and disadvantages - and thats what I think you have - a natural disadvantage to processing information in a way that suits a STEM major. Perhaps if you were in an interior design course - you might flourish. Or if you did nursing - you might be great at it naturally - but you've chosen a STEM major and are now finding it difficult. You haven't spoken about other classmates - so I have no idea of their opinion of the perceived difficulty of the
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, I did a search, and I didn't find anything<mask> please advise me<mask> that has been covered.  Seems like a common thread,<mask> maybe I'm bad at searching. [NEWLINE] [NEWLINE] Title basically says it all.  Whether it is religious head gear, beards, special diets, prayer time, or anything else... I don't think your prison should have to accommodate it. <mask> you wanted to keep observing that, you should have tried harder to stay out of jail.  I don't have any figures on<mask> much is spent on this,<mask> I suspect it is not cheap.  Even<mask> it was only slightly more expensive or even the same cost<mask> an inmate with no religious obligations, it is an inconvenience for the prison staff,<mask> that is only a minor issue. [NEWLINE] [NEWLINE] My biggest issue is that people in prison are there<mask> they did something illegal and they have a debt to pay to society.  We should keep them alive and in no worse health than<mask> they went in,<mask> we shouldn't be concerned about keeping them overly comfortable.  It is supposed to make you uncomfortable<mask> a deterrent to doing another crime. [NEWLINE] [NEWLINE] I won't be convinced by any argument that ignoring their religious obligations amounts to cruel and unusual punishment. <mask> lighting came from the sky and struck them for eating bacon or taking off their headgear then I might agree<mask> the fact is that no harm has ever come to a person from ignoring their religious obligations. [NEWLINE] [NEWLINE] [NEWLINE] Furthermore, ignoring all religious obligations put all prisoners on equal grounds, which a commendable goal.  No special treatment for anyone. [NEWLINE] [NEWLINE] I'm open to having my view changed, and I'll be here for a<mask>.  Thanks. [NEWLINE] [NEWLINE] EDIT: not sure about accuracy,<mask> this one article claims that Kosher food for prisoners costs more than 2X the cost of standard food.  It<mask> creates a black market for this "special food". [NEWLINE] [URL] /?p=all [NEWLINE] [NEWLINE] Final edit: my view has been partially changed and two deltas have been awarded.  I now believe that religious obligations may be met<mask> the inmate pays the cost difference between his needs and the needs of an inmate with no special obligations and<mask> he/ she maintains good behavior.  The inmate may pay the cost difference from his personal fortune, his family may pay it, or he can work a job within the prison to pay the difference.  I believe that taxpayers paying for your religious obligations amounts to an unconstitutional establishment of religion,<mask> actively
Label encoding: <s>Yes, I did a search, and I didn't find anything but please advise me if that has been covered.  Seems like a common thread, but maybe I'm bad at searching. [NEWLINE] [NEWLINE] Title basically says it all.  Whether it is religious head gear, beards, special diets, prayer time, or anything else... I don't think your prison should have to accommodate it.  If you wanted to keep observing that, you should have tried harder to stay out of jail.  I don't have any figures on how much is spent on this, but I suspect it is not cheap.  Even if it was only slightly more expensive or even the same cost as an inmate with no religious obligations, it is an inconvenience for the prison staff, but that is only a minor issue. [NEWLINE] [NEWLINE] My biggest issue is that people in prison are there because they did something illegal and they have a debt to pay to society.  We should keep them alive and in no worse health than when they went in, but we shouldn't be concerned about keeping them overly comfortable.  It is supposed to make you uncomfortable as a deterrent to doing another crime. [NEWLINE] [NEWLINE] I won't be convinced by any argument that ignoring their religious obligations amounts to cruel and unusual punishment.  If lighting came from the sky and struck them for eating bacon or taking off their headgear then I might agree but the fact is that no harm has ever come to a person from ignoring their religious obligations. [NEWLINE] [NEWLINE] [NEWLINE] Furthermore, ignoring all religious obligations put all prisoners on equal grounds, which a commendable goal.  No special treatment for anyone. [NEWLINE] [NEWLINE] I'm open to having my view changed, and I'll be here for a while.  Thanks. [NEWLINE] [NEWLINE] EDIT: not sure about accuracy, but this one article claims that Kosher food for prisoners costs more than 2X the cost of standard food.  It also creates a black market for this "special food". [NEWLINE] [URL] /?p=all [NEWLINE] [NEWLINE] Final edit: my view has been partially changed and two deltas have been awarded.  I now believe that religious obligations may be met if the inmate pays the cost difference between his needs and the needs of an inmate with no special obligations and if he/ she maintains good behavior.  The inmate may pay the cost difference from his personal fortune, his family may pay it, or he can work a job within the prison to pay the difference.  I believe that taxpayers paying for your religious obligations amounts to an unconstitutional establishment of religion, but actively
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I want to say two things: [NEWLINE] [NEWLINE] First, let me tell you a story about the first time I ever did shrooms. [NEWLINE] [NEWLINE] I was raised in a really, really abusive household. My father raped me and beat me and locked me in closets and mentally battered me. My mom was an alcoholic. My sister hated me. I hated myself.<mask> I did shrooms, all of that was different. It was the first time in my life I actually felt<mask> 'being happy' was. It was completely new. Things were brighter, I could laugh, I *felt* things. It was<mask> the first time that I thought I could actually have sex with someone and not feel ashamed. [NEWLINE] [NEWLINE] I don't advocate for drug use,<mask> I don't think that temporary happiness is bad. I couldn't just 'fix it',<mask> it was my life. I couldn't 'fix' my mother or 'fix' my father. I was too afraid to call the police, and my dad manipulated me into thinking that it was *normal* anyway. I thought everyone's life was like mine, and everyone wanted to kill themselves.<mask> this gave me an experience to hold on to. Did I abuse the drugs in the future looking for this feeling? Yeah. Did it keep me alive? I like to think<mask>. I honestly think that<mask> I had never felt this ephemeral happiness,<mask> I had honestly thought that life was just the misery I lived in, I would have killed myself. [NEWLINE] [NEWLINE] <mask>, I'd like to talk about your view on unhappiness. Depression and anxiety are actual disorders. They are the result of a lot of things,<mask> mainly there's a lack of certain chemicals in your brain. This is<mask> SSRIs and other antidepressants and anti anxiety medication exists. With therapy, drugs, and other things, it's totally possible to recover from depression (just<mask> you know: they say you're in remission from depression. Not cured). [NEWLINE] [NEWLINE] <mask>, you can't just tell someone to 'fix' their lives. People are complicated, and their reasons for doing things are complicated. A person in an abusive relationship (usually) can't just get up and leave, and often don't realize they're in it to begin with. A college student who's depressed<mask> they don't have enough serotonin in their brain can't just 'fix it' and make it better. Sure, yes, you can exercise and you can eat certain things and you can try real hard and
Label encoding: <s>I want to say two things: [NEWLINE] [NEWLINE] First, let me tell you a story about the first time I ever did shrooms. [NEWLINE] [NEWLINE] I was raised in a really, really abusive household. My father raped me and beat me and locked me in closets and mentally battered me. My mom was an alcoholic. My sister hated me. I hated myself. When I did shrooms, all of that was different. It was the first time in my life I actually felt what 'being happy' was. It was completely new. Things were brighter, I could laugh, I *felt* things. It was also the first time that I thought I could actually have sex with someone and not feel ashamed. [NEWLINE] [NEWLINE] I don't advocate for drug use, but I don't think that temporary happiness is bad. I couldn't just 'fix it', because it was my life. I couldn't 'fix' my mother or 'fix' my father. I was too afraid to call the police, and my dad manipulated me into thinking that it was *normal* anyway. I thought everyone's life was like mine, and everyone wanted to kill themselves. But this gave me an experience to hold on to. Did I abuse the drugs in the future looking for this feeling? Yeah. Did it keep me alive? I like to think so. I honestly think that if I had never felt this ephemeral happiness, if I had honestly thought that life was just the misery I lived in, I would have killed myself. [NEWLINE] [NEWLINE] Secondly, I'd like to talk about your view on unhappiness. Depression and anxiety are actual disorders. They are the result of a lot of things, but mainly there's a lack of certain chemicals in your brain. This is why SSRIs and other antidepressants and anti anxiety medication exists. With therapy, drugs, and other things, it's totally possible to recover from depression (just so you know: they say you're in remission from depression. Not cured). [NEWLINE] [NEWLINE] Therefore, you can't just tell someone to 'fix' their lives. People are complicated, and their reasons for doing things are complicated. A person in an abusive relationship (usually) can't just get up and leave, and often don't realize they're in it to begin with. A college student who's depressed because they don't have enough serotonin in their brain can't just 'fix it' and make it better. Sure, yes, you can exercise and you can eat certain things and you can try real hard and
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know I am and that dissonance is<mask>'s driving me nuts, I must come to a firm conclusion, that's<mask> I came here. [NEWLINE] [NEWLINE] <mask> I said that one cannot safely claim<mask>'s right or wrong in any culture, I forgot to include that in my view it only applies to cultures not of their own. That depends on the moral standards of which culture you're inserted in, which can vary slightly or greatly from others,<mask> it was my fault. [NEWLINE] [NEWLINE] The second point is<mask> really good, I said that immersion in a culture can change someone's personality,<mask> said culture can cleanse itself from unwanted aspects,<mask> changing their own perception of culture perpetually,<mask> is a really interesting insight on<mask> cultures evolve. This is really an eye-opened, cultures can reinvent themselves<mask> excluding and including behaviors based on their perceptions of<mask>'s wrong and right. This one hurt my head<mask> I liked it, it helped clarify things a lot! [NEWLINE] [NEWLINE] About the annoying behaviors, it goes from the origin of such behaviors to more specific things,<mask> anything I say can be easily simplified by 'it's your opinion' (which I am willing to change). I could include excessive claiming of one's sexuality, self-harming, extreme displays of individuality, unnecessary acceptance and assimilation of (in the case of males) stereotypical bad habits from women and things that could configure practical abuse of freedom, things that I know aren't decisive characteristics of gay people specially<mask> they are all<mask> done by heterosexual people<mask><mask> it is dumb to assume that there isn't difference in hetero and homosexual people. Thing is, those differences should be limited to natural traits<mask> it all falls down to the fact that people have different personalities, which is hard for me to use<mask> an argument to myself to justify<mask> annoys me. I know the whole 'acceptance' thing,<mask> there's limits for things, I mean, it would be morally correct of me to accept two men kissing? Yes it is,<mask> it<mask> means I have to accept, for example, someone who earns much more attention by having negative traits<mask> automatically proclaiming himself<mask> an example of homosexual? [NEWLINE] Maybe in his morality it is okay to call himself a fag and have sex with everyone he sees,<mask> that could bring shame to other homosexuals by a plethora of reasons and do I have to accept it? I don't know<mask> his morality says it's ok, normal gay people (normal<mask> in... different from the flamboy
Label encoding: <s>I know I am and that dissonance is what's driving me nuts, I must come to a firm conclusion, that's why I came here. [NEWLINE] [NEWLINE] When I said that one cannot safely claim what's right or wrong in any culture, I forgot to include that in my view it only applies to cultures not of their own. That depends on the moral standards of which culture you're inserted in, which can vary slightly or greatly from others, so it was my fault. [NEWLINE] [NEWLINE] The second point is also really good, I said that immersion in a culture can change someone's personality, but said culture can cleanse itself from unwanted aspects, thus changing their own perception of culture perpetually, what is a really interesting insight on how cultures evolve. This is really an eye-opened, cultures can reinvent themselves while excluding and including behaviors based on their perceptions of what's wrong and right. This one hurt my head but I liked it, it helped clarify things a lot! [NEWLINE] [NEWLINE] About the annoying behaviors, it goes from the origin of such behaviors to more specific things, so anything I say can be easily simplified by 'it's your opinion' (which I am willing to change). I could include excessive claiming of one's sexuality, self-harming, extreme displays of individuality, unnecessary acceptance and assimilation of (in the case of males) stereotypical bad habits from women and things that could configure practical abuse of freedom, things that I know aren't decisive characteristics of gay people specially because they are all also done by heterosexual people even though it is dumb to assume that there isn't difference in hetero and homosexual people. Thing is, those differences should be limited to natural traits but it all falls down to the fact that people have different personalities, which is hard for me to use as an argument to myself to justify what annoys me. I know the whole 'acceptance' thing, but there's limits for things, I mean, it would be morally correct of me to accept two men kissing? Yes it is, but it also means I have to accept, for example, someone who earns much more attention by having negative traits while automatically proclaiming himself as an example of homosexual? [NEWLINE] Maybe in his morality it is okay to call himself a fag and have sex with everyone he sees, but that could bring shame to other homosexuals by a plethora of reasons and do I have to accept it? I don't know because his morality says it's ok, normal gay people (normal as in... different from the flamboy
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] For a longer-term example, the auto industry in the USA was done in by increasing union-derived benefits which they had to pay for, draining more and more from the companies until they could no longer support themselves. [ENDQ] [NEWLINE] This may have contributed<mask> probably wasn't the main reason for the auto industry to collapse. One of the main reasons was<mask> of the [URL] which saw a rise in crude oil, which lead to a rise in gas prices.<mask> the price of gas increased the general buying trend moved away from the gas-guzzling SUV, that the "Big Three" American companies were focused on producing, to the lighter compact car which companies like Honda was producing. "The Big Three" had moved away from these due to breaking about even on the cars and decided to chase the bigger profits of the SUV's in the 90s. [NEWLINE] [NEWLINE] [STARTQ] <mask> the history of their marques, many long running cars have been discontinued or relegated to fleet sales,[75][76][77]<mask> GM, Ford and DaimlerChrysler shifted away resources from midsize and compact cars to lead the "SUV Craze".<mask> the late 1990s, over half of their profits have come from light trucks and SUVs,<mask> they often could not break even on compact cars unless the buyer chose options.[78] Ron Harbour, in releasing the Oliver Wyman’s 2008 Harbour Report, stated that many small “econoboxes” of the past acted<mask> loss leaders,<mask> were designed to bring customers to the brand in the hopes they would stay loyal and move up to more profitable models. The report estimated that an automaker needed to sell ten small cars to make the same profit<mask> one big vehicle, and that they had to produce small and mid-size cars profitably to succeed, something that the Detroit three have not<mask> done.[79] SUV sales peaked in 1999<mask> have not returned to that level ever<mask>, due to higher gas prices. [ENDQ] [NEWLINE] [URL] #United_States [NEWLINE] [NEWLINE] [STARTQ] UAW Leadership granted concessions to its unions in order to win labor peace, a benefit not calculated by the UAW's many critics.[21] The UAW has claimed that the primary cause of the automotive sector's weakness was substantially more expensive fuel costs[22] linked to the 2003-2008 oil crisis which caused customers to turn away from large sport utility vehicles (SUVs) and pickup trucks,[23] the main market of the American "Big Three" (General Motors, Ford,
Label encoding: <s> [STARTQ] For a longer-term example, the auto industry in the USA was done in by increasing union-derived benefits which they had to pay for, draining more and more from the companies until they could no longer support themselves. [ENDQ] [NEWLINE] This may have contributed but probably wasn't the main reason for the auto industry to collapse. One of the main reasons was because of the [URL] which saw a rise in crude oil, which lead to a rise in gas prices. As the price of gas increased the general buying trend moved away from the gas-guzzling SUV, that the "Big Three" American companies were focused on producing, to the lighter compact car which companies like Honda was producing. "The Big Three" had moved away from these due to breaking about even on the cars and decided to chase the bigger profits of the SUV's in the 90s. [NEWLINE] [NEWLINE] [STARTQ] Despite the history of their marques, many long running cars have been discontinued or relegated to fleet sales,[75][76][77] as GM, Ford and DaimlerChrysler shifted away resources from midsize and compact cars to lead the "SUV Craze". Since the late 1990s, over half of their profits have come from light trucks and SUVs, while they often could not break even on compact cars unless the buyer chose options.[78] Ron Harbour, in releasing the Oliver Wyman’s 2008 Harbour Report, stated that many small “econoboxes” of the past acted as loss leaders, but were designed to bring customers to the brand in the hopes they would stay loyal and move up to more profitable models. The report estimated that an automaker needed to sell ten small cars to make the same profit as one big vehicle, and that they had to produce small and mid-size cars profitably to succeed, something that the Detroit three have not yet done.[79] SUV sales peaked in 1999 but have not returned to that level ever since, due to higher gas prices. [ENDQ] [NEWLINE] [URL] #United_States [NEWLINE] [NEWLINE] [STARTQ] UAW Leadership granted concessions to its unions in order to win labor peace, a benefit not calculated by the UAW's many critics.[21] The UAW has claimed that the primary cause of the automotive sector's weakness was substantially more expensive fuel costs[22] linked to the 2003-2008 oil crisis which caused customers to turn away from large sport utility vehicles (SUVs) and pickup trucks,[23] the main market of the American "Big Three" (General Motors, Ford,
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First, I have to say it's very brave of you to come here and ask for us to challenge this view. Religiously motivated beliefs are always sensitive, and having a serious and mature discussion about them is difficult at best.<mask> too many people never really take the time to question their beliefs about the nature of the universe -<mask><mask> important they may be! -<mask> I'm always delighted to meet someone who does. [NEWLINE] [NEWLINE] Okay, let's see. You believe birds and dogs don't have a common ancestor.<mask> they share many similar features - two eyes, two nostrils, a mouth, four limbs and a tail, brain in the skull, lungs and heart in the ribcage, all other vital organs in the abdomen... I could go on. They are similar to each other in a way that they are not similar to, say, a jellyfish or a carrot. [NEWLINE] [NEWLINE] Of course this doesn't prove anything. Such similarity doesn't have to indicate common descent, just a common designer.<mask> all animals were made by the same being in the same way, you'd expect some similarities to show up.<mask> you would probably not expect any meaningful pattern in the similarities. [NEWLINE] [NEWLINE] <mask> would you tell whether two species are related? Well, presumably you'd look at various traits of those animals - ideally, their DNA,<mask> you can get it - and see<mask> similar they are. The more two species have in common, the more recently they must have diverged. And<mask> you take a large number of species, you can draw a [phylogenetic tree]( [URL] ) showing<mask> closely related they are to one another. This is like a "tree of life" picture you may have seen, with the leaves representing currently living species and the branches representing their various ancestors. [NEWLINE] [NEWLINE] Now,<mask> the world was created 6,000 years ago with the animals fully-formed, I would expect to see those trees "cut off" at that point. You'd have several leaves joining up into larger branches, and then stopping.<mask> there were no ancestors past that point, any similarities between animals are<mask> they live in similar environments, not<mask> they have family relations. Which means the similarities between animals present at the creation of the world should be more-or-less random,<mask> the similarities between their descendants should follow their family structure. [NEWLINE] [NEWLINE] <mask> we don't see this. The tree structure continues all the way back, and there's no visible cutoff around the 6000-year mark. Or anywhere
Label encoding: <s>First, I have to say it's very brave of you to come here and ask for us to challenge this view. Religiously motivated beliefs are always sensitive, and having a serious and mature discussion about them is difficult at best. But too many people never really take the time to question their beliefs about the nature of the universe - despite how important they may be! - so I'm always delighted to meet someone who does. [NEWLINE] [NEWLINE] Okay, let's see. You believe birds and dogs don't have a common ancestor. Yet they share many similar features - two eyes, two nostrils, a mouth, four limbs and a tail, brain in the skull, lungs and heart in the ribcage, all other vital organs in the abdomen... I could go on. They are similar to each other in a way that they are not similar to, say, a jellyfish or a carrot. [NEWLINE] [NEWLINE] Of course this doesn't prove anything. Such similarity doesn't have to indicate common descent, just a common designer. If all animals were made by the same being in the same way, you'd expect some similarities to show up. But you would probably not expect any meaningful pattern in the similarities. [NEWLINE] [NEWLINE] How would you tell whether two species are related? Well, presumably you'd look at various traits of those animals - ideally, their DNA, if you can get it - and see how similar they are. The more two species have in common, the more recently they must have diverged. And if you take a large number of species, you can draw a [phylogenetic tree]( [URL] ) showing how closely related they are to one another. This is like a "tree of life" picture you may have seen, with the leaves representing currently living species and the branches representing their various ancestors. [NEWLINE] [NEWLINE] Now, if the world was created 6,000 years ago with the animals fully-formed, I would expect to see those trees "cut off" at that point. You'd have several leaves joining up into larger branches, and then stopping. Since there were no ancestors past that point, any similarities between animals are because they live in similar environments, not because they have family relations. Which means the similarities between animals present at the creation of the world should be more-or-less random, while the similarities between their descendants should follow their family structure. [NEWLINE] [NEWLINE] But we don't see this. The tree structure continues all the way back, and there's no visible cutoff around the 6000-year mark. Or anywhere
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Rothko (1903-1970) was a Latvian-American 'abstract-expressionist' painter,<mask> he repeatedly rejected this label. He is most known for his geometric and rectangle-based paintings, such<mask> [Rust and Blue]( [URL].jpg), [Magenta, Black, Green on Orange]( [URL] /%27Magenta%2C_Black%2C_Green_on_Orange%27%2C_oil_on_canvas_painting_by_Mark_Rothko%2C_1947%2C_Museum_of_Modern_Art.jpg), or [No.14]( [URL].jpg/765px-Rothko_No_14.jpg). [NEWLINE] [NEWLINE] This style of painting does not seem like it required much effort, which alone causes me to appreciate the art less.<mask> to me, the actual art looks like only a bunch of different coloured rectangles, overlaid. I dislike this<mask> it does not fulfill either a) sending a message or motif; or b) looking artistically pleasing and meaningful. [NEWLINE] [NEWLINE] They are<mask> known for being very large canvasses, and they are. Rothko is quoted<mask> saying "The reason I paint them [largely],<mask>, [...] is precisely<mask> I want to be very intimate and human." I do not see<mask> these paintings, which are based around shapes (particularly rectangles), are any more humanizing than the work of other Abstract Expressionists such<mask> Jackson Pollock. [NEWLINE] [NEWLINE] He<mask> discusses an 'experience' of sorts: "The people who weep before my pictures are having the same religious experience I had<mask> I painted them. And<mask> you,<mask> you say, are moved only by their color relationship, then you miss the point." I do not have some spiritual experience<mask> I view his work.<mask><mask>, I do<mask> he dislikes in this quote; the only thing I can glean from the paintings are the color association.<mask> is the point that I am missing? [NEWLINE] [NEWLINE] In general, I am looking for an answer that<mask> does not *have* to resolve all of these difficulties,<mask> I am looking for one that gives some way to find meaning and value in Rothko's later work. I will reply<mask> best<mask> I can to any answers. [NEWLINE] [NEWLINE] EDIT: I'd just like to introduce a clarification and slight rewrite of my second-to-last point on the experience. I wrote that I do not feel and spiritual
Label encoding: <s>Rothko (1903-1970) was a Latvian-American 'abstract-expressionist' painter, though he repeatedly rejected this label. He is most known for his geometric and rectangle-based paintings, such as [Rust and Blue]( [URL].jpg), [Magenta, Black, Green on Orange]( [URL] /%27Magenta%2C_Black%2C_Green_on_Orange%27%2C_oil_on_canvas_painting_by_Mark_Rothko%2C_1947%2C_Museum_of_Modern_Art.jpg), or [No.14]( [URL].jpg/765px-Rothko_No_14.jpg). [NEWLINE] [NEWLINE] This style of painting does not seem like it required much effort, which alone causes me to appreciate the art less. But to me, the actual art looks like only a bunch of different coloured rectangles, overlaid. I dislike this because it does not fulfill either a) sending a message or motif; or b) looking artistically pleasing and meaningful. [NEWLINE] [NEWLINE] They are also known for being very large canvasses, and they are. Rothko is quoted as saying "The reason I paint them [largely], however, [...] is precisely because I want to be very intimate and human." I do not see how these paintings, which are based around shapes (particularly rectangles), are any more humanizing than the work of other Abstract Expressionists such as Jackson Pollock. [NEWLINE] [NEWLINE] He also discusses an 'experience' of sorts: "The people who weep before my pictures are having the same religious experience I had when I painted them. And if you, as you say, are moved only by their color relationship, then you miss the point." I do not have some spiritual experience when I view his work. In fact, I do what he dislikes in this quote; the only thing I can glean from the paintings are the color association. What is the point that I am missing? [NEWLINE] [NEWLINE] In general, I am looking for an answer that while does not *have* to resolve all of these difficulties, but I am looking for one that gives some way to find meaning and value in Rothko's later work. I will reply as best as I can to any answers. [NEWLINE] [NEWLINE] EDIT: I'd just like to introduce a clarification and slight rewrite of my second-to-last point on the experience. I wrote that I do not feel and spiritual
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Well yeah, of course you can. Its pretty easy to work backwards<mask> you are sure you already know the answer. [ENDQ] [NEWLINE] K,<mask> you agree with me on those. Moving on: [NEWLINE] [NEWLINE] 1. We have more chronic diseases<mask> people are living longer, and people haven't evolved to deal with these types of diseases from old age. Again, that's something that has occurred mostly due to the free-market making food more abundant, and medical science is still catching up<mask> new diseases are harder to treat. [NEWLINE] [NEWLINE] 2. Most of these diagnostic procedures are unnecessary, and are performed strictly in defense of ambulance chasing trial lawyers. [NEWLINE] [NEWLINE] 3. Thanks to the Baby Boom Generation due to the US Government's involvement in WW2 which caused all those soldiers to come home and fuck like rabbits. And the aforementioned increase in lifespan due to technological and agricultural prowess. [NEWLINE] [NEWLINE] [STARTQ] It has really world examples of efficacy in countries that are comparable to ours demographically and economically. [ENDQ] [NEWLINE] And those countries experience long lines, less technologically advanced services, and [people *still* end up dying from lack of care.]( [URL] )<mask> are you<mask> adamant about defending a system with obvious flaws instead of advocating for experimentation with new systems? Again, I'm saying that the US's system *isn't* a free-market system. I'll definitely agree with you that the US Government and crony-capitalists suck. I'm not comparing the US to Europe. I'm comparing these other systems to the system that *used* to exist, and that has<mask> been banned by Governments.<mask> you're actually in Medical School, then you should be pro-experimentation. That's<mask> we derive new knowledge, from pre-clinical and clinical experiments. [NEWLINE] [NEWLINE] [STARTQ] That for most disease outcomes european social systems produce the same or better outcomes at a lower price point. [ENDQ] [NEWLINE] And these statistics are tracked by the government, which many times leave out unfavorable data points, [like babies that are born prematurely.]( [URL] /) [NEWLINE] [NEWLINE] [STARTQ] <mask>, before I waste more time out of studying for boards on my internal medicine rotations, do you really want to have your view changed or are you just interested in convincing others that libertarianism has all the answers? [ENDQ] [NEWLINE] Lolz, i'm not a libertarian; i'm an anarcho-capitalist. Don't blame me for wasting your time, take some personal responsibility. And stop bringing up the fact that you're studying to be a doctor, like it
Label encoding: <s> [STARTQ] Well yeah, of course you can. Its pretty easy to work backwards when you are sure you already know the answer. [ENDQ] [NEWLINE] K, so you agree with me on those. Moving on: [NEWLINE] [NEWLINE] 1. We have more chronic diseases because people are living longer, and people haven't evolved to deal with these types of diseases from old age. Again, that's something that has occurred mostly due to the free-market making food more abundant, and medical science is still catching up because new diseases are harder to treat. [NEWLINE] [NEWLINE] 2. Most of these diagnostic procedures are unnecessary, and are performed strictly in defense of ambulance chasing trial lawyers. [NEWLINE] [NEWLINE] 3. Thanks to the Baby Boom Generation due to the US Government's involvement in WW2 which caused all those soldiers to come home and fuck like rabbits. And the aforementioned increase in lifespan due to technological and agricultural prowess. [NEWLINE] [NEWLINE] [STARTQ] It has really world examples of efficacy in countries that are comparable to ours demographically and economically. [ENDQ] [NEWLINE] And those countries experience long lines, less technologically advanced services, and [people *still* end up dying from lack of care.]( [URL] ) Why are you so adamant about defending a system with obvious flaws instead of advocating for experimentation with new systems? Again, I'm saying that the US's system *isn't* a free-market system. I'll definitely agree with you that the US Government and crony-capitalists suck. I'm not comparing the US to Europe. I'm comparing these other systems to the system that *used* to exist, and that has since been banned by Governments. If you're actually in Medical School, then you should be pro-experimentation. That's how we derive new knowledge, from pre-clinical and clinical experiments. [NEWLINE] [NEWLINE] [STARTQ] That for most disease outcomes european social systems produce the same or better outcomes at a lower price point. [ENDQ] [NEWLINE] And these statistics are tracked by the government, which many times leave out unfavorable data points, [like babies that are born prematurely.]( [URL] /) [NEWLINE] [NEWLINE] [STARTQ] Also, before I waste more time out of studying for boards on my internal medicine rotations, do you really want to have your view changed or are you just interested in convincing others that libertarianism has all the answers? [ENDQ] [NEWLINE] Lolz, i'm not a libertarian; i'm an anarcho-capitalist. Don't blame me for wasting your time, take some personal responsibility. And stop bringing up the fact that you're studying to be a doctor, like it
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0850, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> the price of your neighborhood goes up, you have to uproot your entire life with minimal notice and move to another part of town or another town entirely, and for people living paycheck to paycheck<mask> surviving, they may not be able to absorb the sudden imposition of the transaction costs of moving. [ENDQ] [NEWLINE] This isn't any different than a variety of things that we actually do enforce.  I have friends who didn't carry insurance prior to Obamacare that now work another job to pay for that (and other things,<mask> it's not like you can get a job working 8 hours a week). <mask> gas prices went up, people couldn't afford to drive some of their gas guzzling cars<mask><mask> couldn't afford to replace them with something more efficient.  Hell, we<mask> do this with property taxes. [NEWLINE] [NEWLINE] There are<mask> benefits to displacing low rent properties with higher rent.  This is a big controversy in my city right now (it's a sizable city,<mask> not New York size).  For the last 40 or<mask> years, Downtown has been a place you *do not* want to live in<mask> you can afford not to.  It's dirty, the houses (which were once very nice, and are actually recognized for the architecture) are in disrepair, and the amount of commerce is severely hampered by the income of the surrounding neighborhood.  Recently, developers have been buying these houses, renovating them, and then charging rent that could easily be triple<mask> it was previously.  The people that used to live there can no longer afford it and have to move, *<mask> * the number and quality of the businesses has skyrocketed.  It's become an attraction.  It's a place to go on Friday nights, rather than the kind of place<mask> you lock your windows<mask> you have to stop for more than a minute. [NEWLINE] [NEWLINE] [STARTQ] it's not reasonable to expect everyone to come up with first and last month's rent + security deposit + moving costs with minimal notice [ENDQ] [NEWLINE] First and foremost, assuming you've been a good tenant, moving costs are the only one of those you'll have to pay.  You'll get your security deposit back, and your first month's rent should be payed<mask> or slightly before you move in.  Moving fees should be minimal<mask> well, it's only $20 for a day rental on a 10' Uhaul plus $0.80/mile. <mask> you cannot afford that, then you should have been in a cheaper apartment already
Label encoding: <s> [STARTQ] If the price of your neighborhood goes up, you have to uproot your entire life with minimal notice and move to another part of town or another town entirely, and for people living paycheck to paycheck but surviving, they may not be able to absorb the sudden imposition of the transaction costs of moving. [ENDQ] [NEWLINE] This isn't any different than a variety of things that we actually do enforce.  I have friends who didn't carry insurance prior to Obamacare that now work another job to pay for that (and other things, but it's not like you can get a job working 8 hours a week).  When gas prices went up, people couldn't afford to drive some of their gas guzzling cars but also couldn't afford to replace them with something more efficient.  Hell, we also do this with property taxes. [NEWLINE] [NEWLINE] There are also benefits to displacing low rent properties with higher rent.  This is a big controversy in my city right now (it's a sizable city, though not New York size).  For the last 40 or so years, Downtown has been a place you *do not* want to live in if you can afford not to.  It's dirty, the houses (which were once very nice, and are actually recognized for the architecture) are in disrepair, and the amount of commerce is severely hampered by the income of the surrounding neighborhood.  Recently, developers have been buying these houses, renovating them, and then charging rent that could easily be triple what it was previously.  The people that used to live there can no longer afford it and have to move, * but * the number and quality of the businesses has skyrocketed.  It's become an attraction.  It's a place to go on Friday nights, rather than the kind of place where you lock your windows if you have to stop for more than a minute. [NEWLINE] [NEWLINE] [STARTQ] it's not reasonable to expect everyone to come up with first and last month's rent + security deposit + moving costs with minimal notice [ENDQ] [NEWLINE] First and foremost, assuming you've been a good tenant, moving costs are the only one of those you'll have to pay.  You'll get your security deposit back, and your first month's rent should be payed when or slightly before you move in.  Moving fees should be minimal as well, it's only $20 for a day rental on a 10' Uhaul plus $0.80/mile.  If you cannot afford that, then you should have been in a cheaper apartment already
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's an aspect of the evangelical movement,<mask> not really mainstream or that common. A lot of the ways they indoctrinated the children were not with biblical things<mask> with activities like smashing cups saying government. [ENDQ] [NEWLINE] I beg to differ on it being uncommon. The practices she conducted are actually fairly common among many denominations<mask> dealing with camp. [NEWLINE] [NEWLINE] I've talked to many different people from different sects of christianity including baptist, pentacostal, and evangelical and have received enough feedback that<mask> it comes to church camp, they push the boundaries of<mask> they normally behave. Camps set goals to get<mask> many people saved<mask> possible and will have a worship seminar every night at a lot of camps. [NEWLINE] [NEWLINE] <mask>,<mask> talking to these people they have admitted to doing rituals similar to the smashing cups and dancing around with war paint and look back upon those rituals with embarrasement. The rituals were fluff and did not have a lasting impression on the persons. [NEWLINE] [NEWLINE] <mask> did have a lasting impression was something that was just<mask> powerful<mask> more subtle. Many people<mask> at a church camp will either speak in tongues or attempt to, they'll pledge allegiance to the Bible, and even get on their hands and needs till their pouring tears from their eyes. All of these things<mask> looked back upon without any regret. [NEWLINE] [NEWLINE] Subtle things like pledging your alliegence to something you're too young to comprehend is just<mask> radical<mask> smashing cups dramatically without fully understanding<mask>.<mask>, they are not treated the same over time. The pledges and dramatic worship techniques these children perform stick with them<mask> the fluff typically just goes away. [NEWLINE] [NEWLINE] [STARTQ] <mask> you believe that the Bible is used<mask> a propaganda tool like Mein Kampf you should cite a time in the film<mask> the bible was used like that, not just make the statement that it is<mask>. [ENDQ] [NEWLINE] I guess I did not make myself clear. I did not say the film pointed it out, I'm saying someone I've spoken to has pointed out my method and said it can be applied to Mein Kampf. [NEWLINE] [NEWLINE] I've never read Mein Kampf<mask> I can't add any support for or against it. I plan on investigating it in the future<mask>. [NEWLINE] [NEWLINE] [STARTQ] Sounds like charisma, combined with good training. Did you see that the bible verses were especially effective or used especially effectively for some reason? [ENDQ] [NEWLINE] <mask><mask><mask> we're arguing here is over the use paraphrasing context or stating it directly. The
Label encoding: <s> [STARTQ] It's an aspect of the evangelical movement, but not really mainstream or that common. A lot of the ways they indoctrinated the children were not with biblical things but with activities like smashing cups saying government. [ENDQ] [NEWLINE] I beg to differ on it being uncommon. The practices she conducted are actually fairly common among many denominations while dealing with camp. [NEWLINE] [NEWLINE] I've talked to many different people from different sects of christianity including baptist, pentacostal, and evangelical and have received enough feedback that when it comes to church camp, they push the boundaries of how they normally behave. Camps set goals to get as many people saved as possible and will have a worship seminar every night at a lot of camps. [NEWLINE] [NEWLINE] Also, while talking to these people they have admitted to doing rituals similar to the smashing cups and dancing around with war paint and look back upon those rituals with embarrasement. The rituals were fluff and did not have a lasting impression on the persons. [NEWLINE] [NEWLINE] What did have a lasting impression was something that was just as powerful but more subtle. Many people while at a church camp will either speak in tongues or attempt to, they'll pledge allegiance to the Bible, and even get on their hands and needs till their pouring tears from their eyes. All of these things when looked back upon without any regret. [NEWLINE] [NEWLINE] Subtle things like pledging your alliegence to something you're too young to comprehend is just as radical as smashing cups dramatically without fully understanding why. However, they are not treated the same over time. The pledges and dramatic worship techniques these children perform stick with them while the fluff typically just goes away. [NEWLINE] [NEWLINE] [STARTQ] If you believe that the Bible is used as a propaganda tool like Mein Kampf you should cite a time in the film when the bible was used like that, not just make the statement that it is because. [ENDQ] [NEWLINE] I guess I did not make myself clear. I did not say the film pointed it out, I'm saying someone I've spoken to has pointed out my method and said it can be applied to Mein Kampf. [NEWLINE] [NEWLINE] I've never read Mein Kampf so I can't add any support for or against it. I plan on investigating it in the future though. [NEWLINE] [NEWLINE] [STARTQ] Sounds like charisma, combined with good training. Did you see that the bible verses were especially effective or used especially effectively for some reason? [ENDQ] [NEWLINE] I think what we're arguing here is over the use paraphrasing context or stating it directly. The
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I heard Mike Huckabee say this during the first Republican Debate, and until I went online the next morning, I couldn't imagine anyone would disagree with him. People were saying this was shameful, and embarrassing to the US military, a ridiculous simplification. To be clear, I am no fan of Huckabee, and I have nothing against our military or militaries in general. That said,<mask><mask> he was dead on. [NEWLINE] [NEWLINE] <mask><mask> most people who don't accept this do<mask><mask> they think it's crass, and brutal. Well, yeah,<mask> it's<mask> absolutely true. That's<mask> the military has all those rifles and artillery and tanks and battleships and bombs and chemical weapons and knives and humvees and machine guns: for the killing of people and the breaking of things. Sure, not every member of the military's job is to be a killer,<mask> those people are there to support the killers. The cooks, mechanics, engineers, and secretaries are all there to let everyone else kill people and break things<mask> safely and efficiently<mask> possible. [NEWLINE] [NEWLINE] Again, I have absolutely no problem with this from a moral perspective. I am certainly not condemning anybody, just stating facts. Most people I've seen disagree with Huckabee are just dancing around this. "Soldiers exist to protect the United States and her interests!" Sure, using violence or the threat of violence. "The army doesn't just kill people, they developed the Internet!" Yeah,<mask> a weapon to coordinate their violence in the most efficient way possible. The internet we have now is just an unintended side effect. [NEWLINE] [NEWLINE] The US military is in a bit of a unique position,<mask> they haven't had a lot of opportunities to do their job recently. Our military is<mask> badass, there are not a lot of people with enough courage or stupidity to take us on.<mask>, a lot of time is spent running practice drills and handing out food to people after earthquakes. That's great,<mask> it's all a displacement activity until they need to do their real job. It's a sideshow, a distraction. The alternative would be these men sitting around, waiting for something to need destroying. We don't keep our military around and spend billions of dollars for disaster relief. We keep them around to kill people and break things. That's their real purpose, simple<mask> that. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.
Label encoding: <s>I heard Mike Huckabee say this during the first Republican Debate, and until I went online the next morning, I couldn't imagine anyone would disagree with him. People were saying this was shameful, and embarrassing to the US military, a ridiculous simplification. To be clear, I am no fan of Huckabee, and I have nothing against our military or militaries in general. That said, I think he was dead on. [NEWLINE] [NEWLINE] I think most people who don't accept this do so because they think it's crass, and brutal. Well, yeah, but it's also absolutely true. That's why the military has all those rifles and artillery and tanks and battleships and bombs and chemical weapons and knives and humvees and machine guns: for the killing of people and the breaking of things. Sure, not every member of the military's job is to be a killer, but those people are there to support the killers. The cooks, mechanics, engineers, and secretaries are all there to let everyone else kill people and break things as safely and efficiently as possible. [NEWLINE] [NEWLINE] Again, I have absolutely no problem with this from a moral perspective. I am certainly not condemning anybody, just stating facts. Most people I've seen disagree with Huckabee are just dancing around this. "Soldiers exist to protect the United States and her interests!" Sure, using violence or the threat of violence. "The army doesn't just kill people, they developed the Internet!" Yeah, as a weapon to coordinate their violence in the most efficient way possible. The internet we have now is just an unintended side effect. [NEWLINE] [NEWLINE] The US military is in a bit of a unique position, since they haven't had a lot of opportunities to do their job recently. Our military is so badass, there are not a lot of people with enough courage or stupidity to take us on. Thus, a lot of time is spent running practice drills and handing out food to people after earthquakes. That's great, but it's all a displacement activity until they need to do their real job. It's a sideshow, a distraction. The alternative would be these men sitting around, waiting for something to need destroying. We don't keep our military around and spend billions of dollars for disaster relief. We keep them around to kill people and break things. That's their real purpose, simple as that. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just a heads up, the fastest way to show that you know nothing about music is to say, "I like all kinds of music."  This shines through even more<mask> you only list radio songs for all of your examples.  There are<mask> many kinds of music that even professional music critics could not possibly experience every kind of music. [NEWLINE] [NEWLINE] All rudeness aside<mask>, high quality modern musicians have to know<mask> much more than musicians of the past.  The software that you dismiss<mask> casually can take several college courses just to get good at using one program. [NEWLINE] [NEWLINE] Learning<mask> to discover new music is probably the best way to change your ideas.  I would suggest subscribing to a few of the music subreddits other than r/music.  r/music is a huge useless circle-jerk of Queen and the Beatles.  Try /r/listentothis<mask> a jumping off point. [NEWLINE] [NEWLINE] Just<mask> an example of<mask> broad one genre of music can be, take a moment to look at www.mapofmetal.com.  This is a good demonstration of hundreds of different bands in dozens of subgenres just of metal music.  Now realize that this could be done with every genre of music, many of which you've never even heard of. [NEWLINE] [NEWLINE] <mask> for modern good music, here are a few suggestions to get you started: [NEWLINE] [NEWLINE] Thrice (post-punk): [The Artist in the Ambulance]( [URL] )  or [Digital Sea]( [URL] ) [NEWLINE] [NEWLINE] Kishi Bashi (no idea the genre) [Bright Whites]( [URL] ) [NEWLINE] [NEWLINE] Wintergatan (instrumental): [starmachine]( [URL] ) [NEWLINE] [NEWLINE] Sonata Arctica (Power Metal): [Victoria's Secret]( [URL] ) [NEWLINE] [NEWLINE] Stephen Swartz (Dubstep): [Bullet Train]( [URL] ) [NEWLINE] [NEWLINE] Whiskers (Dubstep / Chiptune): [Game Boy]( [URL] ) [NEWLINE] [NEWLINE] Chipzel (chiptune): [Focus]( [URL] ) [NEWLINE] [NEWLINE] Tally Hall (upbeat rock): [Ruler of Everything]( [URL] -dsW8A) [NEWLINE] [NEWLINE] Bright Eyes (Basically<mask> Bob Dylan had musical talent and could sing): [Bowl of Oranges]( [URL] ) or [Take It Easy]( [URL] ) [NEWLINE] [NEWLINE] The Limousines (Electro-pop): [Very Busy People]( [URL] ) [NEWLINE] [NEWLINE] Jon Gomm (Singer/songwriter): [Passionflower]( [URL] ) (The first 50 seconds are wanky,<mask>
Label encoding: <s>Just a heads up, the fastest way to show that you know nothing about music is to say, "I like all kinds of music."  This shines through even more when you only list radio songs for all of your examples.  There are so many kinds of music that even professional music critics could not possibly experience every kind of music. [NEWLINE] [NEWLINE] All rudeness aside though, high quality modern musicians have to know so much more than musicians of the past.  The software that you dismiss so casually can take several college courses just to get good at using one program. [NEWLINE] [NEWLINE] Learning how to discover new music is probably the best way to change your ideas.  I would suggest subscribing to a few of the music subreddits other than r/music.  r/music is a huge useless circle-jerk of Queen and the Beatles.  Try /r/listentothis as a jumping off point. [NEWLINE] [NEWLINE] Just as an example of how broad one genre of music can be, take a moment to look at www.mapofmetal.com.  This is a good demonstration of hundreds of different bands in dozens of subgenres just of metal music.  Now realize that this could be done with every genre of music, many of which you've never even heard of. [NEWLINE] [NEWLINE] As for modern good music, here are a few suggestions to get you started: [NEWLINE] [NEWLINE] Thrice (post-punk): [The Artist in the Ambulance]( [URL] )  or [Digital Sea]( [URL] ) [NEWLINE] [NEWLINE] Kishi Bashi (no idea the genre) [Bright Whites]( [URL] ) [NEWLINE] [NEWLINE] Wintergatan (instrumental): [starmachine]( [URL] ) [NEWLINE] [NEWLINE] Sonata Arctica (Power Metal): [Victoria's Secret]( [URL] ) [NEWLINE] [NEWLINE] Stephen Swartz (Dubstep): [Bullet Train]( [URL] ) [NEWLINE] [NEWLINE] Whiskers (Dubstep / Chiptune): [Game Boy]( [URL] ) [NEWLINE] [NEWLINE] Chipzel (chiptune): [Focus]( [URL] ) [NEWLINE] [NEWLINE] Tally Hall (upbeat rock): [Ruler of Everything]( [URL] -dsW8A) [NEWLINE] [NEWLINE] Bright Eyes (Basically if Bob Dylan had musical talent and could sing): [Bowl of Oranges]( [URL] ) or [Take It Easy]( [URL] ) [NEWLINE] [NEWLINE] The Limousines (Electro-pop): [Very Busy People]( [URL] ) [NEWLINE] [NEWLINE] Jon Gomm (Singer/songwriter): [Passionflower]( [URL] ) (The first 50 seconds are wanky, but
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> the general issue is a lot of the things you suggest are easier said than done. Few examples... [NEWLINE] [NEWLINE] [STARTQ] Sometimes shit happens and people need help,<mask> six months to a year should be enough time for anyone to get out of this category,<mask> many don't. [ENDQ] [NEWLINE] That's not really reasonable to say, every situation is different. [NEWLINE] [NEWLINE] [STARTQ] It's not difficult to see<mask> a degree will be worth the salary raise or<mask> a loan will actually solve your current problems and their are more than enough resources online and in libraries to learn about proper finance skills in order to do<mask>. [ENDQ] [NEWLINE] Loans usually aren't an option, its the only choice. That's one of the biggest problems<mask>, due to some circumstance, someone can't make money for some period of time and are forced to get a loan to cover mandatory living expenses for themselves or their families.<mask> they don't make much money, its easy for the debt to spiral out of control. [NEWLINE] [NEWLINE] <mask> for learning to better manage their finances,<mask><mask> people don't have the time? Maybe its just a few hours,<mask> there are lots of people who work multiple jobs or have other responsibilities. Again, much easier said than done. [NEWLINE] [NEWLINE] [STARTQ] I think problems like a botched health care system, over-imprisonment for drug sentences, and a generally rigged system such<mask> having to pay to cash checks, not being able to afford the bulk deals, or own a home are factors that plague the lower class and rigidize social mobility,<mask> it does not take away their opportunity to pull themselves up. [ENDQ] [NEWLINE] <mask> it makes it much harder. I don't think the issue is that its impossible,<mask> that its exceedingly difficult. [NEWLINE] [NEWLINE] [STARTQ] There's no question that some situations are more difficult than others,<mask> there is<mask> no question that examples of social mobility exist in both directions: rich kids wasting away a fortune and migrant farm workers becoming well-paid astronauts. [ENDQ] [NEWLINE] A lot easier to lose money than it is to make money. [NEWLINE] [NEWLINE] [STARTQ] they would prefer to work a minimum wage job instead of applying for better positions--I can't say. [ENDQ] [NEWLINE] That is not the case, ever. Nobody decides that they would rather make less money<mask> they like working a minimum wage job. The problem is they are either unable to take those positions due to a lack of education, or there are just no such positions. Consider something like a McDonalds.<mask> many managerial positions are there compared to the minimum wage jobs?
Label encoding: <s>I think the general issue is a lot of the things you suggest are easier said than done. Few examples... [NEWLINE] [NEWLINE] [STARTQ] Sometimes shit happens and people need help, but six months to a year should be enough time for anyone to get out of this category, yet many don't. [ENDQ] [NEWLINE] That's not really reasonable to say, every situation is different. [NEWLINE] [NEWLINE] [STARTQ] It's not difficult to see if a degree will be worth the salary raise or if a loan will actually solve your current problems and their are more than enough resources online and in libraries to learn about proper finance skills in order to do so. [ENDQ] [NEWLINE] Loans usually aren't an option, its the only choice. That's one of the biggest problems where, due to some circumstance, someone can't make money for some period of time and are forced to get a loan to cover mandatory living expenses for themselves or their families. If they don't make much money, its easy for the debt to spiral out of control. [NEWLINE] [NEWLINE] As for learning to better manage their finances, what if people don't have the time? Maybe its just a few hours, but there are lots of people who work multiple jobs or have other responsibilities. Again, much easier said than done. [NEWLINE] [NEWLINE] [STARTQ] I think problems like a botched health care system, over-imprisonment for drug sentences, and a generally rigged system such as having to pay to cash checks, not being able to afford the bulk deals, or own a home are factors that plague the lower class and rigidize social mobility, but it does not take away their opportunity to pull themselves up. [ENDQ] [NEWLINE] But it makes it much harder. I don't think the issue is that its impossible, but that its exceedingly difficult. [NEWLINE] [NEWLINE] [STARTQ] There's no question that some situations are more difficult than others, but there is also no question that examples of social mobility exist in both directions: rich kids wasting away a fortune and migrant farm workers becoming well-paid astronauts. [ENDQ] [NEWLINE] A lot easier to lose money than it is to make money. [NEWLINE] [NEWLINE] [STARTQ] they would prefer to work a minimum wage job instead of applying for better positions--I can't say. [ENDQ] [NEWLINE] That is not the case, ever. Nobody decides that they would rather make less money because they like working a minimum wage job. The problem is they are either unable to take those positions due to a lack of education, or there are just no such positions. Consider something like a McDonalds. How many managerial positions are there compared to the minimum wage jobs?
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Climate change or global warming<mask> it previously was known<mask> is a hot potatoe and not a very clean cut science to predict. There are thousands of climate scientists working on it, trying to predict<mask> we are heading and<mask> fast we are getting there. To make things even harder, [people with special interests are joining the battle]( [URL] #Climate_change) throwing in false baits and helping naysayers gain traction. [NEWLINE] [NEWLINE] All this,<mask> we hear more and more about [reports on ocean acidification is beginning to take effect]( [URL] ), and<mask> usual, [faster than we had predicted.]( [URL] #Rate) [NEWLINE] [NEWLINE] I am currently actively volunteering for green party politics (not in the US), trying to make a difference.<mask> I notice is that the big bad wolf amongst "us" is the burning of fossil fuels. It's like everybody is on a rampage against oil and it's being preached that<mask> we get rid of the burning of fossil fuels we'll live happier ever after.<mask><mask><mask><mask> tipped me over the edge was realizing that this is just a media stunt. Of course fossil fuels is a part of the problem,<mask> say we'd rid the world of oil by tomorrow, it would acutally only rid us of [13% of the problem]( [URL] #two). All those focusing on "saving the world" aren't even acknowledging the rest of the problem. God forbid I mention people should eat less meat,<mask><mask> I do, you'll see the greenest of environmentalists turn red in rage<mask> I'm asking them to change something with themselves instead of the big bad wolf. [NEWLINE] [NEWLINE] Anyway, I believe that we are doomed<mask> : [NEWLINE] [NEWLINE] * The world is going under faster and faster [NEWLINE] * People don't care and keep consuming [NEWLINE] * Those who care don't seem to really want to change or are too disoriented to see the big picture [NEWLINE] * We won't actually START making a difference until it starts hurting,<mask><mask> it's hurting we'll already be dead. [NEWLINE] [NEWLINE] Please change my view! :/ [NEWLINE] [NEWLINE] PS, I apologize beforehand for any grammatical errors or unstructured sentences. English isn't my primary language. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one,
Label encoding: <s>Climate change or global warming as it previously was known as is a hot potatoe and not a very clean cut science to predict. There are thousands of climate scientists working on it, trying to predict where we are heading and how fast we are getting there. To make things even harder, [people with special interests are joining the battle]( [URL] #Climate_change) throwing in false baits and helping naysayers gain traction. [NEWLINE] [NEWLINE] All this, while we hear more and more about [reports on ocean acidification is beginning to take effect]( [URL] ), and as usual, [faster than we had predicted.]( [URL] #Rate) [NEWLINE] [NEWLINE] I am currently actively volunteering for green party politics (not in the US), trying to make a difference. What I notice is that the big bad wolf amongst "us" is the burning of fossil fuels. It's like everybody is on a rampage against oil and it's being preached that if we get rid of the burning of fossil fuels we'll live happier ever after. But I think what tipped me over the edge was realizing that this is just a media stunt. Of course fossil fuels is a part of the problem, but say we'd rid the world of oil by tomorrow, it would acutally only rid us of [13% of the problem]( [URL] #two). All those focusing on "saving the world" aren't even acknowledging the rest of the problem. God forbid I mention people should eat less meat, because when I do, you'll see the greenest of environmentalists turn red in rage because I'm asking them to change something with themselves instead of the big bad wolf. [NEWLINE] [NEWLINE] Anyway, I believe that we are doomed because : [NEWLINE] [NEWLINE] * The world is going under faster and faster [NEWLINE] * People don't care and keep consuming [NEWLINE] * Those who care don't seem to really want to change or are too disoriented to see the big picture [NEWLINE] * We won't actually START making a difference until it starts hurting, but when it's hurting we'll already be dead. [NEWLINE] [NEWLINE] Please change my view! :/ [NEWLINE] [NEWLINE] PS, I apologize beforehand for any grammatical errors or unstructured sentences. English isn't my primary language. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one,
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I want to impress that I do not believe that you can compare wearing a burqa to having to wear a shirt. This is like comparing first degree murder with stealing candy. They're both crimes, and they're both bad,<mask> one is definitively worse than the other. [NEWLINE] [NEWLINE] A burqa is designed to dehumanize and remove a woman's individuality. Objectively, you should be able to see that this is a different degree of "wrongness" than requiring that a woman wear a top in public even<mask> you object to the latter<mask> well. Wearing a burqa prevents many of the subtle contextual clues that humans use in daily communication and limits their ability to read and convey body language. [NEWLINE] [NEWLINE] <mask> for it being a societal norm, of course there are women who choose to wear a burqa and are happy just<mask> there were women who bound their feet and were happy to marry<mask> well. There are<mask> a vast number of women who are forced to wear a burqa out of fear or coercion by their family or society. [NEWLINE] [NEWLINE] My point is that they are both incorrect,<mask> at gross ends of a scale which is ever sliding depending on cultural context. Is it hypocritical to support one and not support the other? I don't believe<mask>. To compare it to another issue that no one can seem to agree on, let's look at taxes. Most rational people agree that there should be some type of tax to help pay for all the things that we enjoy using taxes for.<mask><mask> much tax? You might be happy paying 20% tax<mask> you would probably be very angry at paying 90% tax. [NEWLINE] [NEWLINE] Most modern societies believe that there should be some level of clothing required for being in public. Of course, there are the anarchists who believe we should all go around naked<mask> well.<mask>, we vary from society to society (and from individual to individual) on<mask> much is too much. The vast majority of us can agree that 90% is too much whether we're talking about taxes or clothing. [NEWLINE] [NEWLINE] <mask><mask> about<mask> you get down to 20%. It's pretty reasonable.<mask><mask> is 25%, or 15%. Once you get down to that level, you have less of a "That's wrong! We have to change it!" reaction<mask> honestly you're reaching the point<mask> people start negotiating and aren't really sure<mask> they stand. The vast majority of people don't feel oppressed,<mask> they may sympathize with those who are in the minority who do feel oppressed.<mask>, at this
Label encoding: <s>I want to impress that I do not believe that you can compare wearing a burqa to having to wear a shirt. This is like comparing first degree murder with stealing candy. They're both crimes, and they're both bad, but one is definitively worse than the other. [NEWLINE] [NEWLINE] A burqa is designed to dehumanize and remove a woman's individuality. Objectively, you should be able to see that this is a different degree of "wrongness" than requiring that a woman wear a top in public even if you object to the latter as well. Wearing a burqa prevents many of the subtle contextual clues that humans use in daily communication and limits their ability to read and convey body language. [NEWLINE] [NEWLINE] As for it being a societal norm, of course there are women who choose to wear a burqa and are happy just as there were women who bound their feet and were happy to marry as well. There are also a vast number of women who are forced to wear a burqa out of fear or coercion by their family or society. [NEWLINE] [NEWLINE] My point is that they are both incorrect, but at gross ends of a scale which is ever sliding depending on cultural context. Is it hypocritical to support one and not support the other? I don't believe so. To compare it to another issue that no one can seem to agree on, let's look at taxes. Most rational people agree that there should be some type of tax to help pay for all the things that we enjoy using taxes for. But how much tax? You might be happy paying 20% tax but you would probably be very angry at paying 90% tax. [NEWLINE] [NEWLINE] Most modern societies believe that there should be some level of clothing required for being in public. Of course, there are the anarchists who believe we should all go around naked as well. However, we vary from society to society (and from individual to individual) on how much is too much. The vast majority of us can agree that 90% is too much whether we're talking about taxes or clothing. [NEWLINE] [NEWLINE] But what about when you get down to 20%. It's pretty reasonable. But so is 25%, or 15%. Once you get down to that level, you have less of a "That's wrong! We have to change it!" reaction because honestly you're reaching the point where people start negotiating and aren't really sure where they stand. The vast majority of people don't feel oppressed, but they may sympathize with those who are in the minority who do feel oppressed. However, at this
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>With the advent of self-driving vehicles, the unavoidable will become clear: people are terrible drivers, and operating your own car is unacceptably reckless<mask> a better alternative exists.  I see the coming timeline like this: (copied from a reply to another post) [NEWLINE] [NEWLINE] [STARTQ] 2-5 years: The last major technological hurdles (driving in rural/poorly documented areas, driving in adverse conditions, cost) are resolved. Cars are now demonstratively better drivers than humans in all situations. (note: may be a very liberal estimate.) [ENDQ] 4-6 years: The first round of legal cases involving driverless cars is settled, producing a precedent that makes driving your own car very risky. A collision between two vehicles, one self driving the other not, almost always results in fault to the driver. Causing an accident<mask> operating a car with unused self-driving capability makes drivers extremely vulnerable to being sued. [NEWLINE] 5-10 years: Safety studies, overwhelmingly favorable to self-driving cars, lead to the option becoming mandatory on all new vehicles. insurance companies, burned by litigation, offer premium rates to those who never switch off the driverless option,<mask> increasing rates on drivers who elect to operate their cars manually. Soon the difference between these rates becomes enormous. [NEWLINE] 10-15 years: Commercial driving is entirely automated. Cabs, buses, trucks, trains, "driver" becomes an obsolete profession. The savings in both wages and liability is simply too tremendous to allow any non-automated fleet to remain competitive. [NEWLINE] 15-20 years: Studies conclusively show that the only traffic casualties that still occur are exclusively due to human operator error. It becomes evident that driving your own car is unthinkably dangerous, like drunk driving at night with no headlights or seatbelts. Safety laws are passed that effectively outlaw operating your own vehicle. [NEWLINE] [NEWLINE] By the time my nephew is 15-16, controlling a car will be something that only hobbyists do, and never on public roads.  Very few cars will be privately owned, rather they will be operated by private or municipal transportation services. [NEWLINE] The age of the personal automobile is ending. CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than
Label encoding: <s>With the advent of self-driving vehicles, the unavoidable will become clear: people are terrible drivers, and operating your own car is unacceptably reckless if a better alternative exists.  I see the coming timeline like this: (copied from a reply to another post) [NEWLINE] [NEWLINE] [STARTQ] 2-5 years: The last major technological hurdles (driving in rural/poorly documented areas, driving in adverse conditions, cost) are resolved. Cars are now demonstratively better drivers than humans in all situations. (note: may be a very liberal estimate.) [ENDQ] 4-6 years: The first round of legal cases involving driverless cars is settled, producing a precedent that makes driving your own car very risky. A collision between two vehicles, one self driving the other not, almost always results in fault to the driver. Causing an accident while operating a car with unused self-driving capability makes drivers extremely vulnerable to being sued. [NEWLINE] 5-10 years: Safety studies, overwhelmingly favorable to self-driving cars, lead to the option becoming mandatory on all new vehicles. insurance companies, burned by litigation, offer premium rates to those who never switch off the driverless option, while increasing rates on drivers who elect to operate their cars manually. Soon the difference between these rates becomes enormous. [NEWLINE] 10-15 years: Commercial driving is entirely automated. Cabs, buses, trucks, trains, "driver" becomes an obsolete profession. The savings in both wages and liability is simply too tremendous to allow any non-automated fleet to remain competitive. [NEWLINE] 15-20 years: Studies conclusively show that the only traffic casualties that still occur are exclusively due to human operator error. It becomes evident that driving your own car is unthinkably dangerous, like drunk driving at night with no headlights or seatbelts. Safety laws are passed that effectively outlaw operating your own vehicle. [NEWLINE] [NEWLINE] By the time my nephew is 15-16, controlling a car will be something that only hobbyists do, and never on public roads.  Very few cars will be privately owned, rather they will be operated by private or municipal transportation services. [NEWLINE] The age of the personal automobile is ending. CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1)<mask> do standards of physical attractiveness relate with your arbitrary standard regarding peoples' past experiences? [ENDQ] [NEWLINE] They are all equally valid. For any reason and at any time you get to decide your level of attraction to any facet of another human being. [NEWLINE] [NEWLINE] [STARTQ] 2)<mask> do you have against women who enjoy a perfectly healthy activity? [ENDQ] [NEWLINE] They are free to do whatever,<mask> that doesn't mean I'd want to date them. My personal feeling against them, is that I don't enjoy the notion that 5+ other men have done<mask> I'm doing. I don't want to have to think of the dicks she's sucked on or the semen she's had to clean off her body from every other guy before me. The notion of a guy sucking on her body in the same places I do or intend to is wholly unappealing to me. [NEWLINE] [NEWLINE] Now,<mask> you can convince me that there is some unknown value there for me, I'd happily change my view.<mask> from<mask> I see<mask> (<mask> is the case for me and millions of others) those things are off-putting, then selecting from the pool of women available that don't evoke those thoughts is the right choice. Not just for me,<mask> probably for her<mask> well. She shouldn't have to be with someone that puts that much thought into her past and counts it against her. Socially, there should be no issue with us finding partners that suite our needs whatever they may be. My need happens to be a partner that can't fill a minivan with other men she's been with. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] 3)<mask><mask> your standard prevents you from finding someone who fits you perfectly<mask><mask><mask> looks, personality, values, etc.? [NEWLINE] Then she's gotten a solid 80% match rate, and I'm aiming for 90% and higher? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Worrying about<mask> many partners your spouse had is no longer a big problem<mask><mask><mask> STDs go. [ENDQ] [NEWLINE] Agreed. I'd take the same exact precautions with all women<mask> their purported number of sexual partners<mask>. Not going to trust that the one guy she slept with is clean, nor the 5 guys, nor that she is actually 100% telling the truth. High risk, no reward. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] **I'd suggest instead of limiting your options needlessly - which is only hurting yourself - instead learn to deal with it.** [ENDQ] [NEWLINE] This goes to the heart of my OP. I don't bring that level
Label encoding: <s> [STARTQ] 1) How do standards of physical attractiveness relate with your arbitrary standard regarding peoples' past experiences? [ENDQ] [NEWLINE] They are all equally valid. For any reason and at any time you get to decide your level of attraction to any facet of another human being. [NEWLINE] [NEWLINE] [STARTQ] 2) What do you have against women who enjoy a perfectly healthy activity? [ENDQ] [NEWLINE] They are free to do whatever, but that doesn't mean I'd want to date them. My personal feeling against them, is that I don't enjoy the notion that 5+ other men have done what I'm doing. I don't want to have to think of the dicks she's sucked on or the semen she's had to clean off her body from every other guy before me. The notion of a guy sucking on her body in the same places I do or intend to is wholly unappealing to me. [NEWLINE] [NEWLINE] Now, if you can convince me that there is some unknown value there for me, I'd happily change my view. But from what I see if ( as is the case for me and millions of others) those things are off-putting, then selecting from the pool of women available that don't evoke those thoughts is the right choice. Not just for me, but probably for her as well. She shouldn't have to be with someone that puts that much thought into her past and counts it against her. Socially, there should be no issue with us finding partners that suite our needs whatever they may be. My need happens to be a partner that can't fill a minivan with other men she's been with. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] 3) What if your standard prevents you from finding someone who fits you perfectly as far as looks, personality, values, etc.? [NEWLINE] Then she's gotten a solid 80% match rate, and I'm aiming for 90% and higher? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Worrying about how many partners your spouse had is no longer a big problem as far as STDs go. [ENDQ] [NEWLINE] Agreed. I'd take the same exact precautions with all women despite their purported number of sexual partners though. Not going to trust that the one guy she slept with is clean, nor the 5 guys, nor that she is actually 100% telling the truth. High risk, no reward. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] **I'd suggest instead of limiting your options needlessly - which is only hurting yourself - instead learn to deal with it.** [ENDQ] [NEWLINE] This goes to the heart of my OP. I don't bring that level
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ok, your wording had thrown me off. I don't think "intent" exists in any crime name except<mask> an addon, like assault with intent to kill. From here on I'll just talk conspiracy and attempted. [ENDQ] [NEWLINE] You may be outside the US.  In the US,<mask> you plot to kill someone and take reasonable steps on that journey to do<mask>, you've committed "intent to commit murder." [NEWLINE] [NEWLINE] <mask> you do it with other people you've committed "conspiracy to commit murder." [NEWLINE] [NEWLINE] <mask> you attempt it (<mask> fail) you've committed "attempted murder." [NEWLINE] [NEWLINE] And<mask> you do it, you've committed "murder." [NEWLINE] [NEWLINE] [STARTQ] I still have objections to your internal consistency on this point. Lets say I hire a killer, have a committed a crime now, or only<mask> the killer does the job? Now<mask><mask> I try to hire an undercover cop<mask> a killer? All he can do is tell me not to do it and inform the would be victim? [ENDQ] [NEWLINE] You've talked about killing someone and taken steps towards that goal,<mask> there is no victim<mask>. [NEWLINE] [NEWLINE] <mask> is this any different with current laws?  It's only a charge that would be added on later and<mask> there was an undercover cop, then you still have a way to stop the crime from happening. [NEWLINE] [NEWLINE] [STARTQ] <mask> does one test for this? Who designs the test? Who administers the test? [ENDQ] [NEWLINE] Same way any test is done, psychologists, courthouses. [NEWLINE] [NEWLINE] [STARTQ] <mask> long can they deny you this right? [ENDQ] [NEWLINE] Never.  You begin adulthood<mask> soon<mask> you are self sufficient to decide this is something you want to undertake. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] You state some 30 year olds should be protected.<mask> other age based rights can we deny? The ability to deny someone the right to consent to sex is at least a major obstacle to their right to reproduction. [ENDQ] [NEWLINE] Who said we're denying them the right?  We're just removing legal consent from the equation.  That 30 year old,<mask> they are incapable of taking and passing this test is no more able to consent than a 5 year old. [NEWLINE] [NEWLINE] <mask> we use age<mask> a metric of<mask> is okay. <mask> the actual ABILITY to consent is<mask> is important, NOT age. [NEWLINE] [NEWLINE] [STARTQ] <mask> about the right to vote? [ENDQ] [NEWLINE] Not sure I even really believe in voting.  I like the idea of a meritocracy\technocracy a lot more. 
Label encoding: <s> [STARTQ] Ok, your wording had thrown me off. I don't think "intent" exists in any crime name except as an addon, like assault with intent to kill. From here on I'll just talk conspiracy and attempted. [ENDQ] [NEWLINE] You may be outside the US.  In the US, if you plot to kill someone and take reasonable steps on that journey to do so, you've committed "intent to commit murder." [NEWLINE] [NEWLINE] If you do it with other people you've committed "conspiracy to commit murder." [NEWLINE] [NEWLINE] If you attempt it ( but fail) you've committed "attempted murder." [NEWLINE] [NEWLINE] And if you do it, you've committed "murder." [NEWLINE] [NEWLINE] [STARTQ] I still have objections to your internal consistency on this point. Lets say I hire a killer, have a committed a crime now, or only when the killer does the job? Now what if I try to hire an undercover cop as a killer? All he can do is tell me not to do it and inform the would be victim? [ENDQ] [NEWLINE] You've talked about killing someone and taken steps towards that goal, but there is no victim yet. [NEWLINE] [NEWLINE] How is this any different with current laws?  It's only a charge that would be added on later and if there was an undercover cop, then you still have a way to stop the crime from happening. [NEWLINE] [NEWLINE] [STARTQ] How does one test for this? Who designs the test? Who administers the test? [ENDQ] [NEWLINE] Same way any test is done, psychologists, courthouses. [NEWLINE] [NEWLINE] [STARTQ] How long can they deny you this right? [ENDQ] [NEWLINE] Never.  You begin adulthood as soon as you are self sufficient to decide this is something you want to undertake. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] You state some 30 year olds should be protected. What other age based rights can we deny? The ability to deny someone the right to consent to sex is at least a major obstacle to their right to reproduction. [ENDQ] [NEWLINE] Who said we're denying them the right?  We're just removing legal consent from the equation.  That 30 year old, if they are incapable of taking and passing this test is no more able to consent than a 5 year old. [NEWLINE] [NEWLINE] Yet we use age as a metric of what is okay.  When the actual ABILITY to consent is what is important, NOT age. [NEWLINE] [NEWLINE] [STARTQ] What about the right to vote? [ENDQ] [NEWLINE] Not sure I even really believe in voting.  I like the idea of a meritocracy\technocracy a lot more. 
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> it's impossible to prove that a fetus is not a person [ENDQ] [NEWLINE] It's easy to prove<mask> a fetus is or is not a person. The hard part is deciding<mask> a person *is,*<mask> the word person is a fuzzy concept, and trying to define that concept? Good luck - ask an nineteenth century plantation owner<mask> black people are the same species<mask> white people... [NEWLINE] **** [NEWLINE] Lets try a dictionary definition. "a human being regarded<mask> an individual." [NEWLINE] [NEWLINE] <mask> do you define human being? Saying they must be a child of a human is just circular. Physical taxonomy is problematic<mask> it comes to inclusivity - by the time you've ruled in all the possible variation, including those from deformity (whether by birth or accident) you've probably included a lot of apes, and quite possibly other mammals.<mask> about DNA? Well, that's our best bet,<mask> there's a lot of pitfalls there too. You can't just say take a DNA map and say those who share a 99.9%(it's around that level) of commonality with this are classified<mask> human. Those with downs syndrome have 47 chromasomes instead of 46,<mask> you have to account for those sorts of things, too. [NEWLINE] [NEWLINE] And that's the easy part. We still have to deal with "regarded<mask> an individual." Lets skip'regarded'<mask> it's a nightmare - regarded by whom? Etc. [NEWLINE] [NEWLINE] Okay,<mask> on to "Individual" - well, there's two meanings to individual - single or separate. [NEWLINE] [NEWLINE] <mask> you mean single? That's fine. a fertilized, single cell zygote with human DNA is a person.<mask><mask> about people with chimerism either by birth or due to organ transplant? They have two completely separate sets of DNA. By that measure, they are not individuals. [NEWLINE] [NEWLINE] <mask> you mean "separate"<mask>... a gestating fetus is definitely not separate; Until they're born, children are distinctly parasitic and quickly die<mask> separated from the host.<mask><mask>, this is frequently the delineation used in law. Bonus fact: the difference between a miscarriage and a stillbirth is whether it could have survived on it's own.<mask> you *really* push the parasite angle, most children are incapable of fending for themselves and could be argued not to be real people until somewhere between 3-5 years of age. [NEWLINE] [NEWLINE] And that's just by one definition. There are many others.
Label encoding: <s> [STARTQ] I think it's impossible to prove that a fetus is not a person [ENDQ] [NEWLINE] It's easy to prove if a fetus is or is not a person. The hard part is deciding what a person *is,* because the word person is a fuzzy concept, and trying to define that concept? Good luck - ask an nineteenth century plantation owner if black people are the same species as white people... [NEWLINE] **** [NEWLINE] Lets try a dictionary definition. "a human being regarded as an individual." [NEWLINE] [NEWLINE] How do you define human being? Saying they must be a child of a human is just circular. Physical taxonomy is problematic when it comes to inclusivity - by the time you've ruled in all the possible variation, including those from deformity (whether by birth or accident) you've probably included a lot of apes, and quite possibly other mammals. How about DNA? Well, that's our best bet, but there's a lot of pitfalls there too. You can't just say take a DNA map and say those who share a 99.9%(it's around that level) of commonality with this are classified as human. Those with downs syndrome have 47 chromasomes instead of 46, so you have to account for those sorts of things, too. [NEWLINE] [NEWLINE] And that's the easy part. We still have to deal with "regarded as an individual." Lets skip'regarded' because it's a nightmare - regarded by whom? Etc. [NEWLINE] [NEWLINE] Okay, so on to "Individual" - well, there's two meanings to individual - single or separate. [NEWLINE] [NEWLINE] If you mean single? That's fine. a fertilized, single cell zygote with human DNA is a person. But what about people with chimerism either by birth or due to organ transplant? They have two completely separate sets of DNA. By that measure, they are not individuals. [NEWLINE] [NEWLINE] If you mean "separate" though... a gestating fetus is definitely not separate; Until they're born, children are distinctly parasitic and quickly die if separated from the host. In fact, this is frequently the delineation used in law. Bonus fact: the difference between a miscarriage and a stillbirth is whether it could have survived on it's own. If you *really* push the parasite angle, most children are incapable of fending for themselves and could be argued not to be real people until somewhere between 3-5 years of age. [NEWLINE] [NEWLINE] And that's just by one definition. There are many others.
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Edit:<mask> I say *lie*, I mean falsehood, not real, meaningless, false promise, void, not-quite-well-thought-through, in that sense. I do not mean that founding fathers of countries are liars. [NEWLINE] [NEWLINE] **Setup** [NEWLINE] [NEWLINE] From my observations I define a right<mask> follows: A *right* is an option or a contract offered by a specific *benefactor* to a specific *beneficiary*, in which the benefactor incurs an obligation to carry out the *underlying promise*<mask> the right is excercised by the beneficiary (terms and conditions may apply). [NEWLINE] [NEWLINE] Example: A governement issues a citizen a right to vote in an election and<mask> sees to it that the citizen can vote should he decide to do<mask>. [NEWLINE] [NEWLINE] Some implications: [NEWLINE] [NEWLINE] - A right is owned by (in the sense of a binding obligation) the benefactor. [NEWLINE] [NEWLINE] -<mask> the benefactor is unable to carry out the underlying promise in the specified terms, the right is void and is for all practical purpose non-existent. [NEWLINE] [NEWLINE] - A right without a benefactor of a beneficiary is void. [NEWLINE] [NEWLINE] - A right issued by the benficiary is meaningless and such contract has different names (forcing, extortion). [NEWLINE] [NEWLINE] **Main argument** [NEWLINE] [NEWLINE] Human rights "commonly understood<mask> inalienable fundamental rights to which a person is inherently entitled simply<mask> she or he is a human being" (Wiki) are a lie and<mask> such are void. There is no benefactor issuing such rights. [NEWLINE] [NEWLINE] Inb4 society is the benefactor - society is in this case a fataly vague name. [NEWLINE] [NEWLINE] Inb4 a state/governement is the benefactor - this makes the right in question either a civil right or a meaningless right<mask> the state is impotent to cary it out. [NEWLINE] [NEWLINE] **Example** [NEWLINE] [NEWLINE] The best example of the non-sense I observe is the alleged *right to life/live*. It is issued by the beneficiary, it has no specific benefactor and no one within humanity even has the power to carry it out in the most pragmatic sense. [NEWLINE] [NEWLINE] I understand it<mask> one of three cases: [NEWLINE] [NEWLINE] Right to an unspecified lenghth of biological life, valid after a person is born. - *tautology* [NEWLINE] [NEWLINE] Right to be protected from dying<mask><mask><mask> possible. This is not only criticaly vague<mask><mask> not a right. - *linguistic vandalism* [NEWLINE] [NEWLINE]
Label encoding: <s>Edit: When I say *lie*, I mean falsehood, not real, meaningless, false promise, void, not-quite-well-thought-through, in that sense. I do not mean that founding fathers of countries are liars. [NEWLINE] [NEWLINE] **Setup** [NEWLINE] [NEWLINE] From my observations I define a right as follows: A *right* is an option or a contract offered by a specific *benefactor* to a specific *beneficiary*, in which the benefactor incurs an obligation to carry out the *underlying promise* if the right is excercised by the beneficiary (terms and conditions may apply). [NEWLINE] [NEWLINE] Example: A governement issues a citizen a right to vote in an election and thus sees to it that the citizen can vote should he decide to do so. [NEWLINE] [NEWLINE] Some implications: [NEWLINE] [NEWLINE] - A right is owned by (in the sense of a binding obligation) the benefactor. [NEWLINE] [NEWLINE] - If the benefactor is unable to carry out the underlying promise in the specified terms, the right is void and is for all practical purpose non-existent. [NEWLINE] [NEWLINE] - A right without a benefactor of a beneficiary is void. [NEWLINE] [NEWLINE] - A right issued by the benficiary is meaningless and such contract has different names (forcing, extortion). [NEWLINE] [NEWLINE] **Main argument** [NEWLINE] [NEWLINE] Human rights "commonly understood as inalienable fundamental rights to which a person is inherently entitled simply because she or he is a human being" (Wiki) are a lie and as such are void. There is no benefactor issuing such rights. [NEWLINE] [NEWLINE] Inb4 society is the benefactor - society is in this case a fataly vague name. [NEWLINE] [NEWLINE] Inb4 a state/governement is the benefactor - this makes the right in question either a civil right or a meaningless right if the state is impotent to cary it out. [NEWLINE] [NEWLINE] **Example** [NEWLINE] [NEWLINE] The best example of the non-sense I observe is the alleged *right to life/live*. It is issued by the beneficiary, it has no specific benefactor and no one within humanity even has the power to carry it out in the most pragmatic sense. [NEWLINE] [NEWLINE] I understand it as one of three cases: [NEWLINE] [NEWLINE] Right to an unspecified lenghth of biological life, valid after a person is born. - *tautology* [NEWLINE] [NEWLINE] Right to be protected from dying as long as possible. This is not only criticaly vague but also not a right. - *linguistic vandalism* [NEWLINE] [NEWLINE]
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] At a basic level, studios are closing<mask> not enough people are buying their games. [ENDQ] [NEWLINE] <mask> their games aren't offering anything innovative in terms of game play. Not<mask> they're distinctly or inherently sexist. Missing your target demographic with a bad game is not the same thing<mask> saying your target demographic is waning or disappearing<mask> of sexism. Figuring out<mask> sales are lost is a good way to implement things towards the effect of getting them back. [NEWLINE] [NEWLINE] [STARTQ] First, feminists ARE doing that, and every year more and more games are released from avowed feminist designers, pushing interesting ideas and expanding the scope of gaming. I can think of a dozen feminist games in 2013, compared to a half-dozen in 2012, and less than 3 in 2011. [ENDQ] [NEWLINE] I've<mask> to see or hear of one game that had left it's mark on the gaming community, that was a definitive AAA title, indie or otherwise.<mask> cite your source please.<mask><mask> I've heard entirely to the contrary, Anita Sarkeesian is taking the cult hit nature of Mirror's Edge and tampering with the control scheme to make it more accessible. That's great for money, and in this specific instance it is a control scheme,<mask> the key point here is "accessibility" which in gaming translates almost directly to "casual" which<mask><mask> is a determent to the hobby. I don't want my select super secret boys club<mask> many would advocate. I want my games to stay a quality I've come to expect<mask> someone who's spent most of his life playing games, and instead of everyone hopping ship to improve a demographic reach. [NEWLINE] [NEWLINE] [STARTQ] Finally, on a basic economic level, mobile isn't booming<mask> lots of people are spending $5; the bulk of successful mobile is F2P. Mobile is booming<mask> its most dedicated consumers (a disproportionate amount of whom are women) are willing to spend significantly MORE than $60 on a game (i.e., $20 a week on Candy Crush). The market is there, and it is very, very hungry. [ENDQ] [NEWLINE] They're very very hungry for casual games, which is almost a completely different demographic then anyone who's playing a AAA title.<mask> an example, my father loved playing an MMO with me growing up.<mask> the game fell into obscurity and we both stopped playing, he never picked up a game again. Now this is more of a representation of<mask> a casual mobile gamer is. Sure they'll spend their hundreds of dollars on that single game<mask> they don
Label encoding: <s> [STARTQ] At a basic level, studios are closing because not enough people are buying their games. [ENDQ] [NEWLINE] Because their games aren't offering anything innovative in terms of game play. Not because they're distinctly or inherently sexist. Missing your target demographic with a bad game is not the same thing as saying your target demographic is waning or disappearing because of sexism. Figuring out how sales are lost is a good way to implement things towards the effect of getting them back. [NEWLINE] [NEWLINE] [STARTQ] First, feminists ARE doing that, and every year more and more games are released from avowed feminist designers, pushing interesting ideas and expanding the scope of gaming. I can think of a dozen feminist games in 2013, compared to a half-dozen in 2012, and less than 3 in 2011. [ENDQ] [NEWLINE] I've yet to see or hear of one game that had left it's mark on the gaming community, that was a definitive AAA title, indie or otherwise. So cite your source please. In fact I've heard entirely to the contrary, Anita Sarkeesian is taking the cult hit nature of Mirror's Edge and tampering with the control scheme to make it more accessible. That's great for money, and in this specific instance it is a control scheme, but the key point here is "accessibility" which in gaming translates almost directly to "casual" which IMO is a determent to the hobby. I don't want my select super secret boys club as many would advocate. I want my games to stay a quality I've come to expect as someone who's spent most of his life playing games, and instead of everyone hopping ship to improve a demographic reach. [NEWLINE] [NEWLINE] [STARTQ] Finally, on a basic economic level, mobile isn't booming because lots of people are spending $5; the bulk of successful mobile is F2P. Mobile is booming because its most dedicated consumers (a disproportionate amount of whom are women) are willing to spend significantly MORE than $60 on a game (i.e., $20 a week on Candy Crush). The market is there, and it is very, very hungry. [ENDQ] [NEWLINE] They're very very hungry for casual games, which is almost a completely different demographic then anyone who's playing a AAA title. As an example, my father loved playing an MMO with me growing up. When the game fell into obscurity and we both stopped playing, he never picked up a game again. Now this is more of a representation of what a casual mobile gamer is. Sure they'll spend their hundreds of dollars on that single game but they don
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are numerous good reasons for the existence of interest. [NEWLINE] [NEWLINE] 1) To counter inflation.  A positive rate of inflation means that money loses value over time.  The idea of a loan is that the lender forfeits the ability to use his money today<mask> that the borrower can use it.  Interest rates mean that at the bare minimum,<mask> I decide to forgo $100 of goods today that I can still buy that real value of $100 worth of goods next year. [NEWLINE] [NEWLINE] 2) Risk premium.  This accounts for<mask> certain people get different rates on things like car loans or mortgages.  It is very rare for ordinary folks to buy a home with cash, which is<mask> the mortgage exists. <mask> bankruptcy is a thing, for each loan someone gives out, they need to be aware that some proportion of those loans will not be fully repaid.  This means that the other borrowers shoulder some of that cost the same way that a certain small fraction of markup in supermarkets comes from the need to cover shoplifted goods. [NEWLINE] [NEWLINE] 3) Bonds are vital securities.  Things like savings bonds and t-bills are vital to the fiscal health of the country.  The government effectively borrows money through these vehicles.  The interest rate on a bond is inversely related to its price,<mask><mask> you buy a bond, it has a price and a face value.  You pay cash for the price of the bond and<mask> it matures, the government buys it back for the face value.  These bond interest rates are directly linked with other rates, like car loans or mortgages. [NEWLINE] [NEWLINE] 4) Supply and demand.  The money market works just like any other market.  The "quantity" axis is the amount of money in the system, and the "price" axis is the interest rate.  There is a downward sloping demand curve representing the demand for money and a vertical supply curve representing the money supply.  The intersection of these curves represents the equilibrium interest rate. <mask> interest rates are too low, we see that correspond with a demand for a greater amount of money than exists in the system.  Over in the bonds market, we notice that the return on a bond is inversely proportional to its price.  The increased demand for money corresponds with a decreased demand for bonds,<mask> the price of bonds will fall.  This decrease in bond price corresponds to an increase in their interest rates, which corrects the disequilibrium in the money market.  In consumer terms
Label encoding: <s>There are numerous good reasons for the existence of interest. [NEWLINE] [NEWLINE] 1) To counter inflation.  A positive rate of inflation means that money loses value over time.  The idea of a loan is that the lender forfeits the ability to use his money today so that the borrower can use it.  Interest rates mean that at the bare minimum, if I decide to forgo $100 of goods today that I can still buy that real value of $100 worth of goods next year. [NEWLINE] [NEWLINE] 2) Risk premium.  This accounts for why certain people get different rates on things like car loans or mortgages.  It is very rare for ordinary folks to buy a home with cash, which is why the mortgage exists.  Because bankruptcy is a thing, for each loan someone gives out, they need to be aware that some proportion of those loans will not be fully repaid.  This means that the other borrowers shoulder some of that cost the same way that a certain small fraction of markup in supermarkets comes from the need to cover shoplifted goods. [NEWLINE] [NEWLINE] 3) Bonds are vital securities.  Things like savings bonds and t-bills are vital to the fiscal health of the country.  The government effectively borrows money through these vehicles.  The interest rate on a bond is inversely related to its price, because when you buy a bond, it has a price and a face value.  You pay cash for the price of the bond and when it matures, the government buys it back for the face value.  These bond interest rates are directly linked with other rates, like car loans or mortgages. [NEWLINE] [NEWLINE] 4) Supply and demand.  The money market works just like any other market.  The "quantity" axis is the amount of money in the system, and the "price" axis is the interest rate.  There is a downward sloping demand curve representing the demand for money and a vertical supply curve representing the money supply.  The intersection of these curves represents the equilibrium interest rate.  If interest rates are too low, we see that correspond with a demand for a greater amount of money than exists in the system.  Over in the bonds market, we notice that the return on a bond is inversely proportional to its price.  The increased demand for money corresponds with a decreased demand for bonds, so the price of bonds will fall.  This decrease in bond price corresponds to an increase in their interest rates, which corrects the disequilibrium in the money market.  In consumer terms
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Completely untrue. I'll repost a comment I made further down the thread. BTW,<mask> you read the post, you'll see that I went out of my way to make sure that my post didn't contain anything that could be offensive, such<mask> ablelist words like "crazy" [NEWLINE] [NEWLINE] I posted a thread in SRSMen a<mask> back asking for help for a friend of mine who was straying into misogynistic attitudes due to severe abuse by several women over the course of his life. (Abusive mother, abusive ex, second ex who cheated on him). I reposted it to /r/feminism and got a much more supportive and useful response [link]( [URL] /). The posts are exactly the same btw. [NEWLINE] [NEWLINE] Want to see<mask> /r/SRSMen did? They banned me, and comments like this got upvoted [NEWLINE] [NEWLINE] [STARTQ] Two relationships don't go the way he wanted and he turns into a misogynistic shitlord MRA? Sounds like he was an asshole to begin with. Maybe he was the problem in those relationships to start with. And honestly, you don't hold it against him for being an MRA after "<mask> he's been through"? Are you serious?<mask> the fuck? [ENDQ] [NEWLINE] <mask><mask><mask> SRS, going through horrible abuse (including a mother) is his fault! Gee whiz.<mask> a great community [NEWLINE] [NEWLINE] and then the mod reply [NEWLINE] [NEWLINE] [STARTQ] Seriously? You get banned from here for shitposting and<mask> you make a new account to continue your misogyny apologetics instead of doing<mask> the mods told you to do in order to get your account unbanned? [ENDQ] [NEWLINE] [STARTQ] Your posts and comments are just absolutely DRIPPING with a lack of recognition for your (and your friend's) toxic male privilege.<mask> lovely that you feel<mask> charitable towards him for hating on women! He's one of your good old boys, right? You KNOW he's a good chap inside<mask><mask> he spews poison against women.<mask> could you possibly abandon him<mask> all he's done is hate women and become a loyal patriarchy soldier? Not like he's caused you any personal harm or offence... not like you can be expected to put yourself in women's shoes and feel even for just a minute the bone-deep horror of<mask> another man who thinks we're less than shit. [ENDQ] [NEWLINE] [STARTQ] No, of course not. He's your friend. Ho ho, let us retire to the library for some cigars and brandy,
Label encoding: <s>Completely untrue. I'll repost a comment I made further down the thread. BTW, if you read the post, you'll see that I went out of my way to make sure that my post didn't contain anything that could be offensive, such as ablelist words like "crazy" [NEWLINE] [NEWLINE] I posted a thread in SRSMen a while back asking for help for a friend of mine who was straying into misogynistic attitudes due to severe abuse by several women over the course of his life. (Abusive mother, abusive ex, second ex who cheated on him). I reposted it to /r/feminism and got a much more supportive and useful response [link]( [URL] /). The posts are exactly the same btw. [NEWLINE] [NEWLINE] Want to see what /r/SRSMen did? They banned me, and comments like this got upvoted [NEWLINE] [NEWLINE] [STARTQ] Two relationships don't go the way he wanted and he turns into a misogynistic shitlord MRA? Sounds like he was an asshole to begin with. Maybe he was the problem in those relationships to start with. And honestly, you don't hold it against him for being an MRA after " what he's been through"? Are you serious? What the fuck? [ENDQ] [NEWLINE] So according to SRS, going through horrible abuse (including a mother) is his fault! Gee whiz. What a great community [NEWLINE] [NEWLINE] and then the mod reply [NEWLINE] [NEWLINE] [STARTQ] Seriously? You get banned from here for shitposting and so you make a new account to continue your misogyny apologetics instead of doing what the mods told you to do in order to get your account unbanned? [ENDQ] [NEWLINE] [STARTQ] Your posts and comments are just absolutely DRIPPING with a lack of recognition for your (and your friend's) toxic male privilege. How lovely that you feel so charitable towards him for hating on women! He's one of your good old boys, right? You KNOW he's a good chap inside even though he spews poison against women. How could you possibly abandon him when all he's done is hate women and become a loyal patriarchy soldier? Not like he's caused you any personal harm or offence... not like you can be expected to put yourself in women's shoes and feel even for just a minute the bone-deep horror of yet another man who thinks we're less than shit. [ENDQ] [NEWLINE] [STARTQ] No, of course not. He's your friend. Ho ho, let us retire to the library for some cigars and brandy,
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Although you personally do not like tipping, the custom exists for good reason.<mask> people did not tip and servers had to earn their living from only wages, food prices would increase to compensate the additional cost required to sustain the servers. <mask><mask> to this, not tipping is typically frowned upon<mask> the stiff is disregarding the server's humanity. [NEWLINE] [NEWLINE] Eliminating the tip would not save money for anyone in the long run<mask> restaurants will simply up the price of their food. Imagine a scenario in which one had to pay $10 + at least 20% tips,<mask> raising the price to $12.<mask> tipping ceased, the restaurant would simply raise the initial price of the food to $12. The consumer still has to pay the same amount.<mask>, under the tip system, servers can try to provide better service<mask> they have an incentive for doing<mask>. Tipping helps both the consumer and the server, and<mask> such, it is more preferrable to tip than provide a base wage. [NEWLINE] [NEWLINE] The server is paid a base wage of +/- $2/hour to ensure that they walk home with at least something. It's a protection of their income stability, much like<mask> many salespeople of supermarkets like Walmart are paid much lower than<mask> the skill demands<mask> salespeople make commission. The base wage for a server is not meant to sustain the server. It is only meant to be used<mask> a backup<mask> income from tips is too low. Servers actually earn their money from providing and enhancing a good service by taking away plates and bringing food to the customer. Keep in mind that<mask> consumers pay the price tag for the meal, they are paying only for the cook's labor and the cost of the ingredients. Tips are not factored in<mask> they provide an incentive for servers to give giving a better service.<mask> you may<mask><mask> cooks should be tipped too under this logic<mask> then cooks have an incentive to cook better, the action is simply unrealistic to implement<mask> a cook spends most of his or her time optimizing efficiency in the kitchen. To have to deal with tips is to immediately handicap the maximum achievement the cooks can accomplish. [NEWLINE] [NEWLINE] People who serve are doing<mask> to sustain their families. They try to provide the best service<mask> they need something from you: money. A person who does not tip goves everyone else around him or her that he or she is rich,<mask> chooses to spend cheaply anyway. It's a rude gesture, especially<mask> someone else (the server) is expending a reasonable effort
Label encoding: <s>Although you personally do not like tipping, the custom exists for good reason. If people did not tip and servers had to earn their living from only wages, food prices would increase to compensate the additional cost required to sustain the servers.  In addition to this, not tipping is typically frowned upon because the stiff is disregarding the server's humanity. [NEWLINE] [NEWLINE] Eliminating the tip would not save money for anyone in the long run because restaurants will simply up the price of their food. Imagine a scenario in which one had to pay $10 + at least 20% tips, thus raising the price to $12. If tipping ceased, the restaurant would simply raise the initial price of the food to $12. The consumer still has to pay the same amount. However, under the tip system, servers can try to provide better service because they have an incentive for doing so. Tipping helps both the consumer and the server, and as such, it is more preferrable to tip than provide a base wage. [NEWLINE] [NEWLINE] The server is paid a base wage of +/- $2/hour to ensure that they walk home with at least something. It's a protection of their income stability, much like how many salespeople of supermarkets like Walmart are paid much lower than what the skill demands because salespeople make commission. The base wage for a server is not meant to sustain the server. It is only meant to be used as a backup if income from tips is too low. Servers actually earn their money from providing and enhancing a good service by taking away plates and bringing food to the customer. Keep in mind that when consumers pay the price tag for the meal, they are paying only for the cook's labor and the cost of the ingredients. Tips are not factored in because they provide an incentive for servers to give giving a better service. While you may argue that cooks should be tipped too under this logic because then cooks have an incentive to cook better, the action is simply unrealistic to implement because a cook spends most of his or her time optimizing efficiency in the kitchen. To have to deal with tips is to immediately handicap the maximum achievement the cooks can accomplish. [NEWLINE] [NEWLINE] People who serve are doing so to sustain their families. They try to provide the best service because they need something from you: money. A person who does not tip goves everyone else around him or her that he or she is rich, but chooses to spend cheaply anyway. It's a rude gesture, especially when someone else (the server) is expending a reasonable effort
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, I'd like to point out that "modern art" is a specific period of art from ~1860 to ~1970. You're referring to contemporary abstract expressionist art. [NEWLINE] [NEWLINE] The formal definition of art is that it contains at least one of the [five elements of art]( [URL] ) [NEWLINE] [NEWLINE] In reference to the blank canvas, you have to look at it<mask> an installation; it's not a painting of nothing, it's more a sculpture involving a white canvas.<mask><mask>, the paragraph accompanying it is<mask> part of this installation. Once you've accepted that, it's easy to see<mask> the blank canvas meets the first definition. It utilizes texture (the canvas, being an intentional part of the installation, is a deliberate use of texture), space, shape (namely squares), and value. [NEWLINE] [NEWLINE] You can take those criteria and apply them to any work. For example, Jackson Pollock's works, which often come under fire, undoubtedly contain several elements of art.<mask> do Mark Rothko's works.<mask><mask> many people (including myself) dislike this genre of art (personally I find it to be low effort), it's not like you can say that it doesn't fit the definition of art. [NEWLINE] [NEWLINE] Now<mask> for *all* of this style of art being terrible,<mask><mask> that's a very close-minded generalization.<mask> you're making a generalization about *all* of this art, I'll argue to the specific case of the blank canvas. [NEWLINE] [NEWLINE] The fact that we can even have a conversation about whether or not the blank canvas is art means it's doing its job. It is meant to challenge your perception of<mask> art is. A similar work would be [The Treachery of Images]( [URL].jpg) by René Magritte. ("Ceci n'est pas une pipe" means "This is not a pipe.") It presents a realistic portrayal of a pipe,<mask> reminds you that it's *not* a pipe, it's just a *painting* of a pipe.<mask><mask> is it that people consider *Treachery* to be art<mask> the white canvas not to be art? Is it simply<mask> the artist didn't paint anything?<mask> that's basically discounting a piece of art<mask> it has less effort involved, which has proven time and time again to be a very poor way of thinking about things.<mask> you need proof, just look at Van Gogh. [NEWLINE] [NEWLINE] Once you accept that it
Label encoding: <s>First of all, I'd like to point out that "modern art" is a specific period of art from ~1860 to ~1970. You're referring to contemporary abstract expressionist art. [NEWLINE] [NEWLINE] The formal definition of art is that it contains at least one of the [five elements of art]( [URL] ) [NEWLINE] [NEWLINE] In reference to the blank canvas, you have to look at it as an installation; it's not a painting of nothing, it's more a sculpture involving a white canvas. In fact, the paragraph accompanying it is also part of this installation. Once you've accepted that, it's easy to see how the blank canvas meets the first definition. It utilizes texture (the canvas, being an intentional part of the installation, is a deliberate use of texture), space, shape (namely squares), and value. [NEWLINE] [NEWLINE] You can take those criteria and apply them to any work. For example, Jackson Pollock's works, which often come under fire, undoubtedly contain several elements of art. So do Mark Rothko's works. Even though many people (including myself) dislike this genre of art (personally I find it to be low effort), it's not like you can say that it doesn't fit the definition of art. [NEWLINE] [NEWLINE] Now as for *all* of this style of art being terrible, I think that's a very close-minded generalization. Since you're making a generalization about *all* of this art, I'll argue to the specific case of the blank canvas. [NEWLINE] [NEWLINE] The fact that we can even have a conversation about whether or not the blank canvas is art means it's doing its job. It is meant to challenge your perception of what art is. A similar work would be [The Treachery of Images]( [URL].jpg) by René Magritte. ("Ceci n'est pas une pipe" means "This is not a pipe.") It presents a realistic portrayal of a pipe, but reminds you that it's *not* a pipe, it's just a *painting* of a pipe. So why is it that people consider *Treachery* to be art but the white canvas not to be art? Is it simply because the artist didn't paint anything? Because that's basically discounting a piece of art because it has less effort involved, which has proven time and time again to be a very poor way of thinking about things. If you need proof, just look at Van Gogh. [NEWLINE] [NEWLINE] Once you accept that it
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] * Most discoveries (especially in physics) are made by a team. Agreed. It is difficult to implement and enforce. [NEWLINE] * Apple’s “list scrolling and document translation, scaling, and rotation on a touchscreen display” [patent]( [URL] ;Sect2=HITOFF&amp;d=PALL&amp;p=1&amp;u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&amp;r=1&amp;f=G&amp;l=50&amp;s1=7,469,381.PN.&amp;OS=PN/7,469,381&amp;RS=PN/7,469,381) for example, is not an actual physical thing or invention, or a concoction. [NEWLINE] * Biological patents don't patent ideas. They patent molecules. [NEWLINE] I just wanted to point out that patenting in science is not unheard of. [NEWLINE] * Most science is government funded. Agreed,<mask><mask> the government owns the patent, doesn't it mean more money for the government to invest on further research? [NEWLINE] * Patenting ideas would be horribly unenforceable. Tough,<mask> not impossible. [NEWLINE] * Paid journals would hate it. I don't think scientists will have a problem with free journals circulating their work. [NEWLINE] * Scientists don't actually want discoveries to be patented. Most people like to get credit and recognition for their work, even<mask> they might not want to make money off of it. [NEWLINE] * Scientists don't really want the money. This is a ridiculous generalization. Money is one of the factors for any job. [NEWLINE] * Most discoveries aren't 'thinking the way no one has thought before', they're 'applying well known methods to something that no one has gotten around to doing this on<mask> '. Similarly most inventions are not made from scratch, [NEWLINE] for example : [NEWLINE] [STARTQ] Someone built the first MP3 player, sure...<mask> really, all they did was combine flash someone else built with a DAC someone else built, and an audio compression format someone else invented... Precious little is created without leaning heavily on those who came before you. (lifted from [this]( [URL] )). [ENDQ] [NEWLINE] * This would destroy thousands of jobs. It would create a lot more. Imagine the number of lawyers needed. Not necessarily a good thing i suppose,<mask> Science will be much more lucrative and appealing<mask> there was
Label encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] * Most discoveries (especially in physics) are made by a team. Agreed. It is difficult to implement and enforce. [NEWLINE] * Apple’s “list scrolling and document translation, scaling, and rotation on a touchscreen display” [patent]( [URL] ;Sect2=HITOFF&amp;d=PALL&amp;p=1&amp;u=%2Fnetahtml%2FPTO%2Fsrchnum.htm&amp;r=1&amp;f=G&amp;l=50&amp;s1=7,469,381.PN.&amp;OS=PN/7,469,381&amp;RS=PN/7,469,381) for example, is not an actual physical thing or invention, or a concoction. [NEWLINE] * Biological patents don't patent ideas. They patent molecules. [NEWLINE] I just wanted to point out that patenting in science is not unheard of. [NEWLINE] * Most science is government funded. Agreed, but if the government owns the patent, doesn't it mean more money for the government to invest on further research? [NEWLINE] * Patenting ideas would be horribly unenforceable. Tough, but not impossible. [NEWLINE] * Paid journals would hate it. I don't think scientists will have a problem with free journals circulating their work. [NEWLINE] * Scientists don't actually want discoveries to be patented. Most people like to get credit and recognition for their work, even if they might not want to make money off of it. [NEWLINE] * Scientists don't really want the money. This is a ridiculous generalization. Money is one of the factors for any job. [NEWLINE] * Most discoveries aren't 'thinking the way no one has thought before', they're 'applying well known methods to something that no one has gotten around to doing this on yet '. Similarly most inventions are not made from scratch, [NEWLINE] for example : [NEWLINE] [STARTQ] Someone built the first MP3 player, sure... but really, all they did was combine flash someone else built with a DAC someone else built, and an audio compression format someone else invented... Precious little is created without leaning heavily on those who came before you. (lifted from [this]( [URL] )). [ENDQ] [NEWLINE] * This would destroy thousands of jobs. It would create a lot more. Imagine the number of lawyers needed. Not necessarily a good thing i suppose, but Science will be much more lucrative and appealing if there was
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You know<mask>, I mostly agree with you. <mask> can one person do to change this vicious cycle?  My life feels like an individual ant being forced to go with the flow of the colony.  The major question for me is<mask> the direction of the colony (human nature) is truly beyond hope or<mask> it could maybe change for the better.  Greed is ingrained in human nature it's hard to ever imagine that changing.  I too fear humans will someday go extinct<mask><mask><mask> of our own action,<mask> maybe not... [NEWLINE] [NEWLINE] In the past 5-10 years it's become increasingly popular to "go green".  I've actually been surprised to see<mask> much mainstream it has become to care about the environment.  Nonetheless, that greed still exists and corporations are doing everything they can to sustain their business and ensure things like oil don't become irrelevant.  The ability for a few to control the masses is extremely depressing to me.  Perhaps the problem is beyond a few individuals and it is our will to resist change that is the root cause. [NEWLINE] [NEWLINE] I, too, eat<mask> much meat<mask> I want and don't think twice about taking road trips.  My personal view is that people are going to do<mask> they want to do.  That said, instead of hoping to change human nature, perhaps we need to work on better alternatives. <mask> humans are in a pinch, we are blessed with ingenuity.  It's sad to think we may have to wait until the oil reserves dry up before changing,<mask> I have a strong belief we are capable of creating better tools. <mask><mask><mask>, I don't think humans will easily go extinct. <mask><mask> we're extremely crafty and will find a way to survive.  It would be great to see the negative aspects of our nature disappear,<mask> they probably won't.  That doesn't mean all is lost<mask><mask> don't give up on us! [NEWLINE] [NEWLINE] I don't think you should hold any guilt for doing<mask> most everyone else does too.  The important part is you actually recognize the issue and think twice about it.  I can relate a ton to this attitude,<mask> that doesn't mean we should give up.  It feels hopeless now,<mask> that's only<mask> it's all or nothing.  Our choices at the time being are don't eat meat or don't drive.  Those don't seem very practical now do they?  I'm confident that<mask> enough people are able to *recognize* the
Label encoding: <s>You know what, I mostly agree with you.  What can one person do to change this vicious cycle?  My life feels like an individual ant being forced to go with the flow of the colony.  The major question for me is if the direction of the colony (human nature) is truly beyond hope or if it could maybe change for the better.  Greed is ingrained in human nature it's hard to ever imagine that changing.  I too fear humans will someday go extinct as a result of our own action, but maybe not... [NEWLINE] [NEWLINE] In the past 5-10 years it's become increasingly popular to "go green".  I've actually been surprised to see how much mainstream it has become to care about the environment.  Nonetheless, that greed still exists and corporations are doing everything they can to sustain their business and ensure things like oil don't become irrelevant.  The ability for a few to control the masses is extremely depressing to me.  Perhaps the problem is beyond a few individuals and it is our will to resist change that is the root cause. [NEWLINE] [NEWLINE] I, too, eat as much meat as I want and don't think twice about taking road trips.  My personal view is that people are going to do what they want to do.  That said, instead of hoping to change human nature, perhaps we need to work on better alternatives.  When humans are in a pinch, we are blessed with ingenuity.  It's sad to think we may have to wait until the oil reserves dry up before changing, but I have a strong belief we are capable of creating better tools.  Because of this, I don't think humans will easily go extinct.  I think we're extremely crafty and will find a way to survive.  It would be great to see the negative aspects of our nature disappear, but they probably won't.  That doesn't mean all is lost though so don't give up on us! [NEWLINE] [NEWLINE] I don't think you should hold any guilt for doing what most everyone else does too.  The important part is you actually recognize the issue and think twice about it.  I can relate a ton to this attitude, but that doesn't mean we should give up.  It feels hopeless now, but that's only because it's all or nothing.  Our choices at the time being are don't eat meat or don't drive.  Those don't seem very practical now do they?  I'm confident that if enough people are able to *recognize* the
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In one form or another, everyone is benefiting from the government. Rich people who receive tax breaks are, in essence, getting money from the government. Last year they had to pay $2M in taxes, this year they have to pay $1.5M. [NEWLINE] [NEWLINE] <mask>, should rich people not be allowed to vote? [NEWLINE] [NEWLINE] The government offers a tax incentive to a group of businesses to operate in a particular state.<mask><mask><mask> of this tax incentive, and the subsequent opening of these businesses, 200,000 jobs are created. Everyone who was hired<mask><mask><mask> of that tax incentive owes their job to the government. [NEWLINE] [NEWLINE] Should those people not be allowed to vote? [NEWLINE] [NEWLINE] Military personnel are paid directly by the government. They are LITERALLY taking money from the government.<mask> would someone who is in the military ever vote for someone who is in favor of reducing the size of the military? [NEWLINE] [NEWLINE] Should soldiers be forbidden to vote? [NEWLINE] [NEWLINE] Consider the possibility of removing the vote from people who are on welfare. Without a vote, lots of candidates are elected who don't support our social safety net. It's then eliminated and welfare is eliminated. Now, everyone who was on welfare is no longer on welfare. Sure, they're dirt poor and probably living on the street in droves and droves, creating chaos, crime and probably rampant disease,<mask> there they are. [NEWLINE] [NEWLINE] <mask> they are no longer on welfare.<mask> now they get to vote! [NEWLINE] [NEWLINE] Next election cycle, a whole new group of officials come in, see the terrible problem with the poor, and vote to reinstate the social safety net. This puts a bunch of people on welfare. Who are now not allowed to vote. [NEWLINE] [NEWLINE] Second verse, same<mask> the first. [NEWLINE] [NEWLINE] Voting is a right granted by virtue of citizenship. Trying to dole out votes to people who are "worthy" of voting has a long history of abuse and is not something that should ever be considered. Say nothing of the fact that voting is complicated and there's many things to consider<mask> choosing a candidate. Probably no one on welfare would consider voting for a candidate who wants to repeal welfare. Just like no one who owns a gun would vote for someone who wants to ban guns and no one who's gay would vote for someone who is openly homophobic. [NEWLINE] [NEWLINE] That's<mask> voting works. Just<mask> you are benefiting from the government in some form does not mean that you are incapable of making a rational decision. And there is literally
Label encoding: <s>In one form or another, everyone is benefiting from the government. Rich people who receive tax breaks are, in essence, getting money from the government. Last year they had to pay $2M in taxes, this year they have to pay $1.5M. [NEWLINE] [NEWLINE] So, should rich people not be allowed to vote? [NEWLINE] [NEWLINE] The government offers a tax incentive to a group of businesses to operate in a particular state. As a result of this tax incentive, and the subsequent opening of these businesses, 200,000 jobs are created. Everyone who was hired as a result of that tax incentive owes their job to the government. [NEWLINE] [NEWLINE] Should those people not be allowed to vote? [NEWLINE] [NEWLINE] Military personnel are paid directly by the government. They are LITERALLY taking money from the government. Why would someone who is in the military ever vote for someone who is in favor of reducing the size of the military? [NEWLINE] [NEWLINE] Should soldiers be forbidden to vote? [NEWLINE] [NEWLINE] Consider the possibility of removing the vote from people who are on welfare. Without a vote, lots of candidates are elected who don't support our social safety net. It's then eliminated and welfare is eliminated. Now, everyone who was on welfare is no longer on welfare. Sure, they're dirt poor and probably living on the street in droves and droves, creating chaos, crime and probably rampant disease, but there they are. [NEWLINE] [NEWLINE] But they are no longer on welfare. So now they get to vote! [NEWLINE] [NEWLINE] Next election cycle, a whole new group of officials come in, see the terrible problem with the poor, and vote to reinstate the social safety net. This puts a bunch of people on welfare. Who are now not allowed to vote. [NEWLINE] [NEWLINE] Second verse, same as the first. [NEWLINE] [NEWLINE] Voting is a right granted by virtue of citizenship. Trying to dole out votes to people who are "worthy" of voting has a long history of abuse and is not something that should ever be considered. Say nothing of the fact that voting is complicated and there's many things to consider when choosing a candidate. Probably no one on welfare would consider voting for a candidate who wants to repeal welfare. Just like no one who owns a gun would vote for someone who wants to ban guns and no one who's gay would vote for someone who is openly homophobic. [NEWLINE] [NEWLINE] That's how voting works. Just because you are benefiting from the government in some form does not mean that you are incapable of making a rational decision. And there is literally
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From your OP: [NEWLINE] [NEWLINE] [STARTQ] Price matching is evidence that the store (Store A) could realistically be charging you less and still profit<mask> the competing store (Store B) can do it, and this Store A can too<mask> you call them out on it. This leads me to believe that Store A has no issues with gouging me for<mask> much<mask> it can get<mask><mask><mask> I am not the wiser. CMV, and tell me<mask> should I support that? [ENDQ] [NEWLINE] Some more food for thought.  There are many aspects that influence a retail store's pricing.  Seasonal items, cashflow, inventory/storage space, new products, discontinued products, outdated technology, all of which can influence<mask> much a company charges for a product, not just "<mask> much they can afford to charge." [NEWLINE] [NEWLINE] A company buys products wholesale from the manufacturer, then sells it at a markup to the consumer.  Still, it's impossible to guess precisely<mask> many items you will sell of each product.  Most american retailers tend to overstock items<mask> they don't run out, and liquidate them to clear shelf and warehouse space. <mask><mask> they may sell these last items at a net loss, they figure it's better to slightly overstock than it is to sell out.  Whether this is to maximize profit or a marketing/image strategy, I'm not sure.  The point is that you can't definitively say that they could have sold their entire stock at the lowest advertized price.  For example,<mask> they bought 1000 TVs at $100 each, sold 900 at $150, then sold the last 100 at $50 to get them out of the warehouse and off the shelfs to make room for the next year's model.  In total, on their $10,000 investment, they grossed $14,000.  That's their bottom line.  Take out the overhead, labor, buildings, maintenance, energy, etc. and you're left with a pretty thin profit margin. [NEWLINE] [NEWLINE] The price matching scheme makes sense in this model<mask> that retailer wants to move that product, and it's a way of assuring the customer that it was a good decision.  There will be the occasional sale across town, or even from the same store,<mask> the retailer will have to discount lots of customers' money back and will make zero money or even take a loss on those items sold 30 days prior. <mask> just like with the TVs, they're looking at the big picture of
Label encoding: <s>From your OP: [NEWLINE] [NEWLINE] [STARTQ] Price matching is evidence that the store (Store A) could realistically be charging you less and still profit because the competing store (Store B) can do it, and this Store A can too if you call them out on it. This leads me to believe that Store A has no issues with gouging me for as much as it can get as long as I am not the wiser. CMV, and tell me why should I support that? [ENDQ] [NEWLINE] Some more food for thought.  There are many aspects that influence a retail store's pricing.  Seasonal items, cashflow, inventory/storage space, new products, discontinued products, outdated technology, all of which can influence how much a company charges for a product, not just " how much they can afford to charge." [NEWLINE] [NEWLINE] A company buys products wholesale from the manufacturer, then sells it at a markup to the consumer.  Still, it's impossible to guess precisely how many items you will sell of each product.  Most american retailers tend to overstock items so they don't run out, and liquidate them to clear shelf and warehouse space.  Even though they may sell these last items at a net loss, they figure it's better to slightly overstock than it is to sell out.  Whether this is to maximize profit or a marketing/image strategy, I'm not sure.  The point is that you can't definitively say that they could have sold their entire stock at the lowest advertized price.  For example, If they bought 1000 TVs at $100 each, sold 900 at $150, then sold the last 100 at $50 to get them out of the warehouse and off the shelfs to make room for the next year's model.  In total, on their $10,000 investment, they grossed $14,000.  That's their bottom line.  Take out the overhead, labor, buildings, maintenance, energy, etc. and you're left with a pretty thin profit margin. [NEWLINE] [NEWLINE] The price matching scheme makes sense in this model because that retailer wants to move that product, and it's a way of assuring the customer that it was a good decision.  There will be the occasional sale across town, or even from the same store, where the retailer will have to discount lots of customers' money back and will make zero money or even take a loss on those items sold 30 days prior.  But just like with the TVs, they're looking at the big picture of
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0736, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's exactly it. Neither of your sources control for other factors. The most cited sources of these numbers, including the sources of your articles are the Consad Report for the US Department of labor, and an independent report done by the American Association of University Women (which blatantly conducted the study with an agenda in mind,<mask> I'm just gonna assume they didn't let their bias affect the results or the conclusion they drew). [NEWLINE] [NEWLINE] After adjusting for occupation, the CONSAD report found an unexplained gap between 4.8% and 7.1% (let's say 7 cents on the dollar), and AAUW found a gap of 6.6 cents on the dollar unexplained by occupational differences. [NEWLINE] [NEWLINE] I'm just gonna examine the AAUW report and<mask> they account for the occupational differences. In the report itself, the categories they use to classify occupation are pretty broad. In particular, there is an "Other Occupation" category and an "Other White Collar Occupation" category. [NEWLINE] [NEWLINE] From the report itself: [NEWLINE] [STARTQ] 1 [ENDQ] The category “other white-collar occupations” includes social scientists and related workers (except clinical, counseling, and school psychologists); lawyers, judges, and related workers; education, training, and library [NEWLINE] occupations (except primary, secondary, and special education school teachers); arts, design, entertainment, sports, and media occupations (except commercial and industrial designers, fashion designers, and ﬂoral [NEWLINE] designers); social science research assistants; and law clerks [NEWLINE] [NEWLINE] [STARTQ] The category “other occupations” includes drafters; protective service occupations; food preparation- and serving-related occupations; personal care; service occupations (except supervisors, animal care and service [ENDQ] workers, and entertainment attendants and related workers); farming, ﬁshing, and forestry occupations; construction and extraction occupations; installation, maintenance, and repair occupations; production occupations; [NEWLINE] transportation and material moving occupations (except air transportation workers); military speciﬁc occupations; farm/ranch/other agricultural managers; farmers and ranchers; cartographers and photogrammetrists; [NEWLINE] surveyors; athletes and sports competitors; coaches and scouts; umpire/referee/other sports ofﬁcials; and emergency medical technicians/paramedics [NEWLINE] [NEWLINE] Can you see<mask> men may tend to take different jobs than women even within their individual categories, and<mask> these jobs may lend itself to more pay? I applaud their attempt to actually account for the pay gap by asking for
Label encoding: <s>That's exactly it. Neither of your sources control for other factors. The most cited sources of these numbers, including the sources of your articles are the Consad Report for the US Department of labor, and an independent report done by the American Association of University Women (which blatantly conducted the study with an agenda in mind, but I'm just gonna assume they didn't let their bias affect the results or the conclusion they drew). [NEWLINE] [NEWLINE] After adjusting for occupation, the CONSAD report found an unexplained gap between 4.8% and 7.1% (let's say 7 cents on the dollar), and AAUW found a gap of 6.6 cents on the dollar unexplained by occupational differences. [NEWLINE] [NEWLINE] I'm just gonna examine the AAUW report and how they account for the occupational differences. In the report itself, the categories they use to classify occupation are pretty broad. In particular, there is an "Other Occupation" category and an "Other White Collar Occupation" category. [NEWLINE] [NEWLINE] From the report itself: [NEWLINE] [STARTQ] 1 [ENDQ] The category “other white-collar occupations” includes social scientists and related workers (except clinical, counseling, and school psychologists); lawyers, judges, and related workers; education, training, and library [NEWLINE] occupations (except primary, secondary, and special education school teachers); arts, design, entertainment, sports, and media occupations (except commercial and industrial designers, fashion designers, and ﬂoral [NEWLINE] designers); social science research assistants; and law clerks [NEWLINE] [NEWLINE] [STARTQ] The category “other occupations” includes drafters; protective service occupations; food preparation- and serving-related occupations; personal care; service occupations (except supervisors, animal care and service [ENDQ] workers, and entertainment attendants and related workers); farming, ﬁshing, and forestry occupations; construction and extraction occupations; installation, maintenance, and repair occupations; production occupations; [NEWLINE] transportation and material moving occupations (except air transportation workers); military speciﬁc occupations; farm/ranch/other agricultural managers; farmers and ranchers; cartographers and photogrammetrists; [NEWLINE] surveyors; athletes and sports competitors; coaches and scouts; umpire/referee/other sports ofﬁcials; and emergency medical technicians/paramedics [NEWLINE] [NEWLINE] Can you see how men may tend to take different jobs than women even within their individual categories, and how these jobs may lend itself to more pay? I applaud their attempt to actually account for the pay gap by asking for
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a couple of things to think about. [NEWLINE] [NEWLINE] 1) Removing the assistance is not likely going to affect the birth rates of those in poverty.<mask><mask><mask>,<mask> the birth rates between those on welfare and the general population aren't that different. And second,<mask> that isn't the case around the world. For example, Ethiopia has the [6th highest birth rate in the world]( [URL] ), and<mask> [44% of its entire population is malnourished]( [URL] ). [NEWLINE] [NEWLINE] 2) Related to the above: [birth rates go down<mask> education/wealth go up]( [URL] ).<mask>, arguably,<mask> you want these people to have less children, we should give them *more* assistance. [NEWLINE] [NEWLINE] 3) Removing children from "unfit" parents is exponentially more expensive than welfare/cash assistance. In Ohio, the foster care rate is between $10 and $200 *per day*. That doesn't even count emergency clothing vouchers, respite care, sending them to managed care facilities (even more expensive), and other expenditures, including the management of their case. And let's not pretend that there aren't already *too many* kids in foster care currently.<mask> would all these extra kids go? [NEWLINE] [NEWLINE] 4) You invoked "society" without really examining<mask> society cares about: future workers, e.g. more babies. In 2010, the US birth rate [fell to its lowest<mask> 1920]( [URL] ).<mask><mask>, the US population would be *in decline*<mask> it were not for the fact that we get millions of (legal) immigrants each year.<mask><mask><mask><mask> the children of these welfare parents eventually become tax-paying citizens (which all of them will, income tax aside), then technically "society" wants them all. [NEWLINE] [NEWLINE] 5) Good luck "disallowing" parents to have more children, no matter<mask> the circumstances entail. [NEWLINE] [NEWLINE] Ultimately, you seem caught up in the "fairness" argument on an individual level.<mask> comforting, it is important to acknowledge that<mask> is "fair" is arbitrary and largely meaningless. Any given welfare recipient might receive more money than they pay out in taxes (<mask> who actually knows?),<mask> A) the money circulates/stimulates the economy regardless, B) the mere existence of a safety net allows everyone to make riskier investments/grants more class mobility.<mask> you sign up for insurance on Monday and crash your car on Tuesday, is that unfair? Or is that insurance working<mask> intended
Label encoding: <s>There are a couple of things to think about. [NEWLINE] [NEWLINE] 1) Removing the assistance is not likely going to affect the birth rates of those in poverty. First of all, because the birth rates between those on welfare and the general population aren't that different. And second, because that isn't the case around the world. For example, Ethiopia has the [6th highest birth rate in the world]( [URL] ), and yet [44% of its entire population is malnourished]( [URL] ). [NEWLINE] [NEWLINE] 2) Related to the above: [birth rates go down as education/wealth go up]( [URL] ). So, arguably, if you want these people to have less children, we should give them *more* assistance. [NEWLINE] [NEWLINE] 3) Removing children from "unfit" parents is exponentially more expensive than welfare/cash assistance. In Ohio, the foster care rate is between $10 and $200 *per day*. That doesn't even count emergency clothing vouchers, respite care, sending them to managed care facilities (even more expensive), and other expenditures, including the management of their case. And let's not pretend that there aren't already *too many* kids in foster care currently. Where would all these extra kids go? [NEWLINE] [NEWLINE] 4) You invoked "society" without really examining what society cares about: future workers, e.g. more babies. In 2010, the US birth rate [fell to its lowest since 1920]( [URL] ). In fact, the US population would be *in decline* if it were not for the fact that we get millions of (legal) immigrants each year. So as long as the children of these welfare parents eventually become tax-paying citizens (which all of them will, income tax aside), then technically "society" wants them all. [NEWLINE] [NEWLINE] 5) Good luck "disallowing" parents to have more children, no matter what the circumstances entail. [NEWLINE] [NEWLINE] Ultimately, you seem caught up in the "fairness" argument on an individual level. While comforting, it is important to acknowledge that what is "fair" is arbitrary and largely meaningless. Any given welfare recipient might receive more money than they pay out in taxes ( but who actually knows?), but A) the money circulates/stimulates the economy regardless, B) the mere existence of a safety net allows everyone to make riskier investments/grants more class mobility. If you sign up for insurance on Monday and crash your car on Tuesday, is that unfair? Or is that insurance working as intended
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0644, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm gonna reference some Aristotle here on<mask> there isn't a large council of CEOs. Aristotle concluded the best governments in order are Monarchy, followed by Aristocracy, followed by Democracy. The inverse of these<mask> are the opposite. The worst form of government is Tyranny (corrupted Monarchy), followed by Oligarchy (corrupted Aristocracy) and Ochlocracy (Mob Rule, corrupted Democracy). They are ranked by<mask> many people something has to go through to get things done, with the fewer being the better. Now think of some governments.<mask> a monarch wants something done, it gets done right away, no questions asked.<mask> an aristocracy wants something done, it can be done fairly quick. Democracy can slowly move and at some point be at a total stand still (think US Congress) which can easily put a company under.<mask> power corrupts people, a monarch is most likely to corrupt into a tyranny<mask> that is risky. An aristocracy is fairly likely to corrupt into an oligarchy, and a democracy is least likely to develop into mob rule(rule by the needy), which makes it the safest. that's<mask> the government is run that way, to avoid corruption. a business<mask> needs things to get done. corruption at the head (in the sense of rule) can be avoided<mask><mask> the company is in trouble, the CEO is in trouble. they typically have little job security(they screw up, they're removed) and  it is much smaller than a government<mask> it can afford to have fewer people at the helm. the head of a business has little to gain from becoming tyrannical<mask> much to lose (the opposite of a government monarchy).<mask> you have a few good decision makers(an elected board) running things, things can be done swiftly and well. the more people you have in there the slower things go and the more likely it is to fail (<mask> often ideas can only work<mask> done 100% and democracy leads to things being half-assed<mask> its difficult to get that many people to agree).<mask> having 67 equally in charge decision makers, that company will not run smoothly<mask> you'll have 67 people making individual decisions that ultimately leave the company nowhere<mask> it everyone disagrees on everything. having a powerful head with a board to balance the power and help makes for the most effective way to get things done. that's<mask> the power is with one person. the company decides people's merit, and that in itself is gray<mask> someones pay comes down to<mask> important they think a job
Label encoding: <s>I'm gonna reference some Aristotle here on why there isn't a large council of CEOs. Aristotle concluded the best governments in order are Monarchy, followed by Aristocracy, followed by Democracy. The inverse of these though are the opposite. The worst form of government is Tyranny (corrupted Monarchy), followed by Oligarchy (corrupted Aristocracy) and Ochlocracy (Mob Rule, corrupted Democracy). They are ranked by how many people something has to go through to get things done, with the fewer being the better. Now think of some governments. If a monarch wants something done, it gets done right away, no questions asked. If an aristocracy wants something done, it can be done fairly quick. Democracy can slowly move and at some point be at a total stand still (think US Congress) which can easily put a company under. Because power corrupts people, a monarch is most likely to corrupt into a tyranny so that is risky. An aristocracy is fairly likely to corrupt into an oligarchy, and a democracy is least likely to develop into mob rule(rule by the needy), which makes it the safest. that's why the government is run that way, to avoid corruption. a business however needs things to get done. corruption at the head (in the sense of rule) can be avoided because if the company is in trouble, the CEO is in trouble. they typically have little job security(they screw up, they're removed) and  it is much smaller than a government so it can afford to have fewer people at the helm. the head of a business has little to gain from becoming tyrannical but much to lose (the opposite of a government monarchy). if you have a few good decision makers(an elected board) running things, things can be done swiftly and well. the more people you have in there the slower things go and the more likely it is to fail ( because often ideas can only work if done 100% and democracy leads to things being half-assed because its difficult to get that many people to agree). So having 67 equally in charge decision makers, that company will not run smoothly because you'll have 67 people making individual decisions that ultimately leave the company nowhere because it everyone disagrees on everything. having a powerful head with a board to balance the power and help makes for the most effective way to get things done. that's why the power is with one person. the company decides people's merit, and that in itself is gray as someones pay comes down to how important they think a job
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually, he *can* do both, and judges do both just about every single day. [NEWLINE] [NEWLINE] They are bound by the caselaw that came before them,<mask> they have leeway<mask> it comes to certain discretionary items: for instance, bond conditions (determined based on flight risk and threat to community, etc.) and sentencing (determined based on seriousness of crime, history of defendant, sentencing guidelines, and argument by counsel). I'm not sure of your knowledge of the justice system<mask> I'll expound just a little bit here, forgive me<mask> it's unnecessary. [NEWLINE] [NEWLINE] In regards to sentencing; in the United States, judges decide criminal sentences based on the criminal sentencing guidelines, which amounts to a series of checks that help them analyze a defendant's role in the commission of a certain crime<mask> well<mask> the defendant's criminal history. Essentially, the prosecutor<mask> well<mask> the defense attorney will "score" the Defendant and compare the numbers they come up with; there is room for argument here, which is<mask> prep for sentencing can be a laborious procedure for an attorney. The "score" that each attorney comes up with will correspond to a table of values, for each of which there is a *high* end and a *low* end. The judge can, based on his/her take on the case and the arguments of counsel, decide<mask> within that bracket to sentence the Defendant. [NEWLINE] [NEWLINE] I wholeheartedly disagree that for a law to be "just" that it must carry with it a universal standard (assuming you mean the sentence/punishment here). The rigidity you suggest is untenable for two reasons:<mask> each and every single court case is unique and has factors both mitigating and aggravating; and, I would posit, this flexibility is one of the core tenets of the advocative justice system (i.e. you choose someone to represent you in a court of law<mask> you trust they will fight like hell for you, knowing that on the other side there is someone doing the same thing to fight like hell against you). The nature of the system necessitates a certain amount of "judgment calls." [NEWLINE] [NEWLINE] <mask><mask><mask> stare decisis goes, a good way to look at it,<mask><mask><mask>, is to view caselaw<mask> the material that fills in the "negative space" around which the judge determines a sentence: they cannot go outside the boundaries of the caselaw the precedes them.<mask><mask><mask> they are operating within those boundaries,<mask>, and within
Label encoding: <s>Actually, he *can* do both, and judges do both just about every single day. [NEWLINE] [NEWLINE] They are bound by the caselaw that came before them, but they have leeway when it comes to certain discretionary items: for instance, bond conditions (determined based on flight risk and threat to community, etc.) and sentencing (determined based on seriousness of crime, history of defendant, sentencing guidelines, and argument by counsel). I'm not sure of your knowledge of the justice system so I'll expound just a little bit here, forgive me if it's unnecessary. [NEWLINE] [NEWLINE] In regards to sentencing; in the United States, judges decide criminal sentences based on the criminal sentencing guidelines, which amounts to a series of checks that help them analyze a defendant's role in the commission of a certain crime as well as the defendant's criminal history. Essentially, the prosecutor as well as the defense attorney will "score" the Defendant and compare the numbers they come up with; there is room for argument here, which is why prep for sentencing can be a laborious procedure for an attorney. The "score" that each attorney comes up with will correspond to a table of values, for each of which there is a *high* end and a *low* end. The judge can, based on his/her take on the case and the arguments of counsel, decide where within that bracket to sentence the Defendant. [NEWLINE] [NEWLINE] I wholeheartedly disagree that for a law to be "just" that it must carry with it a universal standard (assuming you mean the sentence/punishment here). The rigidity you suggest is untenable for two reasons: because each and every single court case is unique and has factors both mitigating and aggravating; and, I would posit, this flexibility is one of the core tenets of the advocative justice system (i.e. you choose someone to represent you in a court of law because you trust they will fight like hell for you, knowing that on the other side there is someone doing the same thing to fight like hell against you). The nature of the system necessitates a certain amount of "judgment calls." [NEWLINE] [NEWLINE] As far as stare decisis goes, a good way to look at it, in my opinion, is to view caselaw as the material that fills in the "negative space" around which the judge determines a sentence: they cannot go outside the boundaries of the caselaw the precedes them. So long as they are operating within those boundaries, though, and within
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> sociologists are really your go-to source for study of society, and<mask> they look at something like racism they need to use their jargon to further specify<mask> exactly they're discussing. [ENDQ] [NEWLINE] The question of affirmative action,<mask><mask><mask>, is **not** the purview of sociologists.  It's the purview of citizens everywhere of a democracy. <mask> we're talking about political questions, I don't think sociologists can lay claim to the territory, and then insist people use their terminology.  Political discussion has to take the terminology of the common people, or else language is being used to exclude. [NEWLINE] [NEWLINE] I wouldn't walk into a sociology classroom and insist they use my definition. <mask><mask> it comes to political questions, it's an arena<mask> they have no special power. [NEWLINE] [NEWLINE] To connect it to the OP, you say my definition works in "day-to-day use" - "reverse racism" is a "day-to-day" term, and that's the situation you'd expect to hear it in. <mask> I say "reverse racism is racism", then we're probably in a "day-to-day" situation. [NEWLINE] [NEWLINE] [STARTQ] There's no difference between botanists and sociologists in this scenario. [ENDQ] [NEWLINE] The difference, and part of<mask> I'm coming from, is this. <mask><mask> politics corrodes everything it touches.  Scientific disciplines exposed to politics are liable to turn into partisan bickering and employ poor, motivated arguments. <mask> the correct classification of rhubarb had big implications for political questions, I'd start to suspect botanists' definitions too.  You'd start to see academic papers about rhubarb with long tangents about abortion and shit. [NEWLINE] [NEWLINE] With that in mind, I see the sociological definition of racism<mask> motivated by political ideology.  They could have used another term for "power + prejudice", or invented a term.  By using a previously-common word, they're doing 2 things.  One, they're taking advantage of that word's negative connotations and trying to channel it towards specific things.  Two, they're making a normative statement about which type of racism is worse.  It's not just neutral, disinterested defining of terms. [NEWLINE] [NEWLINE] <mask> we start talking about affirmative action, sociologists are entering the realm of the common voters, and they should use the language of the common voters, rather than insist common voters use the terminology of some
Label encoding: <s> [STARTQ] But sociologists are really your go-to source for study of society, and when they look at something like racism they need to use their jargon to further specify what exactly they're discussing. [ENDQ] [NEWLINE] The question of affirmative action, in my opinion, is **not** the purview of sociologists.  It's the purview of citizens everywhere of a democracy.  When we're talking about political questions, I don't think sociologists can lay claim to the territory, and then insist people use their terminology.  Political discussion has to take the terminology of the common people, or else language is being used to exclude. [NEWLINE] [NEWLINE] I wouldn't walk into a sociology classroom and insist they use my definition.  But when it comes to political questions, it's an arena where they have no special power. [NEWLINE] [NEWLINE] To connect it to the OP, you say my definition works in "day-to-day use" - "reverse racism" is a "day-to-day" term, and that's the situation you'd expect to hear it in.  If I say "reverse racism is racism", then we're probably in a "day-to-day" situation. [NEWLINE] [NEWLINE] [STARTQ] There's no difference between botanists and sociologists in this scenario. [ENDQ] [NEWLINE] The difference, and part of where I'm coming from, is this.  I think politics corrodes everything it touches.  Scientific disciplines exposed to politics are liable to turn into partisan bickering and employ poor, motivated arguments.  If the correct classification of rhubarb had big implications for political questions, I'd start to suspect botanists' definitions too.  You'd start to see academic papers about rhubarb with long tangents about abortion and shit. [NEWLINE] [NEWLINE] With that in mind, I see the sociological definition of racism as motivated by political ideology.  They could have used another term for "power + prejudice", or invented a term.  By using a previously-common word, they're doing 2 things.  One, they're taking advantage of that word's negative connotations and trying to channel it towards specific things.  Two, they're making a normative statement about which type of racism is worse.  It's not just neutral, disinterested defining of terms. [NEWLINE] [NEWLINE] When we start talking about affirmative action, sociologists are entering the realm of the common voters, and they should use the language of the common voters, rather than insist common voters use the terminology of some
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0705, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My opinion? It depends. Imagine<mask> someone told you that they don't like you<mask> you talk too much. You're now very likely going to take that<mask> the truth, and apply that to your next relationship (which could be with someone who actually LIKES people who talk too much). That feedback becomes something you try to "improve on." Should it really be improved<mask>? I say, who cares<mask> 1 person thinks. They aren't a bad person, you aren't a bad person, you're just a bad combination. Sure you might want to know<mask> there are any common complaints about yourself, to see<mask> there's anything you should legitimately "fix,"<mask> I don't really think anyone needs to be fixed unless they're doing it for themselves. I believe there's someone out there for everyone. [NEWLINE] [NEWLINE] A quick thing to think about: Let's say you've gone to a certain doctor a few times. Do you fret over the idea of telling them you're switching doctors? No, probably not. [NEWLINE] [NEWLINE] <mask>, look at it like this. You have the opportunity to give your feedback to certain "people," like for example, restaurants, about<mask> they should improve,<mask> that restaurant is showing that they're fully open to the idea of critique,<mask> it's acceptable in this case. Are individual people the same way? Some are, some aren't. That leaves me with my original point of, "It depends." It depends on the person. Do you think they can handle the feedback? Are you sure the feedback would even do any good?<mask> they're<mask> "crazy" that you feel the need to tell them that, chances are you'll have a "crazy" reaction, and it most likely won't even register to them.<mask> they're not "crazy," they'll either be aware of the incompatibility, or they might ask you for genuine feedback, which you can then give to them without much worry. [NEWLINE] [NEWLINE] Now, a more objective standpoint. Perhaps "commitments" should simply be an equal balance.<mask> you tell someone you WILL be at their event, it's technically "customary" to let them know<mask> you can't come.<mask> you never say you're going to be there, it's ok to not let them know<mask> you can't come.<mask> you tell someone you want to spend the rest of your life with them, it's "customary" to tell them<mask> you don't.<mask> you're casually seeing someone and never mention you're exclusive
Label encoding: <s>My opinion? It depends. Imagine if someone told you that they don't like you because you talk too much. You're now very likely going to take that as the truth, and apply that to your next relationship (which could be with someone who actually LIKES people who talk too much). That feedback becomes something you try to "improve on." Should it really be improved though? I say, who cares what 1 person thinks. They aren't a bad person, you aren't a bad person, you're just a bad combination. Sure you might want to know if there are any common complaints about yourself, to see if there's anything you should legitimately "fix," but I don't really think anyone needs to be fixed unless they're doing it for themselves. I believe there's someone out there for everyone. [NEWLINE] [NEWLINE] A quick thing to think about: Let's say you've gone to a certain doctor a few times. Do you fret over the idea of telling them you're switching doctors? No, probably not. [NEWLINE] [NEWLINE] Also, look at it like this. You have the opportunity to give your feedback to certain "people," like for example, restaurants, about what they should improve, but that restaurant is showing that they're fully open to the idea of critique, so it's acceptable in this case. Are individual people the same way? Some are, some aren't. That leaves me with my original point of, "It depends." It depends on the person. Do you think they can handle the feedback? Are you sure the feedback would even do any good? If they're so "crazy" that you feel the need to tell them that, chances are you'll have a "crazy" reaction, and it most likely won't even register to them. If they're not "crazy," they'll either be aware of the incompatibility, or they might ask you for genuine feedback, which you can then give to them without much worry. [NEWLINE] [NEWLINE] Now, a more objective standpoint. Perhaps "commitments" should simply be an equal balance. If you tell someone you WILL be at their event, it's technically "customary" to let them know if you can't come. If you never say you're going to be there, it's ok to not let them know if you can't come. If you tell someone you want to spend the rest of your life with them, it's "customary" to tell them if you don't. If you're casually seeing someone and never mention you're exclusive
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's examine the [aforementioned scene]( [URL] #t=55) of his "blunder." [NEWLINE] [NEWLINE] Fight break out and he heads out. Lands on the skiff and Luke immediately uses his superior weaponry to destroy his blaster. Not phased, he uses momentary distraction to tie Luke up with fibercord, sealing the fight.<mask>, the gunner on the sail barge decides that is a good opportunity to fire, Luke is able to move his lightsaber slightly to deflect the shot, forcing Fett to dodge and release tension on his fibercord, freeing Luke. Luke goes to deal with the newly arrived sandskiff. Boba gets up, and takes aim at Luke with his wristrocket,<mask> gets hit from behind (mistakenly) by Han, causing his jetpack to malfunction and lead to his (movie) demise. [NEWLINE] [NEWLINE] Breaking that down, Boba did well.<mask> immediate setback, he takes the advantage,<mask> is foiled by friendly fire. Then again he is ready to take the match point,<mask> is hit from behind and subsequently suffered equipment malfunction. Now, the argument can be made that he should have taken care of Han first,<mask> really who can fault him for being more concerned with the Jedi swinging around a lightsaber than the blind man. [NEWLINE] [NEWLINE] <mask> this is no evidence for a good reputation in itself, it certainly doesn't show incompetence. [NEWLINE] [NEWLINE] Looking back at<mask> happens at ESB, it really is brilliance. He correctly predicts the Falcon didn't flee the system and is hiding,<mask> he hides<mask> well.<mask><mask><mask>, he is able to track the Falcon (who he has been personally told not to destroy,<mask> deliver to Vader), and extrapolate the intended target of Bespin. From there is able to alert Vader and setup an ambush to capture them<mask> cleanly<mask> possible. Remember there are two bounties here: 1 is on Han from Jabba, the other is on the delivery of the Falcon to the Empire, which he does to the letter. [NEWLINE] [NEWLINE] During their time on Bespin, he is confident enough to convince Vader to give him Han (<mask> completing the double-bounty),<mask> well<mask> challenge him for compensation should he not survive the freezing process. Just think about that, he stands up to Vader, the most notorious Villian of the Empire. Following this, he holds off Luke from getting to his friends,<mask><mask> NOT killing or hurting him (remember Vader wants Luke, and doesn't want the
Label encoding: <s>Let's examine the [aforementioned scene]( [URL] #t=55) of his "blunder." [NEWLINE] [NEWLINE] Fight break out and he heads out. Lands on the skiff and Luke immediately uses his superior weaponry to destroy his blaster. Not phased, he uses momentary distraction to tie Luke up with fibercord, sealing the fight. However, the gunner on the sail barge decides that is a good opportunity to fire, Luke is able to move his lightsaber slightly to deflect the shot, forcing Fett to dodge and release tension on his fibercord, freeing Luke. Luke goes to deal with the newly arrived sandskiff. Boba gets up, and takes aim at Luke with his wristrocket, but gets hit from behind (mistakenly) by Han, causing his jetpack to malfunction and lead to his (movie) demise. [NEWLINE] [NEWLINE] Breaking that down, Boba did well. Despite immediate setback, he takes the advantage, but is foiled by friendly fire. Then again he is ready to take the match point, but is hit from behind and subsequently suffered equipment malfunction. Now, the argument can be made that he should have taken care of Han first, but really who can fault him for being more concerned with the Jedi swinging around a lightsaber than the blind man. [NEWLINE] [NEWLINE] While this is no evidence for a good reputation in itself, it certainly doesn't show incompetence. [NEWLINE] [NEWLINE] Looking back at what happens at ESB, it really is brilliance. He correctly predicts the Falcon didn't flee the system and is hiding, so he hides as well. As a result, he is able to track the Falcon (who he has been personally told not to destroy, but deliver to Vader), and extrapolate the intended target of Bespin. From there is able to alert Vader and setup an ambush to capture them as cleanly as possible. Remember there are two bounties here: 1 is on Han from Jabba, the other is on the delivery of the Falcon to the Empire, which he does to the letter. [NEWLINE] [NEWLINE] During their time on Bespin, he is confident enough to convince Vader to give him Han ( thus completing the double-bounty), as well as challenge him for compensation should he not survive the freezing process. Just think about that, he stands up to Vader, the most notorious Villian of the Empire. Following this, he holds off Luke from getting to his friends, while also NOT killing or hurting him (remember Vader wants Luke, and doesn't want the
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many moons ago, I had a male friend who had an iphone right<mask> they first came out.  Obviously he was flaunting it a bit, and was used to fellow professionals making positive comments about it, and was flattered. [NEWLINE] [NEWLINE] Not too long afterward, he had to take the bus due to his car breaking down.  He had his iphone out and was messing around with it,<mask> another guy on the bus got into his personal space, was heckling my friend, oggling the phone, then stated that he wanted to sell it for money. [NEWLINE] [NEWLINE] After telling me this story, he complained that he felt he couldn't really have his iphone visible in public.  He perceived people on the street<mask> being potential threats.  He could be robbed, he could be injured in the theft.  Every glance, be they interest, jealousy, or curiosity, instead read to him<mask> a potential threat. [NEWLINE] [NEWLINE] It was for me, a good comparison to<mask> I feel about catcalls.  In college, I had an evening<mask> I was walking back from a study group at a local coffeehouse.  It was a football weekend, and the drunks were out in full force.  It<mask> was just before midterms, I was exhausted, and I had to grab my month's rent. <mask> I stood at the ATM withdrawing the cash, two older, drunken guys came up to the glass enclosure and started making really crude comments.  I grabbed my money and left.  They followed. [NEWLINE] [NEWLINE] That walk home was the scariest mile I've ever walked.  I had two drunk guys around me, getting in my personal space, making comments that had a threat of sexual violence,<mask> well<mask> comments showing they were aware I had a large quantity of cash on me.  The more I ignored them, the angrier and more threatening they became, until they were downright saying that I was a bitch who deserved to be raped and robbed. [NEWLINE] [NEWLINE] The weird thing was,<mask> I told this story afterwards, most of my male friends couldn't see<mask> this was<mask> terrifying.  The guys were just having a little fun, they said, I should take it<mask> a compliment.  Hell, my own father didn't have too much issue with it. [NEWLINE] [NEWLINE] I honestly don't understand the disconnect with some men.  Is it really that they haven't been in a situation like this and can't empathize?  Or is it that
Label encoding: <s>Many moons ago, I had a male friend who had an iphone right when they first came out.  Obviously he was flaunting it a bit, and was used to fellow professionals making positive comments about it, and was flattered. [NEWLINE] [NEWLINE] Not too long afterward, he had to take the bus due to his car breaking down.  He had his iphone out and was messing around with it, when another guy on the bus got into his personal space, was heckling my friend, oggling the phone, then stated that he wanted to sell it for money. [NEWLINE] [NEWLINE] After telling me this story, he complained that he felt he couldn't really have his iphone visible in public.  He perceived people on the street as being potential threats.  He could be robbed, he could be injured in the theft.  Every glance, be they interest, jealousy, or curiosity, instead read to him as a potential threat. [NEWLINE] [NEWLINE] It was for me, a good comparison to how I feel about catcalls.  In college, I had an evening where I was walking back from a study group at a local coffeehouse.  It was a football weekend, and the drunks were out in full force.  It also was just before midterms, I was exhausted, and I had to grab my month's rent.  As I stood at the ATM withdrawing the cash, two older, drunken guys came up to the glass enclosure and started making really crude comments.  I grabbed my money and left.  They followed. [NEWLINE] [NEWLINE] That walk home was the scariest mile I've ever walked.  I had two drunk guys around me, getting in my personal space, making comments that had a threat of sexual violence, as well as comments showing they were aware I had a large quantity of cash on me.  The more I ignored them, the angrier and more threatening they became, until they were downright saying that I was a bitch who deserved to be raped and robbed. [NEWLINE] [NEWLINE] The weird thing was, when I told this story afterwards, most of my male friends couldn't see why this was so terrifying.  The guys were just having a little fun, they said, I should take it as a compliment.  Hell, my own father didn't have too much issue with it. [NEWLINE] [NEWLINE] I honestly don't understand the disconnect with some men.  Is it really that they haven't been in a situation like this and can't empathize?  Or is it that
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The basic issue is a lack of empathy,<mask><mask>. <mask> you get better and better at something, you have to overcome difficulties.  Actually, it's the reverse--you get better and better *by* overcoming difficulties.  You've had to work your butt off to go through all that, and<mask> you see someone else experiencing those same difficulties, making the same mistakes you were making, you get it.  You understand that they're at a certain point in their development, maybe stuck there,<mask> you see the larger context of their growth.  You get it. [NEWLINE] [NEWLINE] <mask> you only get it<mask> you've been through it yourself.  Or<mask> you're very good at using your imagination to put yourself into a 2nd person perspective. [NEWLINE] [NEWLINE] I'm a musician (I've even played on the X-Factor!), and I'm a LOT less critical of other players than I used to be.  I've<mask> noticed,<mask> I've begun hanging around musicians who are more and more talented, *they* are less and less critical of other musicians--<mask> we understand the difficulties those people face, we know<mask> it takes to "just be good", and that sometimes it doesn't always work out that you're<mask> good<mask> you'd like to be at any given moment.  I'm continually noticing an inverse relationship between a person's skill at something, and<mask> critical they are of others doing the same thing. [NEWLINE] [NEWLINE] <mask><mask> most of us are vaguely aware of this on some intuitive level, and I might even go<mask> far<mask> to suspect that the other youtubers are subconsciously *inferring* a lack of musical skill on your part *from your criticism*, which would make sense out of their statements.  After all, they don't *know* you can't do it better, do they? [NEWLINE] [NEWLINE]...do they? [NEWLINE] [NEWLINE] <mask> it's not that you *can't* criticize, it's that<mask> you do, it's a criticism that's not worth much.  Sure, there is value in getting a beginner's untainted opinion,<mask> it sounds like you didn't present it that way.  Essentially you shat all over that process of empathy and understanding, something that a person who's made it past that point in the road would not do. [NEWLINE] [NEWLINE] **tl;dr**--people who *can* do better will probably not criticize<mask> they know<mask> it's like to be at that point in the journey, and that gives them empathy
Label encoding: <s>The basic issue is a lack of empathy, I think.  As you get better and better at something, you have to overcome difficulties.  Actually, it's the reverse--you get better and better *by* overcoming difficulties.  You've had to work your butt off to go through all that, and when you see someone else experiencing those same difficulties, making the same mistakes you were making, you get it.  You understand that they're at a certain point in their development, maybe stuck there, but you see the larger context of their growth.  You get it. [NEWLINE] [NEWLINE] But you only get it if you've been through it yourself.  Or if you're very good at using your imagination to put yourself into a 2nd person perspective. [NEWLINE] [NEWLINE] I'm a musician (I've even played on the X-Factor!), and I'm a LOT less critical of other players than I used to be.  I've also noticed, as I've begun hanging around musicians who are more and more talented, *they* are less and less critical of other musicians-- because we understand the difficulties those people face, we know what it takes to "just be good", and that sometimes it doesn't always work out that you're as good as you'd like to be at any given moment.  I'm continually noticing an inverse relationship between a person's skill at something, and how critical they are of others doing the same thing. [NEWLINE] [NEWLINE] I think most of us are vaguely aware of this on some intuitive level, and I might even go so far as to suspect that the other youtubers are subconsciously *inferring* a lack of musical skill on your part *from your criticism*, which would make sense out of their statements.  After all, they don't *know* you can't do it better, do they? [NEWLINE] [NEWLINE]...do they? [NEWLINE] [NEWLINE] So it's not that you *can't* criticize, it's that if you do, it's a criticism that's not worth much.  Sure, there is value in getting a beginner's untainted opinion, but it sounds like you didn't present it that way.  Essentially you shat all over that process of empathy and understanding, something that a person who's made it past that point in the road would not do. [NEWLINE] [NEWLINE] **tl;dr**--people who *can* do better will probably not criticize because they know what it's like to be at that point in the journey, and that gives them empathy
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi, I work extensively with web and mobile,<mask> I'm just going to say this. [NEWLINE] [NEWLINE] ***Websites are cheap. Really cheap.*** [NEWLINE] [NEWLINE] A domain registration runs at 10$ a year. A shared host or cheap dedicated runs at 5$ a month, no problem, and can usually handle sub 50k uniques a month without too much trouble. A mid level dedicated runs at 100$ a month or<mask>, and can often handle 100k page serves an hour. Costs scale up fairly dramatically after that,<mask> otherwise... they're cheap. [NEWLINE] [NEWLINE] <mask> adblock killed off all ads on the internet, and removed all ad revenue period, it would of course be a huge blow.<mask>, it wouldn't really reduce the amount of free websites. It's far more likely it would just distribute users more evenly over websites. Large company websites are run predominantly off company revenue, and are not necessarily generating revenue themselves. Reddit runs in the red.<mask><mask> would happen? Not tons. [NEWLINE] [NEWLINE] Large serving sites like youtube would almost certainly go down.<mask> they didn't, they'd probably be grabbed by major music or tv labels<mask> a vehicle to push their own content predominantly. Sports sites would be largely unaffected - ad revenue is supporting revenue, not primary for them. News sites may or may not go down, depending. Things like New York Times would not. Sites like Kotaku or Verge would probably subsist on article payoffs from companies without too much trouble. Content farms like all those '10 things you need to Xxx for' would mostly die or paywall, thank god. Webcomics would be largely unaffected - free or cheap hosting is too easily available and webcomics with only a few exceptions are low traffic, and those exceptions can generally leverage many income streams other than ads (and do). Reddit? Reddit I have no idea, by all rights it should have died already. [NEWLINE] [NEWLINE] In general, the web would see the fall of content farms and aggregate sites (who subsist on ad revenue primarily), and a number of large content community sites would<mask> hit the dust or become commercialized. Sites dealing with extreme traffic of video, and to a lesser extent images would<mask> take a huge hit and content sharing would become drastically more decentralized. Some small dross sites and bloggers that exist to push ads and affiliate links would die. Otherwise, probably not much effect. [NEWLINE] [NEWLINE] It's hard to kill the free web via ads<mask> literally every person connected to the net can host a small site
Label encoding: <s>Hi, I work extensively with web and mobile, so I'm just going to say this. [NEWLINE] [NEWLINE] ***Websites are cheap. Really cheap.*** [NEWLINE] [NEWLINE] A domain registration runs at 10$ a year. A shared host or cheap dedicated runs at 5$ a month, no problem, and can usually handle sub 50k uniques a month without too much trouble. A mid level dedicated runs at 100$ a month or so, and can often handle 100k page serves an hour. Costs scale up fairly dramatically after that, but otherwise... they're cheap. [NEWLINE] [NEWLINE] If adblock killed off all ads on the internet, and removed all ad revenue period, it would of course be a huge blow. However, it wouldn't really reduce the amount of free websites. It's far more likely it would just distribute users more evenly over websites. Large company websites are run predominantly off company revenue, and are not necessarily generating revenue themselves. Reddit runs in the red. So what would happen? Not tons. [NEWLINE] [NEWLINE] Large serving sites like youtube would almost certainly go down. If they didn't, they'd probably be grabbed by major music or tv labels as a vehicle to push their own content predominantly. Sports sites would be largely unaffected - ad revenue is supporting revenue, not primary for them. News sites may or may not go down, depending. Things like New York Times would not. Sites like Kotaku or Verge would probably subsist on article payoffs from companies without too much trouble. Content farms like all those '10 things you need to Xxx for' would mostly die or paywall, thank god. Webcomics would be largely unaffected - free or cheap hosting is too easily available and webcomics with only a few exceptions are low traffic, and those exceptions can generally leverage many income streams other than ads (and do). Reddit? Reddit I have no idea, by all rights it should have died already. [NEWLINE] [NEWLINE] In general, the web would see the fall of content farms and aggregate sites (who subsist on ad revenue primarily), and a number of large content community sites would also hit the dust or become commercialized. Sites dealing with extreme traffic of video, and to a lesser extent images would also take a huge hit and content sharing would become drastically more decentralized. Some small dross sites and bloggers that exist to push ads and affiliate links would die. Otherwise, probably not much effect. [NEWLINE] [NEWLINE] It's hard to kill the free web via ads when literally every person connected to the net can host a small site
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd imagine that a cave painting would survive quite a bit more easily than other forms of media,<mask><mask> there's no form of translation needed. I'm not arguing that piracy would be a bad form of perpetuating archiving, per se,<mask> the more steps you have to go through in order to render that which is being archived visible and accessible, the greater chance it will be lost somewhere along the way. Take Egyptian hieroglyphics, for example - without the key the Rosetta Stone provided, we wouldn't be able to access any of the data they stored. (Same applies for many other ancient languages.) Moving to digital video, we have to add in many more layers. [NEWLINE] [NEWLINE] Remember,<mask> you start to encode things more and more, they get less and less accessible. Assuming civilization survives remotely close to<mask> -is (and, of course,<mask> it doesn't,<mask> are future societies supposed to reverse-engineer our machines without any sort of access to any of the information we've provided in our own tongue?), we'd need in the future to keep working computers on hand, programs on hand which can decode the files in question into a format we can understand, screens to project those files, speakers or headphones to hear those files, sources of power<mask> that the computer can turn on in the first place, an assurance that the operating systems we have today will not lock sometime in the future, meaning we will no longer be able to access any of the pirated data - the list just goes on and on. Compare that to cave paintings, say, which require only a few things to be preserved - that the paint does not disintegrate or peel off<mask> tens of thousands of years pass, say, or that the cave does not collapse. [NEWLINE] [NEWLINE] <mask>,<mask> you're implying that I'm saying that piracy will *not* be an effective form of archival, that's one hundred percent not<mask> I'm saying. I fully believe that it's a valid method of keeping things stored for future generations, especially<mask> more and more people download GoT or Breaking Bad episodes.<mask>,<mask> I said above: is it possible for it to be the *most* effective? There are too many difficult-to-control variables - variables outside the scope of the piracy in the first place, variables that affect the storage of all digital media, not just pirated media - to be able to argue it will be, and that tens of thousands of years from now the only remnants of our civilization will be thousands
Label encoding: <s>I'd imagine that a cave painting would survive quite a bit more easily than other forms of media, given that there's no form of translation needed. I'm not arguing that piracy would be a bad form of perpetuating archiving, per se, but the more steps you have to go through in order to render that which is being archived visible and accessible, the greater chance it will be lost somewhere along the way. Take Egyptian hieroglyphics, for example - without the key the Rosetta Stone provided, we wouldn't be able to access any of the data they stored. (Same applies for many other ancient languages.) Moving to digital video, we have to add in many more layers. [NEWLINE] [NEWLINE] Remember, as you start to encode things more and more, they get less and less accessible. Assuming civilization survives remotely close to as -is (and, of course, if it doesn't, how are future societies supposed to reverse-engineer our machines without any sort of access to any of the information we've provided in our own tongue?), we'd need in the future to keep working computers on hand, programs on hand which can decode the files in question into a format we can understand, screens to project those files, speakers or headphones to hear those files, sources of power so that the computer can turn on in the first place, an assurance that the operating systems we have today will not lock sometime in the future, meaning we will no longer be able to access any of the pirated data - the list just goes on and on. Compare that to cave paintings, say, which require only a few things to be preserved - that the paint does not disintegrate or peel off as tens of thousands of years pass, say, or that the cave does not collapse. [NEWLINE] [NEWLINE] So, when you're implying that I'm saying that piracy will *not* be an effective form of archival, that's one hundred percent not what I'm saying. I fully believe that it's a valid method of keeping things stored for future generations, especially as more and more people download GoT or Breaking Bad episodes. However, as I said above: is it possible for it to be the *most* effective? There are too many difficult-to-control variables - variables outside the scope of the piracy in the first place, variables that affect the storage of all digital media, not just pirated media - to be able to argue it will be, and that tens of thousands of years from now the only remnants of our civilization will be thousands
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**The Third Chimpanzee**]( [URL] %20Third%20Chimpanzee): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] ___The Third Chimpanzee: The Evolution and Future of the Human Animal___ is a broad-focus book by academic and popular science author [Jared Diamond]( [URL] ), which explores concepts relating to the animal origins of human behavior, including cultural characteristics and those features often regarded<mask> particularly unique to humans. It further explores the question of<mask> [Homo sapiens]( [URL] ) came to dominate its closest relatives, such<mask> [chimpanzees]( [URL] ), and<mask> one group of humans ([Eurasians]( [URL] )) came to dominate others ([Indigenous peoples of the Americas]( [URL] ), for example). In answering these questions, Diamond (a professor in the fields of [physiology]( [URL] ) and [geography]( [URL] )) applies a variety of [biological]( [URL] ) and [anthropological]( [URL] ) arguments to reject traditional [hegemonic]( [URL] ) views that the dominant peoples came from "superior" genetic stock and argues instead that those peoples who came to dominate others did<mask><mask> of advantages found in their local environment which allowed them to develop larger populations, wider immunities to disease, and superior technologies for [agriculture]( [URL] ) and [warfare]( [URL] ). [ENDQ] [NEWLINE] [STARTQ] ==== [ENDQ] [NEWLINE] [STARTQ] [**Image**]( [URL].jpg) [^(i)]( [URL] :Thirdchimpcover.jpg) [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Jared ^Diamond]( [URL] ) ^| [^Homininae]( [URL] ) ^| [^Chimpanzee]( [URL] ) ^| [^Bonobo]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cjb4pv7) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cjb4pv7)^. ^Will ^<mask> ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods
Label encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**The Third Chimpanzee**]( [URL] %20Third%20Chimpanzee): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] ___The Third Chimpanzee: The Evolution and Future of the Human Animal___ is a broad-focus book by academic and popular science author [Jared Diamond]( [URL] ), which explores concepts relating to the animal origins of human behavior, including cultural characteristics and those features often regarded as particularly unique to humans. It further explores the question of how [Homo sapiens]( [URL] ) came to dominate its closest relatives, such as [chimpanzees]( [URL] ), and why one group of humans ([Eurasians]( [URL] )) came to dominate others ([Indigenous peoples of the Americas]( [URL] ), for example). In answering these questions, Diamond (a professor in the fields of [physiology]( [URL] ) and [geography]( [URL] )) applies a variety of [biological]( [URL] ) and [anthropological]( [URL] ) arguments to reject traditional [hegemonic]( [URL] ) views that the dominant peoples came from "superior" genetic stock and argues instead that those peoples who came to dominate others did so because of advantages found in their local environment which allowed them to develop larger populations, wider immunities to disease, and superior technologies for [agriculture]( [URL] ) and [warfare]( [URL] ). [ENDQ] [NEWLINE] [STARTQ] ==== [ENDQ] [NEWLINE] [STARTQ] [**Image**]( [URL].jpg) [^(i)]( [URL] :Thirdchimpcover.jpg) [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Jared ^Diamond]( [URL] ) ^| [^Homininae]( [URL] ) ^| [^Chimpanzee]( [URL] ) ^| [^Bonobo]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+cjb4pv7) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+cjb4pv7)^. ^Will ^ also ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some fair points. Let me go down the list: [NEWLINE] [NEWLINE] About the 60% statistic: I don't agree that 60% is "barely over half". There is a 20% difference between 60% and 40%. In 10 people, yes, only one more person needs to do something one way or another to make it a 60/40 split,<mask> in large numbers, that is still a resoundingly high proportion. [NEWLINE] [NEWLINE] Think about this: don't you think most wealthy people try to instill the value of money or proper money management in their children? This is anecdotal,<mask> from the inheritance-children I've met and seen in the media all seem to lose sight of "proper money management",<mask> they ever tried to manage it properly in the first place. [NEWLINE] [NEWLINE] To your point on their travel: I'm just not a fan of someone living it up off money they did nothing to earn.<mask>,<mask> you ask: "wouldn't you want your kids to have that feeling?", the answer is yes, absolutely,<mask> not<mask> it means getting that feeling from money that was handed to them through the chance of them being born<mask> a child of mine. [NEWLINE] [NEWLINE] To your point on my assumption:<mask><mask>, it is an assumption. I concede that.<mask>, your saying that "there are<mask> many people who would do many amazing things with an influx of cash" is an equally lost point. I can just<mask> equally say "there are<mask> many people who would do many terrible things with an influx of cash" and we're at a standstill.<mask> I said in my original post, I would love to instill in them the feeling of ambition and hardwork,<mask><mask><mask> it's difficult for a person to struggle forward<mask> they start out with the wind at their backs. A good quote I've heard is: "some people are born on third base thinking they've scored a triple". [NEWLINE] [NEWLINE] About the tycoon comment: Again, you're right. I should not have said "nearly every tycoon". I don't know that. I meant that a lot of people with considerable wealth have publicly commented on their belief that they shouldn't leave much or even anything to their children<mask> they die. Here is a list of 15, actually: [URL] [NEWLINE] [NEWLINE] <mask> to your closing: I mentioned in my original post that I want to give the money to charity,<mask> your closing is nullified... I don't mean to be rude,<mask> you were acting on the
Label encoding: <s>Some fair points. Let me go down the list: [NEWLINE] [NEWLINE] About the 60% statistic: I don't agree that 60% is "barely over half". There is a 20% difference between 60% and 40%. In 10 people, yes, only one more person needs to do something one way or another to make it a 60/40 split, but in large numbers, that is still a resoundingly high proportion. [NEWLINE] [NEWLINE] Think about this: don't you think most wealthy people try to instill the value of money or proper money management in their children? This is anecdotal, but from the inheritance-children I've met and seen in the media all seem to lose sight of "proper money management", if they ever tried to manage it properly in the first place. [NEWLINE] [NEWLINE] To your point on their travel: I'm just not a fan of someone living it up off money they did nothing to earn. So, when you ask: "wouldn't you want your kids to have that feeling?", the answer is yes, absolutely, but not if it means getting that feeling from money that was handed to them through the chance of them being born as a child of mine. [NEWLINE] [NEWLINE] To your point on my assumption: I agree, it is an assumption. I concede that. However, your saying that "there are so many people who would do many amazing things with an influx of cash" is an equally lost point. I can just as equally say "there are so many people who would do many terrible things with an influx of cash" and we're at a standstill. As I said in my original post, I would love to instill in them the feeling of ambition and hardwork, but I think it's difficult for a person to struggle forward if they start out with the wind at their backs. A good quote I've heard is: "some people are born on third base thinking they've scored a triple". [NEWLINE] [NEWLINE] About the tycoon comment: Again, you're right. I should not have said "nearly every tycoon". I don't know that. I meant that a lot of people with considerable wealth have publicly commented on their belief that they shouldn't leave much or even anything to their children when they die. Here is a list of 15, actually: [URL] [NEWLINE] [NEWLINE] As to your closing: I mentioned in my original post that I want to give the money to charity, so your closing is nullified... I don't mean to be rude, but you were acting on the
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here is a different angle on it that contains spoilers for the books and the shows.  I don't really care about the depiction of rape in a show that already depicts a fair amount of really horrible stuff, it really doesn't matter. <mask>, most of this stuff depicted in the show happened in the books<mask> well and they had some sort of purpose, be it for character development, or crafting the brutal GoT universe. [NEWLINE] [NEWLINE] The scene in question<mask>, does neither of this (<mask> you consider the show to still parallel the books at least, this argument breaks down<mask> you believe the show and books have taken a divergent path).  In the books the scene does not happen.  Sansa at the end of the last book is off in the Vale with Littlefinger and never even meets Ramsay.  Instead in the books a character called Jeyne Pool (<mask><mask> that's the spelling, haven't read the books in a<mask> ) is married to Ramsay under the pretence that she is actually Arya Stark.  Now in the books Jeyne is horribly abused by Ramsay and it's much, much worse than it is in the show, there's even some implied forced bestiality.  It's really, really nasty. <mask> Ramsay's treatment of Jeyne acts<mask> a catalyst for Theon/Reek to finally stand up to Ramsay and aid in a rescue attempt of her.  It acts<mask> an important character development tool. [NEWLINE] [NEWLINE] Next we can look at Sansa.  In the books she is developed into much of a stronger character,<mask> in the show.  At the end she begins to realise that she can<mask> become a player in the Game of Thrones and is actually being set up to challenge Littlefinger.  After being a weak character, she is finally beginning to stand up for herself and come into her own.  The rape scene in the show is disgusting<mask> it essentially nullifies this character development by just knocking her back down to being a weak character.  It goes<mask> strongly against<mask> the books and the show have been building her up to be it is really just bad writing. [NEWLINE] [NEWLINE] <mask> basically here we have a scene put into the show for shock value alone, that instead of adding to the story, actually detracts from it by derailing some really exciting development of a main character.  The derailing of Sansa's character development leaves this scene<mask> nothing more than a tasteless shock scene that unnecessarily tackles a very sensitive topic with no good
Label encoding: <s>Here is a different angle on it that contains spoilers for the books and the shows.  I don't really care about the depiction of rape in a show that already depicts a fair amount of really horrible stuff, it really doesn't matter.  However, most of this stuff depicted in the show happened in the books as well and they had some sort of purpose, be it for character development, or crafting the brutal GoT universe. [NEWLINE] [NEWLINE] The scene in question though, does neither of this ( if you consider the show to still parallel the books at least, this argument breaks down if you believe the show and books have taken a divergent path).  In the books the scene does not happen.  Sansa at the end of the last book is off in the Vale with Littlefinger and never even meets Ramsay.  Instead in the books a character called Jeyne Pool ( I think that's the spelling, haven't read the books in a while ) is married to Ramsay under the pretence that she is actually Arya Stark.  Now in the books Jeyne is horribly abused by Ramsay and it's much, much worse than it is in the show, there's even some implied forced bestiality.  It's really, really nasty.  However Ramsay's treatment of Jeyne acts as a catalyst for Theon/Reek to finally stand up to Ramsay and aid in a rescue attempt of her.  It acts as an important character development tool. [NEWLINE] [NEWLINE] Next we can look at Sansa.  In the books she is developed into much of a stronger character, as in the show.  At the end she begins to realise that she can also become a player in the Game of Thrones and is actually being set up to challenge Littlefinger.  After being a weak character, she is finally beginning to stand up for herself and come into her own.  The rape scene in the show is disgusting as it essentially nullifies this character development by just knocking her back down to being a weak character.  It goes so strongly against what the books and the show have been building her up to be it is really just bad writing. [NEWLINE] [NEWLINE] So basically here we have a scene put into the show for shock value alone, that instead of adding to the story, actually detracts from it by derailing some really exciting development of a main character.  The derailing of Sansa's character development leaves this scene as nothing more than a tasteless shock scene that unnecessarily tackles a very sensitive topic with no good
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Can a woman? [ENDQ] [NEWLINE] A woman cannot. [NEWLINE] [NEWLINE] And that is irrelevant to whether a man can or cannot decide his own value. [NEWLINE] [NEWLINE] Your claim was that a man can do such thing. [NEWLINE] [NEWLINE] I provided a demonstration<mask><mask> this is impossible. [NEWLINE] [NEWLINE] You completely ignored it. [NEWLINE] [NEWLINE] [STARTQ] There is a system called patriarchy (or kyriarchy,<mask> you prefer) which has set up value criteria for men and women. Seeing<mask> that women were considered property and / or extensions of the men in their lives for large periods during the creation of this system, you can't possibly<mask><mask> men did not put their own value judgements into this system. [ENDQ] [NEWLINE] Actually, I can possibly argue on that<mask> well. [NEWLINE] [NEWLINE] On two points, actually: [NEWLINE] [NEWLINE] 1. You are making the fallacy of using "women" and "men" to refer to different concepts, instead of using them<mask> gender variations of the same concept. [NEWLINE] Let me explain better:<mask> you say that "women were considered property and / or extensions of the men in their lives", you are talking about individuals. Each individual woman is enslaved, not "all women"<mask> a collective. [NEWLINE] <mask> you imagined "all women"<mask> a massive entity moved by a single purpose, you would have to concede that it would be pretty much impossible to enslave. [NEWLINE] <mask><mask><mask><mask>,<mask> you talk about "men deciding their own value", you are actually referring to "all men"<mask> a single entity that is able to move with a single purpose. [NEWLINE] <mask>? [NEWLINE] <mask> makes "men" a single overmind that can decide its own destiny,<mask> "women" are not a single overmind that "agreed to be enslaved<mask><mask><mask> having majority in numbers"? [NEWLINE] Or, the other way around,<mask> you agree that "women"<mask> a collection of independent individuals is unable to fight an oppressive social status quo, you would have to apply the same rules<mask> you talk about "men": each man is an independent individual who cannot simply decide his new role/value in society. [NEWLINE] In other words,<mask> you talk about "men" being in power or able to decide their own value in society, you are committing a form of ambiguous collective fallacy: [URL].HTM [NEWLINE] [NEWLINE] 2.<mask> you say "during the creation of this system" you are making a big assumption: that the sexism inherent in human society was "created" instead of simply being emergent (for example, tied to sexual dim
Label encoding: <s> [STARTQ] Can a woman? [ENDQ] [NEWLINE] A woman cannot. [NEWLINE] [NEWLINE] And that is irrelevant to whether a man can or cannot decide his own value. [NEWLINE] [NEWLINE] Your claim was that a man can do such thing. [NEWLINE] [NEWLINE] I provided a demonstration as why this is impossible. [NEWLINE] [NEWLINE] You completely ignored it. [NEWLINE] [NEWLINE] [STARTQ] There is a system called patriarchy (or kyriarchy, if you prefer) which has set up value criteria for men and women. Seeing as that women were considered property and / or extensions of the men in their lives for large periods during the creation of this system, you can't possibly argue that men did not put their own value judgements into this system. [ENDQ] [NEWLINE] Actually, I can possibly argue on that as well. [NEWLINE] [NEWLINE] On two points, actually: [NEWLINE] [NEWLINE] 1. You are making the fallacy of using "women" and "men" to refer to different concepts, instead of using them as gender variations of the same concept. [NEWLINE] Let me explain better: when you say that "women were considered property and / or extensions of the men in their lives", you are talking about individuals. Each individual woman is enslaved, not "all women" as a collective. [NEWLINE] If you imagined "all women" as a massive entity moved by a single purpose, you would have to concede that it would be pretty much impossible to enslave. [NEWLINE] On the other hand, when you talk about "men deciding their own value", you are actually referring to "all men" as a single entity that is able to move with a single purpose. [NEWLINE] Why? [NEWLINE] What makes "men" a single overmind that can decide its own destiny, but "women" are not a single overmind that "agreed to be enslaved in spite of having majority in numbers"? [NEWLINE] Or, the other way around, if you agree that "women" as a collection of independent individuals is unable to fight an oppressive social status quo, you would have to apply the same rules when you talk about "men": each man is an independent individual who cannot simply decide his new role/value in society. [NEWLINE] In other words, when you talk about "men" being in power or able to decide their own value in society, you are committing a form of ambiguous collective fallacy: [URL].HTM [NEWLINE] [NEWLINE] 2. When you say "during the creation of this system" you are making a big assumption: that the sexism inherent in human society was "created" instead of simply being emergent (for example, tied to sexual dim
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a lot of movements for accepting previously criticized groups such<mask> fat acceptance, nerd acceptance,... You get the idea. [NEWLINE] [NEWLINE] Sure, those movements make people feel better about themselves<mask> it makes them content with who they are thinking there's nothing wrong with it and not try to improve. There's a reason<mask> they were traditionally thought of<mask> weird. [NEWLINE] [NEWLINE] For example, the fat acceptance movement makes obese people think there is nothing wrong with their bodies and ignore the health risks of being obese. [NEWLINE] [NEWLINE] <mask> some nerds possess moderate social skills, a lot of them don't have social skills good enough to make one friend in an anime convention of thousands of people.<mask><mask> a poll used for an anime Family Feud in an anime convention, at least 60% make 0 friends in a 3 day convention with people of similar interests. That's absurd. [NEWLINE] [NEWLINE] I can't speak for fat acceptance<mask><mask> a previous nerd with no social skills, it was the feeling of rejection that made me improve and push myself to hone my social skills.<mask> the acceptance movements were around back then I would be happier<mask> a nerd<mask> I'd still have 0 friends. That's not to say that I am no longer a nerd, I still have all my previous interests<mask> I know<mask> to talk about things, I know not to correct people too much, I know<mask> to make small talk, and in general be able to deal with people. [NEWLINE] [NEWLINE] <mask> for fat acceptance, it may not be an issue in a country<mask> people pay for their own healthcare<mask> in a country like Canada<mask> everyone pays for healthcare, this is an issue that affects everyone. Here's a link that discusses the issue even in the US: [URL] / [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy
Label encoding: <s>There are a lot of movements for accepting previously criticized groups such as fat acceptance, nerd acceptance,... You get the idea. [NEWLINE] [NEWLINE] Sure, those movements make people feel better about themselves but it makes them content with who they are thinking there's nothing wrong with it and not try to improve. There's a reason why they were traditionally thought of as weird. [NEWLINE] [NEWLINE] For example, the fat acceptance movement makes obese people think there is nothing wrong with their bodies and ignore the health risks of being obese. [NEWLINE] [NEWLINE] While some nerds possess moderate social skills, a lot of them don't have social skills good enough to make one friend in an anime convention of thousands of people. According to a poll used for an anime Family Feud in an anime convention, at least 60% make 0 friends in a 3 day convention with people of similar interests. That's absurd. [NEWLINE] [NEWLINE] I can't speak for fat acceptance but as a previous nerd with no social skills, it was the feeling of rejection that made me improve and push myself to hone my social skills. If the acceptance movements were around back then I would be happier as a nerd but I'd still have 0 friends. That's not to say that I am no longer a nerd, I still have all my previous interests but I know when to talk about things, I know not to correct people too much, I know how to make small talk, and in general be able to deal with people. [NEWLINE] [NEWLINE] As for fat acceptance, it may not be an issue in a country where people pay for their own healthcare but in a country like Canada where everyone pays for healthcare, this is an issue that affects everyone. Here's a link that discusses the issue even in the US: [URL] / [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To add to this... [NEWLINE] [NEWLINE] [STARTQ] Second... [ENDQ] [NEWLINE] <mask> there is even one case of someone getting married and understanding this commitment then your point is proven wrong, OP. [NEWLINE] [NEWLINE] For example, I was married at 19. I am now. 21, still happily married, and see no future<mask> I will not be happily married to the same person. We both understand deeply the commitment we made to each other, and have the emotional understanding and maturity to handle it. At any age it CAN be understood.<mask><mask> you shouldn't get into a serious relationship<mask> you don't understand<mask> a serious relationship entails,<mask> isn't that really obvious? That's like saying "This exam on externalities in the tourist economics will be much easier<mask> you understand externalities in tourist economics." Yup, that's pretty clear in the definition. [NEWLINE] [NEWLINE] [STARTQ] Third... High school is often<mask> decides the course of someone's life.. [ENDQ] [NEWLINE] I 100% disagree. People have far too much control over their lives to say any point in everyone's life will control their direction. There are too many outliers to avoid this. Not everyone who drops out of High School leads a poor life. My mum finished it later and got an education in Computer Science and had a job with a Gas company for many years making over $80k a year in the 80s. Adjust for inflation... It's a well paying job even<mask> you don't. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] High School is not often<mask> decides the corse of ones life.<mask> anything, college is. [NEWLINE] [NEWLINE] <mask><mask> literally all of your arguments COULD be transferred to say College or University. And surely you don't think that right? Surely the majority nowadays don't find the person they intend to marry in their first two years of College.<mask><mask> bother? Those first two years you're busy partying, no emotional maturity there. It's a better argument for your third point too, OP. College is more expensive and time consuming<mask> you have less money and time to spend on frivolous relationships. We should just not date Until everything in life is figured out and stable! [NEWLINE] [NEWLINE] Except that's ridiculous. Life is never inexpensive, we never have time, we always have growing up to do, and we are ALWAYS changing paths<mask> we grow and change. I know people who have changed their majors three times. I know people who've dropped out of High School and gone to lead successful lives. [NEWLINE] [NEWLINE] None of OPs arguments are good<mask> it leaves no time to be good
Label encoding: <s>To add to this... [NEWLINE] [NEWLINE] [STARTQ] Second... [ENDQ] [NEWLINE] If there is even one case of someone getting married and understanding this commitment then your point is proven wrong, OP. [NEWLINE] [NEWLINE] For example, I was married at 19. I am now. 21, still happily married, and see no future where I will not be happily married to the same person. We both understand deeply the commitment we made to each other, and have the emotional understanding and maturity to handle it. At any age it CAN be understood. I agree you shouldn't get into a serious relationship if you don't understand what a serious relationship entails, but isn't that really obvious? That's like saying "This exam on externalities in the tourist economics will be much easier when you understand externalities in tourist economics." Yup, that's pretty clear in the definition. [NEWLINE] [NEWLINE] [STARTQ] Third... High school is often what decides the course of someone's life.. [ENDQ] [NEWLINE] I 100% disagree. People have far too much control over their lives to say any point in everyone's life will control their direction. There are too many outliers to avoid this. Not everyone who drops out of High School leads a poor life. My mum finished it later and got an education in Computer Science and had a job with a Gas company for many years making over $80k a year in the 80s. Adjust for inflation... It's a well paying job even if you don't. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] High School is not often what decides the corse of ones life. If anything, college is. [NEWLINE] [NEWLINE] In fact literally all of your arguments COULD be transferred to say College or University. And surely you don't think that right? Surely the majority nowadays don't find the person they intend to marry in their first two years of College. So why bother? Those first two years you're busy partying, no emotional maturity there. It's a better argument for your third point too, OP. College is more expensive and time consuming so you have less money and time to spend on frivolous relationships. We should just not date Until everything in life is figured out and stable! [NEWLINE] [NEWLINE] Except that's ridiculous. Life is never inexpensive, we never have time, we always have growing up to do, and we are ALWAYS changing paths as we grow and change. I know people who have changed their majors three times. I know people who've dropped out of High School and gone to lead successful lives. [NEWLINE] [NEWLINE] None of OPs arguments are good because it leaves no time to be good
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] ethnic minorities cannot discriminate<mask> of their skin color and/or nationality. [ENDQ] [NEWLINE] Not quite.<mask><mask> you may have misunderstood one or more of the arguments that have lead some to suggest that minorities can't be  racist or practice racism. *Anyone* can discriminate or act bigoted or discriminatory.<mask> from a sociological standpoint, *racism* is more than simple discrimination. [NEWLINE] [NEWLINE] Sure: some will quote dictionary.com and say that 'racism' is any discrimination or bigotry based on race.<mask> that 'by the book' definition doesn't really help us address root causes of societal inequality. There's a sketch on the TV show *Upright Citizens Brigade* in which a racist character looks to find a group to direct his racism toward that won't be damaging and unacceptable and settles on the Laplanders (residents of the extreme north of Norway, Sweden, and Finland - about 200,000 people). The joke is that a white, middle-American guy being racist toward Laplanders is almost<mask> good<mask> being non-racist: in that his racism will be very unlikely to affect anyone in his daily life (of course,<mask> this is comedy, he suddenly encounters Laplanders everywhere and is forced to again confront his racism). [NEWLINE] [NEWLINE] A member of a minority group being 'racist' (bigoted or discriminatory) toward white people is just like a white guy discriminating against Laplanders. On the one hand -society's general preference for white people makes his discrimination more or less non-effective.<mask> few minorities are in positions of authority<mask> they could even effectively discriminate against majority groups; fewer still will achieve those positions by being overtly discriminatory against the majority. Only ultra-insular communities would ever permit a minority person to attain a rank sufficient to discriminate against majority persons: think ultra-orthodox religious communities or radical racial groups. Mainstream whites aren't trying to join Hasidic Jewish synagogues or the Black Panthers any more than blacks are trying to join the KKK. [NEWLINE] [NEWLINE] Sure: excluding *anyone* from *any* group based on race is the dictionary definition of 'racism'.<mask> you have to ask yourself - does excluding members of a majority group from insular, race-based identification societies damage those excluded in any way? Aside from hurting their feelings, are members of the majority racial or ethnic groups damaged in a systematic fashion by bigotry and discrimination by members of minority groups?<mask> not, I don't think that 'racism' is a correct label to
Label encoding: <s> [STARTQ] ethnic minorities cannot discriminate because of their skin color and/or nationality. [ENDQ] [NEWLINE] Not quite. I think you may have misunderstood one or more of the arguments that have lead some to suggest that minorities can't be  racist or practice racism. *Anyone* can discriminate or act bigoted or discriminatory. But from a sociological standpoint, *racism* is more than simple discrimination. [NEWLINE] [NEWLINE] Sure: some will quote dictionary.com and say that 'racism' is any discrimination or bigotry based on race. But that 'by the book' definition doesn't really help us address root causes of societal inequality. There's a sketch on the TV show *Upright Citizens Brigade* in which a racist character looks to find a group to direct his racism toward that won't be damaging and unacceptable and settles on the Laplanders (residents of the extreme north of Norway, Sweden, and Finland - about 200,000 people). The joke is that a white, middle-American guy being racist toward Laplanders is almost as good as being non-racist: in that his racism will be very unlikely to affect anyone in his daily life (of course, because this is comedy, he suddenly encounters Laplanders everywhere and is forced to again confront his racism). [NEWLINE] [NEWLINE] A member of a minority group being 'racist' (bigoted or discriminatory) toward white people is just like a white guy discriminating against Laplanders. On the one hand -society's general preference for white people makes his discrimination more or less non-effective. So few minorities are in positions of authority where they could even effectively discriminate against majority groups; fewer still will achieve those positions by being overtly discriminatory against the majority. Only ultra-insular communities would ever permit a minority person to attain a rank sufficient to discriminate against majority persons: think ultra-orthodox religious communities or radical racial groups. Mainstream whites aren't trying to join Hasidic Jewish synagogues or the Black Panthers any more than blacks are trying to join the KKK. [NEWLINE] [NEWLINE] Sure: excluding *anyone* from *any* group based on race is the dictionary definition of 'racism'. But you have to ask yourself - does excluding members of a majority group from insular, race-based identification societies damage those excluded in any way? Aside from hurting their feelings, are members of the majority racial or ethnic groups damaged in a systematic fashion by bigotry and discrimination by members of minority groups? If not, I don't think that 'racism' is a correct label to
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Overrated isn't a criticism of media. It doesn't say anything about<mask> good a piece of media is,<mask> it excels or falters, or really anything. [NEWLINE] [NEWLINE] <mask> it is a criticism of the public's response to that media. Overrated is saying that most people rate something more highly than it deserves. This can be useful<mask> the critic saying this is someone whose opinion you respect,<mask> it gives you more insight into the public perception and<mask> helps you parse common reactions. [NEWLINE] [NEWLINE] To that end calling something overrated can be very useful, especially<mask> there's a solid explanation alongside the claim. For example I'll use the anime Attack on Titan (AoT). AoT was extremely popular among newer fans to the medium and in the very large 13-25 year old male demographic.<mask> you listened to the average criticism you would get the impression that it's an extremely highly acclaimed show - one of the very best. [NEWLINE] [NEWLINE] This is<mask> calling something overrated can be useful.<mask> someone was looking for a new show to watch and saw all the great rating AoT gets they might decide to check it. Someone explaining that the show is overrated<mask> it's aimed at a very large demographic (that is<mask> known for severe ratings - "best thing ever!" / "worst thing ever!") can help give some context and perspective. At that point<mask> you like the genre/premise and it still seems interesting you can watch it with more realistic expectations, and<mask> you aren't a huge fan of the genre/premise you can decide not to watch it<mask> it's not actually mindblowingly good. [NEWLINE] [NEWLINE] A good contrast to this would be the anime Madoka Magica (Madoka). This is an anime that isn't in a very popular genre and, at first glance, doesn't seem that interesting.<mask>, it recieved incredibly reviews and nearly everyone agreed that it was amazing.<mask> people agreed that it wasn't overrated, it would be something worth checking out even<mask> you're not a large fan of the genre/premise. [NEWLINE] [NEWLINE] In short, overrated<mask> a criticism just helps you understand people's response. It doesn't say anything about a piece of media directly,<mask> it does help you understand it better by giving you a better perspective. It's kind of like looking at a shadow to figure out<mask> something looks like - it's not always correct and it's not the same<mask> looking at it directly,<mask> it's a potentially useful piece of additional
Label encoding: <s>Overrated isn't a criticism of media. It doesn't say anything about how good a piece of media is, where it excels or falters, or really anything. [NEWLINE] [NEWLINE] But it is a criticism of the public's response to that media. Overrated is saying that most people rate something more highly than it deserves. This can be useful when the critic saying this is someone whose opinion you respect, because it gives you more insight into the public perception and also helps you parse common reactions. [NEWLINE] [NEWLINE] To that end calling something overrated can be very useful, especially if there's a solid explanation alongside the claim. For example I'll use the anime Attack on Titan (AoT). AoT was extremely popular among newer fans to the medium and in the very large 13-25 year old male demographic. If you listened to the average criticism you would get the impression that it's an extremely highly acclaimed show - one of the very best. [NEWLINE] [NEWLINE] This is where calling something overrated can be useful. If someone was looking for a new show to watch and saw all the great rating AoT gets they might decide to check it. Someone explaining that the show is overrated because it's aimed at a very large demographic (that is also known for severe ratings - "best thing ever!" / "worst thing ever!") can help give some context and perspective. At that point if you like the genre/premise and it still seems interesting you can watch it with more realistic expectations, and if you aren't a huge fan of the genre/premise you can decide not to watch it because it's not actually mindblowingly good. [NEWLINE] [NEWLINE] A good contrast to this would be the anime Madoka Magica (Madoka). This is an anime that isn't in a very popular genre and, at first glance, doesn't seem that interesting. However, it recieved incredibly reviews and nearly everyone agreed that it was amazing. Because people agreed that it wasn't overrated, it would be something worth checking out even if you're not a large fan of the genre/premise. [NEWLINE] [NEWLINE] In short, overrated as a criticism just helps you understand people's response. It doesn't say anything about a piece of media directly, but it does help you understand it better by giving you a better perspective. It's kind of like looking at a shadow to figure out what something looks like - it's not always correct and it's not the same as looking at it directly, but it's a potentially useful piece of additional
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is taken from another thread that answers the question: [NEWLINE] [NEWLINE] **"<mask> is Kanye considered to be<mask> great."** [NEWLINE] [NEWLINE] "The first and biggest reason is his discography. Not only does he have arguably the greatest discography in hip-hop<mask> it<mask> is among the best in the entire world of music. His first five solo albums are considered classics among hip hop heads everywhere and<mask> his Yeezus got mixed responses, it is on its way to being a classic<mask> well. His fifth studio album, MBDTF is considered by some to be the greatest hip-hop album of all time and is personally mine and many others favorite album ever. [NEWLINE] [NEWLINE] Not only does he have solo album success<mask> his collab album with Jay-Z (Watch the Throne) was generally well received. His other collab album Cruel Summer,<mask> not up to his normal standards had a few big hits, including Mercy, undoubtedly the biggest song of last summer. [NEWLINE] [NEWLINE] His production work is legendary<mask> well.<mask><mask> to producing the majority of his own body of work, he<mask> produced many hits for Jay-Z like "Lucifer", "Heart of the City", and "Izzo(H.O.V.A)" among many others. He<mask> produced the entirety of Common's album Be, which many consider to be Common's greatest. [NEWLINE] [NEWLINE] His lyrics are often clever, and<mask> some may be considered "nothing special" especially on Yeezus, many times he speaks about meaningful and personal topics like on "All Falls Down",  "Hey Mama", "Everything I Am", "Welcome to Heartbreak", "Gorgeous, and "New Slaves" (to name only one from each album.) [NEWLINE] [NEWLINE] This only scratches the surface of the greatness that is Kanye West. Kanye significantly influenced hip-hop by breaking down the gangsta image of rappers. Each album has a different sound, and often lead the way in breaking down the sonic boundaries of hip-hop like on 808s and Heartbreak and Yeezus. He is a commercial and critical success, and has influenced the majority of rappers on the scene today. [NEWLINE] [NEWLINE] <mask> many people see<mask> egotistical and arrogant, his fans see<mask> self confidence. This self confidence infects us through his music and encourages us to believe in ourselves. He dreams big and in turn inspires his listeners to do the same. He has had a significant and meaningful impact on my life, and many others<mask> well."</s>
Label encoding: <s>This is taken from another thread that answers the question: [NEWLINE] [NEWLINE] **" Why is Kanye considered to be so great."** [NEWLINE] [NEWLINE] "The first and biggest reason is his discography. Not only does he have arguably the greatest discography in hip-hop but it also is among the best in the entire world of music. His first five solo albums are considered classics among hip hop heads everywhere and although his Yeezus got mixed responses, it is on its way to being a classic as well. His fifth studio album, MBDTF is considered by some to be the greatest hip-hop album of all time and is personally mine and many others favorite album ever. [NEWLINE] [NEWLINE] Not only does he have solo album success but his collab album with Jay-Z (Watch the Throne) was generally well received. His other collab album Cruel Summer, although not up to his normal standards had a few big hits, including Mercy, undoubtedly the biggest song of last summer. [NEWLINE] [NEWLINE] His production work is legendary as well. In addition to producing the majority of his own body of work, he also produced many hits for Jay-Z like "Lucifer", "Heart of the City", and "Izzo(H.O.V.A)" among many others. He also produced the entirety of Common's album Be, which many consider to be Common's greatest. [NEWLINE] [NEWLINE] His lyrics are often clever, and while some may be considered "nothing special" especially on Yeezus, many times he speaks about meaningful and personal topics like on "All Falls Down",  "Hey Mama", "Everything I Am", "Welcome to Heartbreak", "Gorgeous, and "New Slaves" (to name only one from each album.) [NEWLINE] [NEWLINE] This only scratches the surface of the greatness that is Kanye West. Kanye significantly influenced hip-hop by breaking down the gangsta image of rappers. Each album has a different sound, and often lead the way in breaking down the sonic boundaries of hip-hop like on 808s and Heartbreak and Yeezus. He is a commercial and critical success, and has influenced the majority of rappers on the scene today. [NEWLINE] [NEWLINE] What many people see as egotistical and arrogant, his fans see as self confidence. This self confidence infects us through his music and encourages us to believe in ourselves. He dreams big and in turn inspires his listeners to do the same. He has had a significant and meaningful impact on my life, and many others as well."</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to assume you don't have kids. I would have said the exact same thing until I had one. Mine is 13 months. Just learning to walk. Once they start walking, it's completely out of control. They are constantly zipping off in every direction. And<mask> they do, they are going to want to touch everything. And rip those things off the shelves. And once that stuff is on the floor, they are going to want to see just<mask> far they can scatter that shit. Cause that's very amusing to a young child. Hell, it's amusing to me too. Except, now I've got to be the one to clean it up. [NEWLINE] [NEWLINE] A kid is not going to hold his/her parent's hand for any stretch of time. He's going to scream bloody murder. He's going to throw a damn fit. Then you're going to say, "ok, just stay right over there."<mask> of course he won't. He's gone. Try shopping for food, or whatever, and having to deal with that. It doesn't happen. A parent in a store with a young kid has one of two options: 1) Shop, which is<mask> you went to the store for, or; 2) Watch<mask> the kid is doing. There is no in between ground.<mask> they are very little, you can strap them into the cart.<mask> they are too big for that, all bets are off. [NEWLINE] [NEWLINE] <mask>, it is simply a common courtesy to others. A loose, wandering 4 year old is going to annoy the shit out of everybody else nearby: climbing on their carts, pulling on their clothes, getting in their way, etc. You would be extremely annoyed<mask> someone did this to you,<mask> you criticize the means by which to restrain this activity. [NEWLINE] [NEWLINE] Finally, safety. I'm not one of those who think that everyone in the world is out to kidnap my kid. They aren't.<mask> some people would. Even<mask> nobody would, it is going to involve me having to go to security and have them help search out<mask> the hell this kid went to. [NEWLINE] [NEWLINE] Conclusion:<mask> I can keep a lengthy, 20 ft. rope tied to my kid, its the best of all worlds. That guy can wander 20 ft in any direction; he may explore whatever he chooses; he can enjoy 20 ft of independence. And I can get shit done without having to worry that DCF will be called. </s>
Label encoding: <s>I'm going to assume you don't have kids. I would have said the exact same thing until I had one. Mine is 13 months. Just learning to walk. Once they start walking, it's completely out of control. They are constantly zipping off in every direction. And when they do, they are going to want to touch everything. And rip those things off the shelves. And once that stuff is on the floor, they are going to want to see just how far they can scatter that shit. Cause that's very amusing to a young child. Hell, it's amusing to me too. Except, now I've got to be the one to clean it up. [NEWLINE] [NEWLINE] A kid is not going to hold his/her parent's hand for any stretch of time. He's going to scream bloody murder. He's going to throw a damn fit. Then you're going to say, "ok, just stay right over there." But of course he won't. He's gone. Try shopping for food, or whatever, and having to deal with that. It doesn't happen. A parent in a store with a young kid has one of two options: 1) Shop, which is what you went to the store for, or; 2) Watch what the kid is doing. There is no in between ground. When they are very little, you can strap them into the cart. When they are too big for that, all bets are off. [NEWLINE] [NEWLINE] Additionally, it is simply a common courtesy to others. A loose, wandering 4 year old is going to annoy the shit out of everybody else nearby: climbing on their carts, pulling on their clothes, getting in their way, etc. You would be extremely annoyed if someone did this to you, yet you criticize the means by which to restrain this activity. [NEWLINE] [NEWLINE] Finally, safety. I'm not one of those who think that everyone in the world is out to kidnap my kid. They aren't. But some people would. Even if nobody would, it is going to involve me having to go to security and have them help search out where the hell this kid went to. [NEWLINE] [NEWLINE] Conclusion: If I can keep a lengthy, 20 ft. rope tied to my kid, its the best of all worlds. That guy can wander 20 ft in any direction; he may explore whatever he chooses; he can enjoy 20 ft of independence. And I can get shit done without having to worry that DCF will be called. </s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I will tackle the IUD part. [NEWLINE] [NEWLINE] The reason they don't want to cover the IUD is<mask> in the event that you do get pregnant on the IUD, it could potentially cause a spontaneous miscarriage, due to a piece of metal/plastic being in the uterus<mask> the baby is growing. This,<mask><mask> Hobby Lobby, is an abortion,<mask> I do argue with their reasoning behind that. [NEWLINE] [NEWLINE] <mask> you look at it, getting pregnant in general can cause a spontaneous miscarriage,<mask> they are trying to eliminate all the things that can cause a miscarriage, then they should just tell people to not get pregnant at all,<mask> honestly every pregnancy could potentially result in a miscarriage. [NEWLINE] [NEWLINE] With Hobby Lobby not wanting to cover specific medications due to their religious views, its opening up an entire can of worms for other religiously owned companies.<mask><mask><mask>, health care should cover every aspect of a persons health, up to and including the morning after pill, which is not an abortion pill, its almost like a hyper dose of birth control that keeps an egg from attaching, which some hormonal birth controls do anyways,<mask> Hobby Lobby does cover them. [NEWLINE] [NEWLINE] Now, this is<mask> its going to start getting sticky, they are saying that every child deserves to not be aborted, that is their whole stance on this, pro-life.<mask>, their health insurance will continue to cover viagra for men. You see, men don't have to worry about getting pregnant,<mask> they can get boners and sleep with all the women they want. Now, honestly,<mask> god wanted a mans penis to work after a certain age, he wouldn't have allowed Erectile dysfunction. I don't think that health care should cover that,<mask> an erection and sex is a privilege, they should have to pay out of their own pocket. [NEWLINE] [NEWLINE] Bit silly don't you think? [NEWLINE] [NEWLINE] Then you have Jehovah witnesses, they don't believe in blood transfusions, it is forbidden by their religion.<mask><mask> they try to fight to not have to cover that<mask> their religion doesn't believe in it. [NEWLINE] [NEWLINE] In the end it really isn't about just pro-life vs pro-choice, its now a battle of which religion will chose to opt out of medical coverage based upon x-reasoning. Plan-B could potentially be just the beginning. [NEWLINE] [NEWLINE] Sorry, I got a bit rambly, let me know<mask> something I said didn't make any sense. </s>
Label encoding: <s>I will tackle the IUD part. [NEWLINE] [NEWLINE] The reason they don't want to cover the IUD is because in the event that you do get pregnant on the IUD, it could potentially cause a spontaneous miscarriage, due to a piece of metal/plastic being in the uterus where the baby is growing. This, according to Hobby Lobby, is an abortion, though I do argue with their reasoning behind that. [NEWLINE] [NEWLINE] If you look at it, getting pregnant in general can cause a spontaneous miscarriage, if they are trying to eliminate all the things that can cause a miscarriage, then they should just tell people to not get pregnant at all, because honestly every pregnancy could potentially result in a miscarriage. [NEWLINE] [NEWLINE] With Hobby Lobby not wanting to cover specific medications due to their religious views, its opening up an entire can of worms for other religiously owned companies. In my opinion, health care should cover every aspect of a persons health, up to and including the morning after pill, which is not an abortion pill, its almost like a hyper dose of birth control that keeps an egg from attaching, which some hormonal birth controls do anyways, but Hobby Lobby does cover them. [NEWLINE] [NEWLINE] Now, this is where its going to start getting sticky, they are saying that every child deserves to not be aborted, that is their whole stance on this, pro-life. But, their health insurance will continue to cover viagra for men. You see, men don't have to worry about getting pregnant, so they can get boners and sleep with all the women they want. Now, honestly, if god wanted a mans penis to work after a certain age, he wouldn't have allowed Erectile dysfunction. I don't think that health care should cover that, because an erection and sex is a privilege, they should have to pay out of their own pocket. [NEWLINE] [NEWLINE] Bit silly don't you think? [NEWLINE] [NEWLINE] Then you have Jehovah witnesses, they don't believe in blood transfusions, it is forbidden by their religion. What if they try to fight to not have to cover that because their religion doesn't believe in it. [NEWLINE] [NEWLINE] In the end it really isn't about just pro-life vs pro-choice, its now a battle of which religion will chose to opt out of medical coverage based upon x-reasoning. Plan-B could potentially be just the beginning. [NEWLINE] [NEWLINE] Sorry, I got a bit rambly, let me know if something I said didn't make any sense. </s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I can't race my horse down a highway<mask> that'd be dangerous to myself and those around me. [ENDQ] [NEWLINE] In all honesty it's really just a danger to yourself and incompetent drivers. Horses racing down a highway are really only a danger to other drivers<mask> they don't know<mask> to avoid moving objects (in which case, they shouldn't be driving in the first place). I'd have no problem<mask> someone wanted to ride their horse down the 401,<mask> it's a risk that you took, especially<mask> your horse gets started and knocks you off its back. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask> good of a driver you are, you aren't better than a centrally controlled system that efficiently routes vehicles... able to drive your own vehicle will have to go by the wayside for all<mask> an emergency option to ensure our modes of transportation can be efficient. [ENDQ] [NEWLINE] Of course I'm not,<mask><mask> exactly do you expect people to operate a vehicle in an emergency later down the road<mask> they have no experience doing<mask>? That sounds a lot more dangerous to me than<mask> only the dangerous, inexperienced drivers were forced to use self-driving cars.<mask> here in Toronto we can't even get our traffic lights in sync.<mask> makes you think we'd be able to get a system that efficient routes vehicles going to hundred of iterations of different locations at different times? [NEWLINE] [NEWLINE] [STARTQ] <mask> your issue is that you'll continue to pay taxes for a system that you won't be using, getting a more modern transportation system in place actually benefits you. [ENDQ] [NEWLINE] Personally<mask><mask> there are more ridiculous uses of my tax dollars being spent by the government than road repair. For instance, I'd rather everyone pay for an equal share for road costs,<mask> only individuals in support of providing welfare to those that are not physically or mentally disabled should pay taxes to go towards that,<mask> that's a different conversation for a different thread. [NEWLINE] [NEWLINE] [STARTQ] The only way to CMV here is to demonstrate that the positives vastly outweigh the negatives. [ENDQ] [NEWLINE] Aside from a traffic routing system that we have no proof of even working effectively and decreased human error (<mask> the fact that humans in the future won't even know<mask> to operate their cars in the case of an emergency) I fail to see<mask> these positives you mention are. In my eyes,<mask> I've never caused an accident or jeopardized a human life<mask> driving, society isn't any more safer<mask> I'm driving or<mask> a computer is. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I can't race my horse down a highway because that'd be dangerous to myself and those around me. [ENDQ] [NEWLINE] In all honesty it's really just a danger to yourself and incompetent drivers. Horses racing down a highway are really only a danger to other drivers if they don't know how to avoid moving objects (in which case, they shouldn't be driving in the first place). I'd have no problem if someone wanted to ride their horse down the 401, but it's a risk that you took, especially when your horse gets started and knocks you off its back. [NEWLINE] [NEWLINE] [STARTQ] Regardless of how good of a driver you are, you aren't better than a centrally controlled system that efficiently routes vehicles... able to drive your own vehicle will have to go by the wayside for all but an emergency option to ensure our modes of transportation can be efficient. [ENDQ] [NEWLINE] Of course I'm not, but how exactly do you expect people to operate a vehicle in an emergency later down the road if they have no experience doing so? That sounds a lot more dangerous to me than if only the dangerous, inexperienced drivers were forced to use self-driving cars. Besides here in Toronto we can't even get our traffic lights in sync. What makes you think we'd be able to get a system that efficient routes vehicles going to hundred of iterations of different locations at different times? [NEWLINE] [NEWLINE] [STARTQ] If your issue is that you'll continue to pay taxes for a system that you won't be using, getting a more modern transportation system in place actually benefits you. [ENDQ] [NEWLINE] Personally I think there are more ridiculous uses of my tax dollars being spent by the government than road repair. For instance, I'd rather everyone pay for an equal share for road costs, but only individuals in support of providing welfare to those that are not physically or mentally disabled should pay taxes to go towards that, but that's a different conversation for a different thread. [NEWLINE] [NEWLINE] [STARTQ] The only way to CMV here is to demonstrate that the positives vastly outweigh the negatives. [ENDQ] [NEWLINE] Aside from a traffic routing system that we have no proof of even working effectively and decreased human error ( despite the fact that humans in the future won't even know how to operate their cars in the case of an emergency) I fail to see what these positives you mention are. In my eyes, because I've never caused an accident or jeopardized a human life while driving, society isn't any more safer if I'm driving or if a computer is. [NEWLINE] </s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the main ax to grind is the negative historical narrative associated with things that are "appropriated". For instance, large lips have been mocked in many aspects of entertainment historically (particularly the 18th century), most notably minstrelsy and "black face". There are<mask> images drawn noting the differences in the facial features of blacks with comparison to whites, many of which exaggerate and mock the large lips, large noses, large rear-ends and "nappy hair" of blacks. (there is a particular image that comes to mind that has been widely used to illustrate this concept,<mask> I cannot seem to find it).<mask>, it seems that the very things that are mocked on black bodies, are praised on white bodies. In some instances, the very thing that is mocked on black bodies is praised<mask> being "originated" by a given white individual. Black people have always had big lips. There is nothing inherently new about big lips. Kylie Jenner did not "start" big lips. [NEWLINE] [NEWLINE] Basically, big lips, large asses, hip hop..none of these things BELONG to blacks.<mask>,<mask> they are deemed "less than" on black bodies, or "ghetto" in black society and then praised on white bodies or accepted by the greater white population just<mask> they are being fed by a white person (see kylie jenner's lips and iggy azalea's "rapping") there is an undercurrent of disrespect. It leaves blacks in a place of, "wait I thought big lips were funny? They're 'in' now?<mask> changed?" Just under the surface there is the whole "they love being black<mask> would hate being black" level of discourse that reinforces that your bodies are not ok, your BLACKNESS is not ok, and<mask> of that Kylie Jenner's big lips are beautiful<mask> they are on a white body. [NEWLINE] [NEWLINE] This is just my interpretation of<mask> the larger discourse is, I may be wrong,<mask><mask><mask><mask> there was more respect for the originating culture, and/or less criticism for the same characteristics that are praised on white bodies, it would be less of an issue.  Basically<mask> everyone stopped saying that iggy azalea brought big asses back and kylie jenner started big lips no one would have a problem.<mask><mask> I've stated, I may be wrong...this is a large concept that has no real "answer" per se. [NEWLINE] </s>
Label encoding: <s>I think the main ax to grind is the negative historical narrative associated with things that are "appropriated". For instance, large lips have been mocked in many aspects of entertainment historically (particularly the 18th century), most notably minstrelsy and "black face". There are also images drawn noting the differences in the facial features of blacks with comparison to whites, many of which exaggerate and mock the large lips, large noses, large rear-ends and "nappy hair" of blacks. (there is a particular image that comes to mind that has been widely used to illustrate this concept, but I cannot seem to find it). However, it seems that the very things that are mocked on black bodies, are praised on white bodies. In some instances, the very thing that is mocked on black bodies is praised as being "originated" by a given white individual. Black people have always had big lips. There is nothing inherently new about big lips. Kylie Jenner did not "start" big lips. [NEWLINE] [NEWLINE] Basically, big lips, large asses, hip hop..none of these things BELONG to blacks. However, when they are deemed "less than" on black bodies, or "ghetto" in black society and then praised on white bodies or accepted by the greater white population just because they are being fed by a white person (see kylie jenner's lips and iggy azalea's "rapping") there is an undercurrent of disrespect. It leaves blacks in a place of, "wait I thought big lips were funny? They're 'in' now? What changed?" Just under the surface there is the whole "they love being black but would hate being black" level of discourse that reinforces that your bodies are not ok, your BLACKNESS is not ok, and because of that Kylie Jenner's big lips are beautiful because they are on a white body. [NEWLINE] [NEWLINE] This is just my interpretation of what the larger discourse is, I may be wrong, but I think if there was more respect for the originating culture, and/or less criticism for the same characteristics that are praised on white bodies, it would be less of an issue.  Basically if everyone stopped saying that iggy azalea brought big asses back and kylie jenner started big lips no one would have a problem. But as I've stated, I may be wrong...this is a large concept that has no real "answer" per se. [NEWLINE] </s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, I'm Canadian and this post is specifically talking about Canada.<mask> you're not familiar with our politics this probably isn't very relevant to you. [NEWLINE] [NEWLINE] This idea was inspired by an incident that occurred about fifteen years ago,<mask> the mayor of my city (Toronto) called in soldiers to help shovel snow and use their armoured vehicles to clear roads for emergency services after a severe blizzard which trapped many people indoors and shut down the city's streets. He was much ridiculed for this decision<mask> it was seen<mask> a waste of resources. [NEWLINE] [NEWLINE] <mask><mask> Canada's army is an underutilized resource. We're already paying for their room and board, plus training and equipment and<mask> on,<mask> they wait to defend us<mask> necessary.<mask> we have things like old bridges that desperately need repair, terrible roads in the Far North, and<mask> on.<mask> can't the trained, physically strong people who have volunteered to serve their country be used for infrastructure work? Or going even further,<mask> can't the medical airlift teams that saved people wounded by IEDs in Afghanistan be used to save people who have heart attacks in their remote village in Nunavut? I'm sure there are more things than that that the military could do in peacetime,<mask> the ones I gave are just examples. [NEWLINE] [NEWLINE] <mask><mask> a lot of people in Canada have an instinctive fear of the military, especially left-wing people (and I assure you I'm very left-wing)<mask> they associate it with Americanism and warmongering,<mask><mask><mask> that fear could be allayed<mask> they were used in the manner I propose. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>First off, I'm Canadian and this post is specifically talking about Canada. If you're not familiar with our politics this probably isn't very relevant to you. [NEWLINE] [NEWLINE] This idea was inspired by an incident that occurred about fifteen years ago, where the mayor of my city (Toronto) called in soldiers to help shovel snow and use their armoured vehicles to clear roads for emergency services after a severe blizzard which trapped many people indoors and shut down the city's streets. He was much ridiculed for this decision because it was seen as a waste of resources. [NEWLINE] [NEWLINE] I think Canada's army is an underutilized resource. We're already paying for their room and board, plus training and equipment and so on, while they wait to defend us if necessary. Meanwhile we have things like old bridges that desperately need repair, terrible roads in the Far North, and so on. Why can't the trained, physically strong people who have volunteered to serve their country be used for infrastructure work? Or going even further, why can't the medical airlift teams that saved people wounded by IEDs in Afghanistan be used to save people who have heart attacks in their remote village in Nunavut? I'm sure there are more things than that that the military could do in peacetime, but the ones I gave are just examples. [NEWLINE] [NEWLINE] I think a lot of people in Canada have an instinctive fear of the military, especially left-wing people (and I assure you I'm very left-wing) because they associate it with Americanism and warmongering, but I think that fear could be allayed if they were used in the manner I propose. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>WARNING Princess Mononoke Spoilers ahead!!! [NEWLINE] [NEWLINE] Which films have you watched? I would say my two favorites are Spirited Away and Princess Mononoke. With all of their films you have to first accept the fantastical world in which they live. There are ghosts, creatures, and spells. I'll try and give you a run down for<mask> I love about Mononoke aside from the music and artwork (which<mask><mask>, are fantastic). I was going to do both<mask> this one ended up long enough. [NEWLINE] [NEWLINE] Princess Mononoke is a real out and out fantasty/action movie. It has very strong environmental themes, anti-war themes, and a number of wonderfully grey characters. Aside from Ashitaka everyone is flawed, proud, and stuck thinking that their way is the only way. No character is all evil or all good. There are strong women and men and each side truly believes in its cause. [NEWLINE] [NEWLINE] The entire opening sequence is excellent. Ashitaka, the prince of a hidden and ancient people, spies something huge and evil coming towards his town. Him and his badass elk Yakul chase the thing down and kill it before it can harm the village<mask> at his own sacrifice. He must leave and try and find a cure to the evil the beast poisoned him with. That evil (stemming from a bullet) is at the base of this movie. All at once it causes great suffering,<mask> it<mask> gives Ashitaka power and strength. [NEWLINE] [NEWLINE] There follows a long excellent journey<mask> he meets various people. It all culminates in the boars' suicide attack and the attempted killing of the forest god, a creature that is perfectly loved and serene. I don't know<mask> you've watched this movie,<mask> every time this scene comes around it affects me. It comes down to a war between nature and industry and<mask> the heroes of nature and torn down in their folly you can feel<mask> it affects many of the characters involved. This got a bit rambly and I have to run, might look it back over later. [NEWLINE] [NEWLINE] <mask> a quick note Spirited Away has a number of big themes<mask> well<mask> being a great film for children and adults (like the Pixar greats) and a stellar and compelling main character. Is it number 10 of all time like the list today on /r/movies put it? I don't know.<mask> I do think they're both excellent.</s>
Label encoding: <s>WARNING Princess Mononoke Spoilers ahead!!! [NEWLINE] [NEWLINE] Which films have you watched? I would say my two favorites are Spirited Away and Princess Mononoke. With all of their films you have to first accept the fantastical world in which they live. There are ghosts, creatures, and spells. I'll try and give you a run down for what I love about Mononoke aside from the music and artwork (which I agree, are fantastic). I was going to do both but this one ended up long enough. [NEWLINE] [NEWLINE] Princess Mononoke is a real out and out fantasty/action movie. It has very strong environmental themes, anti-war themes, and a number of wonderfully grey characters. Aside from Ashitaka everyone is flawed, proud, and stuck thinking that their way is the only way. No character is all evil or all good. There are strong women and men and each side truly believes in its cause. [NEWLINE] [NEWLINE] The entire opening sequence is excellent. Ashitaka, the prince of a hidden and ancient people, spies something huge and evil coming towards his town. Him and his badass elk Yakul chase the thing down and kill it before it can harm the village but at his own sacrifice. He must leave and try and find a cure to the evil the beast poisoned him with. That evil (stemming from a bullet) is at the base of this movie. All at once it causes great suffering, but it also gives Ashitaka power and strength. [NEWLINE] [NEWLINE] There follows a long excellent journey where he meets various people. It all culminates in the boars' suicide attack and the attempted killing of the forest god, a creature that is perfectly loved and serene. I don't know if you've watched this movie, but every time this scene comes around it affects me. It comes down to a war between nature and industry and as the heroes of nature and torn down in their folly you can feel how it affects many of the characters involved. This got a bit rambly and I have to run, might look it back over later. [NEWLINE] [NEWLINE] As a quick note Spirited Away has a number of big themes as well as being a great film for children and adults (like the Pixar greats) and a stellar and compelling main character. Is it number 10 of all time like the list today on /r/movies put it? I don't know. But I do think they're both excellent.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] To say that cultures *should* act in a way that is consistent with themselves is to make an ethical claim- that any culture is right in being the way it is.  The corollary being that it is wrong to enforce one cultural norm on another culture. [ENDQ] [NEWLINE] [STARTQ] <mask> suppose I belong to an imperialist culture that,<mask> a part of its culture, defines the good and spreads it at the end of a gun?  Then I am only acting in accord with my culture, in which case neither I or my culture is wrong. <mask> look! I'm forcibly dismantling the cultural norms of other cultures in my imperialism. [ENDQ] [NEWLINE] [STARTQ] That's a conflict that normative cultural relativism doesn't address, and is powerless to address. [ENDQ] [NEWLINE] Obviously everyone is acting<mask><mask> their own values and will attempt to act in accordance with their values. I dislike it<mask> children are married. I empathize with them.<mask> I will act in accordance with my preferences and work to end child marriage in Yemen.<mask>, I recognize that my preferences are my preferences. There is nothing written in the sky which tells us that certain preferences are objectively superior to others. All I know is that I have preferences and I will act in accordance with them. I do not presume that nature happens to conform to my beliefs. [NEWLINE] [NEWLINE] <mask> some culture seeks to spread itself through the gun I don't like it. That doesn't mean that isn't objectively wrong. [NEWLINE] [NEWLINE] [STARTQ] Further, there are absurd conclusions we must face in accepting cultural relativism.  The x-people live next to the y-people under nearly identical conditions.  The x-people castrate and blind one in three children, the y-people do not.  Are the x-people wrong in committing this practice?<mask> you think<mask>, you're not really a cultural relativist. [ENDQ] [NEWLINE] No, nobody is 'wrong' in any practice. There is no right and wrong.<mask> you believe that there is then prove it. Show some evidence that a practice is wrong. And before you answer, no, 'intuition' is not objective evidence<mask> different people in different cultures have different intuitions. [NEWLINE] [NEWLINE] I would work to stop blinding children in other countries<mask> I am made sad by other people suffering, and I work to make myself happy. That's the point of my life. I do not pretend that I have special godlike knowledge of<mask> is objectively right and wrong.</s>
Label encoding: <s> [STARTQ] To say that cultures *should* act in a way that is consistent with themselves is to make an ethical claim- that any culture is right in being the way it is.  The corollary being that it is wrong to enforce one cultural norm on another culture. [ENDQ] [NEWLINE] [STARTQ] But suppose I belong to an imperialist culture that, as a part of its culture, defines the good and spreads it at the end of a gun?  Then I am only acting in accord with my culture, in which case neither I or my culture is wrong.  But look! I'm forcibly dismantling the cultural norms of other cultures in my imperialism. [ENDQ] [NEWLINE] [STARTQ] That's a conflict that normative cultural relativism doesn't address, and is powerless to address. [ENDQ] [NEWLINE] Obviously everyone is acting according to their own values and will attempt to act in accordance with their values. I dislike it when children are married. I empathize with them. Therefore I will act in accordance with my preferences and work to end child marriage in Yemen. However, I recognize that my preferences are my preferences. There is nothing written in the sky which tells us that certain preferences are objectively superior to others. All I know is that I have preferences and I will act in accordance with them. I do not presume that nature happens to conform to my beliefs. [NEWLINE] [NEWLINE] When some culture seeks to spread itself through the gun I don't like it. That doesn't mean that isn't objectively wrong. [NEWLINE] [NEWLINE] [STARTQ] Further, there are absurd conclusions we must face in accepting cultural relativism.  The x-people live next to the y-people under nearly identical conditions.  The x-people castrate and blind one in three children, the y-people do not.  Are the x-people wrong in committing this practice? If you think so, you're not really a cultural relativist. [ENDQ] [NEWLINE] No, nobody is 'wrong' in any practice. There is no right and wrong. If you believe that there is then prove it. Show some evidence that a practice is wrong. And before you answer, no, 'intuition' is not objective evidence because different people in different cultures have different intuitions. [NEWLINE] [NEWLINE] I would work to stop blinding children in other countries because I am made sad by other people suffering, and I work to make myself happy. That's the point of my life. I do not pretend that I have special godlike knowledge of what is objectively right and wrong.</s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not a native English speaker,<mask> I speak English better than my native tongue, Portuguese.  Still, I feel awkward<mask> all hell<mask> I speak Portuguese words -- usually names -- in English.  The normal American accent I have seems wrong, and switching languages altogether seems even wronger. [NEWLINE] [NEWLINE] That said, I feel like some words should be spoken with a normal American accent and some should be spoken with their original language's phonemes.  Spanish words like "Latina" should *definitely* be in American.  It's an English word now, borrowed from Spanish. <mask><mask><mask><mask>, French words and phrases like "raison d'être" are *not* loanwords<mask> actual French words and should be pronounced in French -- perhaps not with an exaggerated Parisian accent,<mask> with the right French phonemes emulated in English -- perhaps with the American R rather than the French R,<mask> still with the French vowels. [NEWLINE] [NEWLINE] It seems that the distinction is between languages,<mask> it's really between whether a word is a foreign word or a borrowed word. <mask> I order a tatsuta-age, I say it with Japanese pronunciation. <mask> I order sushi, I say it with American pronunciation.  (The tatsuta-age set at my local "nicer" Japanese restaurant comes with better side dishes.  I don't think that's relevant,<mask> it *is* delicious.) [NEWLINE] [NEWLINE] This extends to geographic names<mask> well.  You say "Japan" with an American accent; you don't say "Nippon" like in actual Japanese.  That's<mask> "Japan" is not actually a Japanese word! <mask><mask><mask><mask>, place names without American pronunciations should be pronounced somewhat close to the original<mask> not exaggeratedly<mask>.  I tell people I was born in Reeyow dee Djuhnehrow,<mask> it drives me crazy<mask> my father's birthplace is pronounced "Sahw Pawlow".  It's "Suhoong Pahwlow".  You gotta get that right!  And it drives me even crazier<mask> people pronounce "Brazil" with an emphasized S and L,<mask> in Portuguese it's actually pronounced "Braziu". [NEWLINE] [NEWLINE] Basically: pronounce loanwords normally in English, and for foreign words, approximate them<mask> well<mask> you can in English without slipping into a foreign accent.</s>
Label encoding: <s>I'm not a native English speaker, but I speak English better than my native tongue, Portuguese.  Still, I feel awkward as all hell when I speak Portuguese words -- usually names -- in English.  The normal American accent I have seems wrong, and switching languages altogether seems even wronger. [NEWLINE] [NEWLINE] That said, I feel like some words should be spoken with a normal American accent and some should be spoken with their original language's phonemes.  Spanish words like "Latina" should *definitely* be in American.  It's an English word now, borrowed from Spanish.  On the other hand, French words and phrases like "raison d'être" are *not* loanwords but actual French words and should be pronounced in French -- perhaps not with an exaggerated Parisian accent, but with the right French phonemes emulated in English -- perhaps with the American R rather than the French R, but still with the French vowels. [NEWLINE] [NEWLINE] It seems that the distinction is between languages, but it's really between whether a word is a foreign word or a borrowed word.  When I order a tatsuta-age, I say it with Japanese pronunciation.  When I order sushi, I say it with American pronunciation.  (The tatsuta-age set at my local "nicer" Japanese restaurant comes with better side dishes.  I don't think that's relevant, but it *is* delicious.) [NEWLINE] [NEWLINE] This extends to geographic names as well.  You say "Japan" with an American accent; you don't say "Nippon" like in actual Japanese.  That's because "Japan" is not actually a Japanese word!  On the other hand, place names without American pronunciations should be pronounced somewhat close to the original but not exaggeratedly so.  I tell people I was born in Reeyow dee Djuhnehrow, but it drives me crazy when my father's birthplace is pronounced "Sahw Pawlow".  It's "Suhoong Pahwlow".  You gotta get that right!  And it drives me even crazier when people pronounce "Brazil" with an emphasized S and L, because in Portuguese it's actually pronounced "Braziu". [NEWLINE] [NEWLINE] Basically: pronounce loanwords normally in English, and for foreign words, approximate them as well as you can in English without slipping into a foreign accent.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The plea for equal rights, is bullshit<mask> we already have equal rights, I can't marry a man. And gays can marry the opposite sex.<mask> our rights are quite equal. It's just I want to marry someone I can. [ENDQ] [NEWLINE] <mask> this is the right way to think about it, then you would have to say that prior to the removal of prohibitions against interracial marriage, everyone had equal rights,<mask> everyone was equally free to marry someone of the same race. Doesn't that seem like a pretty counter-intuitive way to describe that situation? Doesn't it seem like prohibitions on interracial marriage raises worrying issues of inequality? [NEWLINE] [NEWLINE] [STARTQ] Which brings me to the reason<mask> marriage exists: it's the societies tool to support its own reproduction. That's the reason<mask> families have reduced tax and some other bonuses. You might say that not all families have children,<mask> they just enjoy the doubt. And<mask> being married they have a higher chance of having a child. [ENDQ] [NEWLINE] <mask> there are other problems with this argument, the main problem is that we do (and have always) allowed heterosexual couples to marry who have no intention of producing a child, and who are not physically capable of producing a child. It's not,<mask> you suggest, that we give such people the benefit of the doubt - there is no doubt at all that a post-menopausal woman will not be able to conceive. Yes, a great many people who marry are fertile and do produce children,<mask> not all of them do and we have never considered a couple "less married"<mask> they are unwilling or unable to procreate. [NEWLINE] [NEWLINE] [STARTQ] Now,<mask> same sex couples can't have children in any natural way, and most of them don't want to (here comes in the fact that we don't know<mask> problems that might cause to the child,<mask> I'll leave it), I see no reason for them to marry. [ENDQ] [NEWLINE] Then you lack imagination, to put it frankly. Try asking infertile or elderly heterosexual couples<mask> they chose to marry,<mask> being unable or unwilling to have children. You might<mask> consider that some same-sex couples do<mask><mask> raise children - some of them have children via sperm donation or surrogacy, or from a previous relationship with someone of the opposite sex. In those cases, don't you think it's important to protect the children they are raising by allowing them to marry their partners? [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] The plea for equal rights, is bullshit because we already have equal rights, I can't marry a man. And gays can marry the opposite sex. So our rights are quite equal. It's just I want to marry someone I can. [ENDQ] [NEWLINE] If this is the right way to think about it, then you would have to say that prior to the removal of prohibitions against interracial marriage, everyone had equal rights, since everyone was equally free to marry someone of the same race. Doesn't that seem like a pretty counter-intuitive way to describe that situation? Doesn't it seem like prohibitions on interracial marriage raises worrying issues of inequality? [NEWLINE] [NEWLINE] [STARTQ] Which brings me to the reason why marriage exists: it's the societies tool to support its own reproduction. That's the reason why families have reduced tax and some other bonuses. You might say that not all families have children, but they just enjoy the doubt. And while being married they have a higher chance of having a child. [ENDQ] [NEWLINE] While there are other problems with this argument, the main problem is that we do (and have always) allowed heterosexual couples to marry who have no intention of producing a child, and who are not physically capable of producing a child. It's not, as you suggest, that we give such people the benefit of the doubt - there is no doubt at all that a post-menopausal woman will not be able to conceive. Yes, a great many people who marry are fertile and do produce children, but not all of them do and we have never considered a couple "less married" if they are unwilling or unable to procreate. [NEWLINE] [NEWLINE] [STARTQ] Now, as same sex couples can't have children in any natural way, and most of them don't want to (here comes in the fact that we don't know what problems that might cause to the child, but I'll leave it), I see no reason for them to marry. [ENDQ] [NEWLINE] Then you lack imagination, to put it frankly. Try asking infertile or elderly heterosexual couples why they chose to marry, despite being unable or unwilling to have children. You might also consider that some same-sex couples do in fact raise children - some of them have children via sperm donation or surrogacy, or from a previous relationship with someone of the opposite sex. In those cases, don't you think it's important to protect the children they are raising by allowing them to marry their partners? [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with a language like Python is that,<mask> it is easy to learn and understand, it gives newcomers the wrong idea about<mask> programming works. Python is a scripting language, it was designed solely for programmers to churn out<mask> much code<mask> possible without having to worry about things such<mask> compilers, data types, or optimization. It is important for new programmers to not only learn the process of writing code,<mask> it is<mask> important to have a basic understanding of<mask> a computer interprets the code you write. [NEWLINE] [NEWLINE] [NEWLINE] I find that Java is the best balance of learning<mask> to program<mask> well<mask> learning<mask> programs work. It provides enough abstraction of tedious computer tasks, such<mask> memory management,<mask> exposing you to some of the lower-level aspects of computing that you need to know to be a successful programmer, such<mask><mask> data is represented in memory (ints vs floats, etc.),<mask> a compiler works,<mask> references are and<mask> they work, and a better idea on<mask> I/O works, which is really important to understand. [NEWLINE] [NEWLINE] Java<mask> prepares you to expand out into other languages.It introduces you to C-style syntax, which is used by 80% of programming languages out there. Most importantly, it strongly encourages object-oriented programming by introducing inheritance, encapsulation, and polymorphism in a way that is easy for a beginner to understand. [NEWLINE] [NEWLINE] Going back to your car analogy, think of Java<mask> learning to drive in a stick shift. You'll have to learn<mask> to shift properly along with actually driving,<mask> in the process, you learn more about<mask> an engine and transmission works, and you'll be able to drive more efficiently, learning automatic will be easy for you, and<mask> your car breaks down, you might have a better idea of<mask>'s wrong<mask> you know more about<mask> it works under the hood. [NEWLINE] [NEWLINE] Java (C# is the same way too) gives you the best of both worlds. It teaches you<mask> not to just think like a programmer,<mask> think like a computer too. Python won't tell you<mask> your data is being stored, nor will it tell you that the list comprehension you just used is actually a loop in disguise, and Java won't bug you about managing your memory, operating systems, or confuse you with its syntax (I'm looking at you, C++!)<mask> well<mask> better prepare you to work with other programming languages and platforms.</s>
Label encoding: <s>The problem with a language like Python is that, while it is easy to learn and understand, it gives newcomers the wrong idea about how programming works. Python is a scripting language, it was designed solely for programmers to churn out as much code as possible without having to worry about things such as compilers, data types, or optimization. It is important for new programmers to not only learn the process of writing code, but it is also important to have a basic understanding of how a computer interprets the code you write. [NEWLINE] [NEWLINE] [NEWLINE] I find that Java is the best balance of learning how to program as well as learning how programs work. It provides enough abstraction of tedious computer tasks, such as memory management, while exposing you to some of the lower-level aspects of computing that you need to know to be a successful programmer, such as how data is represented in memory (ints vs floats, etc.), how a compiler works, what references are and how they work, and a better idea on how I/O works, which is really important to understand. [NEWLINE] [NEWLINE] Java also prepares you to expand out into other languages.It introduces you to C-style syntax, which is used by 80% of programming languages out there. Most importantly, it strongly encourages object-oriented programming by introducing inheritance, encapsulation, and polymorphism in a way that is easy for a beginner to understand. [NEWLINE] [NEWLINE] Going back to your car analogy, think of Java as learning to drive in a stick shift. You'll have to learn how to shift properly along with actually driving, but in the process, you learn more about how an engine and transmission works, and you'll be able to drive more efficiently, learning automatic will be easy for you, and if your car breaks down, you might have a better idea of what's wrong because you know more about how it works under the hood. [NEWLINE] [NEWLINE] Java (C# is the same way too) gives you the best of both worlds. It teaches you how not to just think like a programmer, but think like a computer too. Python won't tell you how your data is being stored, nor will it tell you that the list comprehension you just used is actually a loop in disguise, and Java won't bug you about managing your memory, operating systems, or confuse you with its syntax (I'm looking at you, C++!) as well as better prepare you to work with other programming languages and platforms.</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That bit is the problem I feel.<mask> it comes to religiously motivated moral arguments, from the perspective of the non-religious person to say that something is 'wrong' is just your opinion, and not something I am required to share. [ENDQ] [NEWLINE] <mask><mask><mask><mask> that religion has a monopoly on this<mask>. Take something like slavery or murder. Many religious texts outline<mask> these things are ok,<mask> society has largely decided they are not. Are those just commonly held opinions?<mask> makes them right? [NEWLINE] [NEWLINE] More interestingly<mask> we are not making those decisions based on religion,<mask> are we basing those decisions on? Aren't all of our morals merely opinions? [NEWLINE] [NEWLINE] [STARTQ] Of course religious people probably think they are just doing<mask> is right and that they are trying to save people etc. [ENDQ] [NEWLINE] Much the same way that secular society creates laws and rules.<mask> makes those fundamentally right? We have a social understanding that speed limits are essentially just guidelines now,<mask><mask> did we decide that particular law can be broken?<mask> I exploit a legal loophole and benefit by not breaking any laws I am legally right,<mask> am I morally right? [NEWLINE] [NEWLINE] [STARTQ] I can ask them to stop. [ENDQ] [NEWLINE] We can ask someone who is doing something that we do not agree with or feel is harming us to stop. Who decides<mask> they should?<mask> both sides feel that they are right, is it not the opinion that has the moral majority and<mask> the law in a secular society behind it that wins?<mask> happens<mask> a paradigm shift has not reached critical mass?<mask> society has a negative collective opinion on it, does that mean it is wrong?<mask> that collective opinion has changed,<mask> the legal framework hasn't, does that make it right? [NEWLINE] [NEWLINE] [STARTQ] I'm sure they would not be equally understanding<mask> I were to go all richard dawkins and try and'save them from religion' or something<mask>. [ENDQ] [NEWLINE] Conversion is not just a religious issue. In politics, people at different ends of the political spectrum frequently disagree with each other, and often enact laws that the entire jurisdiction has to follow, even<mask> the people within that jurisdiction do not agree with them. Those differences of opinions have very real consequences on very real lives. [NEWLINE] [NEWLINE] Something being morally right or wrong is entirely based on an opinion, and the popularity of that opinion regardless to its source is<mask> dictates who agrees with you. </s>
Label encoding: <s> [STARTQ] That bit is the problem I feel. When it comes to religiously motivated moral arguments, from the perspective of the non-religious person to say that something is 'wrong' is just your opinion, and not something I am required to share. [ENDQ] [NEWLINE] I do not think that religion has a monopoly on this though. Take something like slavery or murder. Many religious texts outline when these things are ok, but society has largely decided they are not. Are those just commonly held opinions? What makes them right? [NEWLINE] [NEWLINE] More interestingly if we are not making those decisions based on religion, what are we basing those decisions on? Aren't all of our morals merely opinions? [NEWLINE] [NEWLINE] [STARTQ] Of course religious people probably think they are just doing what is right and that they are trying to save people etc. [ENDQ] [NEWLINE] Much the same way that secular society creates laws and rules. What makes those fundamentally right? We have a social understanding that speed limits are essentially just guidelines now, but why did we decide that particular law can be broken? If I exploit a legal loophole and benefit by not breaking any laws I am legally right, but am I morally right? [NEWLINE] [NEWLINE] [STARTQ] I can ask them to stop. [ENDQ] [NEWLINE] We can ask someone who is doing something that we do not agree with or feel is harming us to stop. Who decides if they should? If both sides feel that they are right, is it not the opinion that has the moral majority and therefore the law in a secular society behind it that wins? What happens when a paradigm shift has not reached critical mass? If society has a negative collective opinion on it, does that mean it is wrong? If that collective opinion has changed, but the legal framework hasn't, does that make it right? [NEWLINE] [NEWLINE] [STARTQ] I'm sure they would not be equally understanding if I were to go all richard dawkins and try and'save them from religion' or something though. [ENDQ] [NEWLINE] Conversion is not just a religious issue. In politics, people at different ends of the political spectrum frequently disagree with each other, and often enact laws that the entire jurisdiction has to follow, even if the people within that jurisdiction do not agree with them. Those differences of opinions have very real consequences on very real lives. [NEWLINE] [NEWLINE] Something being morally right or wrong is entirely based on an opinion, and the popularity of that opinion regardless to its source is what dictates who agrees with you. </s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0661, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It sounds very much like you're prioritizing the social education of the bully over safety and mental clarity of the victim. I can't fathom<mask>. Sure, it will suck<mask> someone grows up thinking it's okay to hate gay people.<mask> sucks even more is to be gay and experience hatred just for existing. I for one would prefer bullies to grow up underexposed to gay people/culture rather than gays grow up timid, repressed, and afraid,<mask> they even get that far along without committing suicide. I get that you're pointing out that it will be more good for the general public<mask> bullies grow up knowing their behavior isn't okay than<mask> gay people segregate themselves<mask> far enough along in your argument at some point you're asking someone to be a martyr on an individual level and put up with bullying<mask> that these shitty kids can learn a lesson. That's not okay. [ENDQ] [NEWLINE] The reality is that gay children are going to regular schools, even<mask> we had a school like this for every region in the UK, it's only going to enforce the "us and them" mentality. Even<mask> it were the case these schools would be popular,<mask> would LEAs or local governments deal with the underlying issues,<mask> they could just shunt off the "problem" to another school? Problem solved. [NEWLINE] [NEWLINE] [STARTQ] In that case perhaps the existence of such a school and potential high numbers of attendance would give courage to those struggling with coming out. [ENDQ] [NEWLINE] It's not<mask><mask> gay role-models and public personalities do not exist. There could just be a chance that gay students are already aware of other gay students in their schools, or the wider local community. The idea behind the schools isn't including a high enrolment rate, which would be difficult to achieve. The Harvey Milk school in New York, that is being offered<mask> a template, only has over 100 students. [NEWLINE] [NEWLINE] [STARTQ] In a perfect world there would be no bullies.<mask> we recognize this is not a perfect world we do our best to mitigate the instances of, and negative effects of bullying on our youth.<mask> this proposed school were to cut down on the instances of abuse (which at this point would be hard to<mask><mask> it won't at least a little bit) then it I'd a good thing. [ENDQ] [NEWLINE] And that is absolutely not something we could do to bring to regular schools instead?</s><pad>
Label encoding: <s> [STARTQ] It sounds very much like you're prioritizing the social education of the bully over safety and mental clarity of the victim. I can't fathom why. Sure, it will suck if someone grows up thinking it's okay to hate gay people. What sucks even more is to be gay and experience hatred just for existing. I for one would prefer bullies to grow up underexposed to gay people/culture rather than gays grow up timid, repressed, and afraid, if they even get that far along without committing suicide. I get that you're pointing out that it will be more good for the general public if bullies grow up knowing their behavior isn't okay than if gay people segregate themselves but far enough along in your argument at some point you're asking someone to be a martyr on an individual level and put up with bullying so that these shitty kids can learn a lesson. That's not okay. [ENDQ] [NEWLINE] The reality is that gay children are going to regular schools, even if we had a school like this for every region in the UK, it's only going to enforce the "us and them" mentality. Even if it were the case these schools would be popular, why would LEAs or local governments deal with the underlying issues, when they could just shunt off the "problem" to another school? Problem solved. [NEWLINE] [NEWLINE] [STARTQ] In that case perhaps the existence of such a school and potential high numbers of attendance would give courage to those struggling with coming out. [ENDQ] [NEWLINE] It's not as if gay role-models and public personalities do not exist. There could just be a chance that gay students are already aware of other gay students in their schools, or the wider local community. The idea behind the schools isn't including a high enrolment rate, which would be difficult to achieve. The Harvey Milk school in New York, that is being offered as a template, only has over 100 students. [NEWLINE] [NEWLINE] [STARTQ] In a perfect world there would be no bullies. Because we recognize this is not a perfect world we do our best to mitigate the instances of, and negative effects of bullying on our youth. If this proposed school were to cut down on the instances of abuse (which at this point would be hard to argue that it won't at least a little bit) then it I'd a good thing. [ENDQ] [NEWLINE] And that is absolutely not something we could do to bring to regular schools instead?</s><pad>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The Russian....empire have been over long ago [ENDQ] [NEWLINE] You should tell this to the people of the Ukraine, the people of Georgia, the people of Dagestan, and the people of Chechnya. <mask><mask> the Polish would<mask> be interested to learn this<mask> well. [NEWLINE] [NEWLINE] [STARTQ]...Japanese... [ENDQ] [NEWLINE] The Japanese invaded more countries than anyone else has in the last 100 years<mask> to say the USA has invaded more is just a false statement. [NEWLINE] [NEWLINE] [STARTQ] British empire have been over long ago [ENDQ] [NEWLINE] Long ago?  You must be young.  1982 was only 32 years ago.  The Canada Act of 1982 marks one of the final acts of the British Empire. <mask> returning Hong Kong to PR China was the final act of the Empire and that was in 1997. [NEWLINE] [NEWLINE] I bet the Irish and the Scottish are glad the British Empire is over....oh wait. [NEWLINE] [NEWLINE] <mask> the British Empire ended in 1997 and that was "long ago" by your standards, then I can ignore every american invasion before 1997. [NEWLINE] [NEWLINE] [STARTQ] the US just took their place. [ENDQ] [NEWLINE] <mask><mask>? <mask> did we colonize? [NEWLINE] [NEWLINE] BTW the British have 14 overseas territories, compare that to the 4 that the United States has.  There are<mask> 15 nations who still hold Queen Elizabeth II<mask> the reigning [constitutional monarch]( [URL] ). [NEWLINE] [NEWLINE] [STARTQ] You<mask> forget the US was who lead the invasions of afghanistan and iran plus others<mask> saying it was a coalition is disingenuous. [ENDQ] [NEWLINE] The US never invaded Iran,<mask> I am not the one being disingenuous (you are). [NEWLINE] [NEWLINE] [STARTQ] The US<mask> keeps military bases in a large part of the worl[1]. [ENDQ] [NEWLINE] <mask> does this have to do with my implication that you think I am implying? [NEWLINE] [NEWLINE] [STARTQ] I don't know of any recent one that surpasses US in leadership, size nor number of deaths[2] Do you? [ENDQ] [NEWLINE] Care you explain your double standard?  British are excused till 1997 (which was "long ago"), Japanese are excused till 1945, Russia is excused<mask><mask> they continue to expand,<mask> the record of the USA goes back two hundred years? [NEWLINE] [NEWLINE] [STARTQ] US is hated more in many countries. [ENDQ] [NEWLINE] Another non-validated generalized statement, thanks for the continued ambiguity.  </s><pad><pad>
Label encoding: <s> [STARTQ] The Russian....empire have been over long ago [ENDQ] [NEWLINE] You should tell this to the people of the Ukraine, the people of Georgia, the people of Dagestan, and the people of Chechnya.  I think the Polish would also be interested to learn this as well. [NEWLINE] [NEWLINE] [STARTQ]...Japanese... [ENDQ] [NEWLINE] The Japanese invaded more countries than anyone else has in the last 100 years so to say the USA has invaded more is just a false statement. [NEWLINE] [NEWLINE] [STARTQ] British empire have been over long ago [ENDQ] [NEWLINE] Long ago?  You must be young.  1982 was only 32 years ago.  The Canada Act of 1982 marks one of the final acts of the British Empire.  Though returning Hong Kong to PR China was the final act of the Empire and that was in 1997. [NEWLINE] [NEWLINE] I bet the Irish and the Scottish are glad the British Empire is over....oh wait. [NEWLINE] [NEWLINE] Since the British Empire ended in 1997 and that was "long ago" by your standards, then I can ignore every american invasion before 1997. [NEWLINE] [NEWLINE] [STARTQ] the US just took their place. [ENDQ] [NEWLINE] How so?  Where did we colonize? [NEWLINE] [NEWLINE] BTW the British have 14 overseas territories, compare that to the 4 that the United States has.  There are also 15 nations who still hold Queen Elizabeth II as the reigning [constitutional monarch]( [URL] ). [NEWLINE] [NEWLINE] [STARTQ] You also forget the US was who lead the invasions of afghanistan and iran plus others so saying it was a coalition is disingenuous. [ENDQ] [NEWLINE] The US never invaded Iran, so I am not the one being disingenuous (you are). [NEWLINE] [NEWLINE] [STARTQ] The US also keeps military bases in a large part of the worl[1]. [ENDQ] [NEWLINE] What does this have to do with my implication that you think I am implying? [NEWLINE] [NEWLINE] [STARTQ] I don't know of any recent one that surpasses US in leadership, size nor number of deaths[2] Do you? [ENDQ] [NEWLINE] Care you explain your double standard?  British are excused till 1997 (which was "long ago"), Japanese are excused till 1945, Russia is excused even though they continue to expand, yet the record of the USA goes back two hundred years? [NEWLINE] [NEWLINE] [STARTQ] US is hated more in many countries. [ENDQ] [NEWLINE] Another non-validated generalized statement, thanks for the continued ambiguity.  </s><pad><pad>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is important to understand<mask><mask> right wing/left wing has meant over time, and the context of different countries.  Canada's right wing looks a lot more like America's left wing than its right wing. [NEWLINE] [NEWLINE] <mask>....  Cheap education has always been a universal policy in the United States.  There were free/cheap single room schoolhouses back in the time of settlers and indians.  The left has recently been the champion of teachers unions, and in the last few decades "free universal university" has been a darling issue of the hard left. <mask> you can't attribute education, historically, to the left wing.  Keep in mind the very first schools were inventions of the church. [NEWLINE] [NEWLINE] Human Rights.  The foundational human rights document in the usa is the constitution which was, and is, bipartisan. <mask> you look to the human rights progress of the 1960s (<mask> the USA was already a prosperous country) there were a lot of democrats who were just<mask> racist<mask> the republicans.  Lincoln was a Republican.  Again in the last 30 years "human rights" has become a left issue,<mask> this isn't historically true. [NEWLINE] [NEWLINE] Research funds -<mask><mask> completely that this is a partisan issue.  Can you provide support? [NEWLINE] [NEWLINE] Cheap healthcare - this is a left wing issue. <mask><mask><mask> my general policy arguments still apply.  I<mask> don't think that the "left wing" in american politics today is much better than the right wing.  Hillary isn't proposing single payer.  The basic model of people paying for insurance is<mask> both sides want. [NEWLINE] [NEWLINE] Workers rights -<mask><mask> on this one you are historically correct (the "new deal" would never have happened under the republicans),<mask><mask><mask> the contemporary situation is different.  Republicans arn't calling to roll back workplace health and safety legislation, they arn't calling to repeal the minimum wage, they arn't trying to stop private sector collective bargaining.  The contemporary disputes are over whether public sector unions are a good thing, and whether the minimum wage should be raised higher (which republicans think will hurt the poor).  This is one of those issues<mask> time brought the republicans to the democrat's rough position, there have been times<mask> the democrats have come over to the republican side<mask> well.</s>
Label encoding: <s>It is important to understand how what right wing/left wing has meant over time, and the context of different countries.  Canada's right wing looks a lot more like America's left wing than its right wing. [NEWLINE] [NEWLINE] BUT....  Cheap education has always been a universal policy in the United States.  There were free/cheap single room schoolhouses back in the time of settlers and indians.  The left has recently been the champion of teachers unions, and in the last few decades "free universal university" has been a darling issue of the hard left.  However you can't attribute education, historically, to the left wing.  Keep in mind the very first schools were inventions of the church. [NEWLINE] [NEWLINE] Human Rights.  The foundational human rights document in the usa is the constitution which was, and is, bipartisan.  If you look to the human rights progress of the 1960s ( when the USA was already a prosperous country) there were a lot of democrats who were just as racist as the republicans.  Lincoln was a Republican.  Again in the last 30 years "human rights" has become a left issue, but this isn't historically true. [NEWLINE] [NEWLINE] Research funds - I disagree completely that this is a partisan issue.  Can you provide support? [NEWLINE] [NEWLINE] Cheap healthcare - this is a left wing issue.  However I think my general policy arguments still apply.  I also don't think that the "left wing" in american politics today is much better than the right wing.  Hillary isn't proposing single payer.  The basic model of people paying for insurance is what both sides want. [NEWLINE] [NEWLINE] Workers rights - I think on this one you are historically correct (the "new deal" would never have happened under the republicans), but I think the contemporary situation is different.  Republicans arn't calling to roll back workplace health and safety legislation, they arn't calling to repeal the minimum wage, they arn't trying to stop private sector collective bargaining.  The contemporary disputes are over whether public sector unions are a good thing, and whether the minimum wage should be raised higher (which republicans think will hurt the poor).  This is one of those issues where time brought the republicans to the democrat's rough position, there have been times when the democrats have come over to the republican side as well.</s>
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would first like to say that you are comparing apples to oranges, and you seem to understand a little (<mask> in your #3),<mask> haven't<mask> realized the breadth of the problem. Like you say in your #3, pie is a very malleable concept. A pie is a large class of foods.  A cake is not; it is a very specific food.<mask> I were to say that chicken is better than toast, you might ask me<mask> the chicken was prepared and<mask> the toast was prepared<mask> that you might agree or disagree. (alternatively, you might hate toast)  Either way, a pie can be described anywhere between Apples and Sugar in Crust, to Cheese and Pureed Tomato on doughy bread, whereas a cake is a sugary bread with frosting on it.  A cake is a much more specific type of food than a pie, and<mask> not something you should compare against. [NEWLINE] [NEWLINE] <mask>, just for fun (and<mask> people here seem to like iterated lists<mask> much), here's one: [NEWLINE] [NEWLINE] [STARTQ] 1.  Pies are filled with fruit, which is yummy and reasonably healthy even<mask> sweetened and put into pie. [ENDQ] 2.  A good cake can be very good,<mask> an average cake is not<mask> good<mask> even a poorly made pie. [NEWLINE] 3.  Pie is a very malleable concept.  Savoury pies (such<mask> meat pies or pizza pies) are<mask> amazing. [NEWLINE] 4.  Pumpkin pie is the best smell in the world. [NEWLINE] [NEWLINE] 1. Not all pies are filled with fruit, and there are fruit (and alcohol) filled cakes which can be just<mask> healthy (or get you drunker than) any pie. [NEWLINE] 2. A 'good' cake is often measured in both appearance and taste; a very good looking cake will not likely taste good, and a respectably-looking cake will usually taste good; a dessert pie was made to taste good; many of them are essentially cookie crust with sugary slop in the middle.  Looks are not a factor for a pie. [NEWLINE] 3. Addressed above. [NEWLINE] 4.<mask> you like that, you should take a sniff of pumpkin cookies, pumpkin muffins, and pumpkin cakes (they exist!). [NEWLINE] [NEWLINE] tl;dr; An Xbox is clearly superior to a Galaxy 5.</s>
Label encoding: <s>I would first like to say that you are comparing apples to oranges, and you seem to understand a little ( as in your #3), but haven't yet realized the breadth of the problem. Like you say in your #3, pie is a very malleable concept. A pie is a large class of foods.  A cake is not; it is a very specific food. If I were to say that chicken is better than toast, you might ask me how the chicken was prepared and how the toast was prepared so that you might agree or disagree. (alternatively, you might hate toast)  Either way, a pie can be described anywhere between Apples and Sugar in Crust, to Cheese and Pureed Tomato on doughy bread, whereas a cake is a sugary bread with frosting on it.  A cake is a much more specific type of food than a pie, and therefore not something you should compare against. [NEWLINE] [NEWLINE] Also, just for fun (and because people here seem to like iterated lists so much), here's one: [NEWLINE] [NEWLINE] [STARTQ] 1.  Pies are filled with fruit, which is yummy and reasonably healthy even when sweetened and put into pie. [ENDQ] 2.  A good cake can be very good, but an average cake is not as good as even a poorly made pie. [NEWLINE] 3.  Pie is a very malleable concept.  Savoury pies (such as meat pies or pizza pies) are also amazing. [NEWLINE] 4.  Pumpkin pie is the best smell in the world. [NEWLINE] [NEWLINE] 1. Not all pies are filled with fruit, and there are fruit (and alcohol) filled cakes which can be just as healthy (or get you drunker than) any pie. [NEWLINE] 2. A 'good' cake is often measured in both appearance and taste; a very good looking cake will not likely taste good, and a respectably-looking cake will usually taste good; a dessert pie was made to taste good; many of them are essentially cookie crust with sugary slop in the middle.  Looks are not a factor for a pie. [NEWLINE] 3. Addressed above. [NEWLINE] 4. If you like that, you should take a sniff of pumpkin cookies, pumpkin muffins, and pumpkin cakes (they exist!). [NEWLINE] [NEWLINE] tl;dr; An Xbox is clearly superior to a Galaxy 5.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Of course,<mask> Viagra is very cheap and other methods of artificial insemination are very expensive. Using Viagra to have a baby is more natural than other methods available today. [ENDQ] [NEWLINE] I get the point you're making,<mask> it's not convincing enough. ED is still covered<mask> you're not trying to conceive<mask> well, which your point neglects. [NEWLINE] [NEWLINE] [STARTQ] I'm pretty sure you are wrong here. ED is largely preventable through different lifestyle choices,<mask> I do not believe that it is largely treatable. [ENDQ] [NEWLINE] I read [this]( [URL]?ref=health). Here's a quick excerpt: [NEWLINE] [NEWLINE] “There is increasing evidence that we can reverse erectile dysfunction with lifestyle changes,” says Dr. Drogo K. Montague, director of the Center for Genitourinary Reconstruction in the Glickman Urological and Kidney Institute at Cleveland Clinic. [NEWLINE] [NEWLINE] In a recent study of men with E.D., or at risk for developing it, researchers in Italy found that the men could improve their erections by losing weight, improving their diet and exercising more frequently. After two years of significant lifestyle changes, 58 percent of the men had normal erectile function,<mask><mask> the study, which was published in The Journal of Sexual Medicine in January. [NEWLINE] [NEWLINE] [STARTQ] There are young people who use Viagra and I'm 100% confident (even without proof) that there are many people alive today who were only born<mask> of Viagra. You can't make that claim about birth control unless you include birth control failures. Plus the abstinence only crowd does have a point, abstinence is the only absolutely 100% effective non-surgical pregnancy prevention option. [ENDQ] To sum up, Viagra is primarily used for fun,<mask> a subset use it to procreate. Contraceptives are used only for fun (ignoring for now the health benefits of the pill for some women and assuming they can get it another way).<mask> I can understand,<mask> not agreeing with, the moral position for providing Viagra<mask> not contraceptives. [NEWLINE] [NEWLINE] This is interesting. This is more convincing to me. I MAY be able to understand covering Viagra<mask> used for procreation,<mask> not for fun,<mask> birth control isn't covered for fun. Do you think you could address my 'cycle of poverty' point?</s>
Label encoding: <s> [STARTQ] Of course, but Viagra is very cheap and other methods of artificial insemination are very expensive. Using Viagra to have a baby is more natural than other methods available today. [ENDQ] [NEWLINE] I get the point you're making, but it's not convincing enough. ED is still covered if you're not trying to conceive as well, which your point neglects. [NEWLINE] [NEWLINE] [STARTQ] I'm pretty sure you are wrong here. ED is largely preventable through different lifestyle choices, but I do not believe that it is largely treatable. [ENDQ] [NEWLINE] I read [this]( [URL]?ref=health). Here's a quick excerpt: [NEWLINE] [NEWLINE] “There is increasing evidence that we can reverse erectile dysfunction with lifestyle changes,” says Dr. Drogo K. Montague, director of the Center for Genitourinary Reconstruction in the Glickman Urological and Kidney Institute at Cleveland Clinic. [NEWLINE] [NEWLINE] In a recent study of men with E.D., or at risk for developing it, researchers in Italy found that the men could improve their erections by losing weight, improving their diet and exercising more frequently. After two years of significant lifestyle changes, 58 percent of the men had normal erectile function, according to the study, which was published in The Journal of Sexual Medicine in January. [NEWLINE] [NEWLINE] [STARTQ] There are young people who use Viagra and I'm 100% confident (even without proof) that there are many people alive today who were only born because of Viagra. You can't make that claim about birth control unless you include birth control failures. Plus the abstinence only crowd does have a point, abstinence is the only absolutely 100% effective non-surgical pregnancy prevention option. [ENDQ] To sum up, Viagra is primarily used for fun, but a subset use it to procreate. Contraceptives are used only for fun (ignoring for now the health benefits of the pill for some women and assuming they can get it another way). So I can understand, while not agreeing with, the moral position for providing Viagra but not contraceptives. [NEWLINE] [NEWLINE] This is interesting. This is more convincing to me. I MAY be able to understand covering Viagra if used for procreation, but not for fun, if birth control isn't covered for fun. Do you think you could address my 'cycle of poverty' point?</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you're really referring to 9/11 and really still don't know the very valid reasons that Osama bin Laden had to be pissed off at the USA and the CIA, you're really ignoring<mask> American foreign policy has absolutely shaped modern extremist Islam. Essentially, it wouldn't exist had we not been mucking about in the middle east<mask> the end of World War 2. [ENDQ] [NEWLINE] [STARTQ] To be clear, I'm not condoning the 9/11 attacks,<mask> you have to understand the reasons<mask> they were upset to begin with, and<mask> you just say "Oh it's just Islamic extremism!" it's tantamount to Bush saying "Well they just hate our freedom." It doesn't actually say anything. It's a sound-bite, not an argument. It does nothing to address the extremely complex world of international politics. [ENDQ] [NEWLINE] Good job blaming the victim. [NEWLINE] [NEWLINE] Do you even understand<mask> the US was in the Middle East? And more importantly, who Osama actually declared war on? [NEWLINE] [NEWLINE] It wasn't to meddle in affairs. The US-Saudi relationship actually began<mask> *Franklin D. Roosevelt* met the founder of Saudi Arabia, King Abdul-aziz, who in exchange for military bases/depots for refueling ships and planes during World War 2, the US would offer protection. [NEWLINE] [NEWLINE] It had nothing to do with Israel or oil - the relationship predates all of that. [NEWLINE] [NEWLINE] Osama bin Laden did not just declare war on the US - he declared war on Saudi Arabia<mask> well. To Osama and other followers like him, the Saudi government is an illegitimate ruler over the Holy Lands of Islam.<mask> Saddam invaded Kuwait, Osama approached the Saudi royal family (through is family's connections) offering his forces in Afghanistan to help fight Saddam. The Saudi royal family refused, and allowed the international coalition led by the US to fight Saddam from Saudi territory. [NEWLINE] [NEWLINE] To Osama and other hardline Islamists, this was the final straw: the Saudi royal family had committed religious treason by allowing an infidel army to base its troops on holy Islamic soil. [NEWLINE] [NEWLINE] <mask> please, enlighten us more with<mask> Osama was<mask> interested in the US<mask> the US had been around for decades in the area and Osama and other hardline Islamists that predated him rarely ever cared about it until the last few decades.</s>
Label encoding: <s> [STARTQ] If you're really referring to 9/11 and really still don't know the very valid reasons that Osama bin Laden had to be pissed off at the USA and the CIA, you're really ignoring how American foreign policy has absolutely shaped modern extremist Islam. Essentially, it wouldn't exist had we not been mucking about in the middle east since the end of World War 2. [ENDQ] [NEWLINE] [STARTQ] To be clear, I'm not condoning the 9/11 attacks, but you have to understand the reasons why they were upset to begin with, and when you just say "Oh it's just Islamic extremism!" it's tantamount to Bush saying "Well they just hate our freedom." It doesn't actually say anything. It's a sound-bite, not an argument. It does nothing to address the extremely complex world of international politics. [ENDQ] [NEWLINE] Good job blaming the victim. [NEWLINE] [NEWLINE] Do you even understand why the US was in the Middle East? And more importantly, who Osama actually declared war on? [NEWLINE] [NEWLINE] It wasn't to meddle in affairs. The US-Saudi relationship actually began when *Franklin D. Roosevelt* met the founder of Saudi Arabia, King Abdul-aziz, who in exchange for military bases/depots for refueling ships and planes during World War 2, the US would offer protection. [NEWLINE] [NEWLINE] It had nothing to do with Israel or oil - the relationship predates all of that. [NEWLINE] [NEWLINE] Osama bin Laden did not just declare war on the US - he declared war on Saudi Arabia as well. To Osama and other followers like him, the Saudi government is an illegitimate ruler over the Holy Lands of Islam. When Saddam invaded Kuwait, Osama approached the Saudi royal family (through is family's connections) offering his forces in Afghanistan to help fight Saddam. The Saudi royal family refused, and allowed the international coalition led by the US to fight Saddam from Saudi territory. [NEWLINE] [NEWLINE] To Osama and other hardline Islamists, this was the final straw: the Saudi royal family had committed religious treason by allowing an infidel army to base its troops on holy Islamic soil. [NEWLINE] [NEWLINE] So please, enlighten us more with why Osama was so interested in the US when the US had been around for decades in the area and Osama and other hardline Islamists that predated him rarely ever cared about it until the last few decades.</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Not moving, getting citizenship. People move countries for a variety of reasons all the time. Sometimes they stay for months, years, or the rest of their lives without getting citizenship. Getting citizenship is a much more significant matter. [ENDQ] [NEWLINE] It's significant,<mask> isn't a criminal act, and shouldn't be subject to any government sanction whatsoever. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> you are willing to conceed that getting citizenship in another country that is at war with your country can allow your original country to terminate your citizenship against your will we can work with that. [ENDQ] [NEWLINE] I'll conceed it should be legally sanctionable, up to and including prosecution for treason.  Revocation of citizenship is a stupid punishment to apply for it<mask>.  A much more severe punishment (life in prison or the death penalty) is appropriate, and basically moots the revocation of citizenship. [NEWLINE] [NEWLINE] [STARTQ] I would ask you: isn't it worse to go to another country and then take up arms against your country? [ENDQ] [NEWLINE] Yes, it's treason, punishable in nearly all countries by the most severe punishment allowed under law. [NEWLINE] [NEWLINE] [STARTQ] <mask> i was an american after 9/11 and met an afghan girl, married her, and moved to afghanistan i would technically be moving to a country at war with mine<mask> i am still loyal to the USA [ENDQ] [NEWLINE] <mask> we're gonna get technical about<mask> this would play out in a court of law, you would not be moving to a country with which the US is at war, I mean, unless you managed to move to and naturalize with the Taliban government Afghanistan in like the 3 months before we drove them out of power.  And yes,<mask> you tried doing that, anyone with any knowledge of the world would tell you you're not allowed to join up with a government the US is at war with<mask> a US citizen. <mask> you genuinely didn't do it except for love and extreme stupidity, you might get off of the treason charge on grounds of not actually aiding or comforting the enemy. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask><mask><mask> i actively take up arms I am clearly demonstrating a desire to remove myself from USA and join whatever ISIS is trying to create in the middle east. Isn't that worse? [ENDQ] [NEWLINE] It is a worse crime, and gets the most severe punishment allowed by law.</s>
Label encoding: <s> [STARTQ] Not moving, getting citizenship. People move countries for a variety of reasons all the time. Sometimes they stay for months, years, or the rest of their lives without getting citizenship. Getting citizenship is a much more significant matter. [ENDQ] [NEWLINE] It's significant, but isn't a criminal act, and shouldn't be subject to any government sanction whatsoever. [NEWLINE] [NEWLINE] [STARTQ] However if you are willing to conceed that getting citizenship in another country that is at war with your country can allow your original country to terminate your citizenship against your will we can work with that. [ENDQ] [NEWLINE] I'll conceed it should be legally sanctionable, up to and including prosecution for treason.  Revocation of citizenship is a stupid punishment to apply for it though.  A much more severe punishment (life in prison or the death penalty) is appropriate, and basically moots the revocation of citizenship. [NEWLINE] [NEWLINE] [STARTQ] I would ask you: isn't it worse to go to another country and then take up arms against your country? [ENDQ] [NEWLINE] Yes, it's treason, punishable in nearly all countries by the most severe punishment allowed under law. [NEWLINE] [NEWLINE] [STARTQ] If i was an american after 9/11 and met an afghan girl, married her, and moved to afghanistan i would technically be moving to a country at war with mine yet i am still loyal to the USA [ENDQ] [NEWLINE] If we're gonna get technical about how this would play out in a court of law, you would not be moving to a country with which the US is at war, I mean, unless you managed to move to and naturalize with the Taliban government Afghanistan in like the 3 months before we drove them out of power.  And yes, if you tried doing that, anyone with any knowledge of the world would tell you you're not allowed to join up with a government the US is at war with as a US citizen.  If you genuinely didn't do it except for love and extreme stupidity, you might get off of the treason charge on grounds of not actually aiding or comforting the enemy. [NEWLINE] [NEWLINE] [STARTQ] On the other hand when i actively take up arms I am clearly demonstrating a desire to remove myself from USA and join whatever ISIS is trying to create in the middle east. Isn't that worse? [ENDQ] [NEWLINE] It is a worse crime, and gets the most severe punishment allowed by law.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Some other important factors too; [ENDQ] [STARTQ] The behavioral response to giving someone $500 is very different then giving someone $1000 and then taxing them $500 at a later date. They will spend the $1000 and then pay the taxes with income later which creates an inflationary effect. [ENDQ] [STARTQ] There is currently no list in existence that could be used<mask> the basis of UBI, the US government doesn't have a list of citizens and no real mechanism to build one. [ENDQ] [STARTQ] Less significant is that UBI is more prone to fraud and would require new administration and overhead in order to establish, CBI is administered by the tax system based on withholding (an easy way to consider the difference for fraud is that to cheat UBI you have to lie, to cheat CBI you, your employer, your employers accountant and your bank have to lie). [ENDQ] [STARTQ] UBI has a stronger disincentive for labor. CBI has the taper which encourages a constant increase in income and generally has a private income requirement such that everyone would remain connected with the labor force in some capacity. This is significant<mask> it effectively recreates the poverty problem in a different way, you would end up with a permanent underclass of people who never work and have abysmal mobility with their higher income peers. [ENDQ] [NEWLINE] 1) Clawback is not the proper way to pay for the UBI then, obviously. It would be rather foolish to give everyone a 10k UBI and then count that<mask> a taxable income. [NEWLINE] [NEWLINE] 2)<mask> social security isn't a thing? I'm pretty sure we could easily figure out who *our citizens are<mask> there was an impetus to do<mask>. [NEWLINE] [NEWLINE] 3) There would be less systems to defraud than current welfare systems, and less'monitoring' needed to see who's getting<mask> and<mask> much. Once you've established citizenship, there's no other things that need to be monitored outside of death. [NEWLINE] [NEWLINE] 4) Currently there's not enough jobs out there for everyone anyway.<mask> then is actually going to encourage more labor? Or more production? Tying it to labor,<mask> actual human input into useful production has been trending downwards for decades due to automation (and isn't likely to stop anytime soon), essentially means you're forcing the creation of 'do nothing' jobs in order to receive your money.</s>
Label encoding: <s> [STARTQ] Some other important factors too; [ENDQ] [STARTQ] The behavioral response to giving someone $500 is very different then giving someone $1000 and then taxing them $500 at a later date. They will spend the $1000 and then pay the taxes with income later which creates an inflationary effect. [ENDQ] [STARTQ] There is currently no list in existence that could be used as the basis of UBI, the US government doesn't have a list of citizens and no real mechanism to build one. [ENDQ] [STARTQ] Less significant is that UBI is more prone to fraud and would require new administration and overhead in order to establish, CBI is administered by the tax system based on withholding (an easy way to consider the difference for fraud is that to cheat UBI you have to lie, to cheat CBI you, your employer, your employers accountant and your bank have to lie). [ENDQ] [STARTQ] UBI has a stronger disincentive for labor. CBI has the taper which encourages a constant increase in income and generally has a private income requirement such that everyone would remain connected with the labor force in some capacity. This is significant as it effectively recreates the poverty problem in a different way, you would end up with a permanent underclass of people who never work and have abysmal mobility with their higher income peers. [ENDQ] [NEWLINE] 1) Clawback is not the proper way to pay for the UBI then, obviously. It would be rather foolish to give everyone a 10k UBI and then count that as a taxable income. [NEWLINE] [NEWLINE] 2) So social security isn't a thing? I'm pretty sure we could easily figure out who *our citizens are if there was an impetus to do so. [NEWLINE] [NEWLINE] 3) There would be less systems to defraud than current welfare systems, and less'monitoring' needed to see who's getting what and how much. Once you've established citizenship, there's no other things that need to be monitored outside of death. [NEWLINE] [NEWLINE] 4) Currently there's not enough jobs out there for everyone anyway. What then is actually going to encourage more labor? Or more production? Tying it to labor, when actual human input into useful production has been trending downwards for decades due to automation (and isn't likely to stop anytime soon), essentially means you're forcing the creation of 'do nothing' jobs in order to receive your money.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agreed!!  I've mulled the issue myself a few times<mask> for me, i end up with notdomoduro's question;<mask> would I go?  You are likely to find that in every case you would be trading freedom for security, high taxes for high cost of living, etc... [NEWLINE] [NEWLINE] The old saying "no one wants to know<mask> the sausage is made" applies in this situation; the more you know about<mask> your government runs, the less you like it.  This applies to all national governments, no exception that I can think of.  Before you move to some other nation I would encourage that you research that nation to a degree that you know<mask> much about that nation<mask> you do about the U.S. [NEWLINE] [NEWLINE] <mask> you think the U.S. has problems with wrongful execution, think about the fact that many countries do not have freedom of press, and<mask> there is no way that you could know<mask> screwed up they are. [NEWLINE] [NEWLINE] There are plenty of places<mask> you can achieve a sense of personal safety,<mask> in most cases that is<mask> you are part of the privilege class and there is someone else in that country who is oppressed for your benefit.  Would you feel morally right about moving somewhere<mask> you are the 1% and the 99% get kicked in the balls<mask> that you can have a better life? [NEWLINE] [NEWLINE] The rest of the "developed" world is pretty much<mask> in-the-pocket of the corporate universe<mask> the U.S. is (<mask><mask> ), and the U.S. has among the most transparent justice systems in the world. <mask><mask> it is horrible that we have nearly the highest incarceration rate in the world,<mask> the people in jail are there for laws that are on the books, by a jury of your peers (albeit the lowest common denominator of citizens that end up on jury), and there is some sense of responsibility on the officials that the vast majority of inmates survive their sentence. <mask><mask><mask><mask> you can say that to be the case in more than a dozen countries. [NEWLINE] [NEWLINE] One thing I will encourage, without a doubt, is that you should NEVER give up your U.S. citizenship. <mask><mask> there is nothing in the world<mask> valuable<mask> a valid U.S. passport.</s>
Label encoding: <s>Agreed!!  I've mulled the issue myself a few times but for me, i end up with notdomoduro's question; where would I go?  You are likely to find that in every case you would be trading freedom for security, high taxes for high cost of living, etc... [NEWLINE] [NEWLINE] The old saying "no one wants to know how the sausage is made" applies in this situation; the more you know about how your government runs, the less you like it.  This applies to all national governments, no exception that I can think of.  Before you move to some other nation I would encourage that you research that nation to a degree that you know as much about that nation as you do about the U.S. [NEWLINE] [NEWLINE] If you think the U.S. has problems with wrongful execution, think about the fact that many countries do not have freedom of press, and so there is no way that you could know how screwed up they are. [NEWLINE] [NEWLINE] There are plenty of places where you can achieve a sense of personal safety, but in most cases that is because you are part of the privilege class and there is someone else in that country who is oppressed for your benefit.  Would you feel morally right about moving somewhere where you are the 1% and the 99% get kicked in the balls so that you can have a better life? [NEWLINE] [NEWLINE] The rest of the "developed" world is pretty much as in-the-pocket of the corporate universe as the U.S. is ( imho ), and the U.S. has among the most transparent justice systems in the world.  I think it is horrible that we have nearly the highest incarceration rate in the world, but the people in jail are there for laws that are on the books, by a jury of your peers (albeit the lowest common denominator of citizens that end up on jury), and there is some sense of responsibility on the officials that the vast majority of inmates survive their sentence.  I do not think you can say that to be the case in more than a dozen countries. [NEWLINE] [NEWLINE] One thing I will encourage, without a doubt, is that you should NEVER give up your U.S. citizenship.  IMHO there is nothing in the world as valuable as a valid U.S. passport.</s>
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While you may not fully understand the effects of certain actions or words on your own being. Those that do usually suffer from mental illnesses that can cause devestation reflection of an event or memory<mask> brought up. I have ptsd and<mask> i've learned to recognize<mask> i will be triggered it still happens sometimes. [NEWLINE] [NEWLINE] there are people that use this term to gain some sort of control over<mask> others say (tumblrsjw's) and radical feminists of sorts. I first hand know that triggers do exist and<mask><mask> you should at least acknowledge that they do for some people. sometimes they aren't even words<mask> a situation that will trigger somebody. a few nights ago someone was insulting me<mask> I told my friend "I hurt myself too much" and that friends sister went onto her account and started saying that i was a drama queen, that i'm a scrawny white guy who can't even get a job, and other things. That led to a flashback, I am mentally not fit to be in the work force<mask> me "not being able to get a job" was a blow to my self esteem that made me spiral downwards into depressive thoughts and finally i had a flashback. now, this is just an example<mask> let me point out a few key things. [NEWLINE] [NEWLINE] 1) a part of the reason was<mask> i was already stressed out at the time, i was put into a corner of drama from a few different people before that happened. [NEWLINE] [NEWLINE] 2)<mask> there was no stress, I may have been able to identify that i may be close to having an episode. [NEWLINE] [NEWLINE] 3) it wasn't 'one word' that caused this, it was a mix of paragraphs of insults. [NEWLINE] [NEWLINE] Triggers do exist and for some they can be far different than others, i do believe someone bringing up rape could 'possibly' trigger a victim into a flashback of<mask> happened,<mask> i doubt that just the word alone would cause it. now<mask> they went to see a movie and there was a disturbing rape scene then that would be more of an understandable trigger. In closing, i somewhat agree with you stating people should be able to overcome our issues.<mask> that doesn't happen out of nowhere in a split second. people struggle for years before they can achieve being better. </s><pad>
Label encoding: <s>While you may not fully understand the effects of certain actions or words on your own being. Those that do usually suffer from mental illnesses that can cause devestation reflection of an event or memory when brought up. I have ptsd and while i've learned to recognize when i will be triggered it still happens sometimes. [NEWLINE] [NEWLINE] there are people that use this term to gain some sort of control over what others say (tumblrsjw's) and radical feminists of sorts. I first hand know that triggers do exist and i think you should at least acknowledge that they do for some people. sometimes they aren't even words but a situation that will trigger somebody. a few nights ago someone was insulting me because I told my friend "I hurt myself too much" and that friends sister went onto her account and started saying that i was a drama queen, that i'm a scrawny white guy who can't even get a job, and other things. That led to a flashback, I am mentally not fit to be in the work force so me "not being able to get a job" was a blow to my self esteem that made me spiral downwards into depressive thoughts and finally i had a flashback. now, this is just an example but let me point out a few key things. [NEWLINE] [NEWLINE] 1) a part of the reason was because i was already stressed out at the time, i was put into a corner of drama from a few different people before that happened. [NEWLINE] [NEWLINE] 2) if there was no stress, I may have been able to identify that i may be close to having an episode. [NEWLINE] [NEWLINE] 3) it wasn't 'one word' that caused this, it was a mix of paragraphs of insults. [NEWLINE] [NEWLINE] Triggers do exist and for some they can be far different than others, i do believe someone bringing up rape could 'possibly' trigger a victim into a flashback of what happened, but i doubt that just the word alone would cause it. now if they went to see a movie and there was a disturbing rape scene then that would be more of an understandable trigger. In closing, i somewhat agree with you stating people should be able to overcome our issues. but that doesn't happen out of nowhere in a split second. people struggle for years before they can achieve being better. </s><pad>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well,<mask><mask> it depends... There are many different societies in the world, many of them with very different cultures, circumstances, history, economies, infrastructures and public attitudes. A system of public/private balance that works great in one place might be a complete disaster elsewhere. [NEWLINE] [NEWLINE] In practice, I don't think it's possible to have an entirely perfect system anywhere in the world - life is just too messy, the world too complex, and people's interests and opinions vary too wildly.<mask>, things change all the time and you never quite know exactly<mask> the future will bring.<mask>, at best, you can aim for an optimal compromise. And<mask> different countries or regions face very different challenges, I believe they have different optimal compromises. [NEWLINE] [NEWLINE] That being said:<mask> skeptical<mask> I am of the abilities of "pure capitalism" on its own to shape a society worth living in, there is<mask> something to be said for the importance of free enterprise. It's generally a good thing for people to have the ability to invest their energy and resources in the things they choose. There is of course the risk of failure,<mask> doing something you believe is inherently rewarding, and happy citizens generally make for healthy societies.<mask>, fantastic innovations sometimes result from this - innovations that might never have come about or at least might never have been widely adopted<mask> the only way for them to do<mask> would have been by way of politics or government planning. [NEWLINE] [NEWLINE] <mask> to get back to your question, personally I would lean towards more of a mix of public and private sectors. The public sector is no more a universal solution for the world's problems than the free market is.<mask> some things (transportation infrastructure, water, sanitation, etc.) are highly beneficial to society even<mask> they are not financially profitable in and of themselves, and for those things, public ownership generally makes sense<mask><mask><mask> I'm concerned.<mask><mask> it's OK for those things to make a financial loss,<mask> they provide significant value even<mask> they don't provide profit. [NEWLINE] [NEWLINE] (By the way - the traditional forms of capitalism are not necessarily the only way of ensuring free enterprise...<mask> that's a different story, and I don't want to get too far offtopic. Highly recommended viewing on the subject: [URL] )</s>
Label encoding: <s>Well, IMHO it depends... There are many different societies in the world, many of them with very different cultures, circumstances, history, economies, infrastructures and public attitudes. A system of public/private balance that works great in one place might be a complete disaster elsewhere. [NEWLINE] [NEWLINE] In practice, I don't think it's possible to have an entirely perfect system anywhere in the world - life is just too messy, the world too complex, and people's interests and opinions vary too wildly. Also, things change all the time and you never quite know exactly what the future will bring. Therefore, at best, you can aim for an optimal compromise. And since different countries or regions face very different challenges, I believe they have different optimal compromises. [NEWLINE] [NEWLINE] That being said: As skeptical as I am of the abilities of "pure capitalism" on its own to shape a society worth living in, there is also something to be said for the importance of free enterprise. It's generally a good thing for people to have the ability to invest their energy and resources in the things they choose. There is of course the risk of failure, but doing something you believe is inherently rewarding, and happy citizens generally make for healthy societies. Also, fantastic innovations sometimes result from this - innovations that might never have come about or at least might never have been widely adopted if the only way for them to do so would have been by way of politics or government planning. [NEWLINE] [NEWLINE] So to get back to your question, personally I would lean towards more of a mix of public and private sectors. The public sector is no more a universal solution for the world's problems than the free market is. But some things (transportation infrastructure, water, sanitation, etc.) are highly beneficial to society even if they are not financially profitable in and of themselves, and for those things, public ownership generally makes sense as far as I'm concerned. I think it's OK for those things to make a financial loss, because they provide significant value even if they don't provide profit. [NEWLINE] [NEWLINE] (By the way - the traditional forms of capitalism are not necessarily the only way of ensuring free enterprise... but that's a different story, and I don't want to get too far offtopic. Highly recommended viewing on the subject: [URL] )</s>
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1) In that case, Abrahamic religions are deeply concerned with<mask> the world was created, which is<mask> theories like the big bang and evolution are huge problems to them. [NEWLINE] [NEWLINE] I do not include Buddhism and Hinduism in the argument<mask><mask> your point may hold true, I have a different set of issues with those religions that do not hold with the points I have made here. [NEWLINE] [NEWLINE] 2)<mask> instead of "Divine Law" call it "Universal Law",<mask> stripping any divine attachment to it and still leaving it<mask> a credible story to teach morality. [NEWLINE] [NEWLINE] 3) Maybe i need to word this better. I am not saying to not fear the act of dying<mask> it can be scary and painful, it is death itself that i see no reason to fear, not the act and process of dying. [NEWLINE] [NEWLINE] 4) Creating community is not the problem, religion has done very well to bring people together in some cases, and in some cases the opposite is true. It is the preaching and blind following that I see<mask> a problem in religion, not the communal aspect of it. [NEWLINE] [NEWLINE] 5) I have, daily. I am not saying that i have reached the be-all end-all of understanding, and I do not see myself<mask> better than anyone else<mask> of my views. I am open to the views of others, I will listen and understand the,<mask> that does not mean I will accept them. [NEWLINE] [NEWLINE] [STARTQ] And would you really call Descartes "lack proper intellect and reasoning in terms of philosophy" compared to yourself? [ENDQ] [NEWLINE] <mask> Descartes is from a different time than I am he has not been opened to the same resources and science that we are now. I am in absolutely no way saying that I know more than him,<mask> it is possible that<mask> he were alive today he may have a very different mindset. I guess i just can't find the proper way to say it without sounding ignorant (maybe there are no words to do<mask> ),<mask><mask> I am attempting to explain is that some people have been blinded by religion, no matter<mask> vast their knowledge may be they may still have that need to cling to it,<mask> I am in no way saying that<mask> I don't have that need I am better than they are.</s>
Label encoding: <s>1) In that case, Abrahamic religions are deeply concerned with how the world was created, which is why theories like the big bang and evolution are huge problems to them. [NEWLINE] [NEWLINE] I do not include Buddhism and Hinduism in the argument because while your point may hold true, I have a different set of issues with those religions that do not hold with the points I have made here. [NEWLINE] [NEWLINE] 2) So instead of "Divine Law" call it "Universal Law", thus stripping any divine attachment to it and still leaving it as a credible story to teach morality. [NEWLINE] [NEWLINE] 3) Maybe i need to word this better. I am not saying to not fear the act of dying as it can be scary and painful, it is death itself that i see no reason to fear, not the act and process of dying. [NEWLINE] [NEWLINE] 4) Creating community is not the problem, religion has done very well to bring people together in some cases, and in some cases the opposite is true. It is the preaching and blind following that I see as a problem in religion, not the communal aspect of it. [NEWLINE] [NEWLINE] 5) I have, daily. I am not saying that i have reached the be-all end-all of understanding, and I do not see myself as better than anyone else because of my views. I am open to the views of others, I will listen and understand the, but that does not mean I will accept them. [NEWLINE] [NEWLINE] [STARTQ] And would you really call Descartes "lack proper intellect and reasoning in terms of philosophy" compared to yourself? [ENDQ] [NEWLINE] As Descartes is from a different time than I am he has not been opened to the same resources and science that we are now. I am in absolutely no way saying that I know more than him, as it is possible that if he were alive today he may have a very different mindset. I guess i just can't find the proper way to say it without sounding ignorant (maybe there are no words to do so ), but what I am attempting to explain is that some people have been blinded by religion, no matter how vast their knowledge may be they may still have that need to cling to it, but I am in no way saying that because I don't have that need I am better than they are.</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0847, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A gun carried openly by anyone other than a police officer serves no purpose. [ENDQ] [NEWLINE] That's a rather binary sort of statement. It may serve the purpose of psychological reinforcement for the person carrying the weapon. You may disagree,<mask> that doesn't make the purpose invalid. [NEWLINE] [NEWLINE] [STARTQ] It doesn't act<mask> a crime deterrent,<mask> would it? Crimes will be committed either way, and even<mask> a crime was committed and a citizen open carrying just happened to be near by (a statistical anomaly), that citizen couldn't use the gun unless the crime called for such an extreme measure. [ENDQ] [NEWLINE] You don't believe it's possible for a criminal in a gas station to abandon a robbery due to a person in line with a handgun on their hip? [NEWLINE] [NEWLINE] Simply<mask> crimes are committed and will be committed<mask><mask> open carry, doesn't necessarily mean that open carry doesn't ever serve<mask> a deterrent. Carrying a gun out in the open is a pretty good detergent against many types of crimes. Crimes of opportunity or chance against someone who is open carrying are less likely to happen or to be successful.<mask> you were a criminal and had the choice between robbing a person with a gun and a person without a gun, who would you choose? [NEWLINE] [NEWLINE] <mask>, laws very<mask> to<mask> a weapon can be used. A situation doesn't necessarily have to be extreme<mask> you might believe. [NEWLINE] [NEWLINE] [STARTQ] Open carrying draws attention, creates a public sense of fear not safety...<mask> would anyone feel safe around an unknown citizen who is openly carrying a lethal weapon. [ENDQ] [NEWLINE] <mask><mask> that it draws attention<mask> I see that<mask> a positive thing.<mask> good is a possible deterrent<mask> the deterrent can't be detected? [NEWLINE] [NEWLINE] I don't have any fear towards people who open carry. I assume that anyone willing to strap a gun to their belt is more likely to abide by the law than say a criminal with an illegal and concealed weapon.<mask> in the world would a person who is going to do something illegal open carry and have all that attention drawn to them?<mask> you were going to commit a crime would you open carry?<mask> you were allowed to have a gun, for whatever reason, would you open carry? I sure wouldn't. I'd hide it and this hide my intentions. [NEWLINE] </s>
Label encoding: <s> [STARTQ] A gun carried openly by anyone other than a police officer serves no purpose. [ENDQ] [NEWLINE] That's a rather binary sort of statement. It may serve the purpose of psychological reinforcement for the person carrying the weapon. You may disagree, but that doesn't make the purpose invalid. [NEWLINE] [NEWLINE] [STARTQ] It doesn't act as a crime deterrent, why would it? Crimes will be committed either way, and even if a crime was committed and a citizen open carrying just happened to be near by (a statistical anomaly), that citizen couldn't use the gun unless the crime called for such an extreme measure. [ENDQ] [NEWLINE] You don't believe it's possible for a criminal in a gas station to abandon a robbery due to a person in line with a handgun on their hip? [NEWLINE] [NEWLINE] Simply because crimes are committed and will be committed regardless of open carry, doesn't necessarily mean that open carry doesn't ever serve as a deterrent. Carrying a gun out in the open is a pretty good detergent against many types of crimes. Crimes of opportunity or chance against someone who is open carrying are less likely to happen or to be successful. If you were a criminal and had the choice between robbing a person with a gun and a person without a gun, who would you choose? [NEWLINE] [NEWLINE] Also, laws very as to when a weapon can be used. A situation doesn't necessarily have to be extreme as you might believe. [NEWLINE] [NEWLINE] [STARTQ] Open carrying draws attention, creates a public sense of fear not safety... why would anyone feel safe around an unknown citizen who is openly carrying a lethal weapon. [ENDQ] [NEWLINE] I agree that it draws attention but I see that as a positive thing. What good is a possible deterrent if the deterrent can't be detected? [NEWLINE] [NEWLINE] I don't have any fear towards people who open carry. I assume that anyone willing to strap a gun to their belt is more likely to abide by the law than say a criminal with an illegal and concealed weapon. Why in the world would a person who is going to do something illegal open carry and have all that attention drawn to them? If you were going to commit a crime would you open carry? If you were allowed to have a gun, for whatever reason, would you open carry? I sure wouldn't. I'd hide it and this hide my intentions. [NEWLINE] </s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok. Building 7. [NEWLINE] [NEWLINE] NTSC says it was destroyed by fire that melted beams that caused it to collapse.<mask> you could point to any other steel building pre or post 9/11 that collapsed into itself in a single event due to fire, show me. The whole premise doesn't stack up and neither does the structural physics of the event. [NEWLINE] [NEWLINE] Free fall collapse. [NEWLINE] [NEWLINE] Newtons 3rd law states that for every force there must be an equal opposing force. The top half of tower 2 should have met a large amount of resistance from the lower floors on its way down, the weight alone would not have caused it to fall at such a speed<mask> the floors below would have been providing an opposing force which kept it standing in the first place. [NEWLINE] [NEWLINE] In terms of debris falling at different speeds, this is explainable due to acceleration. Something does not begin falling at a constant speed, it accelerates.<mask> debris falling faster than the tower was likely ejected before or<mask> the tower began to fall. Wind resistance<mask> plays a part. An apple and a feather do not fall at the same speed. [NEWLINE] [NEWLINE] Eye witness accounts. [NEWLINE] [NEWLINE] <mask><mask> with you here. These are anecdotal reports and<mask><mask><mask> I am aware there's no scientific evidence to back these up,<mask> have seen data somewhere from seismometers used to measure earthquake that does show large seismic events that do not correspond with the timing of the plane crashes. I'd need to research this more before I could back up these accounts. [NEWLINE] [NEWLINE] Thermite. [NEWLINE] [NEWLINE] Again this needs a lot more investigation. Unfortunately the evidence that could prove this true or false has long<mask> been destroyed. The girders and debris from the site were immediately removed from the scene of the crime and disposed of.<mask> was this done?<mask> isn't any of the material available for analysts, apart from the few samples of dust collected by bystanders?<mask> can you explain the areas of extreme thermal activity present in debris the for weeks during the cleanup and the surgical cuts to the I-beams? [NEWLINE] [NEWLINE] I read through the articles you posted and all of them seem to use deliberate obfuscation and the kind of false logic, I referred to in previous comments, to confuse the reader into believing the information presented<mask> fact.</s>
Label encoding: <s>Ok. Building 7. [NEWLINE] [NEWLINE] NTSC says it was destroyed by fire that melted beams that caused it to collapse. If you could point to any other steel building pre or post 9/11 that collapsed into itself in a single event due to fire, show me. The whole premise doesn't stack up and neither does the structural physics of the event. [NEWLINE] [NEWLINE] Free fall collapse. [NEWLINE] [NEWLINE] Newtons 3rd law states that for every force there must be an equal opposing force. The top half of tower 2 should have met a large amount of resistance from the lower floors on its way down, the weight alone would not have caused it to fall at such a speed since the floors below would have been providing an opposing force which kept it standing in the first place. [NEWLINE] [NEWLINE] In terms of debris falling at different speeds, this is explainable due to acceleration. Something does not begin falling at a constant speed, it accelerates. Therefore debris falling faster than the tower was likely ejected before or as the tower began to fall. Wind resistance also plays a part. An apple and a feather do not fall at the same speed. [NEWLINE] [NEWLINE] Eye witness accounts. [NEWLINE] [NEWLINE] I agree with you here. These are anecdotal reports and as far as I am aware there's no scientific evidence to back these up, although have seen data somewhere from seismometers used to measure earthquake that does show large seismic events that do not correspond with the timing of the plane crashes. I'd need to research this more before I could back up these accounts. [NEWLINE] [NEWLINE] Thermite. [NEWLINE] [NEWLINE] Again this needs a lot more investigation. Unfortunately the evidence that could prove this true or false has long since been destroyed. The girders and debris from the site were immediately removed from the scene of the crime and disposed of. Why was this done? Why isn't any of the material available for analysts, apart from the few samples of dust collected by bystanders? How can you explain the areas of extreme thermal activity present in debris the for weeks during the cleanup and the surgical cuts to the I-beams? [NEWLINE] [NEWLINE] I read through the articles you posted and all of them seem to use deliberate obfuscation and the kind of false logic, I referred to in previous comments, to confuse the reader into believing the information presented as fact.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I've argued, it's not only beneficial for lower class. A meritocracy would lead to a more prosperous society for everyone involved. That includes your children. [ENDQ] [NEWLINE] <mask><mask> with your premise here.<mask> makes you believe that welfare programs increase the prosperity of a nation? [NEWLINE] [NEWLINE] [STARTQ] Look at it this way, do you want your children to earn their own greatness in a society that is great, or do you want them to live off your work, unearned, in a society that is less prosperous than it could be? [ENDQ] [NEWLINE] I want my children to have the happiest lives possible,<mask> my father wanted for me. [NEWLINE] [NEWLINE] From a very young age my father made it clear, that<mask> I wanted to have<mask> he had, be<mask> successful and wealthy<mask> he was, I would have to get a specific degree, work specific jobs, and prove myself worthy. Over the years I have put in the time, and effort, nobody can say I didn't earn<mask> I have today, and the same will be true for my children. They will not receive all I have unless they prove themselves worthy of it. That being said they will still receive enough to be comfortable. [NEWLINE] [NEWLINE] <mask> my children hopefully will achieve something great in a society that is already great, instead of having my fathers hard work, and my hard work confiscated by a government that can't run a single department within its budget. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> that the inheritance stuff is<mask> ingrained<mask> it was the way society was formed. Without society, there was no concept of fairness, and people took<mask> they could with violence, and those who thrived saw that even<mask> their children tried to earn things fairly, it'd never work. [ENDQ] [NEWLINE] This may be correct, who knows. Its irrelevant<mask>. [NEWLINE] [NEWLINE] [STARTQ] <mask> the fact that we have an established society now, means there that there are new possibilities. People fail to recognize that,<mask> we are still rooted in tradition,<mask> modern society and modern markets could allow a fair chance for everyone. [ENDQ] [NEWLINE] Everyone does have a fair chance at success. Some people may get a headstart,<mask> that doesn't mean a poor black kid from Hawaii can't become president, or a couple nerds can't drop out of college and create Microsoft</s>
Label encoding: <s> [STARTQ] As I've argued, it's not only beneficial for lower class. A meritocracy would lead to a more prosperous society for everyone involved. That includes your children. [ENDQ] [NEWLINE] I disagree with your premise here. What makes you believe that welfare programs increase the prosperity of a nation? [NEWLINE] [NEWLINE] [STARTQ] Look at it this way, do you want your children to earn their own greatness in a society that is great, or do you want them to live off your work, unearned, in a society that is less prosperous than it could be? [ENDQ] [NEWLINE] I want my children to have the happiest lives possible, as my father wanted for me. [NEWLINE] [NEWLINE] From a very young age my father made it clear, that if I wanted to have what he had, be as successful and wealthy as he was, I would have to get a specific degree, work specific jobs, and prove myself worthy. Over the years I have put in the time, and effort, nobody can say I didn't earn what I have today, and the same will be true for my children. They will not receive all I have unless they prove themselves worthy of it. That being said they will still receive enough to be comfortable. [NEWLINE] [NEWLINE] So my children hopefully will achieve something great in a society that is already great, instead of having my fathers hard work, and my hard work confiscated by a government that can't run a single department within its budget. [NEWLINE] [NEWLINE] [STARTQ] I think that the inheritance stuff is so ingrained because it was the way society was formed. Without society, there was no concept of fairness, and people took what they could with violence, and those who thrived saw that even if their children tried to earn things fairly, it'd never work. [ENDQ] [NEWLINE] This may be correct, who knows. Its irrelevant though. [NEWLINE] [NEWLINE] [STARTQ] But the fact that we have an established society now, means there that there are new possibilities. People fail to recognize that, because we are still rooted in tradition, but modern society and modern markets could allow a fair chance for everyone. [ENDQ] [NEWLINE] Everyone does have a fair chance at success. Some people may get a headstart, but that doesn't mean a poor black kid from Hawaii can't become president, or a couple nerds can't drop out of college and create Microsoft</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From a moral/practical standpoint, of course Qui-Gon should have freed both Anakin and Schmi -<mask> you have to remember that the world that Qui-Gon lives in is not a practical world. Everyone he meets has a fate and a purpose in the eyes of this weird puppet-master called "The Force". [NEWLINE] [NEWLINE] In Schmi's case she was a test for Anakin. Anakin is always overly attached to Schmi (and all the motherly figures in his life, like Padme). Eventually Schmi's death starts his real descent into the dark side. Qui-Gon didn't know this specifically,<mask> his prophetic powers probably gave him enough information to know that Anakin's ability to overcome his attachment to his mother was vital and figured that sooner was better than later. Remember, Qui-Gon always knew the prophecy about Anakin and probably came to Tatooine under some kind of instruction from the force. Even<mask> Qui-Gon had known that Schmi's death would have made Anakin go apeshit he probably would not have wanted to bring her along anyway. We see Anakin's obsessive need to control the lives of those around him based on his actions surrounding his visions of Padme's death.<mask> Schmi had been around, in Qui-Gon's eyes at least, she would be another distraction.<mask><mask> she would have been on Coruscant during both the Battle of Coruscant<mask> thousands of burning capital ships rained down on the planet and during the execution of order 66.<mask> she wasn't dead by that point, Palpatine would have made sure some accident happened given his designs on Anakin. [NEWLINE] [NEWLINE] <mask>, to summarize - from a simple moral standpoint of jedi good, slavery bad - sure, Qui-Gon might have felt compelled to go vigilante on Watto - it certainly would not have been out of character.<mask>, Qui-Gon was much more concerned with saving the force itself through the vessel of Anakin,<mask> he decided that out of the way on Tatooine would be the best place for Schmi. Bringing her to Coruscant would probably not have saved her life or Anakin's soul.</s>
Label encoding: <s>From a moral/practical standpoint, of course Qui-Gon should have freed both Anakin and Schmi - but you have to remember that the world that Qui-Gon lives in is not a practical world. Everyone he meets has a fate and a purpose in the eyes of this weird puppet-master called "The Force". [NEWLINE] [NEWLINE] In Schmi's case she was a test for Anakin. Anakin is always overly attached to Schmi (and all the motherly figures in his life, like Padme). Eventually Schmi's death starts his real descent into the dark side. Qui-Gon didn't know this specifically, but his prophetic powers probably gave him enough information to know that Anakin's ability to overcome his attachment to his mother was vital and figured that sooner was better than later. Remember, Qui-Gon always knew the prophecy about Anakin and probably came to Tatooine under some kind of instruction from the force. Even if Qui-Gon had known that Schmi's death would have made Anakin go apeshit he probably would not have wanted to bring her along anyway. We see Anakin's obsessive need to control the lives of those around him based on his actions surrounding his visions of Padme's death. If Schmi had been around, in Qui-Gon's eyes at least, she would be another distraction. In fact she would have been on Coruscant during both the Battle of Coruscant when thousands of burning capital ships rained down on the planet and during the execution of order 66. If she wasn't dead by that point, Palpatine would have made sure some accident happened given his designs on Anakin. [NEWLINE] [NEWLINE] So, to summarize - from a simple moral standpoint of jedi good, slavery bad - sure, Qui-Gon might have felt compelled to go vigilante on Watto - it certainly would not have been out of character. However, Qui-Gon was much more concerned with saving the force itself through the vessel of Anakin, so he decided that out of the way on Tatooine would be the best place for Schmi. Bringing her to Coruscant would probably not have saved her life or Anakin's soul.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> a lot of people have said the same things, or given a personal example. I was allowed into the gifted program in junior high (American 8th grade)<mask> low test scores and a poor performance in math. My parents wanted me in the gifted program<mask> I was vastly under-challenged in English and history. They told me I was allowed in on the condition I work extra hard at math and science, and that I would finally be able to work at my level in English and history. [NEWLINE] [NEWLINE] And I was. I couldn't grasp the logic problems. I needed extra help in those areas.<mask>, I was given the opportunity to be a part of a program that allowed me personal mentoring and attention to the areas I loved and excelled at.<mask> I did good work in<mask> I had previously been unchallenged with, it brought focus to the areas I was not<mask> good in. I could not compete with others in this area, and my parents hired a tutor to help me. [NEWLINE] [NEWLINE] And then I excelled in all areas. I went from a B or C student in math to  scoring a perfect score on the Regents exam (a state exam in NY) in math; something I NEVER would have been able to do<mask> not given the confidence that I was smart in<mask> I was good at, and<mask> my parents hadn't had a teacher explain to them that I would excel with extra attention in the area I needed help with. [NEWLINE] [NEWLINE] I think it's important to rely on the "strengths and weaknesses" of some students. Like a lot of other people here, I could say<mask> I was "exceptional" in certain areas---reading at a much higher level, writing at a much higher level, etc. I had already proven I was serious about study, and the gifted program was a way for me to prove<mask> I could do in a controlled environment with extra attention. It<mask> forced me to come to terms with others' abilities---I could explain metaphor, theme, and meaning to people who never grasped it, and some of them took the time to explain logic to me. It was a fantastic opportunity,<mask> one I very realistically had to earn with a lot of hard work and some convincing. </s>
Label encoding: <s>I think a lot of people have said the same things, or given a personal example. I was allowed into the gifted program in junior high (American 8th grade) despite low test scores and a poor performance in math. My parents wanted me in the gifted program because I was vastly under-challenged in English and history. They told me I was allowed in on the condition I work extra hard at math and science, and that I would finally be able to work at my level in English and history. [NEWLINE] [NEWLINE] And I was. I couldn't grasp the logic problems. I needed extra help in those areas. But, I was given the opportunity to be a part of a program that allowed me personal mentoring and attention to the areas I loved and excelled at. When I did good work in what I had previously been unchallenged with, it brought focus to the areas I was not so good in. I could not compete with others in this area, and my parents hired a tutor to help me. [NEWLINE] [NEWLINE] And then I excelled in all areas. I went from a B or C student in math to  scoring a perfect score on the Regents exam (a state exam in NY) in math; something I NEVER would have been able to do if not given the confidence that I was smart in what I was good at, and if my parents hadn't had a teacher explain to them that I would excel with extra attention in the area I needed help with. [NEWLINE] [NEWLINE] I think it's important to rely on the "strengths and weaknesses" of some students. Like a lot of other people here, I could say how I was "exceptional" in certain areas---reading at a much higher level, writing at a much higher level, etc. I had already proven I was serious about study, and the gifted program was a way for me to prove what I could do in a controlled environment with extra attention. It also forced me to come to terms with others' abilities---I could explain metaphor, theme, and meaning to people who never grasped it, and some of them took the time to explain logic to me. It was a fantastic opportunity, but one I very realistically had to earn with a lot of hard work and some convincing. </s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Even in fields like Computer Science and Engineering,<mask> you get the highest degree in the field, it's a Doctor of Philosophy degree in ____. This is important,<mask> it signifies that you have been taught not only facts in your discipline<mask><mask><mask> to think about the discipline. [NEWLINE] [NEWLINE] I'm getting my PhD in Statistics in a couple of months, and one of the best single classes I ever had was a discussion of<mask> "probability" means.<mask> it turns out, that simple point makes up a large part of the difference between the two "schools" of statistical thought - frequentists thing it's something that can be hypothetically observed (even<mask> not practically observed to the point<mask> you could get a continuous measurement), bayesians think probability represents the strength of your belief that something will occur. Neither one is wrong, and it's an arcane point<mask> you aren't a statistician,<mask> it's critically important<mask> you're going to be thinking about models and data with regards to probability. [NEWLINE] [NEWLINE] Without philosophy<mask> an academic discipline in and of itself, it would be a lot harder to build a framework for discussing these types of issues.<mask> I would agree that it's not something that everyone should have to major in,<mask><mask> that a general philosophy class (<mask> well taught) can do wonders for your ability to argue a point, to understand nuances and shades of meaning, and to think critically about a topic (even a highly theoretical point). [NEWLINE] [NEWLINE] It may not be relevant to your everyday life to discuss whether the world exists outside of our perception, whether our perception of the world is accurate, and whether there is any meaning to life at all - you still have to make dinner, do laundry, and go to work the next day; that doesn't mean that it isn't important to consider those questions. Doing<mask> led to cognitive psychology, neuroscience, and pragmatic approaches to treating people with depression. Philosophy is just the way we get from one idea to the next - some of these ideas lead to experiments, new innovations, practical knowledge. [NEWLINE] [NEWLINE] <mask>, philosophy can be fun... the philosophy club at my university, in the tradition of the ancients, has meetings<mask> they drink and argue until they're too drunk to argue. :)</s>
Label encoding: <s>Even in fields like Computer Science and Engineering, when you get the highest degree in the field, it's a Doctor of Philosophy degree in ____. This is important, because it signifies that you have been taught not only facts in your discipline but also how to think about the discipline. [NEWLINE] [NEWLINE] I'm getting my PhD in Statistics in a couple of months, and one of the best single classes I ever had was a discussion of what "probability" means. As it turns out, that simple point makes up a large part of the difference between the two "schools" of statistical thought - frequentists thing it's something that can be hypothetically observed (even if not practically observed to the point where you could get a continuous measurement), bayesians think probability represents the strength of your belief that something will occur. Neither one is wrong, and it's an arcane point if you aren't a statistician, but it's critically important if you're going to be thinking about models and data with regards to probability. [NEWLINE] [NEWLINE] Without philosophy as an academic discipline in and of itself, it would be a lot harder to build a framework for discussing these types of issues. While I would agree that it's not something that everyone should have to major in, I think that a general philosophy class ( if well taught) can do wonders for your ability to argue a point, to understand nuances and shades of meaning, and to think critically about a topic (even a highly theoretical point). [NEWLINE] [NEWLINE] It may not be relevant to your everyday life to discuss whether the world exists outside of our perception, whether our perception of the world is accurate, and whether there is any meaning to life at all - you still have to make dinner, do laundry, and go to work the next day; that doesn't mean that it isn't important to consider those questions. Doing so led to cognitive psychology, neuroscience, and pragmatic approaches to treating people with depression. Philosophy is just the way we get from one idea to the next - some of these ideas lead to experiments, new innovations, practical knowledge. [NEWLINE] [NEWLINE] Also, philosophy can be fun... the philosophy club at my university, in the tradition of the ancients, has meetings where they drink and argue until they're too drunk to argue. :)</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I understand<mask> you feel this way,<mask> I do disagree. I can only draw on my own life<mask> an example. [NEWLINE] [NEWLINE] The very fact that I enjoy being on CMV is that I find it refreshing and necessary for me to listen to and engage with people who hold other views. Being on Reddit is awesome,<mask> the majority of people on here do not believe<mask> I do about MANY things.<mask>... still a good thing, and a welcome thing in my life. I don't think I am an exception to this. [NEWLINE] [NEWLINE] Even on Facebook, I do advocate for people to respectfully listen to each other's point of view, and it is a great skill to learn. My family is a mix of Moderate Liberals, Socialist hippies and extreme Conservatives - all among my own family. You wouldn't be very happy<mask> you didn't learn to get along and listen, and respectfully disagree. A few arguments here and there, and then we work it out and move on. [NEWLINE] [NEWLINE] <mask><mask><mask> Internet dating goes, I just went on the "last blind date" of my life recently. At the age of 49 I just wasn't meeting people socially, and had a hard time just finding time to meet new people. Went on a blind date for coffee and met someone very different from me:  he loves hiking, backpacking, he is a truck driver, he was married before, he is Conservative, and has OCD. [NEWLINE] I am an Accountant, prefer indoors to outdoors, hate bugs and dirt, have never been married, am very liberal about most things and laid back about my things, my money, my house, etc. and guess<mask>? [NEWLINE] We hit it off, and just celebrated our one year wedding anniversary. [NEWLINE] [NEWLINE] Being with people who are just like you, who think like you and act like you... is BORING. Personal growth comes from challenges and every day I get a different perspective than my own that helps me see the world around me in a new light. And it's quite nice, actually. [NEWLINE] [NEWLINE] <mask>, no.<mask><mask> with you. Some people have issues with change and social issues,<mask> the Internet and Social media only shines light on that, it doesn't cause it. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I understand why you feel this way, but I do disagree. I can only draw on my own life as an example. [NEWLINE] [NEWLINE] The very fact that I enjoy being on CMV is that I find it refreshing and necessary for me to listen to and engage with people who hold other views. Being on Reddit is awesome, but the majority of people on here do not believe as I do about MANY things. But... still a good thing, and a welcome thing in my life. I don't think I am an exception to this. [NEWLINE] [NEWLINE] Even on Facebook, I do advocate for people to respectfully listen to each other's point of view, and it is a great skill to learn. My family is a mix of Moderate Liberals, Socialist hippies and extreme Conservatives - all among my own family. You wouldn't be very happy if you didn't learn to get along and listen, and respectfully disagree. A few arguments here and there, and then we work it out and move on. [NEWLINE] [NEWLINE] As far as Internet dating goes, I just went on the "last blind date" of my life recently. At the age of 49 I just wasn't meeting people socially, and had a hard time just finding time to meet new people. Went on a blind date for coffee and met someone very different from me:  he loves hiking, backpacking, he is a truck driver, he was married before, he is Conservative, and has OCD. [NEWLINE] I am an Accountant, prefer indoors to outdoors, hate bugs and dirt, have never been married, am very liberal about most things and laid back about my things, my money, my house, etc. and guess what? [NEWLINE] We hit it off, and just celebrated our one year wedding anniversary. [NEWLINE] [NEWLINE] Being with people who are just like you, who think like you and act like you... is BORING. Personal growth comes from challenges and every day I get a different perspective than my own that helps me see the world around me in a new light. And it's quite nice, actually. [NEWLINE] [NEWLINE] So, no. I disagree with you. Some people have issues with change and social issues, but the Internet and Social media only shines light on that, it doesn't cause it. [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that different skills have differing degrees to which people's natural "talent" for it makes it easier for them to learn it, for one thing. [NEWLINE] [NEWLINE] And sometimes, that puts an almost unassailable barrier in place for someone to get even to the level of "a decent practitioner" in the skill. [NEWLINE] [NEWLINE] Many of these things are really only something that you can do really well at<mask> you're young.<mask> you're not quick enough at learning to be a gymnast that you can get to a high level of skill<mask> you're still under 30, you'll probably never be good enough that someone would consider you a "professional gymnast". [NEWLINE] [NEWLINE] For another example, take physics. I went to arguably the best school in the world for this topic, Caltech, and physics was really the "elite" major that many many people really wanted to succeed at. Many many freshmen were *really* motivated, to the point<mask> they practically killed themselves putting in hours for homework, extra recitation sessions, professor office hours, the whole thing. [NEWLINE] [NEWLINE] And eventually... most of them fail.<mask> you can't learn fast enough in that particular field, and<mask> you don't have that "spark" that makes you stand out in the crowd, you'll never convince someone that you will make any significant new contribution to the field, and that's the bare minimum prerequisite for getting a thesis adviser, which is the bare minimum needed to get a PhD, which is the bare minimum needed to be considered a "professional physicist". [NEWLINE] [NEWLINE] I'm not talking about just the top level of the "elite" here. The minimum bar to be accepted in the profession is just fantastically high. [NEWLINE] [NEWLINE] It takes being *really* brilliant, *and* really motivated, *and* having that special spark that lets you visualize<mask>'s going on with the really weird ways that we now understand the world works. [NEWLINE] [NEWLINE] Math is similar. Can most anyone with a decent level of intelligence get good enough at math to be an accountant? Yeah, probably a lot of people can do that.<mask> being good enough to be "a mathematician"? Even a mediocre mathematician? That takes enough talent to have some hope of advancing the field.</s>
Label encoding: <s>I think that different skills have differing degrees to which people's natural "talent" for it makes it easier for them to learn it, for one thing. [NEWLINE] [NEWLINE] And sometimes, that puts an almost unassailable barrier in place for someone to get even to the level of "a decent practitioner" in the skill. [NEWLINE] [NEWLINE] Many of these things are really only something that you can do really well at when you're young. If you're not quick enough at learning to be a gymnast that you can get to a high level of skill while you're still under 30, you'll probably never be good enough that someone would consider you a "professional gymnast". [NEWLINE] [NEWLINE] For another example, take physics. I went to arguably the best school in the world for this topic, Caltech, and physics was really the "elite" major that many many people really wanted to succeed at. Many many freshmen were *really* motivated, to the point where they practically killed themselves putting in hours for homework, extra recitation sessions, professor office hours, the whole thing. [NEWLINE] [NEWLINE] And eventually... most of them fail. If you can't learn fast enough in that particular field, and if you don't have that "spark" that makes you stand out in the crowd, you'll never convince someone that you will make any significant new contribution to the field, and that's the bare minimum prerequisite for getting a thesis adviser, which is the bare minimum needed to get a PhD, which is the bare minimum needed to be considered a "professional physicist". [NEWLINE] [NEWLINE] I'm not talking about just the top level of the "elite" here. The minimum bar to be accepted in the profession is just fantastically high. [NEWLINE] [NEWLINE] It takes being *really* brilliant, *and* really motivated, *and* having that special spark that lets you visualize what's going on with the really weird ways that we now understand the world works. [NEWLINE] [NEWLINE] Math is similar. Can most anyone with a decent level of intelligence get good enough at math to be an accountant? Yeah, probably a lot of people can do that. But being good enough to be "a mathematician"? Even a mediocre mathematician? That takes enough talent to have some hope of advancing the field.</s>
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not a *bad* thing to do.<mask> it is a huge time commitment, and there can be better things to do with your time. [NEWLINE] [NEWLINE] <mask> are your goals? [NEWLINE] [NEWLINE] <mask> you're going for premed, all medical schools care about are your GPA and MCAT. They don't care at all about<mask> many classes taken or<mask> many degrees you've gotten. The type of GPA they're looking for is [STARTQ] 3.6, and ideally &gt;3.8, even at a competitive university. You'll probably have to settle for a lower ranked medical school<mask> it's any lower than that. [ENDQ] [NEWLINE] <mask> you're looking for work in engineering, internships are an absolute must.<mask><mask>, it reflects badly on you<mask> you have a large amount of classes<mask> minimal practical experience,<mask> it implies that you're not well-rounded.<mask> not internships, then at least undergraduate research is needed. I've talked to a professor who used to manage a company, and he told me that he once had an applicant who had straight A+'s through undergraduate<mask> minimal practical experience, and he rejected the applicant straight out<mask> the guy obviously had ample time to do actually useful things<mask> instead spent all his time on classes. [NEWLINE] [NEWLINE] And another thing about internships and undergraduate research: succeeding at these are far far more fulfilling than succeeding in classes. In classes, problems and projects are usually designed<mask> that they complement<mask> you learned.<mask> you do internships or research, you'll find that the things you did in class feel a lot more contrived and fake<mask> of that. You'll be getting a depth of knowledge and work on solving actual problems. [NEWLINE] [NEWLINE] And<mask> you're looking towards graduate school (this is the thing I can speak most about,<mask> I'm in it), you absolutely need undergraduate research experience to get into even a mediocre graduate school. The graduate chair at my graduate school told our group before our interviews, "We don't need someone that can fill out a test.<mask> use is someone who can fill out a test? Does filling out a test make anything? Did you cure a disease or make a product by filling out a test? Then who cares about classes!"</s>
Label encoding: <s>It's not a *bad* thing to do. But it is a huge time commitment, and there can be better things to do with your time. [NEWLINE] [NEWLINE] What are your goals? [NEWLINE] [NEWLINE] If you're going for premed, all medical schools care about are your GPA and MCAT. They don't care at all about how many classes taken or how many degrees you've gotten. The type of GPA they're looking for is [STARTQ] 3.6, and ideally &gt;3.8, even at a competitive university. You'll probably have to settle for a lower ranked medical school if it's any lower than that. [ENDQ] [NEWLINE] If you're looking for work in engineering, internships are an absolute must. In fact, it reflects badly on you if you have a large amount of classes but minimal practical experience, since it implies that you're not well-rounded. If not internships, then at least undergraduate research is needed. I've talked to a professor who used to manage a company, and he told me that he once had an applicant who had straight A+'s through undergraduate but minimal practical experience, and he rejected the applicant straight out because the guy obviously had ample time to do actually useful things but instead spent all his time on classes. [NEWLINE] [NEWLINE] And another thing about internships and undergraduate research: succeeding at these are far far more fulfilling than succeeding in classes. In classes, problems and projects are usually designed so that they complement what you learned. When you do internships or research, you'll find that the things you did in class feel a lot more contrived and fake because of that. You'll be getting a depth of knowledge and work on solving actual problems. [NEWLINE] [NEWLINE] And if you're looking towards graduate school (this is the thing I can speak most about, since I'm in it), you absolutely need undergraduate research experience to get into even a mediocre graduate school. The graduate chair at my graduate school told our group before our interviews, "We don't need someone that can fill out a test. What use is someone who can fill out a test? Does filling out a test make anything? Did you cure a disease or make a product by filling out a test? Then who cares about classes!"</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I live in Denmark<mask> the priority is rehabilitation instead of punishment. [NEWLINE] <mask> in America<mask><mask> that crime is based around, and the laws about crime are based around massmurderers, it's a big country, with many criminals<mask> the amount of mass murderers is big<mask> well.<mask> the media will focus in on these tradegies, like Columbine, Sandy Hook, and whatever mass murder there have been made. That puts it out of perspective and then people will think that the death penalty is needed,<mask> they think that they have<mask> many dangerous people. It gets blown out of proportion and it fucks it all up. [NEWLINE] [NEWLINE] Now with the death penalty there is 1 good thing about it. It kills the criminal and they can no longer directly hurt people. The bad things it does<mask> far outweighs it. It doesn't really help to kill the person<mask> the idealogy and thought process often stay in the country, and then some weak-minded person sees it and agrees, which might (emphasis on the might) create another killer. [NEWLINE] With the death penalty you can risk killing an innocent person (which is a quite big thing in the States with overzealous prosecuters just wanting to get a promotion), which<mask><mask><mask> makes every pro-deathpenalty fella a murderer (indirectly). [NEWLINE] [NEWLINE] It's a hard subject to discuss and to fix<mask> it is complicated. Do you focus on the family of the victim? Or do you try<mask> much<mask> possible to fix the criminals,<mask> let's face it, most criminals are just people in trouble (mental trouble<mask> in being raised by them, or pushed into the trouble with being extremely poor). [NEWLINE] <mask><mask> that rehabilitation works best in general,<mask> I don't think that rehabilitation works on people like Jeffret Dahmer, Wayne Gacy, Charles Manson, and very intelligent killers. [NEWLINE] It's better to base your judicial system around the mayority of crime and not the rare massmurder committed by a psycho with a messed up world view. [NEWLINE] [NEWLINE] Sorry for being harsh before, I'm a blunt person and<mask> people misunderstand me I make myself<mask> clear<mask> possible, and many times<mask> very condescending. Apologies for that.</s>
Label encoding: <s>I live in Denmark where the priority is rehabilitation instead of punishment. [NEWLINE] But in America I think that crime is based around, and the laws about crime are based around massmurderers, it's a big country, with many criminals so the amount of mass murderers is big as well. So the media will focus in on these tradegies, like Columbine, Sandy Hook, and whatever mass murder there have been made. That puts it out of perspective and then people will think that the death penalty is needed, because they think that they have so many dangerous people. It gets blown out of proportion and it fucks it all up. [NEWLINE] [NEWLINE] Now with the death penalty there is 1 good thing about it. It kills the criminal and they can no longer directly hurt people. The bad things it does however far outweighs it. It doesn't really help to kill the person because the idealogy and thought process often stay in the country, and then some weak-minded person sees it and agrees, which might (emphasis on the might) create another killer. [NEWLINE] With the death penalty you can risk killing an innocent person (which is a quite big thing in the States with overzealous prosecuters just wanting to get a promotion), which in my opinion makes every pro-deathpenalty fella a murderer (indirectly). [NEWLINE] [NEWLINE] It's a hard subject to discuss and to fix because it is complicated. Do you focus on the family of the victim? Or do you try as much as possible to fix the criminals, because let's face it, most criminals are just people in trouble (mental trouble as in being raised by them, or pushed into the trouble with being extremely poor). [NEWLINE] I think that rehabilitation works best in general, but I don't think that rehabilitation works on people like Jeffret Dahmer, Wayne Gacy, Charles Manson, and very intelligent killers. [NEWLINE] It's better to base your judicial system around the mayority of crime and not the rare massmurder committed by a psycho with a messed up world view. [NEWLINE] [NEWLINE] Sorry for being harsh before, I'm a blunt person and when people misunderstand me I make myself as clear as possible, and many times also very condescending. Apologies for that.</s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 170629 injured Canadians in 2010 might disagree with your statement that driving doesn't directly harm people. Dumping oil in my lake wouldn't affect nearly that many people. [ENDQ] [NEWLINE] **My** driving doesn't directly harm people. It doesn't even directly harm people (save for pollution that it creates). Call me naive<mask> I only worry about<mask> it is I do,<mask> I'm the only person that I can (and should be able to) control. [NEWLINE] [NEWLINE] [STARTQ] The point isn't that you are in a self driving car, it's that everyone is in a self driving car, which would make your driving experience immeasurably safer. [ENDQ] [NEWLINE] <mask> are you suggesting that I wouldn't be<mask> everyone else is?<mask> that's the case then I'm totally open to the idea.<mask>,<mask> I'm included in the everyone then I don't care that the driving experience is more safe simply<mask> it hasn't been unsafe for me<mask> far. I used the example on another comment,<mask> it's similar to trying to sell someone in Alberta tsunami insurance. They've never been threatened by it and don't see themselves threatened by it,<mask><mask> should they buy into it<mask> it would be "safer"? [NEWLINE] [NEWLINE] [STARTQ] I have to pay taxes<mask> others can be on welfare<mask> it's not benefit to me. There are plenty of things that we<mask> a society have decided are worth it and we pay for. [ENDQ] [NEWLINE] Again, another argument for another time,<mask> don't use the collective "we" assuming it applies to everyone in society. Some people in society think it's worth it,<mask> not everyone. There were a number of times in history that a society decided something was right, even<mask> it's citizens (and sometimes the majority) would disagree. [NEWLINE] [NEWLINE] [STARTQ] You would forgo this for everyone<mask> you enjoy driving? [ENDQ] [NEWLINE] Yes. At least<mask> far I would. There's been absolutely no benefit to me personally that I can derive from anything you've said. Again, I'm all for safer roads<mask> we put others in self-driving cars,<mask><mask> should I the individual be open to the idea<mask> I don't have a problem driving my own car?</s>
Label encoding: <s> [STARTQ] 170629 injured Canadians in 2010 might disagree with your statement that driving doesn't directly harm people. Dumping oil in my lake wouldn't affect nearly that many people. [ENDQ] [NEWLINE] **My** driving doesn't directly harm people. It doesn't even directly harm people (save for pollution that it creates). Call me naive but I only worry about what it is I do, because I'm the only person that I can (and should be able to) control. [NEWLINE] [NEWLINE] [STARTQ] The point isn't that you are in a self driving car, it's that everyone is in a self driving car, which would make your driving experience immeasurably safer. [ENDQ] [NEWLINE] So are you suggesting that I wouldn't be while everyone else is? If that's the case then I'm totally open to the idea. However, if I'm included in the everyone then I don't care that the driving experience is more safe simply because it hasn't been unsafe for me thus far. I used the example on another comment, but it's similar to trying to sell someone in Alberta tsunami insurance. They've never been threatened by it and don't see themselves threatened by it, so why should they buy into it because it would be "safer"? [NEWLINE] [NEWLINE] [STARTQ] I have to pay taxes so others can be on welfare but it's not benefit to me. There are plenty of things that we as a society have decided are worth it and we pay for. [ENDQ] [NEWLINE] Again, another argument for another time, but don't use the collective "we" assuming it applies to everyone in society. Some people in society think it's worth it, but not everyone. There were a number of times in history that a society decided something was right, even if it's citizens (and sometimes the majority) would disagree. [NEWLINE] [NEWLINE] [STARTQ] You would forgo this for everyone because you enjoy driving? [ENDQ] [NEWLINE] Yes. At least so far I would. There's been absolutely no benefit to me personally that I can derive from anything you've said. Again, I'm all for safer roads if we put others in self-driving cars, but why should I the individual be open to the idea if I don't have a problem driving my own car?</s>
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Based on my experiences,<mask><mask> with some of your remarks on *<mask> * education should do,<mask><mask><mask> you have the *<mask> * wrong. <mask> I fully agree with you that education in this country doesn't stress the development of critical thinking skills, this is precisely<mask> there *is* a turn towards trying to break down the student-teacher dichotomy and construct a discourse based around the students' own experiences.  The way to get people to think critically about the world around them is not to let them wallow in the own experience,<mask> rather to force them to consider things from new perspectives. <mask> a grad student I've TAed at several top tier universities - schools that presumably attract the "cream of the crop" -<mask> in even in these universities many of the students are entirely incapable of thinking beyond the boundaries of their own lives.  Learning to think critically requires something external pushing the learner far beyond their comfort zone, it requires something showing them that many things they think are clear cases of black &amp; white are,<mask><mask>, entirely gray.  This can't happen unless there is a division between student and teacher, with the latter regarded<mask> a source of knowledge.  (It is<mask>, of course, critical that the teacher is dedicated to their job and puts in the time and effort to be very deliberate with their pedagogy). <mask> students see their teachers<mask> equals or friends, they will consider their own ideas<mask> on par with they're being taught,<mask> the fact that<mask> they're being taught is the culmination of centuries of human thought and experimentation. [NEWLINE] [NEWLINE] Yes, there are certainly cases<mask> creativity should be fostered (in arts, music, writing classes),<mask><mask> I consistently deal with students who insist on reading fifth century texts through their own contemporary lenses - and then insist that their own ideas (born out of an utter lack of knowledge on the subject) are no less reasonable than ideas developed by scholars who have devoted their entire lives to learning everything about the subject, it represents a failure of education, and it's a failure which results from teaching students that all ideas are equally valid, even ideas born of a lack of knowledge.</s>
Label encoding: <s>Based on my experiences, I agree with some of your remarks on * what * education should do, but I think you have the * how * wrong.  While I fully agree with you that education in this country doesn't stress the development of critical thinking skills, this is precisely because there *is* a turn towards trying to break down the student-teacher dichotomy and construct a discourse based around the students' own experiences.  The way to get people to think critically about the world around them is not to let them wallow in the own experience, but rather to force them to consider things from new perspectives.  As a grad student I've TAed at several top tier universities - schools that presumably attract the "cream of the crop" - yet in even in these universities many of the students are entirely incapable of thinking beyond the boundaries of their own lives.  Learning to think critically requires something external pushing the learner far beyond their comfort zone, it requires something showing them that many things they think are clear cases of black &amp; white are, in fact, entirely gray.  This can't happen unless there is a division between student and teacher, with the latter regarded as a source of knowledge.  (It is also, of course, critical that the teacher is dedicated to their job and puts in the time and effort to be very deliberate with their pedagogy).  If students see their teachers as equals or friends, they will consider their own ideas as on par with they're being taught, despite the fact that what they're being taught is the culmination of centuries of human thought and experimentation. [NEWLINE] [NEWLINE] Yes, there are certainly cases where creativity should be fostered (in arts, music, writing classes), but when I consistently deal with students who insist on reading fifth century texts through their own contemporary lenses - and then insist that their own ideas (born out of an utter lack of knowledge on the subject) are no less reasonable than ideas developed by scholars who have devoted their entire lives to learning everything about the subject, it represents a failure of education, and it's a failure which results from teaching students that all ideas are equally valid, even ideas born of a lack of knowledge.</s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am stuck in a rut at my current job and need to do something about it. I want to actually do something I feel like I can make a difference with (I really want to be a vet)<mask> I grew up in a household<mask> money was always tight and learned to not bite off more than I can chew. Being in a debt and owing people money is a huge deal to me and I try to avoid it at all costs. [NEWLINE] [NEWLINE] I wasn't to go to school to be a vet<mask> I feel like it will only make things worse by piling on student debts or not being able to find a job in my field (or both).<mask> accurate is this assumption? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] **EDIT: I thank you all for replying to this topic. I honestly wasnt expecting to change the profession<mask> looking at these posts and knowing myself, I would be better suited to studying to be a VET TECH and enroll at community college for the first couple of years and transfer.  I am 26 living in Utah and have put off doing any sort of higher education just based on the whole debt/job field issue<mask> I am going to start making some changes and speak to some career advisors at SLCC (the local community college here) and get started. [NEWLINE] I cant thank you all enough for all the advice given to me!!** [NEWLINE] ____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I am stuck in a rut at my current job and need to do something about it. I want to actually do something I feel like I can make a difference with (I really want to be a vet) but I grew up in a household where money was always tight and learned to not bite off more than I can chew. Being in a debt and owing people money is a huge deal to me and I try to avoid it at all costs. [NEWLINE] [NEWLINE] I wasn't to go to school to be a vet but I feel like it will only make things worse by piling on student debts or not being able to find a job in my field (or both). How accurate is this assumption? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] **EDIT: I thank you all for replying to this topic. I honestly wasnt expecting to change the profession but looking at these posts and knowing myself, I would be better suited to studying to be a VET TECH and enroll at community college for the first couple of years and transfer.  I am 26 living in Utah and have put off doing any sort of higher education just based on the whole debt/job field issue but I am going to start making some changes and speak to some career advisors at SLCC (the local community college here) and get started. [NEWLINE] I cant thank you all enough for all the advice given to me!!** [NEWLINE] ____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've spent the last ten years in sales. Five of those were<mask> am appliance salesperson. The other five<mask> a CSR. You wouldn't believe the number of uninformed customers I help daily. Whether I was  commissioned it not, I quickly learned the best long term sales technique is to sell the customer<mask> they want/need, no more, and no less. This creates customer loyalty. In order to do that, I need to make sure they know<mask> they are getting into. [NEWLINE] [NEWLINE] For example, your "informed" customer loves Consumer Reports. With appliances, they are quite inaccurate. Once they rated a gas range<mask> #1 and the same model in a different color<mask> #8. I even compared parts diagrams to be sure. They were identical other than the paint. These are your "informed" customers. The rest know even less. [NEWLINE] [NEWLINE] Now, I work in a non commissioned sales position. The majority of my colleagues agree I am better at my job than the rest of the department combined. This is<mask> I strive to be the best at anything I do. I have six other people in my department who do not. All they care about is making the customer go away. [NEWLINE] [NEWLINE] <mask> am I useful? For starters, I physically show people<mask> shitty the $11 gallon of paint is compared to the $20 gallon. Without a salesperson, commissioned or not, many customers will grab the cheapest can on the shelf<mask> they are "all the  same." Now they use this paint, have terrible coverage, and need to spend more to finish the job. Now they're mad,<mask> they  buy  a different product at a competitor. [NEWLINE] [NEWLINE] My job is to explain the features and benefits of products and let them choose. Without the information I provide (which comes from years of experience), people will not end up with the products they actually want. [NEWLINE] [NEWLINE] <mask> nothing else, most people still want someone to talk to.<mask> I work (a large retail chain) the number one complaint is not being able to find a salesperson. Maybe you don't need a salesperson,<mask> the majority of people still do. </s>
Label encoding: <s>I've spent the last ten years in sales. Five of those were as am appliance salesperson. The other five as a CSR. You wouldn't believe the number of uninformed customers I help daily. Whether I was  commissioned it not, I quickly learned the best long term sales technique is to sell the customer what they want/need, no more, and no less. This creates customer loyalty. In order to do that, I need to make sure they know what they are getting into. [NEWLINE] [NEWLINE] For example, your "informed" customer loves Consumer Reports. With appliances, they are quite inaccurate. Once they rated a gas range as #1 and the same model in a different color as #8. I even compared parts diagrams to be sure. They were identical other than the paint. These are your "informed" customers. The rest know even less. [NEWLINE] [NEWLINE] Now, I work in a non commissioned sales position. The majority of my colleagues agree I am better at my job than the rest of the department combined. This is because I strive to be the best at anything I do. I have six other people in my department who do not. All they care about is making the customer go away. [NEWLINE] [NEWLINE] Why am I useful? For starters, I physically show people how shitty the $11 gallon of paint is compared to the $20 gallon. Without a salesperson, commissioned or not, many customers will grab the cheapest can on the shelf because they are "all the  same." Now they use this paint, have terrible coverage, and need to spend more to finish the job. Now they're mad, so they  buy  a different product at a competitor. [NEWLINE] [NEWLINE] My job is to explain the features and benefits of products and let them choose. Without the information I provide (which comes from years of experience), people will not end up with the products they actually want. [NEWLINE] [NEWLINE] If nothing else, most people still want someone to talk to. Where I work (a large retail chain) the number one complaint is not being able to find a salesperson. Maybe you don't need a salesperson, but the majority of people still do. </s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I can only begin to imagine the terrible gambler's fallacy that must be involved<mask> you have leveled up an awesome character and you're loathe to part with it forever [ENDQ] [NEWLINE] I've thought about this one a lot.  I don't think an awesome leveled character makes the game more addictive.  For two examples: World of Warcraft and Fallout 3. [NEWLINE] [NEWLINE] I have 3070 hours on my main in wow, with a whole slew of achievements, and with gear near max level for the current expansion. <mask> the thing that keeps me playing has nothing to do with<mask> awesome my character is.  It's<mask> there is to do with my character.  I've played all the content in the latest patch and I'm considering walking away<mask> there's not much to do. <mask> I lose my character permanently, well, there are other games I can play.  The max level is really independent of the fun you have playing content, and the entire game reinforces that. [NEWLINE] [NEWLINE] This is made even more obvious with fallout 3.  To get a perfect level 30 character (all stats at 10, all skills at 100, all the good perks for your build, all the side perks optimized) takes an investment of about 40 hours (for me).  Once you've got it,<mask>, the fun's over.  You can play on your super character and kill some wimpy raiders, maybe get chased around by some albino radscorpions. <mask><mask> Bethesda released another DLC unlocking 10 more levels, would<mask><mask> that now I have to buy it<mask> otherwise my level 30 character goes to waste?  Well, no.  Their game has taught me pretty effectively<mask> their game is about.  I know<mask> fun I'm going to have playing new content in the game, and I can pretty effectively get this same fun replaying the original content from level 1 to 30 (<mask> the game is about progression and collection for me). [NEWLINE] [NEWLINE] <mask><mask> most games in the industry have this property.  Collection and progression is one of the elements of<mask> makes them fun; it's not the keystone of addiction.</s>
Label encoding: <s> [STARTQ] I can only begin to imagine the terrible gambler's fallacy that must be involved if you have leveled up an awesome character and you're loathe to part with it forever [ENDQ] [NEWLINE] I've thought about this one a lot.  I don't think an awesome leveled character makes the game more addictive.  For two examples: World of Warcraft and Fallout 3. [NEWLINE] [NEWLINE] I have 3070 hours on my main in wow, with a whole slew of achievements, and with gear near max level for the current expansion.  But the thing that keeps me playing has nothing to do with how awesome my character is.  It's what there is to do with my character.  I've played all the content in the latest patch and I'm considering walking away because there's not much to do.  If I lose my character permanently, well, there are other games I can play.  The max level is really independent of the fun you have playing content, and the entire game reinforces that. [NEWLINE] [NEWLINE] This is made even more obvious with fallout 3.  To get a perfect level 30 character (all stats at 10, all skills at 100, all the good perks for your build, all the side perks optimized) takes an investment of about 40 hours (for me).  Once you've got it, though, the fun's over.  You can play on your super character and kill some wimpy raiders, maybe get chased around by some albino radscorpions.  But if Bethesda released another DLC unlocking 10 more levels, would I think that now I have to buy it because otherwise my level 30 character goes to waste?  Well, no.  Their game has taught me pretty effectively what their game is about.  I know what fun I'm going to have playing new content in the game, and I can pretty effectively get this same fun replaying the original content from level 1 to 30 ( since the game is about progression and collection for me). [NEWLINE] [NEWLINE] I think most games in the industry have this property.  Collection and progression is one of the elements of what makes them fun; it's not the keystone of addiction.</s>
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The company may potentially be moving "dangerous" people into the house. I live in a very quite residential neighborhood, and the thought of drug addicts being housed near me is disturbing. Even<mask> the residents are well behaved, there is going to be traffic of their "friends" who I'm assuming will be similar in nature. The company won't say who may be living there. It could be just infirm elderly people,<mask> it could<mask> be "sexual predators" for all I know. [ENDQ] [NEWLINE] <mask> the group home is for recovering addicts, you can assume some sort of supervision.<mask> do you fear from an addict under supervision? These aren't dealers or drug lords. I'm not quite sure the "friends" would hang around the neighborhood and cause trouble [NEWLINE] [NEWLINE] <mask>, I've never heard of a group home for sexual predators. You are making things up now. [NEWLINE] [NEWLINE] [STARTQ] My house will lose value<mask> people find out that it is near a group home. Even<mask> it isn't logical, I'm<mask><mask><mask> someone is looking to buy my house they will be put off by the presence of group home nearby. [ENDQ] [NEWLINE] Are you required to disclose the presence of the group home? There is no reason they need to know. [NEWLINE] [NEWLINE] [STARTQ] The business buying the house is a tax-free organization. I don't exactly know<mask> that means,<mask> I assume that they won't be paying property taxes. That means that they will be consuming resources<mask> not paying for them, and my share of the community taxes will increase. [ENDQ] [NEWLINE] <mask> type of unpaid resources will they be consuming? Property taxes pay for things like schools and libraries. This group home is providing a public service and helping to reduce crime and drug usage. [NEWLINE] [NEWLINE] [STARTQ] Even<mask> this particular place only houses people who are complete angels, this will set a precedent and open up the neighborhood for other homes, and there is no guarantee that the other homes will be run<mask> well. [ENDQ] [NEWLINE] <mask> are you worried about precedent? Your neighborhood is already allowing these types of home<mask> no precedent is necessary. Group homes don't need to compete like stores or restaurants. </s>
Label encoding: <s> [STARTQ] The company may potentially be moving "dangerous" people into the house. I live in a very quite residential neighborhood, and the thought of drug addicts being housed near me is disturbing. Even if the residents are well behaved, there is going to be traffic of their "friends" who I'm assuming will be similar in nature. The company won't say who may be living there. It could be just infirm elderly people, but it could also be "sexual predators" for all I know. [ENDQ] [NEWLINE] If the group home is for recovering addicts, you can assume some sort of supervision. What do you fear from an addict under supervision? These aren't dealers or drug lords. I'm not quite sure the "friends" would hang around the neighborhood and cause trouble [NEWLINE] [NEWLINE] Secondly, I've never heard of a group home for sexual predators. You are making things up now. [NEWLINE] [NEWLINE] [STARTQ] My house will lose value when people find out that it is near a group home. Even if it isn't logical, I'm assuming that when someone is looking to buy my house they will be put off by the presence of group home nearby. [ENDQ] [NEWLINE] Are you required to disclose the presence of the group home? There is no reason they need to know. [NEWLINE] [NEWLINE] [STARTQ] The business buying the house is a tax-free organization. I don't exactly know what that means, but I assume that they won't be paying property taxes. That means that they will be consuming resources but not paying for them, and my share of the community taxes will increase. [ENDQ] [NEWLINE] What type of unpaid resources will they be consuming? Property taxes pay for things like schools and libraries. This group home is providing a public service and helping to reduce crime and drug usage. [NEWLINE] [NEWLINE] [STARTQ] Even if this particular place only houses people who are complete angels, this will set a precedent and open up the neighborhood for other homes, and there is no guarantee that the other homes will be run as well. [ENDQ] [NEWLINE] Why are you worried about precedent? Your neighborhood is already allowing these types of home so no precedent is necessary. Group homes don't need to compete like stores or restaurants. </s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I posted on a similar thread a few days ago and continued to think about it.  One of the things I brought up<mask> didn't really unpack is the difference in values and the limited time we have available in a day, and<mask> these two interact. [NEWLINE] [NEWLINE] It takes time and energy to exercise, eat a balanced diet and maintain a healthy weight.  For some people, this is more than others,<mask> that's somewhat beside the point. <mask> person X does not particularly value diet/exercise/not being fat, then that person will not invest their time and energy into this endeavor.  They may be spending time doing other things they subjectively view<mask> more worthwhile (working longer hours, volunteering, caring for family, playing videogames, etc.). [NEWLINE] [NEWLINE] <mask> another example, my apartment is frequently messy.  Apparently I don't value neatness enough to spend enough of my time cleaning up in order to have a neat and organized apartment.  Does this mean that I am stupid or lazy?  Some people might think<mask>.  I would say that I work long hours and I'm tired<mask> I get home and prefer to watch TV and play with my cat.  It doesn't matter to me<mask> my apartment is neat.  For person X who is fat, maybe it doesn't matter to them that they are fat (or it doesn't matter enough to invest their resources in changing it) -<mask> that is a pretty socially taboo thing to say in current culture anyway. [NEWLINE] [NEWLINE] We could come up with thousands examples of differences in values/<mask> people want to spend their time and<mask> that can possibly adversely influence their life. <mask><mask><mask>, the only reason this is such a big issue with obesity is that you can see it on their body and in this culture 'fat' is an undesirable state.  There is plainly visible evidence that a person POSSIBLY (<mask> we<mask> can't know just by looking at someone<mask> they are fat) has a different set of values.  They may be doing something more or less 'important' with their time, and the importance of that thing is subjective.</s>
Label encoding: <s>I posted on a similar thread a few days ago and continued to think about it.  One of the things I brought up but didn't really unpack is the difference in values and the limited time we have available in a day, and how these two interact. [NEWLINE] [NEWLINE] It takes time and energy to exercise, eat a balanced diet and maintain a healthy weight.  For some people, this is more than others, but that's somewhat beside the point.  If person X does not particularly value diet/exercise/not being fat, then that person will not invest their time and energy into this endeavor.  They may be spending time doing other things they subjectively view as more worthwhile (working longer hours, volunteering, caring for family, playing videogames, etc.). [NEWLINE] [NEWLINE] As another example, my apartment is frequently messy.  Apparently I don't value neatness enough to spend enough of my time cleaning up in order to have a neat and organized apartment.  Does this mean that I am stupid or lazy?  Some people might think so.  I would say that I work long hours and I'm tired when I get home and prefer to watch TV and play with my cat.  It doesn't matter to me if my apartment is neat.  For person X who is fat, maybe it doesn't matter to them that they are fat (or it doesn't matter enough to invest their resources in changing it) - but that is a pretty socially taboo thing to say in current culture anyway. [NEWLINE] [NEWLINE] We could come up with thousands examples of differences in values/ how people want to spend their time and how that can possibly adversely influence their life.  In my opinion, the only reason this is such a big issue with obesity is that you can see it on their body and in this culture 'fat' is an undesirable state.  There is plainly visible evidence that a person POSSIBLY ( because we also can't know just by looking at someone why they are fat) has a different set of values.  They may be doing something more or less 'important' with their time, and the importance of that thing is subjective.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here is the problem with this attitude... it is pure, unadulterated entitlement. You don't want to be represented in a democratic system, you want a democratic system to represent you and the will of the majority be damned. [NEWLINE] [NEWLINE] You aren't entitled to a representative who agrees with you, you have no right to guarantee that your view will be represented in federal politics. It's the point of representative democracy, ALL politicians are going to tend toward the centre based on<mask> views are commonly held and your view is an extreme, untenable position that is never going to get popular support in the current political climate. [NEWLINE] [NEWLINE] You are essentially saying that your vote is worthless, not<mask> it isn't counted,<mask><mask> the side you agree with loses the election. And this is the flaw in<mask> you are espousing. Your vote isn't worthless, it is is worth exactly<mask> much<mask> everyone elses. The fact you choose to squander your vote on a position that you know is going to fail is a failure on your part to use your vote effectively, not a demonstration that your vote is worthless. [NEWLINE] [NEWLINE] It is completely within your power to change it. First, you don't have 1 vote, you have several. The general election is only one battle,<mask> primaries have far lower turnout and many states let you vote in one or even both. That is your chance, vote for a candidate who espouses your views in the primary and you show that a person with those views has support. Democracy on most issues isn't binary, it's a bell curve... every vote shifts the curve a certain direction and the peak of the curve is the prevailing political wisdom, which shifts all the time. You are only<mask> powerless<mask> you choose to be. [NEWLINE] [NEWLINE] TL;DR: Your view is equating the right to vote with a right to have your views represented and that is not the way it works. You are guaranteed a vote, and that vote is not worthless, it simply does not guarantee that your chosen candidate is going to win. You have a right to vote, that vote is only worthless<mask> you misuse it </s>
Label encoding: <s>Here is the problem with this attitude... it is pure, unadulterated entitlement. You don't want to be represented in a democratic system, you want a democratic system to represent you and the will of the majority be damned. [NEWLINE] [NEWLINE] You aren't entitled to a representative who agrees with you, you have no right to guarantee that your view will be represented in federal politics. It's the point of representative democracy, ALL politicians are going to tend toward the centre based on what views are commonly held and your view is an extreme, untenable position that is never going to get popular support in the current political climate. [NEWLINE] [NEWLINE] You are essentially saying that your vote is worthless, not because it isn't counted, but because the side you agree with loses the election. And this is the flaw in what you are espousing. Your vote isn't worthless, it is is worth exactly as much as everyone elses. The fact you choose to squander your vote on a position that you know is going to fail is a failure on your part to use your vote effectively, not a demonstration that your vote is worthless. [NEWLINE] [NEWLINE] It is completely within your power to change it. First, you don't have 1 vote, you have several. The general election is only one battle, but primaries have far lower turnout and many states let you vote in one or even both. That is your chance, vote for a candidate who espouses your views in the primary and you show that a person with those views has support. Democracy on most issues isn't binary, it's a bell curve... every vote shifts the curve a certain direction and the peak of the curve is the prevailing political wisdom, which shifts all the time. You are only as powerless as you choose to be. [NEWLINE] [NEWLINE] TL;DR: Your view is equating the right to vote with a right to have your views represented and that is not the way it works. You are guaranteed a vote, and that vote is not worthless, it simply does not guarantee that your chosen candidate is going to win. You have a right to vote, that vote is only worthless if you misuse it </s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I can understand your feelings on this, let's be fair about a few points. [NEWLINE] [NEWLINE] [STARTQ] Gay people can't get married -<mask>?<mask> God says it's a sin. [ENDQ] [NEWLINE] This has more to do with the fact that it has never been societally accepted in almost any culture under religion or atheism before roughly ten years ago. Give it time. [NEWLINE] [NEWLINE] [STARTQ] God says the world is 6000 years old<mask> we shouldn't teach evolution in schools.<mask><mask> about the overwhelming - virtually undeniable evidence that supports evolution, or Earth being billions of years old? Doesn't matter,<mask> God. [ENDQ] [NEWLINE] Not sure<mask> you're saying here,<mask> creationism isn't taught in schools.<mask> at any rate, please don't get the impression that all Christians believe this. Maybe many do<mask> you're from<mask> lots of us believe in Evolution right alongside you. A comment about babies and bathwater might be appropriate here. [NEWLINE] [NEWLINE] [STARTQ] That which can be asserted without evidence can be dismissed without evidence [ENDQ] [STARTQ] [ENDQ] [STARTQ] everything I see around me *tells me that religion is man-made*, and it frustrates me that in the 21st century we have to bend to the will of those whose belief requires no substantiation. [ENDQ] [NEWLINE] (emphasis mine, naturally) I read your second statement to mean roughly that the'sense you get' is that religion is man-made. In essence, you believe it.<mask> you may have evopsych explanations that make sense for you in terms of<mask> religion might have evolved, you will have no evidence to prove that it did<mask> come to be by that way.<mask> of that, your belief falls under the same limitations<mask> mine, and is not superior to it. We have both chosen to believe for our own reasons, not all of them based on evidence, and we both think we're the one that's right. [NEWLINE] [NEWLINE] <mask> that makes you feel bad, just think about this. My belief may enjoy historical popularity, at least on a surface level,<mask> yours most likely will enjoy that popularity in the future,<mask> you have that to look forward to. [NEWLINE] </s>
Label encoding: <s>While I can understand your feelings on this, let's be fair about a few points. [NEWLINE] [NEWLINE] [STARTQ] Gay people can't get married - why? Because God says it's a sin. [ENDQ] [NEWLINE] This has more to do with the fact that it has never been societally accepted in almost any culture under religion or atheism before roughly ten years ago. Give it time. [NEWLINE] [NEWLINE] [STARTQ] God says the world is 6000 years old so we shouldn't teach evolution in schools. But what about the overwhelming - virtually undeniable evidence that supports evolution, or Earth being billions of years old? Doesn't matter, because God. [ENDQ] [NEWLINE] Not sure what you're saying here, as creationism isn't taught in schools. But at any rate, please don't get the impression that all Christians believe this. Maybe many do where you're from but lots of us believe in Evolution right alongside you. A comment about babies and bathwater might be appropriate here. [NEWLINE] [NEWLINE] [STARTQ] That which can be asserted without evidence can be dismissed without evidence [ENDQ] [STARTQ] [ENDQ] [STARTQ] everything I see around me *tells me that religion is man-made*, and it frustrates me that in the 21st century we have to bend to the will of those whose belief requires no substantiation. [ENDQ] [NEWLINE] (emphasis mine, naturally) I read your second statement to mean roughly that the'sense you get' is that religion is man-made. In essence, you believe it. While you may have evopsych explanations that make sense for you in terms of how religion might have evolved, you will have no evidence to prove that it did indeed come to be by that way. Because of that, your belief falls under the same limitations as mine, and is not superior to it. We have both chosen to believe for our own reasons, not all of them based on evidence, and we both think we're the one that's right. [NEWLINE] [NEWLINE] If that makes you feel bad, just think about this. My belief may enjoy historical popularity, at least on a surface level, but yours most likely will enjoy that popularity in the future, so you have that to look forward to. [NEWLINE] </s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's something I can attest to<mask> a guy who spent a good bit of his younger years in a private school. [NEWLINE] A lot of good teachers quit from the public schools due to red tape and excessive amounts of paperwork involved. These teachers, the ones of talked to (who are some of the best I've ever had)<mask><mask> they much prefer teaching<mask> they have more freedom to be creative and instill real knowledge and fun activities, not run through a curriculum<mask> quickly<mask> possible to meet state standards. I worked an internship last year with a teacher at a smaller private school and he was able to have his students do debates, projects, plays, and art. He was able to make them love and respect him, and they all learned a lot-Even the at-risk kids. I can tell you it's a lot more effective than handing out a blue diamond book every week and teaching to the test. [NEWLINE] [NEWLINE] <mask> that doesn't get you, here's another point-The public school near my house was massive and I was barely even a number to them. Private schools are much smaller and more personal. Don't you think a lot of at-risk kids would benefit from having their teacher know them<mask> more than just a name on a list of grades with 500-1000 other names on it? They don't have<mask> much opportunity to sit in the back of the class and get away with murder. It<mask> helps foster discussion,<mask> smaller class sizes are easier to work with than massive ones. [NEWLINE] [NEWLINE] <mask>, in regards to bullying. Oftentimes, draconian state laws basically prohibit anything resembling coherent thought on that topic. I'm sure<mask> you've been on reddit<mask><mask><mask> I have, you've heard plenty of "No tolerance policy" stories. I've met people in real life that have had that happen to them. It's not the least bit uncommon. Private schools are allowed to make their own rules and can handle bullying in a rational way (It<mask> helps that you would probably personally know the administration official you go to for it, whereas that wouldn't be<mask> common in a large public school). [NEWLINE] </s>
Label encoding: <s>Here's something I can attest to as a guy who spent a good bit of his younger years in a private school. [NEWLINE] A lot of good teachers quit from the public schools due to red tape and excessive amounts of paperwork involved. These teachers, the ones of talked to (who are some of the best I've ever had) argue that they much prefer teaching where they have more freedom to be creative and instill real knowledge and fun activities, not run through a curriculum as quickly as possible to meet state standards. I worked an internship last year with a teacher at a smaller private school and he was able to have his students do debates, projects, plays, and art. He was able to make them love and respect him, and they all learned a lot-Even the at-risk kids. I can tell you it's a lot more effective than handing out a blue diamond book every week and teaching to the test. [NEWLINE] [NEWLINE] If that doesn't get you, here's another point-The public school near my house was massive and I was barely even a number to them. Private schools are much smaller and more personal. Don't you think a lot of at-risk kids would benefit from having their teacher know them as more than just a name on a list of grades with 500-1000 other names on it? They don't have as much opportunity to sit in the back of the class and get away with murder. It also helps foster discussion, as smaller class sizes are easier to work with than massive ones. [NEWLINE] [NEWLINE] Also, in regards to bullying. Oftentimes, draconian state laws basically prohibit anything resembling coherent thought on that topic. I'm sure if you've been on reddit as long as I have, you've heard plenty of "No tolerance policy" stories. I've met people in real life that have had that happen to them. It's not the least bit uncommon. Private schools are allowed to make their own rules and can handle bullying in a rational way (It also helps that you would probably personally know the administration official you go to for it, whereas that wouldn't be as common in a large public school). [NEWLINE] </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My claim would be that the fact that the former is unlikely to be published is a massive problem in our system of scientific publications.  Whether or not it is publishable, it's science and it's vital that it occurs.  The failure to publish it reflects the fact that publications have purposes other than promoting science (ie making money and furthering careers of academics). [NEWLINE] [NEWLINE] [STARTQ] Throwing a ball up in the air and having it fall back down doesn't make you a scientist<mask> you aren't confirming a fact that it still up for debate [ENDQ] [NEWLINE] It is still up for debate and it is science.  Gravity is still a theory and is still subject to being falsified.  The extent to which confirmatory findings help improve our certainty is extraordinarily limited<mask> not quite zero. [NEWLINE] [NEWLINE] [STARTQ] Your definition of scientist, by including people who throw balls in the air, makes "scientist" a completely useless label. It dilutes the word to the point that it should not even be a term anymore. [ENDQ] [NEWLINE] <mask> stated, it's super lame science.  I would go back to my previous example.  My walking to the fridge is technically exercise and my confirming gravity still works on Spalding balls on January 27 2015 is technically science.  I would never call a person who walks to the fridge an athlete<mask> they aren't doing much athletics.  I'm doing less than the average person, I'm not getting paid to do it, and I'm not getting recognized to do it.  It would be absurd to call me an athlete<mask><mask> I am doing something minimally athletic.  You need to do something much more athletic than usual to get called an athlete.  You need to do something much more scientific than the average person to get called a scientist.  We can argue whether the cutoff should be something like top 33%, 10%, top 1%, or whatever. [NEWLINE] [NEWLINE] Bill Nye does much more science than the average person, and goes way beyond testing whether balls will still fall.  The ski instructor does much more athletics than the average person, and goes way beyond walking to the fridge.</s>
Label encoding: <s>My claim would be that the fact that the former is unlikely to be published is a massive problem in our system of scientific publications.  Whether or not it is publishable, it's science and it's vital that it occurs.  The failure to publish it reflects the fact that publications have purposes other than promoting science (ie making money and furthering careers of academics). [NEWLINE] [NEWLINE] [STARTQ] Throwing a ball up in the air and having it fall back down doesn't make you a scientist because you aren't confirming a fact that it still up for debate [ENDQ] [NEWLINE] It is still up for debate and it is science.  Gravity is still a theory and is still subject to being falsified.  The extent to which confirmatory findings help improve our certainty is extraordinarily limited but not quite zero. [NEWLINE] [NEWLINE] [STARTQ] Your definition of scientist, by including people who throw balls in the air, makes "scientist" a completely useless label. It dilutes the word to the point that it should not even be a term anymore. [ENDQ] [NEWLINE] As stated, it's super lame science.  I would go back to my previous example.  My walking to the fridge is technically exercise and my confirming gravity still works on Spalding balls on January 27 2015 is technically science.  I would never call a person who walks to the fridge an athlete because they aren't doing much athletics.  I'm doing less than the average person, I'm not getting paid to do it, and I'm not getting recognized to do it.  It would be absurd to call me an athlete even though I am doing something minimally athletic.  You need to do something much more athletic than usual to get called an athlete.  You need to do something much more scientific than the average person to get called a scientist.  We can argue whether the cutoff should be something like top 33%, 10%, top 1%, or whatever. [NEWLINE] [NEWLINE] Bill Nye does much more science than the average person, and goes way beyond testing whether balls will still fall.  The ski instructor does much more athletics than the average person, and goes way beyond walking to the fridge.</s>
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Imagine<mask> the government had this capability during McCarthyism.<mask> do you think that would have gone?<mask><mask>,<mask> a college student, you had a curiosity in communism or Marxism and you did your research on the internet? Then, 20 years later after you become a senator or something, that search history pops back up<mask> evidence that you're a communist sympathizer. [ENDQ] [NEWLINE] Considering<mask> widely it's done,<mask><mask> a paranoid-driven government would have a similar potential-communist file about everyone. I'm not worried about a totalitarian takeover of the government. Being worried about a McCarthyism level of paranoia and imprisonment seems like a giant leap to be worried.<mask><mask> enough government officials would have skeletons in their closet,<mask> that being "counter-revolutionary" or "now or having ever been a member of the communist party" wouldn't be a threat. [NEWLINE] [NEWLINE] [STARTQ] <mask> could you be made to do<mask> the government knew you look at incest porn?<mask> about the fact that you cheat on your wife with dudes from Craigslist? [ENDQ] [NEWLINE] I don't think the government cares about that. I'd feel uncomfortable<mask> a guy sitting at a computer was watching me on webcam<mask> I looked at incest porn or something,<mask> I don't think the NSA has the manpower to do such a thing, and they simply have a CTRL-F type of thing searching out searches or conversations linked to terrorist activity.<mask> I'm put in the file of "incest porn watchers" with millions of other people, with millions of other files of various skeletons in closets, I wouldn't really mind. Again, I don't think this is the case and<mask><mask> they'd be looking solely for threats to the government or safety of citizens. I'm not worried about a computer database having my information,<mask> I'd be miffed at an actual person spying on me. I don't think the latter is the case,<mask>. [NEWLINE] [NEWLINE] <mask> it comes to public officials, I honestly wouldn't mind<mask> all of their information was an open book considering the rampant corruption; I don't intend on being a public official.</s>
Label encoding: <s> [STARTQ] Imagine if the government had this capability during McCarthyism. How do you think that would have gone? What if, as a college student, you had a curiosity in communism or Marxism and you did your research on the internet? Then, 20 years later after you become a senator or something, that search history pops back up as evidence that you're a communist sympathizer. [ENDQ] [NEWLINE] Considering how widely it's done, I think a paranoid-driven government would have a similar potential-communist file about everyone. I'm not worried about a totalitarian takeover of the government. Being worried about a McCarthyism level of paranoia and imprisonment seems like a giant leap to be worried. I think enough government officials would have skeletons in their closet, so that being "counter-revolutionary" or "now or having ever been a member of the communist party" wouldn't be a threat. [NEWLINE] [NEWLINE] [STARTQ] What could you be made to do if the government knew you look at incest porn? What about the fact that you cheat on your wife with dudes from Craigslist? [ENDQ] [NEWLINE] I don't think the government cares about that. I'd feel uncomfortable if a guy sitting at a computer was watching me on webcam while I looked at incest porn or something, but I don't think the NSA has the manpower to do such a thing, and they simply have a CTRL-F type of thing searching out searches or conversations linked to terrorist activity. If I'm put in the file of "incest porn watchers" with millions of other people, with millions of other files of various skeletons in closets, I wouldn't really mind. Again, I don't think this is the case and I think they'd be looking solely for threats to the government or safety of citizens. I'm not worried about a computer database having my information, but I'd be miffed at an actual person spying on me. I don't think the latter is the case, though. [NEWLINE] [NEWLINE] When it comes to public officials, I honestly wouldn't mind if all of their information was an open book considering the rampant corruption; I don't intend on being a public official.</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Offering them<mask> voluntary courses,<mask>, detracts from genuine "art" by inflating the market, for lack of a better word, with trained individuals. [ENDQ] [NEWLINE] <mask>, are you saying that the fact that anyone can learn<mask> to become a competent draughtsman or oil painter invalidates art<mask> a whole? I'm looking for some clarification<mask> I'm not entirely sure I understand which view you want changed. [NEWLINE] [NEWLINE] Anyway, I'm an artist and I don't see much difference between learning foundations in art and learning lab skills in chemistry classes. Foundations provide you with some basic tools of understanding and experience with a variety of media that you can then apply to your own work. For example, before Picasso began experimenting with cubism he had already [mastered]( [URL].jpg) his medium in the more traditional sense. You can see elements of cubist experimentation throughout his blue, rose, and African periods<mask> he was clearly exploring a new form of expression.<mask> without those traditional foundations he would have been struggling with elementary things like color theory and paint mixing or<mask> to properly load and stroke the brush. He may have eventually gotten<mask> he wanted to go without the traditional foundation he learned at his father's knee,<mask> the road would certainly have been rougher. [NEWLINE] [NEWLINE] There are certainly thousands (<mask> not hundreds of thousands) of artists who never took a formal class or apprenticed in their lives,<mask> I don't think it necessarily follows that education in the arts stifles creativity. There are far too many examples of artists with a traditional education/apprenticeship who invent new movements, styles, and forms of art for this to be the case. Anecdotally, I'll offer up this tidbit:<mask> I took art classes I had my "classwork" and my own projects going at the same time. The same was true for pretty much everyone else I knew. I've been able to apply everything I learned in basic drawing, color theory, etc. to any other form of art I've tried, from digital painting to photography to knitting.</s>
Label encoding: <s> [STARTQ] Offering them as voluntary courses, however, detracts from genuine "art" by inflating the market, for lack of a better word, with trained individuals. [ENDQ] [NEWLINE] So, are you saying that the fact that anyone can learn how to become a competent draughtsman or oil painter invalidates art as a whole? I'm looking for some clarification because I'm not entirely sure I understand which view you want changed. [NEWLINE] [NEWLINE] Anyway, I'm an artist and I don't see much difference between learning foundations in art and learning lab skills in chemistry classes. Foundations provide you with some basic tools of understanding and experience with a variety of media that you can then apply to your own work. For example, before Picasso began experimenting with cubism he had already [mastered]( [URL].jpg) his medium in the more traditional sense. You can see elements of cubist experimentation throughout his blue, rose, and African periods where he was clearly exploring a new form of expression. But without those traditional foundations he would have been struggling with elementary things like color theory and paint mixing or how to properly load and stroke the brush. He may have eventually gotten where he wanted to go without the traditional foundation he learned at his father's knee, but the road would certainly have been rougher. [NEWLINE] [NEWLINE] There are certainly thousands ( if not hundreds of thousands) of artists who never took a formal class or apprenticed in their lives, but I don't think it necessarily follows that education in the arts stifles creativity. There are far too many examples of artists with a traditional education/apprenticeship who invent new movements, styles, and forms of art for this to be the case. Anecdotally, I'll offer up this tidbit: When I took art classes I had my "classwork" and my own projects going at the same time. The same was true for pretty much everyone else I knew. I've been able to apply everything I learned in basic drawing, color theory, etc. to any other form of art I've tried, from digital painting to photography to knitting.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] got you laid [ENDQ] [NEWLINE] I'm looking to marry and settle down; sex is the least of my concerns. The problem is that I'm a young person in a first-world country and the global bankers think that I'm disposable. Plus, I live in a society that demands hard work and I'll be demonized<mask> I decide that greed isn't in my best interests. I'd gladly remain a virgin<mask> the love of my life was asexual. [NEWLINE] [NEWLINE] [STARTQ] just like we normally treat both the symptoms of a disease and the disease itself [ENDQ] [NEWLINE] The problem is that with this disease of disadvantaging the vast majority of people must be cured or else one group of extremists will replace another. [NEWLINE] [NEWLINE] [STARTQ] the issues you lump together seem very heterogeneous [ENDQ] [NEWLINE] At the heart of them, the issues all boil down to the fact that the global system is not only unfair<mask> perverse; countries that invest in their citizens [like Canada or much of Europe]( [URL] /) not only are unable to help the US see the light thanks to the brainwashed media and pseudo-political system,<mask> are encouraged to fix<mask> ain't broken thanks to the great hoax of the "European" (actually has its origin in the 2000s housing bust in Florida and elsewhere) "debt" (explain<mask> countries like Canada, Australia, Sweden and Denmark not only aren't utopias<mask> destroying<mask> makes them<mask> much better off than the US<mask><mask><mask> smaller, more homogeneous populations, AAA credit ratings, and excellent economic fundamentals. It's about lowering the standards of the average worker<mask> that people feel pressure to indenture themselves) "crisis" (Spain had a debt-to-GDP ratio that was lower than the US before the downturn and only suffered<mask>, like all civilized countries, it extended aid to its unemployed). Western civilization has largely been a history of increased material comfort, and<mask> things have gone backwards it has been due to war or plague, not due to politicians (openly in the Nordic countries, for instance) deciding that their countries are too developed and too educated.</s>
Label encoding: <s> [STARTQ] got you laid [ENDQ] [NEWLINE] I'm looking to marry and settle down; sex is the least of my concerns. The problem is that I'm a young person in a first-world country and the global bankers think that I'm disposable. Plus, I live in a society that demands hard work and I'll be demonized if I decide that greed isn't in my best interests. I'd gladly remain a virgin if the love of my life was asexual. [NEWLINE] [NEWLINE] [STARTQ] just like we normally treat both the symptoms of a disease and the disease itself [ENDQ] [NEWLINE] The problem is that with this disease of disadvantaging the vast majority of people must be cured or else one group of extremists will replace another. [NEWLINE] [NEWLINE] [STARTQ] the issues you lump together seem very heterogeneous [ENDQ] [NEWLINE] At the heart of them, the issues all boil down to the fact that the global system is not only unfair but perverse; countries that invest in their citizens [like Canada or much of Europe]( [URL] /) not only are unable to help the US see the light thanks to the brainwashed media and pseudo-political system, but are encouraged to fix what ain't broken thanks to the great hoax of the "European" (actually has its origin in the 2000s housing bust in Florida and elsewhere) "debt" (explain why countries like Canada, Australia, Sweden and Denmark not only aren't utopias but destroying what makes them so much better off than the US in spite of smaller, more homogeneous populations, AAA credit ratings, and excellent economic fundamentals. It's about lowering the standards of the average worker so that people feel pressure to indenture themselves) "crisis" (Spain had a debt-to-GDP ratio that was lower than the US before the downturn and only suffered because, like all civilized countries, it extended aid to its unemployed). Western civilization has largely been a history of increased material comfort, and when things have gone backwards it has been due to war or plague, not due to politicians (openly in the Nordic countries, for instance) deciding that their countries are too developed and too educated.</s>
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When Nixon left, he left peacefully and with no promise of a pardon. [NEWLINE] [NEWLINE] You underestimate<mask> ingrained peaceful transition of power is in functioning democracies. 50% of the living governors of Illinois served jail time for corruption related indictments they committed<mask> in office.<mask>, has there *ever* been even a *hint* of a threat that<mask> the Republican won last November that the incumbent Democrat might not hand over power? Of course not! [NEWLINE] [NEWLINE] The 2000 presidential race was the most contentious in the modern era. Do you know the maximum extent of friction that happened<mask> Clinton handed control to the other party? Some of the staff went around and removed the "W" keys from various office keyboards. That is the sum total of disorderliness that came out of that contentious election. [NEWLINE] [NEWLINE] Yes, a peaceful transition is the most important thing in a democracy,<mask> granting full and unconditional immunity is not a part of that. Nixon left peacefully, and he did not have a pardon<mask> he left.<mask><mask> the investigation continued for another month before President Ford decided to end it. [NEWLINE] [NEWLINE] There is no threat from inditing former heads of state, there is a threat from granting a pardon<mask>. [NEWLINE] [NEWLINE] After Nixon, the power of the Executive has grown dramatically. And it has grown the way it has<mask> there is now an assumption that there is no consequences. Reagan,<mask> running against Ford, actively used backchannels with Iran to tell them not to release the hostages until after Carter left office. He then later, in defiance of direct orders from Congress, sold weapons to Iran and used the money to fund fascist rebel groups who were into cocaine and raping nuns. Bush 43 instituted a purge of the Justice Department, firing nearly 40% of prosecutors based on their ideological beliefs. Not to mention the growing unilateral nature of the various presidents' use of force, across party lines. [NEWLINE] [NEWLINE] This is<mask> happens<mask> no one of high importance is held legally responsible for their actions. That is a far bigger threat to our democracies than any worry that they might not respect the results of an election.</s>
Label encoding: <s>When Nixon left, he left peacefully and with no promise of a pardon. [NEWLINE] [NEWLINE] You underestimate how ingrained peaceful transition of power is in functioning democracies. 50% of the living governors of Illinois served jail time for corruption related indictments they committed while in office. Yet, has there *ever* been even a *hint* of a threat that when the Republican won last November that the incumbent Democrat might not hand over power? Of course not! [NEWLINE] [NEWLINE] The 2000 presidential race was the most contentious in the modern era. Do you know the maximum extent of friction that happened when Clinton handed control to the other party? Some of the staff went around and removed the "W" keys from various office keyboards. That is the sum total of disorderliness that came out of that contentious election. [NEWLINE] [NEWLINE] Yes, a peaceful transition is the most important thing in a democracy, but granting full and unconditional immunity is not a part of that. Nixon left peacefully, and he did not have a pardon when he left. In fact the investigation continued for another month before President Ford decided to end it. [NEWLINE] [NEWLINE] There is no threat from inditing former heads of state, there is a threat from granting a pardon however. [NEWLINE] [NEWLINE] After Nixon, the power of the Executive has grown dramatically. And it has grown the way it has because there is now an assumption that there is no consequences. Reagan, while running against Ford, actively used backchannels with Iran to tell them not to release the hostages until after Carter left office. He then later, in defiance of direct orders from Congress, sold weapons to Iran and used the money to fund fascist rebel groups who were into cocaine and raping nuns. Bush 43 instituted a purge of the Justice Department, firing nearly 40% of prosecutors based on their ideological beliefs. Not to mention the growing unilateral nature of the various presidents' use of force, across party lines. [NEWLINE] [NEWLINE] This is what happens when no one of high importance is held legally responsible for their actions. That is a far bigger threat to our democracies than any worry that they might not respect the results of an election.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I find baseball incredibly boring.<mask> in "went to a game, fell asleep halfway through" boring. To me there is literally no more to it than a guy throwing a ball and another guy hitting it with a bat. And they do that for hours! [NEWLINE] [NEWLINE] <mask> I assume I must be missing something,<mask> it seems to be popular. Maybe I just don't understand the significance<mask> some of the small things that are going on - things that seem completely irrelevant to me. [NEWLINE] [NEWLINE] I would imagine it's the same thing for you with basketball. You just see a bunch<mask> guys running back and forth throwing a ball into a basket. Big whoop, right? [NEWLINE] [NEWLINE] <mask><mask> I see basketball, I see plays, I see tactics, I see technically and physically impressive feats. I see<mask> a perfectly executed screen can turn into a roll, which can turn into an open shot on the basket. I see a fast paced sport with constant action, something I don't see in football and baseball, for example,<mask> they seem to break all the time. [NEWLINE] [NEWLINE] You complaint about basketball being a high scoring game, resulting in many games being decided in only the last amount of time. I would<mask><mask> any close game will be decided in the last amount of time.<mask>, this mostly happens<mask> the teams are equally skilled. A much better team can pretty much win the game in the first half. [NEWLINE] <mask> you could<mask><mask> the ability<mask> teams to be able to closely follow each other's score is boring, it<mask> allows teams that are falling behind to have a real shot at getting back,<mask> they change their strategy. [NEWLINE] [NEWLINE] It seems like part of your logic is somewhat circular - I don't like or understand basketball,<mask><mask> it's a bad sport, and<mask> I don't like it. [NEWLINE] [NEWLINE] <mask> it may be true that not *all* fans understand the intricacies, the same can be said for baseball.<mask> for those of us that do, it's a fast paced, interesting, highly tactical sport with many impressive physical feats. </s>
Label encoding: <s>I find baseball incredibly boring. As in "went to a game, fell asleep halfway through" boring. To me there is literally no more to it than a guy throwing a ball and another guy hitting it with a bat. And they do that for hours! [NEWLINE] [NEWLINE] However I assume I must be missing something, since it seems to be popular. Maybe I just don't understand the significance if some of the small things that are going on - things that seem completely irrelevant to me. [NEWLINE] [NEWLINE] I would imagine it's the same thing for you with basketball. You just see a bunch if guys running back and forth throwing a ball into a basket. Big whoop, right? [NEWLINE] [NEWLINE] But when I see basketball, I see plays, I see tactics, I see technically and physically impressive feats. I see how a perfectly executed screen can turn into a roll, which can turn into an open shot on the basket. I see a fast paced sport with constant action, something I don't see in football and baseball, for example, where they seem to break all the time. [NEWLINE] [NEWLINE] You complaint about basketball being a high scoring game, resulting in many games being decided in only the last amount of time. I would argue that any close game will be decided in the last amount of time. Also, this mostly happens when the teams are equally skilled. A much better team can pretty much win the game in the first half. [NEWLINE] While you could argue that the ability if teams to be able to closely follow each other's score is boring, it also allows teams that are falling behind to have a real shot at getting back, if they change their strategy. [NEWLINE] [NEWLINE] It seems like part of your logic is somewhat circular - I don't like or understand basketball, so therefore it's a bad sport, and therefore I don't like it. [NEWLINE] [NEWLINE] While it may be true that not *all* fans understand the intricacies, the same can be said for baseball. But for those of us that do, it's a fast paced, interesting, highly tactical sport with many impressive physical feats. </s>
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Feminists tend to ignore the biological fact (yeah "biofacts") that mothering starts in the womb. [ENDQ] [NEWLINE] "Mothering" in this sense is a role you perform, and I<mask><mask> it takes place only after childbirth. Yes women have the children. That's a biological process and really has little to do with actually raising a child.<mask> a woman had a baby and threw it in the trash, could you really say she'd mothered anything? In one technical sense of the word, sure,<mask><mask><mask> the idea of mothering is tied more to nurturing than the physical action of birthing. Nurturing is providing for the physical, emotional, and educational needs of a child. It's a set of behaviors that anyone can do. Yes,<mask> we were hunters and gatherers, providing for a child's physical needs meant the woman had to be there<mask> your "nipples do not lactate,"<mask> it's a poor excuse in the age of formula and breast pumps. The real question is whether men can adequately fill the role of nurterer, and my answer is that<mask><mask> there's no reason<mask> they shouldn't be able to. [NEWLINE] [NEWLINE] [STARTQ] <mask> today's world may be different we are still animals which have been shaped by evolution both biological and cultural, and to ignore this for political ends seems rather short-sighted. [ENDQ] [NEWLINE] This is an appeal to tradition, and it's not good enough to explain away the wage gap. I can acknowledge that there were once very good reasons that women were the nurturers, logistically speaking. Those reasons don't really exist in today's society, and I see no reason those social structures should continue to pigeonhole people into roles. [NEWLINE] [NEWLINE] [STARTQ] Gender roles may be a social construct,<mask> long-living social constructs are ultimately stable phenotypes. [ENDQ] [NEWLINE] Again, just<mask> that may be a stable phenotype (which<mask><mask> could be debatable with the right information) doesn't mean it's the best phenotype. It's an appeal to tradition, not a good argument. </s>
Label encoding: <s> [STARTQ] Feminists tend to ignore the biological fact (yeah "biofacts") that mothering starts in the womb. [ENDQ] [NEWLINE] "Mothering" in this sense is a role you perform, and I argue that it takes place only after childbirth. Yes women have the children. That's a biological process and really has little to do with actually raising a child. If a woman had a baby and threw it in the trash, could you really say she'd mothered anything? In one technical sense of the word, sure, but I think the idea of mothering is tied more to nurturing than the physical action of birthing. Nurturing is providing for the physical, emotional, and educational needs of a child. It's a set of behaviors that anyone can do. Yes, when we were hunters and gatherers, providing for a child's physical needs meant the woman had to be there because your "nipples do not lactate," but it's a poor excuse in the age of formula and breast pumps. The real question is whether men can adequately fill the role of nurterer, and my answer is that I think there's no reason why they shouldn't be able to. [NEWLINE] [NEWLINE] [STARTQ] While today's world may be different we are still animals which have been shaped by evolution both biological and cultural, and to ignore this for political ends seems rather short-sighted. [ENDQ] [NEWLINE] This is an appeal to tradition, and it's not good enough to explain away the wage gap. I can acknowledge that there were once very good reasons that women were the nurturers, logistically speaking. Those reasons don't really exist in today's society, and I see no reason those social structures should continue to pigeonhole people into roles. [NEWLINE] [NEWLINE] [STARTQ] Gender roles may be a social construct, but long-living social constructs are ultimately stable phenotypes. [ENDQ] [NEWLINE] Again, just because that may be a stable phenotype (which I think could be debatable with the right information) doesn't mean it's the best phenotype. It's an appeal to tradition, not a good argument. </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right?! Applecare is the most amazing customer service I've ever experienced from any company. Ever. [NEWLINE] [NEWLINE] I cannot even say<mask> much applecare has helped me over the last few years. [NEWLINE] [NEWLINE] I had a sony vaoi before my macbook. I bought it<mask> my father bullied me out of getting an apple and I regretted it the moment the thing crapped out four months after getting it. Sent it back to them and they held it for FOUR MONTHS examining it. I had to borrow my mother's laptop<mask> I was a university student and had papers to write. I went weeks without communication and only after the seventh or eighth call asking for an update did the determine it was a dud and mailed me a new one - without the upgrades I'd paid for on the first one. The replacement's hard drive shit the bed my last week of university. I'd JUST finished my last paper (early too thankfully!)<mask> I could party with my friends. Instead I spent the last two days regurgitating the thirty way lit review in the library computer lab crying and looking vaguely homeless<mask> I hadn't showered or changed in 36 hours. I couldn't even bring myself to buy a computer for another six months<mask> I got<mask> angry even thinking about owning another POS like the Sony. [NEWLINE] [NEWLINE] My sister and parents bought me the macbook and it's been glorious. Had trouble with the the screens colours once - just brought it to the genius bar between grad school seminars. The chord ripped? No problem - they replaced it free of charge. Had to replace the battery which died this year and there was no apple stores anywhere in the vicinity - no problem. They had a list of local vendors who honour their warranties. Just dropped it off with them and apple took care of the cost. Kid I was babysitting ripped the headphones out and got the nub stuck in the jack - POST WARRANTY - and they still spent 20 minutes passing it between staffers at the store until one could muscle it out. [NEWLINE] [NEWLINE] Applecare is just bomb dot com.</s>
Label encoding: <s>Right?! Applecare is the most amazing customer service I've ever experienced from any company. Ever. [NEWLINE] [NEWLINE] I cannot even say how much applecare has helped me over the last few years. [NEWLINE] [NEWLINE] I had a sony vaoi before my macbook. I bought it because my father bullied me out of getting an apple and I regretted it the moment the thing crapped out four months after getting it. Sent it back to them and they held it for FOUR MONTHS examining it. I had to borrow my mother's laptop because I was a university student and had papers to write. I went weeks without communication and only after the seventh or eighth call asking for an update did the determine it was a dud and mailed me a new one - without the upgrades I'd paid for on the first one. The replacement's hard drive shit the bed my last week of university. I'd JUST finished my last paper (early too thankfully!) so I could party with my friends. Instead I spent the last two days regurgitating the thirty way lit review in the library computer lab crying and looking vaguely homeless because I hadn't showered or changed in 36 hours. I couldn't even bring myself to buy a computer for another six months because I got so angry even thinking about owning another POS like the Sony. [NEWLINE] [NEWLINE] My sister and parents bought me the macbook and it's been glorious. Had trouble with the the screens colours once - just brought it to the genius bar between grad school seminars. The chord ripped? No problem - they replaced it free of charge. Had to replace the battery which died this year and there was no apple stores anywhere in the vicinity - no problem. They had a list of local vendors who honour their warranties. Just dropped it off with them and apple took care of the cost. Kid I was babysitting ripped the headphones out and got the nub stuck in the jack - POST WARRANTY - and they still spent 20 minutes passing it between staffers at the store until one could muscle it out. [NEWLINE] [NEWLINE] Applecare is just bomb dot com.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You haven't actually made an argument. You've just made some assertions, with no supporting evidence. [NEWLINE] [NEWLINE] [STARTQ] One of the main goals of an HOA are to help maintain or even increase the value of the homes in the neighborhood by preventing people from doing things that would lower property values. [ENDQ] [NEWLINE] Just<mask> HOAs profess to have that<mask> a goal, doesn't mean they *actually* have that<mask> a goal. It<mask> doesn't *necessarily* mean they're actually doing anything to effectively pursue that goal. [NEWLINE] [NEWLINE] <mask>, the shitty knee-jerk argument:<mask><mask> I *want* to do something that would lower property values.<mask><mask> I bought my own home just<mask> I could have the freedom to do this (I don't know, say, have band practices? Paint my house a crazy colour? I'm not a homeowner<mask> I'm spitballing here). HOAs are infringing on my freedom to do this. Obviously HOAs are opt-in<mask> the correct response to me in that situation would be "<mask> don't live<mask> there's an HOA".<mask>, of course, your original statement is "***everyone*** should live<mask> there is one". [NEWLINE] [NEWLINE] [STARTQ] Home owners associations don't have to be expensive for the residents that live under them. They can even be free. [ENDQ] [NEWLINE] Again: Just<mask> they *can* doesn't mean they *are*. Speaking of anecdotes: I've looked into condos that have HOAs around<mask> I live (SF Bay Area, California). Most of them have HOA fees that are somewhere between 1/6th to 1/4th of<mask> the rent is.<mask><mask><mask> I'm concerned, $500/mo HOA fees is insanity. [NEWLINE] [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] It's hard to make a counter argument<mask> I don't actually know<mask> you believe that they are good,<mask> here goes:<mask> the fuck should I *pay for the privilege* of *someone else* telling me<mask> I can or can't do at my house. [NEWLINE] [NEWLINE] That is all</s>
Label encoding: <s>You haven't actually made an argument. You've just made some assertions, with no supporting evidence. [NEWLINE] [NEWLINE] [STARTQ] One of the main goals of an HOA are to help maintain or even increase the value of the homes in the neighborhood by preventing people from doing things that would lower property values. [ENDQ] [NEWLINE] Just because HOAs profess to have that as a goal, doesn't mean they *actually* have that as a goal. It also doesn't *necessarily* mean they're actually doing anything to effectively pursue that goal. [NEWLINE] [NEWLINE] Additionally, the shitty knee-jerk argument: what if I *want* to do something that would lower property values. What if I bought my own home just so I could have the freedom to do this (I don't know, say, have band practices? Paint my house a crazy colour? I'm not a homeowner so I'm spitballing here). HOAs are infringing on my freedom to do this. Obviously HOAs are opt-in so the correct response to me in that situation would be " so don't live where there's an HOA". But, of course, your original statement is "***everyone*** should live where there is one". [NEWLINE] [NEWLINE] [STARTQ] Home owners associations don't have to be expensive for the residents that live under them. They can even be free. [ENDQ] [NEWLINE] Again: Just because they *can* doesn't mean they *are*. Speaking of anecdotes: I've looked into condos that have HOAs around where I live (SF Bay Area, California). Most of them have HOA fees that are somewhere between 1/6th to 1/4th of what the rent is. As far as I'm concerned, $500/mo HOA fees is insanity. [NEWLINE] [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] It's hard to make a counter argument when I don't actually know why you believe that they are good, but here goes: Why the fuck should I *pay for the privilege* of *someone else* telling me what I can or can't do at my house. [NEWLINE] [NEWLINE] That is all</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>On (2), the laws of supply-and-demand aren't applied; they just exist.  Supply-and-demand is true in labor market, whether people want to acknowledge it or not. [NEWLINE] [NEWLINE] In answer to your question,<mask>,<mask> a person can't make enough money to pay for his own necessities, then that's an appropriate place for government and private charity to help out. [NEWLINE] [NEWLINE] (3) The place with 10 minimum wage workers might be able to do more with 9 by investing in technology.  Or, perhaps, replacing some of those unskilled workers with skilled workers.  In Australia, the minmum wage is substantially higher than it is in the US, and places like McDonalds correspondingly employ fewer people. <mask> do they do this?  For example, they use more touch-screens to order instead of dictating orders to minimum-wage workers.  They run fewer "open all night" drive-ins (which might be profitable at a lower minimum wage) and<mask> on. [NEWLINE] [NEWLINE] (4) To the extent a minimum wage merely offsets welfare that a person would otherwise get, this is clearly wrong. <mask>,<mask>, recognize that there are relatively few people affected by a minimum-wage increase,<mask> the increase in the number of people able to buy a product is relatively small.  And, there are occupations<mask> this isn't true at all --<mask> a high-end hotel has to increase its wage to its workers, it's very unlikely that any of those workers will end up staying at the hotel. [NEWLINE] [NEWLINE] On (5),<mask> should it be an employer's responsibility to make sure that their workers aren't on welfare rolls? [NEWLINE] [NEWLINE] Consider this: Let's say there's an unemployed on welfare and a local store is thinking about employing them.  Isn't that person (and society) better off<mask> they get a job, even<mask> it doesn't fully take them off welfare?  After all, that job might not exist<mask> it had to pay enough to take people off welfare.</s>
Label encoding: <s>On (2), the laws of supply-and-demand aren't applied; they just exist.  Supply-and-demand is true in labor market, whether people want to acknowledge it or not. [NEWLINE] [NEWLINE] In answer to your question, though, when a person can't make enough money to pay for his own necessities, then that's an appropriate place for government and private charity to help out. [NEWLINE] [NEWLINE] (3) The place with 10 minimum wage workers might be able to do more with 9 by investing in technology.  Or, perhaps, replacing some of those unskilled workers with skilled workers.  In Australia, the minmum wage is substantially higher than it is in the US, and places like McDonalds correspondingly employ fewer people.  How do they do this?  For example, they use more touch-screens to order instead of dictating orders to minimum-wage workers.  They run fewer "open all night" drive-ins (which might be profitable at a lower minimum wage) and so on. [NEWLINE] [NEWLINE] (4) To the extent a minimum wage merely offsets welfare that a person would otherwise get, this is clearly wrong.  But, also, recognize that there are relatively few people affected by a minimum-wage increase, so the increase in the number of people able to buy a product is relatively small.  And, there are occupations where this isn't true at all -- if a high-end hotel has to increase its wage to its workers, it's very unlikely that any of those workers will end up staying at the hotel. [NEWLINE] [NEWLINE] On (5), WHY should it be an employer's responsibility to make sure that their workers aren't on welfare rolls? [NEWLINE] [NEWLINE] Consider this: Let's say there's an unemployed on welfare and a local store is thinking about employing them.  Isn't that person (and society) better off when they get a job, even when it doesn't fully take them off welfare?  After all, that job might not exist if it had to pay enough to take people off welfare.</s>
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I feel very strongly that the main motivation for these killers isn't the pleasure they get from murdering kids,<mask> the shock and disgust they subject the public to. [ENDQ] [NEWLINE] There are many reasons that these things happen, not one. Many of them genuinely think they are going to get vengeance for wrongs they feel have been committed against them. Others are motivated by the hope for a sick thrill. Some are actually convinced that they can get away with it. [NEWLINE] [NEWLINE] Painting every one with a single brush obfuscates the truth and keeps us from better understanding the matter. That, in turn, keeps us from creating better safeguards against these things<mask> they happen and from creating better systems of preventing them from happening at all. [NEWLINE] [NEWLINE] You are overreacting to the ridiculous coverage. The answer is more moderate, reasonable coverage, not banning any coverage at all. There is a difference between CNN's "let's look at the condiments in the Death-Fridge" 24-hour glamorization and a calm, measured recounting of the facts. [NEWLINE] [NEWLINE] [STARTQ] It might sound drastic,<mask><mask> a law was passed that treated these people like terrorists and forbid the media to give into their demands, I feel that the amount of shootings would go down. CMV, or at least tell my<mask> making a law about it would be a bad idea. [ENDQ] [NEWLINE] Assuming you're an American, your proposed law would be stricken down immediately by the Supreme Court. It would be a blatant violation of the Freedom of the Press outlined in the First Amendment, and the body of case law concerning it is fairly clear. [NEWLINE] [NEWLINE] <mask>'s more, your law would set a *terrifying* precedent, allowing the government to arbitrarily decide<mask> news programs are allowed to cover.<mask> the government decides that the news can't cover protesters, is that ok?<mask> about political corruption?<mask> about an entire war? This kind of erosion of the Freedom of the Press can quickly snowball into a [Ministry of Truth]( [URL] ) situation, which is<mask> our Constitution explicitly forbids it.</s>
Label encoding: <s> [STARTQ] I feel very strongly that the main motivation for these killers isn't the pleasure they get from murdering kids, but the shock and disgust they subject the public to. [ENDQ] [NEWLINE] There are many reasons that these things happen, not one. Many of them genuinely think they are going to get vengeance for wrongs they feel have been committed against them. Others are motivated by the hope for a sick thrill. Some are actually convinced that they can get away with it. [NEWLINE] [NEWLINE] Painting every one with a single brush obfuscates the truth and keeps us from better understanding the matter. That, in turn, keeps us from creating better safeguards against these things when they happen and from creating better systems of preventing them from happening at all. [NEWLINE] [NEWLINE] You are overreacting to the ridiculous coverage. The answer is more moderate, reasonable coverage, not banning any coverage at all. There is a difference between CNN's "let's look at the condiments in the Death-Fridge" 24-hour glamorization and a calm, measured recounting of the facts. [NEWLINE] [NEWLINE] [STARTQ] It might sound drastic, but if a law was passed that treated these people like terrorists and forbid the media to give into their demands, I feel that the amount of shootings would go down. CMV, or at least tell my why making a law about it would be a bad idea. [ENDQ] [NEWLINE] Assuming you're an American, your proposed law would be stricken down immediately by the Supreme Court. It would be a blatant violation of the Freedom of the Press outlined in the First Amendment, and the body of case law concerning it is fairly clear. [NEWLINE] [NEWLINE] What's more, your law would set a *terrifying* precedent, allowing the government to arbitrarily decide what news programs are allowed to cover. If the government decides that the news can't cover protesters, is that ok? How about political corruption? How about an entire war? This kind of erosion of the Freedom of the Press can quickly snowball into a [Ministry of Truth]( [URL] ) situation, which is why our Constitution explicitly forbids it.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I disagree with this fundamentally, I feel like there are a million responses to<mask> you said<mask> I'll just throw one out. [NEWLINE] [NEWLINE] Language communicates ideas in many ways, and some words have a greater capacity to create certain imagery in the mind,<mask> there should be no limitations on the words we use.  Orwell's 1984 showed the government limit the language down to the'simplest and most efficient' words to communicate.  Just gonna leave this quote from one of the characters responsible for simplifying the language: [NEWLINE] [NEWLINE] [STARTQ] "It's a beautiful thing, the Destruction of words. Of course the great wastage is in the verbs and adjectives,<mask> there are hundreds of nouns that can be got rid of<mask> well. It isn't only the synonyms; there are<mask> the antonyms. After all,<mask> justification is there for a word, which is simply the opposite of some other word? A word contains its opposite in itself. Take ‘good,’ for instance.<mask> you have a word like ‘good,’<mask> need is there for a word like ‘bad’? ‘Ungood’ will do just<mask> well – better,<mask> it's an exact opposite, which the other is not. Or again,<mask> you want a stronger version of ‘good,’<mask> sense is there in having a whole string of vague useless words like ‘excellent’ and ‘splendid’ and all the rest of them? ‘Plusgood’ covers the meaning or ‘doubleplusgood’<mask> you want something stronger still. [ENDQ] [NEWLINE] Of course, the book shows that paring down the language clearly is a tool to limit thought.  The thinking being there can't be rebellion<mask> there is no word for it. <mask> relevant is this to your point?  Maybe not fully<mask>,<mask> language is already imprecise enough at describing and understanding the world, I don't see<mask> we should make it even less useful.</s><pad>
Label encoding: <s>I disagree with this fundamentally, I feel like there are a million responses to what you said but I'll just throw one out. [NEWLINE] [NEWLINE] Language communicates ideas in many ways, and some words have a greater capacity to create certain imagery in the mind, so there should be no limitations on the words we use.  Orwell's 1984 showed the government limit the language down to the'simplest and most efficient' words to communicate.  Just gonna leave this quote from one of the characters responsible for simplifying the language: [NEWLINE] [NEWLINE] [STARTQ] "It's a beautiful thing, the Destruction of words. Of course the great wastage is in the verbs and adjectives, but there are hundreds of nouns that can be got rid of as well. It isn't only the synonyms; there are also the antonyms. After all, what justification is there for a word, which is simply the opposite of some other word? A word contains its opposite in itself. Take ‘good,’ for instance. If you have a word like ‘good,’ what need is there for a word like ‘bad’? ‘Ungood’ will do just as well – better, because it's an exact opposite, which the other is not. Or again, if you want a stronger version of ‘good,’ what sense is there in having a whole string of vague useless words like ‘excellent’ and ‘splendid’ and all the rest of them? ‘Plusgood’ covers the meaning or ‘doubleplusgood’ if you want something stronger still. [ENDQ] [NEWLINE] Of course, the book shows that paring down the language clearly is a tool to limit thought.  The thinking being there can't be rebellion if there is no word for it.  How relevant is this to your point?  Maybe not fully so, but language is already imprecise enough at describing and understanding the world, I don't see why we should make it even less useful.</s><pad>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> the other guy mentioned, those specific situations aren't excluded (the logic is supposed to be that child support is for the child's benefit and not the mother,<mask> theory doesn't always translate well into practice). [NEWLINE] [NEWLINE] <mask> even<mask> we did just allow those two situations, I would still<mask><mask> the general idea (always being able to financially abort) should be an option. This really comes down to the fact that I believe nobody should have a child that they cannot afford. It's rational to not have a child (or want to abort<mask> accidental pregnancy occurs)<mask> money is tight. A law that forces otherwise goes against the original intent of the people having sex (I'm presuming that people who have sex do not want a child unless they discuss and agree before hand). [NEWLINE] [NEWLINE] Further, requiring child support payments could be viewed<mask> damaging to the payee. It's a major expense that can financially cripple those who don't have a lot of extra cash (which is quite a lot of people). Especially<mask> you get thrown in prison<mask> you can't make the payment *<mask> still have to make future payments* (<mask> prison obviously ruining your finances even more). [NEWLINE] [NEWLINE] <mask>, I should note that I'm referring solely to cases<mask> the an unwanted pregnancy occurs and not cases<mask> a family with existing children separates (which is very different). [NEWLINE] [NEWLINE] There's<mask> the case of<mask> the father is not informed until after the birth or<mask> the father cannot be found. It's<mask><mask> that in those cases, the father would be absolved from paying child support (unless they want to,<mask> that's something for them to do personally and not a court matter). This is<mask><mask> finances are an issue, the woman always has the power to abort or put the child up for adoption. [NEWLINE] [NEWLINE] As you can tell, I have very strong views against having children that you're knowingly not financially able to support. This is a utilitarian view that won't be popular with many,<mask> I believe to be very rational and maximizes economical fairness.</s>
Label encoding: <s>As the other guy mentioned, those specific situations aren't excluded (the logic is supposed to be that child support is for the child's benefit and not the mother, but theory doesn't always translate well into practice). [NEWLINE] [NEWLINE] But even if we did just allow those two situations, I would still argue that the general idea (always being able to financially abort) should be an option. This really comes down to the fact that I believe nobody should have a child that they cannot afford. It's rational to not have a child (or want to abort if accidental pregnancy occurs) when money is tight. A law that forces otherwise goes against the original intent of the people having sex (I'm presuming that people who have sex do not want a child unless they discuss and agree before hand). [NEWLINE] [NEWLINE] Further, requiring child support payments could be viewed as damaging to the payee. It's a major expense that can financially cripple those who don't have a lot of extra cash (which is quite a lot of people). Especially if you get thrown in prison if you can't make the payment * yet still have to make future payments* ( despite prison obviously ruining your finances even more). [NEWLINE] [NEWLINE] However, I should note that I'm referring solely to cases where the an unwanted pregnancy occurs and not cases where a family with existing children separates (which is very different). [NEWLINE] [NEWLINE] There's also the case of when the father is not informed until after the birth or if the father cannot be found. It's my opinion that in those cases, the father would be absolved from paying child support (unless they want to, but that's something for them to do personally and not a court matter). This is because if finances are an issue, the woman always has the power to abort or put the child up for adoption. [NEWLINE] [NEWLINE] As you can tell, I have very strong views against having children that you're knowingly not financially able to support. This is a utilitarian view that won't be popular with many, but I believe to be very rational and maximizes economical fairness.</s>
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd like to support OP's arguments by way of strategic reasoning. [NEWLINE] [NEWLINE] To prevent the use of hard drugs, the best course is to bring people face to face with the reality of<mask> such can cause.  Sheltering them from the harsh realities by focusing on an enforcement approach to hide the problem away in prisons and poor neighborhoods only allows ignorant people to romanticize the effects<mask> that gullible and foolish people can be too afraid to get help<mask> they need it.  The system *encourages* abuse. [NEWLINE] [NEWLINE] Via legalization, organized crime can be run out of business and a market monitored, with controlled supply.  Without any dealers around, it's possible to manipulate prices and supply to force the weening of entire swats of geography, from addictive substances.  The evidence for this is in the fact that<mask> taxes cause price increases, people quit smoking at an increasing rate and many people smoke less. [NEWLINE] [NEWLINE] Finally, there's a point of proof here. <mask> it's not possible to eliminate the black market then the current approach will be an indefinite, increasing liability.  Sooner or later, this model will fail<mask> it will simply become too expensive. <mask>,<mask> the black market can be eliminated entirely, then by legalization it can be driven out of business.  That, we're good at; forcing aversion to many self-destructive behaviors among a sea of individuals diverse in every conceivable way, not<mask> much. [NEWLINE] [NEWLINE] I typically do not argue for the legalization of hard drugs<mask> we are not<mask> at that inevitable moment<mask> this system will simply fail. <mask>, that moment is mathematically guaranteed to come<mask> the population increases with users of hard drugs maintaining a proportionate presence.  This, in a system<mask> abuse rates rarely drop and never fail to increase again.  Our current strategy will financially collapse, guaranteed,<mask> plain mathematical fact.  We need to have a plan in place ahead of time to not only implement the best replacement system<mask><mask> make the transition<mask> uneventfully<mask> possible.  </s>
Label encoding: <s>I'd like to support OP's arguments by way of strategic reasoning. [NEWLINE] [NEWLINE] To prevent the use of hard drugs, the best course is to bring people face to face with the reality of what such can cause.  Sheltering them from the harsh realities by focusing on an enforcement approach to hide the problem away in prisons and poor neighborhoods only allows ignorant people to romanticize the effects so that gullible and foolish people can be too afraid to get help when they need it.  The system *encourages* abuse. [NEWLINE] [NEWLINE] Via legalization, organized crime can be run out of business and a market monitored, with controlled supply.  Without any dealers around, it's possible to manipulate prices and supply to force the weening of entire swats of geography, from addictive substances.  The evidence for this is in the fact that as taxes cause price increases, people quit smoking at an increasing rate and many people smoke less. [NEWLINE] [NEWLINE] Finally, there's a point of proof here.  If it's not possible to eliminate the black market then the current approach will be an indefinite, increasing liability.  Sooner or later, this model will fail because it will simply become too expensive.  However, if the black market can be eliminated entirely, then by legalization it can be driven out of business.  That, we're good at; forcing aversion to many self-destructive behaviors among a sea of individuals diverse in every conceivable way, not so much. [NEWLINE] [NEWLINE] I typically do not argue for the legalization of hard drugs because we are not yet at that inevitable moment when this system will simply fail.  However, that moment is mathematically guaranteed to come as the population increases with users of hard drugs maintaining a proportionate presence.  This, in a system where abuse rates rarely drop and never fail to increase again.  Our current strategy will financially collapse, guaranteed, as plain mathematical fact.  We need to have a plan in place ahead of time to not only implement the best replacement system but also make the transition as uneventfully as possible.  </s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are more components to hyphenating a word than just for column width. Hyphens can alleviate problems with [Widows and Orphans]( [URL] ) on a page, or<mask> has already been mentioned with real-estate on newspapers, magazines, publications, or journal-entries. [NEWLINE] [NEWLINE] Widows and orphans in typography can cause issues with physical media<mask> a single sentence is printed at the top of a clean page. This can be a waste of paper, which<mask> the media is printed millions of times can certainly add up in cost. [NEWLINE] [NEWLINE] Look at the example listed by [Wikipedia about justification and line-wrapping:]( [URL] #Justification_and_line-wrapping) [NEWLINE] [NEWLINE] [STARTQ] We,<mask>, the [ENDQ] [NEWLINE] [STARTQ] representatives of the United [ENDQ] [NEWLINE] [STARTQ] States of America [real-estate] [ENDQ] [NEWLINE] versus [NEWLINE] [NEWLINE] [STARTQ] We,<mask>, the represen- [ENDQ] [NEWLINE] [STARTQ] tatives of the United States [ENDQ] [NEWLINE] [STARTQ] of America [free real-estate!!!] [ENDQ] [NEWLINE] ----- [NEWLINE] Picture a whole corporation employing an association of **pompous, ostentatious, and tenacious aristocracy.** These bourgeoisie bludgeon, assault, lacerate, and persecute these *`meek widows and orphans.`* Small in size the weak are *pushed* and *pulled* in the sway of the robust company.<mask> ever the bourgeoisie does directly impacts the result of the entire corporation. Their word is law.<mask> the orphans and widows are neglected and are paid minimal wages, the aristocrats earn 774 times<mask> much for exactly the same work. [NEWLINE] [NEWLINE] One day the Union comes through and sees the conditions of the workers and decides the best course of action is to cut the wages of the bourgeoisie<mask> everyone receives the same treatment. The new ^bour—geoi—sie now takes up the same room and can no longer bully those of smaller stature. This saves the corporation in total cost<mask> well<mask> real-estate. [NEWLINE] [NEWLINE] ---- [NEWLINE] </s>
Label encoding: <s>There are more components to hyphenating a word than just for column width. Hyphens can alleviate problems with [Widows and Orphans]( [URL] ) on a page, or as has already been mentioned with real-estate on newspapers, magazines, publications, or journal-entries. [NEWLINE] [NEWLINE] Widows and orphans in typography can cause issues with physical media when a single sentence is printed at the top of a clean page. This can be a waste of paper, which if the media is printed millions of times can certainly add up in cost. [NEWLINE] [NEWLINE] Look at the example listed by [Wikipedia about justification and line-wrapping:]( [URL] #Justification_and_line-wrapping) [NEWLINE] [NEWLINE] [STARTQ] We, therefore, the [ENDQ] [NEWLINE] [STARTQ] representatives of the United [ENDQ] [NEWLINE] [STARTQ] States of America [real-estate] [ENDQ] [NEWLINE] versus [NEWLINE] [NEWLINE] [STARTQ] We, therefore, the represen- [ENDQ] [NEWLINE] [STARTQ] tatives of the United States [ENDQ] [NEWLINE] [STARTQ] of America [free real-estate!!!] [ENDQ] [NEWLINE] ----- [NEWLINE] Picture a whole corporation employing an association of **pompous, ostentatious, and tenacious aristocracy.** These bourgeoisie bludgeon, assault, lacerate, and persecute these *`meek widows and orphans.`* Small in size the weak are *pushed* and *pulled* in the sway of the robust company. What ever the bourgeoisie does directly impacts the result of the entire corporation. Their word is law. While the orphans and widows are neglected and are paid minimal wages, the aristocrats earn 774 times as much for exactly the same work. [NEWLINE] [NEWLINE] One day the Union comes through and sees the conditions of the workers and decides the best course of action is to cut the wages of the bourgeoisie so everyone receives the same treatment. The new ^bour—geoi—sie now takes up the same room and can no longer bully those of smaller stature. This saves the corporation in total cost as well as real-estate. [NEWLINE] [NEWLINE] ---- [NEWLINE] </s>
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I would rather lose my rights,<mask> not be killed then keep my rights and die.<mask><mask> many will agree. [ENDQ] [NEWLINE] [STARTQ] I do not understand this bit, add a quote? [ENDQ] [NEWLINE] You would rather. You would rather choose to lose the right to choose rather than die,<mask> you still want the choice to choose<mask> you would rather have? [NEWLINE] [NEWLINE] [STARTQ] This is taking it out of context. Following a party is optional and all the party will do is change the way the country is governed. [ENDQ] Religion changes the way you live. You have no choice in the matter of the rules you must follow in religion.<mask> you go against them you are sinning and must go to hell. [NEWLINE] [NEWLINE] Lol man....<mask><mask><mask><mask> following God wasn't optional everyone would be following God right now. They technically have the choice not to follow God. [NEWLINE] [NEWLINE] <mask> for not having a say in the rules? Do you know<mask> many branches of Christianity there are?<mask> do you think they all came to be?<mask> people chose<mask> the rules are.<mask> technically sin can be wiped away by confession. [NEWLINE] [NEWLINE] [STARTQ] Can you show me a video which will explain<mask> it says in the bible that homosexuality is wrong? [ENDQ] [NEWLINE] For now<mask><mask> most faiths still argue it's wrong, with varying ranges of<mask> wrong. I'm sure that will change eventually<mask> and for now homosexually is still morally debatable. [NEWLINE] [NEWLINE] [STARTQ] Arguably many people in Africa are experiencing more pain than happiness. Does this mean that no one in Africa should give birth? [ENDQ] [NEWLINE] This is arguable. It could be morally wrong to sentence a person to live a horrible life. [NEWLINE] [NEWLINE] [STARTQ] <mask> I will have kids. I will have kids<mask> it keeps the species going. That's the main reason people have kids. Maybe not on the outside [ENDQ] [NEWLINE] Kinda stupid<mask> the race will continue without you and it might benefit<mask> you don't have children. Plus you don't really have anything to back this up with. </s>
Label encoding: <s> [STARTQ] I would rather lose my rights, but not be killed then keep my rights and die. I think many will agree. [ENDQ] [NEWLINE] [STARTQ] I do not understand this bit, add a quote? [ENDQ] [NEWLINE] You would rather. You would rather choose to lose the right to choose rather than die, yet you still want the choice to choose what you would rather have? [NEWLINE] [NEWLINE] [STARTQ] This is taking it out of context. Following a party is optional and all the party will do is change the way the country is governed. [ENDQ] Religion changes the way you live. You have no choice in the matter of the rules you must follow in religion. If you go against them you are sinning and must go to hell. [NEWLINE] [NEWLINE] Lol man.... first of all if following God wasn't optional everyone would be following God right now. They technically have the choice not to follow God. [NEWLINE] [NEWLINE] As for not having a say in the rules? Do you know how many branches of Christianity there are? How do you think they all came to be? Because people chose what the rules are. Also technically sin can be wiped away by confession. [NEWLINE] [NEWLINE] [STARTQ] Can you show me a video which will explain why it says in the bible that homosexuality is wrong? [ENDQ] [NEWLINE] For now I think most faiths still argue it's wrong, with varying ranges of how wrong. I'm sure that will change eventually though and for now homosexually is still morally debatable. [NEWLINE] [NEWLINE] [STARTQ] Arguably many people in Africa are experiencing more pain than happiness. Does this mean that no one in Africa should give birth? [ENDQ] [NEWLINE] This is arguable. It could be morally wrong to sentence a person to live a horrible life. [NEWLINE] [NEWLINE] [STARTQ] But I will have kids. I will have kids because it keeps the species going. That's the main reason people have kids. Maybe not on the outside [ENDQ] [NEWLINE] Kinda stupid since the race will continue without you and it might benefit if you don't have children. Plus you don't really have anything to back this up with. </s>
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Think of the sports teams. [NEWLINE] [NEWLINE] My team<mask> to state or got close nearly every year. [NEWLINE] [NEWLINE] My football coach would get really stressed<mask> he couldn't fit in enough practice time in<mask> it was dark.  That was with morning and weekend practices. [NEWLINE] [NEWLINE] I wasn't on the Basketball team<mask> I do know that there wasn't enough time that the gym was open.  The freshman boys practiced 7:30-10:00pm<mask> the freshman girls practiced from 5:30-7:30am. [NEWLINE] [NEWLINE] 9-5 would wipe out the desirable after school time.  And create problems<mask> the teams aren't getting enough sleep<mask> all the practices are before school. [NEWLINE] [NEWLINE] After football ended I was on the Swim team.  My school did not have its own pool,<mask> we had to rent out lanes at the public schools' Natatorium.  Scheduling was a huge pain, trying to fit 4, sometimes 5 schools, and a club team into this pool. [NEWLINE] [NEWLINE] Now morning swim practice was worse than morning football.   Coaches hated it, janitors hated it, lifeguards hated it, swimmers hated it. [NEWLINE] [NEWLINE] I was<mask> on the track team. [NEWLINE] [NEWLINE] I ran long-distance and threw. [NEWLINE] [NEWLINE] Before 9 o'clock it's often dark.  After 5 it's often dark.  Is it really such a good idea to send High-Schoolers out to run sidewalks in the dark‽  Have you ever thrown a javelin or tried to recover a discus in the dark‽ [NEWLINE] [NEWLINE] I was in Band, it's hard enough to work around sports.  I feel sorry for anyone who has to do it with wacky scheduling. [NEWLINE] [NEWLINE] In short it's just not practical to have school set with the work day.  At least<mask> extra activities are concerned.   School isn't a daycare.  It has been set at the time it is not arbitrarily,<mask> selected<mask> a balance or compromise that works the best consistently.</s>
Label encoding: <s>Think of the sports teams. [NEWLINE] [NEWLINE] My team when to state or got close nearly every year. [NEWLINE] [NEWLINE] My football coach would get really stressed when he couldn't fit in enough practice time in because it was dark.  That was with morning and weekend practices. [NEWLINE] [NEWLINE] I wasn't on the Basketball team but I do know that there wasn't enough time that the gym was open.  The freshman boys practiced 7:30-10:00pm while the freshman girls practiced from 5:30-7:30am. [NEWLINE] [NEWLINE] 9-5 would wipe out the desirable after school time.  And create problems when the teams aren't getting enough sleep because all the practices are before school. [NEWLINE] [NEWLINE] After football ended I was on the Swim team.  My school did not have its own pool, so we had to rent out lanes at the public schools' Natatorium.  Scheduling was a huge pain, trying to fit 4, sometimes 5 schools, and a club team into this pool. [NEWLINE] [NEWLINE] Now morning swim practice was worse than morning football.   Coaches hated it, janitors hated it, lifeguards hated it, swimmers hated it. [NEWLINE] [NEWLINE] I was also on the track team. [NEWLINE] [NEWLINE] I ran long-distance and threw. [NEWLINE] [NEWLINE] Before 9 o'clock it's often dark.  After 5 it's often dark.  Is it really such a good idea to send High-Schoolers out to run sidewalks in the dark‽  Have you ever thrown a javelin or tried to recover a discus in the dark‽ [NEWLINE] [NEWLINE] I was in Band, it's hard enough to work around sports.  I feel sorry for anyone who has to do it with wacky scheduling. [NEWLINE] [NEWLINE] In short it's just not practical to have school set with the work day.  At least when extra activities are concerned.   School isn't a daycare.  It has been set at the time it is not arbitrarily, but selected as a balance or compromise that works the best consistently.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To caveat:<mask> we lived in a free market, I'd agree.<mask> we don't. [NEWLINE] [NEWLINE] Some people have already made the argument that the 1% rely on government infrastructure to make their money, and I'm going to make a related argument. I don't think it matters<mask> you use public roads to make your fortune, or hire people educated by public education,<mask> once you begin colluding with the government to make public goods serve you disproportionately, that's<mask> you can ask serious questions about whether the 1% deserve their money.<mask>, the question is: Do members of the 1% receive protection or assistance from the government that makes their professions more lucrative? Yes, they do. [NEWLINE] [NEWLINE] [Here's]( [URL] /) an infographic breakdown of the professions held by the 1%. Let's look at one of the larger blocks: Lawyers. [NEWLINE] [NEWLINE] Lawyers make money by participating in trials, and providing legal services to wealthy corporations and individuals who want to be compliant or appear compliant with the law. These laws are written by congresspeople -- many of whom are former lawyers themselves, and will return to legal practice<mask> they fail to be re-elected. The expanding federal and state code that all businesses must adhere to are a direct result of the tacit partnership between the legal profession and the federal government. The cost of this partnership to U.S. businesses can be read about in this report called [10,000 Commandments]( [URL] ). The cost for American businesses to comply with these regulations is $1.8 trillion. Not all of this money goes to lawyers,<mask> a good chunk surely does. [NEWLINE] [NEWLINE] This public/private cronyism is pervasive throughout the U.S. economy. It's hard to find a fortune that hasn't been built through collusion with federal and state governments. [NEWLINE] [NEWLINE] TL;DR: Private individuals and businesses collude with the government to rig the game in their favor against their competitors and against consumers.<mask> they do that, it's completely ridiculous for them to claim their fortunes are theirs alone. [NEWLINE] </s>
Label encoding: <s>To caveat: If we lived in a free market, I'd agree. But we don't. [NEWLINE] [NEWLINE] Some people have already made the argument that the 1% rely on government infrastructure to make their money, and I'm going to make a related argument. I don't think it matters if you use public roads to make your fortune, or hire people educated by public education, but once you begin colluding with the government to make public goods serve you disproportionately, that's when you can ask serious questions about whether the 1% deserve their money. So, the question is: Do members of the 1% receive protection or assistance from the government that makes their professions more lucrative? Yes, they do. [NEWLINE] [NEWLINE] [Here's]( [URL] /) an infographic breakdown of the professions held by the 1%. Let's look at one of the larger blocks: Lawyers. [NEWLINE] [NEWLINE] Lawyers make money by participating in trials, and providing legal services to wealthy corporations and individuals who want to be compliant or appear compliant with the law. These laws are written by congresspeople -- many of whom are former lawyers themselves, and will return to legal practice when they fail to be re-elected. The expanding federal and state code that all businesses must adhere to are a direct result of the tacit partnership between the legal profession and the federal government. The cost of this partnership to U.S. businesses can be read about in this report called [10,000 Commandments]( [URL] ). The cost for American businesses to comply with these regulations is $1.8 trillion. Not all of this money goes to lawyers, but a good chunk surely does. [NEWLINE] [NEWLINE] This public/private cronyism is pervasive throughout the U.S. economy. It's hard to find a fortune that hasn't been built through collusion with federal and state governments. [NEWLINE] [NEWLINE] TL;DR: Private individuals and businesses collude with the government to rig the game in their favor against their competitors and against consumers. When they do that, it's completely ridiculous for them to claim their fortunes are theirs alone. [NEWLINE] </s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Psychology is not a valid science<mask> it cannot observe or test empirical data. [ENDQ] [NEWLINE] <mask><mask>. Good psychology only observes and tests empirical data.<mask><mask><mask> you may have a problem with are the theorized mechanisms behind those observations and data. [NEWLINE] [NEWLINE] Every scientific discipline that studies things about which little is known must rely on some degree of theoretical speculation.<mask> those speculations are always based on the best data available and are amenable to change<mask> new data is obtained. There are plenty of phenomena in physics, astronomy, biology, chemistry, etc that have not been directly observed<mask> have been inferred<mask> the rest of the evidence we have collected points very strongly toward a certain conclusion. Psychology is no different. The object of study of psychology might be complex in a unique way<mask> by no means is it impossible to study. [NEWLINE] [NEWLINE] By repeatedly observing human behavior under very specific and constrained conditions, psychologists can narrow down the potential factors underlying a given behavior, often with great precision. Granted, at the end of the day, we call these theoretical psychological constructs things like motivation, self-esteem, depression, personality, and<mask> on and it seems like you might have a problem with that. There is no "thing" in the human body or brain that corresponds directly to one's self-esteem and I can see<mask> some people might be skeptical of trusting such a concept.<mask> in a laboratory, and often in real life, experimenters can manipulate this thing we call self-esteem and predict the results with surprising accuracy based on the body of knowledge that psychological scientists have collected. [NEWLINE] [NEWLINE] You can take issue with the identification of psychological constructs and theoretical speculation in the field of psychology, many psychologists certainly do<mask> well.<mask> I really don't think you can say that psychology is not a valid science<mask> it can't observe or test empirical data. Good psychologists do nothing<mask> empirical observation.<mask> with all science, it's the interpretation of the results or empirical observation that gets tricky. And especially<mask><mask> studying something<mask> complex<mask> human cognition and behavior.</s>
Label encoding: <s> [STARTQ] Psychology is not a valid science because it cannot observe or test empirical data. [ENDQ] [NEWLINE] I disagree. Good psychology only observes and tests empirical data. I think what you may have a problem with are the theorized mechanisms behind those observations and data. [NEWLINE] [NEWLINE] Every scientific discipline that studies things about which little is known must rely on some degree of theoretical speculation. But those speculations are always based on the best data available and are amenable to change as new data is obtained. There are plenty of phenomena in physics, astronomy, biology, chemistry, etc that have not been directly observed but have been inferred because the rest of the evidence we have collected points very strongly toward a certain conclusion. Psychology is no different. The object of study of psychology might be complex in a unique way but by no means is it impossible to study. [NEWLINE] [NEWLINE] By repeatedly observing human behavior under very specific and constrained conditions, psychologists can narrow down the potential factors underlying a given behavior, often with great precision. Granted, at the end of the day, we call these theoretical psychological constructs things like motivation, self-esteem, depression, personality, and so on and it seems like you might have a problem with that. There is no "thing" in the human body or brain that corresponds directly to one's self-esteem and I can see how some people might be skeptical of trusting such a concept. But in a laboratory, and often in real life, experimenters can manipulate this thing we call self-esteem and predict the results with surprising accuracy based on the body of knowledge that psychological scientists have collected. [NEWLINE] [NEWLINE] You can take issue with the identification of psychological constructs and theoretical speculation in the field of psychology, many psychologists certainly do as well. But I really don't think you can say that psychology is not a valid science because it can't observe or test empirical data. Good psychologists do nothing but empirical observation. As with all science, it's the interpretation of the results or empirical observation that gets tricky. And especially so when studying something as complex as human cognition and behavior.</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I want a law that all debt would have to be tied to income- any debt which was above<mask> a person needed to live a safe and happy life wouldn't be legally enforceable and would be abolished. [ENDQ] [NEWLINE] Honestly, I don't think this (which seems to be the core part of your view) even makes sense. Debt is a long term liability and has nothing to do with your annual income.<mask> you can't service your debt on<mask> you make this year then the term of the debt can be lengthened and you can pay back smaller amounts over longer periods of time. The only limit to this is that the borrower must be able to service the annual interest on the debt lest his liability grow over time. [NEWLINE] [NEWLINE] [STARTQ] No one should feel worried that due to debts they have that they will lose their family, home, car, stuff like that. [ENDQ] [NEWLINE] I don't think you are very familiar with<mask> debt repayment and bankruptcy work (at least in the US, apologies<mask> you are writing from somewhere else). For most types of debt the primary residence, a car (up to a certain value), and a surprisingly large amount of personal property, are *exempt* from bankruptcy proceedings. A lender can't just take everything you own to service your debt. [NEWLINE] [NEWLINE] In reality it is only for secured debt (mortgages, car loans)<mask> a specific piece of property can be seized. And in those cases the remainder of your assets are protected. [NEWLINE] [NEWLINE] [NEWLINE] Honestly<mask><mask> you should change your view on the basis that a framework for mitigating the impact on debt repayment by the poor already exists and that simply expanding those protections will achieve much of<mask> you want without requiring a "debt jubilee" which would be economically disruptive. [NEWLINE] [NEWLINE] I mean<mask> more debt was made dischargeable in bankruptcy and the magnitude of exemptions was increased would this not remove much of the burden the exists for getting out of unpayable debt?<mask><mask><mask> not do this vs your unnecessarily radical proposal. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I want a law that all debt would have to be tied to income- any debt which was above what a person needed to live a safe and happy life wouldn't be legally enforceable and would be abolished. [ENDQ] [NEWLINE] Honestly, I don't think this (which seems to be the core part of your view) even makes sense. Debt is a long term liability and has nothing to do with your annual income. If you can't service your debt on what you make this year then the term of the debt can be lengthened and you can pay back smaller amounts over longer periods of time. The only limit to this is that the borrower must be able to service the annual interest on the debt lest his liability grow over time. [NEWLINE] [NEWLINE] [STARTQ] No one should feel worried that due to debts they have that they will lose their family, home, car, stuff like that. [ENDQ] [NEWLINE] I don't think you are very familiar with how debt repayment and bankruptcy work (at least in the US, apologies if you are writing from somewhere else). For most types of debt the primary residence, a car (up to a certain value), and a surprisingly large amount of personal property, are *exempt* from bankruptcy proceedings. A lender can't just take everything you own to service your debt. [NEWLINE] [NEWLINE] In reality it is only for secured debt (mortgages, car loans) where a specific piece of property can be seized. And in those cases the remainder of your assets are protected. [NEWLINE] [NEWLINE] [NEWLINE] Honestly I think you should change your view on the basis that a framework for mitigating the impact on debt repayment by the poor already exists and that simply expanding those protections will achieve much of what you want without requiring a "debt jubilee" which would be economically disruptive. [NEWLINE] [NEWLINE] I mean if more debt was made dischargeable in bankruptcy and the magnitude of exemptions was increased would this not remove much of the burden the exists for getting out of unpayable debt? If so why not do this vs your unnecessarily radical proposal. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I actually think you might have misunderstood my argument a bit. I understand your confusion<mask> most other commenters are arguing in the way you described. Your original view was that you should not be judged harshly for criticizing a culture. This is different than criticizing a specific practice that is one small aspect or occurrence within a group or culture. This is the difference between criticizing a person/group for engaging in a behavior and criticizing the behavior itself. This is<mask> I have placed such a large emphasis on understanding the culture and context. You have to understand the *<mask> * something occurs in order to place blame or criticize. That is<mask> is difficult to untangle. [NEWLINE] [NEWLINE] I'm not saying you have to be an expert to disagree with a cultural outcome.<mask>, you must be an expert to critique a culture. There's a difference between disagreeing with terrorism ( I do) and criticizing the Muslim religion/culture for terrorism. [NEWLINE] [NEWLINE] To criticize or place blame upon the culture or group is<mask> I have issue with, not criticizing a particular practice. [NEWLINE] [NEWLINE] For instance, you can refer to the example of black incarceration and crime I used<mask> an exemplar. The cultural practice or outcome is wrong,<mask> I can't blame the group or culture otherwise criticize the culture without extensive understanding into<mask> are the causes of it. [NEWLINE] [NEWLINE] I.E. [NEWLINE] [NEWLINE] Critiquing practice: Committing crime is wrong. [NEWLINE] [NEWLINE] Critiquing culture/group:Black culture is wrong for causing<mask> much crime. [NEWLINE] [NEWLINE] Do you blame American culture for serial killers public gunmen (school shootings) or is it an unavoidable aspect of society, caused by another phenomenon entirely, or any other reason America is not able to criticized for this? Not most Americans are serial killers and not most Muslims are terrorists. Before you can even begin to untangle this issue you need to have a deep understanding of the culture and the origins of the practice. [NEWLINE] [NEWLINE] Sorry for any confusion, try rereading my argument with this difference in mind and see<mask> anything changes.</s>
Label encoding: <s>I actually think you might have misunderstood my argument a bit. I understand your confusion since most other commenters are arguing in the way you described. Your original view was that you should not be judged harshly for criticizing a culture. This is different than criticizing a specific practice that is one small aspect or occurrence within a group or culture. This is the difference between criticizing a person/group for engaging in a behavior and criticizing the behavior itself. This is why I have placed such a large emphasis on understanding the culture and context. You have to understand the * why * something occurs in order to place blame or criticize. That is what is difficult to untangle. [NEWLINE] [NEWLINE] I'm not saying you have to be an expert to disagree with a cultural outcome. However, you must be an expert to critique a culture. There's a difference between disagreeing with terrorism ( I do) and criticizing the Muslim religion/culture for terrorism. [NEWLINE] [NEWLINE] To criticize or place blame upon the culture or group is what I have issue with, not criticizing a particular practice. [NEWLINE] [NEWLINE] For instance, you can refer to the example of black incarceration and crime I used as an exemplar. The cultural practice or outcome is wrong, but I can't blame the group or culture otherwise criticize the culture without extensive understanding into what are the causes of it. [NEWLINE] [NEWLINE] I.E. [NEWLINE] [NEWLINE] Critiquing practice: Committing crime is wrong. [NEWLINE] [NEWLINE] Critiquing culture/group:Black culture is wrong for causing so much crime. [NEWLINE] [NEWLINE] Do you blame American culture for serial killers public gunmen (school shootings) or is it an unavoidable aspect of society, caused by another phenomenon entirely, or any other reason America is not able to criticized for this? Not most Americans are serial killers and not most Muslims are terrorists. Before you can even begin to untangle this issue you need to have a deep understanding of the culture and the origins of the practice. [NEWLINE] [NEWLINE] Sorry for any confusion, try rereading my argument with this difference in mind and see if anything changes.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Stonewall was full of trans* people both before and during the raid that led to the riots. The movement<mask> we know it wouldn't have have started<mask> it weren't for trans* people [ENDQ] [NEWLINE] This is simply absolutely false.  This myth was even debunked many years later by the woman who started it:  Sylvia Rivera, who said, and I quote, "The Stonewall wasn't a bar for drag queens. Everybody keeps saying it was....<mask> you were a drag queen, you could get into the Stonewall<mask> they knew you. And only a certain number of drag queens were allowed into the Stonewall at that time. [NEWLINE] [NEWLINE] Before you continue to spread this misinformation, I suggest you do a little reading: [NEWLINE] [NEWLINE] [URL] /#sthash.CouHR4jz.dpuf [NEWLINE] [NEWLINE] [STARTQ] It makes me<mask> sad to see you take this attitude, and it goes deeper than a sense of betrayal [ENDQ] [NEWLINE] Oh<mask> pitiful stuff.  Come on, make an argument, don't just appeal to emotion.  Just<mask> it upsets you that the causes shouldn't be linked, doesn't mean that they should. [NEWLINE] [NEWLINE] [STARTQ] We want you to succeed, too.<mask> do you feel a sense of hate or scorn that's<mask> strong you need to actively work against trans* people? [ENDQ] [NEWLINE] Are you serious?  Read some of my other posts<mask> I mention repeatedly that I am 100% pro trans rights. <mask> I point out that the reason the groups should work separately is<mask> the rights you are looking for a very different than the rights gay people are working towards.  That<mask> we can be a community together, we make no sense<mask> a *political movement* together. [NEWLINE] [NEWLINE] I mean, you're trying to make this all about you and having your feelings hurt, and not about actually having a debate on the merits.  I'm happy to do that with you,<mask> please don't put words in my mouth. [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] Stonewall was full of trans* people both before and during the raid that led to the riots. The movement as we know it wouldn't have have started if it weren't for trans* people [ENDQ] [NEWLINE] This is simply absolutely false.  This myth was even debunked many years later by the woman who started it:  Sylvia Rivera, who said, and I quote, "The Stonewall wasn't a bar for drag queens. Everybody keeps saying it was.... If you were a drag queen, you could get into the Stonewall if they knew you. And only a certain number of drag queens were allowed into the Stonewall at that time. [NEWLINE] [NEWLINE] Before you continue to spread this misinformation, I suggest you do a little reading: [NEWLINE] [NEWLINE] [URL] /#sthash.CouHR4jz.dpuf [NEWLINE] [NEWLINE] [STARTQ] It makes me so sad to see you take this attitude, and it goes deeper than a sense of betrayal [ENDQ] [NEWLINE] Oh what pitiful stuff.  Come on, make an argument, don't just appeal to emotion.  Just because it upsets you that the causes shouldn't be linked, doesn't mean that they should. [NEWLINE] [NEWLINE] [STARTQ] We want you to succeed, too. Why do you feel a sense of hate or scorn that's so strong you need to actively work against trans* people? [ENDQ] [NEWLINE] Are you serious?  Read some of my other posts where I mention repeatedly that I am 100% pro trans rights.  Where I point out that the reason the groups should work separately is because the rights you are looking for a very different than the rights gay people are working towards.  That while we can be a community together, we make no sense as a *political movement* together. [NEWLINE] [NEWLINE] I mean, you're trying to make this all about you and having your feelings hurt, and not about actually having a debate on the merits.  I'm happy to do that with you, but please don't put words in my mouth. [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you're a utilitarian, they appear to be the same,<mask> from a utilitarian standpoint, all of the factors are the same.  All that matters for a utilitarian is "net utility." [NEWLINE] [NEWLINE] <mask>, a lot of people's moral intuition says that they're not and that actively and intentionally murdering someone is worse than pulling the trolley switch. [NEWLINE] [NEWLINE] <mask><mask> you'd enjoy looking at the [principle of double effect]( [URL] ). <mask> was formulated by Thomas Aquinas and still used today by some philosophers and theologians, mainly in discussions of medical ethics and war ethics.  Essentially, it says that an action is morally good<mask> foreseen negative consequences<mask> : [NEWLINE] [NEWLINE] 1) the nature of the act is good or neutral itself [NEWLINE] [NEWLINE] 2) the good is intended and bad is not [NEWLINE] [NEWLINE] 3) the good outweighs the bad [NEWLINE] [NEWLINE] Now, a utilitarian would only accept the third of these principals,<mask><mask> you're willing to accept that these are morally relevant features for some other legitimate moral frameworks, you should be able to see the differences between the trolley problem and the transplant problem. [NEWLINE] [NEWLINE] In the trolley problem, all three points are fulfilled.  Pulling a switch is not morally good or bad on its own, and the death of the solitary man,<mask> foreseen, is not intended.  In the transplant problem, killing the innocent man is normally seen<mask> a violation of the first principle.  Killing the man to take his organs<mask> violates the second point,<mask> his death is both intended and necessary to achieve the goals, whereas in the trolley problem<mask> you throw the switch, the only intent is to save the five, not to kill the one. [NEWLINE] [NEWLINE] Again, utilitarians like yourself and other consequentialist ethicists probably will not agree that things like "intentions" matter at all,<mask> I hope you can at least see  there are some legitimately recognized philosophical distinctions between the trolley problem and the transplant problem, not just psychological ones.</s>
Label encoding: <s>If you're a utilitarian, they appear to be the same, because from a utilitarian standpoint, all of the factors are the same.  All that matters for a utilitarian is "net utility." [NEWLINE] [NEWLINE] However, a lot of people's moral intuition says that they're not and that actively and intentionally murdering someone is worse than pulling the trolley switch. [NEWLINE] [NEWLINE] I think you'd enjoy looking at the [principle of double effect]( [URL] ).  If was formulated by Thomas Aquinas and still used today by some philosophers and theologians, mainly in discussions of medical ethics and war ethics.  Essentially, it says that an action is morally good despite foreseen negative consequences if : [NEWLINE] [NEWLINE] 1) the nature of the act is good or neutral itself [NEWLINE] [NEWLINE] 2) the good is intended and bad is not [NEWLINE] [NEWLINE] 3) the good outweighs the bad [NEWLINE] [NEWLINE] Now, a utilitarian would only accept the third of these principals, but if you're willing to accept that these are morally relevant features for some other legitimate moral frameworks, you should be able to see the differences between the trolley problem and the transplant problem. [NEWLINE] [NEWLINE] In the trolley problem, all three points are fulfilled.  Pulling a switch is not morally good or bad on its own, and the death of the solitary man, while foreseen, is not intended.  In the transplant problem, killing the innocent man is normally seen as a violation of the first principle.  Killing the man to take his organs also violates the second point, since his death is both intended and necessary to achieve the goals, whereas in the trolley problem when you throw the switch, the only intent is to save the five, not to kill the one. [NEWLINE] [NEWLINE] Again, utilitarians like yourself and other consequentialist ethicists probably will not agree that things like "intentions" matter at all, but I hope you can at least see  there are some legitimately recognized philosophical distinctions between the trolley problem and the transplant problem, not just psychological ones.</s>
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was a leash kid in the 90s! [NEWLINE] [NEWLINE] My mom did it for 3 great reasons. [NEWLINE] [NEWLINE] 1) She slipped a disk giving birth to mean. She then had a botched surgery to repair it that left her with chronic pain. Which meant behind over to constantly hold my hand at times physically impossible. [NEWLINE] [NEWLINE] 2) I LOVED running away. My favourite game was playing in traffic. One day<mask> she was recovering from one of her back surgeries she looked out the window to discover I'd escaped my grandmother's care and was making a bee-line for the road in front of our house. She got out of bed, hauled ass out to the front year<mask> she couldn't pick me up.<mask> she'd try to shuffle me back towards the house.<mask> I was pretty quick and would try to dart around her or between her legs. I was pushing us closer and closer to the road<mask> I'd keep getting around her. Eventually she kicked me over and pinned me to the ground with her foot until my grandmother (who was at the time dealing with a cut my sister got from falling into the coffee table) heard her hollering and came and scooped me up. [NEWLINE] [NEWLINE] I'm really glad child services never came around. That wouldn't have looked good. [NEWLINE] [NEWLINE] 3) my mom had another small child to deal with. My sister's barely a year older than me. My sister wasn't a take-off-and-explore kid<mask> she needed attention sometimes too and with my penchant for taking off whenever at all possible it was just safer to have me wear "the backpack". My sister was a pretty shy leg-hugger kid which mean it often took my mom a lot of coaxing to get her to interact with other kids or people.<mask> she was convincing my sister to join the other kids on the playground I'd be doing the cha-cha straight for the three lane roadway. The only thing keeping me from getting too far was my backpack restraint.</s>
Label encoding: <s>I was a leash kid in the 90s! [NEWLINE] [NEWLINE] My mom did it for 3 great reasons. [NEWLINE] [NEWLINE] 1) She slipped a disk giving birth to mean. She then had a botched surgery to repair it that left her with chronic pain. Which meant behind over to constantly hold my hand at times physically impossible. [NEWLINE] [NEWLINE] 2) I LOVED running away. My favourite game was playing in traffic. One day while she was recovering from one of her back surgeries she looked out the window to discover I'd escaped my grandmother's care and was making a bee-line for the road in front of our house. She got out of bed, hauled ass out to the front year but she couldn't pick me up. So she'd try to shuffle me back towards the house. But I was pretty quick and would try to dart around her or between her legs. I was pushing us closer and closer to the road because I'd keep getting around her. Eventually she kicked me over and pinned me to the ground with her foot until my grandmother (who was at the time dealing with a cut my sister got from falling into the coffee table) heard her hollering and came and scooped me up. [NEWLINE] [NEWLINE] I'm really glad child services never came around. That wouldn't have looked good. [NEWLINE] [NEWLINE] 3) my mom had another small child to deal with. My sister's barely a year older than me. My sister wasn't a take-off-and-explore kid but she needed attention sometimes too and with my penchant for taking off whenever at all possible it was just safer to have me wear "the backpack". My sister was a pretty shy leg-hugger kid which mean it often took my mom a lot of coaxing to get her to interact with other kids or people. While she was convincing my sister to join the other kids on the playground I'd be doing the cha-cha straight for the three lane roadway. The only thing keeping me from getting too far was my backpack restraint.</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From a completely economic standpoint, the minimum wage is an artificial bottom line that creates deadweight loss in a free market;<mask> a person is paid more than their work is valued then the employer may choose not to hire more employees and may raise their prices in accordance with the raising of minimum wages to match profits. I still believe a basic minimum wage should be in place (I am making a little above minimum wage<mask> that may contribute to it)<mask> the U.S. market is not perfectly competitive and oligopolies can manipulate prices and wages more than a competitive company can.<mask> obviously the argument is going to come up about<mask> the line should be drawn<mask> I do not believe the answer is to continuously raise it until everyone is happy<mask> that really isn't ever going to happen. My argument against minimum wage increase is shoddy and you guys can rip it apart and tell me<mask> I'm wrong<mask> I firmly believe that the steep progressive tax rate in the U.S. harms our economy more than it helps. A progressive tax punishes innovation and hard work, it is a deterrent for efficiency. These are the things that are going to stimulate the economy and provide for economic growth, not collecting a few extra bucks from those who have worked for it. (Here's the argument that not everyone worked for their money which I will acknowledge and which I cannot come up with a full response<mask> again, there are those who didn't make their own money<mask> there are those that did). Another argument that I forsee is that the people who make a lot of money do not need all of that money.<mask> just<mask> they did not make all of that money does not mean that money is not entitled to them. I can't just take someone else's money<mask> he has more than me even<mask> I need it more.<mask> an aside, the current tax code is over 70k words and extremely obfuscated causing people hours and hours of frustration<mask> filling out their tax forms. A flat tax would simplify the process greatly.</s>
Label encoding: <s>From a completely economic standpoint, the minimum wage is an artificial bottom line that creates deadweight loss in a free market; if a person is paid more than their work is valued then the employer may choose not to hire more employees and may raise their prices in accordance with the raising of minimum wages to match profits. I still believe a basic minimum wage should be in place (I am making a little above minimum wage so that may contribute to it) because the U.S. market is not perfectly competitive and oligopolies can manipulate prices and wages more than a competitive company can. So obviously the argument is going to come up about where the line should be drawn but I do not believe the answer is to continuously raise it until everyone is happy because that really isn't ever going to happen. My argument against minimum wage increase is shoddy and you guys can rip it apart and tell me why I'm wrong but I firmly believe that the steep progressive tax rate in the U.S. harms our economy more than it helps. A progressive tax punishes innovation and hard work, it is a deterrent for efficiency. These are the things that are going to stimulate the economy and provide for economic growth, not collecting a few extra bucks from those who have worked for it. (Here's the argument that not everyone worked for their money which I will acknowledge and which I cannot come up with a full response but again, there are those who didn't make their own money but there are those that did). Another argument that I forsee is that the people who make a lot of money do not need all of that money. However just because they did not make all of that money does not mean that money is not entitled to them. I can't just take someone else's money because he has more than me even if I need it more. As an aside, the current tax code is over 70k words and extremely obfuscated causing people hours and hours of frustration while filling out their tax forms. A flat tax would simplify the process greatly.</s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My only objections are on a practical level (<mask> I personally don't think I could ever eat cat). [NEWLINE] [NEWLINE] Cats are generally pretty small and don't have a lot of meat on them. They<mask> require quite a bit of food to keep them going and that food isn't cheap. [NEWLINE] [NEWLINE] With dogs, they're larger<mask> from<mask> I've read about dog, it tends to be rather greasy and doesn't taste terribly good. You<mask> run into the feed problem. [NEWLINE] [NEWLINE] Basically, both types of animals require a lot of expensive maintenance and don't really give you a lot of decent quality meat at the end. They're useful animals to people (cats catch mice and are companions, dogs do a wide array of different jobs and can<mask> be companions)<mask> they're not really suited for eating. [NEWLINE] [NEWLINE] <mask> you look at areas<mask> consumption of these animals is frequent, the animals are rarely raised specifically for food. They're often strays that are snatched up or even just family pets that were sold off by owners who didn't want them or even straight up kid(pet?)napped. [NEWLINE] [NEWLINE] That's<mask>, economically, paying to feed and keep a cat or dog for the length of time it would take for one to go from birth to "eatin' size" would require an asking price that would be *way* above<mask> people would consider worth it for the actual product. [NEWLINE] [NEWLINE] That means that the only way there can actually be a market is<mask> the people who sell the meat have to sell meat they obtain through less scrupulous ways that tend to encourage the maltreatment of the animals. The market can't really exist any other way. [NEWLINE] [NEWLINE] The ban is partially due to people feeling like they wouldn't want to eat Fluffy or Mittens<mask> it's<mask> due to the economics of<mask> the market for the product would actually look like in that it can only exist<mask> the animals involved are being maltreated or acquired through less than legal means.</s>
Label encoding: <s>My only objections are on a practical level ( though I personally don't think I could ever eat cat). [NEWLINE] [NEWLINE] Cats are generally pretty small and don't have a lot of meat on them. They also require quite a bit of food to keep them going and that food isn't cheap. [NEWLINE] [NEWLINE] With dogs, they're larger but from what I've read about dog, it tends to be rather greasy and doesn't taste terribly good. You also run into the feed problem. [NEWLINE] [NEWLINE] Basically, both types of animals require a lot of expensive maintenance and don't really give you a lot of decent quality meat at the end. They're useful animals to people (cats catch mice and are companions, dogs do a wide array of different jobs and can also be companions) but they're not really suited for eating. [NEWLINE] [NEWLINE] If you look at areas where consumption of these animals is frequent, the animals are rarely raised specifically for food. They're often strays that are snatched up or even just family pets that were sold off by owners who didn't want them or even straight up kid(pet?)napped. [NEWLINE] [NEWLINE] That's because, economically, paying to feed and keep a cat or dog for the length of time it would take for one to go from birth to "eatin' size" would require an asking price that would be *way* above what people would consider worth it for the actual product. [NEWLINE] [NEWLINE] That means that the only way there can actually be a market is if the people who sell the meat have to sell meat they obtain through less scrupulous ways that tend to encourage the maltreatment of the animals. The market can't really exist any other way. [NEWLINE] [NEWLINE] The ban is partially due to people feeling like they wouldn't want to eat Fluffy or Mittens but it's also due to the economics of what the market for the product would actually look like in that it can only exist if the animals involved are being maltreated or acquired through less than legal means.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your title stated that you thought that gym leaders in general were too easy. It looks like the view that you're actually asking people to change is that Darmanitan is over-powered. [NEWLINE] [NEWLINE] It's hard to really argue with that one. He's got an insane attack stat and the ground type gym leader (who should nominally be strong against him) has heavy pokemon who are hit hard by grass knot.<mask> it helps, he's supposed to beat Brycen. Fire is strong against Ice. Beating Elesa isn't a huge surprise,<mask>  electric types are generally frail,<mask> fast (and flying types generally fit a similar mold).<mask> you're using him<mask> much<mask> you say and his level is above par with the gym leaders, he should be destroying them. Luckily for you, the last gym should provide him with more of a challenge,<mask> dragon resists all of his non-fighting attacks and can hit hard back. [NEWLINE] [NEWLINE] <mask> he's powerful to the point<mask> it's not fun, there's always the easy answer of not using him. Otherwise, he's going to plow through everything that isn't a physical wall that resists his attacks, of which there aren't many in the game (<mask> a gym leader that takes little damage from heavy hitters and is generally equipped with healing items is a pain in the ass to fight). [NEWLINE] [NEWLINE] There's been a pokemon this powerful in pretty much every pokemon game. The first two generations had Kadabra, who was fast, had a great special attack, and got access to all the moves he needed to dominate (especially in gen 2,<mask> he could learn the then-special elemental punches). Gens 3 and 4 had the birds, Swellow and Staraptor, both of whom were fast, had reasonable move diversity, and were strong. Gen 6 has... pretty much every new pokemon that isn't Aromatisse. Seriously, the gen 6 mons are awesome and strong. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Your title stated that you thought that gym leaders in general were too easy. It looks like the view that you're actually asking people to change is that Darmanitan is over-powered. [NEWLINE] [NEWLINE] It's hard to really argue with that one. He's got an insane attack stat and the ground type gym leader (who should nominally be strong against him) has heavy pokemon who are hit hard by grass knot. If it helps, he's supposed to beat Brycen. Fire is strong against Ice. Beating Elesa isn't a huge surprise, since  electric types are generally frail, but fast (and flying types generally fit a similar mold). If you're using him as much as you say and his level is above par with the gym leaders, he should be destroying them. Luckily for you, the last gym should provide him with more of a challenge, since dragon resists all of his non-fighting attacks and can hit hard back. [NEWLINE] [NEWLINE] If he's powerful to the point where it's not fun, there's always the easy answer of not using him. Otherwise, he's going to plow through everything that isn't a physical wall that resists his attacks, of which there aren't many in the game ( because a gym leader that takes little damage from heavy hitters and is generally equipped with healing items is a pain in the ass to fight). [NEWLINE] [NEWLINE] There's been a pokemon this powerful in pretty much every pokemon game. The first two generations had Kadabra, who was fast, had a great special attack, and got access to all the moves he needed to dominate (especially in gen 2, where he could learn the then-special elemental punches). Gens 3 and 4 had the birds, Swellow and Staraptor, both of whom were fast, had reasonable move diversity, and were strong. Gen 6 has... pretty much every new pokemon that isn't Aromatisse. Seriously, the gen 6 mons are awesome and strong. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0618, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You missed the point of my comment. [NEWLINE] [NEWLINE] [STARTQ] Americans of European and Americans of African extraction both have long and storied histories. Both groups have achievements and skeletons in their closets. [ENDQ] [NEWLINE] This is vague to the point of meaningless. [NEWLINE] [NEWLINE] [STARTQ] Europeans have done a lot of horrifying things, that is simply undeniable.<mask>, they're not all evil, they've done a lot of wonderful things too. [ENDQ] [NEWLINE] <mask> really vague. And not based on anything that I said. [NEWLINE] [NEWLINE] [STARTQ] Everyone should be allowed to be proud of who they are. It's simply racist to tell someone that they're not allowed to be proud (like everyone else) simply<mask> of their skin colour. [ENDQ] [NEWLINE] Again, missing the point. [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] Black people in America were forcefully relocated from their home, many being of different heritages.<mask><mask> they came to America, their *entire culture was destroyed.* They were unified due to slavery, segregation, and discrimination in the U.S. [NEWLINE] [NEWLINE] This is not true for American Europeans. [NEWLINE] [NEWLINE] There is nothing that unifies Americans with European descent *based on being white*. Well, maybe<mask> you want to count being the benefactors of racism and discrimination, be my guest. They all still have ties back to their respective countries,<mask> there is otherwise nothing that unifies white Americans on the basis of their skin. [NEWLINE] [NEWLINE] No shared heritage.<mask>?<mask> many have ancestors from different parts of Europe. [NEWLINE] [NEWLINE] *That doesn't mean they don't have things to be proud of.* [NEWLINE] [NEWLINE] They are still American, and American culture exists.<mask> American culture =/= white culture.<mask> you're proud of being American, you're not proud of being white. Many people from different heritages make up America. [NEWLINE] [NEWLINE] There is a very important historical and social context that your comment lacked, and is<mask> made it vague to the point of being meaningless. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>You missed the point of my comment. [NEWLINE] [NEWLINE] [STARTQ] Americans of European and Americans of African extraction both have long and storied histories. Both groups have achievements and skeletons in their closets. [ENDQ] [NEWLINE] This is vague to the point of meaningless. [NEWLINE] [NEWLINE] [STARTQ] Europeans have done a lot of horrifying things, that is simply undeniable. But, they're not all evil, they've done a lot of wonderful things too. [ENDQ] [NEWLINE] Also really vague. And not based on anything that I said. [NEWLINE] [NEWLINE] [STARTQ] Everyone should be allowed to be proud of who they are. It's simply racist to tell someone that they're not allowed to be proud (like everyone else) simply because of their skin colour. [ENDQ] [NEWLINE] Again, missing the point. [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] Black people in America were forcefully relocated from their home, many being of different heritages. But when they came to America, their *entire culture was destroyed.* They were unified due to slavery, segregation, and discrimination in the U.S. [NEWLINE] [NEWLINE] This is not true for American Europeans. [NEWLINE] [NEWLINE] There is nothing that unifies Americans with European descent *based on being white*. Well, maybe if you want to count being the benefactors of racism and discrimination, be my guest. They all still have ties back to their respective countries, but there is otherwise nothing that unifies white Americans on the basis of their skin. [NEWLINE] [NEWLINE] No shared heritage. Why? Because many have ancestors from different parts of Europe. [NEWLINE] [NEWLINE] *That doesn't mean they don't have things to be proud of.* [NEWLINE] [NEWLINE] They are still American, and American culture exists. But American culture =/= white culture. When you're proud of being American, you're not proud of being white. Many people from different heritages make up America. [NEWLINE] [NEWLINE] There is a very important historical and social context that your comment lacked, and is what made it vague to the point of being meaningless. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Underrepresentation is a symptom of problems, which is<mask> it gets attention.Basically, there's no "good" reason<mask> there would be a gender disparity in STEM fields.<mask> there is a disparity, we know it's<mask> of a problem somewhere, which may be that "employers are discriminating against women" or that "women don't see a career in this<mask> realistic" or even that "girls are taught not to pursue this field." Not all of these are the employers' fault, and all of them happen to some extent or other (and, to a far lesser degree, to boys and men.) [NEWLINE] [NEWLINE] [STARTQ] The current approach is "forcing" the proportion of women to increase, by means of: - gender-specific student grants, - positions reserved for women, - lower physical requirements, - etc. [ENDQ] [NEWLINE] There is a reason for these measures, which (you are correct) is not about getting at the heart of the problem of discrimination.<mask> ending discrimination is a difficult, long-term goal that doesn't have an easy solution, ending *the effects of discrimination on women now* is something we *can* try to fix. All these measures are about equalizing the opportunities available between men and women entering the workforce. Note that they are not, in all circumstances, perfect at this.<mask> women in the US, today, do have a right to not experience discrimination, and protecting that right can mean legal action. [NEWLINE] [NEWLINE] Note that your points about "the animosity of male coworkers" is comparing to an ideal, discrimination-free society. The fact is that women were *already* discriminated against in the workplace. Before affirmative action, there already *were* stereotypes about "sleeping her way up the ladder"<mask> people just assumed that women got the job for being women. The facts of discrimination, in that sense, haven't changed,<mask> there is still better equality of opportunity to get those jobs in the first place. </s>
Label encoding: <s>Underrepresentation is a symptom of problems, which is why it gets attention.Basically, there's no "good" reason why there would be a gender disparity in STEM fields. If there is a disparity, we know it's because of a problem somewhere, which may be that "employers are discriminating against women" or that "women don't see a career in this as realistic" or even that "girls are taught not to pursue this field." Not all of these are the employers' fault, and all of them happen to some extent or other (and, to a far lesser degree, to boys and men.) [NEWLINE] [NEWLINE] [STARTQ] The current approach is "forcing" the proportion of women to increase, by means of: - gender-specific student grants, - positions reserved for women, - lower physical requirements, - etc. [ENDQ] [NEWLINE] There is a reason for these measures, which (you are correct) is not about getting at the heart of the problem of discrimination. While ending discrimination is a difficult, long-term goal that doesn't have an easy solution, ending *the effects of discrimination on women now* is something we *can* try to fix. All these measures are about equalizing the opportunities available between men and women entering the workforce. Note that they are not, in all circumstances, perfect at this. But women in the US, today, do have a right to not experience discrimination, and protecting that right can mean legal action. [NEWLINE] [NEWLINE] Note that your points about "the animosity of male coworkers" is comparing to an ideal, discrimination-free society. The fact is that women were *already* discriminated against in the workplace. Before affirmative action, there already *were* stereotypes about "sleeping her way up the ladder" where people just assumed that women got the job for being women. The facts of discrimination, in that sense, haven't changed, but there is still better equality of opportunity to get those jobs in the first place. </s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> it would be immoral to bring a child into the world without being capable of providing for it.<mask><mask> it is<mask> immoral to force a man to pay for a child he didn't want.<mask> a woman can't raise the child without the father's support the moral thing to do would be to abort the child. [ENDQ] [NEWLINE] Sure, sure.<mask> on a societal scale, we can't ignore results - it isn't enough to focus on<mask> is moral. [NEWLINE] [NEWLINE] The likely practical outcome of this policy would be more children growing up with less financial support.<mask> of a variety of reasons: [NEWLINE] [NEWLINE] 1. The mother is wishful and thinks the father will "come around". [NEWLINE] 2. The mother sees abortion<mask> immoral. [NEWLINE] 3. The mother has a hard time deciding, and accepts the "default" - not to medically intervene, i.e., carry to term. [NEWLINE] 4. The mother has medical concerns that make abortion risky for some reason. [NEWLINE] [NEWLINE] and on the father's side [NEWLINE] [NEWLINE] 1. Having a foolproof method of getting out of child support means men don't need to be<mask> careful about using condoms ("<mask> she gets pregnant, I'll sign a piece of paper"). That means more unsafe sex, leading to both more pregnancies and more STDs and<mask> forth. [NEWLINE] [NEWLINE] We don't live in a perfect world. It would be nice<mask> men could have all the sex they want, and<mask> the woman happens to get pregnant<mask> taking precautions, they can avoid that affecting the rest of their life. That's great for the men.<mask> for society, it means more children growing up with less financial support. Which we know causes serious problems. [NEWLINE] [NEWLINE] Now sure, you have some guesses about this all working out for the best.<mask> even in the **most** optimistic case, you are talking about an experiment that no human society has ever done, with potentially grevious results.</s>
Label encoding: <s> [STARTQ] I think it would be immoral to bring a child into the world without being capable of providing for it. I think it is also immoral to force a man to pay for a child he didn't want. If a woman can't raise the child without the father's support the moral thing to do would be to abort the child. [ENDQ] [NEWLINE] Sure, sure. But on a societal scale, we can't ignore results - it isn't enough to focus on what is moral. [NEWLINE] [NEWLINE] The likely practical outcome of this policy would be more children growing up with less financial support. Because of a variety of reasons: [NEWLINE] [NEWLINE] 1. The mother is wishful and thinks the father will "come around". [NEWLINE] 2. The mother sees abortion as immoral. [NEWLINE] 3. The mother has a hard time deciding, and accepts the "default" - not to medically intervene, i.e., carry to term. [NEWLINE] 4. The mother has medical concerns that make abortion risky for some reason. [NEWLINE] [NEWLINE] and on the father's side [NEWLINE] [NEWLINE] 1. Having a foolproof method of getting out of child support means men don't need to be as careful about using condoms (" if she gets pregnant, I'll sign a piece of paper"). That means more unsafe sex, leading to both more pregnancies and more STDs and so forth. [NEWLINE] [NEWLINE] We don't live in a perfect world. It would be nice if men could have all the sex they want, and if the woman happens to get pregnant despite taking precautions, they can avoid that affecting the rest of their life. That's great for the men. But for society, it means more children growing up with less financial support. Which we know causes serious problems. [NEWLINE] [NEWLINE] Now sure, you have some guesses about this all working out for the best. But even in the **most** optimistic case, you are talking about an experiment that no human society has ever done, with potentially grevious results.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A drunk driver is not evil on any level. I would assume in a majority of circumstances a drunk driver has no desire to kill anyone -<mask> even one that would muse the legitimacy of the word evil would probably not label a drunk driver<mask> such. [ENDQ] [STARTQ] The "<mask> " is absolutely necessary<mask> evaluating one's actions on a moral level. [ENDQ] [NEWLINE] Perhaps I was unclear, I wasn't saying the drunk driver's actions were evil. I was saying they were immoral. Morality isn't just "evil" and "not evil." That drunk driver's actions are *morally wrong*<mask><mask> the<mask>, making the<mask> not absolutely necessary<mask> evaluating actions on a moral level. [NEWLINE] [NEWLINE] [STARTQ] <mask> a man slaughters a family... it is easy to call this person evil on the surface.<mask> the man had a terrible car accident 5 years before... and it materializes in him becoming suddenly very mentally ill and violent...<mask> it makes his actions no less right or easy to deal with... it may effect ones judgement on calling him "evil". [ENDQ] [NEWLINE] I'm not saying that the<mask> must always be discounted or that it never matters, rather I am positing that there are some actions<mask> immoral that they can be considered evil without regarding the<mask>. We might reevaluate in light of new information, sure,<mask> I can't see any kind of justification for attempted genocide that would make me think I needed to not use the evil label. [NEWLINE] [NEWLINE] [STARTQ] This is still the definition of a very abstract concept, and I am defending the position that the concept<mask> a whole is inadequate in representing a person or their actions. [ENDQ] [NEWLINE] Morality itself is an abstract concept, and unless you're saying that we can't evaluate any actions on a moral level at all I don't see any issue with the word evil describing something "very immoral." It's always going to be somewhat subjective and arbitrary (unless you believe in absolute morality).</s>
Label encoding: <s> [STARTQ] A drunk driver is not evil on any level. I would assume in a majority of circumstances a drunk driver has no desire to kill anyone - so even one that would muse the legitimacy of the word evil would probably not label a drunk driver as such. [ENDQ] [STARTQ] The " why " is absolutely necessary when evaluating one's actions on a moral level. [ENDQ] [NEWLINE] Perhaps I was unclear, I wasn't saying the drunk driver's actions were evil. I was saying they were immoral. Morality isn't just "evil" and "not evil." That drunk driver's actions are *morally wrong* regardless of the why, making the why not absolutely necessary when evaluating actions on a moral level. [NEWLINE] [NEWLINE] [STARTQ] If a man slaughters a family... it is easy to call this person evil on the surface. If the man had a terrible car accident 5 years before... and it materializes in him becoming suddenly very mentally ill and violent... while it makes his actions no less right or easy to deal with... it may effect ones judgement on calling him "evil". [ENDQ] [NEWLINE] I'm not saying that the why must always be discounted or that it never matters, rather I am positing that there are some actions so immoral that they can be considered evil without regarding the why. We might reevaluate in light of new information, sure, but I can't see any kind of justification for attempted genocide that would make me think I needed to not use the evil label. [NEWLINE] [NEWLINE] [STARTQ] This is still the definition of a very abstract concept, and I am defending the position that the concept as a whole is inadequate in representing a person or their actions. [ENDQ] [NEWLINE] Morality itself is an abstract concept, and unless you're saying that we can't evaluate any actions on a moral level at all I don't see any issue with the word evil describing something "very immoral." It's always going to be somewhat subjective and arbitrary (unless you believe in absolute morality).</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They are all equally valid. [ENDQ] [NEWLINE] Is this true for all examples?  For example, Joe only dates women who are under 400 pounds<mask> Bob only dates women who idolize Hitler.  Are you really saying those two things are equal?  Or have you just arbitrarily decided that your particular values are equally valid to physical attractiveness?  And<mask><mask>, isn't your whole argument circular? [NEWLINE] [NEWLINE] [STARTQ] I don't want to have to think of the dicks she's sucked on or the semen she's had to clean off her body from every other guy before me. The notion of a guy sucking on her body in the same places I do or intend to is wholly unappealing to me. [ENDQ] [NEWLINE] This doesn't explain your view at all.  A woman who has been with one guy 1000 times is probably more likely to have been touched everywhere or had a lot of experience with semen than a woman who had six short stints with six different guys.  Remember, your view is against the number sexual partners, not the total amount of sexual experience. [NEWLINE] [NEWLINE] [STARTQ] Now,<mask> you can convince me that there is some unknown value there for me, I'd happily change my view. [ENDQ] [NEWLINE] Experienced women are more likely to be good at sex and less likely to have hangups, for starters. [NEWLINE] [NEWLINE] [STARTQ] This goes to the heart of my OP. I don't bring that level of "advice" to anyone for any preference they have. Sexual history alone seems to earn this "helpful" tip of blaming it on insecurities and needing fixing. [ENDQ] [NEWLINE] You almost admitted yourself.  You're worried<mask> you try something with your partner, it won't be novel enough for her, right?  You don't want to worry<mask> she's had that particular thing happen to her before, right? [NEWLINE] [NEWLINE] I mean,<mask> it's not about insecurity, then<mask> is it about?</s>
Label encoding: <s> [STARTQ] They are all equally valid. [ENDQ] [NEWLINE] Is this true for all examples?  For example, Joe only dates women who are under 400 pounds while Bob only dates women who idolize Hitler.  Are you really saying those two things are equal?  Or have you just arbitrarily decided that your particular values are equally valid to physical attractiveness?  And if so, isn't your whole argument circular? [NEWLINE] [NEWLINE] [STARTQ] I don't want to have to think of the dicks she's sucked on or the semen she's had to clean off her body from every other guy before me. The notion of a guy sucking on her body in the same places I do or intend to is wholly unappealing to me. [ENDQ] [NEWLINE] This doesn't explain your view at all.  A woman who has been with one guy 1000 times is probably more likely to have been touched everywhere or had a lot of experience with semen than a woman who had six short stints with six different guys.  Remember, your view is against the number sexual partners, not the total amount of sexual experience. [NEWLINE] [NEWLINE] [STARTQ] Now, if you can convince me that there is some unknown value there for me, I'd happily change my view. [ENDQ] [NEWLINE] Experienced women are more likely to be good at sex and less likely to have hangups, for starters. [NEWLINE] [NEWLINE] [STARTQ] This goes to the heart of my OP. I don't bring that level of "advice" to anyone for any preference they have. Sexual history alone seems to earn this "helpful" tip of blaming it on insecurities and needing fixing. [ENDQ] [NEWLINE] You almost admitted yourself.  You're worried if you try something with your partner, it won't be novel enough for her, right?  You don't want to worry if she's had that particular thing happen to her before, right? [NEWLINE] [NEWLINE] I mean, if it's not about insecurity, then what is it about?</s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The SRY system makes it nice and clean binary, sure,<mask> that doesn't mean it's the best way to determine sex.  The presence of at least one Y chromosome is clean and binary too,<mask> that doesn't mean *that's* the best way.  Even the article for SRY casts it<mask> part of a larger whole.  There's a section titled "Influence on sex" that says SRY "plays an important role in sex determination."  Biology is a complicated field, and<mask> the presence of an SRY gene is<mask> on/off (<mask> it can be on either an X or a Y chromosome), I don't think SRY is a terribly useful measure of sex in many circumstances. [NEWLINE] [NEWLINE] It all depends on<mask> you *need* to know the sex, and<mask> pointed out elsewhere there are a whole host of reasons.  For instance, many treatments and conditions depend on the hormonal profile, which are most certainly not always aligned with SRY state.  SRY-carrying individuals can still have ovarian tissue. [NEWLINE] [NEWLINE] There's no "issue" in saying that SRY is binary or even saying that SRY sex is binary,<mask> you tried to inject SRY into a larger discussion about sex and gender.  You can't pretend that SRY alone has any relevance outside of very specific medical questions.  That's not politics or policy, that's just acknowledging that the thing we lump into "male and female" is pretty complex, and in order to treat someone - that is, in order to practice and understand biology correctly - we have to understand the many factors that fall under that category. [NEWLINE] [NEWLINE] The treatment for an AIS female is going to be very different from that for a typical male, even thought they both have an active SRY gene.  The combinations are myriad.  Acknowledging those differences isn't politics, it's just good biology.</s>
Label encoding: <s>The SRY system makes it nice and clean binary, sure, but that doesn't mean it's the best way to determine sex.  The presence of at least one Y chromosome is clean and binary too, but that doesn't mean *that's* the best way.  Even the article for SRY casts it as part of a larger whole.  There's a section titled "Influence on sex" that says SRY "plays an important role in sex determination."  Biology is a complicated field, and while the presence of an SRY gene is indeed on/off ( although it can be on either an X or a Y chromosome), I don't think SRY is a terribly useful measure of sex in many circumstances. [NEWLINE] [NEWLINE] It all depends on why you *need* to know the sex, and as pointed out elsewhere there are a whole host of reasons.  For instance, many treatments and conditions depend on the hormonal profile, which are most certainly not always aligned with SRY state.  SRY-carrying individuals can still have ovarian tissue. [NEWLINE] [NEWLINE] There's no "issue" in saying that SRY is binary or even saying that SRY sex is binary, but you tried to inject SRY into a larger discussion about sex and gender.  You can't pretend that SRY alone has any relevance outside of very specific medical questions.  That's not politics or policy, that's just acknowledging that the thing we lump into "male and female" is pretty complex, and in order to treat someone - that is, in order to practice and understand biology correctly - we have to understand the many factors that fall under that category. [NEWLINE] [NEWLINE] The treatment for an AIS female is going to be very different from that for a typical male, even thought they both have an active SRY gene.  The combinations are myriad.  Acknowledging those differences isn't politics, it's just good biology.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're advocating "war" on personal choices. It's not up to you or anybody to decide<mask>'s right for someone. Many "healthy" looking people live unhealthy lifestyles in ways you can't imagine. You<mask> seem to believe that this pressure isn't already there? A much heavier focus on trying to be healthy?<mask> the hell does that even mean? <mask> do you care? It's<mask> funny to me that people pretend to have other's best interests at heart<mask> all they really want to to stop looking at things they deem unsightly. I'm basically positing that your motives are selfish, not magnanimous. Trying to control the way people live their lives is a losing battle.<mask> not become a doctor<mask> you're<mask> passionate about educating people about their health?<mask> about we make sure everyone has a roof over their head before we start to focus on those who choose to let their bodies degrade? Ever hear of mental health problems? Those are much more prevalent and [NEWLINE] important to fix. Hell, most binge eating is 90 percent mental anyway. [NEWLINE] [NEWLINE] <mask><mask> you're<mask> totally ignoring the socio-economic factors at play in Mexico and America that have lead to a rise in obese people. It's easy to say we should "pressure" the family of five that can only afford the cheapest, least nutritional, foodstuffs.<mask> it starts to fall apart<mask> you look at the realities... Do you fine people for being overweight? Pay more for healthcare? Send them to jail? Slippery Slope much? Shouldn't we be taxing or regulating the sugar industry instead of decrying its' victims? [NEWLINE] [NEWLINE] Honestly, there's<mask> many issues I have with this viewpoint that I can't even keep my counter points straight. You're trying to make this a binary issue of fat=unhealthy=bad vs. skinny=healthy=good. And it's just<mask> much more complex then that. </s>
Label encoding: <s>You're advocating "war" on personal choices. It's not up to you or anybody to decide what's right for someone. Many "healthy" looking people live unhealthy lifestyles in ways you can't imagine. You also seem to believe that this pressure isn't already there? A much heavier focus on trying to be healthy? What the hell does that even mean?  Why do you care? It's so funny to me that people pretend to have other's best interests at heart when all they really want to to stop looking at things they deem unsightly. I'm basically positing that your motives are selfish, not magnanimous. Trying to control the way people live their lives is a losing battle. Why not become a doctor if you're so passionate about educating people about their health? How about we make sure everyone has a roof over their head before we start to focus on those who choose to let their bodies degrade? Ever hear of mental health problems? Those are much more prevalent and [NEWLINE] important to fix. Hell, most binge eating is 90 percent mental anyway. [NEWLINE] [NEWLINE] I think you're also totally ignoring the socio-economic factors at play in Mexico and America that have lead to a rise in obese people. It's easy to say we should "pressure" the family of five that can only afford the cheapest, least nutritional, foodstuffs. But it starts to fall apart when you look at the realities... Do you fine people for being overweight? Pay more for healthcare? Send them to jail? Slippery Slope much? Shouldn't we be taxing or regulating the sugar industry instead of decrying its' victims? [NEWLINE] [NEWLINE] Honestly, there's so many issues I have with this viewpoint that I can't even keep my counter points straight. You're trying to make this a binary issue of fat=unhealthy=bad vs. skinny=healthy=good. And it's just so much more complex then that. </s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> it actually does is put a lid on property values<mask> assessed for tax purposes. You have a house that has a Tax Assessment of $250k,<mask> is listed for $500k?<mask> it sells, you get $500k (less various fees). [ENDQ] [NEWLINE] The listing for $500k will reflect the new taxes the owner will have to pay. <mask> the new owner were to keep the 250k tax assessment, they'd be willing to pay more, up to the net present value of the difference in property tax. [NEWLINE] [NEWLINE] <mask> a 250k assessment difference translates to $5000 a year in property tax, that will have a NPV in the neighborhood of $100,000. <mask> the seller loses an effective revenue stream worth $100,000 by selling. <mask> the new owner were to get the lower assessment, they'd be willing to pay $600k instead of $500k. [NEWLINE] [NEWLINE] [STARTQ] <mask> is it you claim that it screws people who want to move in or out? By subjecting them to market price tax assessment<mask> they buy a new place?<mask> would that change<mask> the people who don't move were subjected to the same market forces? [ENDQ] [NEWLINE] <mask><mask> the same total property tax would be assessed, it means that new owners pay a far higher share of total tax than they would in a system<mask> everyone pays the same percent of the house's value in property tax. [NEWLINE] [NEWLINE] To use the numbers from your example,<mask> you have 2 identical houses next door to each other, one assessed at 250k, one at 500k, the person in the 250k assessed house is only paying 1/3 their fair share, and the person in the 500k assessed house is paying 2/3 their fair share.  I would raise the tax on the 250k assessed house to cut it on the 500k assessed house,<mask> taxes should treat similarly disposed people and assets similarly.</s>
Label encoding: <s> [STARTQ] What it actually does is put a lid on property values as assessed for tax purposes. You have a house that has a Tax Assessment of $250k, but is listed for $500k? When it sells, you get $500k (less various fees). [ENDQ] [NEWLINE] The listing for $500k will reflect the new taxes the owner will have to pay.  If the new owner were to keep the 250k tax assessment, they'd be willing to pay more, up to the net present value of the difference in property tax. [NEWLINE] [NEWLINE] If a 250k assessment difference translates to $5000 a year in property tax, that will have a NPV in the neighborhood of $100,000.  So the seller loses an effective revenue stream worth $100,000 by selling.  If the new owner were to get the lower assessment, they'd be willing to pay $600k instead of $500k. [NEWLINE] [NEWLINE] [STARTQ] How is it you claim that it screws people who want to move in or out? By subjecting them to market price tax assessment when they buy a new place? How would that change if the people who don't move were subjected to the same market forces? [ENDQ] [NEWLINE] Assuming that the same total property tax would be assessed, it means that new owners pay a far higher share of total tax than they would in a system where everyone pays the same percent of the house's value in property tax. [NEWLINE] [NEWLINE] To use the numbers from your example, if you have 2 identical houses next door to each other, one assessed at 250k, one at 500k, the person in the 250k assessed house is only paying 1/3 their fair share, and the person in the 500k assessed house is paying 2/3 their fair share.  I would raise the tax on the 250k assessed house to cut it on the 500k assessed house, because taxes should treat similarly disposed people and assets similarly.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mean, of course they are discriminatory.  To discriminate just means to categorize and treat each category differently.  Everyone discriminates all the time, and it isn't a bad thing - I discriminate against restaurants that have shitty food, and I don't go back. [NEWLINE] [NEWLINE] Whether its *morally wrong* to discriminate in this instance is<mask> you need to be arguing.  And<mask><mask> there are several completely different questions here: [NEWLINE] [NEWLINE] 1. Is it morally wrong for the government to provide women only shelters, and not men only shelters (or at least enough gender-neutral shelters to accommodate)?  I would say yes - this is morally wrong and creates an unnecessary disadvantage, which the government has no business doing. [NEWLINE] [NEWLINE] 2. Is it morally wrong for private charity to provide women only shelters, and not men only shelters (or at least enough gender-neutral shelters to accommodate)?  I would say no - private charities have the right to spend money<mask> they want, and<mask> it<mask> happens that more charities want to spend money on women than on men, then<mask><mask> that's unfortunate<mask> understandable, and really the only response is to start new charities<mask> you feel strongly about it. [NEWLINE] [NEWLINE] 3. Is it morally wrong for private business to provide women only gyms, and not men only gyms?  I would say no - businesses have the right to cater to their chosen clientele, and<mask> it<mask> happens that women only gyms are profitable, and men only gyms are not, then more power to them who discover they can provide a desired service and make some money. <mask><mask><mask><mask>, it does seem to skirt rather closely to anti-discrimination laws (can you imagine a gym that only allows white customers?)<mask> gender has a weird relationship with such laws (obviously, gendered bathrooms and locker rooms and such things are still the norm, unlike any other protected class).</s>
Label encoding: <s>I mean, of course they are discriminatory.  To discriminate just means to categorize and treat each category differently.  Everyone discriminates all the time, and it isn't a bad thing - I discriminate against restaurants that have shitty food, and I don't go back. [NEWLINE] [NEWLINE] Whether its *morally wrong* to discriminate in this instance is what you need to be arguing.  And I think there are several completely different questions here: [NEWLINE] [NEWLINE] 1. Is it morally wrong for the government to provide women only shelters, and not men only shelters (or at least enough gender-neutral shelters to accommodate)?  I would say yes - this is morally wrong and creates an unnecessary disadvantage, which the government has no business doing. [NEWLINE] [NEWLINE] 2. Is it morally wrong for private charity to provide women only shelters, and not men only shelters (or at least enough gender-neutral shelters to accommodate)?  I would say no - private charities have the right to spend money how they want, and if it so happens that more charities want to spend money on women than on men, then I think that's unfortunate but understandable, and really the only response is to start new charities if you feel strongly about it. [NEWLINE] [NEWLINE] 3. Is it morally wrong for private business to provide women only gyms, and not men only gyms?  I would say no - businesses have the right to cater to their chosen clientele, and if it so happens that women only gyms are profitable, and men only gyms are not, then more power to them who discover they can provide a desired service and make some money.  On the other hand, it does seem to skirt rather closely to anti-discrimination laws (can you imagine a gym that only allows white customers?) but gender has a weird relationship with such laws (obviously, gendered bathrooms and locker rooms and such things are still the norm, unlike any other protected class).</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The study you linked still shows that men are treated the same<mask> women at best; and only for certain crimes (violent). There are still huge discrepancies - men being more than twice<mask> likely to be sentenced - for property and drug offences. It then goes on to say that men are sentenced to more than 4 more years for violent crimes<mask> less than a year for property and drug offences (<mask> still getting a harsher sentence). I don't see<mask>, in any way, this refutes the previous commenter's point. [NEWLINE] [NEWLINE] I<mask> find it comical that you blame bad statistics,<mask> the study you linked has one of the most selective datasets I've ever seen. It only uses data from selected urban areas in Texas for the year 1991. Here's [a study]( [URL].cfm?abstract_id=2144002) that used Federal crime statistics from 4 bureaus from the years 2001-2008. It's findings speak for themselves. [NEWLINE] [NEWLINE] <mask> for the second point, I don't see<mask> just<mask> a feminist says that Men and Women *should* be treated equally changes the fact that they *aren't*. [NEWLINE] [NEWLINE] [STARTQ] Women are 'weak' and 'fragile', are told it's "their fault for not protecting themselves"? [ENDQ] [NEWLINE] I literally laughed out loud<mask> I read that;<mask> the opposite is true. I really don't see your logic here: Women are supposedly weak,<mask> they should be able to protect themselves, and<mask> they can't it's there own fault. Like,<mask>?<mask> they are seen<mask> weak, shouldn't it *not* be their fault<mask> they can't defend themselves? I've never heard anyone say<mask> you are purporting; only the opposite (ie. there was nothing you could do). [NEWLINE] [NEWLINE] <mask> they way you rationalize male sexual assault being not<mask> bad<mask> female sexual assault makes absolutely no sense. </s>
Label encoding: <s>The study you linked still shows that men are treated the same as women at best; and only for certain crimes (violent). There are still huge discrepancies - men being more than twice as likely to be sentenced - for property and drug offences. It then goes on to say that men are sentenced to more than 4 more years for violent crimes but less than a year for property and drug offences ( though still getting a harsher sentence). I don't see how, in any way, this refutes the previous commenter's point. [NEWLINE] [NEWLINE] I also find it comical that you blame bad statistics, when the study you linked has one of the most selective datasets I've ever seen. It only uses data from selected urban areas in Texas for the year 1991. Here's [a study]( [URL].cfm?abstract_id=2144002) that used Federal crime statistics from 4 bureaus from the years 2001-2008. It's findings speak for themselves. [NEWLINE] [NEWLINE] As for the second point, I don't see how just because a feminist says that Men and Women *should* be treated equally changes the fact that they *aren't*. [NEWLINE] [NEWLINE] [STARTQ] Women are 'weak' and 'fragile', are told it's "their fault for not protecting themselves"? [ENDQ] [NEWLINE] I literally laughed out loud when I read that; since the opposite is true. I really don't see your logic here: Women are supposedly weak, so they should be able to protect themselves, and if they can't it's there own fault. Like, what? If they are seen as weak, shouldn't it *not* be their fault if they can't defend themselves? I've never heard anyone say what you are purporting; only the opposite (ie. there was nothing you could do). [NEWLINE] [NEWLINE] Also they way you rationalize male sexual assault being not as bad as female sexual assault makes absolutely no sense. </s>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't agree with any of this, I'm sorry.<mask><mask> you're presuming OP's intent based on<mask> they're saying -- which is,<mask> I've repeatedly argued, the fundamental blinding fallacy of PC notions. The PC mindset mistakes the map for the terrain, and cannot see past its own singular interpretation (more like presumption) of their meaning to the *real* meaning and intent of the speaker -- which I maintain is the only thing that really matters. [NEWLINE] [NEWLINE] It reminds of the habit of wasps in a field to attack the tractor instead of the farmer,<mask> the tractor is the hot thing making the noise;<mask> the tractor is only an inanimate tool, and in any case no harm was consciously intended to begin with. And even<mask> attacking the tractor was somehow successful, it would not prevent the land from being worked, only that particular tool from being used to do it on that particular day by that particular person. PC addresses symptoms, rather than causes, and at least half the time doesn't even recognise symptoms correctly. [NEWLINE] [NEWLINE] The example you provide is not a restraint on expression. I can choose whether to express myself in a manner that someone else might find inappropriate or not, and no one can stop me from doing that. They can try to punish me for it after the fact,<mask> the closest thing PC offers to restraint is that threat, not any *actual* restraint. More, who gets to decide for the rest of us<mask> is offensive and<mask> is not?<mask><mask> I happen to disagree? Do we vote on it, or<mask>? [NEWLINE] [NEWLINE] <mask><mask> we just tried a different tactic, such<mask> approaching the underlying reasons *<mask> * offensive expression occurs. (Or, I would suggest, the that these concepts are inherently subjective to begin with, and can ever be anything else -- another primal fallacy of the PC strategy that is<mask> doomed to failure for that reason.) [NEWLINE] </s>
Label encoding: <s>I don't agree with any of this, I'm sorry. I think you're presuming OP's intent based on what they're saying -- which is, as I've repeatedly argued, the fundamental blinding fallacy of PC notions. The PC mindset mistakes the map for the terrain, and cannot see past its own singular interpretation (more like presumption) of their meaning to the *real* meaning and intent of the speaker -- which I maintain is the only thing that really matters. [NEWLINE] [NEWLINE] It reminds of the habit of wasps in a field to attack the tractor instead of the farmer, because the tractor is the hot thing making the noise; but the tractor is only an inanimate tool, and in any case no harm was consciously intended to begin with. And even if attacking the tractor was somehow successful, it would not prevent the land from being worked, only that particular tool from being used to do it on that particular day by that particular person. PC addresses symptoms, rather than causes, and at least half the time doesn't even recognise symptoms correctly. [NEWLINE] [NEWLINE] The example you provide is not a restraint on expression. I can choose whether to express myself in a manner that someone else might find inappropriate or not, and no one can stop me from doing that. They can try to punish me for it after the fact, but the closest thing PC offers to restraint is that threat, not any *actual* restraint. More, who gets to decide for the rest of us what is offensive and what is not? What if I happen to disagree? Do we vote on it, or what? [NEWLINE] [NEWLINE] What if we just tried a different tactic, such as approaching the underlying reasons * why * offensive expression occurs. (Or, I would suggest, the that these concepts are inherently subjective to begin with, and can ever be anything else -- another primal fallacy of the PC strategy that is also doomed to failure for that reason.) [NEWLINE] </s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can be a Christian parent, take your kid to church every week, and still not force it on them and make them actively identify<mask> "Christian" before they are able to make that choice on their own. The two do not go hand in hand. I am a Christian; my husband is an atheist. The kids come to church with me every Sunday; they sit and watch things like Cosmos with my husband. We decided that we don't want to play tug-of-war with the kids, and neither of us wants to push our beliefs/lack of belief on the kids and make them feel pressured to adapt them<mask> their own (or pretend to do<mask> in an effort to avoid punishment or conflict). We have all kinds of discussions with our kids, not only about Christianity and not believing in any god,<mask><mask> about<mask> other people believe about other gods. We encourage our kids to ask the hard questions, rather than shying away and denying those natural doubts and curiosity, and think things through for themselves, rather than just telling them<mask> to think and believe. For now, they are pondering those hard questions. Both of them will tell you that they believe in God,<mask> for some reason (and<mask><mask> I know<mask> this seed was planted...) they both struggle with the idea that you can believe in God and still acknowledge the scientific facts of evolution, cosmology, deep time, etc., and that you don't have to pick and choose which one to believe. [NEWLINE] [NEWLINE] My point is, taking your kids to church does not mean you stuff your faith down their throats until they drown in it. [NEWLINE] [NEWLINE] Do you think that atheist parents are wrong to raise their children in *their* worldview? Do you think they should make sure to educate their kids about all the world's religions too and let them choose? Or is this a standard to which you only hold the religious?</s>
Label encoding: <s>You can be a Christian parent, take your kid to church every week, and still not force it on them and make them actively identify as "Christian" before they are able to make that choice on their own. The two do not go hand in hand. I am a Christian; my husband is an atheist. The kids come to church with me every Sunday; they sit and watch things like Cosmos with my husband. We decided that we don't want to play tug-of-war with the kids, and neither of us wants to push our beliefs/lack of belief on the kids and make them feel pressured to adapt them as their own (or pretend to do so in an effort to avoid punishment or conflict). We have all kinds of discussions with our kids, not only about Christianity and not believing in any god, but also about what other people believe about other gods. We encourage our kids to ask the hard questions, rather than shying away and denying those natural doubts and curiosity, and think things through for themselves, rather than just telling them what to think and believe. For now, they are pondering those hard questions. Both of them will tell you that they believe in God, but for some reason (and I think I know where this seed was planted...) they both struggle with the idea that you can believe in God and still acknowledge the scientific facts of evolution, cosmology, deep time, etc., and that you don't have to pick and choose which one to believe. [NEWLINE] [NEWLINE] My point is, taking your kids to church does not mean you stuff your faith down their throats until they drown in it. [NEWLINE] [NEWLINE] Do you think that atheist parents are wrong to raise their children in *their* worldview? Do you think they should make sure to educate their kids about all the world's religions too and let them choose? Or is this a standard to which you only hold the religious?</s>
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Perhaps this is your experience,<mask> there are many people who do deny this very fact. I have even met liberal open minded types who deny this<mask> they have never experienced it first hand. It is a very hard thing to understand<mask> you have had little interaction with minorities. [ENDQ] [NEWLINE] One of those liberals here. Let me explain: Race is weird. Like, really, really weird<mask> a concept. It doesn't reflect a scientific reality and,<mask><mask> the phrase is massively overused, it *is* a perfect example of a social construct.<mask><mask><mask><mask>, humans are complex. You are the incredibly complex and unique sum of your genes (which don't correlate well at all with our concept of race) and years and years of  unique experiences. <mask> we attempt to reduce the incredible complexity of the human experience to just a single metric that can't even be scientifically measured, we run into problems. Reductionist thinking can be useful,<mask> *that* sort of reductionism is incredibly dehumanizing. [NEWLINE] [NEWLINE] <mask> groups in America, white people on average have it better off than black people (<mask> not East Asians, etc.). This is undeniable.<mask> many go beyond that statement and attempt to claim that white *people* have it better off. This does not follow<mask> it ignores the individual experiences of real humans.<mask> you tell a *white person* they have it better than a black person simply<mask> of skin color, the white person is justified in taking offense. They do<mask><mask> you are dismissing their humanity and seeing them<mask> nothing more than their race. [NEWLINE] [NEWLINE] Which is racist to the extreme. [NEWLINE] [NEWLINE] Related anecdote: My white father-in-law escaped a death camp (the rest of his family perished) just to find himself<mask> a penniless refugee in the barren wasteland of Winnipeg. I pity the fool who educates him about his privilege.  </s>
Label encoding: <s> [STARTQ] Perhaps this is your experience, however there are many people who do deny this very fact. I have even met liberal open minded types who deny this because they have never experienced it first hand. It is a very hard thing to understand if you have had little interaction with minorities. [ENDQ] [NEWLINE] One of those liberals here. Let me explain: Race is weird. Like, really, really weird as a concept. It doesn't reflect a scientific reality and, even though the phrase is massively overused, it *is* a perfect example of a social construct. On the other hand, humans are complex. You are the incredibly complex and unique sum of your genes (which don't correlate well at all with our concept of race) and years and years of  unique experiences.  When we attempt to reduce the incredible complexity of the human experience to just a single metric that can't even be scientifically measured, we run into problems. Reductionist thinking can be useful, but *that* sort of reductionism is incredibly dehumanizing. [NEWLINE] [NEWLINE] As groups in America, white people on average have it better off than black people ( although not East Asians, etc.). This is undeniable. But many go beyond that statement and attempt to claim that white *people* have it better off. This does not follow as it ignores the individual experiences of real humans. When you tell a *white person* they have it better than a black person simply because of skin color, the white person is justified in taking offense. They do so because you are dismissing their humanity and seeing them as nothing more than their race. [NEWLINE] [NEWLINE] Which is racist to the extreme. [NEWLINE] [NEWLINE] Related anecdote: My white father-in-law escaped a death camp (the rest of his family perished) just to find himself as a penniless refugee in the barren wasteland of Winnipeg. I pity the fool who educates him about his privilege.  </s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe that birth control,<mask> generally a quality of life improvement, is not a "basic human right"<mask> many people would argue. The largest argument I have heard is one of women's rights, in that women should have control over their own reproductive choices, and birth control is a means to that end. I believe that the basic human right associated with reproductive choice is the right to refuse sexual intercourse. People are entirely free to not have children, and that freedom should be protected<mask> a human right.<mask>, I don't believe that people are entitled to having sex without consequences. There are many things in this world that are quality of life improvements,<mask> that doesn't necesitate that they be provided free of charge and considered fundamental to decent human existance.<mask> an aside, I understand that birth control pills are often prescribed<mask> a treatment for certain conditions. I would consider that<mask> medicine and a basic human right. I am speaking only to birth control to be used for the purpose of recreational sex.<mask> there you have it, please CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I believe that birth control, while generally a quality of life improvement, is not a "basic human right" as many people would argue. The largest argument I have heard is one of women's rights, in that women should have control over their own reproductive choices, and birth control is a means to that end. I believe that the basic human right associated with reproductive choice is the right to refuse sexual intercourse. People are entirely free to not have children, and that freedom should be protected as a human right. However, I don't believe that people are entitled to having sex without consequences. There are many things in this world that are quality of life improvements, but that doesn't necesitate that they be provided free of charge and considered fundamental to decent human existance. As an aside, I understand that birth control pills are often prescribed as a treatment for certain conditions. I would consider that as medicine and a basic human right. I am speaking only to birth control to be used for the purpose of recreational sex. So there you have it, please CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Scientists in the past did not have the benefit of our current understanding during their lifetimes. [ENDQ] [NEWLINE] This is a good point,<mask> then again, did they have any evidence for<mask> they *did* believe? [NEWLINE] [NEWLINE] [STARTQ] Subjectivity is allowed. Personal tastes are not the same<mask> saying a sky fairy exists at all. [ENDQ] [NEWLINE] <mask><mask> the "sky fairy" that they happen to believe in happens to conform with<mask> is currently understood in the scientific community? THen it would appear to be an issue that they believe something *without* evidence,<mask> opposed to *<mask><mask><mask> * evidence.<mask> the question is, is believing something without evidence,<mask> not<mask><mask><mask> it, acceptable under your dogma?<mask> you think of "religion," you're obviously conjuring up images of common religions like Christianity,<mask><mask><mask> a religion provided a scientific explanation for everything known by science,<mask> a religious explanation for everything else? Most Jews accept science<mask> truth, and understand much of their religious texts to have been written figuratively. Can they be true scientists? [NEWLINE] [NEWLINE] Darwin was still religious, even<mask> he contradicted the account of Genesis. In his mind, Christianity enveloped the theory of evolution and natural selection. He was right on the border that you described; he knew some things said by the Bible were not true,<mask> he clung to much of the rest of his religion. Was he not a scientist? [NEWLINE] [NEWLINE] [STARTQ] No, I'm saying religious scientists are hypocrites. [ENDQ] [NEWLINE] All religious scientists of all religions are hypocrites?<mask><mask> I were a scientist who believed in a goddess who simply "ignited" the big bang by establishing the laws of the universe and then disappeared without a trace from our reality. Is there evidence against her existence? No. Is there evidence *for* her existence? Not really.<mask><mask> do I stand?</s><pad>
Label encoding: <s> [STARTQ] Scientists in the past did not have the benefit of our current understanding during their lifetimes. [ENDQ] [NEWLINE] This is a good point, but then again, did they have any evidence for what they *did* believe? [NEWLINE] [NEWLINE] [STARTQ] Subjectivity is allowed. Personal tastes are not the same as saying a sky fairy exists at all. [ENDQ] [NEWLINE] What if the "sky fairy" that they happen to believe in happens to conform with what is currently understood in the scientific community? THen it would appear to be an issue that they believe something *without* evidence, as opposed to * in spite of * evidence. So the question is, is believing something without evidence, but not in spite of it, acceptable under your dogma? When you think of "religion," you're obviously conjuring up images of common religions like Christianity, but what if a religion provided a scientific explanation for everything known by science, but a religious explanation for everything else? Most Jews accept science as truth, and understand much of their religious texts to have been written figuratively. Can they be true scientists? [NEWLINE] [NEWLINE] Darwin was still religious, even if he contradicted the account of Genesis. In his mind, Christianity enveloped the theory of evolution and natural selection. He was right on the border that you described; he knew some things said by the Bible were not true, but he clung to much of the rest of his religion. Was he not a scientist? [NEWLINE] [NEWLINE] [STARTQ] No, I'm saying religious scientists are hypocrites. [ENDQ] [NEWLINE] All religious scientists of all religions are hypocrites? What if I were a scientist who believed in a goddess who simply "ignited" the big bang by establishing the laws of the universe and then disappeared without a trace from our reality. Is there evidence against her existence? No. Is there evidence *for* her existence? Not really. So where do I stand?</s><pad>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] For one, the poor countries countries most likely to be impacted by climate change<mask> have very little power to affect change on the issue. The combined action or inaction of the US, Europe, and BRIC is<mask> will shape the impact of climate change. [ENDQ] [NEWLINE] <mask><mask>,<mask> you are saying something different from OP. OP stated; [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask><mask> the number one concern for the human race<mask> a whole should be climate change, and I find it hard to justify any other causes [ENDQ] [NEWLINE] [NEWLINE] You want to make the argument that not everyone is *needed* to effect change.  I am making the argument that not everyone should be *expected* to effect change<mask> it does not meet their best interests.  The two are not mutually exclusive. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] And second, you seem to think we can't both combat climate change and world hunger. [ENDQ] [NEWLINE] I didn't say that we couldn't.  I used world hunger<mask> an argument to show that Climate change is not necessarily at the top of the list in everyone's eyes.  Specifically those who would die from hunger sooner than they would die from climate change. [NEWLINE] [NEWLINE] [STARTQ] <mask> to summarize, it's ridiculous to assert that concern about climate change on the part of citizens of wealthy countries is simply self-interest. [ENDQ] [NEWLINE] It is survival which is the purest form of self interest.  I don't see that<mask> ridiculous at all. <mask><mask>, I don't see<mask> it could be contended.  I am not saying that the wealthy nations should *not* fight to protect the world from climate change.  I am only saying that each person/group/nation will fight for the most pressing survival issues in the order with which they threaten *their* lives. [NEWLINE] [NEWLINE] <mask> you hated the military comparison, it sounds like you countered  reading or understanding it.</s>
Label encoding: <s> [STARTQ] For one, the poor countries countries most likely to be impacted by climate change also have very little power to affect change on the issue. The combined action or inaction of the US, Europe, and BRIC is what will shape the impact of climate change. [ENDQ] [NEWLINE] I agree, but you are saying something different from OP. OP stated; [NEWLINE] [NEWLINE] [STARTQ] Thus, I think the number one concern for the human race as a whole should be climate change, and I find it hard to justify any other causes [ENDQ] [NEWLINE] [NEWLINE] You want to make the argument that not everyone is *needed* to effect change.  I am making the argument that not everyone should be *expected* to effect change because it does not meet their best interests.  The two are not mutually exclusive. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] And second, you seem to think we can't both combat climate change and world hunger. [ENDQ] [NEWLINE] I didn't say that we couldn't.  I used world hunger as an argument to show that Climate change is not necessarily at the top of the list in everyone's eyes.  Specifically those who would die from hunger sooner than they would die from climate change. [NEWLINE] [NEWLINE] [STARTQ] So to summarize, it's ridiculous to assert that concern about climate change on the part of citizens of wealthy countries is simply self-interest. [ENDQ] [NEWLINE] It is survival which is the purest form of self interest.  I don't see that as ridiculous at all.  In fact, I don't see how it could be contended.  I am not saying that the wealthy nations should *not* fight to protect the world from climate change.  I am only saying that each person/group/nation will fight for the most pressing survival issues in the order with which they threaten *their* lives. [NEWLINE] [NEWLINE] While you hated the military comparison, it sounds like you countered  reading or understanding it.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0652, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I could write two dozen things in opposition to this,<mask> just to gather a few basic points. [NEWLINE] [NEWLINE] -You should understand,<mask> you don't(seems like you don't), that<mask> you're referring is [eugenics]( [URL] ), a field of thought that has existed for about<mask><mask><mask> Darwin's theory of evolution has has a long, *very* sordid history.  Most notably, it was one of the primary justifications for Nazi Germany's genocidal practices. [NEWLINE] [NEWLINE] -<mask> it's attractiveness to certain political persuasions, [overpopulation is simply not a concern with legitimate scientific backing]( [URL] ). <mask><mask>,<mask> the world continues to increase on its path of modernization, democratization and secularism, *under*population is the issue we will have to be concerned.  In highly educated, relatively prosperous societies, birth rates become lower and lower, to the point<mask> they are now beginning to become lower than death rates in some places.  In America,<mask> you excluded immigration, the population would not be growing.  In Japan, it's shrinking and aging outright.  It's only<mask> of the poorer places in the world that population continues to grow substantially, and even then, the worldwide birth rate is shrinking slowly every year. [NEWLINE] [NEWLINE] -There are any number of political oppositions I have to this<mask> I won't get to deep into them<mask> ultimately that's my ethical viewpoint vs. yours.  Needless to say,<mask><mask> measuring the value of a human being wholly by their ability to produce and work for their country is a fascist viewpoint, and the ability of an individual to pass any kind of mental or psychological test(which undoubtedly would be designed by rich white people) is strongly dependent on the situation and opportunities they grew up with,<mask> it doesn't even remotely approach an accurate genetic test.</s><pad>
Label encoding: <s>I could write two dozen things in opposition to this, but just to gather a few basic points. [NEWLINE] [NEWLINE] -You should understand, if you don't(seems like you don't), that what you're referring is [eugenics]( [URL] ), a field of thought that has existed for about as long as Darwin's theory of evolution has has a long, *very* sordid history.  Most notably, it was one of the primary justifications for Nazi Germany's genocidal practices. [NEWLINE] [NEWLINE] - Despite it's attractiveness to certain political persuasions, [overpopulation is simply not a concern with legitimate scientific backing]( [URL] ).  In fact, if the world continues to increase on its path of modernization, democratization and secularism, *under*population is the issue we will have to be concerned.  In highly educated, relatively prosperous societies, birth rates become lower and lower, to the point where they are now beginning to become lower than death rates in some places.  In America, if you excluded immigration, the population would not be growing.  In Japan, it's shrinking and aging outright.  It's only because of the poorer places in the world that population continues to grow substantially, and even then, the worldwide birth rate is shrinking slowly every year. [NEWLINE] [NEWLINE] -There are any number of political oppositions I have to this but I won't get to deep into them as ultimately that's my ethical viewpoint vs. yours.  Needless to say, I think measuring the value of a human being wholly by their ability to produce and work for their country is a fascist viewpoint, and the ability of an individual to pass any kind of mental or psychological test(which undoubtedly would be designed by rich white people) is strongly dependent on the situation and opportunities they grew up with, so it doesn't even remotely approach an accurate genetic test.</s><pad>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Historically, that's<mask> they used to do, sure. [NEWLINE] [NEWLINE] The (legitimate*) purpose of the military is and has always been to defend the public from foreign threats. Now, you're right that the most normal way to do that is to kill people and break things. Killing the damned Nazi that tried to take my house or breaking the tank he's driving is a perfectly reasonable way to protect people from a big foreign threat that manifests in a simple way, often in a uniform or within defined borders. [NEWLINE] <mask> those sorts of threats aren't very common any more, are they? Modern threats are a good deal less complicated. [NEWLINE] [NEWLINE] Take Vietnam^. The military tried their standard kill-people-break-stuff routine.<mask><mask>, they did it way better than they've ever done it before.<mask> 'd that go? Did it mitigate the supposed threat from the NVA? Nup. [NEWLINE] [NEWLINE] And most modern tasks for the defence force are like that. Aside from China, which is deterred by MAD, all of America's threats, from Al Qaeda to African civil war, are unconventional. That means the US military needs to adopt a more sophisticated approach than Mike Huckabee is capable of understanding with his pea brain. This might be something like counter-insurgency, perhaps. These would involve much more than mere violence. EG, in Afghanistan the strategy is clear and build - political science is just<mask> important<mask> violence! [NEWLINE] [NEWLINE] This is<mask> people make fun of Mike Huckabee: he's too stupid or dishonest to be capable of understanding complex things like basic strategy. [NEWLINE] [NEWLINE] *The military can<mask> be used for a bunch of other things, like invading and toppling democracies<mask> a welfare handout to American big business. [NEWLINE] [NEWLINE] ^Whether or not Vietnam was a real threat is immaterial. That's a decision for the politicians.</s>
Label encoding: <s>Historically, that's what they used to do, sure. [NEWLINE] [NEWLINE] The (legitimate*) purpose of the military is and has always been to defend the public from foreign threats. Now, you're right that the most normal way to do that is to kill people and break things. Killing the damned Nazi that tried to take my house or breaking the tank he's driving is a perfectly reasonable way to protect people from a big foreign threat that manifests in a simple way, often in a uniform or within defined borders. [NEWLINE] But those sorts of threats aren't very common any more, are they? Modern threats are a good deal less complicated. [NEWLINE] [NEWLINE] Take Vietnam^. The military tried their standard kill-people-break-stuff routine. In fact, they did it way better than they've ever done it before. How 'd that go? Did it mitigate the supposed threat from the NVA? Nup. [NEWLINE] [NEWLINE] And most modern tasks for the defence force are like that. Aside from China, which is deterred by MAD, all of America's threats, from Al Qaeda to African civil war, are unconventional. That means the US military needs to adopt a more sophisticated approach than Mike Huckabee is capable of understanding with his pea brain. This might be something like counter-insurgency, perhaps. These would involve much more than mere violence. EG, in Afghanistan the strategy is clear and build - political science is just as important as violence! [NEWLINE] [NEWLINE] This is why people make fun of Mike Huckabee: he's too stupid or dishonest to be capable of understanding complex things like basic strategy. [NEWLINE] [NEWLINE] *The military can also be used for a bunch of other things, like invading and toppling democracies as a welfare handout to American big business. [NEWLINE] [NEWLINE] ^Whether or not Vietnam was a real threat is immaterial. That's a decision for the politicians.</s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, given identical starting conditions there would be a 30-60% chance that rain would occur within the area covered by the prediction. It's not terribly uncommon for there to be a very high percentage chance for the whole area to be hit with rain<mask> for a small sub-area to have a very low chance. [NEWLINE] [NEWLINE] To complicate matters relationships never have the same starting conditions. A truly random pairing would result a 45% chance of cheating and a 4% chance of false paternity,<mask> the odds vary based on the traits of the people involved and the situations that occur. [NEWLINE] [NEWLINE] <mask> there are conditions or personality traits that pose a risk of infidelity or false paternity then you can be adults about it, have a discussion, and get the test without it harming the relationship unduly.<mask> those conditions are not present, then instead of there being a 45/4 it's something more akin to a 5/.05 likelihood then you already have functional certainty.<mask> get the expensive test?<mask> functionally accuse your partner of the decidedly unlikely case of false paternity? [NEWLINE] [NEWLINE] I would be rather upset<mask> someone accused me of cheating on my taxes, something that I don't do<mask> more than 4% of other people do. Does the fact that other people cheat on their taxes mean that I necessarily must cheat on mine? The accusation hurts<mask> I don't, and<mask><mask> I know it's the IRS' job that doesn't change the fact that I would be upset<mask> they were to audit me.<mask> that is true for me,<mask> could a woman not be upset of I accused her of something even more serious and even more unlikely based on *absolutely nothing* except that it has occurred somewhere to someone else entirely that one time? [NEWLINE] [NEWLINE] I don't get the benefit in cases<mask> the conditions that correlate to the low end of the bell curve are present.</s>
Label encoding: <s>Yes, given identical starting conditions there would be a 30-60% chance that rain would occur within the area covered by the prediction. It's not terribly uncommon for there to be a very high percentage chance for the whole area to be hit with rain but for a small sub-area to have a very low chance. [NEWLINE] [NEWLINE] To complicate matters relationships never have the same starting conditions. A truly random pairing would result a 45% chance of cheating and a 4% chance of false paternity, but the odds vary based on the traits of the people involved and the situations that occur. [NEWLINE] [NEWLINE] If there are conditions or personality traits that pose a risk of infidelity or false paternity then you can be adults about it, have a discussion, and get the test without it harming the relationship unduly. If those conditions are not present, then instead of there being a 45/4 it's something more akin to a 5/.05 likelihood then you already have functional certainty. Why get the expensive test? Why functionally accuse your partner of the decidedly unlikely case of false paternity? [NEWLINE] [NEWLINE] I would be rather upset if someone accused me of cheating on my taxes, something that I don't do but more than 4% of other people do. Does the fact that other people cheat on their taxes mean that I necessarily must cheat on mine? The accusation hurts because I don't, and even though I know it's the IRS' job that doesn't change the fact that I would be upset if they were to audit me. If that is true for me, how could a woman not be upset of I accused her of something even more serious and even more unlikely based on *absolutely nothing* except that it has occurred somewhere to someone else entirely that one time? [NEWLINE] [NEWLINE] I don't get the benefit in cases where the conditions that correlate to the low end of the bell curve are present.</s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The two are certainly related,<mask> I'm not saying Islamic terrorism and Radical Islam are the same thing. Radical Islam is a term that describes the social, cultural, and political movement towards extreme interpretations of Islam. Saying Radical Islam doesn't exist is like saying "Fundamentalist Christians" don't exist.<mask> a Christian extremist bombed an abortion center, it would make sense to discuss the fundamentalist culture that their beliefs originated from to see<mask> it could have shaped their behavior. Further,<mask> we discovered that a significant number of extremist Christians *supported* the bombing, and some actively provided money and political support to promote more bombings, it would be foolish to write off this trend<mask> "they're just a bunch of isolated nutjobs." Are they religious fanatics? Sure.<mask> that doesn't mean they don't represent a significant cultural movement. [NEWLINE] [NEWLINE] [STARTQ] There is no such thing<mask> "Islamic Politics" either. Any more than there is such a thing<mask> "Christian Politics" or "Jewish Politics." [ENDQ] [NEWLINE] <mask> you haven't been paying attention to the republican party over the past 20 years, you might have missed the fact that Christian politics play a **huge** role in America's political landscape. Conservative Christians are a hugely influential, highly motivated, and remarkably consistent block of voters. Issues like abortion, which almost completely rest on religious beliefs, have been a major divider between republicans and democrats.  Whether someone is talking about "liberal politics", "Islamic politics" or "gamer politics", it is referring to the political issues and beliefs commonly important within that cultural group. [NEWLINE] [NEWLINE] Cultural movements happen, and often they become intertwined with politics. I do not support the paranoia and thinly-veiled racism that many people have towards Muslims.<mask> Radical Islam *does* exist, and it has had tangible effects on the political landscape of the Middle East.</s>
Label encoding: <s>The two are certainly related, but I'm not saying Islamic terrorism and Radical Islam are the same thing. Radical Islam is a term that describes the social, cultural, and political movement towards extreme interpretations of Islam. Saying Radical Islam doesn't exist is like saying "Fundamentalist Christians" don't exist. If a Christian extremist bombed an abortion center, it would make sense to discuss the fundamentalist culture that their beliefs originated from to see how it could have shaped their behavior. Further, if we discovered that a significant number of extremist Christians *supported* the bombing, and some actively provided money and political support to promote more bombings, it would be foolish to write off this trend as "they're just a bunch of isolated nutjobs." Are they religious fanatics? Sure. But that doesn't mean they don't represent a significant cultural movement. [NEWLINE] [NEWLINE] [STARTQ] There is no such thing as "Islamic Politics" either. Any more than there is such a thing as "Christian Politics" or "Jewish Politics." [ENDQ] [NEWLINE] If you haven't been paying attention to the republican party over the past 20 years, you might have missed the fact that Christian politics play a **huge** role in America's political landscape. Conservative Christians are a hugely influential, highly motivated, and remarkably consistent block of voters. Issues like abortion, which almost completely rest on religious beliefs, have been a major divider between republicans and democrats.  Whether someone is talking about "liberal politics", "Islamic politics" or "gamer politics", it is referring to the political issues and beliefs commonly important within that cultural group. [NEWLINE] [NEWLINE] Cultural movements happen, and often they become intertwined with politics. I do not support the paranoia and thinly-veiled racism that many people have towards Muslims. But Radical Islam *does* exist, and it has had tangible effects on the political landscape of the Middle East.</s>
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Toilets are private already (from<mask> I have seen),<mask> making whole bathrooms for every subgroup that cares about one would be difficult. I'm talking about bathrooms like we have in airports, gyms, schools and large restaurant. The discomfort of being on the toilet/seen washing yourself near someone of our own gender who may or may not be homosexual is not something ingrained to our society. Most people don't expect a stranger to be gay<mask> it is rather rare, relatively speaking. [NEWLINE] [NEWLINE] In a perfect world there would be private bathrooms everywhere for each who wants one.<mask> in our world, even shop owners who are pro LGBT rights will not be bothered constructing a whole third bathroom in a mainstream establishment for the off chance of it being used by more than one or two intersex persons a year. [NEWLINE] [NEWLINE] Real life example: My family owns a restaurant. It has two single-stall toilets connected with a narrow room with a single sink and mirror. Constructing a third toilet stall would be completely unfeasible<mask> we would have to pretty much expand our space. And even<mask> it were possible or even cheap it would never happen,<mask> in the 7+ years we have it I haven't noticed a single intersex or transgendered person. [NEWLINE] [NEWLINE] Now you could<mask><mask> we could scratch off the male/female signs from the toilet doors which, honestly, wouldn't make any difference to most people<mask> you don't normally share the rooms with a second person anyway. I myself have used both often enough and even customers sometimes choose to just go into the opposites sex toilet instead of waiting. [NEWLINE] [NEWLINE] <mask><mask> benefit would making both toilets permanently unisex give us, other than weirding out any conservative/small-minded/teenage customers we have? Even<mask> their reasons to be put off are reasons I don't understand.</s>
Label encoding: <s>Toilets are private already (from what I have seen), but making whole bathrooms for every subgroup that cares about one would be difficult. I'm talking about bathrooms like we have in airports, gyms, schools and large restaurant. The discomfort of being on the toilet/seen washing yourself near someone of our own gender who may or may not be homosexual is not something ingrained to our society. Most people don't expect a stranger to be gay because it is rather rare, relatively speaking. [NEWLINE] [NEWLINE] In a perfect world there would be private bathrooms everywhere for each who wants one. But in our world, even shop owners who are pro LGBT rights will not be bothered constructing a whole third bathroom in a mainstream establishment for the off chance of it being used by more than one or two intersex persons a year. [NEWLINE] [NEWLINE] Real life example: My family owns a restaurant. It has two single-stall toilets connected with a narrow room with a single sink and mirror. Constructing a third toilet stall would be completely unfeasible as we would have to pretty much expand our space. And even if it were possible or even cheap it would never happen, because in the 7+ years we have it I haven't noticed a single intersex or transgendered person. [NEWLINE] [NEWLINE] Now you could argue that we could scratch off the male/female signs from the toilet doors which, honestly, wouldn't make any difference to most people as you don't normally share the rooms with a second person anyway. I myself have used both often enough and even customers sometimes choose to just go into the opposites sex toilet instead of waiting. [NEWLINE] [NEWLINE] But what benefit would making both toilets permanently unisex give us, other than weirding out any conservative/small-minded/teenage customers we have? Even if their reasons to be put off are reasons I don't understand.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I smoke nearly every day, usually in the late evening. [NEWLINE] [NEWLINE] I<mask> hold a job, go to school, socialize with friends, and lead a pretty normal life. I eat well, exercise, and have a healthy relationship with my boyfriend, who<mask> smokes. [NEWLINE] [NEWLINE] For me, smoking is a great stress-reliever. I spend a lot of time worrying about the state of the world, and it really gets me down. Weed helps me to focus on the positive in my life, like my friends, boyfriend, job, studies, and future. It makes things fun that wouldn't normally be fun, like reading for class. [NEWLINE] [NEWLINE] I promise you, not all stoners are like<mask> you see on /r/trees (<mask> many of us are friendly to bits and pride ourselves upon that) or on movies like Harold and Kumar. [NEWLINE] [NEWLINE] An activity like smoking pot does not define me, just like a social drinker is not defined by<mask> they spend their free time. There *ARE* people who let it define them, like the stereotypical stoners that many complain about.<mask>, this happens with a lot of people and a lot of hobbies. I see a ton of people into things like DnD (which is fun, yeah!) or CoD who make it ALL THAT THEY ARE. I know, that can get annoying. Just like some people from the Internet who quote memes all day. [NEWLINE] [NEWLINE] My main point here is that you can partake in a hobby without letting it define you or take over your whole life. I guarantee you that you probably know more people who smoke pot than you think you do...they just don't broadcast it<mask> it's a hobby, not who they are. [NEWLINE] [NEWLINE] Hope that helps -<mask> you need me to clarify, feel free to ask. :) </s>
Label encoding: <s>I smoke nearly every day, usually in the late evening. [NEWLINE] [NEWLINE] I also hold a job, go to school, socialize with friends, and lead a pretty normal life. I eat well, exercise, and have a healthy relationship with my boyfriend, who also smokes. [NEWLINE] [NEWLINE] For me, smoking is a great stress-reliever. I spend a lot of time worrying about the state of the world, and it really gets me down. Weed helps me to focus on the positive in my life, like my friends, boyfriend, job, studies, and future. It makes things fun that wouldn't normally be fun, like reading for class. [NEWLINE] [NEWLINE] I promise you, not all stoners are like what you see on /r/trees ( although many of us are friendly to bits and pride ourselves upon that) or on movies like Harold and Kumar. [NEWLINE] [NEWLINE] An activity like smoking pot does not define me, just like a social drinker is not defined by how they spend their free time. There *ARE* people who let it define them, like the stereotypical stoners that many complain about. However, this happens with a lot of people and a lot of hobbies. I see a ton of people into things like DnD (which is fun, yeah!) or CoD who make it ALL THAT THEY ARE. I know, that can get annoying. Just like some people from the Internet who quote memes all day. [NEWLINE] [NEWLINE] My main point here is that you can partake in a hobby without letting it define you or take over your whole life. I guarantee you that you probably know more people who smoke pot than you think you do...they just don't broadcast it because it's a hobby, not who they are. [NEWLINE] [NEWLINE] Hope that helps - if you need me to clarify, feel free to ask. :) </s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Honestly I never aimed to be super fit. I wasn't on the hugely large side<mask> I was obese (about 16st, 5'7, male). I got tonsillitis and basically couldn't eat for a weekend except for stuff like ice cream and ended up losing about half a stone and thought well,<mask> not keep going? [NEWLINE] [NEWLINE] I absorbed<mask> much info about it<mask> I could<mask> I've literally never tried to lose weight and I've seen people fail at it<mask> I started tracking calories with myfitnesspal and slowly got down the scale. Some days it was really hard, other days just passed. Eventually I got to slap bang in the middle of my ideal BMI and I was happy. [NEWLINE] [NEWLINE] Thing is, I really was perfectly happy with that,<mask> part of me was curious with<mask> I could take it.<mask> I got some cheap weights and a squat rack, borrowed a bench from a friend and put it all in my garage. Once again I wanted to do it right<mask> I read lots of stuff and started doing some lifting. The thing is, it feels really good to do that. Like, seeing strength progression is immensely satisfying. [NEWLINE] [NEWLINE] Now I'm about 70kg and I'm close to<mask> I want to be. I might go another 10kg<mask> that's pretty much the limit, I don't want to be hulking or be a sculpted bro,<mask> every progression has been immensely satisfying, whether it was finding out I could comfortably do pull-ups or just being able to swing my kids around more or being middle of the pack in a run. [NEWLINE] [NEWLINE] Is my goal to be super fit and healthy? No, not at all. Getting fitter is just a result of lots of little things coming together and the satisfaction of achieving goals that you know were really hard work.</s>
Label encoding: <s>Honestly I never aimed to be super fit. I wasn't on the hugely large side but I was obese (about 16st, 5'7, male). I got tonsillitis and basically couldn't eat for a weekend except for stuff like ice cream and ended up losing about half a stone and thought well, why not keep going? [NEWLINE] [NEWLINE] I absorbed as much info about it as I could as I've literally never tried to lose weight and I've seen people fail at it so I started tracking calories with myfitnesspal and slowly got down the scale. Some days it was really hard, other days just passed. Eventually I got to slap bang in the middle of my ideal BMI and I was happy. [NEWLINE] [NEWLINE] Thing is, I really was perfectly happy with that, but part of me was curious with where I could take it. So I got some cheap weights and a squat rack, borrowed a bench from a friend and put it all in my garage. Once again I wanted to do it right so I read lots of stuff and started doing some lifting. The thing is, it feels really good to do that. Like, seeing strength progression is immensely satisfying. [NEWLINE] [NEWLINE] Now I'm about 70kg and I'm close to where I want to be. I might go another 10kg but that's pretty much the limit, I don't want to be hulking or be a sculpted bro, but every progression has been immensely satisfying, whether it was finding out I could comfortably do pull-ups or just being able to swing my kids around more or being middle of the pack in a run. [NEWLINE] [NEWLINE] Is my goal to be super fit and healthy? No, not at all. Getting fitter is just a result of lots of little things coming together and the satisfaction of achieving goals that you know were really hard work.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The only issue I see here is that you're talking about teaching something in a classroom setting which, I'm sure we can agree, is far more effectively learned through experience. [NEWLINE] [NEWLINE] Classrooms teach basic, objective facts<mask> they're the groundwork of greater knowledge.  Even the most savvy of Googlers needs to know the basis of<mask> they're researching before they can know<mask> direction to direct their prowess on the search engines. <mask><mask> that schools should be evolving more with the times and acknowledging the fact that being able to find something on the internet isn't "cheating,"<mask> rather wise resource management,<mask><mask><mask> that's a matter of simply adjusting the standards by which they evaluate students, rather than changing the subject matter entirely. [NEWLINE] [NEWLINE] A large part of school,<mask>, is attempting to provide the most objective picture of the world without editorializing, which can be tough on the internet.  Something like history class comes to mind.  School can teach basic facts and consequences that are going to be tough to find on the internet, knowing that you have a trustworthy source for your information. [NEWLINE] [NEWLINE] With something like math, it's incredibly important to understand<mask> things come out the way they do, rather than just being able to punch it into a graphing calculator and seeing<mask> comes out.  It greatly increases your grasp of the concept behind<mask> something integrates the way it does, which allows you to make better conclusions about it. [NEWLINE] [NEWLINE] I would amend the stance to say that schools need to accept the fact that there are new, valid tools to gaining knowledge that didn't used to exist, and simply integrate that into our expectations of students, rather than saying we need to actively attempt *teaching* them<mask> to use this technology.  The fact is that nothing can teach you those things better than just doing it.</s>
Label encoding: <s>The only issue I see here is that you're talking about teaching something in a classroom setting which, I'm sure we can agree, is far more effectively learned through experience. [NEWLINE] [NEWLINE] Classrooms teach basic, objective facts because they're the groundwork of greater knowledge.  Even the most savvy of Googlers needs to know the basis of what they're researching before they can know what direction to direct their prowess on the search engines.  I agree that schools should be evolving more with the times and acknowledging the fact that being able to find something on the internet isn't "cheating," but rather wise resource management, but I think that's a matter of simply adjusting the standards by which they evaluate students, rather than changing the subject matter entirely. [NEWLINE] [NEWLINE] A large part of school, also, is attempting to provide the most objective picture of the world without editorializing, which can be tough on the internet.  Something like history class comes to mind.  School can teach basic facts and consequences that are going to be tough to find on the internet, knowing that you have a trustworthy source for your information. [NEWLINE] [NEWLINE] With something like math, it's incredibly important to understand WHY things come out the way they do, rather than just being able to punch it into a graphing calculator and seeing what comes out.  It greatly increases your grasp of the concept behind why something integrates the way it does, which allows you to make better conclusions about it. [NEWLINE] [NEWLINE] I would amend the stance to say that schools need to accept the fact that there are new, valid tools to gaining knowledge that didn't used to exist, and simply integrate that into our expectations of students, rather than saying we need to actively attempt *teaching* them how to use this technology.  The fact is that nothing can teach you those things better than just doing it.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> that domestic violence against men is more widespread than statistics show or we would think,<mask> to nowhere near the extent to which domestic violence against women occurs. [ENDQ] [NEWLINE] You *think* this is true. Well... [URL] /~mfiebert/assault.htm [NEWLINE] [NEWLINE] Read through this too<mask> you have a strong stomach: [URL] [NEWLINE] [NEWLINE] [STARTQ] <mask>, that shouldn't stop us from trying to reduce it<mask> much<mask> we possibly can. [ENDQ] [NEWLINE] Certainly.<mask> good intentions aren't enough. You have to always keep checking to see whether your methods are effective or not. [NEWLINE] [NEWLINE] [STARTQ] For the problem of domestic violence against men, I'm not sure<mask> else could be done other than to try and lead men to feel less ashamed of it. [ENDQ] [NEWLINE] Awareness raising is a good first step, yes. I<mask> think there needs to be an equal number of men's DV shelters and overall equality in funding to male and female victims. [NEWLINE] [NEWLINE] [STARTQ] <mask> it were generally accepted that some women are buck strong [ENDQ] [NEWLINE] Not just that;<mask> you're holding a knife or a gun or a cast iron skillet, it doesn't matter<mask> you're physically smaller. [NEWLINE] [NEWLINE] [STARTQ] Until then, it's kind of a "shadow problem" that's hard to address<mask> nobody knows<mask> often it happens, much less<mask> and<mask>. [ENDQ] [NEWLINE] Actually, people do.<mask> this research is largely ignored. I highly suggest Googling "erin pizzey domestic violence" Erin Pizzey started the first women's domestic shelter in Britain and there's probably no one on the planet with more firsthand experience than her. Listen to her for a<mask>, you'll see that it's always been obvious that DV is a gender-symmetrical problem,<mask> society doesn't want to hear it.</s>
Label encoding: <s> [STARTQ] I think that domestic violence against men is more widespread than statistics show or we would think, but to nowhere near the extent to which domestic violence against women occurs. [ENDQ] [NEWLINE] You *think* this is true. Well... [URL] /~mfiebert/assault.htm [NEWLINE] [NEWLINE] Read through this too if you have a strong stomach: [URL] [NEWLINE] [NEWLINE] [STARTQ] However, that shouldn't stop us from trying to reduce it as much as we possibly can. [ENDQ] [NEWLINE] Certainly. But good intentions aren't enough. You have to always keep checking to see whether your methods are effective or not. [NEWLINE] [NEWLINE] [STARTQ] For the problem of domestic violence against men, I'm not sure what else could be done other than to try and lead men to feel less ashamed of it. [ENDQ] [NEWLINE] Awareness raising is a good first step, yes. I also think there needs to be an equal number of men's DV shelters and overall equality in funding to male and female victims. [NEWLINE] [NEWLINE] [STARTQ] If it were generally accepted that some women are buck strong [ENDQ] [NEWLINE] Not just that; if you're holding a knife or a gun or a cast iron skillet, it doesn't matter if you're physically smaller. [NEWLINE] [NEWLINE] [STARTQ] Until then, it's kind of a "shadow problem" that's hard to address because nobody knows how often it happens, much less why and where. [ENDQ] [NEWLINE] Actually, people do. But this research is largely ignored. I highly suggest Googling "erin pizzey domestic violence" Erin Pizzey started the first women's domestic shelter in Britain and there's probably no one on the planet with more firsthand experience than her. Listen to her for a while, you'll see that it's always been obvious that DV is a gender-symmetrical problem, but society doesn't want to hear it.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0767, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Recently there was a thread on /r/AskReddit about "reddit's low-point". In this thread, a lot of the most disgusting subreddits were listed. Subreddits in which things like misogyny, rape, racism, anti-semitism, holocaust-denial etc. were openly condoned, agreed to, and spread. [NEWLINE] [NEWLINE] I am not going to link to these subreddits here,<mask> for the record, i am not talking about subreddits that sometimes shred along the border of such things like /r/mensrights and /r/whiterights,<mask> I don't agree with the message of those,<mask><mask> they have a right to exist. [NEWLINE] [NEWLINE] I'm pretty sure<mask> reddit is a private website, they are in no way bound by law to preserve freedom of speech. [NEWLINE] [NEWLINE] I am from Germany,<mask> I am used to some things, mostly holocaust-denial, being excempt from the freedom of speech and it being illegal to voice those things in public, and I somehow don't see it<mask> a bad thing<mask><mask><mask> those exceptions are clearly limited to those things, which is the case in Germany at this moment. [NEWLINE] [NEWLINE] I know the biggest argument against this will be that<mask> reddit ever starts doing this, there is no way to make sure they will not spread their censorship to other things.<mask> I see that<mask> a valid concern, I really don't think it will be a problem: Reddit is a website that lives off the number of viewers, and they really will not risk losing users through censorship.<mask><mask><mask><mask> we still live in a free society<mask> alternatives to reddit can and will be created<mask> it stops being a free place, I really don't think we will have a problem with censorship on reddit. [NEWLINE] [NEWLINE] Change my view!</s>
Label encoding: <s>Recently there was a thread on /r/AskReddit about "reddit's low-point". In this thread, a lot of the most disgusting subreddits were listed. Subreddits in which things like misogyny, rape, racism, anti-semitism, holocaust-denial etc. were openly condoned, agreed to, and spread. [NEWLINE] [NEWLINE] I am not going to link to these subreddits here, but for the record, i am not talking about subreddits that sometimes shred along the border of such things like /r/mensrights and /r/whiterights, while I don't agree with the message of those, I think they have a right to exist. [NEWLINE] [NEWLINE] I'm pretty sure since reddit is a private website, they are in no way bound by law to preserve freedom of speech. [NEWLINE] [NEWLINE] I am from Germany, so I am used to some things, mostly holocaust-denial, being excempt from the freedom of speech and it being illegal to voice those things in public, and I somehow don't see it as a bad thing as long as those exceptions are clearly limited to those things, which is the case in Germany at this moment. [NEWLINE] [NEWLINE] I know the biggest argument against this will be that if reddit ever starts doing this, there is no way to make sure they will not spread their censorship to other things. While I see that as a valid concern, I really don't think it will be a problem: Reddit is a website that lives off the number of viewers, and they really will not risk losing users through censorship. So as long as we still live in a free society where alternatives to reddit can and will be created when it stops being a free place, I really don't think we will have a problem with censorship on reddit. [NEWLINE] [NEWLINE] Change my view!</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Religion was created to answer questions of the unknown, we now have other resources that can answer these questions for us. [ENDQ] [NEWLINE] Except there are certain religious concepts that are outside the realm of science. Can science prove or disprove the existence of a higher power? Can it prove or disprove the idea that everyone has a soul? Or the idea of an afterlife? [NEWLINE] [NEWLINE] [STARTQ] Religious books are outdated. [ENDQ] [NEWLINE] Religious books are nothing more than stories and guidelines with countless different interpretations which have changed over time. The interpretation of religious texts evolves<mask> time goes on,<mask> the overall lessons that they teach are timeless. [NEWLINE] [NEWLINE] [STARTQ] Religion has taught us to fear more than it has to love and has created utter hysteria on the topic of life and death. [ENDQ] [NEWLINE] Maybe based on your interpretation of religious texts,<mask> there are many common interpretations which teach love and compassion for others. You<mask> mentioned that religion teaches that the only two options for people after they die are eternal happiness or eternal damnation. This simply isn't true. Some religions teach reincarnation,<mask> other religions don't say much about<mask> happens after we die at all and focus on<mask> happens<mask> we are alive instead. [NEWLINE] [NEWLINE] [STARTQ] Religion separates humanity. [ENDQ] [NEWLINE] It can, just like anything else can separate humanity. Politics separates humanity. Money separates humanity.<mask> people live separates humanity.<mask> type of car people drive separates humanity. Anything can separate humanity. [NEWLINE] [NEWLINE] [STARTQ] Finally, people who cling to religion lack proper intellect and reasoning. [ENDQ] [NEWLINE] Human intellect has continued to grow throughout history<mask> the fact that religion has been a large part of society for<mask> long.<mask> religion is truly in the way of proper intellect and reasoning like you seem to think, then<mask> have we developed<mask> well to this point in history?</s>
Label encoding: <s> [STARTQ] Religion was created to answer questions of the unknown, we now have other resources that can answer these questions for us. [ENDQ] [NEWLINE] Except there are certain religious concepts that are outside the realm of science. Can science prove or disprove the existence of a higher power? Can it prove or disprove the idea that everyone has a soul? Or the idea of an afterlife? [NEWLINE] [NEWLINE] [STARTQ] Religious books are outdated. [ENDQ] [NEWLINE] Religious books are nothing more than stories and guidelines with countless different interpretations which have changed over time. The interpretation of religious texts evolves as time goes on, while the overall lessons that they teach are timeless. [NEWLINE] [NEWLINE] [STARTQ] Religion has taught us to fear more than it has to love and has created utter hysteria on the topic of life and death. [ENDQ] [NEWLINE] Maybe based on your interpretation of religious texts, but there are many common interpretations which teach love and compassion for others. You also mentioned that religion teaches that the only two options for people after they die are eternal happiness or eternal damnation. This simply isn't true. Some religions teach reincarnation, while other religions don't say much about what happens after we die at all and focus on what happens while we are alive instead. [NEWLINE] [NEWLINE] [STARTQ] Religion separates humanity. [ENDQ] [NEWLINE] It can, just like anything else can separate humanity. Politics separates humanity. Money separates humanity. Where people live separates humanity. What type of car people drive separates humanity. Anything can separate humanity. [NEWLINE] [NEWLINE] [STARTQ] Finally, people who cling to religion lack proper intellect and reasoning. [ENDQ] [NEWLINE] Human intellect has continued to grow throughout history despite the fact that religion has been a large part of society for so long. If religion is truly in the way of proper intellect and reasoning like you seem to think, then how have we developed so well to this point in history?</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you are over reacting. [NEWLINE] [NEWLINE] [STARTQ] You come here open to get your view change, not to see<mask> others feel the same. [ENDQ] [NEWLINE] Yes, I came here to see<mask> people had the same view<mask> the redditor I spoke of,<mask> I<mask> came to discuss the topic with people intelligently and see<mask> they could cmv. (and two people did change my view on the topic.) Am I not allowed to do both? I said "I post here to see<mask> others feel the same and<mask> their *arguments can persuade me* that I am responsible, partially or otherwise. [NEWLINE] [NEWLINE] [STARTQ] The op of this thread, contrary to the rules, did not counter your post. [ENDQ] [NEWLINE] Correct, I am<mask> sorry.<mask> are telling me? Tell him and/or report<mask> you want. [NEWLINE] [NEWLINE] [STARTQ] That's<mask> I reported the threader in order to get it deleted, and you should respect the establishment of CMV, for I fear that it will become a soap boxing thread. Already have I had reported way too many posts, only to see them deleted. [ENDQ] [NEWLINE] Thank you for your service. I respectfully request that you report and not bother me about it. There has been plenty of healthy debate in this thread. You took one line out of one post and used that to report that I do not respect this subreddit. You are wrong. [NEWLINE] [NEWLINE] Please remember the rules yourself: [NEWLINE] [NEWLINE] **Refrain from accusing OP or anyone else of being unwilling to change their view.** [NEWLINE] [NEWLINE] It may not relate to you perfectly,<mask> it's close enough.<mask> you don't like<mask> you see, report it and move on. You are distracting from the point of the thread yourself. Leave the moderating to the mods. Thank you and good night.</s>
Label encoding: <s>I think you are over reacting. [NEWLINE] [NEWLINE] [STARTQ] You come here open to get your view change, not to see if others feel the same. [ENDQ] [NEWLINE] Yes, I came here to see if people had the same view as the redditor I spoke of, but I also came to discuss the topic with people intelligently and see if they could cmv. (and two people did change my view on the topic.) Am I not allowed to do both? I said "I post here to see if others feel the same and if their *arguments can persuade me* that I am responsible, partially or otherwise. [NEWLINE] [NEWLINE] [STARTQ] The op of this thread, contrary to the rules, did not counter your post. [ENDQ] [NEWLINE] Correct, I am so sorry. Why are telling me? Tell him and/or report if you want. [NEWLINE] [NEWLINE] [STARTQ] That's why I reported the threader in order to get it deleted, and you should respect the establishment of CMV, for I fear that it will become a soap boxing thread. Already have I had reported way too many posts, only to see them deleted. [ENDQ] [NEWLINE] Thank you for your service. I respectfully request that you report and not bother me about it. There has been plenty of healthy debate in this thread. You took one line out of one post and used that to report that I do not respect this subreddit. You are wrong. [NEWLINE] [NEWLINE] Please remember the rules yourself: [NEWLINE] [NEWLINE] **Refrain from accusing OP or anyone else of being unwilling to change their view.** [NEWLINE] [NEWLINE] It may not relate to you perfectly, but it's close enough. If you don't like what you see, report it and move on. You are distracting from the point of the thread yourself. Leave the moderating to the mods. Thank you and good night.</s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The thing is, locking your door takes 2 seconds and is a very minor inconvenience in your life. Locking your door has zero effect on your life,<mask> it is pretty reasonable to question<mask> somebody wouldn't do it and call them an idiot<mask> they just couldn't be bothered.<mask>,<mask> women are asked to do to'make themselves safe' and avoid rape are much more of a big deal,<mask> it isn't a fair comparison. [NEWLINE] [NEWLINE] Using one of the most popular examples in this thread of not walking alone after dark, can you imagine<mask> difficult that would make your life<mask> you made that a blanket rule?<mask> I live, it gets dark at about 4:30/5pm in the winter.<mask><mask>, I just stay inside after that? There goes my social life, apart from having people over (<mask> I guess I shouldn't do that either,<mask> statistically most rapes are committed by people known to the victim). In the evenings I often pop over to see my friend for a cup of tea and a chat, and walk the 20 minutes home. That's definitely enough time to get attacked,<mask> should I just stop visiting my friend? [NEWLINE] [NEWLINE] <mask> about work - I don't finish till after 5 and my commute is about an hour<mask> it would definitely be dark by the time I'm walking home from the station. I can't afford a taxi every day, and it's a fairly short distance<mask> it would be kind of stupid to drive anyway.<mask> I get attacked then, am I an idiot for not taking better precautions? [NEWLINE] [NEWLINE] I guess<mask> i'm saying is that it isn't helpful to compare theft and rape. Theft can be prevented with easy and reasonable means. The things women are expected to do which will supposedly prevent rape are not reasonable.  </s>
Label encoding: <s>The thing is, locking your door takes 2 seconds and is a very minor inconvenience in your life. Locking your door has zero effect on your life, so it is pretty reasonable to question why somebody wouldn't do it and call them an idiot if they just couldn't be bothered. However, what women are asked to do to'make themselves safe' and avoid rape are much more of a big deal, so it isn't a fair comparison. [NEWLINE] [NEWLINE] Using one of the most popular examples in this thread of not walking alone after dark, can you imagine how difficult that would make your life if you made that a blanket rule? Where I live, it gets dark at about 4:30/5pm in the winter. So what, I just stay inside after that? There goes my social life, apart from having people over ( but I guess I shouldn't do that either, because statistically most rapes are committed by people known to the victim). In the evenings I often pop over to see my friend for a cup of tea and a chat, and walk the 20 minutes home. That's definitely enough time to get attacked, so should I just stop visiting my friend? [NEWLINE] [NEWLINE] What about work - I don't finish till after 5 and my commute is about an hour so it would definitely be dark by the time I'm walking home from the station. I can't afford a taxi every day, and it's a fairly short distance so it would be kind of stupid to drive anyway. If I get attacked then, am I an idiot for not taking better precautions? [NEWLINE] [NEWLINE] I guess what i'm saying is that it isn't helpful to compare theft and rape. Theft can be prevented with easy and reasonable means. The things women are expected to do which will supposedly prevent rape are not reasonable.  </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're looking at the complaints (<mask> not all of them) from the wrong approach. [NEWLINE] [NEWLINE] The fact that people feel backed into a one or the other, lesser of two evils choice every four years is enough reason for disdain<mask> is,<mask> it is only compounded by a trend that seems to grow every election; executive or legislative. [NEWLINE] [NEWLINE] People are voting for a President or representative and expect them to be their voice in steering the country in the direction the people want.  These people get elected,<mask>, and do<mask> **they** see<mask> right.  People are complaining about that, not<mask> much the man himself. [NEWLINE] [NEWLINE] I never understood<mask> politicians told me<mask> they felt about certain issues.  I don't give a shit<mask> they feel, I want someone who will vote, or in the President's case veto/pass,<mask> the majority wants. [NEWLINE] [NEWLINE] <mask> they rarely do and people complain.  Is the complaining pointless? No. It gets attention. With enough attention change may come about, whether it leads to an internal reform of<mask> things are done, people voting a certain/different way or an all out revolution. [NEWLINE] [NEWLINE] Now to address specific complaints toward Obama and the desire to remove him, this goes back to the idea that<mask><mask> he was elected, he's not living up to the expectations of those who elected him. <mask> I had to buy a car and chose the 'best of<mask> was offered' and ended up with a car that didn't run at all after a week, I'd want a new car.  Just<mask> the salesman said it was great doesn't mean it was or that I should be obligated to keep it. [NEWLINE] [NEWLINE] tl;dr Complaining isn't pointless, it raises awareness and may lead to change.</s>
Label encoding: <s>I think you're looking at the complaints ( though not all of them) from the wrong approach. [NEWLINE] [NEWLINE] The fact that people feel backed into a one or the other, lesser of two evils choice every four years is enough reason for disdain as is, but it is only compounded by a trend that seems to grow every election; executive or legislative. [NEWLINE] [NEWLINE] People are voting for a President or representative and expect them to be their voice in steering the country in the direction the people want.  These people get elected, however, and do what **they** see as right.  People are complaining about that, not so much the man himself. [NEWLINE] [NEWLINE] I never understood why politicians told me how they felt about certain issues.  I don't give a shit how they feel, I want someone who will vote, or in the President's case veto/pass, what the majority wants. [NEWLINE] [NEWLINE] But they rarely do and people complain.  Is the complaining pointless? No. It gets attention. With enough attention change may come about, whether it leads to an internal reform of how things are done, people voting a certain/different way or an all out revolution. [NEWLINE] [NEWLINE] Now to address specific complaints toward Obama and the desire to remove him, this goes back to the idea that even though he was elected, he's not living up to the expectations of those who elected him.  If I had to buy a car and chose the 'best of what was offered' and ended up with a car that didn't run at all after a week, I'd want a new car.  Just because the salesman said it was great doesn't mean it was or that I should be obligated to keep it. [NEWLINE] [NEWLINE] tl;dr Complaining isn't pointless, it raises awareness and may lead to change.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As sexual activity is not inherently harmful to children, there is no basis on which to label all sexual relations between children and adults<mask> abusive. [NEWLINE] [NEWLINE] [STARTQ] A Dutch study published in 1987 found that a sample of boys in paedophilic relationships felt positively about them. And a major<mask> still controversial 1998-2000 meta-study suggests –<mask> J Michael Bailey of Northwestern University, Chicago, says – that such relationships, entered into voluntarily, are "nearly uncorrelated with undesirable outcomes". [ENDQ] [NEWLINE] [STARTQ] Most people find that idea impossible.<mask> writing last year in the peer-reviewed Archives of Sexual Behaviour, Bailey said that<mask> he<mask> found the notion "disturbing", he was forced to recognise that **"persuasive evidence for the harmfulness of paedophilic relationships does not<mask> exist".** [ENDQ] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] A substantial number of people who<mask> children had sex with adults feel positively about the experience, and do not regard it<mask> abusive in any capacity. [NEWLINE] [NEWLINE] Long-Range Effects of Child and Adolescent Sexual Experiences Positive Review", Allie C. Kilpatrick. [NEWLINE] [NEWLINE] &gt;This book will be disturbing to many readers. The assumption that all children are "damaged" by their experiences is challenged by Kilpatrick's finding that 38% of the adult respondents reported the sexual experiences<mask> children to be "pleasant"<mask> only 25% reported them to be "unpleasant." Kilpatrick<mask> found that,<mask> the majority of the women stated that the experience was initiated by the partner, for many (23% of the children 0-14 years and 39% of adolescents 15-17 years) the women reported having been the initiator. Another surprising finding was that only 4% of the respondents reported that they would have liked to have had counseling.</s>
Label encoding: <s>As sexual activity is not inherently harmful to children, there is no basis on which to label all sexual relations between children and adults as abusive. [NEWLINE] [NEWLINE] [STARTQ] A Dutch study published in 1987 found that a sample of boys in paedophilic relationships felt positively about them. And a major if still controversial 1998-2000 meta-study suggests – as J Michael Bailey of Northwestern University, Chicago, says – that such relationships, entered into voluntarily, are "nearly uncorrelated with undesirable outcomes". [ENDQ] [NEWLINE] [STARTQ] Most people find that idea impossible. But writing last year in the peer-reviewed Archives of Sexual Behaviour, Bailey said that while he also found the notion "disturbing", he was forced to recognise that **"persuasive evidence for the harmfulness of paedophilic relationships does not yet exist".** [ENDQ] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] A substantial number of people who as children had sex with adults feel positively about the experience, and do not regard it as abusive in any capacity. [NEWLINE] [NEWLINE] Long-Range Effects of Child and Adolescent Sexual Experiences Positive Review", Allie C. Kilpatrick. [NEWLINE] [NEWLINE] &gt;This book will be disturbing to many readers. The assumption that all children are "damaged" by their experiences is challenged by Kilpatrick's finding that 38% of the adult respondents reported the sexual experiences as children to be "pleasant" while only 25% reported them to be "unpleasant." Kilpatrick also found that, although the majority of the women stated that the experience was initiated by the partner, for many (23% of the children 0-14 years and 39% of adolescents 15-17 years) the women reported having been the initiator. Another surprising finding was that only 4% of the respondents reported that they would have liked to have had counseling.</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] not only negative things come out of the ghetto. [ENDQ] [NEWLINE] That's<mask> I said "on net" a couple posts up. [NEWLINE] [NEWLINE] [STARTQ] <mask> the KKK definitely are not a culture, they are a social movement. [ENDQ] [NEWLINE] Culture: the behaviors and beliefs characteristic of a particular social, ethnic, or age group [NEWLINE] [NEWLINE] I do believe they qualify<mask> a social group. It's a pretty small group these days,<mask> it wasn't always. [NEWLINE] [NEWLINE] [STARTQ] Are you seriously arguing that it is OK to stigmatise (i.e. reduce them into simple, dehumanising qualities) people who definitely are not all bad? [ENDQ] [NEWLINE] It's important to judge people, and groups of people. Just<mask> they aren't all 100% evil doesn't mean we can't judge their choices. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the majority of these people live relatively normal lives,<mask> they go about their own business, without even realising that there are people out there (you) who would characterise the whole culture they are a part of<mask> evil or "toxic" based solely on stereotypes perpetrated by mainstream media. [ENDQ] [NEWLINE] This is just some sort of wishful thinking on your part. I spent a year teaching in the inner city. They *don't* live<mask> anyone else would consider normal lives going about their own business. [NEWLINE] [NEWLINE] Whenever there was a local murder nearby, the joke was that the teachers would all say "don't be from our school, don't be from our school, don't be from our school, damn he's from our school." [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> you should probably get out of your own neighborhood and speak to people in real low income areas, not the ones you see on TV or that you have somehow imagined. [ENDQ] [NEWLINE] right back at ya.</s><pad>
Label encoding: <s> [STARTQ] not only negative things come out of the ghetto. [ENDQ] [NEWLINE] That's why I said "on net" a couple posts up. [NEWLINE] [NEWLINE] [STARTQ] Also the KKK definitely are not a culture, they are a social movement. [ENDQ] [NEWLINE] Culture: the behaviors and beliefs characteristic of a particular social, ethnic, or age group [NEWLINE] [NEWLINE] I do believe they qualify as a social group. It's a pretty small group these days, but it wasn't always. [NEWLINE] [NEWLINE] [STARTQ] Are you seriously arguing that it is OK to stigmatise (i.e. reduce them into simple, dehumanising qualities) people who definitely are not all bad? [ENDQ] [NEWLINE] It's important to judge people, and groups of people. Just because they aren't all 100% evil doesn't mean we can't judge their choices. [NEWLINE] [NEWLINE] [STARTQ] In fact the majority of these people live relatively normal lives, where they go about their own business, without even realising that there are people out there (you) who would characterise the whole culture they are a part of as evil or "toxic" based solely on stereotypes perpetrated by mainstream media. [ENDQ] [NEWLINE] This is just some sort of wishful thinking on your part. I spent a year teaching in the inner city. They *don't* live what anyone else would consider normal lives going about their own business. [NEWLINE] [NEWLINE] Whenever there was a local murder nearby, the joke was that the teachers would all say "don't be from our school, don't be from our school, don't be from our school, damn he's from our school." [NEWLINE] [NEWLINE] [STARTQ] I think you should probably get out of your own neighborhood and speak to people in real low income areas, not the ones you see on TV or that you have somehow imagined. [ENDQ] [NEWLINE] right back at ya.</s><pad>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a solid debate, and everyone is clearly considering other perspectives, which I'm really impressed by. [NEWLINE] [NEWLINE] My contentions are, first, that basic income would not create a "work-free environment" - you have a greater incentive and ability to work productively under basic income than under the safety net. I know, you dealt with this in your chapter on whether or not the welfare trap truly exists,<mask> at the very least basic income recipients have been shown in pilot studies to actually be more productive, and, in a bit of a tired repeated point, take more risks in investments and innovative business attempts which are important for economic growth; and second, that the way I see basic income is this: say automation gets rid of 47% of jobs like that Oxford study showed in the next 50 years: new jobs may open up, that those newly unemployed may be able to train to take,<mask> in the meantime those automated jobs have moved money up the chain to the amassing owners of production (hello Marx). The new jobs those unemployed take will either be redundant pity jobs or expanding new sectors, creating new wealth. Regardless, we have a shitload of money moving up to that 1%, deepening inequality etc etc. I say that we use part of that new money to redistribute into basic income, in a steadily expanding distributed sum, which gets bigger<mask> the amount of overall money going upwards to the owners of production expands. [NEWLINE] [NEWLINE] It's not about the safety net, which can slowly reduce<mask> problems it was designed to solve are taken care of by basic income. It's simply about sharing around the new productivity in a post-scarcity world<mask> there's tonnes of money and we need to decide<mask> to best use it to make our society happy and prosperous.</s>
Label encoding: <s>This is a solid debate, and everyone is clearly considering other perspectives, which I'm really impressed by. [NEWLINE] [NEWLINE] My contentions are, first, that basic income would not create a "work-free environment" - you have a greater incentive and ability to work productively under basic income than under the safety net. I know, you dealt with this in your chapter on whether or not the welfare trap truly exists, but at the very least basic income recipients have been shown in pilot studies to actually be more productive, and, in a bit of a tired repeated point, take more risks in investments and innovative business attempts which are important for economic growth; and second, that the way I see basic income is this: say automation gets rid of 47% of jobs like that Oxford study showed in the next 50 years: new jobs may open up, that those newly unemployed may be able to train to take, but in the meantime those automated jobs have moved money up the chain to the amassing owners of production (hello Marx). The new jobs those unemployed take will either be redundant pity jobs or expanding new sectors, creating new wealth. Regardless, we have a shitload of money moving up to that 1%, deepening inequality etc etc. I say that we use part of that new money to redistribute into basic income, in a steadily expanding distributed sum, which gets bigger as the amount of overall money going upwards to the owners of production expands. [NEWLINE] [NEWLINE] It's not about the safety net, which can slowly reduce as problems it was designed to solve are taken care of by basic income. It's simply about sharing around the new productivity in a post-scarcity world where there's tonnes of money and we need to decide how to best use it to make our society happy and prosperous.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I guess I'm saying that implicitly embedded within the statement "white people shouldn't criticize minorities" is the assumption that we're talking about judging character, rather than objective actions. The statement doesn't really make logical sense otherwise,<mask><mask> we both agree on that. [NEWLINE] [NEWLINE] Kinda like<mask> the phrase "black lives matter" is really "black lives matter *too*." The "all lives matter" movement is really just a failure to understand that we wouldn't have this movement<mask> the implicit too wasn't there. [NEWLINE] [NEWLINE] I suppose it's arguable that these people should just say<mask> they mean,<mask> I<mask> think that we should be willing to give the most charitable interpretation of people's words<mask> we want to claim to be empathetic towards others and cognizant of our biases. [NEWLINE] [NEWLINE] This is getting a little far afield of this CMV,<mask> it's based off of a specific conversation with your friend. I really don't know<mask> she meant her words literally and absolutely, or<mask> you were very clear that you were talking only about the actions and very clear that it wasn't about judging them<mask> a person, or<mask> either of you even had the nuance in mind at the time. [NEWLINE] [NEWLINE] I would<mask><mask> going forward,<mask> you encounter someone saying that white people shouldn't or can't criticize minorities, the best response isn't to reply, that their statement taken literally makes no sense. I'd actually just clarify similarly to<mask> you did just now. And you may find yourself agreeing with them on many points rather than arguing. [NEWLINE] [NEWLINE] <mask> that might not change your view on the concept you specifically set out to argue in this post,<mask> it may change your view of others who say that "white people shouldn't criticize minorities".</s>
Label encoding: <s>I guess I'm saying that implicitly embedded within the statement "white people shouldn't criticize minorities" is the assumption that we're talking about judging character, rather than objective actions. The statement doesn't really make logical sense otherwise, I think we both agree on that. [NEWLINE] [NEWLINE] Kinda like how the phrase "black lives matter" is really "black lives matter *too*." The "all lives matter" movement is really just a failure to understand that we wouldn't have this movement if the implicit too wasn't there. [NEWLINE] [NEWLINE] I suppose it's arguable that these people should just say what they mean, but I also think that we should be willing to give the most charitable interpretation of people's words if we want to claim to be empathetic towards others and cognizant of our biases. [NEWLINE] [NEWLINE] This is getting a little far afield of this CMV, since it's based off of a specific conversation with your friend. I really don't know if she meant her words literally and absolutely, or if you were very clear that you were talking only about the actions and very clear that it wasn't about judging them as a person, or if either of you even had the nuance in mind at the time. [NEWLINE] [NEWLINE] I would argue that going forward, if you encounter someone saying that white people shouldn't or can't criticize minorities, the best response isn't to reply, that their statement taken literally makes no sense. I'd actually just clarify similarly to how you did just now. And you may find yourself agreeing with them on many points rather than arguing. [NEWLINE] [NEWLINE] So that might not change your view on the concept you specifically set out to argue in this post, but it may change your view of others who say that "white people shouldn't criticize minorities".</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Even<mask> you have a genetic pre-disposition to schizophrenia, statistically youu probably wont get it and even<mask> you dont have that pre-disposition you can still become schizophrenic. Genetics play a role in your statistical likelihood<mask> considering a genetic pre-disposition does no actual pre-determine anything, it's still determined by your environment. [NEWLINE] [NEWLINE] [NEWLINE] Likewise a chemical imbalance of dopamine or serotonin can exist<mask> it doesnt follow that you're born with it. Our brains are changing constantly. Whatever you're doing at any time, you're training your brain to be better at it. Through neuroplasticity a certain environment can create a brain which behaves in a certain way, and just like you get better<mask> you practice piano, the more you feel a certain emotion, the more that process is physically occuring in your brain, the more that process is strengthened similar to the physical neurological process of improving your piano skills. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Even if you have a genetic pre-disposition to schizophrenia, statistically youu probably wont get it and even if you dont have that pre-disposition you can still become schizophrenic. Genetics play a role in your statistical likelihood but considering a genetic pre-disposition does no actual pre-determine anything, it's still determined by your environment. [NEWLINE] [NEWLINE] [NEWLINE] Likewise a chemical imbalance of dopamine or serotonin can exist but it doesnt follow that you're born with it. Our brains are changing constantly. Whatever you're doing at any time, you're training your brain to be better at it. Through neuroplasticity a certain environment can create a brain which behaves in a certain way, and just like you get better when you practice piano, the more you feel a certain emotion, the more that process is physically occuring in your brain, the more that process is strengthened similar to the physical neurological process of improving your piano skills. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Could you be looking at this from the wrong angle? You're asking two completely different things: 1)<mask> people don't listen to classical music; 2) Is it<mask> classical music doesn't have a drum set? [NEWLINE] [NEWLINE] The way I see it is that most people stick with<mask> they're used to or exposed to at some formative age. I know someone who listened to a ton of classical music<mask> his father used to play it in the house<mask> he was growing up. His dad was very knowledgeable about the composers, artists, etc. and passed that on to his son. I,<mask><mask><mask><mask>, was exposed to a ton of classical rock (from my friends)<mask> I was growing up. That's still the music that I like the most, and he still likes classical music best.<mask><mask>, I didn't get into classical music at all (I'm still a noob) until I watched Requiem for a Dream and got blown away by the soundtrack - Kronos Quartet did an amazing job.<mask> my friend tried, I was resistant to it<mask> of my own lack of "training." [NEWLINE] [NEWLINE] That's<mask> it is with exposure to music.<mask> you show me<mask>'s "popular" right now I would just stare blankly at the list. The whole Taylor Swift/Nicki Minaj drama meant nothing to me<mask> I don't know a single song by Minaj.<mask> it isn't a matter of the qualities of the music, it's more about exposure and discovery. Classical music isn't a presence the same way pop music is.<mask><mask> that's<mask> accounts for the difference in popularity. [NEWLINE] [NEWLINE] Great post, by the way. Let me know<mask> you think about my $0.02.</s>
Label encoding: <s>Could you be looking at this from the wrong angle? You're asking two completely different things: 1) Why people don't listen to classical music; 2) Is it because classical music doesn't have a drum set? [NEWLINE] [NEWLINE] The way I see it is that most people stick with what they're used to or exposed to at some formative age. I know someone who listened to a ton of classical music because his father used to play it in the house while he was growing up. His dad was very knowledgeable about the composers, artists, etc. and passed that on to his son. I, on the other hand, was exposed to a ton of classical rock (from my friends) when I was growing up. That's still the music that I like the most, and he still likes classical music best. In fact, I didn't get into classical music at all (I'm still a noob) until I watched Requiem for a Dream and got blown away by the soundtrack - Kronos Quartet did an amazing job. Though my friend tried, I was resistant to it because of my own lack of "training." [NEWLINE] [NEWLINE] That's how it is with exposure to music. If you show me what's "popular" right now I would just stare blankly at the list. The whole Taylor Swift/Nicki Minaj drama meant nothing to me because I don't know a single song by Minaj. So it isn't a matter of the qualities of the music, it's more about exposure and discovery. Classical music isn't a presence the same way pop music is. I think that's what accounts for the difference in popularity. [NEWLINE] [NEWLINE] Great post, by the way. Let me know what you think about my $0.02.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] I'm partially persuaded. I'm an academic myself,<mask> far junior to you. The Dean of my College fought to keep his research lab open<mask> being promoted,<mask> he didn't want to stop being a scientist.<mask>, I'm not convinced that being a Dean, for example, means you're still a scientist.<mask><mask> at that point, you're just an administrator. Nevertheless, you raise worthy points about NSF, FDA, DOE posts. Such positions certainly contribute to science in a way that transcends benchwork or running a lab, and are still arguably scientific.<mask> you deserve a Delta for expanding my definition of scientist,<mask> I'm not quite sure NdGT qualifies. [NEWLINE] [NEWLINE] I know,<mask> a scientist, you've got a lot of personal identity invested in being a scientist and will likely still consider yourself a scientist<mask> you retire.<mask> is being a scientist really just about identity? Surely considering oneself a scientist is not enough to qualify them<mask> a scientist. There must be more to it than that. (Think creation scientists.) And I know others think about being a scientists<mask> having a body of knowledge/ expertise.<mask><mask> you get your PhD in biochemistry and spend your life<mask> a bartender, then are you still a scientist? [NEWLINE] [NEWLINE] This is<mask><mask><mask> being a scientist is a job. It's something you do, not who you feel you are or<mask> you know.<mask><mask> you base it on who you feel you are or <mask> you know too many obviously non-scientists get included.<mask>, in that<mask><mask> NdGT is, for all intents and purposed, retired from doing science, I still think he's not a scientist. </s>
Label encoding: <s>&amp;#8710; [NEWLINE] [NEWLINE] I'm partially persuaded. I'm an academic myself, although far junior to you. The Dean of my College fought to keep his research lab open when being promoted, because he didn't want to stop being a scientist. So, I'm not convinced that being a Dean, for example, means you're still a scientist. I think at that point, you're just an administrator. Nevertheless, you raise worthy points about NSF, FDA, DOE posts. Such positions certainly contribute to science in a way that transcends benchwork or running a lab, and are still arguably scientific. So you deserve a Delta for expanding my definition of scientist, but I'm not quite sure NdGT qualifies. [NEWLINE] [NEWLINE] I know, as a scientist, you've got a lot of personal identity invested in being a scientist and will likely still consider yourself a scientist when you retire. But is being a scientist really just about identity? Surely considering oneself a scientist is not enough to qualify them as a scientist. There must be more to it than that. (Think creation scientists.) And I know others think about being a scientists as having a body of knowledge/ expertise. But if you get your PhD in biochemistry and spend your life as a bartender, then are you still a scientist? [NEWLINE] [NEWLINE] This is why I think being a scientist is a job. It's something you do, not who you feel you are or what you know. Because if you base it on who you feel you are or  what you know too many obviously non-scientists get included. So, in that I think NdGT is, for all intents and purposed, retired from doing science, I still think he's not a scientist. </s>
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] I misunderstood my own view, this post helped me clear up my thoughts.<mask><mask> that the Day of Silence demonstrates that many people support LGBT rights, which is beneficial for the movement. Delta awarded for making me realize that. [NEWLINE] [NEWLINE] My issue with the movement (which I did not realize<mask> I posted this) is that it does nothing to combat bullying in general.<mask> you make it unacceptable for kids to be bullied for being gay, bullies will switch back to one of the other timeless insults they've been using for<mask><mask><mask> kids have been assholes to each other, like "you're too fat," "too skinny," "too tall," "too smart," etc. [NEWLINE] [NEWLINE] In other words, bullying will still exist, even<mask> the LGBT movement is successful in attaining their goals. The Day of Silence fails to address<mask> bullying exists, and the environmental factors that lead kids to attack each other. [NEWLINE] [NEWLINE] Peer pressuring each other into being nice all the time doesn't get to the root of the problem; it doesn't keep his alcoholic father from beating him. It just removes an outlet for the bully's pent up anger. That isn't to say that bullying is an appropriate outlet,<mask><mask> you take away one outlet, they will just find another. [NEWLINE] [NEWLINE] To state my issue with the Day of Silence more succinctly... it is too single-minded. They're simply making bullying someone else's problem to deal with.<mask> it's not a trans kids being bullied, then,<mask> someone said below, it's "the fat kids, nerds, kids with glasses, kids with red hair, kids with pimples, orphans, or kids with braces." We need to address the root cause of bullying in general.</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] I misunderstood my own view, this post helped me clear up my thoughts. I agree that the Day of Silence demonstrates that many people support LGBT rights, which is beneficial for the movement. Delta awarded for making me realize that. [NEWLINE] [NEWLINE] My issue with the movement (which I did not realize when I posted this) is that it does nothing to combat bullying in general. If you make it unacceptable for kids to be bullied for being gay, bullies will switch back to one of the other timeless insults they've been using for as long as kids have been assholes to each other, like "you're too fat," "too skinny," "too tall," "too smart," etc. [NEWLINE] [NEWLINE] In other words, bullying will still exist, even if the LGBT movement is successful in attaining their goals. The Day of Silence fails to address why bullying exists, and the environmental factors that lead kids to attack each other. [NEWLINE] [NEWLINE] Peer pressuring each other into being nice all the time doesn't get to the root of the problem; it doesn't keep his alcoholic father from beating him. It just removes an outlet for the bully's pent up anger. That isn't to say that bullying is an appropriate outlet, but when you take away one outlet, they will just find another. [NEWLINE] [NEWLINE] To state my issue with the Day of Silence more succinctly... it is too single-minded. They're simply making bullying someone else's problem to deal with. If it's not a trans kids being bullied, then, as someone said below, it's "the fat kids, nerds, kids with glasses, kids with red hair, kids with pimples, orphans, or kids with braces." We need to address the root cause of bullying in general.</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Can you explain this point a little bit more? [ENDQ] [NEWLINE] <mask> I mean to say is, war is a terrible thing. The destruction and carnage it brings tears apart nations and leaves a battered infrastructure,<mask><mask> to possibly leaving the citizens politically divided. Intervention and arbitration through diplomatic means will inherently favor the arbitrator (in this case America). And<mask> that's all fine and dandy in the short term, America's long and storied history of intervening in other countries has given it a reputation that has given it many enemies. [NEWLINE] [NEWLINE] Fighting for the good of the people in the countries is one thing,<mask> you seem to be arguing that intervening in other countries is an effective foreign policy in general, not that it is a just one. [NEWLINE] [NEWLINE] Even<mask> you mean to<mask><mask> intervention is a better policy than non-intervention for the good of the country intervening, I still disagree. Non-intervention can<mask> bring about a healthier and stronger relationship between two countries,<mask> only<mask> intervention has a detrimental effect,<mask> opposed to non-intervention having a positive effect. [NEWLINE] [NEWLINE] Sri Lanka suffered from a gripping civil war for around 50 years this past century, and only recently resolved it. India, its closest and<mask> powerful neighbor, did not intervene in the civil war.<mask>, today, relations between the two countries are very strong. Would relations be stronger<mask> India intervened on behalf of the victorious side? Possibly,<mask> the fact of the matter is that Sri Lanka is in a state of relative peace and progressive civil rights plus it had friendly relations with its neighbors. [NEWLINE] [NEWLINE] Non-intervention can be<mask> effective a strategy<mask> intervention, plus it usually has the side-effect of leaving the country with a stronger infrastructure and less damage.</s>
Label encoding: <s> [STARTQ] Can you explain this point a little bit more? [ENDQ] [NEWLINE] What I mean to say is, war is a terrible thing. The destruction and carnage it brings tears apart nations and leaves a battered infrastructure, in addition to possibly leaving the citizens politically divided. Intervention and arbitration through diplomatic means will inherently favor the arbitrator (in this case America). And while that's all fine and dandy in the short term, America's long and storied history of intervening in other countries has given it a reputation that has given it many enemies. [NEWLINE] [NEWLINE] Fighting for the good of the people in the countries is one thing, but you seem to be arguing that intervening in other countries is an effective foreign policy in general, not that it is a just one. [NEWLINE] [NEWLINE] Even if you mean to argue that intervention is a better policy than non-intervention for the good of the country intervening, I still disagree. Non-intervention can also bring about a healthier and stronger relationship between two countries, if only because intervention has a detrimental effect, as opposed to non-intervention having a positive effect. [NEWLINE] [NEWLINE] Sri Lanka suffered from a gripping civil war for around 50 years this past century, and only recently resolved it. India, its closest and also powerful neighbor, did not intervene in the civil war. However, today, relations between the two countries are very strong. Would relations be stronger if India intervened on behalf of the victorious side? Possibly, but the fact of the matter is that Sri Lanka is in a state of relative peace and progressive civil rights plus it had friendly relations with its neighbors. [NEWLINE] [NEWLINE] Non-intervention can be as effective a strategy as intervention, plus it usually has the side-effect of leaving the country with a stronger infrastructure and less damage.</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You kind of answered your own question. There are<mask> many different levels of psychiatric disorder that it just can't be limited on a blanket basis.  Should someone who is overtly violent and threatening be limited? Sure, everyone can agree on that.<mask><mask> about the other cases? [NEWLINE] [NEWLINE] Seasonal depression (which is caused by vitamin deficiencies) is a mood disorder. Should someone be limited due to feeling down every now and then? [NEWLINE] [NEWLINE] My wife was diagnosed<mask> bipolar<mask> she was in Highschool (<mask> hormones are at their craziest) and placed on all sorts of medications for it. 10 years later, she decides to get off the medication and come off birth control, and voila! She has absolutely no issues whatsoever. Should she be barred? [NEWLINE] [NEWLINE] PTSD has hundreds of symptoms. I have no problems with a normal life. Play with my kid, work, deal with people daily, my only actual "symptom" of PTSD is I don't like large/loud gatherings of people. Should I be barred from owning firearms<mask> I don't like malls or concerts? Not to mention that soldiers are not the leading group of PTSD sufferers (that title belongs to car accident victims). [NEWLINE] [NEWLINE] <mask> about a woman who has has PTSD being abused by her ex-husband? Should she be barred from protecting herself (or her kids)<mask> he violates that restraining order and tries to break into her apartment to attack her? [NEWLINE] [NEWLINE] Tl;dr version: The severity of mental illnesses can vary<mask> much (and be<mask> over diagnosed) that a blanket law barring ownership would do much more harm than good. A case by case basis,<mask> is already available (and rarely used) by law is a better system. </s>
Label encoding: <s>You kind of answered your own question. There are so many different levels of psychiatric disorder that it just can't be limited on a blanket basis.  Should someone who is overtly violent and threatening be limited? Sure, everyone can agree on that. But what about the other cases? [NEWLINE] [NEWLINE] Seasonal depression (which is caused by vitamin deficiencies) is a mood disorder. Should someone be limited due to feeling down every now and then? [NEWLINE] [NEWLINE] My wife was diagnosed as bipolar when she was in Highschool ( when hormones are at their craziest) and placed on all sorts of medications for it. 10 years later, she decides to get off the medication and come off birth control, and voila! She has absolutely no issues whatsoever. Should she be barred? [NEWLINE] [NEWLINE] PTSD has hundreds of symptoms. I have no problems with a normal life. Play with my kid, work, deal with people daily, my only actual "symptom" of PTSD is I don't like large/loud gatherings of people. Should I be barred from owning firearms because I don't like malls or concerts? Not to mention that soldiers are not the leading group of PTSD sufferers (that title belongs to car accident victims). [NEWLINE] [NEWLINE] What about a woman who has has PTSD being abused by her ex-husband? Should she be barred from protecting herself (or her kids) if he violates that restraining order and tries to break into her apartment to attack her? [NEWLINE] [NEWLINE] Tl;dr version: The severity of mental illnesses can vary so much (and be so over diagnosed) that a blanket law barring ownership would do much more harm than good. A case by case basis, as is already available (and rarely used) by law is a better system. </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Taliban and Al-Qaeda grew out of Mujahideen. Which grew into power during Soviet war in Afghanistan. [NEWLINE] [NEWLINE] [STARTQ] Operation Cyclone was the code name for the United States Central Intelligence Agency (CIA) program to arm and finance the Afghan mujahideen prior to and during the Soviet war in Afghanistan, from 1979 to 1989. [ENDQ] [NEWLINE] [Wiki]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] The origins of al-Qaeda<mask> a network inspiring terrorism around the world and training operatives can be traced to the Soviet War in Afghanistan (December 1979 – February 1989).A CIA program called Operation Cyclone channeled funds through Pakistan's Inter-Services Intelligence agency to the Afghan Mujahideen who were fighting the Soviet occupation. [ENDQ] [NEWLINE] [Wiki]( [URL] #Jihad_in_Afghanistan) [NEWLINE] [NEWLINE] [STARTQ] In August 1996, bin Laden declared war against the United States.[101]<mask> the assurance of President George H.W. Bush to King Fahd in 1990, that all U.S. forces based in Saudi Arabia would be withdrawn once the Iraqi threat had been dealt with, by 1996 the Americans were still there. Bush cited the necessity of dealing with the remnants of Saddam's regime (which Bush had chosen not to destroy). Bin Laden's view was that "the 'evils' of the Middle East arose from America's attempt to take over the region and from its support for Israel. [ENDQ] [STARTQ] [ENDQ] [STARTQ]... [ENDQ] [STARTQ] [ENDQ] [STARTQ] In Afghanistan, bin Laden and al-Qaeda raised money from "donors from the days of the Soviet jihad", and from the Pakistani ISI to establish more training camps for Mujahideen fighters. [ENDQ] [NEWLINE] [Wiki]( [URL] #Sudan_and_return_to_Afghanistan) [NEWLINE] </s>
Label encoding: <s>Taliban and Al-Qaeda grew out of Mujahideen. Which grew into power during Soviet war in Afghanistan. [NEWLINE] [NEWLINE] [STARTQ] Operation Cyclone was the code name for the United States Central Intelligence Agency (CIA) program to arm and finance the Afghan mujahideen prior to and during the Soviet war in Afghanistan, from 1979 to 1989. [ENDQ] [NEWLINE] [Wiki]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] The origins of al-Qaeda as a network inspiring terrorism around the world and training operatives can be traced to the Soviet War in Afghanistan (December 1979 – February 1989).A CIA program called Operation Cyclone channeled funds through Pakistan's Inter-Services Intelligence agency to the Afghan Mujahideen who were fighting the Soviet occupation. [ENDQ] [NEWLINE] [Wiki]( [URL] #Jihad_in_Afghanistan) [NEWLINE] [NEWLINE] [STARTQ] In August 1996, bin Laden declared war against the United States.[101] Despite the assurance of President George H.W. Bush to King Fahd in 1990, that all U.S. forces based in Saudi Arabia would be withdrawn once the Iraqi threat had been dealt with, by 1996 the Americans were still there. Bush cited the necessity of dealing with the remnants of Saddam's regime (which Bush had chosen not to destroy). Bin Laden's view was that "the 'evils' of the Middle East arose from America's attempt to take over the region and from its support for Israel. [ENDQ] [STARTQ] [ENDQ] [STARTQ]... [ENDQ] [STARTQ] [ENDQ] [STARTQ] In Afghanistan, bin Laden and al-Qaeda raised money from "donors from the days of the Soviet jihad", and from the Pakistani ISI to establish more training camps for Mujahideen fighters. [ENDQ] [NEWLINE] [Wiki]( [URL] #Sudan_and_return_to_Afghanistan) [NEWLINE] </s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Betting on there being a God is the smart decision<mask> there might be a pay off,<mask> atheism has no reward. [ENDQ] [NEWLINE] The key word you're stating is betting,<mask> you "bet" on "God" then you aren't actually believing in it, you can't be legitimately convinced that a deity exists simply<mask> you feel like it would be the higher payout. [NEWLINE] [NEWLINE] Let's consider leprechaun gold at the end of the rainbow, it will be more profitable<mask> this is real correct?<mask> by the logic of your view, the average person will honestly and sincerely believe that at the end of every rainbow, there is literally and by all means really, a pot of gold, and this affirmation of belief was based *solely* on the idea that<mask> the gold does exist then it will be more profitable than non existing. Do you see<mask> ridiculous this is *<mask> a reason to accept the belief*? [NEWLINE] [NEWLINE] You aren't really believing in something just<mask> you say you think it's the more profitable choice.<mask> there was a cash prize for believing 2+2 = 7, you might<mask><mask> everyone should just believe 2 +2 is 7 for the money you would otherwise not get by believing 2+2 =4.<mask> there is NOTHING ELSE to convince you that it is *true*. I can say outloud " i believe 2+2 is 7"<mask><mask> i'm not convinced. [NEWLINE] [NEWLINE] To end on that note, do you think an all knowing god will think you are a legitimate believer<mask> the only reason for your belief was that it was the better gamble? I'm telling you that<mask> that's all there is to it, then you do not really believe. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Betting on there being a God is the smart decision because there might be a pay off, while atheism has no reward. [ENDQ] [NEWLINE] The key word you're stating is betting, if you "bet" on "God" then you aren't actually believing in it, you can't be legitimately convinced that a deity exists simply because you feel like it would be the higher payout. [NEWLINE] [NEWLINE] Let's consider leprechaun gold at the end of the rainbow, it will be more profitable if this is real correct? So by the logic of your view, the average person will honestly and sincerely believe that at the end of every rainbow, there is literally and by all means really, a pot of gold, and this affirmation of belief was based *solely* on the idea that if the gold does exist then it will be more profitable than non existing. Do you see how ridiculous this is * as a reason to accept the belief*? [NEWLINE] [NEWLINE] You aren't really believing in something just because you say you think it's the more profitable choice. If there was a cash prize for believing 2+2 = 7, you might argue that everyone should just believe 2 +2 is 7 for the money you would otherwise not get by believing 2+2 =4. But there is NOTHING ELSE to convince you that it is *true*. I can say outloud " i believe 2+2 is 7" even though i'm not convinced. [NEWLINE] [NEWLINE] To end on that note, do you think an all knowing god will think you are a legitimate believer if the only reason for your belief was that it was the better gamble? I'm telling you that if that's all there is to it, then you do not really believe. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Read<mask> I've written carefully. Do you see<mask> I said<mask>? [ENDQ] [NEWLINE] <mask> you do not, then you are simply playing the Devil's Advocate, in which case, I would simply ask that you have clarified that up front, to reduce confusion. [NEWLINE] [NEWLINE] [STARTQ] This is simply not true. There are plenty of situations in which there is much disagreement on morality. [ENDQ] [NEWLINE] Of course there are disagreements on<mask> is moral.  That's one of the functions of a political system, to iron out those disagreements into a system that ideally most,<mask> not all, members of the society can agree upon. [NEWLINE] [NEWLINE] No perfect solution can exist.  In the case of a democratically elected government, from a certain perspective and to a certain extent and using certain definitions of the words involved the OP is right, the tyranny of the democratic majority can victimize a small minority of people.  Not<mask> many<mask> of restrictions imposed by the Constitution,<mask> some yes. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, under voluntarism the tyranny of a small and powerful minority victimizes the majority of people.  All anarchies do.  Power forms spontaneously, it's bred into us by millions upon millions of years of social evolution. [NEWLINE] [NEWLINE] Maybe, some time far in the future,<mask> we've managed to get rid of our evolutionary baggage and we are all better more egalitarian humans, maybe voluntarism will work.  Once you eliminate the "asshole factor" lots of currently unrealistic alternatives begin to work.  Many of which are more efficient systems ignoring the havoc that can be wrecked upon it by essentially bad input. [NEWLINE] [NEWLINE] <mask> we're not living in that time,<mask> these ideas aren't practical possibilities.  Not now.  Not<mask>.</s>
Label encoding: <s> [STARTQ] Read what I've written carefully. Do you see where I said so? [ENDQ] [NEWLINE] If you do not, then you are simply playing the Devil's Advocate, in which case, I would simply ask that you have clarified that up front, to reduce confusion. [NEWLINE] [NEWLINE] [STARTQ] This is simply not true. There are plenty of situations in which there is much disagreement on morality. [ENDQ] [NEWLINE] Of course there are disagreements on what is moral.  That's one of the functions of a political system, to iron out those disagreements into a system that ideally most, if not all, members of the society can agree upon. [NEWLINE] [NEWLINE] No perfect solution can exist.  In the case of a democratically elected government, from a certain perspective and to a certain extent and using certain definitions of the words involved the OP is right, the tyranny of the democratic majority can victimize a small minority of people.  Not as many because of restrictions imposed by the Constitution, but some yes. [NEWLINE] [NEWLINE] On the other hand, under voluntarism the tyranny of a small and powerful minority victimizes the majority of people.  All anarchies do.  Power forms spontaneously, it's bred into us by millions upon millions of years of social evolution. [NEWLINE] [NEWLINE] Maybe, some time far in the future, when we've managed to get rid of our evolutionary baggage and we are all better more egalitarian humans, maybe voluntarism will work.  Once you eliminate the "asshole factor" lots of currently unrealistic alternatives begin to work.  Many of which are more efficient systems ignoring the havoc that can be wrecked upon it by essentially bad input. [NEWLINE] [NEWLINE] But we're not living in that time, so these ideas aren't practical possibilities.  Not now.  Not yet.</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is pure sophistry and you know it. The complexity of religions doesn't prevent you from making comparisons that contain a fair degree of accuracy. No amount of pedantry will save you from the fact that Islam overall contains more potential for violence than Buddhism overall. [NEWLINE] [NEWLINE] One could always cite some fringe Buddhist sect that advocates the killing of every living creature on the planet,<mask><mask> only a tiny tiny fraction of religious adherents believe it, and the basic tenets of Buddhism don't allow for it, you can't tout it<mask> proof that Buddhism is generally<mask> violent<mask> Islam. [NEWLINE] [NEWLINE] I can already imagine the gears turning, "Well ISIS is a tiny fraction of Muslims!" To say that Islam can cause violence is not to say that Islam necessarily causes violence. Thankfully, there are billions of Muslims who've managed to ignore the most horrific parts of the Quran and Hadith.<mask> until that's the case for all Muslims, it will still be valid to point out links between real world violence and the teachings in Islam. [NEWLINE] [NEWLINE] It's<mask> worth mentioning that even<mask> -called "moderate" Muslims believe some pretty terrible shit about apostates, women, homosexuals etc. These attitudes have the potential for violence and its been borne out in places like Indonesia. Indonesia, home of the moderates,<mask> women are caned for being gang-raped.<mask> that isn't religiously-sanctioned state violence I don't know<mask> is. [NEWLINE] [NEWLINE] In any case, it's<mask> the point. I'm not interested in doing the calculus of which religion is more violent controlling for all relevant factors. With respect to the OP's point, it's simply enough to say that religions clearly motivate violence, and<mask> would logically have an effect on violence<mask> they disappeared.</s>
Label encoding: <s>This is pure sophistry and you know it. The complexity of religions doesn't prevent you from making comparisons that contain a fair degree of accuracy. No amount of pedantry will save you from the fact that Islam overall contains more potential for violence than Buddhism overall. [NEWLINE] [NEWLINE] One could always cite some fringe Buddhist sect that advocates the killing of every living creature on the planet, but if only a tiny tiny fraction of religious adherents believe it, and the basic tenets of Buddhism don't allow for it, you can't tout it as proof that Buddhism is generally as violent as Islam. [NEWLINE] [NEWLINE] I can already imagine the gears turning, "Well ISIS is a tiny fraction of Muslims!" To say that Islam can cause violence is not to say that Islam necessarily causes violence. Thankfully, there are billions of Muslims who've managed to ignore the most horrific parts of the Quran and Hadith. But until that's the case for all Muslims, it will still be valid to point out links between real world violence and the teachings in Islam. [NEWLINE] [NEWLINE] It's also worth mentioning that even so -called "moderate" Muslims believe some pretty terrible shit about apostates, women, homosexuals etc. These attitudes have the potential for violence and its been borne out in places like Indonesia. Indonesia, home of the moderates, where women are caned for being gang-raped. If that isn't religiously-sanctioned state violence I don't know what is. [NEWLINE] [NEWLINE] In any case, it's besides the point. I'm not interested in doing the calculus of which religion is more violent controlling for all relevant factors. With respect to the OP's point, it's simply enough to say that religions clearly motivate violence, and so would logically have an effect on violence if they disappeared.</s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yea, somehow I always confuse the names of Cantor and Godel (both great mathematicians!). [NEWLINE] [NEWLINE] Bertrand Russell studied mathematics,<mask> it's hardly fair to just call him a philosopher. The same goes for Whitehead.<mask><mask> Wikipedia, Whitehead went to a philosophy lecture for the first time<mask> he was 63! [NEWLINE] [NEWLINE] [STARTQ] Certainly computer scientists are interested in work done by philosophers on logic, I remember only last semester we had an emeritus professor of computer science attending our proof theory reading group [...] [ENDQ] [NEWLINE] Well, certainly a person can be interested in more than one subject! And a subject like proof theory is more a branch of mathematics than philosophy anyway. [NEWLINE] [NEWLINE] Note that<mask><mask><mask><mask> that philosophy is useless, it can be quite an interesting subject. The borders between philosophy and other subjects aren't very clear in the first place, especially in an area like formal logic.<mask>,<mask><mask><mask> direct applicability to fields like physics/mathematics/biology/computer science goes, philosophy is quite far down the list. Another thing that one often sees, is that a particular philosophical question is debated for a long time (sometimes for centuries),<mask> eventually the question is answered by science or mathematics. For example, the question of the elements, logical consistency, whether two objects can be exactly the same, and more plain questions like zeno's paradox, the raven paradox, etc.<mask><mask><mask> some people would say that instead of doing philosophy, we should simply be doing science &amp; mathematics<mask> keeping the philosophical questions in the back of our head. That will get us answers more quickly than by doing philosophy. Biology and neuroscience will probably have a lot to say about open questions in philosophy in the coming century.</s>
Label encoding: <s>Yea, somehow I always confuse the names of Cantor and Godel (both great mathematicians!). [NEWLINE] [NEWLINE] Bertrand Russell studied mathematics, so it's hardly fair to just call him a philosopher. The same goes for Whitehead. According to Wikipedia, Whitehead went to a philosophy lecture for the first time when he was 63! [NEWLINE] [NEWLINE] [STARTQ] Certainly computer scientists are interested in work done by philosophers on logic, I remember only last semester we had an emeritus professor of computer science attending our proof theory reading group [...] [ENDQ] [NEWLINE] Well, certainly a person can be interested in more than one subject! And a subject like proof theory is more a branch of mathematics than philosophy anyway. [NEWLINE] [NEWLINE] Note that I do not think that philosophy is useless, it can be quite an interesting subject. The borders between philosophy and other subjects aren't very clear in the first place, especially in an area like formal logic. However, as far as direct applicability to fields like physics/mathematics/biology/computer science goes, philosophy is quite far down the list. Another thing that one often sees, is that a particular philosophical question is debated for a long time (sometimes for centuries), but eventually the question is answered by science or mathematics. For example, the question of the elements, logical consistency, whether two objects can be exactly the same, and more plain questions like zeno's paradox, the raven paradox, etc. Because of this some people would say that instead of doing philosophy, we should simply be doing science &amp; mathematics while keeping the philosophical questions in the back of our head. That will get us answers more quickly than by doing philosophy. Biology and neuroscience will probably have a lot to say about open questions in philosophy in the coming century.</s>
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd say that murder is often about power. A person who murders is asserting that their life is worth more than the person that they're killing. Murder out of jealousy (you walk in to see your partner sleeping with someone else) comes from an anger that has its roots in the loss of power. Then there's the case of war, in which murder is used very obviously<mask> a tool of power of the state. [NEWLINE] [NEWLINE] There's<mask> war rape, which has been prevalent throughout most of history. Undoubtedly, there's some sexual energy behind war rape,<mask><mask> it occurs on a massive scale, it's another tool<mask> the power of one group is used to lessen the power of others. [NEWLINE] [NEWLINE] (I consider theft to be a different case - it's generally not motivated out of a wish to cause harm to someone,<mask> instead a wish to increase one's own material wealth.) [NEWLINE] [NEWLINE] <mask><mask><mask> rape is often about power, even on the individual level.<mask> I don't think that means that sex isn't a part of it: we often channel our desires for things through our biological drives. Think about the stress eater, who feels anxiety and worry and attempts to remedy that by consuming calories. Rape,<mask><mask>, is a channeling of the frustration of being powerless into a person's sexual drive. Consider the date rapist who might be both horny and dealing with feelings of rejection and insecurity.<mask> that rapist was only horny, their moral sense might reign them in from forcing themselves on someone.<mask><mask> these feelings of rejection weigh heavily upon them, that need to reclaim power - a powerful need upon itself - merges with the sexual drive and overrides a person's ethical objections to forcing themselves on someone.</s>
Label encoding: <s>I'd say that murder is often about power. A person who murders is asserting that their life is worth more than the person that they're killing. Murder out of jealousy (you walk in to see your partner sleeping with someone else) comes from an anger that has its roots in the loss of power. Then there's the case of war, in which murder is used very obviously as a tool of power of the state. [NEWLINE] [NEWLINE] There's also war rape, which has been prevalent throughout most of history. Undoubtedly, there's some sexual energy behind war rape, but when it occurs on a massive scale, it's another tool where the power of one group is used to lessen the power of others. [NEWLINE] [NEWLINE] (I consider theft to be a different case - it's generally not motivated out of a wish to cause harm to someone, but instead a wish to increase one's own material wealth.) [NEWLINE] [NEWLINE] So I think rape is often about power, even on the individual level. But I don't think that means that sex isn't a part of it: we often channel our desires for things through our biological drives. Think about the stress eater, who feels anxiety and worry and attempts to remedy that by consuming calories. Rape, I think, is a channeling of the frustration of being powerless into a person's sexual drive. Consider the date rapist who might be both horny and dealing with feelings of rejection and insecurity. If that rapist was only horny, their moral sense might reign them in from forcing themselves on someone. But because these feelings of rejection weigh heavily upon them, that need to reclaim power - a powerful need upon itself - merges with the sexual drive and overrides a person's ethical objections to forcing themselves on someone.</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think your first "<mask> " follows.  There are many inefficient food sources that haven't been bred to be companions, and the anthropology on the domestication of dogs doesn't support the claim that their domestication *followed* from their status<mask> inefficient food sources. [NEWLINE] [NEWLINE] The problem I have with your claim is that you're basically adopting a purely moral-relativist view, that eating dogs is morally wrong<mask> "we" empathize with them. [NEWLINE] [NEWLINE] <mask> who is "we"? Some cultures don't empathize with dogs (some have never seen them!).  Is it morally wrong for those cultures to eat dogs?<mask><mask>,<mask>?<mask> not -<mask> not? And are you happy with this being a type of moral relativism? [NEWLINE] [NEWLINE] <mask> others have pointed out, for some cultures<mask> dogs are eaten, there's a  population of dogs-<mask> -pets, and a separate population of dogs-<mask> -food.  By their lights, our reaction to the practice is odd,<mask> we're combining<mask> one category (all dogs)  something that they see<mask> two.  (We do the same thing in the US with rabbits; there are<mask> people who keep fish<mask> pets and eat fish -<mask> I'll concede that in the latter example, they're different species). [NEWLINE] [NEWLINE] <mask> even your empathy argument fails, unless you can explain<mask> our "Western" categorization (all dogs, taken together) should be imposed on a culture with a different categorization ( pets, and food source) and allow us to judge them<mask> committing a moral wrong. They too have empathy (for the pet category) and don't eat them,<mask> they meet your standard.</s><pad><pad><pad>
Label encoding: <s>I don't think your first " therefore " follows.  There are many inefficient food sources that haven't been bred to be companions, and the anthropology on the domestication of dogs doesn't support the claim that their domestication *followed* from their status as inefficient food sources. [NEWLINE] [NEWLINE] The problem I have with your claim is that you're basically adopting a purely moral-relativist view, that eating dogs is morally wrong because "we" empathize with them. [NEWLINE] [NEWLINE] But who is "we"? Some cultures don't empathize with dogs (some have never seen them!).  Is it morally wrong for those cultures to eat dogs? If so, why? If not - why not? And are you happy with this being a type of moral relativism? [NEWLINE] [NEWLINE] As others have pointed out, for some cultures where dogs are eaten, there's a  population of dogs- as -pets, and a separate population of dogs- as -food.  By their lights, our reaction to the practice is odd, because we're combining as one category (all dogs)  something that they see as two.  (We do the same thing in the US with rabbits; there are also people who keep fish as pets and eat fish - but I'll concede that in the latter example, they're different species). [NEWLINE] [NEWLINE] So even your empathy argument fails, unless you can explain why our "Western" categorization (all dogs, taken together) should be imposed on a culture with a different categorization ( pets, and food source) and allow us to judge them as committing a moral wrong. They too have empathy (for the pet category) and don't eat them, so they meet your standard.</s><pad><pad><pad>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First of all, wow,<mask> an awesome, thought-provoking comment thread. Thanks<mask> much /u/Madplato and OP for the discussion.<mask>, I apologize in advance for my English; it's my third fluent language and I've never been confident at it. [NEWLINE] [NEWLINE] I just wanted to add one point, in response to OP's comment: [NEWLINE] [NEWLINE] [STARTQ] I'm not making the positive claim that there is no God or demons, I'm making the claim that it's unreasonable to believe in them, in which case absence of evidence does make it unreasonable. [ENDQ] [NEWLINE] I potentially see a problem with the "absence of evidence" part.<mask> you've said, there is no *scientific* way to prove or disprove the existence of God; we can't find an evidence that is objectively observable by everyone. [NEWLINE] [NEWLINE] For religious people (especially those that believe in a personal God, e.g., Christians), their "evidence" for the existence of God--things that have strongly convicted them of their faith--are personal, subjective life experiences that they've had, which atheists cannot experience or observe first-hand. [NEWLINE] [NEWLINE] You said in another comment thread [NEWLINE] [NEWLINE] [STARTQ] I just don't trust my own [subjective] senses over scientific consensus.<mask> my belief is that I was probably wrong. [ENDQ] [NEWLINE] and I totally understand that (I've<mask> been struggling with the same exact issue).<mask> just<mask> your experience was personal, and other people can't observe it, does that invalidate its credibility? Maybe the answer is yes<mask> you look at it from a purely scientific point of view,<mask> you should<mask> consider the limitations of science in terms of<mask> it can and can't comment on.</s>
Label encoding: <s>First of all, wow, what an awesome, thought-provoking comment thread. Thanks so much /u/Madplato and OP for the discussion. Also, I apologize in advance for my English; it's my third fluent language and I've never been confident at it. [NEWLINE] [NEWLINE] I just wanted to add one point, in response to OP's comment: [NEWLINE] [NEWLINE] [STARTQ] I'm not making the positive claim that there is no God or demons, I'm making the claim that it's unreasonable to believe in them, in which case absence of evidence does make it unreasonable. [ENDQ] [NEWLINE] I potentially see a problem with the "absence of evidence" part. As you've said, there is no *scientific* way to prove or disprove the existence of God; we can't find an evidence that is objectively observable by everyone. [NEWLINE] [NEWLINE] For religious people (especially those that believe in a personal God, e.g., Christians), their "evidence" for the existence of God--things that have strongly convicted them of their faith--are personal, subjective life experiences that they've had, which atheists cannot experience or observe first-hand. [NEWLINE] [NEWLINE] You said in another comment thread [NEWLINE] [NEWLINE] [STARTQ] I just don't trust my own [subjective] senses over scientific consensus. So my belief is that I was probably wrong. [ENDQ] [NEWLINE] and I totally understand that (I've also been struggling with the same exact issue). But just because your experience was personal, and other people can't observe it, does that invalidate its credibility? Maybe the answer is yes if you look at it from a purely scientific point of view, but you should also consider the limitations of science in terms of what it can and can't comment on.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't agree with the OP at all,<mask> I'm going to ask for a massive "citation needed" for the othering going on here.<mask> has any animal psychology study ever concluded that animals can't even comprehend<mask> the nature of an event happening to them is? [ENDQ] [NEWLINE] 1) Most animals do not posses<mask> is known<mask> 'theory of mind', this is the knowledge that someone other than yourself has thoughts and feelings. An animal without this couldn't hope to understand<mask> the human was thinking at the time and would only think "am i or am i not in danger etc". [NEWLINE] [NEWLINE] 2) An example which proves the point and is easily understandable is simply pets and vets. Many pets hate going to vets<mask> the fact that it is for their own good. They do not understand this.<mask> should they then understand<mask> a human is doing whatever disgusting action to them? The only time animals seem to do this<mask> related to beasiality is<mask> males will for lack of a better word hump humans, knowing that they are not the same species. This could be due to hormones, asserting dominance, on this i am not sure. [NEWLINE] [NEWLINE] [STARTQ] <mask> does this hypothetical animal attack and the subsequent killing of the animal even relate to or support your argument..? [ENDQ] [NEWLINE] OP mentioned the animal resisting through attack, depending on the animal this attack could be very serious and once that occurs the future of the animal is dependant on<mask> violent it could be in the future. For dogs this normally ends in euthanasia for example, either<mask><mask><mask> of the violence or from a lack of ownership<mask> it would be confiscated from the owner (<mask> it is illegal in that state/country).</s>
Label encoding: <s> [STARTQ] I don't agree with the OP at all, but I'm going to ask for a massive "citation needed" for the othering going on here. When has any animal psychology study ever concluded that animals can't even comprehend what the nature of an event happening to them is? [ENDQ] [NEWLINE] 1) Most animals do not posses what is known as 'theory of mind', this is the knowledge that someone other than yourself has thoughts and feelings. An animal without this couldn't hope to understand what the human was thinking at the time and would only think "am i or am i not in danger etc". [NEWLINE] [NEWLINE] 2) An example which proves the point and is easily understandable is simply pets and vets. Many pets hate going to vets despite the fact that it is for their own good. They do not understand this. Why should they then understand why a human is doing whatever disgusting action to them? The only time animals seem to do this when related to beasiality is when males will for lack of a better word hump humans, knowing that they are not the same species. This could be due to hormones, asserting dominance, on this i am not sure. [NEWLINE] [NEWLINE] [STARTQ] How does this hypothetical animal attack and the subsequent killing of the animal even relate to or support your argument..? [ENDQ] [NEWLINE] OP mentioned the animal resisting through attack, depending on the animal this attack could be very serious and once that occurs the future of the animal is dependant on how violent it could be in the future. For dogs this normally ends in euthanasia for example, either as a result of the violence or from a lack of ownership as it would be confiscated from the owner ( if it is illegal in that state/country).</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>your reading comprehension is lacking very badly, my friend. that makes it very difficult to have any sort of serious conversation with you.<mask> let me give this one try. [NEWLINE] [NEWLINE] [STARTQ] <mask> the other guy is typically offering her is that he treats her like shit (in one form or another). [ENDQ] [NEWLINE] nope. some guys may<mask> do this.<mask><mask> someone who has actually dated someone really, really shitty who I definitely shouldn't have dated, I have to say that I remember the qualities that made me attracted to him. they didn't make up for his flaws,<mask><mask> him being such a complete shithead not worth dating, some "nice guy" could certainly have observed to an extent<mask> was going on and could have learned from the few things he was doing right and that made me attracted to him. [NEWLINE] [NEWLINE] [STARTQ] People may notice that other people are being nice,<mask> the assumption is that there is an ulterior motive. [ENDQ] [NEWLINE] nope.<mask> a guy is nice to me, I generally assume he's, you know, nice. And I notice.<mask>,<mask>, he asks me out, I reject him, and he insults me, then that's the moment<mask> it becomes clear that he wasn't really a nice guy, and is only superficially nice to people he wants to fuck.<mask> he were really nice, he'd continue to be nice (or at least neutral) after being rejected (and yes, this is possible and people do this. I've been rejected and continued to be civil, and I've rejected several guys who continued to be civil. this isn't much to ask, and<mask> someone were genuinely nice, it would come very easily to them). [NEWLINE] [NEWLINE] </s>
Label encoding: <s>your reading comprehension is lacking very badly, my friend. that makes it very difficult to have any sort of serious conversation with you. but let me give this one try. [NEWLINE] [NEWLINE] [STARTQ] What the other guy is typically offering her is that he treats her like shit (in one form or another). [ENDQ] [NEWLINE] nope. some guys may also do this. but as someone who has actually dated someone really, really shitty who I definitely shouldn't have dated, I have to say that I remember the qualities that made me attracted to him. they didn't make up for his flaws, but despite him being such a complete shithead not worth dating, some "nice guy" could certainly have observed to an extent what was going on and could have learned from the few things he was doing right and that made me attracted to him. [NEWLINE] [NEWLINE] [STARTQ] People may notice that other people are being nice, but the assumption is that there is an ulterior motive. [ENDQ] [NEWLINE] nope. If a guy is nice to me, I generally assume he's, you know, nice. And I notice. IF, however, he asks me out, I reject him, and he insults me, then that's the moment where it becomes clear that he wasn't really a nice guy, and is only superficially nice to people he wants to fuck. if he were really nice, he'd continue to be nice (or at least neutral) after being rejected (and yes, this is possible and people do this. I've been rejected and continued to be civil, and I've rejected several guys who continued to be civil. this isn't much to ask, and if someone were genuinely nice, it would come very easily to them). [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It seems pretty intuitive that<mask> the first one is wrong,<mask> is the second,<mask> you're generalizing a group of people based on a trait they can't control. You don't choose to be born a male,<mask> it's not fair that you're labeled<mask> a potential rapist with examples like the "bowl of M&amp;M's":<mask> I were to use that example, except having the bowl of M&amp;M's be representative of black people, I would be (rightly) called a Neo-Nazi and told to go fuck a Confederate flag or something. [NEWLINE] [NEWLINE] <mask>, I see the "All women live in fear of being raped" being passed around a lot, and I don't quite see the rationale behind it. Considering that only a minority of men rape, it seems pretty unfair to make a statement that holds all men accountable. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>It seems pretty intuitive that if the first one is wrong, so is the second, since you're generalizing a group of people based on a trait they can't control. You don't choose to be born a male, so it's not fair that you're labeled as a potential rapist with examples like the "bowl of M&amp;M's": If I were to use that example, except having the bowl of M&amp;M's be representative of black people, I would be (rightly) called a Neo-Nazi and told to go fuck a Confederate flag or something. [NEWLINE] [NEWLINE] However, I see the "All women live in fear of being raped" being passed around a lot, and I don't quite see the rationale behind it. Considering that only a minority of men rape, it seems pretty unfair to make a statement that holds all men accountable. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The problem is,<mask> they say that,<mask> they are really feeling is great pride and sense of accomplishment, which is decisively not humility. [ENDQ] [NEWLINE] No,<mask> they are expressing is a sense of being unworthy and the recipient of a gift, a gift of talent, a gift of opportunity, a gift of trust from voters.  Now, it's possible you don't believe<mask> they are saying,<mask> that is<mask> they are saying.  "I'm not worthy, thank you, I'll try not to let it go to my head, to stay humble, to keep my feet on the ground<mask> all these wonderful accolades."  That's<mask> it means to say "I feel humbled." [NEWLINE] [NEWLINE] <mask>, even<mask> the meaning were changing,<mask>'s wrong with language changing?  The word "literally" now means both "literally" and "figuratively,"<mask><mask> many people have used it to mean "figuratively" that it's become part of the language.  Language changes, get over it.  "Terrible" used to mean "inspiring terror," which was a good thing.  "Awful" used to mean "inspiring awe," which was a good thing.  Same with "Fearful."  Which is<mask> the King of England,<mask> St. Paul's Cathedral opened in London in the 1600s after the Great Fire, called it awful, terrible, and fearful -- and those were compliments.  Over time, they became insults.  There's no such thing<mask> "denigrating the meaning of a word."  There may be a lot of hypocrisy out there that should stop,<mask> it has nothing to do with policing language.</s>
Label encoding: <s> [STARTQ] The problem is, when they say that, what they are really feeling is great pride and sense of accomplishment, which is decisively not humility. [ENDQ] [NEWLINE] No, what they are expressing is a sense of being unworthy and the recipient of a gift, a gift of talent, a gift of opportunity, a gift of trust from voters.  Now, it's possible you don't believe what they are saying, but that is what they are saying.  "I'm not worthy, thank you, I'll try not to let it go to my head, to stay humble, to keep my feet on the ground despite all these wonderful accolades."  That's what it means to say "I feel humbled." [NEWLINE] [NEWLINE] However, even if the meaning were changing, what's wrong with language changing?  The word "literally" now means both "literally" and "figuratively," because so many people have used it to mean "figuratively" that it's become part of the language.  Language changes, get over it.  "Terrible" used to mean "inspiring terror," which was a good thing.  "Awful" used to mean "inspiring awe," which was a good thing.  Same with "Fearful."  Which is why the King of England, when St. Paul's Cathedral opened in London in the 1600s after the Great Fire, called it awful, terrible, and fearful -- and those were compliments.  Over time, they became insults.  There's no such thing as "denigrating the meaning of a word."  There may be a lot of hypocrisy out there that should stop, but it has nothing to do with policing language.</s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Advocating a color-blind society is an easy thing to do<mask> one belongs to an empowered race; by ignoring race, you ignore your own race,<mask> naturalizing any privileges that come with being your race. Ignoring races makes it easy to ignore very real issues such<mask> institutionalized racism. To fight racism, you have to talk about race. [NEWLINE] [NEWLINE] I'm going to go a bit off your point here,<mask> I've often heard people say things like, "I wish we wouldn't talk about racism<mask> much, it only divides people." This is a bit different than<mask> you argue,<mask> it's related and I'd like to address it anyway. Most would agree that it's important to talk about our history to avoid making the mistakes we've made in the past. Racism is a part of our history,<mask> we have to talk about race. 50 years ago, you could have looked in a US history textbook and there might only be one paragraph about the history of slavery and segregation. Today, the treatment of minorities groups is a central theme in many textbooks. *<mask> American society has become more racially aware, racism has faded from our culture.* Racism is now stigmatized, and overt prejudice and discrimination is nearly dead. [NEWLINE] [NEWLINE] Now to address your point more directly, American culture is far from monolithic or homogeneous. America is a "melting pot" composed of several sub-cultures. Simply identifying with your nationality means identifying with only one aspect of your identity. People may label themselves<mask> that they can identify with their sub-culture or their heritage. To ask people to only identify themselves<mask> American is to ask them to forget their personal culture and history. </s>
Label encoding: <s>Advocating a color-blind society is an easy thing to do when one belongs to an empowered race; by ignoring race, you ignore your own race, thus naturalizing any privileges that come with being your race. Ignoring races makes it easy to ignore very real issues such as institutionalized racism. To fight racism, you have to talk about race. [NEWLINE] [NEWLINE] I'm going to go a bit off your point here, but I've often heard people say things like, "I wish we wouldn't talk about racism so much, it only divides people." This is a bit different than what you argue, but it's related and I'd like to address it anyway. Most would agree that it's important to talk about our history to avoid making the mistakes we've made in the past. Racism is a part of our history, so we have to talk about race. 50 years ago, you could have looked in a US history textbook and there might only be one paragraph about the history of slavery and segregation. Today, the treatment of minorities groups is a central theme in many textbooks. * As American society has become more racially aware, racism has faded from our culture.* Racism is now stigmatized, and overt prejudice and discrimination is nearly dead. [NEWLINE] [NEWLINE] Now to address your point more directly, American culture is far from monolithic or homogeneous. America is a "melting pot" composed of several sub-cultures. Simply identifying with your nationality means identifying with only one aspect of your identity. People may label themselves so that they can identify with their sub-culture or their heritage. To ask people to only identify themselves as American is to ask them to forget their personal culture and history. </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0600, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Throw pillows take up extra space on the couch and are useless, except for aesthetic reasons.<mask><mask> that a couch can look bare without them,<mask> they are not practical<mask> : [NEWLINE] [NEWLINE] *<mask> you want to sit<mask> the throw pillow is, it makes you feel like the princess and the pea,<mask> it only comforts the middle of your back. [NEWLINE] * They are bad for your posture. [NEWLINE] * They reduce the depth of the couch you can sit on [NEWLINE] [NEWLINE] <mask> you want to keep pillow(s) there for sleeping on,<mask> leave them on the couch<mask> you can put them away and save extra couch space? Obviously,<mask> you have extra couch space and you're just using it<mask> a place for storage, then it's fine,<mask> it remains a problem whenever you have guests over<mask> there is no longer enough couch space. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Throw pillows take up extra space on the couch and are useless, except for aesthetic reasons. I agree that a couch can look bare without them, but they are not practical because : [NEWLINE] [NEWLINE] * If you want to sit where the throw pillow is, it makes you feel like the princess and the pea, since it only comforts the middle of your back. [NEWLINE] * They are bad for your posture. [NEWLINE] * They reduce the depth of the couch you can sit on [NEWLINE] [NEWLINE] If you want to keep pillow(s) there for sleeping on, why leave them on the couch when you can put them away and save extra couch space? Obviously, if you have extra couch space and you're just using it as a place for storage, then it's fine, but it remains a problem whenever you have guests over because there is no longer enough couch space. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Your answer assumes that conformity with "arbitrary rules" doesn't in and of itself facilitate communication. I would take issue with that assumption. [ENDQ] [NEWLINE] <mask> I'm understanding correctly, you're claiming that 'proper grammar' inherently facilitates communication. Does this mean that the proper way to phrase something is always less ambiguous than any other way to say it? And<mask> is it about proper grammar that inherently facilitates communication? [NEWLINE] [NEWLINE] [STARTQ] The "benefit" of imposing "proper grammar" is that it decreases the frequency and magnitude of departures from proper grammar in all situations (both those in which "proper grammar" would facilitate communication, and those in which it wouldn't). [ENDQ] [NEWLINE] <mask> you're saying the benefits of imposing 'proper grammar' are that people will use 'improper grammar' less? Unless you're saying that 'proper grammar' is a self-evident good, I'm not sure<mask> exactly the benefits are, or<mask> the benefits are beneficial. [NEWLINE] [NEWLINE] [STARTQ] <mask> the benefits of normalization that come with "proper grammar" impose the "cost" of having some people corrected for their grammar even<mask> they could be otherwise understood, I don't think that we should be<mask> quick to dismiss those arbitrary rules. [ENDQ] [NEWLINE] Another cost would be that language would be more ambiguous<mask> 'proper grammar' is more ambiguous. [NEWLINE] [NEWLINE] [STARTQ] I don't think that we should be<mask> quick to dismiss those arbitrary rules. [ENDQ] [NEWLINE] I never dismissed them; I said that facilitating communication should be of a higher priority than adhering to proper grammar<mask> facilitating communication is the purpose of grammar. I specifically said "This is not to say that proper grammar is bad."</s>
Label encoding: <s> [STARTQ] Your answer assumes that conformity with "arbitrary rules" doesn't in and of itself facilitate communication. I would take issue with that assumption. [ENDQ] [NEWLINE] If I'm understanding correctly, you're claiming that 'proper grammar' inherently facilitates communication. Does this mean that the proper way to phrase something is always less ambiguous than any other way to say it? And what is it about proper grammar that inherently facilitates communication? [NEWLINE] [NEWLINE] [STARTQ] The "benefit" of imposing "proper grammar" is that it decreases the frequency and magnitude of departures from proper grammar in all situations (both those in which "proper grammar" would facilitate communication, and those in which it wouldn't). [ENDQ] [NEWLINE] So you're saying the benefits of imposing 'proper grammar' are that people will use 'improper grammar' less? Unless you're saying that 'proper grammar' is a self-evident good, I'm not sure what exactly the benefits are, or how the benefits are beneficial. [NEWLINE] [NEWLINE] [STARTQ] If the benefits of normalization that come with "proper grammar" impose the "cost" of having some people corrected for their grammar even when they could be otherwise understood, I don't think that we should be so quick to dismiss those arbitrary rules. [ENDQ] [NEWLINE] Another cost would be that language would be more ambiguous when 'proper grammar' is more ambiguous. [NEWLINE] [NEWLINE] [STARTQ] I don't think that we should be so quick to dismiss those arbitrary rules. [ENDQ] [NEWLINE] I never dismissed them; I said that facilitating communication should be of a higher priority than adhering to proper grammar because facilitating communication is the purpose of grammar. I specifically said "This is not to say that proper grammar is bad."</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The AR15 actually does not overpenetrate<mask> you would expect. It tends to tumble after a few walls of overpenetration.<mask> you should always be aware<mask> direction you are shooting, the tumbling of a round will cause it to lose power quicker than a 12 gauge **slug or 00 Buck.** [ENDQ] [NEWLINE] Sure,<mask> compared to the most over penetrating rounds the 12 gauge has to offer, then, yes, the AR isn't<mask> bad. [NEWLINE] [NEWLINE] [STARTQ] The AR15 is much softer recoiling than a shotgun and easier to control than a handgun. You can create an AR15 that's more compact and easier to use in a house buy buying a Short Barreled Rifle tax stamp from the ATF or buying an AR15 Pistol. [ENDQ] [NEWLINE] You can build a magazine fed shotgun with a pistol grip before even having to deal with the ATF and NFA. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> to accessories that may or may not make the firearm "better" the AR15 platform is extremely modular. The ability to change from a.223/5.56 to.300 blackout to.50 Beowulf to cailbers I've never heard of to meet the exact needs of anyone who needs a rifle makes this one platform good for anyone from farmers to competitive shooters to hardline survivalists. [ENDQ] [NEWLINE] This is a great argument for the AR15. [NEWLINE] [NEWLINE] [STARTQ] I really think that the AR15 is a good rifle that meets the needs and wants of a large number of people. I believe that it is the everyman's firearm and helps lessen the number of physical firearms that someone would need just by being modular. [ENDQ] [NEWLINE] <mask> a good argument. </s>
Label encoding: <s> [STARTQ] The AR15 actually does not overpenetrate as you would expect. It tends to tumble after a few walls of overpenetration. While you should always be aware what direction you are shooting, the tumbling of a round will cause it to lose power quicker than a 12 gauge **slug or 00 Buck.** [ENDQ] [NEWLINE] Sure, when compared to the most over penetrating rounds the 12 gauge has to offer, then, yes, the AR isn't as bad. [NEWLINE] [NEWLINE] [STARTQ] The AR15 is much softer recoiling than a shotgun and easier to control than a handgun. You can create an AR15 that's more compact and easier to use in a house buy buying a Short Barreled Rifle tax stamp from the ATF or buying an AR15 Pistol. [ENDQ] [NEWLINE] You can build a magazine fed shotgun with a pistol grip before even having to deal with the ATF and NFA. [NEWLINE] [NEWLINE] [STARTQ] In addition to accessories that may or may not make the firearm "better" the AR15 platform is extremely modular. The ability to change from a.223/5.56 to.300 blackout to.50 Beowulf to cailbers I've never heard of to meet the exact needs of anyone who needs a rifle makes this one platform good for anyone from farmers to competitive shooters to hardline survivalists. [ENDQ] [NEWLINE] This is a great argument for the AR15. [NEWLINE] [NEWLINE] [STARTQ] I really think that the AR15 is a good rifle that meets the needs and wants of a large number of people. I believe that it is the everyman's firearm and helps lessen the number of physical firearms that someone would need just by being modular. [ENDQ] [NEWLINE] Also a good argument. </s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Have you considered that the quicker business cycle may have something to do with the quicker speed of the economy overall? With improved communication (and especially automation of trading), panics can spread more rapidly, which leads to the bursting of bubbles. [NEWLINE] [NEWLINE] <mask> for their creation, using the definition of a bubble<mask> an economy that's heavily invested in an overvalued product/asset, wouldn't the solutions to bubbles be improved valuations standards? From your previous arguments, I'm assuming you'll say consumers can provide the basis for the improved valuations,<mask><mask><mask> the consumers don't know enough about macroeconomics to be able to understand valuations? (<mask> an example,<mask> the public often freaks out about debt,<mask> pays little attention to the debt-to-GDP ratio, which is a better indicator). [NEWLINE] [NEWLINE] <mask> consumers aren't informed enough about macroeconomics to understand incorrect valuations, who's a nonpartisan entity who can be trusted? I'd personally argue for increased regulation (well, more of a return to old levels of regulation). [NEWLINE] [NEWLINE] <mask> would be your source of consumer protection? Would the capitalists be expected to forgo some profit to ensure their industry is sustainable, or are you implying that through bitcoin consumers can be free? [NEWLINE] [NEWLINE] <mask> would you respond to claims that bitcoin is currently a speculative bubble?<mask> I've heard an influx of Chinese 'clients' may provide stability to the currency, it won't keep climbing indefinitely, and<mask> it seems to be a speculative investment equilivent to gold,<mask> will happen to it<mask> it inevitably comes back down? There's no FDIC for it,<mask><mask> 'll prevent a "run on bitcoin"?</s>
Label encoding: <s>Have you considered that the quicker business cycle may have something to do with the quicker speed of the economy overall? With improved communication (and especially automation of trading), panics can spread more rapidly, which leads to the bursting of bubbles. [NEWLINE] [NEWLINE] As for their creation, using the definition of a bubble as an economy that's heavily invested in an overvalued product/asset, wouldn't the solutions to bubbles be improved valuations standards? From your previous arguments, I'm assuming you'll say consumers can provide the basis for the improved valuations, but what if the consumers don't know enough about macroeconomics to be able to understand valuations? ( As an example, how the public often freaks out about debt, but pays little attention to the debt-to-GDP ratio, which is a better indicator). [NEWLINE] [NEWLINE] If consumers aren't informed enough about macroeconomics to understand incorrect valuations, who's a nonpartisan entity who can be trusted? I'd personally argue for increased regulation (well, more of a return to old levels of regulation). [NEWLINE] [NEWLINE] What would be your source of consumer protection? Would the capitalists be expected to forgo some profit to ensure their industry is sustainable, or are you implying that through bitcoin consumers can be free? [NEWLINE] [NEWLINE] How would you respond to claims that bitcoin is currently a speculative bubble? While I've heard an influx of Chinese 'clients' may provide stability to the currency, it won't keep climbing indefinitely, and as it seems to be a speculative investment equilivent to gold, what will happen to it when it inevitably comes back down? There's no FDIC for it, so what 'll prevent a "run on bitcoin"?</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I appreciate that you gave a delta,<mask> you brought up an interesting idea about Batman that got me thinking. [NEWLINE] [NEWLINE] Essentially, it seems like you are asserting that Batman,<mask> a mortal man, is actually a superhuman in that he has superhuman powers of tenacity. True, Batman's thirst for knowing could be his undoing against an Elder God,<mask> consider that his superhuman tenacity would allow him to enlist others in his struggle. Perhaps he convinces another Elder God to fight his fight for him against the other god,<mask> he uses his powers of persuation to 'work his way' up the chain (he convinces Superman, Superman talks to Dr. Manhattan, who talks to Apocalypse, and<mask> on). In other words,<mask> given the proper time to contemplate the problem, he is able to adapt and overcome any attack, even sneaky sinister attacks that play on Batman's thirst for knowing. Perhaps Cthulhu allows him to get away after the first attack, then lets him go only to feed off his thirst for knowing and allow Batman to undo himself.<mask> assume that Batman is able to take enough time, he actually discovers that Cthulhu is doing this and develops a countermeasure.  Batman's tenacity wins. [NEWLINE] [NEWLINE] It seems to me then, that you are submitting that Batman's supreme power of tenacity and will, has the property of ALWAYS allowing him to defeat his foe, no matter the foes power. [NEWLINE] [NEWLINE] You've convinced me good sir. Batman cannot be defeated by anyone,<mask> he is able to withstand the initial attack and go back to use his supreme powers of tenacity and clever oneupsmanship to will him into finding a solution to the problem. </s>
Label encoding: <s>I appreciate that you gave a delta, but you brought up an interesting idea about Batman that got me thinking. [NEWLINE] [NEWLINE] Essentially, it seems like you are asserting that Batman, while a mortal man, is actually a superhuman in that he has superhuman powers of tenacity. True, Batman's thirst for knowing could be his undoing against an Elder God, but consider that his superhuman tenacity would allow him to enlist others in his struggle. Perhaps he convinces another Elder God to fight his fight for him against the other god, because he uses his powers of persuation to 'work his way' up the chain (he convinces Superman, Superman talks to Dr. Manhattan, who talks to Apocalypse, and so on). In other words, if given the proper time to contemplate the problem, he is able to adapt and overcome any attack, even sneaky sinister attacks that play on Batman's thirst for knowing. Perhaps Cthulhu allows him to get away after the first attack, then lets him go only to feed off his thirst for knowing and allow Batman to undo himself. But assume that Batman is able to take enough time, he actually discovers that Cthulhu is doing this and develops a countermeasure.  Batman's tenacity wins. [NEWLINE] [NEWLINE] It seems to me then, that you are submitting that Batman's supreme power of tenacity and will, has the property of ALWAYS allowing him to defeat his foe, no matter the foes power. [NEWLINE] [NEWLINE] You've convinced me good sir. Batman cannot be defeated by anyone, if he is able to withstand the initial attack and go back to use his supreme powers of tenacity and clever oneupsmanship to will him into finding a solution to the problem. </s>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Occupy Wall Street and the Occupy movement in general was not<mask> ineffectual<mask> many people seem to assume. [NEWLINE] [NEWLINE] * There's powerful symbolism in the movement... much more<mask> than your run-of-the-mill street protest or march, the idea of camping out on Wall St. captured the public imagination and held it for several years. We're still talking about it. A powerful, clear symbol can crystallize public action. [NEWLINE] [NEWLINE] * The Occupy movement<mask> has made or led to concrete steps and change. The organizers have been promoting a ["Community Bill of Rights"]( [URL] ), which is a specific set of laws and amendments intended to reverse decisions like *Citizens United* and make it<mask> the wealthy have to obey the same laws<mask> everyone else. Occupy groups are pushing for these legal changes at the city, state and federal level. [NEWLINE] [NEWLINE] * [Occupy Homes]( [URL] ) specifically targets banks threatening foreclosure unfairly, aiming both to rescue homeowners and to change mortgage lending policy in that state and nationwide. [NEWLINE] [NEWLINE] * [Strike Debt]( [URL] #Strike_Debt) an OWS offshoot, is a group that targets unfair/excessive lending practices. One of their main techniques is to buy loans on the loan auction market, and then forgive the borrower's debt.<mask> of Sept. 2014 they had wiped out $19 million in student loans this way. [NEWLINE] [NEWLINE] * There are a number of other specific policy proposals developed by offshoots of the Occupy movement, like concrete plans for a "community banking" system, "Occupy Sandy" created a system for helping communities hit by disasters who get bypassed by FEMA and state aid agencies, etc. </s>
Label encoding: <s>Occupy Wall Street and the Occupy movement in general was not as ineffectual as many people seem to assume. [NEWLINE] [NEWLINE] * There's powerful symbolism in the movement... much more so than your run-of-the-mill street protest or march, the idea of camping out on Wall St. captured the public imagination and held it for several years. We're still talking about it. A powerful, clear symbol can crystallize public action. [NEWLINE] [NEWLINE] * The Occupy movement also has made or led to concrete steps and change. The organizers have been promoting a ["Community Bill of Rights"]( [URL] ), which is a specific set of laws and amendments intended to reverse decisions like *Citizens United* and make it so the wealthy have to obey the same laws as everyone else. Occupy groups are pushing for these legal changes at the city, state and federal level. [NEWLINE] [NEWLINE] * [Occupy Homes]( [URL] ) specifically targets banks threatening foreclosure unfairly, aiming both to rescue homeowners and to change mortgage lending policy in that state and nationwide. [NEWLINE] [NEWLINE] * [Strike Debt]( [URL] #Strike_Debt) an OWS offshoot, is a group that targets unfair/excessive lending practices. One of their main techniques is to buy loans on the loan auction market, and then forgive the borrower's debt. As of Sept. 2014 they had wiped out $19 million in student loans this way. [NEWLINE] [NEWLINE] * There are a number of other specific policy proposals developed by offshoots of the Occupy movement, like concrete plans for a "community banking" system, "Occupy Sandy" created a system for helping communities hit by disasters who get bypassed by FEMA and state aid agencies, etc. </s>
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason that diamonds are chosen for engagement rings is tied to the price. At the time the engagement ring became popular people starting families were growing into mostly one income homes, with a primary bread winner and a housewife. To symbolize taking on this additional expense and staying comfortable the future breadwinner would reduce his spending for an amount of time and spend the money saved on the engagement ring.<mask>, at the time, the engagement ring served<mask> a symbol of sacrifice. [NEWLINE] [NEWLINE] <mask> then, times have changed. Families now for the most part require two incomes to stay afloat.<mask> social pressures still push for the man to be the breadwinner, it is not quite<mask> strong<mask> it was. The sacrifice is less important than it was in a literal sense.<mask>, either spouse may still need to spend from their income in dire situations, or to help their spouse.<mask> one requires medical care, or intends to begin or return to higher education the cost for the family will increase, and it will likely cause the total income to decrease. In families intending to have children this is not only a possibility<mask> an eventuality, either through actual pregnancy or the difficulties that can arise<mask> failing to conceive. [NEWLINE] [NEWLINE] <mask>, this combined with tradition would be the reason that diamonds are usually chosen for engagement rings.<mask><mask><mask> "Diamonds are the inferior gemstone" there are categories<mask> other stones are superior,<mask> there are categories<mask> the diamond is superior. Carbon content by mass, hardness, and price come to mind.<mask> with that I will counter your statement with a declaration of my own. [NEWLINE] [NEWLINE] &gt;Diamonds are not uniformally superior or inferior<mask> a gemstone.</s>
Label encoding: <s>The reason that diamonds are chosen for engagement rings is tied to the price. At the time the engagement ring became popular people starting families were growing into mostly one income homes, with a primary bread winner and a housewife. To symbolize taking on this additional expense and staying comfortable the future breadwinner would reduce his spending for an amount of time and spend the money saved on the engagement ring. So, at the time, the engagement ring served as a symbol of sacrifice. [NEWLINE] [NEWLINE] Since then, times have changed. Families now for the most part require two incomes to stay afloat. While social pressures still push for the man to be the breadwinner, it is not quite as strong as it was. The sacrifice is less important than it was in a literal sense. However, either spouse may still need to spend from their income in dire situations, or to help their spouse. If one requires medical care, or intends to begin or return to higher education the cost for the family will increase, and it will likely cause the total income to decrease. In families intending to have children this is not only a possibility but an eventuality, either through actual pregnancy or the difficulties that can arise when failing to conceive. [NEWLINE] [NEWLINE] So, this combined with tradition would be the reason that diamonds are usually chosen for engagement rings. As far as "Diamonds are the inferior gemstone" there are categories where other stones are superior, but there are categories where the diamond is superior. Carbon content by mass, hardness, and price come to mind. So with that I will counter your statement with a declaration of my own. [NEWLINE] [NEWLINE] &gt;Diamonds are not uniformally superior or inferior as a gemstone.</s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To preface this, I am on mobile at work and have not reread all my political theory for a little<mask>. [NEWLINE] [NEWLINE] I will try and address the logic and reasoning the founders employed for making the amendment process difficult. [NEWLINE] [NEWLINE] 1) The degree of "virtue" (<mask> defined by Machiavelli) is highest at the founding of a nation. Essentially,<mask> he is saying is the nation will be most aware of the difficulties of governance, dangers and benefits of power, the level of morality (which he deems very important) and the relationship between citizens and government. In other words, the inevitable creep of corruption will be lowest at the outset. [NEWLINE] [NEWLINE] 2) The idea of basic human rights that are meant to be eternal and unalienable are thought to be the foundations upon which all other rights are derived from.<mask> those are protected then all other rights come from those. [NEWLINE] [NEWLINE] 3) The Constitution actually depends on being vague. It allows each successive generation to understand and interpret it for the needs of that generation. By doing this, the constitution can remain steady longer thereby insuring the longevity of the country (at least before it destroys itself either through corruption or bad military tactics - again Machiavelli). [NEWLINE] [NEWLINE] 3.1) The founders actually did have foresight by creating a process to amend the constitution. Just slowly, which isn't a bad thing (see the mobocracy comment). [NEWLINE] [NEWLINE] 4) To preserve the Constitution for the longest time possible (and reduce the damaging effects of corruption) it is best to limit changes, this may make it "backward"<mask><mask> the document is vague, the current generation can reinterpret it.</s>
Label encoding: <s>To preface this, I am on mobile at work and have not reread all my political theory for a little while. [NEWLINE] [NEWLINE] I will try and address the logic and reasoning the founders employed for making the amendment process difficult. [NEWLINE] [NEWLINE] 1) The degree of "virtue" ( as defined by Machiavelli) is highest at the founding of a nation. Essentially, what he is saying is the nation will be most aware of the difficulties of governance, dangers and benefits of power, the level of morality (which he deems very important) and the relationship between citizens and government. In other words, the inevitable creep of corruption will be lowest at the outset. [NEWLINE] [NEWLINE] 2) The idea of basic human rights that are meant to be eternal and unalienable are thought to be the foundations upon which all other rights are derived from. If those are protected then all other rights come from those. [NEWLINE] [NEWLINE] 3) The Constitution actually depends on being vague. It allows each successive generation to understand and interpret it for the needs of that generation. By doing this, the constitution can remain steady longer thereby insuring the longevity of the country (at least before it destroys itself either through corruption or bad military tactics - again Machiavelli). [NEWLINE] [NEWLINE] 3.1) The founders actually did have foresight by creating a process to amend the constitution. Just slowly, which isn't a bad thing (see the mobocracy comment). [NEWLINE] [NEWLINE] 4) To preserve the Constitution for the longest time possible (and reduce the damaging effects of corruption) it is best to limit changes, this may make it "backward" but because the document is vague, the current generation can reinterpret it.</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> (correct me<mask> I'm wrong)<mask> it seems your argument for there not being a problem indicates that the government will always make decisions that ultimately benefit the greater good,<mask> there may be moral objections. I can think of times<mask> it does,<mask> I can<mask> think of many times<mask> it doesn't. Are we under a social contract to surrender staunch morals in some cases in hopes (often ill-conceived hopes) that the government will make the best decision? [ENDQ] [NEWLINE] No. I'm saying that the argument "Well, the government should let people opt out of X for religious objections or else people will be forced to violate their own moral codes!" is not convincing,<mask> all government action does this.<mask> we allow governments to exist at all, then we implicitly accept that there are some moral beliefs that are okay to impose on people against their will. Once you've accepted that, then it's just a matter of determining which beliefs are that important. [NEWLINE] [NEWLINE] Governments may impose the wrong beliefs. The Soviet Union might require compulsory conscription to invade Lithuania; this is bad,<mask> people have a right to a democratic government, and this violates that right.<mask> the Soviets were to say "<mask> you're a communist party member and have deep objections, we won't draft you", I'd say that the logic of such an exception would apply equally to non-communists too.<mask> that exception isn't extended, the Soviets would be doing something bad,<mask> it wouldn't be bad<mask> government imposition of morality is inherently evil; it's bad<mask> nobody should be drafted to replace democracy with imperialism, and communist party membership has no legitimate relationship to that. </s>
Label encoding: <s> [STARTQ] Also (correct me if I'm wrong) but it seems your argument for there not being a problem indicates that the government will always make decisions that ultimately benefit the greater good, though there may be moral objections. I can think of times where it does, but I can also think of many times where it doesn't. Are we under a social contract to surrender staunch morals in some cases in hopes (often ill-conceived hopes) that the government will make the best decision? [ENDQ] [NEWLINE] No. I'm saying that the argument "Well, the government should let people opt out of X for religious objections or else people will be forced to violate their own moral codes!" is not convincing, because all government action does this. If we allow governments to exist at all, then we implicitly accept that there are some moral beliefs that are okay to impose on people against their will. Once you've accepted that, then it's just a matter of determining which beliefs are that important. [NEWLINE] [NEWLINE] Governments may impose the wrong beliefs. The Soviet Union might require compulsory conscription to invade Lithuania; this is bad, because people have a right to a democratic government, and this violates that right. If the Soviets were to say " If you're a communist party member and have deep objections, we won't draft you", I'd say that the logic of such an exception would apply equally to non-communists too. If that exception isn't extended, the Soviets would be doing something bad, but it wouldn't be bad because government imposition of morality is inherently evil; it's bad because nobody should be drafted to replace democracy with imperialism, and communist party membership has no legitimate relationship to that. </s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. English,<mask><mask><mask> any scholar can tell, is changing just<mask> it has been<mask> its wee beginnings in Pre-Proto-Indo-European thousands and thousands of years ago. There is certainly evidence that it is still changing. [NEWLINE] [NEWLINE] <mask><mask> we need to separate "written English" and "the English language" in order to deal with your argument. "written English" is a human invention: it is constructed, taught, guided, and reliant on materials and resources. It's a result of writing's usefulness<mask> a method of transferring ideas, and it will continue to function<mask><mask><mask> it's practical. [NEWLINE] [NEWLINE] 2. The goal of language is communication, yes. Language change is a universal, almost-constant process. English will probably inevitably split into a variety of mutually unintelligible new languages, just<mask> Latin, Chinese, Arabic, or any other language in history did. It is not a bad thing, and does not significantly impede the process of societal and scientific change. Consistency in language is not a result of grammar nazis' collective efforts,<mask> rather a result of the necessity of communication. [NEWLINE] [NEWLINE] The video is a further example of the spoken vs. written phenomenon. Spoken language is<mask> it is and will probably always be,<mask><mask><mask> we know: a changing thing that diverges and becomes mutually unintelligible with distance (social, economic, or geographic) and time. The almost unintelligibility of the person in the video is a result of hundreds of years of separation, and I expect in the next couple hundred years "the English language" will be<mask> much an arbitrary political construct<mask> Arabic.</s>
Label encoding: <s>1. English, as far as any scholar can tell, is changing just as it has been since its wee beginnings in Pre-Proto-Indo-European thousands and thousands of years ago. There is certainly evidence that it is still changing. [NEWLINE] [NEWLINE] I think we need to separate "written English" and "the English language" in order to deal with your argument. "written English" is a human invention: it is constructed, taught, guided, and reliant on materials and resources. It's a result of writing's usefulness as a method of transferring ideas, and it will continue to function as long as it's practical. [NEWLINE] [NEWLINE] 2. The goal of language is communication, yes. Language change is a universal, almost-constant process. English will probably inevitably split into a variety of mutually unintelligible new languages, just as Latin, Chinese, Arabic, or any other language in history did. It is not a bad thing, and does not significantly impede the process of societal and scientific change. Consistency in language is not a result of grammar nazis' collective efforts, but rather a result of the necessity of communication. [NEWLINE] [NEWLINE] The video is a further example of the spoken vs. written phenomenon. Spoken language is what it is and will probably always be, as far as we know: a changing thing that diverges and becomes mutually unintelligible with distance (social, economic, or geographic) and time. The almost unintelligibility of the person in the video is a result of hundreds of years of separation, and I expect in the next couple hundred years "the English language" will be as much an arbitrary political construct as Arabic.</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I cannot believe that the outer districts would simply stand for being forced to sacrifice children from their own communities for the sake of a sporting event whose very overt purpose is solely for the entertainment of the elite and the demoralization of the other districts. [ENDQ] [NEWLINE] Terrorist leaders using women and children<mask> suicide bombers with the goal of increasing their own political power [NEWLINE] [NEWLINE] [STARTQ] I don't believe the elite could suddenly become<mask> depraved that they would relish and cheer to watch children brutally murder one another, even<mask> they were somewhat depraved. There is just a cognitive dissonance between their own sense of humanity and desire to live, and their joyous reception of the participants in the games, immediately followed by their cheering on their murdering one another. I can't get past that. I feel like there is meant to be a social commentary here,<mask> I can't figure out<mask> it is, perhaps<mask> someone can point me to it it would help. [ENDQ] [NEWLINE] The Coliseum in Roman times [NEWLINE] [NEWLINE] [STARTQ] Within the games themselves, the way certain groups of tributes gleefully ally with one another and hunt down others seems unbelievable. Alliances probably would happen,<mask> they would be reluctant and untrusting of one another<mask> at the end of the day, everyone knows there can be only one survivor and that at some point they would be murdered by their own ally. I'm not a game theorist<mask> I'm pretty sure it wouldn't happen<mask> depicted in the story. [ENDQ] [NEWLINE] I don't think it's that unbelievable that people would revert to primal Machiavellian logic<mask> they were put in a situation<mask> they were being hunted.  Animals will do pretty much anything for survival.</s>
Label encoding: <s> [STARTQ] I cannot believe that the outer districts would simply stand for being forced to sacrifice children from their own communities for the sake of a sporting event whose very overt purpose is solely for the entertainment of the elite and the demoralization of the other districts. [ENDQ] [NEWLINE] Terrorist leaders using women and children as suicide bombers with the goal of increasing their own political power [NEWLINE] [NEWLINE] [STARTQ] I don't believe the elite could suddenly become so depraved that they would relish and cheer to watch children brutally murder one another, even if they were somewhat depraved. There is just a cognitive dissonance between their own sense of humanity and desire to live, and their joyous reception of the participants in the games, immediately followed by their cheering on their murdering one another. I can't get past that. I feel like there is meant to be a social commentary here, but I can't figure out what it is, perhaps if someone can point me to it it would help. [ENDQ] [NEWLINE] The Coliseum in Roman times [NEWLINE] [NEWLINE] [STARTQ] Within the games themselves, the way certain groups of tributes gleefully ally with one another and hunt down others seems unbelievable. Alliances probably would happen, but they would be reluctant and untrusting of one another since at the end of the day, everyone knows there can be only one survivor and that at some point they would be murdered by their own ally. I'm not a game theorist but I'm pretty sure it wouldn't happen as depicted in the story. [ENDQ] [NEWLINE] I don't think it's that unbelievable that people would revert to primal Machiavellian logic if they were put in a situation where they were being hunted.  Animals will do pretty much anything for survival.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not sure<mask> your response is meant to be sarcastic,<mask> I'll pretend it isn't. [NEWLINE] [NEWLINE] Yes, I'm in the USA.  Boston to be specific. [NEWLINE] [NEWLINE] I know some people are big believers in not having children<mask> of all the kids that need to be adopted.  Makes no sense to me.  Don't get me wrong, I feel bad that there are kids out there without families,<mask> they're not my child.  My child looks a little like me and a little like my wife.  He's a continuation of my DNA.  He's tall like me (for a 7 week old), he has my dimpled chin, and he has my wife's nose.  An adopted child isn't my DNA and isn't my lineage.  Now<mask> you can't physically have children (or don't want to) adopting is a fantastic choice,<mask> to say that you shouldn't have a child of your own<mask> there are thousands of orphaned children in the country is asinine. [NEWLINE] [NEWLINE] The capitalists to prey on and exploit?  Are you saying free market economies are a bad thing?  Chasing profits are<mask> the lights are still on at Reddit.com,<mask> the iPhone was invented, and<mask> cancer medications were developed! [NEWLINE] [NEWLINE] I understand that I'm going to spend hundreds of thousands of dollars to raise a child and that there are hungry people out there,<mask> it's absurd to say that I shouldn't use my money to have a child<mask> of those hungry people. <mask> can't they work to feed themselves like we've all done.  Same goes for the impoverished nations without drinking water.</s>
Label encoding: <s>Not sure if your response is meant to be sarcastic, but I'll pretend it isn't. [NEWLINE] [NEWLINE] Yes, I'm in the USA.  Boston to be specific. [NEWLINE] [NEWLINE] I know some people are big believers in not having children because of all the kids that need to be adopted.  Makes no sense to me.  Don't get me wrong, I feel bad that there are kids out there without families, but they're not my child.  My child looks a little like me and a little like my wife.  He's a continuation of my DNA.  He's tall like me (for a 7 week old), he has my dimpled chin, and he has my wife's nose.  An adopted child isn't my DNA and isn't my lineage.  Now if you can't physically have children (or don't want to) adopting is a fantastic choice, but to say that you shouldn't have a child of your own because there are thousands of orphaned children in the country is asinine. [NEWLINE] [NEWLINE] The capitalists to prey on and exploit?  Are you saying free market economies are a bad thing?  Chasing profits are why the lights are still on at Reddit.com, why the iPhone was invented, and why cancer medications were developed! [NEWLINE] [NEWLINE] I understand that I'm going to spend hundreds of thousands of dollars to raise a child and that there are hungry people out there, but it's absurd to say that I shouldn't use my money to have a child because of those hungry people.  Why can't they work to feed themselves like we've all done.  Same goes for the impoverished nations without drinking water.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0571, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Though my personal political views match those of Bernie Sanders much more than Hillary Clinton, I believe nominating Bernie could ultimately be very damaging for the progressive cause. My main concern with nominating Sanders is that he will alienate all voters except for the left and the end result will be a 48-state Reagan v. Mondale style steamrolling. A GOP presidency, possible supermajority in both houses and 1-3 supreme court nominations could be enough to nullify the limited progress we've had under Obama. [NEWLINE] [NEWLINE] A Clinton presidency would probably look very similar to Obama's, and come with a lot of disappointments. I do think it's possible that we might see progress towards things like a single-payer healthcare system that progressives have wanted all along. [NEWLINE] [NEWLINE] I like Bernie a lot, and would like to vote for him<mask> please CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Though my personal political views match those of Bernie Sanders much more than Hillary Clinton, I believe nominating Bernie could ultimately be very damaging for the progressive cause. My main concern with nominating Sanders is that he will alienate all voters except for the left and the end result will be a 48-state Reagan v. Mondale style steamrolling. A GOP presidency, possible supermajority in both houses and 1-3 supreme court nominations could be enough to nullify the limited progress we've had under Obama. [NEWLINE] [NEWLINE] A Clinton presidency would probably look very similar to Obama's, and come with a lot of disappointments. I do think it's possible that we might see progress towards things like a single-payer healthcare system that progressives have wanted all along. [NEWLINE] [NEWLINE] I like Bernie a lot, and would like to vote for him so please CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I strongly disagree that it should *never* be done<mask> *anyone* is not okay with it. I don't think everyone should have veto power-- I'm okay with majority vote, possibly requiring higher-than-majority approval (eg 3/4 or 2/3 approval). [NEWLINE] [NEWLINE] I'd argue it's a simple matter of net outcome.<mask> most people around you enjoy the music, it should be kept on, even<mask> a few people are annoyed. I don't think we should spend our lives trying to satisfy the least happy/most complainy people,<mask> instead try to increase happiness. [NEWLINE] [NEWLINE] Now-- in practice, our ideas sometimes come out the same. That dickhead at McDonald's listening to his shitty rap-rock through his smartphone tinny speakers should turn that shit off,<mask> everyone except his asshole friends is annoyed by it. [NEWLINE] [NEWLINE] <mask><mask> you apply it at work, you may get an opposite effect. I always hated<mask> most people at a workplace enjoyed listening to the radio,<mask><mask> just *one* person didn't like it, the boss would have us turn it off. I definitely would have been in favor with alternating between on and off,<mask> this idea that everyone in the office holds absolute veto power over something that has the power to increase the happiness and productivity of everyone else...<mask><mask> that's bull. [NEWLINE] [NEWLINE] I'd generally agree that we should weight the opinions of the people who don't like the music/video slightly higher than the people who do-- after all, being annoyed is slightly more negative of an emotion than "not being amused" is positive.<mask> definitely not to the degree you assert.</s>
Label encoding: <s>I strongly disagree that it should *never* be done if *anyone* is not okay with it. I don't think everyone should have veto power-- I'm okay with majority vote, possibly requiring higher-than-majority approval (eg 3/4 or 2/3 approval). [NEWLINE] [NEWLINE] I'd argue it's a simple matter of net outcome. If most people around you enjoy the music, it should be kept on, even if a few people are annoyed. I don't think we should spend our lives trying to satisfy the least happy/most complainy people, but instead try to increase happiness. [NEWLINE] [NEWLINE] Now-- in practice, our ideas sometimes come out the same. That dickhead at McDonald's listening to his shitty rap-rock through his smartphone tinny speakers should turn that shit off, because everyone except his asshole friends is annoyed by it. [NEWLINE] [NEWLINE] But if you apply it at work, you may get an opposite effect. I always hated when most people at a workplace enjoyed listening to the radio, but if just *one* person didn't like it, the boss would have us turn it off. I definitely would have been in favor with alternating between on and off, but this idea that everyone in the office holds absolute veto power over something that has the power to increase the happiness and productivity of everyone else... I think that's bull. [NEWLINE] [NEWLINE] I'd generally agree that we should weight the opinions of the people who don't like the music/video slightly higher than the people who do-- after all, being annoyed is slightly more negative of an emotion than "not being amused" is positive. But definitely not to the degree you assert.</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL] /[1] [NEWLINE] [NEWLINE] The only reason this video is on the front page is<mask> men feel truly oppressed by<mask> they perceive to be a double standard that says they can't hit women even<mask> they're being attacked.<mask> you doubt that this video is purely about gender roles, ask yourself<mask> it would be on the front page<mask> all the parties involved were male. [NEWLINE] [NEWLINE] By feeling oppressed by this dynamic,<mask><mask> most men are fundamentally failing to understand~~ a)<mask> much more common it is for men to be violent towards women,~~ b) the culture of fear and intimidation that such violence breeds, and c)<mask> ridiculous it makes you look to cheer and go "Right on!" in the relatively rare case<mask> a man is justified in hitting a woman. [NEWLINE] [NEWLINE] Anyone remember the last time a story about a woman defending herself against a violent man was on the front page? Me neither. [NEWLINE] [NEWLINE] EDIT: A few people have accurately pointed out that the frequency of cases of women perpetrating violence against men is roughly the same<mask> men perpetrating violence against women. Others have<mask> pointed out that<mask> makes violence against women more pernicious is the greater damage that men are able to inflict. In making the initial point, I was attempting to demonstrate that violence is a way of asserting power or control, and in the case of men committing violence against women, that power dynamic is shifted heavily in their favor. The overall point about men equating violence against women with violence against men stands -- it is not a valid comparison, and most of the men making it seem to be fundamentally failing to understand<mask> those power dynamics make the two situations quite different.</s>
Label encoding: <s> [URL] /[1] [NEWLINE] [NEWLINE] The only reason this video is on the front page is because men feel truly oppressed by what they perceive to be a double standard that says they can't hit women even if they're being attacked. If you doubt that this video is purely about gender roles, ask yourself if it would be on the front page if all the parties involved were male. [NEWLINE] [NEWLINE] By feeling oppressed by this dynamic, I think most men are fundamentally failing to understand~~ a) how much more common it is for men to be violent towards women,~~ b) the culture of fear and intimidation that such violence breeds, and c) how ridiculous it makes you look to cheer and go "Right on!" in the relatively rare case where a man is justified in hitting a woman. [NEWLINE] [NEWLINE] Anyone remember the last time a story about a woman defending herself against a violent man was on the front page? Me neither. [NEWLINE] [NEWLINE] EDIT: A few people have accurately pointed out that the frequency of cases of women perpetrating violence against men is roughly the same as men perpetrating violence against women. Others have also pointed out that what makes violence against women more pernicious is the greater damage that men are able to inflict. In making the initial point, I was attempting to demonstrate that violence is a way of asserting power or control, and in the case of men committing violence against women, that power dynamic is shifted heavily in their favor. The overall point about men equating violence against women with violence against men stands -- it is not a valid comparison, and most of the men making it seem to be fundamentally failing to understand how those power dynamics make the two situations quite different.</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The "rights" to Life, Liberty, and Property are the axioms under which most social contracts have their theoretical basis. These rights are usually poorly defined and in no sense absolute. Obvious examples include: death penalty, prison, and imminent domain. These exceptions are powers granted to the state which give them easily abused<mask> often vital ability to arbitrate disputes and facilitate commerce. Locking up dangerous criminals, and imminent domain to build roads are an obvious example here. [NEWLINE] [NEWLINE] I would contend that any society which allows contracts to violate any of the rights to life, liberty or right to own property would quickly degenerate into oligarchy/aristocracy.<mask> money allows you to make other people your legal slave or to kill them, the rich become the official, not merely de-facto rulers of your society, especially<mask> you apportion the votes of slaves to slave owners/slave owning territory. [NEWLINE] [NEWLINE] <mask> I would agree that the terms rights is perhaps a bit strong. Instead it's a priority system,<mask> most states have the following priority: [NEWLINE] [NEWLINE] 0. The general welfare (broad,<mask> flexible) [NEWLINE] 1. The life of each citizen [NEWLINE] 2. The property rights of each citizen [NEWLINE] 3. The liberty of each citizen [NEWLINE] 4. Other laws [NEWLINE] [NEWLINE] General welfare can trump any right, life trumps property (except in stand your ground<mask> the reverse is true), property trumps liberty (steal and go to jail). [NEWLINE] [NEWLINE] <mask> yeah, they are pretty much values instead of rights,<mask><mask> you don't put those values high in your social contract, everyone ends up indentured to the rich which is a Bad Thing.</s>
Label encoding: <s>The "rights" to Life, Liberty, and Property are the axioms under which most social contracts have their theoretical basis. These rights are usually poorly defined and in no sense absolute. Obvious examples include: death penalty, prison, and imminent domain. These exceptions are powers granted to the state which give them easily abused but often vital ability to arbitrate disputes and facilitate commerce. Locking up dangerous criminals, and imminent domain to build roads are an obvious example here. [NEWLINE] [NEWLINE] I would contend that any society which allows contracts to violate any of the rights to life, liberty or right to own property would quickly degenerate into oligarchy/aristocracy. When money allows you to make other people your legal slave or to kill them, the rich become the official, not merely de-facto rulers of your society, especially if you apportion the votes of slaves to slave owners/slave owning territory. [NEWLINE] [NEWLINE] So I would agree that the terms rights is perhaps a bit strong. Instead it's a priority system, where most states have the following priority: [NEWLINE] [NEWLINE] 0. The general welfare (broad, but flexible) [NEWLINE] 1. The life of each citizen [NEWLINE] 2. The property rights of each citizen [NEWLINE] 3. The liberty of each citizen [NEWLINE] 4. Other laws [NEWLINE] [NEWLINE] General welfare can trump any right, life trumps property (except in stand your ground where the reverse is true), property trumps liberty (steal and go to jail). [NEWLINE] [NEWLINE] So yeah, they are pretty much values instead of rights, but if you don't put those values high in your social contract, everyone ends up indentured to the rich which is a Bad Thing.</s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All favourites are preferences *and* logically exclusive. [NEWLINE] [NEWLINE] A "favourite" is a thing that's been isolated from a list of similar things that share a common characteristic<mask> the one you like best - to the exclusion of 2nd, 3rd and other members of that list. You can't have *all* flavours of icecream<mask> your favourite icecream flavour. [NEWLINE] [NEWLINE] A preference is a greater than-less than comparison, a greater liking for one over the other. [NEWLINE] [NEWLINE] You can like and be friends with many people -<mask> they must exclude the set of people you don't like or value. [NEWLINE] [NEWLINE] You can love and be lovers with many people too -<mask> they must exclude the set you don't love. [NEWLINE] [NEWLINE] <mask> you can't have multiple *favourite* lovers  - "the favourite" is logically exclusive. You can have a favourite blond for sex, a favourite person for erotic sex, a favorite person for companion sex - a favourite person for romantic sex - or a favourite *human* for sex. The favourite person for sex is at the exclusion of other people being your favourites for sex. That's not to say you can't find yourself having sex with your your 2nd or 50th preference. [NEWLINE] [NEWLINE] Yes, jealousy is most often immature, it's<mask> not man-made or unnatural, and it's<mask> not always a mistake for the person who feels it. [NEWLINE] [NEWLINE] Romantic Love *is* possessive and exclusive. Each gives each other *to* each other, willingly e.g. "I'm yours! Take me! I only want you and no other!" [NEWLINE] </s>
Label encoding: <s>All favourites are preferences *and* logically exclusive. [NEWLINE] [NEWLINE] A "favourite" is a thing that's been isolated from a list of similar things that share a common characteristic as the one you like best - to the exclusion of 2nd, 3rd and other members of that list. You can't have *all* flavours of icecream as your favourite icecream flavour. [NEWLINE] [NEWLINE] A preference is a greater than-less than comparison, a greater liking for one over the other. [NEWLINE] [NEWLINE] You can like and be friends with many people - but they must exclude the set of people you don't like or value. [NEWLINE] [NEWLINE] You can love and be lovers with many people too - but they must exclude the set you don't love. [NEWLINE] [NEWLINE] But you can't have multiple *favourite* lovers  - "the favourite" is logically exclusive. You can have a favourite blond for sex, a favourite person for erotic sex, a favorite person for companion sex - a favourite person for romantic sex - or a favourite *human* for sex. The favourite person for sex is at the exclusion of other people being your favourites for sex. That's not to say you can't find yourself having sex with your your 2nd or 50th preference. [NEWLINE] [NEWLINE] Yes, jealousy is most often immature, it's also not man-made or unnatural, and it's also not always a mistake for the person who feels it. [NEWLINE] [NEWLINE] Romantic Love *is* possessive and exclusive. Each gives each other *to* each other, willingly e.g. "I'm yours! Take me! I only want you and no other!" [NEWLINE] </s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Bad analogy. There simply isn't enough sufficient data to claim<mask> you're saying. We have plenty of data that shows the likelihood of developing lung cancer for smokers (15% for life time smoker, 25% for heavy smokers),<mask> no such data is available for storms and global warming. e.g<mask> global warming is linked to increase in hurricane activities or its intensity, then by<mask> much? There simply isn't enough data to make any strong statements [ENDQ] [NEWLINE] That's<mask> the tobacco industry (or asbestos or DDT or...) said for half a century. Fact is that we know that more energy in a system means that it will have to be released in some way, we know it's been getting hotter, we know hurricane damage payouts are rising worryingly in the insurance industry, and we know hurricanes are more damaging<mask> the water is warmer. [NEWLINE] [NEWLINE] [STARTQ] "Mass exodus" from New Orlean is due to nature disasters and ultimately a local and temporary event. We already have these kind of disasters happening over the past few years in third world and developing countries (Hurricane, Tsunami, etc), and we didn't see a mass exodus of immigrants fleeing into the US/Euro.<mask> these events are localized and rare, and a gradually rising sea level isn't going to change that. [ENDQ] [NEWLINE] You're just describing<mask> you want it to be... [NEWLINE] [NEWLINE] It's not<mask> it doesn't cause an instant apocalypse that everything will stay the same forever. People don't get fat instantly the moment they open a package of crisps.<mask><mask> they do<mask> most days of the week<mask> a habit, they will.</s><pad><pad>
Label encoding: <s> [STARTQ] Bad analogy. There simply isn't enough sufficient data to claim what you're saying. We have plenty of data that shows the likelihood of developing lung cancer for smokers (15% for life time smoker, 25% for heavy smokers), but no such data is available for storms and global warming. e.g if global warming is linked to increase in hurricane activities or its intensity, then by how much? There simply isn't enough data to make any strong statements [ENDQ] [NEWLINE] That's what the tobacco industry (or asbestos or DDT or...) said for half a century. Fact is that we know that more energy in a system means that it will have to be released in some way, we know it's been getting hotter, we know hurricane damage payouts are rising worryingly in the insurance industry, and we know hurricanes are more damaging when the water is warmer. [NEWLINE] [NEWLINE] [STARTQ] "Mass exodus" from New Orlean is due to nature disasters and ultimately a local and temporary event. We already have these kind of disasters happening over the past few years in third world and developing countries (Hurricane, Tsunami, etc), and we didn't see a mass exodus of immigrants fleeing into the US/Euro. Because these events are localized and rare, and a gradually rising sea level isn't going to change that. [ENDQ] [NEWLINE] You're just describing what you want it to be... [NEWLINE] [NEWLINE] It's not because it doesn't cause an instant apocalypse that everything will stay the same forever. People don't get fat instantly the moment they open a package of crisps. But if they do so most days of the week as a habit, they will.</s><pad><pad>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> morality is relative, then I could just<mask> easily say that I don't have to respect other cultures. [ENDQ] [NEWLINE] Sorry for the 2 hour wait btw. [NEWLINE] [NEWLINE] Absolutely. In truth, morality is purely a social construct that happens to be subjective in nature. [This post goes more into depth with<mask> I mean.]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] It's possible that there's an alternative to absolutism and relativism that I haven't considered. [ENDQ] [NEWLINE] That *may* be possible.<mask>, that question is<mask> ominous<mask> "<mask> are we here?" There's an infinite number of possibilities<mask> to<mask> we may be here. It may even be for no reason at all. [NEWLINE] [NEWLINE] [STARTQ] <mask> someone has a deep philosophical argument about<mask> constitutes truth, I might even consider absolutism. [ENDQ] [NEWLINE] In order for absolutism to exist, there has to be a moral source. Currently there is no known source of morality. There *may* be a source,<mask> it's currently unknown. [NEWLINE] [NEWLINE] [STARTQ] <mask> I was really getting at with my topic,<mask>, is that<mask><mask> that most people who consider themselves moral relativists do<mask> for irrational reasons.<mask><mask> that it's irrational to be a relativist just<mask> societies don't agree on morality. [ENDQ] [NEWLINE] Morality in and of itself is irrational. The only reason<mask> we utilize it is<mask> that we can establish order and functionality within society. The world is amoral by default. (Until a moral source with "absolute moral powers" proves otherwise.) We create morality in our minds.<mask> there really isn't any.</s>
Label encoding: <s> [STARTQ] If morality is relative, then I could just as easily say that I don't have to respect other cultures. [ENDQ] [NEWLINE] Sorry for the 2 hour wait btw. [NEWLINE] [NEWLINE] Absolutely. In truth, morality is purely a social construct that happens to be subjective in nature. [This post goes more into depth with what I mean.]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] It's possible that there's an alternative to absolutism and relativism that I haven't considered. [ENDQ] [NEWLINE] That *may* be possible. However, that question is as ominous as " why are we here?" There's an infinite number of possibilities as to why we may be here. It may even be for no reason at all. [NEWLINE] [NEWLINE] [STARTQ] If someone has a deep philosophical argument about what constitutes truth, I might even consider absolutism. [ENDQ] [NEWLINE] In order for absolutism to exist, there has to be a moral source. Currently there is no known source of morality. There *may* be a source, but it's currently unknown. [NEWLINE] [NEWLINE] [STARTQ] What I was really getting at with my topic, though, is that I think that most people who consider themselves moral relativists do so for irrational reasons. I think that it's irrational to be a relativist just because societies don't agree on morality. [ENDQ] [NEWLINE] Morality in and of itself is irrational. The only reason why we utilize it is so that we can establish order and functionality within society. The world is amoral by default. (Until a moral source with "absolute moral powers" proves otherwise.) We create morality in our minds. But there really isn't any.</s>
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not trying to make a value judgement here - "worse" implies broader analysis and a value judgement on the whole framework. Directly showing a rape is not necessarily presenting one viewpoint or another. For example, Gaspar Noe's film *Irreversible* features an approximately 11 minute, single shot rape scene, in which the viewer is forced to watch the anal rape of a major female character. This is a very unpleasant scene to watch, and it is certainly disempowering to the victim,<mask> it does not situate itself within the Male Gaze,<mask> the detached camera, which is not mediated by any male character, forces the viewer to endure the suffering rather than the power fantasy. The viewer, like the character, is given no escape from that suffering - not even a single edit. A far less graphic scene can be far more empathetic to the male gaze at the expense of the victim's suffering by mediating the viewer's experience of that suffering through an empowered male character. Again, these aren't "worse" or "better" without broader context and an understanding of the thematic purpose of the scene, and I'm using an extreme example here,<mask> a tamer scene can certainly be more provacative than a graphic one on some levels merely<mask> of<mask> it is shot. [NEWLINE] [NEWLINE] Without being familiar with this show, I'm of course still speaking generally,<mask> it is certainly at least possible that the discomfort people have with the scene is partially due to<mask> the camera mediates it through a male character (which seems to be<mask> /u/Gweena14 is saying).</s>
Label encoding: <s>I'm not trying to make a value judgement here - "worse" implies broader analysis and a value judgement on the whole framework. Directly showing a rape is not necessarily presenting one viewpoint or another. For example, Gaspar Noe's film *Irreversible* features an approximately 11 minute, single shot rape scene, in which the viewer is forced to watch the anal rape of a major female character. This is a very unpleasant scene to watch, and it is certainly disempowering to the victim, but it does not situate itself within the Male Gaze, because the detached camera, which is not mediated by any male character, forces the viewer to endure the suffering rather than the power fantasy. The viewer, like the character, is given no escape from that suffering - not even a single edit. A far less graphic scene can be far more empathetic to the male gaze at the expense of the victim's suffering by mediating the viewer's experience of that suffering through an empowered male character. Again, these aren't "worse" or "better" without broader context and an understanding of the thematic purpose of the scene, and I'm using an extreme example here, but a tamer scene can certainly be more provacative than a graphic one on some levels merely because of how it is shot. [NEWLINE] [NEWLINE] Without being familiar with this show, I'm of course still speaking generally, but it is certainly at least possible that the discomfort people have with the scene is partially due to how the camera mediates it through a male character (which seems to be what /u/Gweena14 is saying).</s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> for his dopey demise at the sarlaac pit, that says more about the dubious action directing than it does about him. Star wars canon explicitly says that stormtroopers are extremely accurate and precise.<mask><mask> confronted by "main character defensive magic", they always miss. These missed shots, along with boba fetts blunder, are silly movie tropes that appear in shoddy action scenes, and shouldn't be taken to have wider canon implications,<mask> opposed to his strongly implied and demonstrated record of competence from the rest of the script. [ENDQ] [NEWLINE] This is pretty bad analysis. You can't wave away the stupid actions of characters and their incompetence<mask> of bad directing or movie tropes.<mask><mask><mask><mask> no one lives their life<mask><mask> they were in a movie controlled by some unseen writer. Some director isn't controlling your actions.<mask> you want to analysis a character, you have to take their actions<mask><mask> you were watching a documentary. After all, I don't see anyone arguing, "he is just bad ass<mask> he is written that way". That would be ludicrous. [NEWLINE] [NEWLINE] [STARTQ] During this meeting, his reputation was such that Vader spoke to him individually about his tendencies, implying that his reputation extends all the way to Vader. [ENDQ] [NEWLINE] Or implying that an underlying told him, "hey, we hear this guy takes no prisoners".<mask>, there is absolutely nothing in the film that states that the bounty hunters are elite in any sense of the word. You have no way of knowing<mask> they are the best of the best of the best or just those who answered a Craig's List ad on short notice. </s>
Label encoding: <s> [STARTQ] As for his dopey demise at the sarlaac pit, that says more about the dubious action directing than it does about him. Star wars canon explicitly says that stormtroopers are extremely accurate and precise. Yet when confronted by "main character defensive magic", they always miss. These missed shots, along with boba fetts blunder, are silly movie tropes that appear in shoddy action scenes, and shouldn't be taken to have wider canon implications, as opposed to his strongly implied and demonstrated record of competence from the rest of the script. [ENDQ] [NEWLINE] This is pretty bad analysis. You can't wave away the stupid actions of characters and their incompetence because of bad directing or movie tropes. The reason is because no one lives their life as if they were in a movie controlled by some unseen writer. Some director isn't controlling your actions. If you want to analysis a character, you have to take their actions as if you were watching a documentary. After all, I don't see anyone arguing, "he is just bad ass because he is written that way". That would be ludicrous. [NEWLINE] [NEWLINE] [STARTQ] During this meeting, his reputation was such that Vader spoke to him individually about his tendencies, implying that his reputation extends all the way to Vader. [ENDQ] [NEWLINE] Or implying that an underlying told him, "hey, we hear this guy takes no prisoners". Also, there is absolutely nothing in the film that states that the bounty hunters are elite in any sense of the word. You have no way of knowing if they are the best of the best of the best or just those who answered a Craig's List ad on short notice. </s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to address each of your gender-related points in turn. [NEWLINE] [NEWLINE] **1 - It is not good to be feared.** It is good to be loved, something which comes much easier<mask> you are not feared, and in general comes easier to women. [NEWLINE] [NEWLINE] **2 - Men are held responsible for their children in ALL states.**<mask><mask>, men can even be held responsible for children who are not theirs ([source]( [URL] )). The fact that women cannot abort (<mask> can give up for adoption) in all states,<mask> highly objectionable, does not make them less privileged than men.<mask>, the idea that there is no societal judgment against "deadbeat dads" is ridiculous on its face. [NEWLINE] [NEWLINE] **3 - Tech and engineering are not anti-woman.**<mask><mask>, there are a great many programs and scholarships trying to get more women into these fields ([examples]( [URL] )). It is women who are *choosing* not to enter these fields, which is their right, and not an injustice at all. [NEWLINE] [NEWLINE] **4 -<mask> men dress the way women do, they are pilloried.** Everyone should be able to dress<mask> they want,<mask> unfortunately everyone can face hate for<mask> they dress. I don't know<mask> you can claim that women have it worse in this regard, seeing<mask> there are to my knowledge no items of clothing which are often found acceptable for a man to wear,<mask> not a woman.<mask><mask><mask><mask>, a man wearing a skirt will often face judgment and violence, perhaps fueled by homophobia, another form of hatred which is predominantly focused on males.</s>
Label encoding: <s>I'm going to address each of your gender-related points in turn. [NEWLINE] [NEWLINE] **1 - It is not good to be feared.** It is good to be loved, something which comes much easier when you are not feared, and in general comes easier to women. [NEWLINE] [NEWLINE] **2 - Men are held responsible for their children in ALL states.** In fact, men can even be held responsible for children who are not theirs ([source]( [URL] )). The fact that women cannot abort ( but can give up for adoption) in all states, while highly objectionable, does not make them less privileged than men. Also, the idea that there is no societal judgment against "deadbeat dads" is ridiculous on its face. [NEWLINE] [NEWLINE] **3 - Tech and engineering are not anti-woman.** In fact, there are a great many programs and scholarships trying to get more women into these fields ([examples]( [URL] )). It is women who are *choosing* not to enter these fields, which is their right, and not an injustice at all. [NEWLINE] [NEWLINE] **4 - When men dress the way women do, they are pilloried.** Everyone should be able to dress how they want, but unfortunately everyone can face hate for how they dress. I don't know how you can claim that women have it worse in this regard, seeing as there are to my knowledge no items of clothing which are often found acceptable for a man to wear, but not a woman. On the other hand, a man wearing a skirt will often face judgment and violence, perhaps fueled by homophobia, another form of hatred which is predominantly focused on males.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, without the editor it wouldn't be<mask> it is today [ENDQ] [NEWLINE] A piece of garbage *with* decent grammar? ^^I'm ^^(mostly) ^^joking ^^here ^^please ^^don't ^^kill ^^me [NEWLINE] [NEWLINE] In all seriousness, you raise a lot of good points.<mask><mask> that my problem here is that I don't really hang around in literary circles,<mask> to speak. My primary exposure to people talking about classics has been from teachers, online articles (tributes, etc.), and online conversations, which all (in my experience) tend towards praise and away from criticism. [NEWLINE] [NEWLINE] Your explanation of<mask> makes a book a classic provides a good reference point (<mask> my problem wasn't exactly with books being considered classics<mask> a whole; I was thinking mainly about the inordinate amount of praise I see them getting). [NEWLINE] [NEWLINE] <mask> for Dickens... well, I don't have a whole lot to say. I probably shouldn't have included him<mask> an example,<mask> I've only read *A Christmas Carol* and the beginning of *A Tale of Two Cities* (along with various excerpts from his books). I found his writing to be overly verbose in general,<mask> I don't really have examples to back it up. Your explanation makes sense; perhaps I'll have to give his work another go with that in mind. [NEWLINE] [NEWLINE] I'm not awarding a delta at the moment<mask>,<mask> your post raised good points, none of it really challenged my originally stated view (that classics are given an undue amount of praise<mask> their flaws are glossed over).</s>
Label encoding: <s> [STARTQ] But, without the editor it wouldn't be what it is today [ENDQ] [NEWLINE] A piece of garbage *with* decent grammar? ^^I'm ^^(mostly) ^^joking ^^here ^^please ^^don't ^^kill ^^me [NEWLINE] [NEWLINE] In all seriousness, you raise a lot of good points. I think that my problem here is that I don't really hang around in literary circles, so to speak. My primary exposure to people talking about classics has been from teachers, online articles (tributes, etc.), and online conversations, which all (in my experience) tend towards praise and away from criticism. [NEWLINE] [NEWLINE] Your explanation of what makes a book a classic provides a good reference point ( although my problem wasn't exactly with books being considered classics as a whole; I was thinking mainly about the inordinate amount of praise I see them getting). [NEWLINE] [NEWLINE] As for Dickens... well, I don't have a whole lot to say. I probably shouldn't have included him as an example, since I've only read *A Christmas Carol* and the beginning of *A Tale of Two Cities* (along with various excerpts from his books). I found his writing to be overly verbose in general, but I don't really have examples to back it up. Your explanation makes sense; perhaps I'll have to give his work another go with that in mind. [NEWLINE] [NEWLINE] I'm not awarding a delta at the moment because, while your post raised good points, none of it really challenged my originally stated view (that classics are given an undue amount of praise while their flaws are glossed over).</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] there's no inherent reason it must be in the form of profit [ENDQ] [NEWLINE] Profit is merely the moment that happens<mask> two people exchange in a mutually beneficial way.  It means all people who participated are better off than before.  I buy a slice of pizza I want, which makes me happy and fed, and the vendor can keep selling slices, which gives him means to employ people, make more pizza, and feed himself. [NEWLINE] [NEWLINE] [STARTQ] Profit is private and accumulated by individuals [ENDQ] [NEWLINE] All things are individuals.  There is no unit of sentience<mask> an individual.  There are social trends and discourses,<mask> there are zero actors<mask> individuals.  No exceptions. [NEWLINE] [NEWLINE] [STARTQ] They are sold to maximize profit. This does not necessarily mean selling it at the lowest price [ENDQ] [NEWLINE] Yes, you're trying to have it one way and not realizing that it's both ways.  Things are sold to maximize profit,<mask> _people must be **willing** to buy it at a price._  That means prices drop to the lowest possible point. <mask> prices are high, there is a very good reason for it... <mask> I'm the only person selling ice cream in the desert, I could charge a thousand dollars a cone. <mask> people want it, it's worth a thousand dollars,<mask> there's no other ice cream around. [NEWLINE] [NEWLINE] [STARTQ] You could sell lower than a competitor to attract more customers [ENDQ] [NEWLINE] Yes, lowering prices. [NEWLINE] [NEWLINE] [STARTQ] you could price high and increase the exchange value of the good [ENDQ] [NEWLINE] Or go out of business,<mask> your selling rates will drop noticeably. </s>
Label encoding: <s> [STARTQ] there's no inherent reason it must be in the form of profit [ENDQ] [NEWLINE] Profit is merely the moment that happens when two people exchange in a mutually beneficial way.  It means all people who participated are better off than before.  I buy a slice of pizza I want, which makes me happy and fed, and the vendor can keep selling slices, which gives him means to employ people, make more pizza, and feed himself. [NEWLINE] [NEWLINE] [STARTQ] Profit is private and accumulated by individuals [ENDQ] [NEWLINE] All things are individuals.  There is no unit of sentience besides an individual.  There are social trends and discourses, but there are zero actors besides individuals.  No exceptions. [NEWLINE] [NEWLINE] [STARTQ] They are sold to maximize profit. This does not necessarily mean selling it at the lowest price [ENDQ] [NEWLINE] Yes, you're trying to have it one way and not realizing that it's both ways.  Things are sold to maximize profit, but _people must be **willing** to buy it at a price._  That means prices drop to the lowest possible point.  If prices are high, there is a very good reason for it...  if I'm the only person selling ice cream in the desert, I could charge a thousand dollars a cone.  If people want it, it's worth a thousand dollars, because there's no other ice cream around. [NEWLINE] [NEWLINE] [STARTQ] You could sell lower than a competitor to attract more customers [ENDQ] [NEWLINE] Yes, lowering prices. [NEWLINE] [NEWLINE] [STARTQ] you could price high and increase the exchange value of the good [ENDQ] [NEWLINE] Or go out of business, since your selling rates will drop noticeably. </s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Bill Nye's argument was "Evolution is the accepted theory, here is all the evidence we have supporting it, and please<mask> you have any actual evidence that contradicts it, such<mask> XYZ, bring it to us and we will change our minds. [NEWLINE] [NEWLINE] Ken Ham's argument was "This book tells me<mask> the truth is, and no amount of evidence will change my mind to the contrary." [NEWLINE] [NEWLINE] The question of the night was "Does the creationist theory hold<mask> a viable model for<mask> we came from?" not "Can a creationist become a scientist,<mask> being a creationist?" [NEWLINE] [NEWLINE] The only point that Ken Ham had to stand on (<mask> he wasn't pointing back to "The bible tells me it is real") was "We weren't around to see the rings on the trees, or the layers of ice and<mask> they were formed,<mask> this is all speculative" and the rebuttal to that from Nye boiled down to "We are extrapolating backwards using<mask> we know to make predictions", or put another way "<mask> something consistently behaves in a certain manner, the longer it happens that way, the more certain I can be that it will continue happening that way, and the more certain I can be that it happened that way in the past." which is<mask> we extrapolate from layers of ice and snow and rock layers with fossils. [NEWLINE] [NEWLINE] <mask> basically: I would<mask><mask> Ken Ham brought no irrefutable evidence that creationism holds up<mask> a scientifically rigorous theory, and Bill Nye brought plenty of evidence that refutes creationism and holds up evolution<mask> a theory.</s>
Label encoding: <s>Bill Nye's argument was "Evolution is the accepted theory, here is all the evidence we have supporting it, and please if you have any actual evidence that contradicts it, such as XYZ, bring it to us and we will change our minds. [NEWLINE] [NEWLINE] Ken Ham's argument was "This book tells me what the truth is, and no amount of evidence will change my mind to the contrary." [NEWLINE] [NEWLINE] The question of the night was "Does the creationist theory hold as a viable model for where we came from?" not "Can a creationist become a scientist, despite being a creationist?" [NEWLINE] [NEWLINE] The only point that Ken Ham had to stand on ( when he wasn't pointing back to "The bible tells me it is real") was "We weren't around to see the rings on the trees, or the layers of ice and how they were formed, so this is all speculative" and the rebuttal to that from Nye boiled down to "We are extrapolating backwards using what we know to make predictions", or put another way " If something consistently behaves in a certain manner, the longer it happens that way, the more certain I can be that it will continue happening that way, and the more certain I can be that it happened that way in the past." which is how we extrapolate from layers of ice and snow and rock layers with fossils. [NEWLINE] [NEWLINE] So basically: I would argue that Ken Ham brought no irrefutable evidence that creationism holds up as a scientifically rigorous theory, and Bill Nye brought plenty of evidence that refutes creationism and holds up evolution as a theory.</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] [invalid comparison of abortional vs. birth pain] [ENDQ] [NEWLINE] It still is the woman's right of body integrity and the right to choose which type of pain she prefers to have. [NEWLINE] [NEWLINE] [STARTQ] *consent to have sex ≠ consent to have a baby* [ENDQ] [NEWLINE] [STARTQ] <mask> is *not* fair to make someone pay money for a choice he didn't have at all. [ENDQ] [NEWLINE] Every human has the choice not to have sex. (We are not talking about rape here.) Don't tell me about this all-men-are-inherently-sex-maniacs-who-can't-refrain-from-copulating-nonsense. [NEWLINE] [NEWLINE] I hope I get you right in that you wish to exempt men from parental care in that they generally don't have to pay child support,<mask> to shift the duty onto the mother.<mask><mask> I don't agree with the reasons you pointed out, it is an interesting concept<mask> let's talk through it. [NEWLINE] [NEWLINE] A child has needs and they cost money.<mask> it was the women's task alone to earn the money, there would be more women in need of social welfare benefit,<mask> many of them wouldn't be able to work and care for a child at once.<mask> who would pay for that? Society<mask> a whole, by paying taxes.<mask> to me it is sensible that the duty is rather on the person who de-facto "made a mistake" by sleeping with someone he didn't want to raise a child with, than on all the other people who really didn't have a choice at all.</s>
Label encoding: <s> [STARTQ] [invalid comparison of abortional vs. birth pain] [ENDQ] [NEWLINE] It still is the woman's right of body integrity and the right to choose which type of pain she prefers to have. [NEWLINE] [NEWLINE] [STARTQ] *consent to have sex ≠ consent to have a baby* [ENDQ] [NEWLINE] [STARTQ] Therefore is *not* fair to make someone pay money for a choice he didn't have at all. [ENDQ] [NEWLINE] Every human has the choice not to have sex. (We are not talking about rape here.) Don't tell me about this all-men-are-inherently-sex-maniacs-who-can't-refrain-from-copulating-nonsense. [NEWLINE] [NEWLINE] I hope I get you right in that you wish to exempt men from parental care in that they generally don't have to pay child support, but to shift the duty onto the mother. Even though I don't agree with the reasons you pointed out, it is an interesting concept so let's talk through it. [NEWLINE] [NEWLINE] A child has needs and they cost money. If it was the women's task alone to earn the money, there would be more women in need of social welfare benefit, because many of them wouldn't be able to work and care for a child at once. So who would pay for that? Society as a whole, by paying taxes. So to me it is sensible that the duty is rather on the person who de-facto "made a mistake" by sleeping with someone he didn't want to raise a child with, than on all the other people who really didn't have a choice at all.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Biologically, your DNA's goal is not to be spread into<mask> many different women<mask> possible.  That would be pointless,<mask> the next man to find any women you've laid seed in will kill your offspring and father his own offspring.  (<mask> the penis is shaped specifically to pump out the seed of any competing males in favor of your own,<mask> that's<mask> the point). Historically,<mask> you were not there to protect your young, there is a very high chance that they would not survive until adulthood. [NEWLINE] [NEWLINE] In order for your DNA to successfully survive long enough to propagate itself, you have to not just create the offspring,<mask> protect them until they are old enough to defend themselves and reproduce on their own. This typically happens at around 12-15 years of age, which is<mask> those were the unofficial ages of adulthood in medieval times (women tended to enter adult hood a little earlier then males). [NEWLINE] [NEWLINE] Now it would<mask> help your DNA<mask> your children weren't poor, degenerate, alcoholic, anti-social, psychotic fuck ups.<mask> they were, then they are less likely to attract mates themselves, and your DNA would not propagate.<mask> you stick around, guiding them and teaching them, supporting them, etc.<mask> that they can grow up to be fully functional human beings that are capable of reproducing. [NEWLINE] [NEWLINE] Monogamy helps in all of this, and it isn't just an accident or coincidence that every single human culture in the history of the world has had the construct of marriage. Marriage is a social adaptation that is a competitive advantage<mask> compared to other species.  </s>
Label encoding: <s>Biologically, your DNA's goal is not to be spread into as many different women as possible.  That would be pointless, because the next man to find any women you've laid seed in will kill your offspring and father his own offspring.  ( Also the penis is shaped specifically to pump out the seed of any competing males in favor of your own, but that's besides the point). Historically, If you were not there to protect your young, there is a very high chance that they would not survive until adulthood. [NEWLINE] [NEWLINE] In order for your DNA to successfully survive long enough to propagate itself, you have to not just create the offspring, but protect them until they are old enough to defend themselves and reproduce on their own. This typically happens at around 12-15 years of age, which is why those were the unofficial ages of adulthood in medieval times (women tended to enter adult hood a little earlier then males). [NEWLINE] [NEWLINE] Now it would also help your DNA if your children weren't poor, degenerate, alcoholic, anti-social, psychotic fuck ups. If they were, then they are less likely to attract mates themselves, and your DNA would not propagate. So you stick around, guiding them and teaching them, supporting them, etc. so that they can grow up to be fully functional human beings that are capable of reproducing. [NEWLINE] [NEWLINE] Monogamy helps in all of this, and it isn't just an accident or coincidence that every single human culture in the history of the world has had the construct of marriage. Marriage is a social adaptation that is a competitive advantage when compared to other species.  </s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Sorry, no. Here you describe a representative democracy(one probably too influenced by the plutocracy), and about to describe a direct democracy. Both are types of democracy. Democracy is best defined<mask> "rule by the people". [ENDQ] [NEWLINE] No to<mask>? Nothing you just said<mask><mask> with. I did describe representative democracy, and then I went on to describe direct democracy. [NEWLINE] [NEWLINE] <mask> you're saying no to my thought that representative democracy is not democracy<mask> democracy is "rule by the people" (<mask> simplistic and elementary-textbookish that may be), I don't think representative democracy is rule by the people. Politicians can tell you whatever they want, and then go on to get elected and completely do the opposite.<mask> this were rule by the people, 1) he wouldn't be allowed to do that, and 2) the people would be able to immediately recall him (they can't). [NEWLINE] [NEWLINE] [STARTQ] You have not experienced enough or investigated enough to make such a statement. [ENDQ] [NEWLINE] Let me rephrase: the only people that can represent me are the ones that I choose to represent me, who would have the same thoughts<mask> I do. [NEWLINE] [NEWLINE] On a side note. You don't know<mask> I've experienced or investigated. This wasn't a productive statement for you to make. [NEWLINE] [NEWLINE] [STARTQ] It can be a democratic system for that one aspect,<mask> not for the rest of your necessities,<mask> not really. [ENDQ] [NEWLINE] Yes. I thought democracy is "rule by the people."<mask> did "oh by the way it has to provide all of your necessities" come in?</s><pad>
Label encoding: <s> [STARTQ] Sorry, no. Here you describe a representative democracy(one probably too influenced by the plutocracy), and about to describe a direct democracy. Both are types of democracy. Democracy is best defined as "rule by the people". [ENDQ] [NEWLINE] No to what? Nothing you just said I disagree with. I did describe representative democracy, and then I went on to describe direct democracy. [NEWLINE] [NEWLINE] If you're saying no to my thought that representative democracy is not democracy because democracy is "rule by the people" ( however simplistic and elementary-textbookish that may be), I don't think representative democracy is rule by the people. Politicians can tell you whatever they want, and then go on to get elected and completely do the opposite. If this were rule by the people, 1) he wouldn't be allowed to do that, and 2) the people would be able to immediately recall him (they can't). [NEWLINE] [NEWLINE] [STARTQ] You have not experienced enough or investigated enough to make such a statement. [ENDQ] [NEWLINE] Let me rephrase: the only people that can represent me are the ones that I choose to represent me, who would have the same thoughts as I do. [NEWLINE] [NEWLINE] On a side note. You don't know what I've experienced or investigated. This wasn't a productive statement for you to make. [NEWLINE] [NEWLINE] [STARTQ] It can be a democratic system for that one aspect, but not for the rest of your necessities, so not really. [ENDQ] [NEWLINE] Yes. I thought democracy is "rule by the people." When did "oh by the way it has to provide all of your necessities" come in?</s><pad>
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Essentially you are suggesting we allow natural selection to take its course.  That a population will become more stable and hardier<mask> we allow natural pressures to do their work.  There's plenty of scientific support for the idea<mask> lets expand that premise. [NEWLINE] [NEWLINE] BRCA1 and BRCA2 are "breast cancer genes".  People with these harmful mutations are 5 times more likely to develop breast cancer than the average woman.  Breast cancer is often fatal, especially<mask> left untreated. <mask> we follow your premise then allowing all women with either mutation die of breast cancer will eliminate the gene from our gene pool and drastically reduce breast cancer rates. <mask>, we have "solved" the problem of breast cancer. [NEWLINE] [NEWLINE] This principle can be extended to any heritable disease, chronic condition or generally unhealthy person. <mask> we simply let everyone afflicted with disease die there would be much less disease in our society<mask> it would be inherited and communicable through far fewer people. [NEWLINE] [NEWLINE] <mask> surely you see the problem with this train of thought.  Its awfully inhumane. <mask> humans most of our interests intersect around the desire to remain alive and to keep the people we love alive.  We have developed all kinds of ingenious ploys to further this goal from medicine, to industry to food and water supply.  Its easy to say "lets not ship food to Africa, they're overpopulated.  Its for their own good, it'll end hunger".  Its much harder to watch your mother die of breast cancer without anyone lifting a finger to help<mask> "It'll end breast cancer".</s>
Label encoding: <s>Essentially you are suggesting we allow natural selection to take its course.  That a population will become more stable and hardier if we allow natural pressures to do their work.  There's plenty of scientific support for the idea so lets expand that premise. [NEWLINE] [NEWLINE] BRCA1 and BRCA2 are "breast cancer genes".  People with these harmful mutations are 5 times more likely to develop breast cancer than the average woman.  Breast cancer is often fatal, especially if left untreated.  If we follow your premise then allowing all women with either mutation die of breast cancer will eliminate the gene from our gene pool and drastically reduce breast cancer rates.  Thus, we have "solved" the problem of breast cancer. [NEWLINE] [NEWLINE] This principle can be extended to any heritable disease, chronic condition or generally unhealthy person.  If we simply let everyone afflicted with disease die there would be much less disease in our society as it would be inherited and communicable through far fewer people. [NEWLINE] [NEWLINE] But surely you see the problem with this train of thought.  Its awfully inhumane.  As humans most of our interests intersect around the desire to remain alive and to keep the people we love alive.  We have developed all kinds of ingenious ploys to further this goal from medicine, to industry to food and water supply.  Its easy to say "lets not ship food to Africa, they're overpopulated.  Its for their own good, it'll end hunger".  Its much harder to watch your mother die of breast cancer without anyone lifting a finger to help because "It'll end breast cancer".</s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>(Sorry it took me<mask> long to respond to this, I had stuff to do today. I understand<mask> you don't respond to it.) [NEWLINE] [NEWLINE] <mask><mask> you are debating a straw-man. I definitely would not interpret the statement "We live in the best time to be alive"<mask> equivalent to "Life has gotten better for every single person in every single area of the world", and<mask><mask><mask><mask> others would either. That second statement is clearly false, and I would never defend it. [NEWLINE] [NEWLINE] Instead, it seems clear to me that OP meant something more along the lines of "the proportion of people in abject poverty has decreased" or "people have on _average_ a higher quality of life today". Or,<mask> other people brought up (Rawl's Veil of Ignorance)<mask> you could choose to be born now or in the past,<mask> you didn't know<mask> you were going to be born, or in<mask> class,<mask> would you choose? I would definitely choose now (I mean, sure,<mask> I could be reborn<mask> Rockefeller, that would be pretty awesome,<mask><mask> are the odds?) [NEWLINE] [NEWLINE] I really think OP asked an essentially statistical question, and you responded with anecdotal data. Yes, things haven't improved for some people, and some countries have gotten worse.<mask><mask> do we compare the gains and losses to get a larger picture of<mask> the world has changed, and<mask> it has changed in a net positive or a net negative direction? That is literally<mask> statistics is for,<mask> I'm confused<mask> you are<mask> against it in this case.</s>
Label encoding: <s>(Sorry it took me so long to respond to this, I had stuff to do today. I understand if you don't respond to it.) [NEWLINE] [NEWLINE] I think you are debating a straw-man. I definitely would not interpret the statement "We live in the best time to be alive" as equivalent to "Life has gotten better for every single person in every single area of the world", and I do not think others would either. That second statement is clearly false, and I would never defend it. [NEWLINE] [NEWLINE] Instead, it seems clear to me that OP meant something more along the lines of "the proportion of people in abject poverty has decreased" or "people have on _average_ a higher quality of life today". Or, as other people brought up (Rawl's Veil of Ignorance) if you could choose to be born now or in the past, but you didn't know where you were going to be born, or in what class, what would you choose? I would definitely choose now (I mean, sure, if I could be reborn as Rockefeller, that would be pretty awesome, but what are the odds?) [NEWLINE] [NEWLINE] I really think OP asked an essentially statistical question, and you responded with anecdotal data. Yes, things haven't improved for some people, and some countries have gotten worse. But how do we compare the gains and losses to get a larger picture of how the world has changed, and if it has changed in a net positive or a net negative direction? That is literally what statistics is for, so I'm confused why you are so against it in this case.</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What do you mean then<mask> you say you want to "let white males have reddit"?<mask> it sounds like you want to make reddit a space<mask> unattractive for all others that virtually no one else bothers sticking around.<mask> not actively, then at least passively. [NEWLINE] [NEWLINE] <mask><mask> it is absolutely bad<mask> reddit is<mask> far squewed to one side of the story that another might<mask> well not exist at all. [NEWLINE] [NEWLINE] <mask> some white, sheltered kid says that racism is over in America, I want a black person to be on reddit<mask> they can relate<mask> they were stopped by the police unreasonably often. [NEWLINE] [NEWLINE] <mask> MRAs talk about<mask> much better it is to be a women in the west, I want a women to be able to comment about<mask> she can't even walk down the street without being verbally harrassed. [NEWLINE] [NEWLINE] <mask> someone claims that homosexuals should just stop being like that, I want a homosexual person to relate<mask> they cried themselves to sleep<mask> they just could not. [NEWLINE] [NEWLINE] <mask> a couple of atheist talk about<mask> religion should best be purged from the earth, I want a religious person on reddit to tell them<mask> their religion changed their life for the better. [NEWLINE] [NEWLINE] I'm not saying these group are always wrong, just that there is always another side to any story, and it is always best to hear them both. Even<mask> the thing that was said in the first placee is right afterall. [NEWLINE] [NEWLINE] <mask><mask> this is absolutely worth some changes.<mask> we loose a few hateful fucks in the process, I won't cry for them.</s>
Label encoding: <s>What do you mean then if you say you want to "let white males have reddit"? Because it sounds like you want to make reddit a space so unattractive for all others that virtually no one else bothers sticking around. If not actively, then at least passively. [NEWLINE] [NEWLINE] I think it is absolutely bad if reddit is so far squewed to one side of the story that another might as well not exist at all. [NEWLINE] [NEWLINE] If some white, sheltered kid says that racism is over in America, I want a black person to be on reddit so they can relate how they were stopped by the police unreasonably often. [NEWLINE] [NEWLINE] If MRAs talk about how much better it is to be a women in the west, I want a women to be able to comment about how she can't even walk down the street without being verbally harrassed. [NEWLINE] [NEWLINE] If someone claims that homosexuals should just stop being like that, I want a homosexual person to relate how they cried themselves to sleep because they just could not. [NEWLINE] [NEWLINE] If a couple of atheist talk about how religion should best be purged from the earth, I want a religious person on reddit to tell them how their religion changed their life for the better. [NEWLINE] [NEWLINE] I'm not saying these group are always wrong, just that there is always another side to any story, and it is always best to hear them both. Even if the thing that was said in the first placee is right afterall. [NEWLINE] [NEWLINE] I think this is absolutely worth some changes. If we loose a few hateful fucks in the process, I won't cry for them.</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL].9w4;_ylu=X3oDMTEzdm5zbTcwBHNlYwNzcgRwb3MDNQRjb2xvA2dxMQR2dGlkA01PVVM5Ml8x/RV=1/RE=1395461014/RO=10/RU=http%3a%2f%2fwww.biblicalarchaeology.org%2fdaily%2fnews%2fdid-the-carthaginians-really-practice-infant-sacrifice%2f/RS=%5EADAYXZYSZL5bgQYTrPWVJjPpNqoYHo- [NEWLINE] [NEWLINE] [URL].9w4;_ylu=X3oDMTEzc2k5MWRoBHNlYwNzcgRwb3MDMQRjb2xvA2dxMQR2dGlkA01PVVM5Ml8x/RV=1/RE=1395461014/RO=10/RU=http%3a%2f%2falencon13.blogspot.com%2f2006%2f06%2fhuman-sacrifice-in-ancient-canaan.html/RS=%5EADA4Pe8llTZrBhX9evhbo_rusXDyjA-</s>
Label encoding: <s> [URL].9w4;_ylu=X3oDMTEzdm5zbTcwBHNlYwNzcgRwb3MDNQRjb2xvA2dxMQR2dGlkA01PVVM5Ml8x/RV=1/RE=1395461014/RO=10/RU=http%3a%2f%2fwww.biblicalarchaeology.org%2fdaily%2fnews%2fdid-the-carthaginians-really-practice-infant-sacrifice%2f/RS=%5EADAYXZYSZL5bgQYTrPWVJjPpNqoYHo- [NEWLINE] [NEWLINE] [URL].9w4;_ylu=X3oDMTEzc2k5MWRoBHNlYwNzcgRwb3MDMQRjb2xvA2dxMQR2dGlkA01PVVM5Ml8x/RV=1/RE=1395461014/RO=10/RU=http%3a%2f%2falencon13.blogspot.com%2f2006%2f06%2fhuman-sacrifice-in-ancient-canaan.html/RS=%5EADA4Pe8llTZrBhX9evhbo_rusXDyjA-</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am the result of endless eons of evolution. For me to bear witness to the *now* in which I currently exist every single form of life that is my lineage have accomplished an incredible feat we've<mask> to observe elsewhere in the universe. [NEWLINE] [NEWLINE] From my mother and father, to the beasts we once were, to the spark-of-life microbe<mask> it began eternities ago reaching into the infinite. Every one of them learned<mask> to survive, to thrive and procreate. I am no stranger to this myself with a beautiful daughter to continue our great endless legacy. [NEWLINE] [NEWLINE] The thought of death frightens some<mask> pondering the soul. Cast the notion of'soul' aside. Our identity is physical. It is just a system of survival confined in our brains.<mask> we gave life to our daughter we didn't create life from lifelessness. We didn't produce a new soul for her to bear. We took a very much alive part of ourselves and created a new vessel for it. We are the same phenomenon reaching ever forward into the infinite. [NEWLINE] [NEWLINE] The only death that frightens me is one that drives ***my*** specific lineage to a halt. The death of a body is nothing. We have died a countless infinite number. The death of our endless reach into the infinite is the only death worth fearing. It would be a sincerely regrettable shame to end a legacy that has lasted<mask> many billions of years. [NEWLINE] [NEWLINE] <mask> you separate the concept of memory and identity from<mask> you define *life*, you can begin to think about your fear a bit differently.</s><pad>
Label encoding: <s>I am the result of endless eons of evolution. For me to bear witness to the *now* in which I currently exist every single form of life that is my lineage have accomplished an incredible feat we've yet to observe elsewhere in the universe. [NEWLINE] [NEWLINE] From my mother and father, to the beasts we once were, to the spark-of-life microbe where it began eternities ago reaching into the infinite. Every one of them learned how to survive, to thrive and procreate. I am no stranger to this myself with a beautiful daughter to continue our great endless legacy. [NEWLINE] [NEWLINE] The thought of death frightens some when pondering the soul. Cast the notion of'soul' aside. Our identity is physical. It is just a system of survival confined in our brains. When we gave life to our daughter we didn't create life from lifelessness. We didn't produce a new soul for her to bear. We took a very much alive part of ourselves and created a new vessel for it. We are the same phenomenon reaching ever forward into the infinite. [NEWLINE] [NEWLINE] The only death that frightens me is one that drives ***my*** specific lineage to a halt. The death of a body is nothing. We have died a countless infinite number. The death of our endless reach into the infinite is the only death worth fearing. It would be a sincerely regrettable shame to end a legacy that has lasted so many billions of years. [NEWLINE] [NEWLINE] When you separate the concept of memory and identity from how you define *life*, you can begin to think about your fear a bit differently.</s><pad>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let's set aside Comcast for the moment,<mask> it is a semi-monopoly in most places. [NEWLINE] [NEWLINE] Let's consider a company that just made the best video game ever. <mask> shouldn't they be able to charge whatever they want for it? You don't have to buy<mask> you don't think it's worth it.  They can set the price lower to get more sales and lower profit or higher for more profit and fewer sales. <mask> either way, they will make<mask> people want to pay for it. [NEWLINE] [NEWLINE] And then, seeing<mask> much they made, another company comes in and invests to make an even better game.  And the same thing begins again. [NEWLINE] [NEWLINE] Then, the first company goes back, invests a ton in a 3rd game, which completely bombs. [NEWLINE] [NEWLINE] Who is harmed by this process?  Nobody is screwed<mask> they don't have to buy the game.  We end up with not just one<mask> two great games<mask> of the profit motive.  The first company is able to survive the failure of game 3<mask> of the profits they made on game 1. [NEWLINE] [NEWLINE] Now, in your scenario, the developer behind the game, who is working a boring job writing code for spreadsheet software takes a look and decides that it's not worth the risk of quitting his job and spending a couple of years developing the killer game.  Or even<mask> he does, company 2 decides that it's not worth the investment for 15%. [NEWLINE] [NEWLINE] Now, Comcast is another story. <mask> essentially a utility,<mask><mask> with regulating the bastards. [NEWLINE] </s>
Label encoding: <s>Let's set aside Comcast for the moment, since it is a semi-monopoly in most places. [NEWLINE] [NEWLINE] Let's consider a company that just made the best video game ever.  Why shouldn't they be able to charge whatever they want for it? You don't have to buy if you don't think it's worth it.  They can set the price lower to get more sales and lower profit or higher for more profit and fewer sales.  But either way, they will make what people want to pay for it. [NEWLINE] [NEWLINE] And then, seeing how much they made, another company comes in and invests to make an even better game.  And the same thing begins again. [NEWLINE] [NEWLINE] Then, the first company goes back, invests a ton in a 3rd game, which completely bombs. [NEWLINE] [NEWLINE] Who is harmed by this process?  Nobody is screwed because they don't have to buy the game.  We end up with not just one but two great games because of the profit motive.  The first company is able to survive the failure of game 3 because of the profits they made on game 1. [NEWLINE] [NEWLINE] Now, in your scenario, the developer behind the game, who is working a boring job writing code for spreadsheet software takes a look and decides that it's not worth the risk of quitting his job and spending a couple of years developing the killer game.  Or even if he does, company 2 decides that it's not worth the investment for 15%. [NEWLINE] [NEWLINE] Now, Comcast is another story.  As essentially a utility, I agree with regulating the bastards. [NEWLINE] </s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0577, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> art is... anything that's hard to make? [ENDQ] [NEWLINE] <mask> I meant was that the paintings seemed to be created in a very basic manner, which has<mask> been corrected,<mask> there is no need to expound on that. And<mask><mask><mask><mask><mask>,<mask> I appreciate most art, I would appreciate *The Ambassadors* more than a childhood drawing of a rocket. [NEWLINE] [NEWLINE] [STARTQ] That's exactly<mask> it is. [ENDQ] [NEWLINE] My first assumption was that there was something more to it than that, which is apparently wrong. It looks interesting<mask> it is,<mask> contains less elements than most other paintings which I do appreciate. [NEWLINE] [NEWLINE] [STARTQ] Just<mask> you don't receive a message doesn't mean it doesn't send one.<mask>, who said art has to be 'pleasing'? Is art only good<mask> it looks pretty? [ENDQ] [NEWLINE] I realize this.<mask>, sending a message is wasted<mask> it is not received, and<mask> others have apparently received it, I have not. This difference led me to conclude that there must be something more to it than I am getting. I made some mistakes in wording; essentially I mean that<mask> there could be a message it is wasted on me, and I want to find out<mask> this is<mask>.<mask> for 'looking pretty,' one part of<mask> I consider artistically pleasing is *complexity,* not just aesthetic beauty (I probably should have clarified in my original post). Something more complex than<mask> I see in Rothko's paintings, which are,<mask> you confirmed, a bunch of different overlaid rectangles.</s>
Label encoding: <s> [STARTQ] So art is... anything that's hard to make? [ENDQ] [NEWLINE] What I meant was that the paintings seemed to be created in a very basic manner, which has since been corrected, so there is no need to expound on that. And as a matter of fact, while I appreciate most art, I would appreciate *The Ambassadors* more than a childhood drawing of a rocket. [NEWLINE] [NEWLINE] [STARTQ] That's exactly what it is. [ENDQ] [NEWLINE] My first assumption was that there was something more to it than that, which is apparently wrong. It looks interesting as it is, but contains less elements than most other paintings which I do appreciate. [NEWLINE] [NEWLINE] [STARTQ] Just because you don't receive a message doesn't mean it doesn't send one. Also, who said art has to be 'pleasing'? Is art only good when it looks pretty? [ENDQ] [NEWLINE] I realize this. However, sending a message is wasted if it is not received, and while others have apparently received it, I have not. This difference led me to conclude that there must be something more to it than I am getting. I made some mistakes in wording; essentially I mean that while there could be a message it is wasted on me, and I want to find out why this is so. As for 'looking pretty,' one part of what I consider artistically pleasing is *complexity,* not just aesthetic beauty (I probably should have clarified in my original post). Something more complex than what I see in Rothko's paintings, which are, as you confirmed, a bunch of different overlaid rectangles.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this is an interesting point,<mask><mask> it illustrates a completely reasonable standard for consent that<mask><mask> falls apart in context of sex in our repressed culture (i.e. esp. in the U.S.). In business,<mask> someone agrees to something, we expect them to follow through. They can technically pull out of an agreement at any time,<mask> they are punished for this. We expect them to not waste our time. I would say that in most business situations, working together means cooperating toward some future end involving mutual benefit. [NEWLINE] [NEWLINE] Sex does not fit this mold<mask> easily. People do not necessarily know<mask> they want, especially<mask> they are inexperienced. In cases like that, even an enthusiastic yes is not good enough. You have to constantly communicate. This is not<mask> easy<mask> simply getting more yes responses - that reduces to the absurd pretty quickly - you actually have to talk and figure out feelings. Sometimes people don't WANT to talk stuff out - they want to rush in and have their partner take responsibility. It is a tough space to navigate. [NEWLINE] [NEWLINE] Aside from that, I am of the opinion that even<mask> you are in the middle of having great consensual sex,<mask> someone wants to stop - you stop. There is no business arrangement that implies<mask> the future will be.<mask> anything, you can decide to stop associating with people that you find have stopping habits that annoy you. [NEWLINE] [NEWLINE] Similarly, I do not want to live in a world<mask> being naked means I owe people sex. That does not seem desirable to me. ; ) [NEWLINE] </s>
Label encoding: <s>I think this is an interesting point, insofar as it illustrates a completely reasonable standard for consent that I think falls apart in context of sex in our repressed culture (i.e. esp. in the U.S.). In business, if someone agrees to something, we expect them to follow through. They can technically pull out of an agreement at any time, but they are punished for this. We expect them to not waste our time. I would say that in most business situations, working together means cooperating toward some future end involving mutual benefit. [NEWLINE] [NEWLINE] Sex does not fit this mold as easily. People do not necessarily know what they want, especially if they are inexperienced. In cases like that, even an enthusiastic yes is not good enough. You have to constantly communicate. This is not as easy as simply getting more yes responses - that reduces to the absurd pretty quickly - you actually have to talk and figure out feelings. Sometimes people don't WANT to talk stuff out - they want to rush in and have their partner take responsibility. It is a tough space to navigate. [NEWLINE] [NEWLINE] Aside from that, I am of the opinion that even if you are in the middle of having great consensual sex, if someone wants to stop - you stop. There is no business arrangement that implies how the future will be. If anything, you can decide to stop associating with people that you find have stopping habits that annoy you. [NEWLINE] [NEWLINE] Similarly, I do not want to live in a world where being naked means I owe people sex. That does not seem desirable to me. ; ) [NEWLINE] </s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I really want to give you a delta<mask> of all the work you've put into your posts, and I completely understand<mask> you're coming from in your second paragraph. Obviously, no one knows everything about me just from<mask> I wrote, and I can't consider every single facet of my personality except<mask> I need to bring it up<mask> it's relevant.<mask>, we run into those issues<mask> nobody can verify anything. A very sticky situation. [NEWLINE] [NEWLINE] I may very well be a narcissist,<mask> I'll look into that. I've always thought that I'm excessively modest,<mask> maybe I'm narcissistic about my modesty? Is that a thing? [NEWLINE] [NEWLINE] The reason I can't award you a delta just<mask> is<mask> of the following: You mention that at some point,<mask> they somehow miss my bad qualities due to some PUA magic, a woman is going to wake up and be hit full force by my negative personality traits. The problem is that I mentioned in my OP the *opposite* situation,<mask> I've noticed that<mask> women give me a chance (and that's<mask> it really feels like sometimes, like I'm basically working super hard to convince them to give me a shot), they'll fall in love with me extremely quickly once they get to know the real me. The reason PUA appeals to me from my perspective is<mask> I can shortcut that first step, I can spend more time showing my true personality. I don't need to keep the PUA shit up forever, just until I'm past that first step. Does that make sense?</s>
Label encoding: <s>I really want to give you a delta because of all the work you've put into your posts, and I completely understand where you're coming from in your second paragraph. Obviously, no one knows everything about me just from what I wrote, and I can't consider every single facet of my personality except when I need to bring it up because it's relevant. Hence, we run into those issues where nobody can verify anything. A very sticky situation. [NEWLINE] [NEWLINE] I may very well be a narcissist, so I'll look into that. I've always thought that I'm excessively modest, but maybe I'm narcissistic about my modesty? Is that a thing? [NEWLINE] [NEWLINE] The reason I can't award you a delta just yet is because of the following: You mention that at some point, if they somehow miss my bad qualities due to some PUA magic, a woman is going to wake up and be hit full force by my negative personality traits. The problem is that I mentioned in my OP the *opposite* situation, where I've noticed that when women give me a chance (and that's what it really feels like sometimes, like I'm basically working super hard to convince them to give me a shot), they'll fall in love with me extremely quickly once they get to know the real me. The reason PUA appeals to me from my perspective is if I can shortcut that first step, I can spend more time showing my true personality. I don't need to keep the PUA shit up forever, just until I'm past that first step. Does that make sense?</s>
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is some competition,<mask> it's not vigorous by any means. The few companies in power still maintain a *very* large profit margin. [NEWLINE] [NEWLINE] <mask> takes this a step further from other inflated industries is that there is that you have to buy the diamond even<mask> you don't gain anything from it. Refusing to buy a ring comes with a very large price<mask> people may criticize you heavily for the choice. For the sake of argument, let's say it's about equal to $11,000. Between the choice of gaining close to nothing of value from the $10,000 diamond, or losing $11,000, you will buy it<mask><mask> it's vastly beyond<mask> the rock is worth. OP claims (and I partially agree) that the diamond companies intentionally set this up<mask> that you would buy something you don't need. [NEWLINE] [NEWLINE] For contrast, let's look at Apple. A good laptop probably costs about $300-600 to make,<mask> they charge around $2000 for a MacBook. This is far in excess of<mask> the computer is worth, and there are only a handful of minor reasons to buy it instead of getting a windows and installing mac OS, saving $1000+ in the process.<mask>, this is not a sham<mask> you have the choice.<mask> I don't like<mask> Apple is doing, I just buy from Microsoft, or I get a phone/tablet instead of a computer. [NEWLINE] [NEWLINE] I don't know of any company that sells diamonds near the real price, and there are no viable substitutes for a lot of people.</s>
Label encoding: <s>There is some competition, but it's not vigorous by any means. The few companies in power still maintain a *very* large profit margin. [NEWLINE] [NEWLINE] What takes this a step further from other inflated industries is that there is that you have to buy the diamond even if you don't gain anything from it. Refusing to buy a ring comes with a very large price since people may criticize you heavily for the choice. For the sake of argument, let's say it's about equal to $11,000. Between the choice of gaining close to nothing of value from the $10,000 diamond, or losing $11,000, you will buy it even though it's vastly beyond what the rock is worth. OP claims (and I partially agree) that the diamond companies intentionally set this up so that you would buy something you don't need. [NEWLINE] [NEWLINE] For contrast, let's look at Apple. A good laptop probably costs about $300-600 to make, yet they charge around $2000 for a MacBook. This is far in excess of what the computer is worth, and there are only a handful of minor reasons to buy it instead of getting a windows and installing mac OS, saving $1000+ in the process. However, this is not a sham because you have the choice. If I don't like what Apple is doing, I just buy from Microsoft, or I get a phone/tablet instead of a computer. [NEWLINE] [NEWLINE] I don't know of any company that sells diamonds near the real price, and there are no viable substitutes for a lot of people.</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask>...all I need to do is present you with one good reason to have IVF that isn't vanity or personal narcissism at work to get the delta? [NEWLINE] [NEWLINE] Okay.... [NEWLINE] [NEWLINE] The Orthodox Jewish community has a lot of problems<mask> they've been through some major genetic bottlenecks and are insular in their breeding habits, leaving them with _very_ high incidence of genetic disorders (Tay-Sachs, for instance).  This has proven such a problem for their community that every child is pre-screened to see<mask> they are carriers of known genetic disorders and efforts made to pair them off with mates that do not share their carrier status (which helps,<mask> is far from fool-proof).  Breeding is a _huge_ issue for their community<mask> they wish to keep their culture and religious practice alive, especially<mask> they believe that Judaism is passed through the mother's lineage - adoption is not an option for them. [NEWLINE] [NEWLINE] <mask> part of this effort, they make heavy use of IVF services, not just to help those with fertility issues have children,<mask> to allow them to screen embryos for genetic disorders.  Modern IVF tech allows doctors to extract cells from the developing embryo at a very early stage and screen them for genetic issues ranging from anuploidy to specific single-gene disorders, prior to implanting them in the mother. [NEWLINE] [NEWLINE] So, for them at least, IVF isn't a matter of vanity or personal narcissism, it's a much-needed tool to help keep their culture alive.</s>
Label encoding: <s>So...all I need to do is present you with one good reason to have IVF that isn't vanity or personal narcissism at work to get the delta? [NEWLINE] [NEWLINE] Okay.... [NEWLINE] [NEWLINE] The Orthodox Jewish community has a lot of problems because they've been through some major genetic bottlenecks and are insular in their breeding habits, leaving them with _very_ high incidence of genetic disorders (Tay-Sachs, for instance).  This has proven such a problem for their community that every child is pre-screened to see if they are carriers of known genetic disorders and efforts made to pair them off with mates that do not share their carrier status (which helps, but is far from fool-proof).  Breeding is a _huge_ issue for their community if they wish to keep their culture and religious practice alive, especially since they believe that Judaism is passed through the mother's lineage - adoption is not an option for them. [NEWLINE] [NEWLINE] As part of this effort, they make heavy use of IVF services, not just to help those with fertility issues have children, but to allow them to screen embryos for genetic disorders.  Modern IVF tech allows doctors to extract cells from the developing embryo at a very early stage and screen them for genetic issues ranging from anuploidy to specific single-gene disorders, prior to implanting them in the mother. [NEWLINE] [NEWLINE] So, for them at least, IVF isn't a matter of vanity or personal narcissism, it's a much-needed tool to help keep their culture alive.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me start with this: a good parent puts the needs of their children above their own.  This isn't true of all parents, certainly,<mask> it's pretty universal for the good ones.  The fact that you are thinking this through makes me suspect that you are more likely to be one<mask> well. [NEWLINE] [NEWLINE] <mask> you have a kid, you are making a decision to forgo a lifestyle with dramatically more disposal income.  You're giving up the ability to just randomly say, "lets go away for the weekend". [NEWLINE] [NEWLINE] You give up the right to sleep in on Saturday<mask> there is a baby the needs attention. [NEWLINE] [NEWLINE] Most of all, you give up the right to think of yourself first. <mask> a fire alarm goes off, good parents don't think "<mask> do I save myself?", they think, "<mask>'s my kid?". [NEWLINE] [NEWLINE] And you give up the right to sleep well at night, instead of worrying about not just the stuff that might happen to you,<mask>, even worse,<mask> might happen to your kid.  And you STILL keep worrying<mask> you are an adult and your kids have kids of their own. [NEWLINE] [NEWLINE] Now, there are plenty of benefits<mask> well,<mask> you don't do it<mask> you have someone to visit you in the nursing home.  Instead you'd trying to build the perfect model airplane, only knowing that you're going to have to wind it up and let it fly, and maybe never see it again. [NEWLINE] [NEWLINE] <mask>, no, I don't think there are only selfish reasons.</s>
Label encoding: <s>Let me start with this: a good parent puts the needs of their children above their own.  This isn't true of all parents, certainly, but it's pretty universal for the good ones.  The fact that you are thinking this through makes me suspect that you are more likely to be one as well. [NEWLINE] [NEWLINE] When you have a kid, you are making a decision to forgo a lifestyle with dramatically more disposal income.  You're giving up the ability to just randomly say, "lets go away for the weekend". [NEWLINE] [NEWLINE] You give up the right to sleep in on Saturday if there is a baby the needs attention. [NEWLINE] [NEWLINE] Most of all, you give up the right to think of yourself first.  When a fire alarm goes off, good parents don't think " how do I save myself?", they think, " where's my kid?". [NEWLINE] [NEWLINE] And you give up the right to sleep well at night, instead of worrying about not just the stuff that might happen to you, but, even worse, what might happen to your kid.  And you STILL keep worrying when you are an adult and your kids have kids of their own. [NEWLINE] [NEWLINE] Now, there are plenty of benefits as well, but you don't do it so you have someone to visit you in the nursing home.  Instead you'd trying to build the perfect model airplane, only knowing that you're going to have to wind it up and let it fly, and maybe never see it again. [NEWLINE] [NEWLINE] So, no, I don't think there are only selfish reasons.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The internet is a giant idea filter. I mean, think about<mask> we all do with it. Aside from looking at funny gifs and pirating our favorite TV shows we spend a pretty disproportionate amount of time being confronted with our own ignorance's and biases and confronting those of others. [NEWLINE] [NEWLINE] The end result is that over time ideas in the real world become drastically filtered through the lens of all these ideas being heavily debated on the internet. Think of yourself in an internet vacuum. You've never been introduced to the internet.<mask> many of the ideas you hold dear aren't there anymore?<mask> much of your social value system is gone? [NEWLINE] [NEWLINE] Me personally, a whole lot of my values and morality are informed by conversations I've had with people online. A whole lot of it actually. Then I go out into the world with those same concepts and educate people who spend less time online like myself. I vote based on those concepts. I'm willing to fight for some of them. I'm willing to change myself completely for others. [NEWLINE] [NEWLINE] /r/CMV is a big extension of that concept of the internet<mask> an idea filter. It's just a small part of the massive debate going on online right now. Just one debate forum among many.<mask> nonetheless places like this are exceptionally important to the overall progress of human society. [NEWLINE] [NEWLINE] It's an idea I've had for a<mask> now and one that<mask><mask> will be strengthened<mask> time goes on and we see society change in a myriad of ways and at a much quicker rate than it has in generations past.</s>
Label encoding: <s>The internet is a giant idea filter. I mean, think about what we all do with it. Aside from looking at funny gifs and pirating our favorite TV shows we spend a pretty disproportionate amount of time being confronted with our own ignorance's and biases and confronting those of others. [NEWLINE] [NEWLINE] The end result is that over time ideas in the real world become drastically filtered through the lens of all these ideas being heavily debated on the internet. Think of yourself in an internet vacuum. You've never been introduced to the internet. How many of the ideas you hold dear aren't there anymore? How much of your social value system is gone? [NEWLINE] [NEWLINE] Me personally, a whole lot of my values and morality are informed by conversations I've had with people online. A whole lot of it actually. Then I go out into the world with those same concepts and educate people who spend less time online like myself. I vote based on those concepts. I'm willing to fight for some of them. I'm willing to change myself completely for others. [NEWLINE] [NEWLINE] /r/CMV is a big extension of that concept of the internet as an idea filter. It's just a small part of the massive debate going on online right now. Just one debate forum among many. But nonetheless places like this are exceptionally important to the overall progress of human society. [NEWLINE] [NEWLINE] It's an idea I've had for a while now and one that I think will be strengthened as time goes on and we see society change in a myriad of ways and at a much quicker rate than it has in generations past.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; = to you. Hi klw, thanks for diving a bit deeper into<mask> hyperrealism really is. I'll say that my view *has* changed based on discussion yesterday,<mask><mask> I would've responded to you before that happened is:<mask><mask><mask> that the process is certainly beneficial to the artist, it should be irrelevant to the viewer. In other words, pointless to the viewer, not the artist.<mask> I were to see the photograph of the red head, it would be about her.<mask><mask><mask><mask><mask> I see the hyperrealistic drawing of said red head, the art is no longer about her,<mask> about the process of having created the piece to begin with (the photograph could easily be adjusted to match the colors he used, and I would say the other changes he made don't change the subject matter substantially<mask><mask> ).<mask> I originally argued, I don't think it should be about the process,<mask> about the subject matter. [NEWLINE] [NEWLINE] Well consider my view changed.<mask><mask> now that art cannot be viewed without context,<mask> it's an essential part, whatever the context is.<mask><mask> I concede now that hyperrealism isn't pointless, I might still<mask><mask> it's *merely* a contemplation on technique (at least for the viewer;<mask> you say, it can server many purposes for the artist).<mask> such, I find it rather uninteresting, boring and unimaginative,<mask> that would simply fall in the realm of my personal opinions. [NEWLINE] [NEWLINE] Thanks for the reply!</s>
Label encoding: <s>&amp;#8710; = to you. Hi klw, thanks for diving a bit deeper into what hyperrealism really is. I'll say that my view *has* changed based on discussion yesterday, but what I would've responded to you before that happened is: while I agree that the process is certainly beneficial to the artist, it should be irrelevant to the viewer. In other words, pointless to the viewer, not the artist. If I were to see the photograph of the red head, it would be about her. If on the other hand I see the hyperrealistic drawing of said red head, the art is no longer about her, but about the process of having created the piece to begin with (the photograph could easily be adjusted to match the colors he used, and I would say the other changes he made don't change the subject matter substantially IMHO ). As I originally argued, I don't think it should be about the process, but about the subject matter. [NEWLINE] [NEWLINE] Well consider my view changed. I think now that art cannot be viewed without context, so it's an essential part, whatever the context is. So while I concede now that hyperrealism isn't pointless, I might still argue that it's *merely* a contemplation on technique (at least for the viewer; as you say, it can server many purposes for the artist). As such, I find it rather uninteresting, boring and unimaginative, but that would simply fall in the realm of my personal opinions. [NEWLINE] [NEWLINE] Thanks for the reply!</s>
Loss: tensor(0.0610, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well right now there isn't a plague of people saying "<mask> I wanted him to do that to me"<mask> the other person is being punished.<mask> most illegal things you would want to do to you cannot be prosecuted without cooperation from the victim (rape and other sexual misconduct, assault, theft...). The big exception being murder, or rather assisted suicide. That is a different debate (and one in which I expect we are, mostly, on the same side of)<mask> even<mask> that is legal I would want to be sure the decision is made in a very safe place, like a hospital, not in a back alley or on a roof top. [NEWLINE] [NEWLINE] <mask> I see this change you propose cause far more harm than it eliminates. I see too many people taking advantage of it<mask> I don't see anyone suffering for its lack. [NEWLINE] [NEWLINE] For examples of currently existing solutions... [NEWLINE] Theft with permission is a gift [NEWLINE] Assault with permission could happen in a boxing ring or a martial arts contest [NEWLINE] Arson... well I'm pretty sure it is illegal to burn your own home (at least in cities) in places<mask> it is legal all the 'victim' would have to say is that they did it themselves and no problem [NEWLINE] Assisted suicide... technically illegal,<mask> nothing illegal about leaving a bottle of nitrogen and an oxygen mask next to someone who knows<mask> to connect them and put the mask on (maybe I'm wrong) [NEWLINE] [NEWLINE] Do you have specific examples of things you want to be legal with permission that there isn't already a workaround for?</s>
Label encoding: <s>Well right now there isn't a plague of people saying " But I wanted him to do that to me" while the other person is being punished. Besides most illegal things you would want to do to you cannot be prosecuted without cooperation from the victim (rape and other sexual misconduct, assault, theft...). The big exception being murder, or rather assisted suicide. That is a different debate (and one in which I expect we are, mostly, on the same side of) but even if that is legal I would want to be sure the decision is made in a very safe place, like a hospital, not in a back alley or on a roof top. [NEWLINE] [NEWLINE] So I see this change you propose cause far more harm than it eliminates. I see too many people taking advantage of it while I don't see anyone suffering for its lack. [NEWLINE] [NEWLINE] For examples of currently existing solutions... [NEWLINE] Theft with permission is a gift [NEWLINE] Assault with permission could happen in a boxing ring or a martial arts contest [NEWLINE] Arson... well I'm pretty sure it is illegal to burn your own home (at least in cities) in places where it is legal all the 'victim' would have to say is that they did it themselves and no problem [NEWLINE] Assisted suicide... technically illegal, but nothing illegal about leaving a bottle of nitrogen and an oxygen mask next to someone who knows how to connect them and put the mask on (maybe I'm wrong) [NEWLINE] [NEWLINE] Do you have specific examples of things you want to be legal with permission that there isn't already a workaround for?</s>
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It is not a myth that there are a couple billion too many people on Earth. We can fix the distribution problem with immigration reform. [ENDQ] [NEWLINE] <mask>, there are too many people,<mask> this can be fixed by...immigration reform? Even<mask> true, that makes this point about overpopulation irrelevant to having kids. [NEWLINE] [NEWLINE] [STARTQ] <mask> this is not taking into account that we are in the age of dwindling fossil fuels. Not only are the critical resources running out, they are creating wars and poisoning the environment. [ENDQ] [NEWLINE] <mask>, resource scarcity is another problem,<mask> immigration reform isn't the fix.<mask> about fossil fuel alternatives?<mask> the opportunity cost of acquiring new resources exceeds the cost of alternate resources, we'll find something better. We always do. We survived before petrochems. [NEWLINE] [NEWLINE] [STARTQ] <mask> are your statistics on this? I feel like I know more people who have been royally fucked up by their parents than not. [ENDQ] [NEWLINE] This is a tu quoque, and both arguments can be dismissed with Newton's flaming laser sword. Still, you might be trying to let the perfect be the enemy of the good. [NEWLINE] [NEWLINE] [STARTQ] I don't just care about myself. I care about everyone who is... [ENDQ] [NEWLINE] <mask>, you don't care about: [NEWLINE] [NEWLINE] * Yourself [NEWLINE] * Future generations [NEWLINE] * The idea of humanity [NEWLINE] [NEWLINE] <mask>, you do care about existing people. Maybe you could just be satisfied that you're not super excited about babies, and who would want an ambivalent parent? That seems like a fair argument.</s>
Label encoding: <s> [STARTQ] It is not a myth that there are a couple billion too many people on Earth. We can fix the distribution problem with immigration reform. [ENDQ] [NEWLINE] So, there are too many people, but this can be fixed by...immigration reform? Even if true, that makes this point about overpopulation irrelevant to having kids. [NEWLINE] [NEWLINE] [STARTQ] But this is not taking into account that we are in the age of dwindling fossil fuels. Not only are the critical resources running out, they are creating wars and poisoning the environment. [ENDQ] [NEWLINE] So, resource scarcity is another problem, but immigration reform isn't the fix. How about fossil fuel alternatives? When the opportunity cost of acquiring new resources exceeds the cost of alternate resources, we'll find something better. We always do. We survived before petrochems. [NEWLINE] [NEWLINE] [STARTQ] What are your statistics on this? I feel like I know more people who have been royally fucked up by their parents than not. [ENDQ] [NEWLINE] This is a tu quoque, and both arguments can be dismissed with Newton's flaming laser sword. Still, you might be trying to let the perfect be the enemy of the good. [NEWLINE] [NEWLINE] [STARTQ] I don't just care about myself. I care about everyone who is... [ENDQ] [NEWLINE] So, you don't care about: [NEWLINE] [NEWLINE] * Yourself [NEWLINE] * Future generations [NEWLINE] * The idea of humanity [NEWLINE] [NEWLINE] But, you do care about existing people. Maybe you could just be satisfied that you're not super excited about babies, and who would want an ambivalent parent? That seems like a fair argument.</s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] All of the circumstances that cause a circumstancial depression can actually cause the imbalance in the brain chemistry, can it not? [ENDQ] [NEWLINE] Can it?  I can't find anything that suggests that it can... [NEWLINE] [NEWLINE] [STARTQ] The antidepressants will then be useful to help you deal with said circumstances.<mask> you do not help these people (with medication or not), they risk just spiraling further into depression.<mask> you catch it<mask> it is not too bad, you can prevent them from getting really ill. [ENDQ] [NEWLINE] I'll humor you<mask> you haven't responded to the first part<mask> :  Even<mask> these circumstances can cause a chemical imbalance, that would mean we have to rethink<mask> we deal with antidepressants.  There would be no excuse for prescribing antidepressants without therapy<mask> the goal would have to be normalizing people and getting them off the drugs, not keep them on them for life<mask><mask> depression were a chronic disease. [NEWLINE] [NEWLINE] [STARTQ] <mask> this imbalance is not present in all cases of clinical depression, should we just not medically treat the people who show no imbalance,<mask> have every symptom of depression? [ENDQ] [NEWLINE] <mask><mask><mask> alternative treatments have been ruled out, I don't see a problem with trialing antidepressants and monitoring patients for side-effects under these circumstances. <mask> you brought up, we know very little about<mask> the brain works; I would put this in the "people who might really need it" category,<mask> with great caution. <mask><mask> that the symptoms of depression alone aren't enough to properly determine a need for antidepressants.</s>
Label encoding: <s> [STARTQ] All of the circumstances that cause a circumstancial depression can actually cause the imbalance in the brain chemistry, can it not? [ENDQ] [NEWLINE] Can it?  I can't find anything that suggests that it can... [NEWLINE] [NEWLINE] [STARTQ] The antidepressants will then be useful to help you deal with said circumstances. If you do not help these people (with medication or not), they risk just spiraling further into depression. If you catch it while it is not too bad, you can prevent them from getting really ill. [ENDQ] [NEWLINE] I'll humor you as you haven't responded to the first part yet :  Even if these circumstances can cause a chemical imbalance, that would mean we have to rethink how we deal with antidepressants.  There would be no excuse for prescribing antidepressants without therapy as the goal would have to be normalizing people and getting them off the drugs, not keep them on them for life as if depression were a chronic disease. [NEWLINE] [NEWLINE] [STARTQ] If this imbalance is not present in all cases of clinical depression, should we just not medically treat the people who show no imbalance, but have every symptom of depression? [ENDQ] [NEWLINE] As long as alternative treatments have been ruled out, I don't see a problem with trialing antidepressants and monitoring patients for side-effects under these circumstances.  As you brought up, we know very little about how the brain works; I would put this in the "people who might really need it" category, but with great caution.  I think that the symptoms of depression alone aren't enough to properly determine a need for antidepressants.</s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wow thanks for all the tips, I'll look into those.  I've been studying philosophy, neurology, and really anything concerning metaphysics (more recently).  Not concerning this topic specifically<mask> I'm reading "The Master Key System" by Charles Haanel and he lays out an amazing understanding and explanation of our consciousness to subconscious to solar plexus interaction....the qualia of all your physiology combined to create consciousness.  The book is more directed at<mask> he calls "right thought" and<mask> using this understanding of your system you are far more effective at life through greater understanding, insight and clarity. [NEWLINE] [NEWLINE] I don't perceive this<mask> mystical anymore based off of the understanding of his work.  I would definitely recommend the book, the pdf is online for download.  I see humans<mask> a solid integrated system<mask> there is no dichotomy of mind/body<mask> unfortunately with our large brains we've developed a way to observe ourselves in a way and that we usually call "ego"......I see it<mask> a whole system of course<mask> I remember a time in my life<mask> I felt<mask><mask> there was a dichotomy.....<mask> it's hard to explain to some people sometimes<mask> usually they are in that first state of growth of consciousness<mask> they are identified with some kind of dichotomy. <mask>, all this can be misunderstood easily<mask> we are using very large concepts and peoples understanding of them may be different than intended.  I hope any of this made sense haha  Anyways, thanks for the response!</s>
Label encoding: <s>Wow thanks for all the tips, I'll look into those.  I've been studying philosophy, neurology, and really anything concerning metaphysics (more recently).  Not concerning this topic specifically but I'm reading "The Master Key System" by Charles Haanel and he lays out an amazing understanding and explanation of our consciousness to subconscious to solar plexus interaction....the qualia of all your physiology combined to create consciousness.  The book is more directed at what he calls "right thought" and how using this understanding of your system you are far more effective at life through greater understanding, insight and clarity. [NEWLINE] [NEWLINE] I don't perceive this as mystical anymore based off of the understanding of his work.  I would definitely recommend the book, the pdf is online for download.  I see humans as a solid integrated system where there is no dichotomy of mind/body but unfortunately with our large brains we've developed a way to observe ourselves in a way and that we usually call "ego"......I see it as a whole system of course but I remember a time in my life when I felt as if there was a dichotomy..... so it's hard to explain to some people sometimes because usually they are in that first state of growth of consciousness where they are identified with some kind of dichotomy.  Also, all this can be misunderstood easily because we are using very large concepts and peoples understanding of them may be different than intended.  I hope any of this made sense haha  Anyways, thanks for the response!</s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No, magical answers are out. [ENDQ] [NEWLINE] <mask>? [NEWLINE] [NEWLINE] [STARTQ] much better than systemic racism,<mask> I've already addressed poverty. [ENDQ] [NEWLINE] <mask><mask>'s your answer? [NEWLINE] [NEWLINE] [STARTQ] except they aren't all in poverty. the white poverty rate is about 10%, the black about 25. Is that good? absolutely not,<mask> it doesn't explain the crime pattern. ANd<mask> you want more poverty programs fine, just make them race neutral. [ENDQ] [NEWLINE] Most of the people who commit crime in this country are in poverty, across racial lines. [NEWLINE] [NEWLINE] The black people who aren't in poverty aren't committing the crime. [NEWLINE] [NEWLINE] [STARTQ] It is literally the definition of discrimination! It is giving people bonus points based on their race. [ENDQ] [NEWLINE] No, it is not literally the definition of discrimination. [NEWLINE] [NEWLINE] [STARTQ] Well, that's simply not true.<mask> is true is that it is that<mask>ians suffer far more, the slots that they would get go to blacks instead.<mask> that just makes affirmative action even more problematic.<mask> on earth is rewarding blacks at the expense of<mask>ians, who have their own long history of racial mistreatment, in any way just? [ENDQ] [NEWLINE] It simply is true. Oh no? White guy didn't make it into Harvard<mask> of affirmative action? Turns out his future is just as bright as it would have been anyway. [NEWLINE] [NEWLINE] The whole Asian thing is just a crutch for people like you to lean on.<mask> fine, lets institute affirmative action for Asian people to alleviate this issue.</s>
Label encoding: <s> [STARTQ] No, magical answers are out. [ENDQ] [NEWLINE] What? [NEWLINE] [NEWLINE] [STARTQ] much better than systemic racism, but I've already addressed poverty. [ENDQ] [NEWLINE] So what's your answer? [NEWLINE] [NEWLINE] [STARTQ] except they aren't all in poverty. the white poverty rate is about 10%, the black about 25. Is that good? absolutely not, but it doesn't explain the crime pattern. ANd if you want more poverty programs fine, just make them race neutral. [ENDQ] [NEWLINE] Most of the people who commit crime in this country are in poverty, across racial lines. [NEWLINE] [NEWLINE] The black people who aren't in poverty aren't committing the crime. [NEWLINE] [NEWLINE] [STARTQ] It is literally the definition of discrimination! It is giving people bonus points based on their race. [ENDQ] [NEWLINE] No, it is not literally the definition of discrimination. [NEWLINE] [NEWLINE] [STARTQ] Well, that's simply not true. What is true is that it is that asians suffer far more, the slots that they would get go to blacks instead. but that just makes affirmative action even more problematic. how on earth is rewarding blacks at the expense of asians, who have their own long history of racial mistreatment, in any way just? [ENDQ] [NEWLINE] It simply is true. Oh no? White guy didn't make it into Harvard because of affirmative action? Turns out his future is just as bright as it would have been anyway. [NEWLINE] [NEWLINE] The whole Asian thing is just a crutch for people like you to lean on. So fine, lets institute affirmative action for Asian people to alleviate this issue.</s>
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They eat poo. [ENDQ] [NEWLINE] Other than being gross (and not all dogs eat shit),<mask> is this a problem to dog ownership<mask> it does not bother someone? [NEWLINE] [NEWLINE] [STARTQ] They are annoying. [ENDQ] [NEWLINE] Personal opinion. Dogs that are properly trained are roughly<mask> annoying to the general public<mask> a child. Should children not be allowed on buses?<mask>, there are all kinds of wild animals that poop in the park<mask> well,<mask> it should be noted that most owners pick up after their animals (making it seem like your problem is more with bad owners than dogs) which leads to the next problem. [NEWLINE] [NEWLINE] <mask> no one were to own dogs,<mask> would happen to the millions out there currently? Surely it is better to have dogs owned than to either kill them all or let them run wild. Many dogs would not know<mask> to even survive in the wild. Most dogs enjoy being owned. [NEWLINE] [NEWLINE] Your point about dangerous dogs it pretty weak considering the risks inherent in nearly anything. [NEWLINE] [NEWLINE] Dogs eating does not directly take away from people eating food.<mask> everyone who was paying to feed a dog would alternatively feed a starving person, then your point would stand,<mask> that just isn't the case. [NEWLINE] [NEWLINE] Your point about greenhouse gases could<mask> be used to<mask><mask> no one should have children (I mean, it's probably worse<mask> that kid might grow up to have children and all of them might keep dogs),<mask> people still do.<mask> makes having children more acceptable than owning a dog? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] They eat poo. [ENDQ] [NEWLINE] Other than being gross (and not all dogs eat shit), why is this a problem to dog ownership if it does not bother someone? [NEWLINE] [NEWLINE] [STARTQ] They are annoying. [ENDQ] [NEWLINE] Personal opinion. Dogs that are properly trained are roughly as annoying to the general public as a child. Should children not be allowed on buses? Additionally, there are all kinds of wild animals that poop in the park as well, but it should be noted that most owners pick up after their animals (making it seem like your problem is more with bad owners than dogs) which leads to the next problem. [NEWLINE] [NEWLINE] If no one were to own dogs, what would happen to the millions out there currently? Surely it is better to have dogs owned than to either kill them all or let them run wild. Many dogs would not know how to even survive in the wild. Most dogs enjoy being owned. [NEWLINE] [NEWLINE] Your point about dangerous dogs it pretty weak considering the risks inherent in nearly anything. [NEWLINE] [NEWLINE] Dogs eating does not directly take away from people eating food. If everyone who was paying to feed a dog would alternatively feed a starving person, then your point would stand, but that just isn't the case. [NEWLINE] [NEWLINE] Your point about greenhouse gases could also be used to argue that no one should have children (I mean, it's probably worse since that kid might grow up to have children and all of them might keep dogs), yet people still do. What makes having children more acceptable than owning a dog? [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Except<mask> you're contrasting is conservative and *progressive*, not conservative and liberal.  Progressives and liberals often have a lot of overlap (in part<mask> early liberalism was by nature a progressive change from the status quo),<mask> they're not the exact same thing. <mask>, neither progressives nor liberals tend to change things just for the sake of changing them: they may be much more *open* to the idea of reinterpreting traditional understandings of Constitutional rights,<mask> that still doesn't say anything about a given example.  Basically, you're using a dictionary definition of "conservative" and a colloquial definition of "liberal"<mask> OP's entire argument was about the original definition of liberalism and its historical focus on civil rights.  Sure,<mask> we conflate "liberal" and "progressive" then restricting gun rights can fall under that umbrella,<mask> at that point<mask> could making it illegal for the religious to hold public office.  (Needless to say, I don't think that would end up being considered a liberal position.) [NEWLINE] [NEWLINE] Perhaps it would help<mask> I rephrased the question: by my understanding,<mask> OP is asking is "<mask> is the restriction of this particular civil right now considered to be a liberal idea,<mask> in virtually all other cases liberals expand and defend civil rights?"  Yes, change can be considered liberal,<mask> there are lots of thing that liberals have no interest in changing and, on this sort of thing, they traditionally advocate a very different sort of change.</s>
Label encoding: <s>Except what you're contrasting is conservative and *progressive*, not conservative and liberal.  Progressives and liberals often have a lot of overlap (in part because early liberalism was by nature a progressive change from the status quo), but they're not the exact same thing.  Moreover, neither progressives nor liberals tend to change things just for the sake of changing them: they may be much more *open* to the idea of reinterpreting traditional understandings of Constitutional rights, but that still doesn't say anything about a given example.  Basically, you're using a dictionary definition of "conservative" and a colloquial definition of "liberal" when OP's entire argument was about the original definition of liberalism and its historical focus on civil rights.  Sure, if we conflate "liberal" and "progressive" then restricting gun rights can fall under that umbrella, but at that point so could making it illegal for the religious to hold public office.  (Needless to say, I don't think that would end up being considered a liberal position.) [NEWLINE] [NEWLINE] Perhaps it would help if I rephrased the question: by my understanding, what OP is asking is " why is the restriction of this particular civil right now considered to be a liberal idea, yet in virtually all other cases liberals expand and defend civil rights?"  Yes, change can be considered liberal, but there are lots of thing that liberals have no interest in changing and, on this sort of thing, they traditionally advocate a very different sort of change.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Whether or not he's been published, I'd<mask><mask> one of the goals of science beyond just discovering the truth of the universe is spreading awareness of these truths to the rest of the human race. Bill Nye may not have done<mask> much of the former,<mask> he has done a ton of the latter. [ENDQ] [NEWLINE] By this rationale, [Morgan Freeman is a scientist too.]( [URL] /) [NEWLINE] Further, the scientific method says absolutely nothing about'spreading truth'. That's philosophy, not science. [NEWLINE] [NEWLINE] [STARTQ] Are you adverse to things having a marketing department or a PR department? [ENDQ] [NEWLINE] No, I work in marketing myself. I build websites for an ad firm. I use analytic data in doing<mask>, and make conscious and substantiated claims about that data: I'm not a scientist for it. I'm a web-developer. [NEWLINE] [NEWLINE] You worked in IT, and have an understanding of the technology: that doesn't make you a scientist. I don't consider myself a 'computer scientist'<mask> I can program javascript and C#. I certainly don't think it makes me qualified to speak on the topic of the internet or computer science at large,<mask> you'd find me doing<mask><mask> you dug through my history. I just don't do it in any official capacity, nor would I let anyone claim I was an expert in a thing to back up their theories. By presenting him<mask> an argument, the President was asserting he was an authority on the subject<mask> a scientist, and he wasn't.</s>
Label encoding: <s> [STARTQ] Whether or not he's been published, I'd argue that one of the goals of science beyond just discovering the truth of the universe is spreading awareness of these truths to the rest of the human race. Bill Nye may not have done as much of the former, but he has done a ton of the latter. [ENDQ] [NEWLINE] By this rationale, [Morgan Freeman is a scientist too.]( [URL] /) [NEWLINE] Further, the scientific method says absolutely nothing about'spreading truth'. That's philosophy, not science. [NEWLINE] [NEWLINE] [STARTQ] Are you adverse to things having a marketing department or a PR department? [ENDQ] [NEWLINE] No, I work in marketing myself. I build websites for an ad firm. I use analytic data in doing so, and make conscious and substantiated claims about that data: I'm not a scientist for it. I'm a web-developer. [NEWLINE] [NEWLINE] You worked in IT, and have an understanding of the technology: that doesn't make you a scientist. I don't consider myself a 'computer scientist' because I can program javascript and C#. I certainly don't think it makes me qualified to speak on the topic of the internet or computer science at large, though you'd find me doing so if you dug through my history. I just don't do it in any official capacity, nor would I let anyone claim I was an expert in a thing to back up their theories. By presenting him as an argument, the President was asserting he was an authority on the subject as a scientist, and he wasn't.</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Much of the value of performance art - especially this type of avant-garde performance art - comes from its unexpected and even shocking qualities.  It can act<mask> social commentary.  It can force the audience to ask themselves<mask> art *is*, and it offers them little or no guidance to make that judgement. [NEWLINE] [NEWLINE] <mask> this particular act has been done before... and it's been done several times.  Marina Abramovic did it.  Yoko Ono did it.  A whole slew of art students have done s worksimilar.  Abramovic's performance pretty firmly accepted<mask> "art" now, at least by the art community,<mask> the shock value is diminished... and<mask> the "art" value is diminished.  It seems like "edgy" artistic endeavors are becoming a way to revive stagnating celebrity careers. [NEWLINE] [NEWLINE] The social experiment value has been muddled<mask> these performances aren't exactly carried out scientifically -<mask>, for instance,<mask> you wanted to draw conclusions about society's objectification of women, or society's objectification of celebrities, or people's willingness to transgress moral boundaries<mask> presented with the opportunity, then you really wouldn't be able to use performance art<mask> "research." <mask> every repetition of this "experiment" just clouds the issue and makes the "experiment" less valuable.   The results have become almost predictable, too... to the point<mask> doing this kind of performance art is kind of like the art equivalent of a high school science fair project.  </s>
Label encoding: <s>Much of the value of performance art - especially this type of avant-garde performance art - comes from its unexpected and even shocking qualities.  It can act as social commentary.  It can force the audience to ask themselves what art *is*, and it offers them little or no guidance to make that judgement. [NEWLINE] [NEWLINE] But this particular act has been done before... and it's been done several times.  Marina Abramovic did it.  Yoko Ono did it.  A whole slew of art students have done s worksimilar.  Abramovic's performance pretty firmly accepted as "art" now, at least by the art community, so the shock value is diminished... and therefore the "art" value is diminished.  It seems like "edgy" artistic endeavors are becoming a way to revive stagnating celebrity careers. [NEWLINE] [NEWLINE] The social experiment value has been muddled because these performances aren't exactly carried out scientifically - so, for instance, if you wanted to draw conclusions about society's objectification of women, or society's objectification of celebrities, or people's willingness to transgress moral boundaries when presented with the opportunity, then you really wouldn't be able to use performance art as "research."  So every repetition of this "experiment" just clouds the issue and makes the "experiment" less valuable.   The results have become almost predictable, too... to the point where doing this kind of performance art is kind of like the art equivalent of a high school science fair project.  </s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Of course, family is more important than job. [ENDQ] [NEWLINE] <mask> that's the case, then maybe that parent should have taken a job with fewer hours or with a schedule more conducive to parenting. [NEWLINE] [NEWLINE] [STARTQ] Dad wants to spend time with his kid [ENDQ] [NEWLINE] Maybe the childless person wants to spend time with a family member who is sick or dying. <mask> is that less valid? [NEWLINE] [NEWLINE] [STARTQ] I would probably value the freetime of a parent over someone childless<mask> they're shaping up a kid for society, they're helping us all [ENDQ] [NEWLINE] *<mask> do you know that the childless person isn't<mask> having an effect on the future of society?  Perhaps he or she is very involved in the lives of nieces/nephews, younger siblings, children of friends?  Just<mask> a person is not a parent directly does not mean that he or she doesn't have influence on the lives of children. [NEWLINE] [NEWLINE] *<mask> do you know that the parent is actually being a GOOD parent? [NEWLINE] [NEWLINE] *<mask> is it the employer's responsibility, and by extension the responsibility of the childless employees, to give special treatment to a person simply<mask> he or she is a parent?  The employer had no say in whether their employee was a parent, the other employees had no say in whether their colleague was a parent.  People who choose to have children are choosing to make a sacrifice - they have no right to expect their employer or colleagues to bear a portion of that sacrifice for them.</s>
Label encoding: <s> [STARTQ] Of course, family is more important than job. [ENDQ] [NEWLINE] If that's the case, then maybe that parent should have taken a job with fewer hours or with a schedule more conducive to parenting. [NEWLINE] [NEWLINE] [STARTQ] Dad wants to spend time with his kid [ENDQ] [NEWLINE] Maybe the childless person wants to spend time with a family member who is sick or dying.  How is that less valid? [NEWLINE] [NEWLINE] [STARTQ] I would probably value the freetime of a parent over someone childless because they're shaping up a kid for society, they're helping us all [ENDQ] [NEWLINE] * How do you know that the childless person isn't also having an effect on the future of society?  Perhaps he or she is very involved in the lives of nieces/nephews, younger siblings, children of friends?  Just because a person is not a parent directly does not mean that he or she doesn't have influence on the lives of children. [NEWLINE] [NEWLINE] * How do you know that the parent is actually being a GOOD parent? [NEWLINE] [NEWLINE] * Why is it the employer's responsibility, and by extension the responsibility of the childless employees, to give special treatment to a person simply because he or she is a parent?  The employer had no say in whether their employee was a parent, the other employees had no say in whether their colleague was a parent.  People who choose to have children are choosing to make a sacrifice - they have no right to expect their employer or colleagues to bear a portion of that sacrifice for them.</s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Infringing on someone's rights is taking something away from them that they previously had. [NEWLINE] [NEWLINE] Not serving someone bacon isn't infringing on a right. Now<mask> he's sitting there eating some, and you walk in and take it away - that would be infringing on his rights. [NEWLINE] [NEWLINE] I mean I'm not going to jerk you off, doesn't mean I'm infringing on your rights to handjobs. [NEWLINE] [NEWLINE] And it's not hurting him either. You have no bacon beforehand, you have no bacon after. [NEWLINE] [NEWLINE] [STARTQ] <mask> not? I'm not forcing him to eat it. I'm not asking him to taste test it for me. All I'm asking him to do is his job. The one he gets paid for.<mask> he doesn't want to cook bacon, maybe he shouldn't have gotten a job cooking bacon? [ENDQ] [NEWLINE] Yeah,<mask><mask> he doesn't want to, he shouldn't be forced to. Yes,<mask><mask> that becoming a breakfast cook would be a bad idea<mask> you refuse to cook bacon,<mask> even<mask> - he has no obligation to you whatsoever.<mask> he doesn't want to serve you he doesn't have to. He's not infringing upon your rights. [NEWLINE] [NEWLINE] I would agree that we need basic rights outlawing discrimination of a few protected categories<mask> it would allow groups of businesses to basically run a racist/sexist/whatever ring,<mask> the idea that<mask> you run a business you should have to break established religious or political views is ridiculous. </s>
Label encoding: <s>Infringing on someone's rights is taking something away from them that they previously had. [NEWLINE] [NEWLINE] Not serving someone bacon isn't infringing on a right. Now if he's sitting there eating some, and you walk in and take it away - that would be infringing on his rights. [NEWLINE] [NEWLINE] I mean I'm not going to jerk you off, doesn't mean I'm infringing on your rights to handjobs. [NEWLINE] [NEWLINE] And it's not hurting him either. You have no bacon beforehand, you have no bacon after. [NEWLINE] [NEWLINE] [STARTQ] Why not? I'm not forcing him to eat it. I'm not asking him to taste test it for me. All I'm asking him to do is his job. The one he gets paid for. If he doesn't want to cook bacon, maybe he shouldn't have gotten a job cooking bacon? [ENDQ] [NEWLINE] Yeah, but if he doesn't want to, he shouldn't be forced to. Yes, I agree that becoming a breakfast cook would be a bad idea if you refuse to cook bacon, but even so - he has no obligation to you whatsoever. If he doesn't want to serve you he doesn't have to. He's not infringing upon your rights. [NEWLINE] [NEWLINE] I would agree that we need basic rights outlawing discrimination of a few protected categories because it would allow groups of businesses to basically run a racist/sexist/whatever ring, but the idea that because you run a business you should have to break established religious or political views is ridiculous. </s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Someone posted a reply here that really drives my point home,<mask> they deleted it. I'm going to post it anyway<mask> I already had the reponse typed up by the time it was gone: [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] &gt;<mask> on top of that, most advice to avoid rape is just terrible. [ENDQ] [NEWLINE] [STARTQ] Not really, saying "<mask> you have to walk home late, go together with someone or go through places<mask> there are a lot of people" is not terrible, it helps women know<mask> to avoid rapists<mask> much<mask> possible. It helps people rather than hurts them [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] That is,<mask><mask>, an example of terrible advice. It makes women feel less safe and doesn't actually give them any more safety. I consider telling people they should live in a constant state of fear a form of harm. It causes anxiety and stress, and those have real, physical effects. Telling a woman that just drives home that they are powerless and weak and always vulnerable. [NEWLINE] [NEWLINE] The vast majority of rapes happen between friends or acquaintances. Some guy grabbing a woman in a dark alleyway and raping her is just not a thing that happens with any frequency.<mask> you want to give women tips on<mask> to not be raped, teach them<mask> to spot creepy behavior of the people they know, and<mask> to get out of situations with friends<mask> they don't feel safe. Teach them it's okay to tell someone they're making her uncomfortable and that she would like to leave.</s>
Label encoding: <s>Someone posted a reply here that really drives my point home, but they deleted it. I'm going to post it anyway because I already had the reponse typed up by the time it was gone: [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] &gt; But on top of that, most advice to avoid rape is just terrible. [ENDQ] [NEWLINE] [STARTQ] Not really, saying " if you have to walk home late, go together with someone or go through places where there are a lot of people" is not terrible, it helps women know how to avoid rapists as much as possible. It helps people rather than hurts them [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] That is, in fact, an example of terrible advice. It makes women feel less safe and doesn't actually give them any more safety. I consider telling people they should live in a constant state of fear a form of harm. It causes anxiety and stress, and those have real, physical effects. Telling a woman that just drives home that they are powerless and weak and always vulnerable. [NEWLINE] [NEWLINE] The vast majority of rapes happen between friends or acquaintances. Some guy grabbing a woman in a dark alleyway and raping her is just not a thing that happens with any frequency. If you want to give women tips on how to not be raped, teach them how to spot creepy behavior of the people they know, and how to get out of situations with friends where they don't feel safe. Teach them it's okay to tell someone they're making her uncomfortable and that she would like to leave.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Pre-licensing statistics will try to get information on the incidence of side effects and the severity.<mask> they might find for instance, 1 in 10,000 normal (i.e. without health conditions which would preclude vaccine use) people have side effects like a headache. And 1 in a 100,000 will have serious side effects like vertigo.<mask> the serious side effects are e.g. heart failure or liver damage, the vaccine won't go to market. [NEWLINE] [NEWLINE] It is then up to the pharmaceutical company to decide whether to put the product up for licensing  and the licensing board to have the final say - is the potential risk worth the benefit. They then follow the drug after it's been released to see<mask> there are any effects (say in a particular ethnic group who might not have have been represented in the clinical trials) that haven't been noted. [NEWLINE] [NEWLINE] Now you were sort of at no more at risk than anyone else and more at risk. Think of it like a peanut allergy - you are more at risk of reacting to peanuts,<mask> within the population in general your risk of having a peanut allergy is no greater than anyone else's. And there's no way to tell until you eat a peanut. [NEWLINE] [NEWLINE] Carrying the metaphor to your next point - not wanting your friends to get this vaccination is like not wanting them to eat peanuts<mask> you had an allergic reaction. Yes, peanuts are harmful to you<mask> their risk of having a peanut allergy is still incredibly low.</s>
Label encoding: <s>Pre-licensing statistics will try to get information on the incidence of side effects and the severity. So they might find for instance, 1 in 10,000 normal (i.e. without health conditions which would preclude vaccine use) people have side effects like a headache. And 1 in a 100,000 will have serious side effects like vertigo. If the serious side effects are e.g. heart failure or liver damage, the vaccine won't go to market. [NEWLINE] [NEWLINE] It is then up to the pharmaceutical company to decide whether to put the product up for licensing  and the licensing board to have the final say - is the potential risk worth the benefit. They then follow the drug after it's been released to see if there are any effects (say in a particular ethnic group who might not have have been represented in the clinical trials) that haven't been noted. [NEWLINE] [NEWLINE] Now you were sort of at no more at risk than anyone else and more at risk. Think of it like a peanut allergy - you are more at risk of reacting to peanuts, but within the population in general your risk of having a peanut allergy is no greater than anyone else's. And there's no way to tell until you eat a peanut. [NEWLINE] [NEWLINE] Carrying the metaphor to your next point - not wanting your friends to get this vaccination is like not wanting them to eat peanuts because you had an allergic reaction. Yes, peanuts are harmful to you but their risk of having a peanut allergy is still incredibly low.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think it's a good idea to base anything on the manifestos of crazy people. <mask><mask> you're equating a scientifically-backed principle with a symbol of a defunct rebel government.  It sounds like<mask> you're saying is that<mask> someone has killed in the name of X, X should not be taught or celebrated. [NEWLINE] [NEWLINE] <mask><mask>,<mask><mask> you'd be hard pressed to find a principle that someone *hasn't* killed someone over.  Just about every religion fits this bill,<mask> well<mask> every governmental system. <mask> you try to suss out the one that hasn't had any blood spilled, you'd be<mask> silly<mask>  that scene from Rain Man in the airport.  (Dustin Hoffman's character won't get on a plane<mask> they all have had crashes.  The only one that hasn't at the time of the film was Qantas, which doesn't go<mask> he needs to be). [NEWLINE] [NEWLINE] Natural selection is the idea that those best fit to their environment will survive through adversity.  That's all.  It's not moral or immoral--it's amoral.  Just<mask> amoral<mask> the idea that water is made up of two hydrogen atoms and one oxygen atom. [NEWLINE] [NEWLINE] The Confederate flag was a symbol used today to harken back to a group of rebels (<mask><mask>, racists).  Whether it symbolizes heritage, hate, or a heritage of hate can be argued,<mask> that isn't scientific.</s>
Label encoding: <s>I don't think it's a good idea to base anything on the manifestos of crazy people.  I think you're equating a scientifically-backed principle with a symbol of a defunct rebel government.  It sounds like what you're saying is that because someone has killed in the name of X, X should not be taught or celebrated. [NEWLINE] [NEWLINE] In fact, I think you'd be hard pressed to find a principle that someone *hasn't* killed someone over.  Just about every religion fits this bill, as well as every governmental system.  If you try to suss out the one that hasn't had any blood spilled, you'd be as silly as  that scene from Rain Man in the airport.  (Dustin Hoffman's character won't get on a plane because they all have had crashes.  The only one that hasn't at the time of the film was Qantas, which doesn't go where he needs to be). [NEWLINE] [NEWLINE] Natural selection is the idea that those best fit to their environment will survive through adversity.  That's all.  It's not moral or immoral--it's amoral.  Just as amoral as the idea that water is made up of two hydrogen atoms and one oxygen atom. [NEWLINE] [NEWLINE] The Confederate flag was a symbol used today to harken back to a group of rebels ( IMO, racists).  Whether it symbolizes heritage, hate, or a heritage of hate can be argued, but that isn't scientific.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What is embracing atheism?<mask> is embracing a lack of a belief? [NEWLINE] [NEWLINE] That gets tired and boring real quick. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] I feel saying "Gun's don't kill people, People with God kill people" is equal to a religious person saying "Guns don't kill people, people without God kill people". [ENDQ] [NEWLINE] Do you not get the point of the former? People have and do kill in the name of their religion and their god(s), and<mask> a million other pointless reasons. [NEWLINE] [NEWLINE] [NEWLINE] Those who are atheists do not kill<mask> they are atheists, they kill<mask> of a million other reasons. [NEWLINE] [NEWLINE] Get the difference? [NEWLINE] [NEWLINE] Good people do good things, bad people do bad things. It takes religion to get good people to do bad things.(paraphrased from [Steven Weinberg]( [URL] )) [NEWLINE] [NEWLINE] Disagree? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It harbours hatred towards a particular group of people; people who hold a different life view than their own. [ENDQ] [NEWLINE] No, the hatred comes from that different life view that is determined on imposing said view on the rest of us. Are you not aware of laws not allowing gays to marry, or constant threat of implementing a law to deny women to choose the option of abortion? [NEWLINE] [NEWLINE] R/atheism is a safe haven. Leave it alone. We are powerless, and don't even have the desire to impose on others,<mask><mask><mask> you don't define fighting religious oppression<mask> imposing on others.</s>
Label encoding: <s>What is embracing atheism? What is embracing a lack of a belief? [NEWLINE] [NEWLINE] That gets tired and boring real quick. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] I feel saying "Gun's don't kill people, People with God kill people" is equal to a religious person saying "Guns don't kill people, people without God kill people". [ENDQ] [NEWLINE] Do you not get the point of the former? People have and do kill in the name of their religion and their god(s), and also a million other pointless reasons. [NEWLINE] [NEWLINE] [NEWLINE] Those who are atheists do not kill because they are atheists, they kill because of a million other reasons. [NEWLINE] [NEWLINE] Get the difference? [NEWLINE] [NEWLINE] Good people do good things, bad people do bad things. It takes religion to get good people to do bad things.(paraphrased from [Steven Weinberg]( [URL] )) [NEWLINE] [NEWLINE] Disagree? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] It harbours hatred towards a particular group of people; people who hold a different life view than their own. [ENDQ] [NEWLINE] No, the hatred comes from that different life view that is determined on imposing said view on the rest of us. Are you not aware of laws not allowing gays to marry, or constant threat of implementing a law to deny women to choose the option of abortion? [NEWLINE] [NEWLINE] R/atheism is a safe haven. Leave it alone. We are powerless, and don't even have the desire to impose on others, as long as you don't define fighting religious oppression as imposing on others.</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There is no such thing<mask> "atheist philosophy". [ENDQ] [NEWLINE] I'll be more clear, then. There are parts of philosophy that deal with religion and the existence of God. These may be interesting for atheists to discuss. [NEWLINE] [NEWLINE] [STARTQ] atheists don't share any common philosophy. We all have our own ideas, our own politics, our own everything that are completely independent of our lack of belief in a god.<mask> there's really just nothing else to talk about. [ENDQ] [NEWLINE] Atheism does not imply a specific worldview,<mask> can you deny that the atheist perspective has led to many of them holding certain ideas about the world, and that these ideas may extend beyond hating on religion? [NEWLINE] [NEWLINE] [STARTQ] Atheism is nothing<mask> a lack of belief. It's not a worldview. It's not its own religion with common teachings.<mask> any discussions we had that were NOT about religion would just be a debate that had nothing to do with atheism. [ENDQ] [NEWLINE] Exactly. Everyone has their own form of atheism. Can one not discuss these things independently of religion? Not everyone stops at "there is no deity." [There are different forms of atheism, each with their own merits.]( [URL] #Concepts) It's narrowminded,<mask><mask><mask>, to have an entire subreddit for atheist discussion and use it to post pictures of stupid Facebook posts and get angry whenever a Muslim cleric rapes a little girl. These are worthy topics, yes,<mask> they are not the only topics. [NEWLINE] </s>
Label encoding: <s> [STARTQ] There is no such thing as "atheist philosophy". [ENDQ] [NEWLINE] I'll be more clear, then. There are parts of philosophy that deal with religion and the existence of God. These may be interesting for atheists to discuss. [NEWLINE] [NEWLINE] [STARTQ] atheists don't share any common philosophy. We all have our own ideas, our own politics, our own everything that are completely independent of our lack of belief in a god. So there's really just nothing else to talk about. [ENDQ] [NEWLINE] Atheism does not imply a specific worldview, but can you deny that the atheist perspective has led to many of them holding certain ideas about the world, and that these ideas may extend beyond hating on religion? [NEWLINE] [NEWLINE] [STARTQ] Atheism is nothing but a lack of belief. It's not a worldview. It's not its own religion with common teachings. So any discussions we had that were NOT about religion would just be a debate that had nothing to do with atheism. [ENDQ] [NEWLINE] Exactly. Everyone has their own form of atheism. Can one not discuss these things independently of religion? Not everyone stops at "there is no deity." [There are different forms of atheism, each with their own merits.]( [URL] #Concepts) It's narrowminded, in my opinion, to have an entire subreddit for atheist discussion and use it to post pictures of stupid Facebook posts and get angry whenever a Muslim cleric rapes a little girl. These are worthy topics, yes, but they are not the only topics. [NEWLINE] </s>
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0537, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Back to your example of long-term guy friends, do you ever feel physical attraction to them, or is that something that never crosses your mind? [ENDQ] [NEWLINE] eh, depends on which one. One I have dated<mask> I wanted to give him a chance<mask> he just wasn't someone I wanted to date. My guy friends have personalities I am cool with being friends with<mask> I don't think I could be anything more with them. They aren't people I could see myself living with. I'd get annoyed by them after spending too much time with them. We don't really agree on certain things I would find important<mask> I was gonna be more than friends with them. [NEWLINE] [NEWLINE] Some of them vary too differently politically and I don't really want that in a husband.<mask> I don't care about it in a friend. [NEWLINE] [NEWLINE] Some of them are too irresponsible and I don't want that in a husband<mask> in a friend, who cares. I don't care who the breadwinner is<mask> I want someone else who can be responsible<mask> I need them to be. [NEWLINE] [NEWLINE] Some of them have anger issues or can be controlling. It doesn't really affect our friendship,<mask> with my parents, I know I am never marrying someone like that. [NEWLINE] [NEWLINE] Im<mask> not going to waste time dating someone I know I won't marry. And I'm not sex with someone I'm not serious about.<mask> my guy friends will remain my guy friends<mask> nothing more. </s>
Label encoding: <s> [STARTQ] Back to your example of long-term guy friends, do you ever feel physical attraction to them, or is that something that never crosses your mind? [ENDQ] [NEWLINE] eh, depends on which one. One I have dated because I wanted to give him a chance but he just wasn't someone I wanted to date. My guy friends have personalities I am cool with being friends with but I don't think I could be anything more with them. They aren't people I could see myself living with. I'd get annoyed by them after spending too much time with them. We don't really agree on certain things I would find important if I was gonna be more than friends with them. [NEWLINE] [NEWLINE] Some of them vary too differently politically and I don't really want that in a husband. But I don't care about it in a friend. [NEWLINE] [NEWLINE] Some of them are too irresponsible and I don't want that in a husband but in a friend, who cares. I don't care who the breadwinner is but I want someone else who can be responsible when I need them to be. [NEWLINE] [NEWLINE] Some of them have anger issues or can be controlling. It doesn't really affect our friendship, but with my parents, I know I am never marrying someone like that. [NEWLINE] [NEWLINE] Im also not going to waste time dating someone I know I won't marry. And I'm not sex with someone I'm not serious about. So my guy friends will remain my guy friends but nothing more. </s>
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0611, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [NEWLINE] I was born into a poor family with a mother who didn't understand<mask> to spend money wisely,<mask><mask><mask> we were constantly broke and most of my possessions were stolen from one place or another by my dad<mask> he wasn't working. Two of my uncles died<mask><mask><mask> of alcohol poisoning and now I'm watching my younger brother fall into the same traps and drink his life away. My sister owes me £20,000, and has no intention of paying it back after I lent her the money from my savings for a house to help her from being evicted and her possessions being auctioned. I have been bullied from primary school into high school and had to quit college during the last year to work more hours to help keep my family afloat. [NEWLINE] [NEWLINE] And you're telling me that<mask> I'm white I have it on easy mode? I've worked hard every day of my life to ensure that today I and my family have something to show for it. Do you have any idea<mask> insulting it is to hear that? [NEWLINE] [NEWLINE] Yeah, being white means there's a slight increase in my odds of being able to do anything with my life,<mask> don't you dare try to claim that money, class and the environment you're raised in aren't orders of magnitude more critical. Racism exists and it sucks,<mask> trying to claim I have it easy<mask> I'm white and completely dismiss everything else is a joke. [NEWLINE] [NEWLINE] EDIT: removed profanity.</s>
Label encoding: <s> [NEWLINE] [NEWLINE] I was born into a poor family with a mother who didn't understand how to spend money wisely, as a result we were constantly broke and most of my possessions were stolen from one place or another by my dad when he wasn't working. Two of my uncles died as a result of alcohol poisoning and now I'm watching my younger brother fall into the same traps and drink his life away. My sister owes me £20,000, and has no intention of paying it back after I lent her the money from my savings for a house to help her from being evicted and her possessions being auctioned. I have been bullied from primary school into high school and had to quit college during the last year to work more hours to help keep my family afloat. [NEWLINE] [NEWLINE] And you're telling me that because I'm white I have it on easy mode? I've worked hard every day of my life to ensure that today I and my family have something to show for it. Do you have any idea how insulting it is to hear that? [NEWLINE] [NEWLINE] Yeah, being white means there's a slight increase in my odds of being able to do anything with my life, but don't you dare try to claim that money, class and the environment you're raised in aren't orders of magnitude more critical. Racism exists and it sucks, but trying to claim I have it easy because I'm white and completely dismiss everything else is a joke. [NEWLINE] [NEWLINE] EDIT: removed profanity.</s>
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm a male and prefer to be searched by a mtf transgender furry with a Swedish accent.  Come on, I don't think we ought to head down that road.  The fact is, different people will have their own reasons for wanting to be segregated by sex in a public restroom.  I'm told by women that they won't even poop freely<mask> there's someone in the stall next to them.  I'm at a point in my male life<mask> I take care of my business without reservation,<mask> I guarantee I'll be much less comfortable doing<mask> with a female next to me. [NEWLINE] [NEWLINE] There are subtle dynamics at play.  A male restroom has urinals, and I don't think a unisex restroom can keep urinals for reasons that are obvious to me. I like urinals. They are quick, and rather enjoyable compared to a stall. <mask> I'm in a stall in a men's restroom, I know<mask> I'm doing, and I figure any other guy in a stall is taking a shit, too. I'm not that interested in shitting next to a woman who isn't shitting.  Or one who is. [NEWLINE] [NEWLINE] Everyone will have their own reasons to be for or against unisex bathrooms, and<mask> I take a stand against is the idea that any of those personal reasons can fall under the umbrella of harassment.  The matters at stake are practical in nature, not political or philosophical.</s>
Label encoding: <s>I'm a male and prefer to be searched by a mtf transgender furry with a Swedish accent.  Come on, I don't think we ought to head down that road.  The fact is, different people will have their own reasons for wanting to be segregated by sex in a public restroom.  I'm told by women that they won't even poop freely when there's someone in the stall next to them.  I'm at a point in my male life where I take care of my business without reservation, but I guarantee I'll be much less comfortable doing so with a female next to me. [NEWLINE] [NEWLINE] There are subtle dynamics at play.  A male restroom has urinals, and I don't think a unisex restroom can keep urinals for reasons that are obvious to me. I like urinals. They are quick, and rather enjoyable compared to a stall.  If I'm in a stall in a men's restroom, I know what I'm doing, and I figure any other guy in a stall is taking a shit, too. I'm not that interested in shitting next to a woman who isn't shitting.  Or one who is. [NEWLINE] [NEWLINE] Everyone will have their own reasons to be for or against unisex bathrooms, and what I take a stand against is the idea that any of those personal reasons can fall under the umbrella of harassment.  The matters at stake are practical in nature, not political or philosophical.</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll grant you that that sentence was poorly worded and worse argued and that a lot of<mask> you are saying is true. I know that sounds like politician's speak,<mask> really, that sentence was stupid. [NEWLINE] [NEWLINE] <mask>, none of<mask> you say challenges my two main points: [NEWLINE] [NEWLINE] * In the modern diplomatically,<mask> more importantly economically interconnected world, war is no longer the best means to influence other countries, having been replaced by economic pressure and other things. [NEWLINE] * Even<mask> American might was needed to maintain peace, that still does not explain the necesity for it to be *this big*. Can you convincingly<mask><mask> cutting the US defence budget by 40-80 billion dollars would severely harm global safety? [NEWLINE] [NEWLINE] I see the Ukraine conlict<mask> confirming these points, not disproving them. Instead of just straight up invading Ukraine<mask> he would have a hundred years ago, Putin tried to do it 'peacefully'. He tried to make it seem like the will of the local people. He threatened sanctions. He did everything *except declare war*. Only after his hand was forced (the shooting down of that airplane causing international outrage) did he actively intervene,<mask> I have pointed out, at great costs to his own economy. And even<mask> using force in this particular case, he still avoids using it in other circumstances. Instead of threatening force to prevent intervention from europe, he instead is using his control over the flow of gas.</s><pad>
Label encoding: <s>I'll grant you that that sentence was poorly worded and worse argued and that a lot of what you are saying is true. I know that sounds like politician's speak, but really, that sentence was stupid. [NEWLINE] [NEWLINE] However, none of what you say challenges my two main points: [NEWLINE] [NEWLINE] * In the modern diplomatically, but more importantly economically interconnected world, war is no longer the best means to influence other countries, having been replaced by economic pressure and other things. [NEWLINE] * Even if American might was needed to maintain peace, that still does not explain the necesity for it to be *this big*. Can you convincingly argue that cutting the US defence budget by 40-80 billion dollars would severely harm global safety? [NEWLINE] [NEWLINE] I see the Ukraine conlict as confirming these points, not disproving them. Instead of just straight up invading Ukraine as he would have a hundred years ago, Putin tried to do it 'peacefully'. He tried to make it seem like the will of the local people. He threatened sanctions. He did everything *except declare war*. Only after his hand was forced (the shooting down of that airplane causing international outrage) did he actively intervene, as I have pointed out, at great costs to his own economy. And even while using force in this particular case, he still avoids using it in other circumstances. Instead of threatening force to prevent intervention from europe, he instead is using his control over the flow of gas.</s><pad>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most of the points I was going to bring up have been said already: [NEWLINE] [NEWLINE] 1. You own the comic and can read it over again, unlike the movie. Most comics readers place high value on ownership of the object, either<mask> they can read it dozens of times throughout their lives or<mask> they simply value the object and being a collector of a rare or special piece of art. [NEWLINE] [NEWLINE] 2. You're not taking your time with the comic, apparently. Checking out comics criticism might help you appreciate<mask> comics fans are doing<mask> reading, and<mask> they enjoy it<mask> much. [NEWLINE] [NEWLINE] 3. The amount of time, energy and creativity that multiple writers and artists at the top of their game put into each comic is alone worth the price on an economic level, like a hand-made or hand-stitched item that might be priced higher than something factory built. [NEWLINE] [NEWLINE] <mask> I haven't seen anyone point out is<mask> often we spend $3 or more on something that is consumed in far less time than 10-15 minutes. Food may be a necessity,<mask> most food we eat is far more expensive than necessary<mask> we enjoy it more. Rather than eat rice and beans and bread at home, we go out and spend $8-10 at a restaurant. Even most things at a fast food restaurant will be more than $3-4, and will be consumed within minutes. Do you ever buy beer? Coffee at a coffee shop? [NEWLINE] </s>
Label encoding: <s>Most of the points I was going to bring up have been said already: [NEWLINE] [NEWLINE] 1. You own the comic and can read it over again, unlike the movie. Most comics readers place high value on ownership of the object, either because they can read it dozens of times throughout their lives or because they simply value the object and being a collector of a rare or special piece of art. [NEWLINE] [NEWLINE] 2. You're not taking your time with the comic, apparently. Checking out comics criticism might help you appreciate what comics fans are doing while reading, and why they enjoy it so much. [NEWLINE] [NEWLINE] 3. The amount of time, energy and creativity that multiple writers and artists at the top of their game put into each comic is alone worth the price on an economic level, like a hand-made or hand-stitched item that might be priced higher than something factory built. [NEWLINE] [NEWLINE] What I haven't seen anyone point out is how often we spend $3 or more on something that is consumed in far less time than 10-15 minutes. Food may be a necessity, but most food we eat is far more expensive than necessary because we enjoy it more. Rather than eat rice and beans and bread at home, we go out and spend $8-10 at a restaurant. Even most things at a fast food restaurant will be more than $3-4, and will be consumed within minutes. Do you ever buy beer? Coffee at a coffee shop? [NEWLINE] </s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] you are suggesting that in order not to increase the price of a meal in order to pay waiters a guaranteed minimum wage, you should increase the price of a meal automatically by 25% in order to pay a decent tip to a server [ENDQ] [NEWLINE] Not exactly. Servers already get an automatic minimum wage (contrary to<mask> you hear on reddit.)<mask> their tips don't meet that amout, the employer, by law, is obligated to subsidize their pay.<mask>, many states don't allow a separate wage for tipped employees at all (California for instance,)<mask> the server is earning more than minimum wage already. My problem with simply raising the wage and the prices correspondingly only benefits the employer, not the server.<mask> volume is extremely high and the server had to work much harder than usual, the server's pay stays the same<mask> the employer receives more profit.<mask> the increase is a separate line item, the server is directly compensated for their additional effort. [NEWLINE] [NEWLINE] [STARTQ] again<mask><mask> it's telling that the American'server' is preferred to a more European 'waiter') [ENDQ] [NEWLINE] The term "server" is relatively new, much<mask> "flight attendant". We used to use the terms "waiter" and "waitress"<mask> have abandoned those primarily<mask> they are a bit sexist to some. Semantically,<mask>, I don't see a fundamental difference between "waiting" on someone and "serving" them. [NEWLINE] </s>
Label encoding: <s> [STARTQ] you are suggesting that in order not to increase the price of a meal in order to pay waiters a guaranteed minimum wage, you should increase the price of a meal automatically by 25% in order to pay a decent tip to a server [ENDQ] [NEWLINE] Not exactly. Servers already get an automatic minimum wage (contrary to what you hear on reddit.) If their tips don't meet that amout, the employer, by law, is obligated to subsidize their pay. Also, many states don't allow a separate wage for tipped employees at all (California for instance,) so the server is earning more than minimum wage already. My problem with simply raising the wage and the prices correspondingly only benefits the employer, not the server. If volume is extremely high and the server had to work much harder than usual, the server's pay stays the same but the employer receives more profit. If the increase is a separate line item, the server is directly compensated for their additional effort. [NEWLINE] [NEWLINE] [STARTQ] again I think it's telling that the American'server' is preferred to a more European 'waiter') [ENDQ] [NEWLINE] The term "server" is relatively new, much as "flight attendant". We used to use the terms "waiter" and "waitress" but have abandoned those primarily because they are a bit sexist to some. Semantically, though, I don't see a fundamental difference between "waiting" on someone and "serving" them. [NEWLINE] </s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Lots of people have already given you links to serious discussions about the causes and consequences of climate change,<mask> allow me to point out a few things myself.  Glaciers are melting all over the world, and except for the two very largest, which are the polar ice caps, the world's glaciers are almost entirely gone, and the ice caps are in the process of melting.  The sea level has risen a bit,<mask> not a large amount<mask><mask> ; it will continue to rise, and coastal flooding will increasingly be a problem. [NEWLINE] The greenhouse effect itself, the fact that certain substances such<mask> glass, used in actual greenhouses, or carbon dioxide in the atmosphere, are transparent to visible wavelengths of light<mask> opaque to infrared radiation and<mask> retain warmth, is very well confirmed by scientific study, and is not in doubt.  The fact that the atmospheric concentration of carbon dioxide has increased substantially in recent decades is<mask> not in doubt.  Only the conclusion has been questioned, that the climate of the Earth will be affected by the increased greenhouse effect caused by the increased amount of greenhouse gases. <mask> this conclusion is inevitable. <mask> you cannot question that the greenhouse effect itself is real, and you cannot question that the quantity of greenhouse gases in the atmosphere has increased (and you really cannot question those things, they are undisputed, observable facts) then I really do not see<mask> you can question the conclusion that the Earth's climate is being affected.</s>
Label encoding: <s>Lots of people have already given you links to serious discussions about the causes and consequences of climate change, but allow me to point out a few things myself.  Glaciers are melting all over the world, and except for the two very largest, which are the polar ice caps, the world's glaciers are almost entirely gone, and the ice caps are in the process of melting.  The sea level has risen a bit, but not a large amount as yet ; it will continue to rise, and coastal flooding will increasingly be a problem. [NEWLINE] The greenhouse effect itself, the fact that certain substances such as glass, used in actual greenhouses, or carbon dioxide in the atmosphere, are transparent to visible wavelengths of light but opaque to infrared radiation and therefore retain warmth, is very well confirmed by scientific study, and is not in doubt.  The fact that the atmospheric concentration of carbon dioxide has increased substantially in recent decades is also not in doubt.  Only the conclusion has been questioned, that the climate of the Earth will be affected by the increased greenhouse effect caused by the increased amount of greenhouse gases.  Yet this conclusion is inevitable.  If you cannot question that the greenhouse effect itself is real, and you cannot question that the quantity of greenhouse gases in the atmosphere has increased (and you really cannot question those things, they are undisputed, observable facts) then I really do not see how you can question the conclusion that the Earth's climate is being affected.</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0647, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry, yeah I kind of forgot to clarify<mask> it relates, its the first time I've used this analogy online (not in person). [NEWLINE] [NEWLINE] "<mask> i do have enough time to enter the house save anyone who might be inside,<mask> is the an issue in the first place?" [NEWLINE] This is my point. We have the ability to save the baby (in most cases).<mask> it's worth the extra effort (childbirth) even<mask> we're not sure<mask> there's really life in it<mask>.<mask><mask> the pro-life / pro-choice argument centers on whether or not you view a fetus<mask> a human being or not. The burning house is meant to illustrate that even<mask> you're not sure that there is life in there to be saved, its worth the effort to save it just in case. [NEWLINE] [NEWLINE] [NEWLINE] "Would you run into a burning house<mask> you thought that no one was inside? Is it right to stop someone from running into a probably-empty burning house<mask> they are not a trained firefighter? (<mask> they don't find anyone could they could burn to death instead of no one)?" This is<mask> the line become a little fuzzier for me. Is it worth saving a possible life at the risk of losing a definite life?<mask><mask> anytime there is a chance to save both lives, that is the route that should be taken,<mask> I'm much more understanding of the pro-choice argument in this context.</s>
Label encoding: <s>Sorry, yeah I kind of forgot to clarify how it relates, its the first time I've used this analogy online (not in person). [NEWLINE] [NEWLINE] " If i do have enough time to enter the house save anyone who might be inside, why is the an issue in the first place?" [NEWLINE] This is my point. We have the ability to save the baby (in most cases). So it's worth the extra effort (childbirth) even if we're not sure if there's really life in it yet. I think the pro-life / pro-choice argument centers on whether or not you view a fetus as a human being or not. The burning house is meant to illustrate that even if you're not sure that there is life in there to be saved, its worth the effort to save it just in case. [NEWLINE] [NEWLINE] [NEWLINE] "Would you run into a burning house if you thought that no one was inside? Is it right to stop someone from running into a probably-empty burning house if they are not a trained firefighter? ( If they don't find anyone could they could burn to death instead of no one)?" This is where the line become a little fuzzier for me. Is it worth saving a possible life at the risk of losing a definite life? I think anytime there is a chance to save both lives, that is the route that should be taken, but I'm much more understanding of the pro-choice argument in this context.</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ok...<mask> they predicted "some harm". For scale they predicted a stubbed toe, in the prevention of a broken knee...<mask> they received was a heart attack, broken knees and stubbed toes. [NEWLINE] [NEWLINE] Not to mention it made the Federal government into "the bad guys",<mask> much<mask> not more<mask> than ending slavery. Considering Jim Crow was in full effect and the Klan was being heralded<mask> "Red blooded Americans"... [NEWLINE] [NEWLINE] By making the Federal Government "The bad guys" the push towards a more Laissez Faire government gained traction. (Unless you wanted a drink.. then the G-men were all about hands on.) Which is at the least a contributing factor to the financial realities of the 1930's. [NEWLINE] [NEWLINE] I'm not saying that prohibition caused the depression,<mask> it did help to create the political climate that allowed it. [NEWLINE] [NEWLINE] Which is something that makes sense in hindsight,<mask> people are simply too short sighted to see it before it happens. [NEWLINE] [NEWLINE] Knowing this is a weakness of human beings the Constitutional Congress intentionally made it difficult to change the constitution in part to mitigate the negative effects of that weakness. [NEWLINE] [NEWLINE] That there is a way to legally and non-violently change the constitution at all is an admission of the Conventions limitations in creating a document that could survive the changes their own shortsightedness. [NEWLINE] [NEWLINE] Not a disastrous oversight at all, instead a spectacularly effective foresight. </s>
Label encoding: <s>ok... so they predicted "some harm". For scale they predicted a stubbed toe, in the prevention of a broken knee... what they received was a heart attack, broken knees and stubbed toes. [NEWLINE] [NEWLINE] Not to mention it made the Federal government into "the bad guys", as much if not more so than ending slavery. Considering Jim Crow was in full effect and the Klan was being heralded as "Red blooded Americans"... [NEWLINE] [NEWLINE] By making the Federal Government "The bad guys" the push towards a more Laissez Faire government gained traction. (Unless you wanted a drink.. then the G-men were all about hands on.) Which is at the least a contributing factor to the financial realities of the 1930's. [NEWLINE] [NEWLINE] I'm not saying that prohibition caused the depression, but it did help to create the political climate that allowed it. [NEWLINE] [NEWLINE] Which is something that makes sense in hindsight, but people are simply too short sighted to see it before it happens. [NEWLINE] [NEWLINE] Knowing this is a weakness of human beings the Constitutional Congress intentionally made it difficult to change the constitution in part to mitigate the negative effects of that weakness. [NEWLINE] [NEWLINE] That there is a way to legally and non-violently change the constitution at all is an admission of the Conventions limitations in creating a document that could survive the changes their own shortsightedness. [NEWLINE] [NEWLINE] Not a disastrous oversight at all, instead a spectacularly effective foresight. </s>
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If conservatives are underrepresented in academia (specifically the sciences), that has more to do with their lack of interest in it then anything else. I don't think any college application asks "are you conservative or liberal?" and then awards more points<mask> you are liberal. I can understand wanting scholarships for people with conservative principles (which already exist),<mask> I don't think we should award any kind of affirmative action (<mask> I'm against affirmative action of any kind except for low socio-economic status). For the most part, there are fewer conservative scientists<mask> they are more interested in studying the bible then studying science (MOST of the time, not ALL the time). I do think its EXTREMELY important to get conservatives to study academics, real history (vs. revisionist history) and science..<mask> I don't think affirmative action will do the trick.<mask><mask> social pressure within churches to find FACTS would do the trick -<mask> that is never going to happen. [NEWLINE] [NEWLINE] The one thing<mask><mask> with is that an academic environment with more opposing views is always a good thing.<mask> I don't think affirmative action is going to change that. Social pressure within the conservative community will change it. [NEWLINE] [NEWLINE] The other thing I want to bring up is that another reason conservatives are underrepresented in academia is<mask> a LOT of the time (not ALL the time) people who enter academia with conservative values change their minds once they are educated. </s>
Label encoding: <s>If conservatives are underrepresented in academia (specifically the sciences), that has more to do with their lack of interest in it then anything else. I don't think any college application asks "are you conservative or liberal?" and then awards more points if you are liberal. I can understand wanting scholarships for people with conservative principles (which already exist), but I don't think we should award any kind of affirmative action ( though I'm against affirmative action of any kind except for low socio-economic status). For the most part, there are fewer conservative scientists because they are more interested in studying the bible then studying science (MOST of the time, not ALL the time). I do think its EXTREMELY important to get conservatives to study academics, real history (vs. revisionist history) and science.. but I don't think affirmative action will do the trick. I think social pressure within churches to find FACTS would do the trick - but that is never going to happen. [NEWLINE] [NEWLINE] The one thing I agree with is that an academic environment with more opposing views is always a good thing. But I don't think affirmative action is going to change that. Social pressure within the conservative community will change it. [NEWLINE] [NEWLINE] The other thing I want to bring up is that another reason conservatives are underrepresented in academia is because a LOT of the time (not ALL the time) people who enter academia with conservative values change their minds once they are educated. </s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the "vote brigading" of SRS is overstated. [NEWLINE] [NEWLINE] SRS has it's own bot that takes a screen shot<mask> comments are linked SRS. This<mask><mask> is an act of good faith in itself. [NEWLINE] [NEWLINE] no. 1 on SRS atm is [this]( [URL] /), and the vote bot says [this]( [URL] ). [NEWLINE] [NEWLINE] The post has been deleted,<mask> reddit enhancement suit implies that the post was at 232|47 (~= 185 upvotes)<mask> deleted. The child post to the the linked post is<mask> problematic, and<mask> it has gone from 4 points to 100 (136|32) points<mask> it has been posted to SRS. [NEWLINE] [NEWLINE] SRS has 41,493 subscribers, the SRS post has 69 comments and 87 upvotes (at the time I'm posting). [NEWLINE] [NEWLINE] <mask>, neither of the "problematic" posts have over 50 downvotes. I don't think these posts have been "brigaded". [NEWLINE] [NEWLINE] Sometimes, non-SRSers downvote things<mask> they become more visible. I'm not subscribed to SRS,<mask> I downvote that kind of stuff<mask> I see it. You cant assume that every downvote came from SRS. [NEWLINE] [NEWLINE] (Yes, I realise that reddit enhancement suite fudges the numbers. I still think it is a relevant rough estimate of the downvotes and upvotes of a post).</s>
Label encoding: <s>I think the "vote brigading" of SRS is overstated. [NEWLINE] [NEWLINE] SRS has it's own bot that takes a screen shot when comments are linked SRS. This IMO is an act of good faith in itself. [NEWLINE] [NEWLINE] no. 1 on SRS atm is [this]( [URL] /), and the vote bot says [this]( [URL] ). [NEWLINE] [NEWLINE] The post has been deleted, but reddit enhancement suit implies that the post was at 232|47 (~= 185 upvotes) when deleted. The child post to the the linked post is also problematic, and yet it has gone from 4 points to 100 (136|32) points since it has been posted to SRS. [NEWLINE] [NEWLINE] SRS has 41,493 subscribers, the SRS post has 69 comments and 87 upvotes (at the time I'm posting). [NEWLINE] [NEWLINE] Yet, neither of the "problematic" posts have over 50 downvotes. I don't think these posts have been "brigaded". [NEWLINE] [NEWLINE] Sometimes, non-SRSers downvote things when they become more visible. I'm not subscribed to SRS, but I downvote that kind of stuff when I see it. You cant assume that every downvote came from SRS. [NEWLINE] [NEWLINE] (Yes, I realise that reddit enhancement suite fudges the numbers. I still think it is a relevant rough estimate of the downvotes and upvotes of a post).</s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> 3D is a perfect extension of pure-CG films [ENDQ] [NEWLINE] Not for all of them.<mask> I went to re-image one of the Blender Open Movies I had to choose between "Elephants Dream" and "Big Buck Bunny". My choice was "Elephants Dream" then,<mask> its format and visuals stroke me<mask> far better suited for 3D than BBB. And the stunning results confirmed this. [NEWLINE] [NEWLINE] Interestingly enough, many of the methods I developed to improve the experience really did pay off. I got incredible response from a lot of people who told me, that usually they get nausea in 3D movies, even<mask> the 3D effect is only weak,<mask> in my version of "Elephants Dream 3D"<mask> the strong 3D they could watch it comfortably without getting simulator/3D sickness. [NEWLINE] [NEWLINE] [STARTQ] <mask> it's something that can be added to the film without too much effort [ENDQ] [NEWLINE] Having done single handedly a complete re-imaged of "Elephants Dream" into steroscopic 3D myself I can tell you, that it's not a simple click of a button. Many things require special care to work in 3D. For example 2D matte paintings must be turned into 3D counterparts. Yes<mask> in CGI matte paintings are used. And you have to do carefull stereoscopic direction to match the stereoscopy with the scene and the action.</s>
Label encoding: <s> [STARTQ] I think 3D is a perfect extension of pure-CG films [ENDQ] [NEWLINE] Not for all of them. When I went to re-image one of the Blender Open Movies I had to choose between "Elephants Dream" and "Big Buck Bunny". My choice was "Elephants Dream" then, because its format and visuals stroke me as far better suited for 3D than BBB. And the stunning results confirmed this. [NEWLINE] [NEWLINE] Interestingly enough, many of the methods I developed to improve the experience really did pay off. I got incredible response from a lot of people who told me, that usually they get nausea in 3D movies, even if the 3D effect is only weak, but in my version of "Elephants Dream 3D" despite the strong 3D they could watch it comfortably without getting simulator/3D sickness. [NEWLINE] [NEWLINE] [STARTQ] as it's something that can be added to the film without too much effort [ENDQ] [NEWLINE] Having done single handedly a complete re-imaged of "Elephants Dream" into steroscopic 3D myself I can tell you, that it's not a simple click of a button. Many things require special care to work in 3D. For example 2D matte paintings must be turned into 3D counterparts. Yes also in CGI matte paintings are used. And you have to do carefull stereoscopic direction to match the stereoscopy with the scene and the action.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I would image that<mask> there were any credible claims that Jesus never existed, we would certainly hear about it, wouldn't we? [ENDQ] [NEWLINE] <mask> would you expect that out of curiosity? Even in the most advanced age of science nobody can run an article that says "GOD IS DEAD/NONEXISTENT" for several reasons: [NEWLINE] [NEWLINE] * Severe backlash from all religious entities [NEWLINE] * Religious people and Christians in particular hold to their beliefs in an unswaying manner<mask> unless something is explicitly proven and in front of their eyes they can believe the opposite.<mask> there is no way to disprove the existence of God, they often cling to the "god of gaps", or a god that hasn't been *quite* disproven<mask>. [NEWLINE] [NEWLINE] <mask> an example of the god of gaps,<mask> growing up in private Christian school it was oft cited that the fossil record was incomplete and there just weren't enough of man's homo ancestors fossils/remains discovered to prove that evolution was real. Now,<mask> a large chunk of those gaps have been filled in it's much more difficult for them to use that argument,<mask> they just drift off to another one fluidly.<mask>, for the same reason you can't sit down with a religious person and have them accept that God doesn't exist, the Jesus doesn't exist argument has been out there for quite some time<mask> largely ignored by those it would impact the most.</s>
Label encoding: <s> [STARTQ] I would image that if there were any credible claims that Jesus never existed, we would certainly hear about it, wouldn't we? [ENDQ] [NEWLINE] Why would you expect that out of curiosity? Even in the most advanced age of science nobody can run an article that says "GOD IS DEAD/NONEXISTENT" for several reasons: [NEWLINE] [NEWLINE] * Severe backlash from all religious entities [NEWLINE] * Religious people and Christians in particular hold to their beliefs in an unswaying manner because unless something is explicitly proven and in front of their eyes they can believe the opposite. As there is no way to disprove the existence of God, they often cling to the "god of gaps", or a god that hasn't been *quite* disproven yet. [NEWLINE] [NEWLINE] As an example of the god of gaps, when growing up in private Christian school it was oft cited that the fossil record was incomplete and there just weren't enough of man's homo ancestors fossils/remains discovered to prove that evolution was real. Now, since a large chunk of those gaps have been filled in it's much more difficult for them to use that argument, so they just drift off to another one fluidly. So, for the same reason you can't sit down with a religious person and have them accept that God doesn't exist, the Jesus doesn't exist argument has been out there for quite some time but largely ignored by those it would impact the most.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0755, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Quotes from Carey's piece: [NEWLINE] [NEWLINE] "controversial vaccine issue" [NEWLINE] [NEWLINE] There isn't a controversy other than the one invented by the anti-vax movement to try and get air time.  The whole tone of the article follows from this assertion in the first paragraph. [NEWLINE] [NEWLINE] "In this growing crisis" [NEWLINE] [NEWLINE] See<mask> I mean?  There is no crisis<mask> they want people to think that there is. [NEWLINE] [NEWLINE] "The anecdotal evidence... must be seriously considered" [NEWLINE] [NEWLINE] <mask><mask> there is mountains of empirical evidence? [NEWLINE] [NEWLINE] "The legitimate concern they and many in the scientific community have that environmental toxins, including those found in vaccines, may be causing autism and other disorders" [NEWLINE] [NEWLINE] These are not legitimate concerns.  This is propaganda. [NEWLINE] [NEWLINE] "With all the doubt that's left hanging on this topic" [NEWLINE] [NEWLINE] There is little room for doubt<mask> you look at the facts. Carey's article is full of "may be" and "<mask> " and "might",<mask> woefully short on facts (and woefully high in misrepresentation of information!). [NEWLINE] [NEWLINE] There is plenty more<mask><mask><mask> I have made point. [NEWLINE] [NEWLINE] I am not surprised that people get angry about it and voice that anger on reddit and elsewhere.  Children (and adults) are suffering needlessly<mask> of the anti-vax movement and that is something that is pretty damn easy to get angry about. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Quotes from Carey's piece: [NEWLINE] [NEWLINE] "controversial vaccine issue" [NEWLINE] [NEWLINE] There isn't a controversy other than the one invented by the anti-vax movement to try and get air time.  The whole tone of the article follows from this assertion in the first paragraph. [NEWLINE] [NEWLINE] "In this growing crisis" [NEWLINE] [NEWLINE] See what I mean?  There is no crisis but they want people to think that there is. [NEWLINE] [NEWLINE] "The anecdotal evidence... must be seriously considered" [NEWLINE] [NEWLINE] Why when there is mountains of empirical evidence? [NEWLINE] [NEWLINE] "The legitimate concern they and many in the scientific community have that environmental toxins, including those found in vaccines, may be causing autism and other disorders" [NEWLINE] [NEWLINE] These are not legitimate concerns.  This is propaganda. [NEWLINE] [NEWLINE] "With all the doubt that's left hanging on this topic" [NEWLINE] [NEWLINE] There is little room for doubt if you look at the facts. Carey's article is full of "may be" and " if " and "might", but woefully short on facts (and woefully high in misrepresentation of information!). [NEWLINE] [NEWLINE] There is plenty more but I think I have made point. [NEWLINE] [NEWLINE] I am not surprised that people get angry about it and voice that anger on reddit and elsewhere.  Children (and adults) are suffering needlessly because of the anti-vax movement and that is something that is pretty damn easy to get angry about. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 538, isn't it owned by ESPN? Just another oddsmaker with suspect impartiality. [ENDQ] [NEWLINE] Huh? 538 is run by Nate Silver. His group has been affiliated with a bunch of different things. He was independent at one point, and then he was working for the New York Times, and is now connected to ESPN.<mask><mask> is that at all relevant? His entire brand rests on giving accurate statistical analysis. [NEWLINE] [NEWLINE] <mask>,<mask> makes you think an oddsmaker will have suspect impartiality? And<mask> you really do think there's something suspect here,<mask> not just look at their track record? [NEWLINE] [NEWLINE] [STARTQ] By liberal, I mean someone who works to make things better for the American people, not someone who talks about it<mask> only helps her special interests, like Wall Street banks. [ENDQ] [NEWLINE] And<mask> makes you think Hillary doesn't fall into that category? Can you define it more specifically or are is exactly the sort of vague narrative based  notion that cannot be possibly given any sort of clear meaning? [NEWLINE] [NEWLINE] You know, I find this really funny. Back in 2012,<mask> Silver was saying things that everyone on the left wanted to hear, they were very happy to talk about his data, and the right was happy to bash him.<mask><mask> soon<mask> he says something that one doesn't like, apparently, that's the time to start attacking his credibility. Humans are fascinatingly inconsistent. </s>
Label encoding: <s> [STARTQ] 538, isn't it owned by ESPN? Just another oddsmaker with suspect impartiality. [ENDQ] [NEWLINE] Huh? 538 is run by Nate Silver. His group has been affiliated with a bunch of different things. He was independent at one point, and then he was working for the New York Times, and is now connected to ESPN. But how is that at all relevant? His entire brand rests on giving accurate statistical analysis. [NEWLINE] [NEWLINE] Also, what makes you think an oddsmaker will have suspect impartiality? And if you really do think there's something suspect here, why not just look at their track record? [NEWLINE] [NEWLINE] [STARTQ] By liberal, I mean someone who works to make things better for the American people, not someone who talks about it but only helps her special interests, like Wall Street banks. [ENDQ] [NEWLINE] And what makes you think Hillary doesn't fall into that category? Can you define it more specifically or are is exactly the sort of vague narrative based  notion that cannot be possibly given any sort of clear meaning? [NEWLINE] [NEWLINE] You know, I find this really funny. Back in 2012, when Silver was saying things that everyone on the left wanted to hear, they were very happy to talk about his data, and the right was happy to bash him. Yet as soon as he says something that one doesn't like, apparently, that's the time to start attacking his credibility. Humans are fascinatingly inconsistent. </s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This system will institutionalize oppression of the vulnerable.  Those who are politically informed tend to be more educated, wealthy and powerful members of society.  People who are uninformed tend to be uneducated, poor and weaker members of society. [NEWLINE] [NEWLINE] Your system would directly lead to a disproportional representation<mask> wealthy, educated and stronger members of society will achieve more political power and influence.  Likewise, those on the bottom will get even less. [NEWLINE] [NEWLINE] At first it might seem right that someone who is politically ignorant gets punished for that ignorance (by not being able to vote)<mask> such a position assumes an overly simplified reason for<mask> people are politically ignorant.   Some people our naturally less intelligent than others, popular media is designed to influence the minds of people who haven't been trained to think critically, political issues are inherently complicated.  There are all kinds of reasons other than simple laziness that would prevent people from passing a test<mask> you describe. [NEWLINE] [NEWLINE] <mask> we all know,<mask> only a portion of the populace is represented in government, you don't have real democracy.  The uneducated voter has always been a weakness of democracy<mask> it is a necessary one.  Without it, you no longer have democracy and you move to a form of fascism/oligarchy/caste system. [NEWLINE] [NEWLINE] <mask> Winston Churchill said "Democracy is the worst form of government, except for all the other ones."</s>
Label encoding: <s>This system will institutionalize oppression of the vulnerable.  Those who are politically informed tend to be more educated, wealthy and powerful members of society.  People who are uninformed tend to be uneducated, poor and weaker members of society. [NEWLINE] [NEWLINE] Your system would directly lead to a disproportional representation as wealthy, educated and stronger members of society will achieve more political power and influence.  Likewise, those on the bottom will get even less. [NEWLINE] [NEWLINE] At first it might seem right that someone who is politically ignorant gets punished for that ignorance (by not being able to vote) but such a position assumes an overly simplified reason for why people are politically ignorant.   Some people our naturally less intelligent than others, popular media is designed to influence the minds of people who haven't been trained to think critically, political issues are inherently complicated.  There are all kinds of reasons other than simple laziness that would prevent people from passing a test as you describe. [NEWLINE] [NEWLINE] As we all know, if only a portion of the populace is represented in government, you don't have real democracy.  The uneducated voter has always been a weakness of democracy but it is a necessary one.  Without it, you no longer have democracy and you move to a form of fascism/oligarchy/caste system. [NEWLINE] [NEWLINE] As Winston Churchill said "Democracy is the worst form of government, except for all the other ones."</s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0639, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the flaw in your judgement is your quantity (time) = quality (cost). [NEWLINE] [NEWLINE] Entertainment isn't necessarily about<mask> much time it takes<mask> rather,<mask> much enjoyment you took from the experience. [NEWLINE] [NEWLINE] <mask> you subtract your time:cost ratio, you're left with the more simple question; Do I enjoy this form of entertainment enough to justify spending $3.50 on another one? [NEWLINE] [NEWLINE] <mask> maybe you just don't enjoy reading comics all that much. There's nothing wrong with that. I've never really been into them either<mask> I know many people who absolutely adore all things comic. Which brings up another point. [NEWLINE] [NEWLINE] The folks who really enjoy them, are likely to read them more than once. Could be<mask> simple<mask> reading the previous issue<mask> they get a new one, or they might read entire series start to finish multiple times much like one would with a favorite book. The point is that they've bought a physical object. Something tangible that they can keep and enjoy anytime they wish. And that revisit of a past issue is much more approachable than with a 500 page novel. [NEWLINE] [NEWLINE] <mask> you take this into account, you may even be able to find peace with your cost:time comparison. After all, you pay $15.00 to watch a 2 hour movie once.<mask> you pay $3.50 to read a comic<mask> often<mask> you like. </s>
Label encoding: <s>I think the flaw in your judgement is your quantity (time) = quality (cost). [NEWLINE] [NEWLINE] Entertainment isn't necessarily about how much time it takes but rather, how much enjoyment you took from the experience. [NEWLINE] [NEWLINE] When you subtract your time:cost ratio, you're left with the more simple question; Do I enjoy this form of entertainment enough to justify spending $3.50 on another one? [NEWLINE] [NEWLINE] So maybe you just don't enjoy reading comics all that much. There's nothing wrong with that. I've never really been into them either but I know many people who absolutely adore all things comic. Which brings up another point. [NEWLINE] [NEWLINE] The folks who really enjoy them, are likely to read them more than once. Could be as simple as reading the previous issue when they get a new one, or they might read entire series start to finish multiple times much like one would with a favorite book. The point is that they've bought a physical object. Something tangible that they can keep and enjoy anytime they wish. And that revisit of a past issue is much more approachable than with a 500 page novel. [NEWLINE] [NEWLINE] When you take this into account, you may even be able to find peace with your cost:time comparison. After all, you pay $15.00 to watch a 2 hour movie once. Yet you pay $3.50 to read a comic as often as you like. </s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You don't have to defend football to me, it's<mask> my favorite sport.  It's just weird to explain<mask> "illegal shift" or "illegal formation"  to the layman.  This is no barrier in soccer. [NEWLINE] [NEWLINE] You get some really good coordinated plays.  These guys practice together.  A lot.  They are very good at knowing<mask> each one is going to be at any time (without seeing them), and getting the ball to their teammates. <mask> you look up "soccer highlights" you'll probably mostly see goals with incredibly individual effort,<mask> sometimes the [teamwork goals]( [URL] ) are the most impressive. [NEWLINE] [NEWLINE] You<mask> see a lot of organized plays on [free kicks]( [URL] ).  Keep in mind these all virtually no margin of error and must be executed perfectly. [NEWLINE] [NEWLINE] <mask> there is a considerable amount of coordination and teammork going on in soccer, that<mask><mask> any American viewer can appreciate. <mask> I said in my parent comment, soccer itself isn't incompatible with American audiences, its the lack of association with teams that kills it.  Americans watch the World Cup in droves,<mask> it hasn't transferred much to the local leagues, the same reason<mask> minor league teams draw small crowds to their big games.  The "World Stage" is in Europe, unlike every other popular sport in the US with a strong fanbase.</s>
Label encoding: <s>You don't have to defend football to me, it's also my favorite sport.  It's just weird to explain what "illegal shift" or "illegal formation"  to the layman.  This is no barrier in soccer. [NEWLINE] [NEWLINE] You get some really good coordinated plays.  These guys practice together.  A lot.  They are very good at knowing where each one is going to be at any time (without seeing them), and getting the ball to their teammates.  If you look up "soccer highlights" you'll probably mostly see goals with incredibly individual effort, but sometimes the [teamwork goals]( [URL] ) are the most impressive. [NEWLINE] [NEWLINE] You also see a lot of organized plays on [free kicks]( [URL] ).  Keep in mind these all virtually no margin of error and must be executed perfectly. [NEWLINE] [NEWLINE] So there is a considerable amount of coordination and teammork going on in soccer, that I think any American viewer can appreciate.  As I said in my parent comment, soccer itself isn't incompatible with American audiences, its the lack of association with teams that kills it.  Americans watch the World Cup in droves, but it hasn't transferred much to the local leagues, the same reason why minor league teams draw small crowds to their big games.  The "World Stage" is in Europe, unlike every other popular sport in the US with a strong fanbase.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your objectives would be much better served by (1) providing a more rigorous science education and (2) creating an independent, nonpartisan advising body of scientists (some sort of Congressional Office of Science and Technology) like the CBO, with the sole job of producing nonpartisan assessments of major science, technology and environmental concerns for congress to analyze. [NEWLINE] [NEWLINE] Elected officials and elites in general tend to be the product of the society in which they reside. That is, the people in power got there in part through some combination of exploiting the system, naturally being a good match for the system and just being lucky. [NEWLINE] [NEWLINE] A better educated populous will naturally elect a representative body with a more scientifically reasonable agenda. [NEWLINE] [NEWLINE] <mask><mask>, it is important to protect scientists from patronage (producing the results that politicians want them to produce, for the sake of serving their agendas). The scientific community is good at policing itself,<mask> there will usually be that one unethical PhD willing to take money to say whatever a politician wants them to say. [NEWLINE] [NEWLINE] Centrally, we need to be vigilant to ensure that people like [John Holdren]( [URL] ) don't get there positions through political games (or at least minimize the extent to which they do). [NEWLINE] [NEWLINE] By the way, I'm not making any accusations about Dr. Holdren, I don't actually know that much about him and<mask> I do know is favorable.</s>
Label encoding: <s>I think your objectives would be much better served by (1) providing a more rigorous science education and (2) creating an independent, nonpartisan advising body of scientists (some sort of Congressional Office of Science and Technology) like the CBO, with the sole job of producing nonpartisan assessments of major science, technology and environmental concerns for congress to analyze. [NEWLINE] [NEWLINE] Elected officials and elites in general tend to be the product of the society in which they reside. That is, the people in power got there in part through some combination of exploiting the system, naturally being a good match for the system and just being lucky. [NEWLINE] [NEWLINE] A better educated populous will naturally elect a representative body with a more scientifically reasonable agenda. [NEWLINE] [NEWLINE] In addition, it is important to protect scientists from patronage (producing the results that politicians want them to produce, for the sake of serving their agendas). The scientific community is good at policing itself, but there will usually be that one unethical PhD willing to take money to say whatever a politician wants them to say. [NEWLINE] [NEWLINE] Centrally, we need to be vigilant to ensure that people like [John Holdren]( [URL] ) don't get there positions through political games (or at least minimize the extent to which they do). [NEWLINE] [NEWLINE] By the way, I'm not making any accusations about Dr. Holdren, I don't actually know that much about him and what I do know is favorable.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was in a depression for years without medication. I don't know<mask> I finally got out of mine,<mask> I was in college and deciding to work part time and take 1 class really helped. For me, I was<mask> unhappy with my environment that a change really helped. I<mask> smoked some marijuana,<mask> I don't know<mask> that helped (I ended up quitting after 6 months<mask> it gave me panic attacks).<mask> really fixed it was finishing college entirely. It was such a toxic environment for me. [NEWLINE] [NEWLINE] <mask> you're interested in a non-pharmaceutical method first, try /r/eood (exercise out of depression). [NEWLINE] [NEWLINE] <mask>, fuck those people who just tell you to get over it. It doesn't work like that. The people minimalizing your depression don't have to live with it.  There's no one solution and you can't just snap out of it. I wouldn't rule out antidepressants (<mask> I ever get depressed again I would try them),<mask><mask> you don't like them you could try other options like therapy or exercise first. This is your health and happinesses. There's no reason to disregard medicine that could help you<mask> of other people's ignorance. Depression takes away<mask> much already. The first step you should take is to stand up for yourself and your own life to get the hell you need, whatever that may be.</s>
Label encoding: <s>I was in a depression for years without medication. I don't know how I finally got out of mine, but I was in college and deciding to work part time and take 1 class really helped. For me, I was so unhappy with my environment that a change really helped. I also smoked some marijuana, but I don't know if that helped (I ended up quitting after 6 months because it gave me panic attacks). What really fixed it was finishing college entirely. It was such a toxic environment for me. [NEWLINE] [NEWLINE] If you're interested in a non-pharmaceutical method first, try /r/eood (exercise out of depression). [NEWLINE] [NEWLINE] Also, fuck those people who just tell you to get over it. It doesn't work like that. The people minimalizing your depression don't have to live with it.  There's no one solution and you can't just snap out of it. I wouldn't rule out antidepressants ( if I ever get depressed again I would try them), but if you don't like them you could try other options like therapy or exercise first. This is your health and happinesses. There's no reason to disregard medicine that could help you because of other people's ignorance. Depression takes away so much already. The first step you should take is to stand up for yourself and your own life to get the hell you need, whatever that may be.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I thought a little more about this, and<mask><mask> I want to add that our main point of contention seems to arise from whether or not the idea of positive/negative rights is still a real 'thing' or issue that exists. Negative rights make sense theoretically<mask> aren't really practicable. Positive rights don't really make sense theoretically<mask> are the best form of practice. [NEWLINE] [NEWLINE] <mask>, I would need to go read up on this again<mask><mask><mask> that another reason for maintaining the conceptualization of negative rights (ie, not that they're just a way to 'aim low') goes back to Foucauldian/biopower stuff.<mask><mask> the basic argument was (and I'm sure this is very simplified) that putting the obligation on government to provide positive goods gives the state more ability and discretion to control the allocation, access, and distribution of those goods (basically...<mask> the state giveth the state taketh away), whereas at least maintaining a conception of negative rights demarcates a clear boundary of<mask> the state cannot take away and cannot ever infringe upon. [NEWLINE] [NEWLINE] Then again, this is sounding dangerously close to libertarianism and I do just want to reiterate<mask><mask> all of this negative rights stuff is pretty stupid in practice.<mask> I like to bring it up at least<mask> the OP has put up for debate<mask> the concept of a 'right' should be.</s>
Label encoding: <s>I thought a little more about this, and I think I want to add that our main point of contention seems to arise from whether or not the idea of positive/negative rights is still a real 'thing' or issue that exists. Negative rights make sense theoretically but aren't really practicable. Positive rights don't really make sense theoretically but are the best form of practice. [NEWLINE] [NEWLINE] Also, I would need to go read up on this again but I think that another reason for maintaining the conceptualization of negative rights (ie, not that they're just a way to 'aim low') goes back to Foucauldian/biopower stuff. I think the basic argument was (and I'm sure this is very simplified) that putting the obligation on government to provide positive goods gives the state more ability and discretion to control the allocation, access, and distribution of those goods (basically... what the state giveth the state taketh away), whereas at least maintaining a conception of negative rights demarcates a clear boundary of what the state cannot take away and cannot ever infringe upon. [NEWLINE] [NEWLINE] Then again, this is sounding dangerously close to libertarianism and I do just want to reiterate I think all of this negative rights stuff is pretty stupid in practice. But I like to bring it up at least when the OP has put up for debate what the concept of a 'right' should be.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We are a nation of laws, not of emotions. Are we really comfortable with replacing the rule of law with judging acts by whether they are "right?" [NEWLINE] [NEWLINE] The law defined his act<mask> illegal.<mask>, the law should apply *blindly.* Some people feel that<mask> he did was immoral,<mask> others feel that it was the height of morality.<mask> in a country of laws that doesn't matter -- we aim to apply the law evenly -- that is,<mask> you're driving at 80 mph in a 40 mph zone then you're guilty of speeding whether its your birthday, you are late work, you're drunk, you're driving to your wife's delivery or you're just in a bad mood. [NEWLINE] [NEWLINE] Take it a bit further. Even<mask> you're comfortable with allowing someone immunity from lawbreaking<mask> he's doing something "right,"<mask><mask><mask> he's wrong?<mask><mask> he THOUGHT it was right,<mask> it was actually wrong? [NEWLINE] [NEWLINE] And he judges<mask> is "right?"<mask> does right even mean? [NEWLINE] [NEWLINE] <mask> about those that think it is "right" to kill abortion doctors? Or those that think it is right to bomb innocent civilians instead of military targets. [NEWLINE] [NEWLINE] That's<mask> we have laws. And that's<mask> you're guilty for breaking them. Even<mask> you think you're doing the "right" thing. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>We are a nation of laws, not of emotions. Are we really comfortable with replacing the rule of law with judging acts by whether they are "right?" [NEWLINE] [NEWLINE] The law defined his act as illegal. Therefore, the law should apply *blindly.* Some people feel that what he did was immoral, while others feel that it was the height of morality. But in a country of laws that doesn't matter -- we aim to apply the law evenly -- that is, if you're driving at 80 mph in a 40 mph zone then you're guilty of speeding whether its your birthday, you are late work, you're drunk, you're driving to your wife's delivery or you're just in a bad mood. [NEWLINE] [NEWLINE] Take it a bit further. Even if you're comfortable with allowing someone immunity from lawbreaking because he's doing something "right," but what if he's wrong? What if he THOUGHT it was right, but it was actually wrong? [NEWLINE] [NEWLINE] And he judges what is "right?" What does right even mean? [NEWLINE] [NEWLINE] What about those that think it is "right" to kill abortion doctors? Or those that think it is right to bomb innocent civilians instead of military targets. [NEWLINE] [NEWLINE] That's why we have laws. And that's why you're guilty for breaking them. Even if you think you're doing the "right" thing. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Parents will either lose or give up the child, sending them to orphanages already bursting at seams. [ENDQ] [NEWLINE] <mask> makes you think we have a problem with too many babies entering orphanages? [NEWLINE] [NEWLINE] That's not a great picture of parentless children in the US at least,<mask> the protesting you're talking about is happening. In the US, any problems with overpopulation in orphanages (of which there are relatively few) or the foster system, are really about older children who are not<mask> desirable for adoption. [NEWLINE] [NEWLINE] For newborn babies, there is a massive demand. Only about 14k babies are put up for adoption any given year. For every child adopted, **36** couples remain on the waiting list. The system<mask> it is could absorb a HUGE number of newborn babies. [NEWLINE] [NEWLINE] For a given newborn child, being put up for adoption doesn't create a large chance of being reliant on taxpayers for food. Our system can handle a vast expansion of adoption. Tons of parents are waiting. And those parents who truly want a child are a far better home than activists who you believe should be guilted into caring for a child. [NEWLINE] [NEWLINE] Of course, we don't have enough adoptive parents should abortion drop to zero and unwanted pregnancy remain at the exact same rate,<mask> that's a separate question that I'd be happy to get into<mask> you really like.</s>
Label encoding: <s> [STARTQ] Parents will either lose or give up the child, sending them to orphanages already bursting at seams. [ENDQ] [NEWLINE] What makes you think we have a problem with too many babies entering orphanages? [NEWLINE] [NEWLINE] That's not a great picture of parentless children in the US at least, where the protesting you're talking about is happening. In the US, any problems with overpopulation in orphanages (of which there are relatively few) or the foster system, are really about older children who are not as desirable for adoption. [NEWLINE] [NEWLINE] For newborn babies, there is a massive demand. Only about 14k babies are put up for adoption any given year. For every child adopted, **36** couples remain on the waiting list. The system as it is could absorb a HUGE number of newborn babies. [NEWLINE] [NEWLINE] For a given newborn child, being put up for adoption doesn't create a large chance of being reliant on taxpayers for food. Our system can handle a vast expansion of adoption. Tons of parents are waiting. And those parents who truly want a child are a far better home than activists who you believe should be guilted into caring for a child. [NEWLINE] [NEWLINE] Of course, we don't have enough adoptive parents should abortion drop to zero and unwanted pregnancy remain at the exact same rate, but that's a separate question that I'd be happy to get into if you really like.</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't understand<mask> sexual orientation is separate from other sexual interests.<mask> is a man being interested in men distinctly different than a man being interested in short women or old people. It seems like gender is just one identifier for attraction<mask> it is magnified to define someone's entire orientation.<mask> isn't there an orientation for someone who is only interested in girls who wear glasses? [NEWLINE] [NEWLINE] Everytime I make a comment with this sentiment it gets rigorously downvoted and I am curious<mask> to<mask>. It seems very reasonable to me. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I don't understand how sexual orientation is separate from other sexual interests. How is a man being interested in men distinctly different than a man being interested in short women or old people. It seems like gender is just one identifier for attraction but it is magnified to define someone's entire orientation. Why isn't there an orientation for someone who is only interested in girls who wear glasses? [NEWLINE] [NEWLINE] Everytime I make a comment with this sentiment it gets rigorously downvoted and I am curious as to why. It seems very reasonable to me. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0879, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The idea that the moment you discover it is the moment you stop being<mask> childish. You've never heard of this? [ENDQ] [NEWLINE] Not really - the only context I've seen it in is that you lose some of your childhood innocence<mask> you realize that Santa isn't real. Which I suppose is an analogous concept,<mask> somehow the spin is quite different. [NEWLINE] [NEWLINE] Again, your whole approach seems centered on that moment<mask> the child starts to doubt - correct me<mask> I'm misreading your posts,<mask> you seem to be saying that a big portion of the rationale behind the Santa belief is that moment<mask> the child *loses* the belief. And I just don't see that<mask> being most people's rationale - it's very common to hear statements like "Yeah, I'm really dreading the day my kids ask me<mask> Santa is real," and a lot of the strategies I hear for coping with this are much more like damage control than like a teaching moment. Occasionally I'll hear people planning on the "Well,<mask> do *you* think" approach, which is closer to<mask> you're talking about.<mask> it doesn't seem to be<mask> common. For most parents, the point of Santa is all the stuff that comes *before* the loss of faith; the latter is just an unfortunate<mask> unavoidable side-effect, not the central teaching moment.</s>
Label encoding: <s> [STARTQ] The idea that the moment you discover it is the moment you stop being so childish. You've never heard of this? [ENDQ] [NEWLINE] Not really - the only context I've seen it in is that you lose some of your childhood innocence when you realize that Santa isn't real. Which I suppose is an analogous concept, but somehow the spin is quite different. [NEWLINE] [NEWLINE] Again, your whole approach seems centered on that moment when the child starts to doubt - correct me if I'm misreading your posts, but you seem to be saying that a big portion of the rationale behind the Santa belief is that moment when the child *loses* the belief. And I just don't see that as being most people's rationale - it's very common to hear statements like "Yeah, I'm really dreading the day my kids ask me if Santa is real," and a lot of the strategies I hear for coping with this are much more like damage control than like a teaching moment. Occasionally I'll hear people planning on the "Well, what do *you* think" approach, which is closer to what you're talking about. But it doesn't seem to be as common. For most parents, the point of Santa is all the stuff that comes *before* the loss of faith; the latter is just an unfortunate yet unavoidable side-effect, not the central teaching moment.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to disagree with this<mask> best I can. [NEWLINE] [NEWLINE] Rule of Law is not the standing threat of violence<mask> much<mask> the standing promise of repercussions.  Yes, violence can be, and often is, a part of that,<mask> the threat of violence is a fundamental fact of life.  There is still threat of violence even without laws; the threat that<mask> you perturb someone bigger/stronger/better armed/more martially capable than you, they're going to do you harm.  There's the threat of the (financial) violence of theft.  That threat never goes away, not really. [NEWLINE] [NEWLINE] The difference is that the Rule of Law, the very concept of Laws and Justice (rather than whoever is most powerful dictating<mask> things will go) is designed around the idea of the threat of violence being applied in just and equitable fashion. <mask> such it is much more than simply a standing threat of violence,<mask> it is a standing threat of violence *exclusively against those who do violence.* [NEWLINE] [NEWLINE] Me pointing a gun at some random person on the street is threat of violence.  Me pointing a gun at someone who is threatening/initiating harm against someone else is (a step towards) justice.  That is the goal the Rule of Law, and that's<mask> makes it more than a simple standing threat.</s>
Label encoding: <s>I'm going to disagree with this as best I can. [NEWLINE] [NEWLINE] Rule of Law is not the standing threat of violence so much as the standing promise of repercussions.  Yes, violence can be, and often is, a part of that, but the threat of violence is a fundamental fact of life.  There is still threat of violence even without laws; the threat that if you perturb someone bigger/stronger/better armed/more martially capable than you, they're going to do you harm.  There's the threat of the (financial) violence of theft.  That threat never goes away, not really. [NEWLINE] [NEWLINE] The difference is that the Rule of Law, the very concept of Laws and Justice (rather than whoever is most powerful dictating how things will go) is designed around the idea of the threat of violence being applied in just and equitable fashion.  As such it is much more than simply a standing threat of violence, because it is a standing threat of violence *exclusively against those who do violence.* [NEWLINE] [NEWLINE] Me pointing a gun at some random person on the street is threat of violence.  Me pointing a gun at someone who is threatening/initiating harm against someone else is (a step towards) justice.  That is the goal the Rule of Law, and that's what makes it more than a simple standing threat.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm saying that I knew there were risks<mask> I took the job. [ENDQ] [NEWLINE] Right, and even<mask> you took every precaution, something might happen.  Well that risk is just inherent to being a woman - they can't make a cost benefit analysis on that<mask>. [NEWLINE] [NEWLINE] [STARTQ] I'm saying, "yes, that is unfair,<mask> that's<mask> you should wear your seatbelt [ENDQ] [NEWLINE] And the problem I have with this is that: [NEWLINE] [NEWLINE] Everyone should wear their seatbelt. <mask> you really want to have this talk with kids, tell ALL kids not to be drunk in public -<mask><mask> is clear, any number of ills can befall you<mask> you are. <mask> the problem, I'd say, is that it doesn't really address the vast number of rapes, which are acquaintance rapes.  I mean, sure, you don't go to bars anymore,<mask> do you drink with your friends?  Have an 'eccentric' family member?  Go on dates? [NEWLINE] [NEWLINE] Any of those situations,<mask> *I* personally don't have "I could be raped" on my mind could a be a situation that ends up with a woman being raped.  The inclination to tell a rape victim to be more careful next time is in almost every situation unnecessary, and in many situations, not even particularly applicable. [NEWLINE] </s>
Label encoding: <s> [STARTQ] I'm saying that I knew there were risks when I took the job. [ENDQ] [NEWLINE] Right, and even if you took every precaution, something might happen.  Well that risk is just inherent to being a woman - they can't make a cost benefit analysis on that though. [NEWLINE] [NEWLINE] [STARTQ] I'm saying, "yes, that is unfair, but that's why you should wear your seatbelt [ENDQ] [NEWLINE] And the problem I have with this is that: [NEWLINE] [NEWLINE] Everyone should wear their seatbelt.  If you really want to have this talk with kids, tell ALL kids not to be drunk in public - because as is clear, any number of ills can befall you if you are.  But the problem, I'd say, is that it doesn't really address the vast number of rapes, which are acquaintance rapes.  I mean, sure, you don't go to bars anymore, but do you drink with your friends?  Have an 'eccentric' family member?  Go on dates? [NEWLINE] [NEWLINE] Any of those situations, where *I* personally don't have "I could be raped" on my mind could a be a situation that ends up with a woman being raped.  The inclination to tell a rape victim to be more careful next time is in almost every situation unnecessary, and in many situations, not even particularly applicable. [NEWLINE] </s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] **Reason 1: Laziness** [ENDQ] [NEWLINE] [STARTQ] A lot of technology today encourages laziness.<mask>, this takes it to a whole new level. [ENDQ] [NEWLINE] Damn kids and their gasoline powered cars. I used to walk to school, up hill, in the snow, both ways. [NEWLINE] [NEWLINE] [STARTQ] **Reason 2: Reliance on Technology** [ENDQ] [NEWLINE] Name one thing that's a "technology" that we invented and still use today that wouldn't be considered something we rely on. [NEWLINE] [NEWLINE] EDIT: This one came out wrong. I don't mean like a slap chop or a chia pet. Any significantly advanced "technology" becomes indispensable to us after it's conception. You cant just put the genie back iin the bottle. [NEWLINE] [NEWLINE] [STARTQ] **Reason 3: Invasion of Privacy** [ENDQ] [NEWLINE] Privacy in public spaces and<mask> interacting with public entities like corporations is already non private. Your information is and has been collected and sold<mask> they were able to do<mask>. That old Sears catalog that they deliver to your home 100 years ago? They sold, and used your information and purchase history to better sell you items.<mask> you dont like a companies practices don't use them. I know this can be hard in some cases<mask> your discussing a product that has to he purchased before use. It's not a malicious line of code on a popular website or something.</s>
Label encoding: <s> [STARTQ] **Reason 1: Laziness** [ENDQ] [NEWLINE] [STARTQ] A lot of technology today encourages laziness. However, this takes it to a whole new level. [ENDQ] [NEWLINE] Damn kids and their gasoline powered cars. I used to walk to school, up hill, in the snow, both ways. [NEWLINE] [NEWLINE] [STARTQ] **Reason 2: Reliance on Technology** [ENDQ] [NEWLINE] Name one thing that's a "technology" that we invented and still use today that wouldn't be considered something we rely on. [NEWLINE] [NEWLINE] EDIT: This one came out wrong. I don't mean like a slap chop or a chia pet. Any significantly advanced "technology" becomes indispensable to us after it's conception. You cant just put the genie back iin the bottle. [NEWLINE] [NEWLINE] [STARTQ] **Reason 3: Invasion of Privacy** [ENDQ] [NEWLINE] Privacy in public spaces and when interacting with public entities like corporations is already non private. Your information is and has been collected and sold since they were able to do so. That old Sears catalog that they deliver to your home 100 years ago? They sold, and used your information and purchase history to better sell you items. If you dont like a companies practices don't use them. I know this can be hard in some cases but your discussing a product that has to he purchased before use. It's not a malicious line of code on a popular website or something.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Alright,<mask> your view is a sort of eye for an eye, retributive justice viewpoint. The common argument against retributive justice is that it's backwards looking. The punishment is guided by a past crime and is designed to atone for damage already committed. The other way to look at punishments is for them to be forward looking (more utilitarian).<mask> such, punishments are guided by the social good they might achieve in the future. [NEWLINE] [NEWLINE] I don't think torturing criminals achieves much social good. It doesn't do anything to really deter serious crime, it doesn't rehabilitate the criminals, and it doesn't give us more security. It does send out the dangerous message that we<mask> a society are fine with torture,<mask><mask><mask> the people are 'bad' enough. There would be no more moral high ground to take against countries that torture their enemies. [NEWLINE] [NEWLINE] The victims of the crime may feel a slight jolt of happiness upon seeing the punishment<mask> it would be incredibly short lived. I'm sure many would feel disgusted that they felt happiness in the suffering of others. Some, like me, would feel no happiness at all. [NEWLINE] [NEWLINE] From the more utilitarian side of the justice debate, I don't quite see the benefits of it.<mask> you aren't ethically against torture for tortures sake, maybe you are against unnecessary torture? </s>
Label encoding: <s>Alright, so your view is a sort of eye for an eye, retributive justice viewpoint. The common argument against retributive justice is that it's backwards looking. The punishment is guided by a past crime and is designed to atone for damage already committed. The other way to look at punishments is for them to be forward looking (more utilitarian). As such, punishments are guided by the social good they might achieve in the future. [NEWLINE] [NEWLINE] I don't think torturing criminals achieves much social good. It doesn't do anything to really deter serious crime, it doesn't rehabilitate the criminals, and it doesn't give us more security. It does send out the dangerous message that we as a society are fine with torture, so long as the people are 'bad' enough. There would be no more moral high ground to take against countries that torture their enemies. [NEWLINE] [NEWLINE] The victims of the crime may feel a slight jolt of happiness upon seeing the punishment but it would be incredibly short lived. I'm sure many would feel disgusted that they felt happiness in the suffering of others. Some, like me, would feel no happiness at all. [NEWLINE] [NEWLINE] From the more utilitarian side of the justice debate, I don't quite see the benefits of it. If you aren't ethically against torture for tortures sake, maybe you are against unnecessary torture? </s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>'s your point? Polygamy exists - fact. Jealousy in Polygamy exists - "fact" [ENDQ] [NEWLINE] My point is that polygamy "works"<mask> it is not about free will, it's close to sexual slavery. The "free version" of it would be multiamory, in which there's no restriction for anyone involved (or at least no arbitrary difference in limitations). [NEWLINE] [NEWLINE] [STARTQ] <mask>, Some stronger males sometimes engage in jealous'mate guarding.' [ENDQ] "More strikingly, pairs or trios of top-ranking males sometimes engaged in cooperative aggression to prevent estrous females from mating with other males,<mask> tolerated each other's mating activities." [NEWLINE] [NEWLINE] That's the same with humans. Some males try to shag<mask> much<mask> possible<mask> can't accept that women do the same. It's know<mask> slut-shaming, or<mask> "I guy that has a lot of sex is a hero, a woman that does is a whore". [NEWLINE] [NEWLINE] Mate guarding is *cheating the system* in an free-love society, in the same way being unfaithful is *cheating the system* in a monogamous relationship. It's about egoism and power-abuse, be it by physical power, be it by economic power, and<mask><mask><mask> dispropriate power -balance it's accepted. </s>
Label encoding: <s> [STARTQ] What's your point? Polygamy exists - fact. Jealousy in Polygamy exists - "fact" [ENDQ] [NEWLINE] My point is that polygamy "works" because it is not about free will, it's close to sexual slavery. The "free version" of it would be multiamory, in which there's no restriction for anyone involved (or at least no arbitrary difference in limitations). [NEWLINE] [NEWLINE] [STARTQ] Yet, Some stronger males sometimes engage in jealous'mate guarding.' [ENDQ] "More strikingly, pairs or trios of top-ranking males sometimes engaged in cooperative aggression to prevent estrous females from mating with other males, but tolerated each other's mating activities." [NEWLINE] [NEWLINE] That's the same with humans. Some males try to shag as much as possible but can't accept that women do the same. It's know as slut-shaming, or as "I guy that has a lot of sex is a hero, a woman that does is a whore". [NEWLINE] [NEWLINE] Mate guarding is *cheating the system* in an free-love society, in the same way being unfaithful is *cheating the system* in a monogamous relationship. It's about egoism and power-abuse, be it by physical power, be it by economic power, and because of this dispropriate power -balance it's accepted. </s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree that the writers unknowingly writing a character to have autism without realizing or acknowledging it reflects a lack of understanding, and that's a shame. :/ [NEWLINE] [NEWLINE] <mask>, I don't even like BBT,<mask> I don't think the jokes are done in a way that dehumanizes or belittles his character (or those with autism or similar diagnoses, whom he represents in a way). The writing is done in a way that acknowledges his differences,<mask> develops him a personality (that's surprisingly deep for a sitcom) and character that's neither negatively portrayed nor stereotypical. [NEWLINE] [NEWLINE] I'd<mask><mask> the jokes are not at his expense. I find him similar to Abed, from Community. It's suggested heavily in the show that Abed has Aspergers Syndrome, and the difficulties of his behavior is even addressed with his interactions with friends and family.<mask> his traits that are considered unusual are portrayed<mask> awesome,<mask> unique,<mask> the creative reflections of a creative human. I see the same thing<mask> I see Sheldon play an archery game with the Wii<mask> pretend to grab arrows from an imaginary quiver. [NEWLINE] [NEWLINE] <mask> anything,<mask>, the constant laugh track may be demeaning. Laughter for every single line Sheldon says would devalue<mask> he has to say, and the laugh track does that a *lot* for him.</s>
Label encoding: <s>I agree that the writers unknowingly writing a character to have autism without realizing or acknowledging it reflects a lack of understanding, and that's a shame. :/ [NEWLINE] [NEWLINE] However, I don't even like BBT, yet I don't think the jokes are done in a way that dehumanizes or belittles his character (or those with autism or similar diagnoses, whom he represents in a way). The writing is done in a way that acknowledges his differences, but develops him a personality (that's surprisingly deep for a sitcom) and character that's neither negatively portrayed nor stereotypical. [NEWLINE] [NEWLINE] I'd argue that the jokes are not at his expense. I find him similar to Abed, from Community. It's suggested heavily in the show that Abed has Aspergers Syndrome, and the difficulties of his behavior is even addressed with his interactions with friends and family. But his traits that are considered unusual are portrayed as awesome, as unique, as the creative reflections of a creative human. I see the same thing when I see Sheldon play an archery game with the Wii but pretend to grab arrows from an imaginary quiver. [NEWLINE] [NEWLINE] If anything, though, the constant laugh track may be demeaning. Laughter for every single line Sheldon says would devalue what he has to say, and the laugh track does that a *lot* for him.</s>
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0673, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can think of no reason any printable character should be excluded from a security standpoint. [NEWLINE] [NEWLINE] I can say that it does prevent one problem. People copy pasting their password. Most sites ask you to confirm your password to ensure you don't enter it incorrectly and waste their time later trying to fix your mistake. Similarly<mask> you copy your password from somewhere<mask> accidentally copy a space before or after, you'll probably spend time trying to figure out<mask> it doesn't work later<mask> you type it out. [NEWLINE] [NEWLINE] Either that it's a lazy holdover from the days<mask> spaces couldn't be in filenames. [NEWLINE] [NEWLINE] I don't disagree that it's not a solid reason<mask><mask> they want to change a policy they have to rework the screening code.<mask> it ain't broken they won't fix it. [NEWLINE] [NEWLINE] <mask> to character length from a data storage standpoint longer passwords isn't really a problem. My guess is that they limit length to make the process of screening potential password faster<mask> anything. Thumbing<mask> one 18+ character password looking for errors is not<mask> bad.<mask> a server is probably doing that thousands of times each minute and that starts to strain it. [NEWLINE] [NEWLINE] You'd be surprised<mask> small repetitive task can really muddy down a server. Ensuring a balance between security and server performance is probably<mask> constraints come from. [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s>I can think of no reason any printable character should be excluded from a security standpoint. [NEWLINE] [NEWLINE] I can say that it does prevent one problem. People copy pasting their password. Most sites ask you to confirm your password to ensure you don't enter it incorrectly and waste their time later trying to fix your mistake. Similarly if you copy your password from somewhere but accidentally copy a space before or after, you'll probably spend time trying to figure out why it doesn't work later when you type it out. [NEWLINE] [NEWLINE] Either that it's a lazy holdover from the days when spaces couldn't be in filenames. [NEWLINE] [NEWLINE] I don't disagree that it's not a solid reason but if they want to change a policy they have to rework the screening code. If it ain't broken they won't fix it. [NEWLINE] [NEWLINE] As to character length from a data storage standpoint longer passwords isn't really a problem. My guess is that they limit length to make the process of screening potential password faster if anything. Thumbing though one 18+ character password looking for errors is not so bad. But a server is probably doing that thousands of times each minute and that starts to strain it. [NEWLINE] [NEWLINE] You'd be surprised how small repetitive task can really muddy down a server. Ensuring a balance between security and server performance is probably where constraints come from. [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's a legitimate,<mask> not a popular, form of argument - who said debate had to be logical? Certainly appealing to logic<mask> discussing mortality is limited -<mask> conclusion, and<mask> comfort, do *you* suppose logic can offer here? [NEWLINE] [NEWLINE] [STARTQ] Thou shall not discredit the person by remarks on character. [ENDQ] [NEWLINE] I made no such slight. Suicide and the feelings and thoughts that surround it are very personal, and ought to be respected. *I* did not use them in this debate, and I absolutely do not consider it a mark on a person's character. Perhaps you should examine your own views on the matter, that lead you to<mask> misread mine. [NEWLINE] [NEWLINE] To expand upon my original comment: [NEWLINE] [NEWLINE] Being suicidal would be otherwise inconsequential... except in this one instance<mask> it's vital: u/DSKNg was offering a personal opinion on death and offering sollace to another - questioning the validity of opinion of someone who is a self-confessed suicidal, is appropriate and expected. [NEWLINE] [NEWLINE] Logic aside, I put it to you - who would you want talking your friend down off the ledge? The fellow suicide risk?! The logician!? [NEWLINE] [NEWLINE] I might not be brave enough to offer my own thoughts on death,<mask> suicide ought not be trivialised, nor celebrated and coddled.</s>
Label encoding: <s>It's a legitimate, if not a popular, form of argument - who said debate had to be logical? Certainly appealing to logic when discussing mortality is limited - what conclusion, and indeed comfort, do *you* suppose logic can offer here? [NEWLINE] [NEWLINE] [STARTQ] Thou shall not discredit the person by remarks on character. [ENDQ] [NEWLINE] I made no such slight. Suicide and the feelings and thoughts that surround it are very personal, and ought to be respected. *I* did not use them in this debate, and I absolutely do not consider it a mark on a person's character. Perhaps you should examine your own views on the matter, that lead you to so misread mine. [NEWLINE] [NEWLINE] To expand upon my original comment: [NEWLINE] [NEWLINE] Being suicidal would be otherwise inconsequential... except in this one instance where it's vital: u/DSKNg was offering a personal opinion on death and offering sollace to another - questioning the validity of opinion of someone who is a self-confessed suicidal, is appropriate and expected. [NEWLINE] [NEWLINE] Logic aside, I put it to you - who would you want talking your friend down off the ledge? The fellow suicide risk?! The logician!? [NEWLINE] [NEWLINE] I might not be brave enough to offer my own thoughts on death, but suicide ought not be trivialised, nor celebrated and coddled.</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**EDIT: View changed.** [NEWLINE] [NEWLINE] I saw an advertisement for a horror movie today, claiming itself to be "the best American horror film<mask> The Blair Witch Project." [NEWLINE] [NEWLINE] This left me wondering:<mask> is Blair Witch<mask> widely exalted in the horror genre?<mask> I watched it a few years ago, I was underwhelmed, bemused, and did not experience a single fright. [NEWLINE] [NEWLINE] Here's my breakdown of the film: [NEWLINE] [NEWLINE] -Daytime: hikers wander around and find creepy little things made by the witch. Trite--we've all seen something like that before. [NEWLINE] [NEWLINE] -Nighttime: Hikers hear a noise, take the camcorder outside to investigate, see *something* (we can hardly see<mask> scared them<mask> it's nighttime and the camcorder quality isn't great at long distance), camera gets pointed at the ground, person filming takes off running and screaming. [NEWLINE] [NEWLINE] This cycle repeats several times until the end of the film. The end was somewhat disturbing and left me wondering,<mask> it still wasn't scary. [NEWLINE] [NEWLINE] TL;DR: Blair Witch is a horror movie in which we see people get scared,<mask> never really get to see<mask> exactly is scaring them.<mask> is it a horror movie at all, and<mask> has it received<mask> much praise within the genre?</s>
Label encoding: <s>**EDIT: View changed.** [NEWLINE] [NEWLINE] I saw an advertisement for a horror movie today, claiming itself to be "the best American horror film since The Blair Witch Project." [NEWLINE] [NEWLINE] This left me wondering: why is Blair Witch so widely exalted in the horror genre? When I watched it a few years ago, I was underwhelmed, bemused, and did not experience a single fright. [NEWLINE] [NEWLINE] Here's my breakdown of the film: [NEWLINE] [NEWLINE] -Daytime: hikers wander around and find creepy little things made by the witch. Trite--we've all seen something like that before. [NEWLINE] [NEWLINE] -Nighttime: Hikers hear a noise, take the camcorder outside to investigate, see *something* (we can hardly see what scared them because it's nighttime and the camcorder quality isn't great at long distance), camera gets pointed at the ground, person filming takes off running and screaming. [NEWLINE] [NEWLINE] This cycle repeats several times until the end of the film. The end was somewhat disturbing and left me wondering, but it still wasn't scary. [NEWLINE] [NEWLINE] TL;DR: Blair Witch is a horror movie in which we see people get scared, but never really get to see what exactly is scaring them. How is it a horror movie at all, and why has it received so much praise within the genre?</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] On the other side, many beleif systems have very specific definitions of<mask> constitutes marraige and those guidelines cant be changed by the legal system. [ENDQ] [NEWLINE] Who gives a fuck?<mask> is this a legitimate greivence at all?<mask> those guidelines *can't* be changed by the legal system then<mask>'s the problem? You don't get to impose your beliefs on everyone else, that doesn't make any sense. "Oh, you can't call them contracts<mask> in my belief system a contract is only defined<mask> being a firey ball of plasma held together by gravity, and<mask> you want to call something else a contract it looks like we'll just have to completely do away with the whole concept." [NEWLINE] [NEWLINE] I don't care<mask> anyone's individual definition of a word is, they don't own the word. [NEWLINE] [NEWLINE] [STARTQ] The solution to this problem is remove marraige from the legal system. [ENDQ] [NEWLINE] There's a much easier solution: make gay marriage legal. [NEWLINE] [NEWLINE] [STARTQ] By law neither straight or gay couples can marry. Marraige would still exist<mask> a religious tradition, and gay people still get equal treatment under the law. [ENDQ] [NEWLINE] I'm married and not religious and I would like to remain married, thanks. Marriage is not an exclusively religious tradition,<mask> evidenced by it existing<mask> a legal one.</s>
Label encoding: <s> [STARTQ] On the other side, many beleif systems have very specific definitions of what constitutes marraige and those guidelines cant be changed by the legal system. [ENDQ] [NEWLINE] Who gives a fuck? How is this a legitimate greivence at all? If those guidelines *can't* be changed by the legal system then what's the problem? You don't get to impose your beliefs on everyone else, that doesn't make any sense. "Oh, you can't call them contracts because in my belief system a contract is only defined as being a firey ball of plasma held together by gravity, and since you want to call something else a contract it looks like we'll just have to completely do away with the whole concept." [NEWLINE] [NEWLINE] I don't care what anyone's individual definition of a word is, they don't own the word. [NEWLINE] [NEWLINE] [STARTQ] The solution to this problem is remove marraige from the legal system. [ENDQ] [NEWLINE] There's a much easier solution: make gay marriage legal. [NEWLINE] [NEWLINE] [STARTQ] By law neither straight or gay couples can marry. Marraige would still exist as a religious tradition, and gay people still get equal treatment under the law. [ENDQ] [NEWLINE] I'm married and not religious and I would like to remain married, thanks. Marriage is not an exclusively religious tradition, as evidenced by it existing as a legal one.</s>
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> do your actions in some way create happier animals?<mask> a large enough group of people eschewed meat, factory farmed animals wouldn't simply be happier. They would simply cease to exist<mask> producers would lower livestock numbers to adjust for reduced demand. [NEWLINE] [NEWLINE] In that sense you haven't created any more happiness. The factory farmed animals will remain<mask> in a smaller capacity. The rest will simply never be born. [NEWLINE] [NEWLINE] But say that same group decided to support free range livestock instead. In that case, both the number of factory farmed animals decrease (<mask> producers adjust supply for reduced demand)<mask> the number of free range animals increase -<mask> increasing the sum of happiness. [NEWLINE] [NEWLINE] <mask> essentially, say we start with 100 unhappy factory farm animals. In scenario 1,<mask> many people turn vegetarian or vegan, you now would get 70 unhappy factory farm animals<mask> no extra happy animals. You've shrunk unhappiness<mask> haven't at all increased happiness. [NEWLINE] [NEWLINE] In scenario 2, you get the same 70 unhappy factory farm animals<mask> now you<mask> produce 30 happy free range animals. You've decreased both the number of unhappy animals in existence<mask><mask> increasing the number of happy animals. [NEWLINE] [NEWLINE] And finally,<mask> is a more realistic campaign that'll actually impact people: telling people to give up meat or telling people to find better sources of meat?</s>
Label encoding: <s>But do your actions in some way create happier animals? If a large enough group of people eschewed meat, factory farmed animals wouldn't simply be happier. They would simply cease to exist as producers would lower livestock numbers to adjust for reduced demand. [NEWLINE] [NEWLINE] In that sense you haven't created any more happiness. The factory farmed animals will remain but in a smaller capacity. The rest will simply never be born. [NEWLINE] [NEWLINE] But say that same group decided to support free range livestock instead. In that case, both the number of factory farmed animals decrease ( as producers adjust supply for reduced demand) while the number of free range animals increase - thus increasing the sum of happiness. [NEWLINE] [NEWLINE] So essentially, say we start with 100 unhappy factory farm animals. In scenario 1, where many people turn vegetarian or vegan, you now would get 70 unhappy factory farm animals but no extra happy animals. You've shrunk unhappiness but haven't at all increased happiness. [NEWLINE] [NEWLINE] In scenario 2, you get the same 70 unhappy factory farm animals but now you also produce 30 happy free range animals. You've decreased both the number of unhappy animals in existence while also increasing the number of happy animals. [NEWLINE] [NEWLINE] And finally, what is a more realistic campaign that'll actually impact people: telling people to give up meat or telling people to find better sources of meat?</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0636, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0574, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] For international issues, we're going to be in disagreement. The general public is much more isolationist than our current policy, and<mask><mask> the public policy should reflect that. [ENDQ] [NEWLINE] That's not really a good thing, especially<mask> trade and relations are vital to most countries survival. [NEWLINE] [NEWLINE] [STARTQ] Referenda would potentially work,<mask> it would have to be much more common than it is now for people to have enough input into the system. [ENDQ] [NEWLINE] <mask> then the solution would be to have more referendums, not to completely overhaul the entire system. [NEWLINE] [NEWLINE] [STARTQ] <mask> I do think that your average person should have some say in the US's operations in other countries, and right now we have very little. [ENDQ] [NEWLINE] And this brings up the danger of people having no clue about these issues. [NEWLINE] [NEWLINE] [STARTQ] The reason a bunch of terrorists want to kill everyone is<mask> the US has been drone bombing them, and finding various other ways of killing their families for years. [ENDQ] [NEWLINE] [NEWLINE] That's a vastly oversimplified reason. [NEWLINE] [NEWLINE] [STARTQ] I didn't want that,<mask> they want to kill me, too. This is unfair. [ENDQ] [NEWLINE] It turns out that even<mask> you don't want it, many others might. Before the start of the Iraq war, between 47-60% of people supported an invasion.</s>
Label encoding: <s> [STARTQ] For international issues, we're going to be in disagreement. The general public is much more isolationist than our current policy, and I think the public policy should reflect that. [ENDQ] [NEWLINE] That's not really a good thing, especially since trade and relations are vital to most countries survival. [NEWLINE] [NEWLINE] [STARTQ] Referenda would potentially work, but it would have to be much more common than it is now for people to have enough input into the system. [ENDQ] [NEWLINE] So then the solution would be to have more referendums, not to completely overhaul the entire system. [NEWLINE] [NEWLINE] [STARTQ] But I do think that your average person should have some say in the US's operations in other countries, and right now we have very little. [ENDQ] [NEWLINE] And this brings up the danger of people having no clue about these issues. [NEWLINE] [NEWLINE] [STARTQ] The reason a bunch of terrorists want to kill everyone is because the US has been drone bombing them, and finding various other ways of killing their families for years. [ENDQ] [NEWLINE] [NEWLINE] That's a vastly oversimplified reason. [NEWLINE] [NEWLINE] [STARTQ] I didn't want that, yet they want to kill me, too. This is unfair. [ENDQ] [NEWLINE] It turns out that even if you don't want it, many others might. Before the start of the Iraq war, between 47-60% of people supported an invasion.</s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The problem isn't that society views women<mask> less than a man,<mask> rather that society has views<mask> to<mask> a man should do and<mask> a woman should do, and<mask> you don't fall into the category society has set for you--man or woman--you're shamed for it. [ENDQ] [NEWLINE] <mask><mask> you're not looking at this quite the right way. Things being "gendered" isn't nearly half the issue. The problem resides in the fact that "gendered" things are valued quite differently in social context. Typical female gender roles, for example, have little to no value and are not an opportunity to build capital (both social and material) leaving women with little weight in society. [NEWLINE] [NEWLINE] I<mask> think you're wrong with the following assertion. Feminism aims, amongst other things, to dismantle gender roles. Now, gender roles do hurt men too<mask> there's no question that it's gonna help them in the end.<mask>, I do agree that the primary focus is women plights,<mask> I don't know<mask> people are expecting from *feminism*. Yes, it's gonna indirectly help men, *no*, it's not their main objective. On another note, I do support the LGBT movement,<mask> being heterosexual. I don't feel they're taking anything away from me. [NEWLINE] </s>
Label encoding: <s> [STARTQ] The problem isn't that society views women as less than a man, but rather that society has views as to what a man should do and what a woman should do, and if you don't fall into the category society has set for you--man or woman--you're shamed for it. [ENDQ] [NEWLINE] I think you're not looking at this quite the right way. Things being "gendered" isn't nearly half the issue. The problem resides in the fact that "gendered" things are valued quite differently in social context. Typical female gender roles, for example, have little to no value and are not an opportunity to build capital (both social and material) leaving women with little weight in society. [NEWLINE] [NEWLINE] I also think you're wrong with the following assertion. Feminism aims, amongst other things, to dismantle gender roles. Now, gender roles do hurt men too so there's no question that it's gonna help them in the end. However, I do agree that the primary focus is women plights, but I don't know what people are expecting from *feminism*. Yes, it's gonna indirectly help men, *no*, it's not their main objective. On another note, I do support the LGBT movement, while being heterosexual. I don't feel they're taking anything away from me. [NEWLINE] </s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is not accurate. [NEWLINE] [NEWLINE] [STARTQ] Before contemporary feminist movement was less than 10 years old, feminist thinkers began to talk about the way in which patriarchy was harmful to men. Without changing our fierce critique of male domination feminist politics expanded to include the recognition that patriarchy stripped men of certain rights, lmposing on them a sexist masculine identity. [ENDQ] [NEWLINE]... [NEWLINE] [NEWLINE] [STARTQ] Feminists who called for a recognition of men<mask> comrades in struggle never received mass media attention. Our theoretical work critiquing the demonization of men<mask> the enemy did not change the perspectives of women who were anti-male. And it was reaction to negative representations of manhood that led to the development of a men's movement that was anti-female. Writing about the "men's liberation movement" I called attention to the opportunism undergirding this movement: [ENDQ] [NEWLINE] [STARTQ] "These men identified themselves<mask> victims of sexism, working to liberate men. They identified rigid sex roles<mask> the primary source of their victimization, and,<mask> they wanted to change the notion of masculinity, they were not particularly concerned with their sexist exploitation and oppression of women." [ENDQ] [NEWLINE] [STARTQ] In many ways the men's movement mirrored the most negative aspects of the women's movement. [ENDQ] [NEWLINE] -- Bell Hooks, *Feminism is for Everybody*, pp 68-69</s>
Label encoding: <s>This is not accurate. [NEWLINE] [NEWLINE] [STARTQ] Before contemporary feminist movement was less than 10 years old, feminist thinkers began to talk about the way in which patriarchy was harmful to men. Without changing our fierce critique of male domination feminist politics expanded to include the recognition that patriarchy stripped men of certain rights, lmposing on them a sexist masculine identity. [ENDQ] [NEWLINE]... [NEWLINE] [NEWLINE] [STARTQ] Feminists who called for a recognition of men as comrades in struggle never received mass media attention. Our theoretical work critiquing the demonization of men as the enemy did not change the perspectives of women who were anti-male. And it was reaction to negative representations of manhood that led to the development of a men's movement that was anti-female. Writing about the "men's liberation movement" I called attention to the opportunism undergirding this movement: [ENDQ] [NEWLINE] [STARTQ] "These men identified themselves as victims of sexism, working to liberate men. They identified rigid sex roles as the primary source of their victimization, and, though they wanted to change the notion of masculinity, they were not particularly concerned with their sexist exploitation and oppression of women." [ENDQ] [NEWLINE] [STARTQ] In many ways the men's movement mirrored the most negative aspects of the women's movement. [ENDQ] [NEWLINE] -- Bell Hooks, *Feminism is for Everybody*, pp 68-69</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0677, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0596, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Two reasons<mask> grammar nazis are *saving* language and promoting globalization: [NEWLINE] [NEWLINE] 1. Consistency. English has been<mask> it is for a very long time.<mask>, there are a lot of resources that are still correct- it doesn't matter<mask> it's from 1930 or 2005, it's probably at least 90% correct. Changing the language causes problems for people in remote places who may only have old textbooks and learning supplies-<mask> they use a book from 1980, right now that's not a problem,<mask><mask> we gradually change the language,<mask> long until it's outdated and useless? [NEWLINE] [NEWLINE] 2. The goal of language is communication. Change is not a consistent, constant, universal thing- promoting or allowing change on a large scale will, over time, result in a variety of mutually unintelligible new pseudo-English languages that will limit people's communication ability to their region. Keeping language consistent, via people like grammar nazis, ensures that the English spoken in place A is at least 90% the same<mask> the English spoken in place B. [NEWLINE] [NEWLINE] <mask>, a quick example of<mask> happens<mask> you *don't* have a limiting factor keeping English spelling/pronunciation consistent and mutually intelligible: [NEWLINE] [NEWLINE] [Listen to it without looking at the subtitles]( [URL] ) [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Two reasons why grammar nazis are *saving* language and promoting globalization: [NEWLINE] [NEWLINE] 1. Consistency. English has been as it is for a very long time. Therefore, there are a lot of resources that are still correct- it doesn't matter if it's from 1930 or 2005, it's probably at least 90% correct. Changing the language causes problems for people in remote places who may only have old textbooks and learning supplies- if they use a book from 1980, right now that's not a problem, but if we gradually change the language, how long until it's outdated and useless? [NEWLINE] [NEWLINE] 2. The goal of language is communication. Change is not a consistent, constant, universal thing- promoting or allowing change on a large scale will, over time, result in a variety of mutually unintelligible new pseudo-English languages that will limit people's communication ability to their region. Keeping language consistent, via people like grammar nazis, ensures that the English spoken in place A is at least 90% the same as the English spoken in place B. [NEWLINE] [NEWLINE] Lastly, a quick example of what happens when you *don't* have a limiting factor keeping English spelling/pronunciation consistent and mutually intelligible: [NEWLINE] [NEWLINE] [Listen to it without looking at the subtitles]( [URL] ) [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I had a larger post written,<mask> allow me to point this out. [NEWLINE] [NEWLINE] This is<mask> the OP wrote, italics mine: [NEWLINE] [NEWLINE] [STARTQ] <mask> one believes that a fetus is a person, then they would be pro-life. This view would imply theat abortion would be the *moral equivalent* to murder (except in cases<mask> the mother's life is at risk). All other issues, such<mask> a woman's right to her body would be secondary,<mask> saving a life - forgoing an abortion- *comes first morally.* [ENDQ] [NEWLINE] <mask> you hold that rights are *made up*, then<mask> do they have any bearing on the issue, morally or otherwise? They *do not exist.* [NEWLINE] [NEWLINE] My point is this: the OP is talking about morality, or ethics, and the top comment (at this moment) I was responding to was making a point about<mask> a system of law, based on rights, had certain implications for abortion. My point is this:<mask> rights are not real, it's not immediately obvious to myself<mask> they can be directly relevant to morality. [NEWLINE] [NEWLINE] (edit: and, frankly,<mask> rights are something we make up, then enforcing your rights - which again, we have manufactured - might actually be counter-productive in pursuing<mask> is morally or ethically correct.)</s>
Label encoding: <s>I had a larger post written, but allow me to point this out. [NEWLINE] [NEWLINE] This is what the OP wrote, italics mine: [NEWLINE] [NEWLINE] [STARTQ] If one believes that a fetus is a person, then they would be pro-life. This view would imply theat abortion would be the *moral equivalent* to murder (except in cases where the mother's life is at risk). All other issues, such as a woman's right to her body would be secondary, as saving a life - forgoing an abortion- *comes first morally.* [ENDQ] [NEWLINE] If you hold that rights are *made up*, then why do they have any bearing on the issue, morally or otherwise? They *do not exist.* [NEWLINE] [NEWLINE] My point is this: the OP is talking about morality, or ethics, and the top comment (at this moment) I was responding to was making a point about how a system of law, based on rights, had certain implications for abortion. My point is this: if rights are not real, it's not immediately obvious to myself how they can be directly relevant to morality. [NEWLINE] [NEWLINE] (edit: and, frankly, if rights are something we make up, then enforcing your rights - which again, we have manufactured - might actually be counter-productive in pursuing what is morally or ethically correct.)</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Honestly your income is more important to your standard of living than your country of residence.<mask> you've got money and can afford to pay for your own health insurance, I bet your standard of living would increase. You'd still be getting good healthcare<mask> your taxes would be lower<mask> you wouldn't be paying for poorer people's healthcare. [NEWLINE] [NEWLINE] <mask> you can't afford your own healthcare, you'd be better off in the UK<mask> richer people will pay for it on your behalf. [NEWLINE] [NEWLINE] In daily life in America, you don't really deal with law enforcement and you'd probably never seen a gun (except for a cop's) outside of a shooting range or hunting area. Neither of those two things would really affect your day to day life at all. [NEWLINE] [NEWLINE] I don't think the GINI coefficient will affect your personal standard of living at all.<mask><mask><mask> poor other people in the country are, your personal income probably won't change that much.<mask> you were a highly skilled well paid worker in the UK, you'd be the same in the US.<mask> you get paid minimun wage in the UK, your wage could go up or down depending on the state you'd move to. [NEWLINE] [NEWLINE] Honestly, your life wouldn't change that much, especially<mask> your middle class or wealthy. [NEWLINE] </s>
Label encoding: <s>Honestly your income is more important to your standard of living than your country of residence. If you've got money and can afford to pay for your own health insurance, I bet your standard of living would increase. You'd still be getting good healthcare but your taxes would be lower because you wouldn't be paying for poorer people's healthcare. [NEWLINE] [NEWLINE] If you can't afford your own healthcare, you'd be better off in the UK where richer people will pay for it on your behalf. [NEWLINE] [NEWLINE] In daily life in America, you don't really deal with law enforcement and you'd probably never seen a gun (except for a cop's) outside of a shooting range or hunting area. Neither of those two things would really affect your day to day life at all. [NEWLINE] [NEWLINE] I don't think the GINI coefficient will affect your personal standard of living at all. Regardless of how poor other people in the country are, your personal income probably won't change that much. If you were a highly skilled well paid worker in the UK, you'd be the same in the US. If you get paid minimun wage in the UK, your wage could go up or down depending on the state you'd move to. [NEWLINE] [NEWLINE] Honestly, your life wouldn't change that much, especially if your middle class or wealthy. [NEWLINE] </s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Look man, here's<mask> it is. [NEWLINE] [NEWLINE] [From the BBC]( [URL] ): [NEWLINE] [NEWLINE] [STARTQ] **Israel attacked a UN-run school housing refugees in Gaza<mask> warnings that civilians were there, the UN has said.** [ENDQ] [NEWLINE] [STARTQ] UN spokesman Chris Gunness said "the world stands disgraced" by the attack, in which 15 died and dozens were hurt. [ENDQ] [NEWLINE] [STARTQ] Mr Gunness, from the UN Relief and Works Agency (Unrwa), told the BBC that **Israel had been told 17 times that the school in the Jabaliya refugee camp was housing the displaced.** [ENDQ] [NEWLINE] [STARTQ] **"The last time was hours before the fatal attack,"** he said. "Our initial assessment is that it was Israeli artillery that hit our school." [ENDQ] [NEWLINE] [STARTQ] The UN says this is the sixth time that one of its schools has been hit by shellfire<mask> this conflict began. And I've been told by UN officials on the ground that they believe Israeli forces were responsible on each occasion. [ENDQ] [NEWLINE] Bottom line, Israel does not give a flying fuck about who they kill.  They are deliberately targeting civilians.  Nearly every death of those 55 Israelis was IDF soldier.  53 of them, to be exact. [NEWLINE] [NEWLINE] <mask> you tell me, who is really running the propaganda show?</s>
Label encoding: <s>Look man, here's how it is. [NEWLINE] [NEWLINE] [From the BBC]( [URL] ): [NEWLINE] [NEWLINE] [STARTQ] **Israel attacked a UN-run school housing refugees in Gaza despite warnings that civilians were there, the UN has said.** [ENDQ] [NEWLINE] [STARTQ] UN spokesman Chris Gunness said "the world stands disgraced" by the attack, in which 15 died and dozens were hurt. [ENDQ] [NEWLINE] [STARTQ] Mr Gunness, from the UN Relief and Works Agency (Unrwa), told the BBC that **Israel had been told 17 times that the school in the Jabaliya refugee camp was housing the displaced.** [ENDQ] [NEWLINE] [STARTQ] **"The last time was hours before the fatal attack,"** he said. "Our initial assessment is that it was Israeli artillery that hit our school." [ENDQ] [NEWLINE] [STARTQ] The UN says this is the sixth time that one of its schools has been hit by shellfire since this conflict began. And I've been told by UN officials on the ground that they believe Israeli forces were responsible on each occasion. [ENDQ] [NEWLINE] Bottom line, Israel does not give a flying fuck about who they kill.  They are deliberately targeting civilians.  Nearly every death of those 55 Israelis was IDF soldier.  53 of them, to be exact. [NEWLINE] [NEWLINE] So you tell me, who is really running the propaganda show?</s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0842, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks. The humor was somewhat unintentional! ;) I wasn't implying that politicians are secretly space-lizards,<mask> I do believe that virtues that make you a better human being might make you worse at some specialized job. [NEWLINE] [NEWLINE] Open to new evidence and open to new ideologies are two different things. They are definitely connected,<mask> ideologies are mostly about ethical goals, secondarily about social theories, and only<mask><mask> about specific factual or causal claims. New evidence,<mask><mask><mask><mask>, primarily concerns factual or causal claims, has a weaker ability to affect our social theories, and has almost no grip on our ethical ideals.<mask> you don't want, in a presidential election, is to vote for someone in favor of an equal society, who a year later encountere some right-wing argument for the first time and becomes an elitist. (Or vice-versa,<mask> you're an elitist.) It doesn't matter<mask> much<mask> you vote for a candidate who promises to raise the sales tax, and then after his administration gathers more data he raises the income tax instead - presumably his goals are still the same,<mask> he changed his mind about the most technically efficient way to reach his goals. "Personal growth" and other sorts of conversion experiences are more about the former than the latter.</s>
Label encoding: <s>Thanks. The humor was somewhat unintentional! ;) I wasn't implying that politicians are secretly space-lizards, but I do believe that virtues that make you a better human being might make you worse at some specialized job. [NEWLINE] [NEWLINE] Open to new evidence and open to new ideologies are two different things. They are definitely connected, but ideologies are mostly about ethical goals, secondarily about social theories, and only lastly about specific factual or causal claims. New evidence, on the other hand, primarily concerns factual or causal claims, has a weaker ability to affect our social theories, and has almost no grip on our ethical ideals. What you don't want, in a presidential election, is to vote for someone in favor of an equal society, who a year later encountere some right-wing argument for the first time and becomes an elitist. (Or vice-versa, if you're an elitist.) It doesn't matter as much if you vote for a candidate who promises to raise the sales tax, and then after his administration gathers more data he raises the income tax instead - presumably his goals are still the same, but he changed his mind about the most technically efficient way to reach his goals. "Personal growth" and other sorts of conversion experiences are more about the former than the latter.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll give it a shot: Batman is our mortal stand-in in the DC Universe. [NEWLINE] [NEWLINE] One can make standalone points based on specific storylines all day, arguing that Wayne Industries' methods of operation are subject to highly diminishing returns,<mask><mask> you are looking for a force that Batman represents that can stand up to the untold mystical powers of the DCverse, it's the potential of the everyman. [NEWLINE] [NEWLINE] Batman is ingenuity, can-do -- hell, the living incarnation of mind over matter, or at least dramatic pragmatism. On the other end (for the purposes of this exposition, bear with me) is Superman, who is the definition of raw paranormal power, and in between the two we see the whole continuum of human to superhuman, the complete gamut of power within many can say "I am"<mask> few can add "<mask> I gave it my all". [NEWLINE] [NEWLINE] Batman is the meticulous, self-made polar opposite to Superman's cosmically overwhelming genetic gifts, and the quintessential anchor of the DC Universe to the idea that we can do whatever we put our minds to. We read fantasy not just to be amazed by the impossible,<mask> to be imbued with hope, and that can be found in seeing that we too are powerful. Aren't we all Batman?</s>
Label encoding: <s>I'll give it a shot: Batman is our mortal stand-in in the DC Universe. [NEWLINE] [NEWLINE] One can make standalone points based on specific storylines all day, arguing that Wayne Industries' methods of operation are subject to highly diminishing returns, but if you are looking for a force that Batman represents that can stand up to the untold mystical powers of the DCverse, it's the potential of the everyman. [NEWLINE] [NEWLINE] Batman is ingenuity, can-do -- hell, the living incarnation of mind over matter, or at least dramatic pragmatism. On the other end (for the purposes of this exposition, bear with me) is Superman, who is the definition of raw paranormal power, and in between the two we see the whole continuum of human to superhuman, the complete gamut of power within many can say "I am" but few can add " because I gave it my all". [NEWLINE] [NEWLINE] Batman is the meticulous, self-made polar opposite to Superman's cosmically overwhelming genetic gifts, and the quintessential anchor of the DC Universe to the idea that we can do whatever we put our minds to. We read fantasy not just to be amazed by the impossible, but to be imbued with hope, and that can be found in seeing that we too are powerful. Aren't we all Batman?</s>
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think this is at the root, really. [NEWLINE] [NEWLINE] "I have never" is a statement of fact.  It might be said at an inappropriate time that could make it racist,<mask> it isn't intrinsically racist, and does not require racism to be spoken. [NEWLINE] [NEWLINE] "I prefer" is the kind of passive, cultural racism that is a problem,<mask> isn't the same<mask> outright discrimination.  It's frowned on,<mask> not outright bigoted necessarily. [NEWLINE] [NEWLINE] "I would never" is out-right bigotry.  A person is precluding the possibility that an individual of X group could ever break out of the stereotype you have built and prove themselves "worthy" of whatever it is we're talking about. [NEWLINE] [NEWLINE] [NEWLINE] An exception is made for classes that are biologically tied to the thing in question: "I would never have sex with a man" is an acceptable statement for a man, and<mask> it can ruffle some sexist jimmies, is<mask> acceptable for any woman or other, more complicated gender identity.  It is true that this is, realistically, an exception and not a different thing altogether.  Plenty of men who "would never" have sex with a man feel that way<mask> of shame, not attraction.  Maybe not a lot,<mask> enough.</s>
Label encoding: <s>I think this is at the root, really. [NEWLINE] [NEWLINE] "I have never" is a statement of fact.  It might be said at an inappropriate time that could make it racist, but it isn't intrinsically racist, and does not require racism to be spoken. [NEWLINE] [NEWLINE] "I prefer" is the kind of passive, cultural racism that is a problem, but isn't the same as outright discrimination.  It's frowned on, but not outright bigoted necessarily. [NEWLINE] [NEWLINE] "I would never" is out-right bigotry.  A person is precluding the possibility that an individual of X group could ever break out of the stereotype you have built and prove themselves "worthy" of whatever it is we're talking about. [NEWLINE] [NEWLINE] [NEWLINE] An exception is made for classes that are biologically tied to the thing in question: "I would never have sex with a man" is an acceptable statement for a man, and though it can ruffle some sexist jimmies, is also acceptable for any woman or other, more complicated gender identity.  It is true that this is, realistically, an exception and not a different thing altogether.  Plenty of men who "would never" have sex with a man feel that way because of shame, not attraction.  Maybe not a lot, but enough.</s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> you're buying Gap, your still getting ripped off: you just have a much lower selection to choose from. That 1 polo for 97 cents *IS* a central part of their inventory. It gets you in the door. It gets you advertising Gap on your clothes. [NEWLINE] [NEWLINE] The only problem is, at your price point, they have an inventory of one. Or an entire outlet of alternative clothing lines labeled "Gap" on "sale"<mask><mask> these items are never sold in the "regular" stores in the first place. [NEWLINE] [NEWLINE] If I want clothes that actually fit well with a wide selection to choose from, I'll pay full price. 5 years out of a $40 pair of jeans is worth it. You're unlikely to find that durablity at the gap<mask>. [NEWLINE] [NEWLINE] <mask><mask><mask> :<mask> you only wan to spend $1, hit up a thrift shop or non-fashion brands. At the gap, you're selection is limited<mask> you only shop the sale racks. [NEWLINE] [NEWLINE] Tl;Dr: *RETAIL CLOTHING* is overvalued. It has nothing to do with being on a "clearance" rack or not. "Clearance" and "Sale" are marketing terms, not statements of value. </s>
Label encoding: <s>If you're buying Gap, your still getting ripped off: you just have a much lower selection to choose from. That 1 polo for 97 cents *IS* a central part of their inventory. It gets you in the door. It gets you advertising Gap on your clothes. [NEWLINE] [NEWLINE] The only problem is, at your price point, they have an inventory of one. Or an entire outlet of alternative clothing lines labeled "Gap" on "sale" even though these items are never sold in the "regular" stores in the first place. [NEWLINE] [NEWLINE] If I want clothes that actually fit well with a wide selection to choose from, I'll pay full price. 5 years out of a $40 pair of jeans is worth it. You're unlikely to find that durablity at the gap though. [NEWLINE] [NEWLINE] THEREFORE : If you only wan to spend $1, hit up a thrift shop or non-fashion brands. At the gap, you're selection is limited when you only shop the sale racks. [NEWLINE] [NEWLINE] Tl;Dr: *RETAIL CLOTHING* is overvalued. It has nothing to do with being on a "clearance" rack or not. "Clearance" and "Sale" are marketing terms, not statements of value. </s>
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It truly does depend on the position. I don't know<mask> you're hiring for.<mask> it's some sort of office position or anything that requires careful, consistent, "white-collar" professional skills then I would understand. [NEWLINE] [NEWLINE] <mask> I work in construction and I can tell you now that spelling, sentence structure, and eloquence on a resume mean nothing to me. I need guys who are quick on their feet, hard working, willing to learn, and professional on a very basic, practical level. We've hired a guy in his 50's without a high-school diploma and no prior experience in construction. Comes in 15 minutes early, leaves 15 minutes after we clock out, and busts his ass every day. He's<mask> quite smart, being very strategic<mask> he organizes and plans day to day tasks. His resume looked like a 4th grader wrote it. [NEWLINE] [NEWLINE] <mask> you're hiring for a position that isn't directly responsible for representing your company on a professional level through his/her writing abilities,<mask><mask> you're missing out on some great people. [NEWLINE] [NEWLINE] Now,<mask> I were hiring a real-estate agent and she/he made all sorts of spelling errors I would take them less seriously. That sort of thing is important in their line of work. </s>
Label encoding: <s>It truly does depend on the position. I don't know what you're hiring for. If it's some sort of office position or anything that requires careful, consistent, "white-collar" professional skills then I would understand. [NEWLINE] [NEWLINE] But I work in construction and I can tell you now that spelling, sentence structure, and eloquence on a resume mean nothing to me. I need guys who are quick on their feet, hard working, willing to learn, and professional on a very basic, practical level. We've hired a guy in his 50's without a high-school diploma and no prior experience in construction. Comes in 15 minutes early, leaves 15 minutes after we clock out, and busts his ass every day. He's also quite smart, being very strategic how he organizes and plans day to day tasks. His resume looked like a 4th grader wrote it. [NEWLINE] [NEWLINE] If you're hiring for a position that isn't directly responsible for representing your company on a professional level through his/her writing abilities, I think you're missing out on some great people. [NEWLINE] [NEWLINE] Now, if I were hiring a real-estate agent and she/he made all sorts of spelling errors I would take them less seriously. That sort of thing is important in their line of work. </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0634, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry, this doesn't hold water.<mask>'s the point of posting statistics<mask> there is no baseline to compare them to? [NEWLINE] [NEWLINE] Of course the statistics the other commenter provided were presented in an even more misleading way. Instead of adding up ALL violent crimes, he presented them one by one.<mask> happens<mask> you add them all together? I ended up with 1.102% which means that out of **100** black people, one of them has committed an act of violence serious enough to go to jail over. Remember, this data only shows convictions and does not show plea deals, or violence that didn't result in arrest. [NEWLINE] [NEWLINE] You<mask> need to factor in the population itself, which includes infants, toddlers and bedridden elderly. Pretty sure OP is not saying that black infants, toddlers, and octogenarians are violent...  Around 33% of the entire US population is under 14 or over 65. Once you factor that in, that means around one in 67 black people were *convicted* for violent crimes. [NEWLINE] [NEWLINE] <mask> you had a 1 in 67 chance of contracting a rare disease would you get tested for it more frequently than a friend who had a 1 in 1,000 chance? Wouldn't you take precautions<mask> you could? let's say you were an actuary...</s>
Label encoding: <s>Sorry, this doesn't hold water. What's the point of posting statistics if there is no baseline to compare them to? [NEWLINE] [NEWLINE] Of course the statistics the other commenter provided were presented in an even more misleading way. Instead of adding up ALL violent crimes, he presented them one by one. What happens when you add them all together? I ended up with 1.102% which means that out of **100** black people, one of them has committed an act of violence serious enough to go to jail over. Remember, this data only shows convictions and does not show plea deals, or violence that didn't result in arrest. [NEWLINE] [NEWLINE] You also need to factor in the population itself, which includes infants, toddlers and bedridden elderly. Pretty sure OP is not saying that black infants, toddlers, and octogenarians are violent...  Around 33% of the entire US population is under 14 or over 65. Once you factor that in, that means around one in 67 black people were *convicted* for violent crimes. [NEWLINE] [NEWLINE] If you had a 1 in 67 chance of contracting a rare disease would you get tested for it more frequently than a friend who had a 1 in 1,000 chance? Wouldn't you take precautions if you could? let's say you were an actuary...</s>
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Were they to agree with Man B, they may<mask> feel compelled to action, to assist those they see Man A or his organization to be manipulating. [ENDQ] [NEWLINE] <mask> yes, they might do that!  That would be a subset of actions in the "not rushing out to church for the salvation of their immortal soul" group, which was the subject of conversation in this case. [NEWLINE] [NEWLINE] Of course there are a lot of other options.  To be clear I'll bring in an example I used elsewhere: [NEWLINE] [NEWLINE] It is impossible to remain functionally agnostic about the existence of a live hand grenade under your seat.  No matter<mask> broad spectrum of action you may imagine, you must ultimately start by either moving or staying.  Inaction is not neutral.  Leaning down to look under the seat is functionally identical to sitting still:<mask> the grenade is there, you wasted time and you're dead. [NEWLINE] [NEWLINE] To go back to the original argument, there is obviously a difference between someone who walks wordlessly away from someone telling them that they were raped and someone who punches them in the face instead.  That fact does not change the other fact that inaction in the face of a rape allegation in the situation that spawned the argument is just another way of siding with the accused.</s>
Label encoding: <s> [STARTQ] Were they to agree with Man B, they may also feel compelled to action, to assist those they see Man A or his organization to be manipulating. [ENDQ] [NEWLINE] Why yes, they might do that!  That would be a subset of actions in the "not rushing out to church for the salvation of their immortal soul" group, which was the subject of conversation in this case. [NEWLINE] [NEWLINE] Of course there are a lot of other options.  To be clear I'll bring in an example I used elsewhere: [NEWLINE] [NEWLINE] It is impossible to remain functionally agnostic about the existence of a live hand grenade under your seat.  No matter what broad spectrum of action you may imagine, you must ultimately start by either moving or staying.  Inaction is not neutral.  Leaning down to look under the seat is functionally identical to sitting still: if the grenade is there, you wasted time and you're dead. [NEWLINE] [NEWLINE] To go back to the original argument, there is obviously a difference between someone who walks wordlessly away from someone telling them that they were raped and someone who punches them in the face instead.  That fact does not change the other fact that inaction in the face of a rape allegation in the situation that spawned the argument is just another way of siding with the accused.</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes,<mask> governments are known for wise and frugal spending decisions not motivated by politics at all. [NEWLINE] [NEWLINE] <mask> we started a draconian estate tax people would simply move to a flag of convenience, setup trusts and other paper entities, and otherwise obfuscate their wealth. You would gain little to no income for your efforts. [NEWLINE] [NEWLINE] Beyond this, even<mask> they had the money in hand, they would want to spend it immediately on things immediately useful to them. They won't be somberly building up social security reserves (especially considering<mask> they got<mask> low in the first place), they will be passing programs to show voters and lobby groups, for votes and money respectively. [NEWLINE] [NEWLINE] This isn't the silver bullet you are looking for, and<mask> anything will create two groups of the wealthy. Those with<mask> much money an army of lawyers and accountants keeps their wealth in their hands generation after generation and first generation wealth that ends up paying serious estate tax for not having an army of financial planners. [NEWLINE] [NEWLINE] [And all of this wouldn't even balance the budget, even<mask> raised to 100% tax.]( [URL] -1QY&amp;list=PL-erRSWG3IoAsjCO0lwmVUB3OcuYaZvLP)</s>
Label encoding: <s>Yes, because governments are known for wise and frugal spending decisions not motivated by politics at all. [NEWLINE] [NEWLINE] if we started a draconian estate tax people would simply move to a flag of convenience, setup trusts and other paper entities, and otherwise obfuscate their wealth. You would gain little to no income for your efforts. [NEWLINE] [NEWLINE] Beyond this, even if they had the money in hand, they would want to spend it immediately on things immediately useful to them. They won't be somberly building up social security reserves (especially considering how they got so low in the first place), they will be passing programs to show voters and lobby groups, for votes and money respectively. [NEWLINE] [NEWLINE] This isn't the silver bullet you are looking for, and if anything will create two groups of the wealthy. Those with so much money an army of lawyers and accountants keeps their wealth in their hands generation after generation and first generation wealth that ends up paying serious estate tax for not having an army of financial planners. [NEWLINE] [NEWLINE] [And all of this wouldn't even balance the budget, even if raised to 100% tax.]( [URL] -1QY&amp;list=PL-erRSWG3IoAsjCO0lwmVUB3OcuYaZvLP)</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would ask, are there any examples of successful communism? Two examples you cited, the kibbutzim and China, I wouldn't count. The kibbutzim are no longer the perfect examples of communism that their founders envisioned. They have introduced wages and can only survive by very generous subsidies from the capitalist Israeli gobernment. China is becoming more capitalist everyday. Compare their society today to<mask> they were more purely communist and few would say that those times were better for your average Chinese citizen. [NEWLINE] [NEWLINE] The only time it really works is in those few communities in the US and elsewhere that you mentioned.<mask> those are voluntary communes. No one is forced to stay and<mask> you don't like it, you only have to go to the next town over.<mask> communism becomes the state government, that idea of individual consent disappears.<mask> preventing emigration may not be<mask> Marx and Engels had in mind, to my knowledge, it has been practiced in every communist country. [NEWLINE] [NEWLINE] <mask>, yes, communism is not<mask> bad<mask> restricted to individual communities. Once it becomes the national government and inescapable to the populace, it is not a good thing. Once it becomes political and seeks global revolution and exporting its system to the rest of the world, it becomes dangerous.</s>
Label encoding: <s>I would ask, are there any examples of successful communism? Two examples you cited, the kibbutzim and China, I wouldn't count. The kibbutzim are no longer the perfect examples of communism that their founders envisioned. They have introduced wages and can only survive by very generous subsidies from the capitalist Israeli gobernment. China is becoming more capitalist everyday. Compare their society today to when they were more purely communist and few would say that those times were better for your average Chinese citizen. [NEWLINE] [NEWLINE] The only time it really works is in those few communities in the US and elsewhere that you mentioned. But those are voluntary communes. No one is forced to stay and if you don't like it, you only have to go to the next town over. When communism becomes the state government, that idea of individual consent disappears. While preventing emigration may not be what Marx and Engels had in mind, to my knowledge, it has been practiced in every communist country. [NEWLINE] [NEWLINE] So, yes, communism is not so bad when restricted to individual communities. Once it becomes the national government and inescapable to the populace, it is not a good thing. Once it becomes political and seeks global revolution and exporting its system to the rest of the world, it becomes dangerous.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Precision**:<mask> Fahrenheit is technically more precise<mask> of its scale,<mask><mask>? Would a hypothetical new temperature system be more useful<mask> you multiplied the Fahrenheit system by 10 (100 F became 1000 new units)? With Celsius, it's not like its scale is<mask> compact that each degree is an absurd difference.<mask><mask>, I personally prefer the fact that each 10 degrees is a more meaningful difference.<mask> my point is that 22 C is still a hell of a lot like 23 C.<mask> the scale really isn't an issue. [NEWLINE] [NEWLINE] **Water**: Is it worth having a scale based on water?<mask> you live anywhere that snows, the marker of 0<mask> "freezing" seems like a very logical and useful one to me. Boiling,<mask><mask>, is less useful in everyday weather. [NEWLINE] [NEWLINE] **Conversion**: and at the end of the day, Celsius and Kelvin are the ones used in science and whatnot.<mask><mask> it's just simpler<mask> we have less temperature systems, especially<mask> we have to convert between scales which is a massive pain. I don't think the benefits of Fahrenheit really matter much (are Canadians worse at precisely telling time than Americans?) and the negatives make Fahrenheit kind of annoying.<mask> I am a pretty biased Canadian.</s>
Label encoding: <s>**Precision**: while Fahrenheit is technically more precise because of its scale, so what? Would a hypothetical new temperature system be more useful if you multiplied the Fahrenheit system by 10 (100 F became 1000 new units)? With Celsius, it's not like its scale is so compact that each degree is an absurd difference. In fact, I personally prefer the fact that each 10 degrees is a more meaningful difference. But my point is that 22 C is still a hell of a lot like 23 C. So the scale really isn't an issue. [NEWLINE] [NEWLINE] **Water**: Is it worth having a scale based on water? If you live anywhere that snows, the marker of 0 as "freezing" seems like a very logical and useful one to me. Boiling, I agree, is less useful in everyday weather. [NEWLINE] [NEWLINE] **Conversion**: and at the end of the day, Celsius and Kelvin are the ones used in science and whatnot. I think it's just simpler if we have less temperature systems, especially if we have to convert between scales which is a massive pain. I don't think the benefits of Fahrenheit really matter much (are Canadians worse at precisely telling time than Americans?) and the negatives make Fahrenheit kind of annoying. But I am a pretty biased Canadian.</s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem with making it a *core* discipline, is that philosophy devoid of context is incredibly abstract and abstract concepts are boring to children of school age. [NEWLINE] [NEWLINE] <mask> we should be doing is teaching logic... in math classes. And to a degree we do. [NEWLINE] [NEWLINE] And teaching ethics... in English class and History class. And again, we do. [NEWLINE] [NEWLINE] And teaching epistomology... in Science class. Ditto. [NEWLINE] [NEWLINE] And teaching critical thinking (which isn't really a field of philosophy,<mask> whatever), in all classes. [NEWLINE] [NEWLINE] Probably theology should be skipped in public schools... and metaphysics, largely for the same reason. [NEWLINE] [NEWLINE] Trying to teach philosophy in primary/secondary schools is a recipe for some teacher thinking it's actually a good idea to teach that Heidegger quote that /u/Bulvye mentioned. It's a recipe for people hating philosophy even more than they already hate it. [NEWLINE] [NEWLINE] Could we use more emphasis on some of the fields of philosophy in schools? Sure.<mask> we *don't* need, and really should avoid, is teaching it<mask> a *discipline* before college, except<mask> perhaps an elective for kids that show an interest (guess<mask>, we do that already too in many places). </s>
Label encoding: <s>The problem with making it a *core* discipline, is that philosophy devoid of context is incredibly abstract and abstract concepts are boring to children of school age. [NEWLINE] [NEWLINE] What we should be doing is teaching logic... in math classes. And to a degree we do. [NEWLINE] [NEWLINE] And teaching ethics... in English class and History class. And again, we do. [NEWLINE] [NEWLINE] And teaching epistomology... in Science class. Ditto. [NEWLINE] [NEWLINE] And teaching critical thinking (which isn't really a field of philosophy, but whatever), in all classes. [NEWLINE] [NEWLINE] Probably theology should be skipped in public schools... and metaphysics, largely for the same reason. [NEWLINE] [NEWLINE] Trying to teach philosophy in primary/secondary schools is a recipe for some teacher thinking it's actually a good idea to teach that Heidegger quote that /u/Bulvye mentioned. It's a recipe for people hating philosophy even more than they already hate it. [NEWLINE] [NEWLINE] Could we use more emphasis on some of the fields of philosophy in schools? Sure. What we *don't* need, and really should avoid, is teaching it as a *discipline* before college, except as perhaps an elective for kids that show an interest (guess what, we do that already too in many places). </s>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0617, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0815, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One extraneous thing I want to say, concerning your comment about a composition's live equivalent, and<mask> one should simulate a fade-out... [NEWLINE] [NEWLINE] Consider the basic idea of a performance versus a recording.<mask> did you get the idea that a performance must be able to properly simulate a recording or that a recording is a less-substantial substitute for a live performance? That may have originally (I mean over a hundred years ago originally) been the case,<mask> an artist has the right the engineer his recording to be solely a recording. There's no rule that it must be performed accurately on a stage, or that there's some standard for similarity between a studio recording and a live performance, below which it's not'real' music. That's a pre-conceived notion a lot of us have.<mask> I'm a musician, and I'm not interested in pop appeal, I may decide to play completely different music on-stage and in-record. [NEWLINE] [NEWLINE] <mask>, a fade-out is a compositional tool just like any other; it can be used cheaply and lazily, or it can be used appropriately and intelligently.<mask><mask> Hey Jude fades out, which is logical,<mask> it ends in a repetitive, joyous, and cyclical manner.</s>
Label encoding: <s>One extraneous thing I want to say, concerning your comment about a composition's live equivalent, and how one should simulate a fade-out... [NEWLINE] [NEWLINE] Consider the basic idea of a performance versus a recording. Where did you get the idea that a performance must be able to properly simulate a recording or that a recording is a less-substantial substitute for a live performance? That may have originally (I mean over a hundred years ago originally) been the case, but an artist has the right the engineer his recording to be solely a recording. There's no rule that it must be performed accurately on a stage, or that there's some standard for similarity between a studio recording and a live performance, below which it's not'real' music. That's a pre-conceived notion a lot of us have. If I'm a musician, and I'm not interested in pop appeal, I may decide to play completely different music on-stage and in-record. [NEWLINE] [NEWLINE] So, a fade-out is a compositional tool just like any other; it can be used cheaply and lazily, or it can be used appropriately and intelligently. I think Hey Jude fades out, which is logical, as it ends in a repetitive, joyous, and cyclical manner.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't really think the number of sexual partners says anything about a persons personality. I've known slutty virgins and prudish nymphomaniacs, the sexual act holds no real meaning<mask> their personal views about sex does. [NEWLINE] [NEWLINE] <mask> you found out your spouse of 10 years had sex with 1000 people before you met, would that change who they have been for the past 10 years? It's been said that whatever behaviors/personalities we maintain for a year become who we are,<mask> at<mask> point does a persons past stop being who they are now? [NEWLINE] [NEWLINE] I dated two girls who had been with 50+ plus guys and neither were any different from any other girl. One was a party girl with a little drug history that had a lot of casual sex. The other simply had 50+ failed attempts at dating. Is it fair to judge her for being dumped to many times? [NEWLINE] [NEWLINE] You can reject someone for whatever reasons you wish<mask><mask> has nothing to do with their current personality than your not really rejecting them<mask> rejecting the opinion of them that only exists in your mind. [NEWLINE] [NEWLINE] That's like firing a hard worker solely<mask> they've had to many previous jobs. The number means far less then the context of that number and their current work ethic.</s>
Label encoding: <s>I don't really think the number of sexual partners says anything about a persons personality. I've known slutty virgins and prudish nymphomaniacs, the sexual act holds no real meaning but their personal views about sex does. [NEWLINE] [NEWLINE] If you found out your spouse of 10 years had sex with 1000 people before you met, would that change who they have been for the past 10 years? It's been said that whatever behaviors/personalities we maintain for a year become who we are, so at what point does a persons past stop being who they are now? [NEWLINE] [NEWLINE] I dated two girls who had been with 50+ plus guys and neither were any different from any other girl. One was a party girl with a little drug history that had a lot of casual sex. The other simply had 50+ failed attempts at dating. Is it fair to judge her for being dumped to many times? [NEWLINE] [NEWLINE] You can reject someone for whatever reasons you wish but if has nothing to do with their current personality than your not really rejecting them but rejecting the opinion of them that only exists in your mind. [NEWLINE] [NEWLINE] That's like firing a hard worker solely because they've had to many previous jobs. The number means far less then the context of that number and their current work ethic.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No it isn't. Cutpeach is saying that a straw purchase is an example of a gun which is "legally purchased before being diverted into unlawful ownership". [NEWLINE] [NEWLINE] A straw purchase is "the illegal purchase of a firearm by one person for another." [ATF]( [URL] ) [NEWLINE] [NEWLINE] [NEWLINE] [Abramski v. United States]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] <mask> a person buys a gun intending to later sell it to someone else, the government often prosecutes the initial buyer under 18 U.S.C. § 922 (a)(6) for making a false statement about the identity of the buyer that is "material to the lawfulness of the sale." These prosecutions rely on the court-created "straw purchaser" doctrine, a legal fiction that treats the ultimate recipient of a firearm<mask> the "actual buyer," and the immediate purchaser<mask> a mere "straw man." [ENDQ] [STARTQ] [ENDQ] &gt; The lower courts uniformly agree that a buyer's intent to resell a gun to someone who cannot lawfully buy it is a fact "material to the lawfulness of the sale."<mask> the Fourth, Sixth, and Eleventh Circuits have split with the Fifth and Ninth Circuits about whether the same is true<mask> the ultimate recipient can lawfully buy a gun.</s>
Label encoding: <s>No it isn't. Cutpeach is saying that a straw purchase is an example of a gun which is "legally purchased before being diverted into unlawful ownership". [NEWLINE] [NEWLINE] A straw purchase is "the illegal purchase of a firearm by one person for another." [ATF]( [URL] ) [NEWLINE] [NEWLINE] [NEWLINE] [Abramski v. United States]( [URL] ) [NEWLINE] [NEWLINE] [STARTQ] When a person buys a gun intending to later sell it to someone else, the government often prosecutes the initial buyer under 18 U.S.C. § 922 (a)(6) for making a false statement about the identity of the buyer that is "material to the lawfulness of the sale." These prosecutions rely on the court-created "straw purchaser" doctrine, a legal fiction that treats the ultimate recipient of a firearm as the "actual buyer," and the immediate purchaser as a mere "straw man." [ENDQ] [STARTQ] [ENDQ] &gt; The lower courts uniformly agree that a buyer's intent to resell a gun to someone who cannot lawfully buy it is a fact "material to the lawfulness of the sale." But the Fourth, Sixth, and Eleventh Circuits have split with the Fifth and Ninth Circuits about whether the same is true when the ultimate recipient can lawfully buy a gun.</s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The only people who should be taxed 80% are people making tens of millions or more.  $500k a year doesn't warrant such heavy taxes, especially<mask> it is EARNED  income.  EARNED. EARNED EARNED EARNED EARNED income. [NEWLINE] [NEWLINE] Lawyers and doctors and small business owners are still working schmucks.  I'm talking about those that own huge amounts of capital who<mask> don't need to perform LABOR to EARN an income. [NEWLINE] [NEWLINE] You want to talk about fair? <mask> is it fair that I am born into a world<mask> I own nothing, and a Walton is born into a world<mask> he owns.1% of the entire universe? [NEWLINE] [NEWLINE] Capital gains taxes are pathetically low, and that's<mask> rich people make their money.  Your $500k a year lawyer is probably paying 35% in taxes.  A Walton is probably paying 17% in taxes<mask> they are all capital gains.  LABOR shouldn't be taxed harshly<mask> WE WANT PEOPLE TO DO LABOR. [NEWLINE] [NEWLINE] LEISURE should be taxed.  Sitting on giant piles of money should be taxed.  Collecting massive rents and siphoning all your laborers wealth should be taxed.</s>
Label encoding: <s>The only people who should be taxed 80% are people making tens of millions or more.  $500k a year doesn't warrant such heavy taxes, especially if it is EARNED  income.  EARNED. EARNED EARNED EARNED EARNED income. [NEWLINE] [NEWLINE] Lawyers and doctors and small business owners are still working schmucks.  I'm talking about those that own huge amounts of capital who therefore don't need to perform LABOR to EARN an income. [NEWLINE] [NEWLINE] You want to talk about fair?  How is it fair that I am born into a world where I own nothing, and a Walton is born into a world where he owns.1% of the entire universe? [NEWLINE] [NEWLINE] Capital gains taxes are pathetically low, and that's how rich people make their money.  Your $500k a year lawyer is probably paying 35% in taxes.  A Walton is probably paying 17% in taxes because they are all capital gains.  LABOR shouldn't be taxed harshly because WE WANT PEOPLE TO DO LABOR. [NEWLINE] [NEWLINE] LEISURE should be taxed.  Sitting on giant piles of money should be taxed.  Collecting massive rents and siphoning all your laborers wealth should be taxed.</s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No problem! [NEWLINE] [NEWLINE] I don't know<mask> anyone else has mentioned it,<mask> the other large scale experiment in anarchism occurred during the Italian Revolution. I can't really speak of it much,<mask> I haven't studied it to the same extent<mask> the Zapatistas. It's remnants is Mondragon, a federation of worker cooperatives in Basque Spain. You might have heard of it. Anyway, during the revolution anarchist took over the area, and instituted self-managment, land redistribution etc.<mask> economic productivity did increase, and there was an increase in personal and political freedom, the group was largely defeated militarily by mussolini's fascists. This was a result that someone in another comment thread suggested would  likely happen<mask> anarchist forms of organization were instituted.<mask>, with the advent of the internet and other forms of globalizing technology, large enough networks have been able to be formed<mask><mask> to be resilient,<mask> seen with the Zapatistas. [NEWLINE] [NEWLINE] [NEWLINE] There are<mask> numerous smaller organizations,<mask> well<mask> temporary organizations (such<mask> the Battle of Seattle) that operate or have operated successfully with anarchist forms of decision making, organization, resource-allocation etc. For instance, cooperative businesses can be seen, in some respects,<mask> anarchist organization. </s>
Label encoding: <s>No problem! [NEWLINE] [NEWLINE] I don't know if anyone else has mentioned it, but the other large scale experiment in anarchism occurred during the Italian Revolution. I can't really speak of it much, as I haven't studied it to the same extent as the Zapatistas. It's remnants is Mondragon, a federation of worker cooperatives in Basque Spain. You might have heard of it. Anyway, during the revolution anarchist took over the area, and instituted self-managment, land redistribution etc. While economic productivity did increase, and there was an increase in personal and political freedom, the group was largely defeated militarily by mussolini's fascists. This was a result that someone in another comment thread suggested would  likely happen if anarchist forms of organization were instituted. However, with the advent of the internet and other forms of globalizing technology, large enough networks have been able to be formed so as to be resilient, as seen with the Zapatistas. [NEWLINE] [NEWLINE] [NEWLINE] There are also numerous smaller organizations, as well as temporary organizations (such as the Battle of Seattle) that operate or have operated successfully with anarchist forms of decision making, organization, resource-allocation etc. For instance, cooperative businesses can be seen, in some respects, as anarchist organization. </s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It can be thought of<mask> a mental problem (<mask> a hormonal one would probably be more accurate)<mask> it is a life threatening medical issues<mask> the only affective cure or treatment is transitioning.  You keep saying you do not understand, and that is really understandabke seeing<mask> you are cis,<mask> your lack of understanding does not change<mask> people work in the real world, and pretty much every relevant medical professional agrees that<mask> they feel is real and transitioning is the only real good option. [NEWLINE] [NEWLINE] [NEWLINE] Sex and gender are not<mask> simple<mask> two chromosomes, development depends on a huge number of different genes and external factors in the womb,<mask> there is a problem with any one of these it can lead to the improper downregulation of some hormones and/or upregulation of the another, this can lead to differences in brain and hormone regulation throught their life, development of the "wrong" genitals for the chromosomes they have, development of genitals that do not conform to either male or female, and a host of other conditions.  Any of these could potentially end up with the person being assigned the wrong gender at birth, and even without things like that humans are very complex biological systems and simplistic view of sex and gender really do not accurately cover everything that happens in real life.</s>
Label encoding: <s>It can be thought of as a mental problem ( although a hormonal one would probably be more accurate) but it is a life threatening medical issues where the only affective cure or treatment is transitioning.  You keep saying you do not understand, and that is really understandabke seeing as you are cis, but your lack of understanding does not change how people work in the real world, and pretty much every relevant medical professional agrees that what they feel is real and transitioning is the only real good option. [NEWLINE] [NEWLINE] [NEWLINE] Sex and gender are not as simple as two chromosomes, development depends on a huge number of different genes and external factors in the womb, if there is a problem with any one of these it can lead to the improper downregulation of some hormones and/or upregulation of the another, this can lead to differences in brain and hormone regulation throught their life, development of the "wrong" genitals for the chromosomes they have, development of genitals that do not conform to either male or female, and a host of other conditions.  Any of these could potentially end up with the person being assigned the wrong gender at birth, and even without things like that humans are very complex biological systems and simplistic view of sex and gender really do not accurately cover everything that happens in real life.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0558, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You do realize that homosexual people are still like, adults, right?  They're really just normal people and should know<mask> to conduct themselves in public.  This doesn't mean they shouldn't kiss or dance with someone of the same sex<mask> heterosexual couples do that all the time. <mask> most gay people won't hit on someone that's straight.  Just doesn't make sense.  No one wants to bark up the wrong tree. [NEWLINE] [NEWLINE] I'm getting married soon and my aunt is gay.  I'm definitely inviting her partner<mask> well.  Sure, her partner is just a really weird person (thinks rocks in jewelry are yearning to be free, etc.),<mask> it's not<mask> she's gay.  It's<mask> she's just a weird-ass person. [NEWLINE] [NEWLINE] That being said, my aunt is part of my family, has been with her partner for at least 7 years now (they had a commitment ceremony a<mask> back,<mask> I was away at college and unable to come) and they are just<mask> committed<mask> any other couple. [NEWLINE] [NEWLINE] Finally, I guess I'm confused<mask> to<mask> you would be uncomfortable with displays of affection between gay and lesbian people.  You state no religious objections,<mask><mask> makes you feel uncomfortable?</s>
Label encoding: <s>You do realize that homosexual people are still like, adults, right?  They're really just normal people and should know how to conduct themselves in public.  This doesn't mean they shouldn't kiss or dance with someone of the same sex because heterosexual couples do that all the time.  But most gay people won't hit on someone that's straight.  Just doesn't make sense.  No one wants to bark up the wrong tree. [NEWLINE] [NEWLINE] I'm getting married soon and my aunt is gay.  I'm definitely inviting her partner as well.  Sure, her partner is just a really weird person (thinks rocks in jewelry are yearning to be free, etc.), but it's not because she's gay.  It's because she's just a weird-ass person. [NEWLINE] [NEWLINE] That being said, my aunt is part of my family, has been with her partner for at least 7 years now (they had a commitment ceremony a while back, but I was away at college and unable to come) and they are just as committed as any other couple. [NEWLINE] [NEWLINE] Finally, I guess I'm confused as to why you would be uncomfortable with displays of affection between gay and lesbian people.  You state no religious objections, so what makes you feel uncomfortable?</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No, it means that that society must have a very good reason for overriding the bond that every other society feels towards dogs. [ENDQ] [NEWLINE] This sounds like a circular argument.  "There is a reason society doesn't eat dogs...<mask> they must have a very good reason".  The same goes for the "People generally didn't..." argument. Until the 19th century, people generally didn't eat tomatoes. Does that mean it was immoral to eat tomatoes? [NEWLINE] [NEWLINE] [STARTQ] Dogs are a very inefficient food source.[...]<mask> eat dog for one day<mask> you could eat cow for 5 days? [ENDQ] [NEWLINE] <mask><mask><mask> it's inefficient? People do lots of things that aren't maximally efficient, and that includes choices of food.<mask> does that create  a *moral* injunction against eating dog, or any other "inefficient" food (eg, most spices, caviar...)? [NEWLINE] [NEWLINE] And your "chain of eating" argument has to be applied globally.<mask> that's your reason, you can't make special pleading for dogs. <mask> about fish? Particularly in the case of large fish, they're even further along the chain from solar energy than dogs are,<mask> even more inefficient,<mask> morally even *worse* than eating dog.  </s>
Label encoding: <s> [STARTQ] No, it means that that society must have a very good reason for overriding the bond that every other society feels towards dogs. [ENDQ] [NEWLINE] This sounds like a circular argument.  "There is a reason society doesn't eat dogs... because they must have a very good reason".  The same goes for the "People generally didn't..." argument. Until the 19th century, people generally didn't eat tomatoes. Does that mean it was immoral to eat tomatoes? [NEWLINE] [NEWLINE] [STARTQ] Dogs are a very inefficient food source.[...] Why eat dog for one day when you could eat cow for 5 days? [ENDQ] [NEWLINE] So what if it's inefficient? People do lots of things that aren't maximally efficient, and that includes choices of food. How does that create  a *moral* injunction against eating dog, or any other "inefficient" food (eg, most spices, caviar...)? [NEWLINE] [NEWLINE] And your "chain of eating" argument has to be applied globally. If that's your reason, you can't make special pleading for dogs.  What about fish? Particularly in the case of large fish, they're even further along the chain from solar energy than dogs are, thus even more inefficient, thus morally even *worse* than eating dog.  </s>
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That was the whole reason for the Waterloo campaign. Napoleon's MO was to isolate his enemies and defeat them in detail.<mask> he, Ney, and Grouchy had managed to defeat Blucher and Wellsley, Napoleon could isolate the remaining Allies and conceivably pull off another Austerlitz. [ENDQ] [NEWLINE] The big problem is that this is 1815 and not 1805, Anti-Napoleon forces have more or less adapted to his lighting fast flanking maneuvers that enabled victories like Austerlitz. Whereas before Napoleon was able to use mass columns to rapidly get around his enemy's flanks, now everybody else is using the same tactics. That's<mask> Austerlitz is really one of the last great napoleonic victories and subsequent battles (Borodino comes to mind) mostly consist of Napoleon attacking headon and the battles turning into bloody attrition fighting rather than his earlier brilliant victories. [NEWLINE] [NEWLINE] [STARTQ] &gt;Wouldn't this have been true of the Allies on the continent<mask> well? [ENDQ] [NEWLINE] Not nearly to the extent of France, even<mask> early<mask> 1812 most of Napoleon's Grand Armee were consisted of non-Frenchmen: Dutch, Italians, Germans, Poles etc. This source of manpower was removed by 1815.</s>
Label encoding: <s> [STARTQ] That was the whole reason for the Waterloo campaign. Napoleon's MO was to isolate his enemies and defeat them in detail. If he, Ney, and Grouchy had managed to defeat Blucher and Wellsley, Napoleon could isolate the remaining Allies and conceivably pull off another Austerlitz. [ENDQ] [NEWLINE] The big problem is that this is 1815 and not 1805, Anti-Napoleon forces have more or less adapted to his lighting fast flanking maneuvers that enabled victories like Austerlitz. Whereas before Napoleon was able to use mass columns to rapidly get around his enemy's flanks, now everybody else is using the same tactics. That's why Austerlitz is really one of the last great napoleonic victories and subsequent battles (Borodino comes to mind) mostly consist of Napoleon attacking headon and the battles turning into bloody attrition fighting rather than his earlier brilliant victories. [NEWLINE] [NEWLINE] [STARTQ] &gt;Wouldn't this have been true of the Allies on the continent as well? [ENDQ] [NEWLINE] Not nearly to the extent of France, even as early as 1812 most of Napoleon's Grand Armee were consisted of non-Frenchmen: Dutch, Italians, Germans, Poles etc. This source of manpower was removed by 1815.</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0635, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Aha! Here we go! You my good friend, can now afford a large collection of music on your computer for free! Say hello spotify! With some ads thrown in, you too can pay fractions of pennies to your artist of choice! [NEWLINE] [NEWLINE] Sarcasm aside, you can obtain the music you want in a legal legitimate manner that benefits (albeit very slightly) the artists and those that make it possible for them to put out music. [NEWLINE] [NEWLINE] Seriously<mask>, music i now feel there is absolutely no good reason to pirate. I can afford 10 dollars a month for streaming music to my comp and my phone. (can only use mobile with 10 buck a month subscription.) I can download the music<mask> i don't have to always have an internet connection and i assume that the stuff just expires and locks up after x time. And the library they have is insanely huge. i was very reluctant at first,<mask> now i'm a believer. [NEWLINE] [NEWLINE] TL;DR Spotify for the computer gives a legitimate way to listen to almost any music for free. You have to have ads. For $10 a month, you have no ads, you can download the music and use it offline, and you can<mask> stream and download the music to a mobile device.</s>
Label encoding: <s>Aha! Here we go! You my good friend, can now afford a large collection of music on your computer for free! Say hello spotify! With some ads thrown in, you too can pay fractions of pennies to your artist of choice! [NEWLINE] [NEWLINE] Sarcasm aside, you can obtain the music you want in a legal legitimate manner that benefits (albeit very slightly) the artists and those that make it possible for them to put out music. [NEWLINE] [NEWLINE] Seriously though, music i now feel there is absolutely no good reason to pirate. I can afford 10 dollars a month for streaming music to my comp and my phone. (can only use mobile with 10 buck a month subscription.) I can download the music so i don't have to always have an internet connection and i assume that the stuff just expires and locks up after x time. And the library they have is insanely huge. i was very reluctant at first, but now i'm a believer. [NEWLINE] [NEWLINE] TL;DR Spotify for the computer gives a legitimate way to listen to almost any music for free. You have to have ads. For $10 a month, you have no ads, you can download the music and use it offline, and you can also stream and download the music to a mobile device.</s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0760, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[This page]( [URL].c?id=pcmcat160300050011) shows<mask> energy for cooking costs very little. [NEWLINE] [NEWLINE] [And here's a Telegraph article]( [URL] ) quoting figures from the Energy Saving Trust saying that "On average, gas ovens cost around £9 a year to run, whereas electric ovens cost around £44 a year". Per meal that is an extremely low cost of around 1 to 6 pence. [NEWLINE] [NEWLINE] I had a metered house a couple of years ago and very little cash (<mask> I was often topping up by £10 or £20 a time and kept a good eye on the meter). My experience agreed with the figures above; it seemed that cooking used pennies a day, compared to heating<mask> it was possible to spend a pound or<mask> a day from basic winter use. [NEWLINE] [NEWLINE] <mask> I don't think the cost of energy for cooking is a real impediment to poor people cooking at home. Of course there is the issue of whether they're aware of the relative cost,<mask> the greater availability of energy meters might help here. [NEWLINE] [NEWLINE] FYI, it's the landlord's responsibility to repair white goods supplied with the property unless stated in a contract,<mask><mask><mask> I'm aware.</s>
Label encoding: <s>[This page]( [URL].c?id=pcmcat160300050011) shows how energy for cooking costs very little. [NEWLINE] [NEWLINE] [And here's a Telegraph article]( [URL] ) quoting figures from the Energy Saving Trust saying that "On average, gas ovens cost around £9 a year to run, whereas electric ovens cost around £44 a year". Per meal that is an extremely low cost of around 1 to 6 pence. [NEWLINE] [NEWLINE] I had a metered house a couple of years ago and very little cash ( so I was often topping up by £10 or £20 a time and kept a good eye on the meter). My experience agreed with the figures above; it seemed that cooking used pennies a day, compared to heating where it was possible to spend a pound or so a day from basic winter use. [NEWLINE] [NEWLINE] So I don't think the cost of energy for cooking is a real impediment to poor people cooking at home. Of course there is the issue of whether they're aware of the relative cost, but the greater availability of energy meters might help here. [NEWLINE] [NEWLINE] FYI, it's the landlord's responsibility to repair white goods supplied with the property unless stated in a contract, as far as I'm aware.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0693, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] i completely agree with OP. Yes<mask> someone dies it is horrific,<mask> have that person be stuck (whether they make good out of it or not) with malformed, unusable, limbs, or mental disabilities is worse in my eyes. Yes people will still care about you and love you,<mask> you aren't who you were before and that is who they truly loved or cared about. (Don't misinterpret<mask> i'm saying<mask> in<mask> you lose a leg or an arm no one will love you) [ENDQ] [STARTQ] In all cases I say the person affected will change. And in my view, for the worse. [ENDQ] [STARTQ] I have no way of knowing<mask> I would act or feel,<mask> many of the things I enjoy doing take effort from bronze and brains and<mask> I was physically incapable or mentally unstable or capable I would be SOL. [ENDQ] [STARTQ] OP summed it up well and more politically correct than me: I understand that death is permanent,<mask> I don't think suffering and being alive is better than being dead. Especially in cases<mask> things don't get better. It's great<mask> you can get over something and move on --<mask> odds are that thing will always be there eating at you, somewhere. And it will make you continually suffer. [ENDQ] [NEWLINE] </s>
Label encoding: <s> [STARTQ] i completely agree with OP. Yes if someone dies it is horrific, but have that person be stuck (whether they make good out of it or not) with malformed, unusable, limbs, or mental disabilities is worse in my eyes. Yes people will still care about you and love you, but you aren't who you were before and that is who they truly loved or cared about. (Don't misinterpret what i'm saying as in if you lose a leg or an arm no one will love you) [ENDQ] [STARTQ] In all cases I say the person affected will change. And in my view, for the worse. [ENDQ] [STARTQ] I have no way of knowing how I would act or feel, but many of the things I enjoy doing take effort from bronze and brains and if I was physically incapable or mentally unstable or capable I would be SOL. [ENDQ] [STARTQ] OP summed it up well and more politically correct than me: I understand that death is permanent, but I don't think suffering and being alive is better than being dead. Especially in cases where things don't get better. It's great if you can get over something and move on -- but odds are that thing will always be there eating at you, somewhere. And it will make you continually suffer. [ENDQ] [NEWLINE] </s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I posted a picture like that, many people would assume that I was gay. This is<mask> it has become much more common to be publicly gay. This assumption **really bothers me**. I don't like that I can't behave<mask> I want to, without people making assumptions, based on stereotypes. [ENDQ] [NEWLINE] It seems that your entire view hinges on this statement. <mask> does it matter<mask> people assume you're gay? <mask> you have a problem with people thinking you might be gay, then that's a problem with you.  That isn't a problem with them, gays, transgenders, society or the "way we are promoting acceptance". [NEWLINE] [NEWLINE] <mask><mask> it wasn't an assumption that you were gay? <mask><mask> you have blond hair<mask> due to the lighting in a specific picture, it looks more brown.  Would it "really bother you"<mask> someone made the assumption that you had brown hair based upon that photo?  Unless you're some kind of hair-Nazi, I'd imagine you wouldn't really give a fuck and the person's incorrect assumption about your hair color would be a complete non-event. <mask><mask>'s the difference between assuming your wrong hair color, and assuming your wrong sexual preference?</s>
Label encoding: <s> [STARTQ] If I posted a picture like that, many people would assume that I was gay. This is because it has become much more common to be publicly gay. This assumption **really bothers me**. I don't like that I can't behave how I want to, without people making assumptions, based on stereotypes. [ENDQ] [NEWLINE] It seems that your entire view hinges on this statement.  Why does it matter if people assume you're gay?  If you have a problem with people thinking you might be gay, then that's a problem with you.  That isn't a problem with them, gays, transgenders, society or the "way we are promoting acceptance". [NEWLINE] [NEWLINE] What if it wasn't an assumption that you were gay?  What if you have blond hair but due to the lighting in a specific picture, it looks more brown.  Would it "really bother you" if someone made the assumption that you had brown hair based upon that photo?  Unless you're some kind of hair-Nazi, I'd imagine you wouldn't really give a fuck and the person's incorrect assumption about your hair color would be a complete non-event.  So what's the difference between assuming your wrong hair color, and assuming your wrong sexual preference?</s>
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a fellow atheist, I totally agree.  I grew up in a Christian household and<mask> did many of my fellow friends who are atheists.  One of the reasons I could never get behind the "indoctrination!" battle cry was<mask> *part of raising kids is indoctrinating them with your values.* [NEWLINE] [NEWLINE] <mask> your values spring from the fountainhead of religion, then you're going to end up passing your religion onto them, even<mask> you attempt to be neutral. [NEWLINE] [NEWLINE] The Amish do something similar to this,<mask> a child comes of age (<mask><mask> its 16), they are sent away to live in the regular, non-Amish world and are allowed to make a decision<mask> to whether or not they want to stay out there or return to their roots.  Many return to their roots, many walk away. [NEWLINE] [NEWLINE] Attempting a neutral upbringing in which you can allow your children to make their own decisions is definitely noble,<mask> unfortunately, *true* neutrality wouldn't allow us to teach our children anything. <mask><mask> they wanted to speak French instead of English?  Too bad, your parents only know English.  Every child's upbringing is strictly limited to the knowledge, morals, and ethics of their parents.</s>
Label encoding: <s>As a fellow atheist, I totally agree.  I grew up in a Christian household and so did many of my fellow friends who are atheists.  One of the reasons I could never get behind the "indoctrination!" battle cry was because *part of raising kids is indoctrinating them with your values.* [NEWLINE] [NEWLINE] If your values spring from the fountainhead of religion, then you're going to end up passing your religion onto them, even if you attempt to be neutral. [NEWLINE] [NEWLINE] The Amish do something similar to this, when a child comes of age ( I think its 16), they are sent away to live in the regular, non-Amish world and are allowed to make a decision as to whether or not they want to stay out there or return to their roots.  Many return to their roots, many walk away. [NEWLINE] [NEWLINE] Attempting a neutral upbringing in which you can allow your children to make their own decisions is definitely noble, but unfortunately, *true* neutrality wouldn't allow us to teach our children anything.  What if they wanted to speak French instead of English?  Too bad, your parents only know English.  Every child's upbringing is strictly limited to the knowledge, morals, and ethics of their parents.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. Your point about social pressure is different to the previous point about 'teaching' which I'd agree with. [NEWLINE] [NEWLINE] 2. Citation please about stereotyping affecting performance in society over a sustained period of time, and not in an isolated lab environment (which was all I could find).<mask> I told someone 'everyone says you're shit at maths' right before they did a maths test, of course that would hurt their performance. [NEWLINE] [NEWLINE] 3. The 'geek' argument is completely unsubstantiated. '<mask> the guys don't want to actually interact with women who are<mask> intelligent or quirky<mask> them. they want to be able to feel smart and powerful by explaining things and talking down to the woman.' Really? Citation please. [NEWLINE] [NEWLINE] 3. Yeah I've heard a lot about this too and resume discrimination could be true. [NEWLINE] [NEWLINE] 4. Well, there is a disproportionate amount of men in STEM fields,<mask> it would seem to be the case for 'a lot' of women.<mask><mask> do you draw the line between social pressures and personal preferences?<mask> is this being attributed to'social pressures'<mask> other gender-dominated fields like car sales or fashion seem to go unquestioned. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [URL] </s>
Label encoding: <s>1. Your point about social pressure is different to the previous point about 'teaching' which I'd agree with. [NEWLINE] [NEWLINE] 2. Citation please about stereotyping affecting performance in society over a sustained period of time, and not in an isolated lab environment (which was all I could find). If I told someone 'everyone says you're shit at maths' right before they did a maths test, of course that would hurt their performance. [NEWLINE] [NEWLINE] 3. The 'geek' argument is completely unsubstantiated.'While the guys don't want to actually interact with women who are as intelligent or quirky as them. they want to be able to feel smart and powerful by explaining things and talking down to the woman.' Really? Citation please. [NEWLINE] [NEWLINE] 3. Yeah I've heard a lot about this too and resume discrimination could be true. [NEWLINE] [NEWLINE] 4. Well, there is a disproportionate amount of men in STEM fields, so it would seem to be the case for 'a lot' of women. But where do you draw the line between social pressures and personal preferences? Why is this being attributed to'social pressures' when other gender-dominated fields like car sales or fashion seem to go unquestioned. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [URL] </s>
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's my problem with your argument for the ring. <mask> should the man be sacrificing for the woman?  It's an archaic argument.  In reality, both people should be willing to sacrifice for each other.  It shouldn't require a one sided sign of dedication.  Particularly for the low income-  there's<mask> much more important things they could spend their money on.  It's not about valuing money more than the dedication-  it's about the other things that can show dedication being far more meaningful than flushing money away, and about the dated idea that it should be a one sided show of dedication. [NEWLINE] [NEWLINE] Same with the wedding itself-  I don't need to spend 10-20K to feel dedicated or for it to be memorable.  I'm not against some form of ceremony/celebration (<mask> it isn't necessary, I know plenty of people who just got a clerk marriage and are totally happy),<mask> the amount of money spent on it is insane.  All I need for it to be memorable is my partner and an oath.  Let my mother and sister be there and it will be even more special.  Beyond that nobody else will add to the experience for me.</s>
Label encoding: <s>Here's my problem with your argument for the ring.  Why should the man be sacrificing for the woman?  It's an archaic argument.  In reality, both people should be willing to sacrifice for each other.  It shouldn't require a one sided sign of dedication.  Particularly for the low income-  there's so much more important things they could spend their money on.  It's not about valuing money more than the dedication-  it's about the other things that can show dedication being far more meaningful than flushing money away, and about the dated idea that it should be a one sided show of dedication. [NEWLINE] [NEWLINE] Same with the wedding itself-  I don't need to spend 10-20K to feel dedicated or for it to be memorable.  I'm not against some form of ceremony/celebration ( although it isn't necessary, I know plenty of people who just got a clerk marriage and are totally happy), but the amount of money spent on it is insane.  All I need for it to be memorable is my partner and an oath.  Let my mother and sister be there and it will be even more special.  Beyond that nobody else will add to the experience for me.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] On average, people put more resources into the world than they take out of it. [ENDQ] [NEWLINE] No, they don't. The fact that we don't have enough resources is one of the great reasons to not have children. [NEWLINE] [NEWLINE] [STARTQ] the benefits of that child becoming a productive member of society are spread around. [ENDQ] [NEWLINE] <mask> benefits, specifically? Your argument seems to be that, generally speaking, the more people there are in the world, the better off it is. That's obviously not true. [NEWLINE] [NEWLINE] [STARTQ] On average, people are glad to be alive. [ENDQ] [NEWLINE] That's irrelevant,<mask> a person who is never born is not *unhappy* to *not* be alive.<mask> the average person's happiness is 6 out of 10, the world's happiness is a 6 no matter<mask> many people it has. (And,<mask><mask>, the more people there are, the more likely it is that everybody--and<mask> the world--will drop to a 5.) [NEWLINE] [NEWLINE] [STARTQ] the odds are pretty good that,<mask> you have a child, you've done something very positive for the world [ENDQ] [NEWLINE] You keep saying things like this and then offering zero explanation or evidence.<mask> exactly are you talking about? [NEWLINE] </s>
Label encoding: <s> [STARTQ] On average, people put more resources into the world than they take out of it. [ENDQ] [NEWLINE] No, they don't. The fact that we don't have enough resources is one of the great reasons to not have children. [NEWLINE] [NEWLINE] [STARTQ] the benefits of that child becoming a productive member of society are spread around. [ENDQ] [NEWLINE] What benefits, specifically? Your argument seems to be that, generally speaking, the more people there are in the world, the better off it is. That's obviously not true. [NEWLINE] [NEWLINE] [STARTQ] On average, people are glad to be alive. [ENDQ] [NEWLINE] That's irrelevant, since a person who is never born is not *unhappy* to *not* be alive. If the average person's happiness is 6 out of 10, the world's happiness is a 6 no matter how many people it has. (And, in fact, the more people there are, the more likely it is that everybody--and thus the world--will drop to a 5.) [NEWLINE] [NEWLINE] [STARTQ] the odds are pretty good that, if you have a child, you've done something very positive for the world [ENDQ] [NEWLINE] You keep saying things like this and then offering zero explanation or evidence. What exactly are you talking about? [NEWLINE] </s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For every song with a positive message like 'Keep ya head up' or 'Changes', he's got 10 more that glorify gang violence and shootings. The idea that he was non-violent is totally at odds with the music he actually made. [NEWLINE] [NEWLINE] A brief glance at his [wikipedia page]( [URL] #Personal_life) will<mask> show you that he was an exceptionally violent person: He assaulted another rapper with a baseball bat in 1993, and assaulted a film director in 1994. In both cases he plead guilty. He was<mask> found guilty of gang raping a woman in a hotel, a crime for which he was sent to prison. [NEWLINE] [NEWLINE] With anybody else, those three crimes would be enough to judge someone<mask> an extremely violent and unsafe person highly unsuited to be a role model for anybody,<mask> for some reason he's earned a reputation for being some kinda hip-hop gandhi. There are plenty of legendary rappers like Mos Def, Nas and Andre 3000 whose positive messages are overlooked in favour of Tupac,<mask> managing to avoid assaulting or raping people. [NEWLINE] The fact that Tupac garners<mask> much respect<mask> he does is symptomatic of massive immaturity in certain parts of the hip-hop community.</s>
Label encoding: <s>For every song with a positive message like 'Keep ya head up' or 'Changes', he's got 10 more that glorify gang violence and shootings. The idea that he was non-violent is totally at odds with the music he actually made. [NEWLINE] [NEWLINE] A brief glance at his [wikipedia page]( [URL] #Personal_life) will also show you that he was an exceptionally violent person: He assaulted another rapper with a baseball bat in 1993, and assaulted a film director in 1994. In both cases he plead guilty. He was also found guilty of gang raping a woman in a hotel, a crime for which he was sent to prison. [NEWLINE] [NEWLINE] With anybody else, those three crimes would be enough to judge someone as an extremely violent and unsafe person highly unsuited to be a role model for anybody, but for some reason he's earned a reputation for being some kinda hip-hop gandhi. There are plenty of legendary rappers like Mos Def, Nas and Andre 3000 whose positive messages are overlooked in favour of Tupac, despite managing to avoid assaulting or raping people. [NEWLINE] The fact that Tupac garners as much respect as he does is symptomatic of massive immaturity in certain parts of the hip-hop community.</s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for including that source!<mask> the graduation rates aren't all that different, it is true that athletes do have higher graduation rates.<mask><mask> it would be interesting to compare<mask> well student athletes do academically vs.<mask> other students do academically. [NEWLINE] [NEWLINE] <mask> do you think about students<mask> volunteer, play sports, participate in clubs, and study every single week?<mask><mask> that probably adds up to 20 hours too. Student athletes, such<mask> the football players probably are only practicing, whereas other athletes who don't have rigorous practice schedules can fill their time doing other things.<mask> both student athletes (football players) and other students have really time consuming schedules. [NEWLINE] [NEWLINE] You do have a good point<mask> - I guess<mask> the athletes don't have time for homework whereas other students do (<mask> I was a varsity student athlete, in three clubs, and volunteering weekly and I could do it just fine) then they might not have<mask> good of grades.<mask> then that shows that school probably wasn't<mask> important to them<mask> their sport was. I really think<mask> admissions criteria are going to be changed for athletes they should at least place an importance on schooling<mask> well, or it's kind of just like a free ride.</s>
Label encoding: <s>Thanks for including that source! While the graduation rates aren't all that different, it is true that athletes do have higher graduation rates. I think it would be interesting to compare how well student athletes do academically vs. how other students do academically. [NEWLINE] [NEWLINE] What do you think about students how volunteer, play sports, participate in clubs, and study every single week? I think that probably adds up to 20 hours too. Student athletes, such as the football players probably are only practicing, whereas other athletes who don't have rigorous practice schedules can fill their time doing other things. Thus both student athletes (football players) and other students have really time consuming schedules. [NEWLINE] [NEWLINE] You do have a good point though - I guess if the athletes don't have time for homework whereas other students do ( although I was a varsity student athlete, in three clubs, and volunteering weekly and I could do it just fine) then they might not have as good of grades. But then that shows that school probably wasn't as important to them as their sport was. I really think if admissions criteria are going to be changed for athletes they should at least place an importance on schooling as well, or it's kind of just like a free ride.</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0742, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] And congratulations for deciding that students and scientists in most of the world should be blocked from scientific research results<mask> you personally don't see a problem. [ENDQ] [NEWLINE] <mask>? You are changing the subject now. First, you were talking about making government funded research results available<mask> that everyone can read the papers without paying money to the journals. Now, you've changed the topic to students and scientists in other countries accessing scientific journals. [NEWLINE] [NEWLINE] Open access journals are not "free." They just charge differently. Instead of charging for subscriptions, they charge authors to publish papers. The publication fees can be high. For US researchers, the cost can be<mask> much<mask> $1000 per article. The cost for researchers in other countries might be lower. [NEWLINE] [NEWLINE] Contrary to<mask> you are suggesting, using only open access journals does not mean that scientific papers would necessarily be more accessible to researchers in students in third world countries. Researchers would still have to pay money to publish their papers in an open access journal. [NEWLINE] [NEWLINE] <mask> you are suggesting just shifts the costs. Instead of universities paying for subscriptions, universities will just pay to publish each article. In the end, universities will still have to pay a lot of money to publish and access papers.</s>
Label encoding: <s> [STARTQ] And congratulations for deciding that students and scientists in most of the world should be blocked from scientific research results because you personally don't see a problem. [ENDQ] [NEWLINE] What? You are changing the subject now. First, you were talking about making government funded research results available so that everyone can read the papers without paying money to the journals. Now, you've changed the topic to students and scientists in other countries accessing scientific journals. [NEWLINE] [NEWLINE] Open access journals are not "free." They just charge differently. Instead of charging for subscriptions, they charge authors to publish papers. The publication fees can be high. For US researchers, the cost can be as much as $1000 per article. The cost for researchers in other countries might be lower. [NEWLINE] [NEWLINE] Contrary to what you are suggesting, using only open access journals does not mean that scientific papers would necessarily be more accessible to researchers in students in third world countries. Researchers would still have to pay money to publish their papers in an open access journal. [NEWLINE] [NEWLINE] What you are suggesting just shifts the costs. Instead of universities paying for subscriptions, universities will just pay to publish each article. In the end, universities will still have to pay a lot of money to publish and access papers.</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It has nothing to do with points. You are attempting to control a discussion by use of force and bullying. [ENDQ] [NEWLINE] Yes,<mask><mask><mask> "public pressure to stop being racist/sexist" isn't quite "bullying". [NEWLINE] [NEWLINE] [STARTQ] There are lots of communities in the world that you don't agree with.<mask> you don't like<mask> is being said somewhere, at some restaurant, at some public setting, and you are in the minority, maybe you should leave. [ENDQ] [NEWLINE] This isn't a table at a restaurant<mask>, it's a public forum which has a noticeable impact on our culture *outside of the internet* and<mask> of that we shouldn't treat the popular discussions<mask> isolated internet things. [NEWLINE] [NEWLINE] [STARTQ] Maybe I'm mistaking you or your view,<mask> you kind of remind me of the dick who goes to a cigar lounge and complains<mask> people are smoking. [ENDQ] [NEWLINE] A more apt comparison is someone who tries to get smoking banned in parks. I'm not even really talking about downvote brigading something like /r/mensrights, which would be pointless and futile,<mask> random sexist and/or racist comments elsewhere on reddit<mask> the point of the sub isn't basically being sexist.</s>
Label encoding: <s> [STARTQ] It has nothing to do with points. You are attempting to control a discussion by use of force and bullying. [ENDQ] [NEWLINE] Yes, though I think "public pressure to stop being racist/sexist" isn't quite "bullying". [NEWLINE] [NEWLINE] [STARTQ] There are lots of communities in the world that you don't agree with. If you don't like what is being said somewhere, at some restaurant, at some public setting, and you are in the minority, maybe you should leave. [ENDQ] [NEWLINE] This isn't a table at a restaurant though, it's a public forum which has a noticeable impact on our culture *outside of the internet* and because of that we shouldn't treat the popular discussions as isolated internet things. [NEWLINE] [NEWLINE] [STARTQ] Maybe I'm mistaking you or your view, but you kind of remind me of the dick who goes to a cigar lounge and complains because people are smoking. [ENDQ] [NEWLINE] A more apt comparison is someone who tries to get smoking banned in parks. I'm not even really talking about downvote brigading something like /r/mensrights, which would be pointless and futile, but random sexist and/or racist comments elsewhere on reddit where the point of the sub isn't basically being sexist.</s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] People drink<mask> much they black out, commit crimes and either die or need medical attention. This happens every weekend in the U.S. to thousands of people and we praise it! We advertise it at football games and put it all over T.V. [ENDQ] [NEWLINE] I don't know<mask> much dying from alcohol poisoning or being an alcoholic (a TRUE alcoholic) is praised on T.V. I believe you are confusing addiction with recreational use. [NEWLINE] [NEWLINE] The issue is that many drugs have heavy addictive qualities (I am not saying ALL drugs, many) -<mask> you may have control over your ability to deal with the cravings, many people will not<mask> of the effects they have on the body.<mask>, they are deemed a risk to cause addiction. I am of the opinion that individuals are far more at risk of being addicted to crack than they are to alcohol. I am<mask> of the opinion that a crack addiction is harmful to the individual. I am finally of the opinion that the harm to that individual will directly/indirectly cause harm to others. [NEWLINE] [NEWLINE] I simply believe that in a risk/reward scenario, there is far more risk in allowing individuals free reign to crack than to alcohol.</s>
Label encoding: <s> [STARTQ] People drink so much they black out, commit crimes and either die or need medical attention. This happens every weekend in the U.S. to thousands of people and we praise it! We advertise it at football games and put it all over T.V. [ENDQ] [NEWLINE] I don't know how much dying from alcohol poisoning or being an alcoholic (a TRUE alcoholic) is praised on T.V. I believe you are confusing addiction with recreational use. [NEWLINE] [NEWLINE] The issue is that many drugs have heavy addictive qualities (I am not saying ALL drugs, many) - while you may have control over your ability to deal with the cravings, many people will not because of the effects they have on the body. Therefore, they are deemed a risk to cause addiction. I am of the opinion that individuals are far more at risk of being addicted to crack than they are to alcohol. I am also of the opinion that a crack addiction is harmful to the individual. I am finally of the opinion that the harm to that individual will directly/indirectly cause harm to others. [NEWLINE] [NEWLINE] I simply believe that in a risk/reward scenario, there is far more risk in allowing individuals free reign to crack than to alcohol.</s>
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When police introduce a new form of speed prevention through easier distribution of tickets or they crack down on speeders, the motivation behind it doesn't actually make a difference. Whether they are doing it to decrease chances of accidents or they are revenue raising, the power still lies completely in the hand of the person behind the wheel. [NEWLINE] [NEWLINE] Even<mask> the police had the intention of revenue raising, that money still goes toward hiring more officers, putting more police cars on the road and getting better technology to combat road accidents. They aren't making anyone speed, they are just making the best of the fact that people are going to speed,<mask> they might<mask> well bring some revenue in from it. [NEWLINE] [NEWLINE] Edit: I should have clarified that I am referring exclusively to traffic offences eg. Speeding, drink driving etc. [NEWLINE] [NEWLINE] Edit: My view has been changed. [NEWLINE] [NEWLINE] I now see that the motivations behind revenue raising contradict the motivations that the police service is created to have.<mask> revenue raising was the goal, the Police service would not only be out for the wrong reasons,<mask> the entire outcome of<mask> they would achieve would be counterproductive to society<mask> a whole. Due to this, revenue raising cannot be seen<mask> morally sound.</s>
Label encoding: <s>When police introduce a new form of speed prevention through easier distribution of tickets or they crack down on speeders, the motivation behind it doesn't actually make a difference. Whether they are doing it to decrease chances of accidents or they are revenue raising, the power still lies completely in the hand of the person behind the wheel. [NEWLINE] [NEWLINE] Even if the police had the intention of revenue raising, that money still goes toward hiring more officers, putting more police cars on the road and getting better technology to combat road accidents. They aren't making anyone speed, they are just making the best of the fact that people are going to speed, so they might as well bring some revenue in from it. [NEWLINE] [NEWLINE] Edit: I should have clarified that I am referring exclusively to traffic offences eg. Speeding, drink driving etc. [NEWLINE] [NEWLINE] Edit: My view has been changed. [NEWLINE] [NEWLINE] I now see that the motivations behind revenue raising contradict the motivations that the police service is created to have. If revenue raising was the goal, the Police service would not only be out for the wrong reasons, but the entire outcome of what they would achieve would be counterproductive to society as a whole. Due to this, revenue raising cannot be seen as morally sound.</s>
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Going from your example<mask>,<mask> could we talk about causation of, say, rape, regarding society<mask> a whole? [NEWLINE] [NEWLINE] Banning porn (or certain kinds, like the UK is doing)? [NEWLINE] [NEWLINE] Harsher ~~laws~~ sentences for sexual abuse crimes? [NEWLINE] [NEWLINE] I hate having the discussion at all<mask> I always get stuck on this point. I know it's not the victim's fault.<mask> like OP,<mask><mask> discussing causation, even<mask> it involves the victim, is important. [NEWLINE] [NEWLINE] I have seen documented cases of women regretting sex they had<mask> tipsy or drunk and then accusing the man of rape out of shame. [NEWLINE] [NEWLINE] I have<mask>, obviously, seen many horrible documented cases, and rather high statistics, of actual rape. [NEWLINE] [NEWLINE] <mask> it's a really shitty topic. A lot of it can't be "proven" in court, especially<mask>, for example, in a lot of famous rape cases (Bill Cosby, Jian Ghomeshi recently), people are accusing each other YEARS or DECADES later.<mask> the fuck are you gonna prove that without a rape kit, bruises, some kind of alibi or proof of<mask> you were, any substantive evidence...</s>
Label encoding: <s>Going from your example though, how could we talk about causation of, say, rape, regarding society as a whole? [NEWLINE] [NEWLINE] Banning porn (or certain kinds, like the UK is doing)? [NEWLINE] [NEWLINE] Harsher ~~laws~~ sentences for sexual abuse crimes? [NEWLINE] [NEWLINE] I hate having the discussion at all because I always get stuck on this point. I know it's not the victim's fault. But like OP, I think discussing causation, even if it involves the victim, is important. [NEWLINE] [NEWLINE] I have seen documented cases of women regretting sex they had while tipsy or drunk and then accusing the man of rape out of shame. [NEWLINE] [NEWLINE] I have also, obviously, seen many horrible documented cases, and rather high statistics, of actual rape. [NEWLINE] [NEWLINE] So it's a really shitty topic. A lot of it can't be "proven" in court, especially since, for example, in a lot of famous rape cases (Bill Cosby, Jian Ghomeshi recently), people are accusing each other YEARS or DECADES later. How the fuck are you gonna prove that without a rape kit, bruises, some kind of alibi or proof of where you were, any substantive evidence...</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Though it seems like you have already been convinced, I'd like to offer an example. Oil boomed in the late 1800's and early 1900's.<mask><mask><mask> (and the invention of the car), oil companies were EVERYWHERE. Texas basically looked like one of those cartoons<mask> the only thing that existed were oil derricks. There were tens of thousands of oil companies out there. [NEWLINE] [NEWLINE] Before too long, the ones that really struck it rich started to simply by up their competitors. The movie There Will Be Blood shows this process very well and is a great movie. Pretty soon the remaining oil companies only numbered less than a dozen, I believe only 6. [NEWLINE] [NEWLINE] The government recognized that these companies had too much power and broke them up.<mask> this happened, the remaining companies all ballooned up<mask> again. Over many years, they consolidated again only this time far larger than they had started. We're basically already back to having about 6 real oil companies in the world. [NEWLINE] [NEWLINE] Here's an illustration of much of this: [NEWLINE] [pic]( [URL].com/std-oil.jpg) [NEWLINE] [NEWLINE] Here's an illustration of the banks doing the same thing: [NEWLINE] [pic]( [URL].jpg)</s>
Label encoding: <s>Though it seems like you have already been convinced, I'd like to offer an example. Oil boomed in the late 1800's and early 1900's. Because of this (and the invention of the car), oil companies were EVERYWHERE. Texas basically looked like one of those cartoons where the only thing that existed were oil derricks. There were tens of thousands of oil companies out there. [NEWLINE] [NEWLINE] Before too long, the ones that really struck it rich started to simply by up their competitors. The movie There Will Be Blood shows this process very well and is a great movie. Pretty soon the remaining oil companies only numbered less than a dozen, I believe only 6. [NEWLINE] [NEWLINE] The government recognized that these companies had too much power and broke them up. When this happened, the remaining companies all ballooned up yet again. Over many years, they consolidated again only this time far larger than they had started. We're basically already back to having about 6 real oil companies in the world. [NEWLINE] [NEWLINE] Here's an illustration of much of this: [NEWLINE] [pic]( [URL].com/std-oil.jpg) [NEWLINE] [NEWLINE] Here's an illustration of the banks doing the same thing: [NEWLINE] [pic]( [URL].jpg)</s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be clear, I'm not an apologist for rude people.  I<mask> don't think it's the responsibility of every or any black person to explain white privilege to you. <mask> people keep yelling at you for something you believe, maybe take a look at some of their reasoned arguments rather than their shrill tone? [NEWLINE] [NEWLINE] All boats rise together, brother in arms, I'm right there with you on socialism,<mask> arguing which movement is more important is silly - they're all important, and we all have our part to play in making the world a better place.  I've felt that sentiment (Your opinions are not valued) and it sucks, and maybe those are just people you don't jibe with.  Don't confuse people with the ideology.  People are fucking nasty and mean, and, you know, sometimes they have good reasons for being nasty and mean.  Malcolm X or some Black Panthers might not like me personally for being white,<mask> I understand and respect their opinions and see<mask> they're coming from.  I'm honestly not that up to date on black intellectuals, or<mask> that's a fair characterization of Malcolm,<mask> pardon my ignorance,<mask> still, ya know?</s>
Label encoding: <s>To be clear, I'm not an apologist for rude people.  I also don't think it's the responsibility of every or any black person to explain white privilege to you.  If people keep yelling at you for something you believe, maybe take a look at some of their reasoned arguments rather than their shrill tone? [NEWLINE] [NEWLINE] All boats rise together, brother in arms, I'm right there with you on socialism, but arguing which movement is more important is silly - they're all important, and we all have our part to play in making the world a better place.  I've felt that sentiment (Your opinions are not valued) and it sucks, and maybe those are just people you don't jibe with.  Don't confuse people with the ideology.  People are fucking nasty and mean, and, you know, sometimes they have good reasons for being nasty and mean.  Malcolm X or some Black Panthers might not like me personally for being white, but I understand and respect their opinions and see where they're coming from.  I'm honestly not that up to date on black intellectuals, or if that's a fair characterization of Malcolm, so pardon my ignorance, but still, ya know?</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What other crimes could those things be applied to? You say "damn near every other crime" and "a lot of other crimes",<mask> all you mention is 'battery'. Assault &amp; Battery is part of nearly every rape,<mask><mask> could you not understand<mask> rape is worse? Rape is like 5 crimes in one. Could you name 5 crimes that result in the feelings you claim are true for every other crime that *aren't a part of your average rape?* [NEWLINE] [NEWLINE] The victim could become pregnant against their will or catch a life-threatening or humiliating STD. Sex (<mask> is supposed to be the best part of life) could become ruined for that person *forever*. Obviously<mask> any of these things happen, maintaining a normal relationship and living your life normally becomes extremely difficult. [NEWLINE] [NEWLINE] Rape essentially turns the most private and protected parts of a person's life into the worst parts of their life. Follow that up by having 25% of the world pretty much think you deserved it somehow by default. [NEWLINE] [NEWLINE] Honestly, can you even name a single other crime that "devalues the victim<mask> a person" and makes "the victim often think they're going to die"? </s>
Label encoding: <s>What other crimes could those things be applied to? You say "damn near every other crime" and "a lot of other crimes", but all you mention is 'battery'. Assault &amp; Battery is part of nearly every rape, so how could you not understand why rape is worse? Rape is like 5 crimes in one. Could you name 5 crimes that result in the feelings you claim are true for every other crime that *aren't a part of your average rape?* [NEWLINE] [NEWLINE] The victim could become pregnant against their will or catch a life-threatening or humiliating STD. Sex ( what is supposed to be the best part of life) could become ruined for that person *forever*. Obviously if any of these things happen, maintaining a normal relationship and living your life normally becomes extremely difficult. [NEWLINE] [NEWLINE] Rape essentially turns the most private and protected parts of a person's life into the worst parts of their life. Follow that up by having 25% of the world pretty much think you deserved it somehow by default. [NEWLINE] [NEWLINE] Honestly, can you even name a single other crime that "devalues the victim as a person" and makes "the victim often think they're going to die"? </s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In baseball, the stats I linked to are for total wins and losses in franchise history, not individual season stats. In soccer it would be a little different<mask> of the structure of leagues,<mask><mask> you either looked only at one particular league (and ignored all games played by teams of that league against out of league teams), or you looked at all soccer teams that play each other, you still end up with a zero sum game. The real reason baseball is great from a statistical standpoint is that baseball fans love data<mask> there is a ton of great info available in several different easy-to-use data sets, whereas in other sports this is not the case<mask> much.  With regards to stocks, you are confusing a stock with the company itself. Coke's stock prices are a function of the demand for Coke stock on the market, not a function of the market share for soft drinks that Coke has. It is possible for every single soft drink manufacturer to see stock prices rise over the same period,<mask> showing that even within an industry, stocks are not a zero sum game. It is not possible for every single soccer team in the MLS to win more games than they lose over the course of a season.</s>
Label encoding: <s>In baseball, the stats I linked to are for total wins and losses in franchise history, not individual season stats. In soccer it would be a little different because of the structure of leagues, but if you either looked only at one particular league (and ignored all games played by teams of that league against out of league teams), or you looked at all soccer teams that play each other, you still end up with a zero sum game. The real reason baseball is great from a statistical standpoint is that baseball fans love data so there is a ton of great info available in several different easy-to-use data sets, whereas in other sports this is not the case as much.  With regards to stocks, you are confusing a stock with the company itself. Coke's stock prices are a function of the demand for Coke stock on the market, not a function of the market share for soft drinks that Coke has. It is possible for every single soft drink manufacturer to see stock prices rise over the same period, thus showing that even within an industry, stocks are not a zero sum game. It is not possible for every single soccer team in the MLS to win more games than they lose over the course of a season.</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Back to horses: in cultures<mask> horses were the primary means of transportation, people were buried with their favourite horse. And I would like to stress again, people do develop<mask> strong bond with their horse and the horse with the owner<mask> is the case with dogs. Still, eating horses is not frowned upon that much. [ENDQ] [NEWLINE] <mask> I said, I generally support fitting horses into my argument<mask> well,<mask><mask><mask> dogs are more firmly in it. [NEWLINE] [NEWLINE] [STARTQ] Considering the dog<mask> part of a family unit, above other animals, especially in cultures<mask> horses were of primary importance... I really doubt it,<mask> it's a doubt nothing else. [ENDQ] [NEWLINE] The historical evidence is pretty overwhelming that dogs were the primary animal that was part of the family unit.  There are exceptions<mask> over most time and area it was dogs. [NEWLINE] [NEWLINE] [STARTQ] And finally to go back to the original point, after this long discussion: the fact that dogs have a special place in the domestic animal kingdom is enough to criticize people for eating dog? [ENDQ] [NEWLINE] I don't know about criticizing people<mask> at least considering it worse than eating a wild or farm animal that was and always has been raised for consumption. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Back to horses: in cultures where horses were the primary means of transportation, people were buried with their favourite horse. And I would like to stress again, people do develop as strong bond with their horse and the horse with the owner as is the case with dogs. Still, eating horses is not frowned upon that much. [ENDQ] [NEWLINE] As I said, I generally support fitting horses into my argument as well, although I think dogs are more firmly in it. [NEWLINE] [NEWLINE] [STARTQ] Considering the dog as part of a family unit, above other animals, especially in cultures where horses were of primary importance... I really doubt it, but it's a doubt nothing else. [ENDQ] [NEWLINE] The historical evidence is pretty overwhelming that dogs were the primary animal that was part of the family unit.  There are exceptions but over most time and area it was dogs. [NEWLINE] [NEWLINE] [STARTQ] And finally to go back to the original point, after this long discussion: the fact that dogs have a special place in the domestic animal kingdom is enough to criticize people for eating dog? [ENDQ] [NEWLINE] I don't know about criticizing people but at least considering it worse than eating a wild or farm animal that was and always has been raised for consumption. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I'm not allowed to objectify or do other sexual acts in person without having consent from the other person, I don't see<mask> that wouldn't<mask> translate over to masturbation. [ENDQ] [NEWLINE] It translates over<mask> you can do whatever you want with your thoughts. And you can do whatever you want in a private place. [NEWLINE] [NEWLINE] [STARTQ] It's the same sexual thoughts, feelings, and intentions [ENDQ] [NEWLINE] No it's not. Again. You have complete and total control over your thoughts.<mask> you have a thought and tell nobody about it. For all intents and purposes, for the rest of the world you have never had that thought. It didn't exist. [NEWLINE] [NEWLINE] I could<mask> imagine you right now eating a plate of bacon. This could be a violation of your faith and religion.<mask> do I need your consent to imagine you eating bacon? [NEWLINE] [NEWLINE] I could imagine you shooting guns at a gun range. You may be stanchly morally opposed to guns. Do I need your consent to imagine you doing this? [NEWLINE] [NEWLINE] The answer. Simply is no.<mask> People can do whatever they want with their thoughts. And to even *attempt* to say otherwise, is laughable.</s>
Label encoding: <s> [STARTQ] If I'm not allowed to objectify or do other sexual acts in person without having consent from the other person, I don't see how that wouldn't also translate over to masturbation. [ENDQ] [NEWLINE] It translates over because you can do whatever you want with your thoughts. And you can do whatever you want in a private place. [NEWLINE] [NEWLINE] [STARTQ] It's the same sexual thoughts, feelings, and intentions [ENDQ] [NEWLINE] No it's not. Again. You have complete and total control over your thoughts. If you have a thought and tell nobody about it. For all intents and purposes, for the rest of the world you have never had that thought. It didn't exist. [NEWLINE] [NEWLINE] I could also imagine you right now eating a plate of bacon. This could be a violation of your faith and religion. But do I need your consent to imagine you eating bacon? [NEWLINE] [NEWLINE] I could imagine you shooting guns at a gun range. You may be stanchly morally opposed to guns. Do I need your consent to imagine you doing this? [NEWLINE] [NEWLINE] The answer. Simply is no. Because People can do whatever they want with their thoughts. And to even *attempt* to say otherwise, is laughable.</s>
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My analysis in that regard isn't that complex. <mask> you are born with male genitals, you are male. <mask> you believe that your male genitals were some sort of error, and that you were "supposed to" have female genitals instead, that is not a rational belief.  There is no such thing<mask> "supposed to" in this context.  You were born male, and you are male. <mask> you believe otherwise, your belief is delusional. [NEWLINE] [NEWLINE] The fact that someone born male may have a brain that resembles brains more commonly found in females does not make that person not male, either.  That belief, too, is delusional. [NEWLINE] [NEWLINE] These irrational beliefs are complicated by the bundle of social expectations we tend to put on males and females respectively.  It may be the case that a male, born with a brain more chemically aligned with a female, may be unprepared to deal with the social construct of "male" and will not fit well into that construct.  This, too, does not make him female.  It makes him an effeminate male, which is a perfectly fine thing to be, and does not require the removal of his genitals.</s>
Label encoding: <s>My analysis in that regard isn't that complex.  If you are born with male genitals, you are male.  If you believe that your male genitals were some sort of error, and that you were "supposed to" have female genitals instead, that is not a rational belief.  There is no such thing as "supposed to" in this context.  You were born male, and you are male.  If you believe otherwise, your belief is delusional. [NEWLINE] [NEWLINE] The fact that someone born male may have a brain that resembles brains more commonly found in females does not make that person not male, either.  That belief, too, is delusional. [NEWLINE] [NEWLINE] These irrational beliefs are complicated by the bundle of social expectations we tend to put on males and females respectively.  It may be the case that a male, born with a brain more chemically aligned with a female, may be unprepared to deal with the social construct of "male" and will not fit well into that construct.  This, too, does not make him female.  It makes him an effeminate male, which is a perfectly fine thing to be, and does not require the removal of his genitals.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No question that has a factual answer is really worth arguing about. :) [NEWLINE] [NEWLINE] Your comment about vacuousness is overly negative. I would say instead that the only meaning we have in life is<mask> we manage to create ourselves: more Sartre than Leibniz.<mask> the necessary corollary is that the only way to create meaning *after* we die is to create other beings who can create meanings of their own. Obviously the logic is still circular with no end state,<mask> at least the opportunity to do something meaningful stays alive,<mask> to speak, even<mask> we ourselves do not. [NEWLINE] [NEWLINE] Then the question becomes:<mask> not let others have the kids? It isn't necessary for everyone to squirt out rug rats, is it? No, of course it isn't.<mask> someone does have to do it, and it's kind of like the soldier volunteering for war<mask> that others don't have to. Falling on a sword is a noble act.<mask>,<mask> nobility isn't sufficiently motivating, there's the fact that your kids are more likely to leave a world more like you would want it, whereas other crappy people's crappy kids will leave a crappy world. </s>
Label encoding: <s>No question that has a factual answer is really worth arguing about. :) [NEWLINE] [NEWLINE] Your comment about vacuousness is overly negative. I would say instead that the only meaning we have in life is what we manage to create ourselves: more Sartre than Leibniz. But the necessary corollary is that the only way to create meaning *after* we die is to create other beings who can create meanings of their own. Obviously the logic is still circular with no end state, but at least the opportunity to do something meaningful stays alive, so to speak, even if we ourselves do not. [NEWLINE] [NEWLINE] Then the question becomes: why not let others have the kids? It isn't necessary for everyone to squirt out rug rats, is it? No, of course it isn't. But someone does have to do it, and it's kind of like the soldier volunteering for war so that others don't have to. Falling on a sword is a noble act. Also, if nobility isn't sufficiently motivating, there's the fact that your kids are more likely to leave a world more like you would want it, whereas other crappy people's crappy kids will leave a crappy world. </s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I didn't make the case that it's always the parents' fault, just that<mask> you isolate the teacher<mask> the source of failure, you're left needing to explain<mask> the other children turned out ok. [NEWLINE] [NEWLINE] People are created with all sorts of problems, and parents do all sorts of crap. <mask> the child is brain-damaged, obviously we can't blame the parenting for their low achievement. <mask> the child is abused, obviously we can.  In large groups, we can see the impact.  On a single case basis it's impossible to isolate nature from nurture,<mask> we can speculate based on studies on groups. [NEWLINE] [NEWLINE] I've heard a lot from both camps,<mask> I don't expect that parenting is<mask> important<mask> genetics - and that goes for ambition and intelligence. [NEWLINE] [NEWLINE] Point is, it's senseless to put the blame on the child, not<mask> it's not their fault,<mask><mask> that kid won't be the last kid to grow up in a home like that, or go to a school like that, and lots more will be even lazier or even dumber and we need systems that can prepare children for adulthood and the workforce.</s>
Label encoding: <s>I didn't make the case that it's always the parents' fault, just that if you isolate the teacher as the source of failure, you're left needing to explain why the other children turned out ok. [NEWLINE] [NEWLINE] People are created with all sorts of problems, and parents do all sorts of crap.  If the child is brain-damaged, obviously we can't blame the parenting for their low achievement.  If the child is abused, obviously we can.  In large groups, we can see the impact.  On a single case basis it's impossible to isolate nature from nurture, but we can speculate based on studies on groups. [NEWLINE] [NEWLINE] I've heard a lot from both camps, but I don't expect that parenting is as important as genetics - and that goes for ambition and intelligence. [NEWLINE] [NEWLINE] Point is, it's senseless to put the blame on the child, not because it's not their fault, but because that kid won't be the last kid to grow up in a home like that, or go to a school like that, and lots more will be even lazier or even dumber and we need systems that can prepare children for adulthood and the workforce.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ]...someone who does believe in god has an inherent sense of comfort in 'knowing' that death is not the end. [ENDQ] [NEWLINE] This is certainly true, and death is<mask> terrifying to me.<mask> really -- it is<mask> it is. We can freak the hell out about death, or we can realize that that makes the time we have all the more meaningful. The belief that we have X amount of time to do<mask> we can makes that time<mask> much *more valuable*, compared to<mask> we had X + infinity. Is X more meaningful on its own or<mask> you dilute it with infinity? [NEWLINE] [NEWLINE] I could play word games and use philosophical arguments to try to get you to buy some rationalization for<mask> atheism is more meaningful.<mask> is truly the case (<mask><mask><mask> ), is that **you are in control of making your own meaning**. People who are religious simply adopt a culture that teaches them<mask> meaning they should pursue. Being an atheist means you are allowed to find it on your own. Sure, that might seem horrible at first,<mask> once you start to build it for yourself, it will be more fulfilling than some meaning you borrowed from someone else.</s>
Label encoding: <s> [STARTQ]...someone who does believe in god has an inherent sense of comfort in 'knowing' that death is not the end. [ENDQ] [NEWLINE] This is certainly true, and death is also terrifying to me. But really -- it is what it is. We can freak the hell out about death, or we can realize that that makes the time we have all the more meaningful. The belief that we have X amount of time to do what we can makes that time so much *more valuable*, compared to if we had X + infinity. Is X more meaningful on its own or if you dilute it with infinity? [NEWLINE] [NEWLINE] I could play word games and use philosophical arguments to try to get you to buy some rationalization for why atheism is more meaningful. What is truly the case ( in my opinion ), is that **you are in control of making your own meaning**. People who are religious simply adopt a culture that teaches them what meaning they should pursue. Being an atheist means you are allowed to find it on your own. Sure, that might seem horrible at first, but once you start to build it for yourself, it will be more fulfilling than some meaning you borrowed from someone else.</s>
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>THIS IS FALSE. [NEWLINE] [NEWLINE] Women do not make less than men for doing the same job.<mask> you take two identical employees, one being male, the other being female, the female makes around 97% of<mask> the male makes. [NEWLINE] [NEWLINE] That 3% is accounted for in the data which shows that men are more likely to ask for a raise than women and are granted slightly more raises<mask><mask><mask>. [NEWLINE] [NEWLINE] <mask> a business could pay women more than men, men would not be hired for jobs. That doesn't happen<mask> the entire argument is based entirely around bad statistics. [NEWLINE] [NEWLINE] The 77% wage gap number is from taking ALL men and ALL women, adding their salaries together, and dividing by the number of individuals. [NEWLINE] [NEWLINE] This means that women are more likely than men to take up lower paying work and/or not work at all (stay at home mothers). [NEWLINE] [NEWLINE] The fact that the 77% number is still being thrown around is absolutely ludicrous to me,<mask> liberals are typically all about "science and logic"<mask> clearly having little-to-no understanding of either. [NEWLINE] [NEWLINE] I'm not conservative either<mask> DAMN this bad math will not die. </s>
Label encoding: <s>THIS IS FALSE. [NEWLINE] [NEWLINE] Women do not make less than men for doing the same job. If you take two identical employees, one being male, the other being female, the female makes around 97% of what the male makes. [NEWLINE] [NEWLINE] That 3% is accounted for in the data which shows that men are more likely to ask for a raise than women and are granted slightly more raises as a result. [NEWLINE] [NEWLINE] If a business could pay women more than men, men would not be hired for jobs. That doesn't happen because the entire argument is based entirely around bad statistics. [NEWLINE] [NEWLINE] The 77% wage gap number is from taking ALL men and ALL women, adding their salaries together, and dividing by the number of individuals. [NEWLINE] [NEWLINE] This means that women are more likely than men to take up lower paying work and/or not work at all (stay at home mothers). [NEWLINE] [NEWLINE] The fact that the 77% number is still being thrown around is absolutely ludicrous to me, because liberals are typically all about "science and logic" while clearly having little-to-no understanding of either. [NEWLINE] [NEWLINE] I'm not conservative either but DAMN this bad math will not die. </s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Often the prologue is a framing the device for the rest of the story, and carefully reading it changes<mask> the reader interprets the story. [NEWLINE] [NEWLINE] A great example of this is The Scarlet Letter.  The prologue sets up<mask> the manuscript was found,<mask><mask> reveals the mindset of the person who found it.  That mindset reveals<mask> their sympathies lie, and changes<mask> the story is read (did I avoid enough spoilers?). [NEWLINE] [NEWLINE] <mask> reading the prologue tips me off to the themes to watch out for, brings up questions I should be asking, and,<mask> I re-read, forces me to re-evaluate conclusions I drew from my first reading.  It's like a friend introducing me to someone I've already met; their experience with the person colors and re-colors my own, adding further depth. [NEWLINE] [NEWLINE] Of course, any bad writer can screw up any part of a good read. <mask> the decision to structure the book a certain way is in the hands of the creator, and I have to trust them.  And let's face it, there's only one true solution to having a great reading experience- read good books.</s>
Label encoding: <s>Often the prologue is a framing the device for the rest of the story, and carefully reading it changes how the reader interprets the story. [NEWLINE] [NEWLINE] A great example of this is The Scarlet Letter.  The prologue sets up how the manuscript was found, but also reveals the mindset of the person who found it.  That mindset reveals where their sympathies lie, and changes how the story is read (did I avoid enough spoilers?). [NEWLINE] [NEWLINE] So reading the prologue tips me off to the themes to watch out for, brings up questions I should be asking, and, if I re-read, forces me to re-evaluate conclusions I drew from my first reading.  It's like a friend introducing me to someone I've already met; their experience with the person colors and re-colors my own, adding further depth. [NEWLINE] [NEWLINE] Of course, any bad writer can screw up any part of a good read.  But the decision to structure the book a certain way is in the hands of the creator, and I have to trust them.  And let's face it, there's only one true solution to having a great reading experience- read good books.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you really like the name Bombay, or do you hate the name Mumbai? [NEWLINE] [NEWLINE] Does your distaste for name changes apply elsewhere?  Does it bother you that we now call North America "America" instead of<mask> it was originally called? <mask> you were alive back in colonial times, would you have been opposed to the English calling Mumbai Bombay?  Does the fact that Germany is called Allemania by most non-germanic european countries bother you? <mask> about that Germans call Germany Deutchland?  Who is wrong there, the Germans for calling it something different than everyone else, or everyone else for calling it something different than it is?  Does this distaste for renaming things extend beyond land?  Does it bother you<mask> they rename products in an attempt to re-brand them ( [URL] )? [NEWLINE] [NEWLINE] It seems clear that it bothers you,<mask><mask><mask> its unclear to pretty much everyone here<mask> it would bother you?  The only rational explanation I can come up with is that you personally prefer the name Bombay to Mumbai -<mask> then your CMV should really be "Bombay is a better name than Mumbai, CMV."</s>
Label encoding: <s>Do you really like the name Bombay, or do you hate the name Mumbai? [NEWLINE] [NEWLINE] Does your distaste for name changes apply elsewhere?  Does it bother you that we now call North America "America" instead of what it was originally called?  If you were alive back in colonial times, would you have been opposed to the English calling Mumbai Bombay?  Does the fact that Germany is called Allemania by most non-germanic european countries bother you?  How about that Germans call Germany Deutchland?  Who is wrong there, the Germans for calling it something different than everyone else, or everyone else for calling it something different than it is?  Does this distaste for renaming things extend beyond land?  Does it bother you when they rename products in an attempt to re-brand them ( [URL] )? [NEWLINE] [NEWLINE] It seems clear that it bothers you, but I think its unclear to pretty much everyone here why it would bother you?  The only rational explanation I can come up with is that you personally prefer the name Bombay to Mumbai - but then your CMV should really be "Bombay is a better name than Mumbai, CMV."</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Since the lives of cattle are unpleasant, on the surface it seems that slaughtering them at an earlier age would minimize suffering and would<mask> be less immoral. [NEWLINE] [NEWLINE] The reality of veal<mask>, is that it is essentially a by-product of the dairy industry.  To keep cows lactating they must constantly be producing calves.  Excess male calves are often used for veal rather than raised<mask> steers, or (very rarely) breeding bulls simply<mask> it takes less time and space away from their more profitable mothers.  A dairy cow generally lives about 5 years until it's milk production wanes,<mask> steers raised for meat are only forced to live 18 months or<mask>, and obviously veal calves get to die very quickly. [NEWLINE] [NEWLINE] <mask><mask> killing baby cows may alleviate more suffering than killing adult cows for food, in practice it is inextricably linked to keeping adult cows alive for food, causing them to suffer far more than they would had they been raised for meat rather than milk.  Add to the fact that many of these calves wouldn't have to be born or suffer at all<mask> not for their essential role in keeping dairy cows lactating.</s>
Label encoding: <s>Since the lives of cattle are unpleasant, on the surface it seems that slaughtering them at an earlier age would minimize suffering and would therefore be less immoral. [NEWLINE] [NEWLINE] The reality of veal however, is that it is essentially a by-product of the dairy industry.  To keep cows lactating they must constantly be producing calves.  Excess male calves are often used for veal rather than raised as steers, or (very rarely) breeding bulls simply because it takes less time and space away from their more profitable mothers.  A dairy cow generally lives about 5 years until it's milk production wanes, while steers raised for meat are only forced to live 18 months or so, and obviously veal calves get to die very quickly. [NEWLINE] [NEWLINE] So while killing baby cows may alleviate more suffering than killing adult cows for food, in practice it is inextricably linked to keeping adult cows alive for food, causing them to suffer far more than they would had they been raised for meat rather than milk.  Add to the fact that many of these calves wouldn't have to be born or suffer at all if not for their essential role in keeping dairy cows lactating.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Homogenous? Before the Labour Party gained real power in Parliament, faith in our national democracy among workers was<mask> dim that<mask> the bourgeoisie marched on the 17th of May - constitution day - flying Norwegian flags - the working people would march on May 1st, saying "strike the Christian cross off the flag and raise it pure and red". [NEWLINE] [NEWLINE] There were bloody confrontations between workers on strike and police forces, resulting in deaths. [NEWLINE] [NEWLINE] Yes, we were ethnically homogeneous,<mask> society was not homogeneous in any political sense. [NEWLINE] [NEWLINE] Yes, we are small,<mask> I don't see the relevance. Every time I say I don't, someone will fault me for not seeing the obvious relevance which they stubbornly refuse to explain to me,<mask> I don't see<mask> this wouldn't scale. [NEWLINE] [NEWLINE] Agree politically? No, we don't. We have a far greater selection of political parties in Europe than an American ever will, and a far wider selection. Norway has climate denialists in government right now. You can vote Communist or Libertarian, and you don't need more than a few thousand votes to actually get a seat in the national parliament.</s>
Label encoding: <s>Homogenous? Before the Labour Party gained real power in Parliament, faith in our national democracy among workers was so dim that although the bourgeoisie marched on the 17th of May - constitution day - flying Norwegian flags - the working people would march on May 1st, saying "strike the Christian cross off the flag and raise it pure and red". [NEWLINE] [NEWLINE] There were bloody confrontations between workers on strike and police forces, resulting in deaths. [NEWLINE] [NEWLINE] Yes, we were ethnically homogeneous, but society was not homogeneous in any political sense. [NEWLINE] [NEWLINE] Yes, we are small, but I don't see the relevance. Every time I say I don't, someone will fault me for not seeing the obvious relevance which they stubbornly refuse to explain to me, but I don't see why this wouldn't scale. [NEWLINE] [NEWLINE] Agree politically? No, we don't. We have a far greater selection of political parties in Europe than an American ever will, and a far wider selection. Norway has climate denialists in government right now. You can vote Communist or Libertarian, and you don't need more than a few thousand votes to actually get a seat in the national parliament.</s>
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Presidents schedules are roughed out weeks in advance. This press conference was likely stuck into his schedule, given it was entirely unexpected. Blocking four hours of golf for the president is probably not easy to manage.<mask> you said, he does deserve to exercise.<mask> I don't think it was exercise. He just got back from a vacation. Golf isn't like bike riding, a solitary pursuit. You can get business done playing golf. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Businessman Glenn Hutchins was<mask> pictured out on the course. [ENDQ] [NEWLINE] Meetings happen on golf courses all the time.<mask><mask>, I doubt there is ever a moment<mask> the president isn't being pressed for some kind of business or governmental request. This may be the case here. Maybe he needed this 'golf meeting' more than he needed to worry about optics at the moment. [NEWLINE] [NEWLINE] <mask>, I would think the president meets this situation more often than you would imagine. Meeting with grieving families of fallen soldiers then going to the Christmas tree lighting ceremony is probably<mask> his day goes normally and<mask> he shut himself away each time he encountered such things, he would never get any work done. [NEWLINE] </s>
Label encoding: <s>Presidents schedules are roughed out weeks in advance. This press conference was likely stuck into his schedule, given it was entirely unexpected. Blocking four hours of golf for the president is probably not easy to manage. As you said, he does deserve to exercise. But I don't think it was exercise. He just got back from a vacation. Golf isn't like bike riding, a solitary pursuit. You can get business done playing golf. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Businessman Glenn Hutchins was also pictured out on the course. [ENDQ] [NEWLINE] Meetings happen on golf courses all the time. In fact, I doubt there is ever a moment when the president isn't being pressed for some kind of business or governmental request. This may be the case here. Maybe he needed this 'golf meeting' more than he needed to worry about optics at the moment. [NEWLINE] [NEWLINE] Also, I would think the president meets this situation more often than you would imagine. Meeting with grieving families of fallen soldiers then going to the Christmas tree lighting ceremony is probably how his day goes normally and if he shut himself away each time he encountered such things, he would never get any work done. [NEWLINE] </s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But it doesn't even have to touch morality.<mask> our goal is to protect the members of society, we should try to discourage all negligence. [NEWLINE] [NEWLINE] Lets look at 4 people. Each drives drunk and crashes. Two made a onetime mistake, are deeply sorry and wont do it again,<mask> the other two drove drunk<mask> one stupid act in a series of reckless behavior. One out of each group hurt someone else in the crash,<mask> the other through sheer luck, didn't. [NEWLINE] [NEWLINE] In our current system, the ones that hurt someone else get a high sentence<mask> the other two a lower one. The outcome for society is one person rightfully locked away from the public, one locked up<mask> he/she isn't a threat, one rightfully set free and one set free<mask> being a threat. [NEWLINE] [NEWLINE] In OPs system, society would be protected from the two persons that are a threat,<mask> the two that aren't could contribute to the advancement of this society. [NEWLINE] [NEWLINE] Now, OPs system isn't practical/possible and has other flaws,<mask><mask> our only goal is to protect the members of society, it would be a better one.</s>
Label encoding: <s>But it doesn't even have to touch morality. If our goal is to protect the members of society, we should try to discourage all negligence. [NEWLINE] [NEWLINE] Lets look at 4 people. Each drives drunk and crashes. Two made a onetime mistake, are deeply sorry and wont do it again, while the other two drove drunk as one stupid act in a series of reckless behavior. One out of each group hurt someone else in the crash, while the other through sheer luck, didn't. [NEWLINE] [NEWLINE] In our current system, the ones that hurt someone else get a high sentence while the other two a lower one. The outcome for society is one person rightfully locked away from the public, one locked up although he/she isn't a threat, one rightfully set free and one set free despite being a threat. [NEWLINE] [NEWLINE] In OPs system, society would be protected from the two persons that are a threat, while the two that aren't could contribute to the advancement of this society. [NEWLINE] [NEWLINE] Now, OPs system isn't practical/possible and has other flaws, but if our only goal is to protect the members of society, it would be a better one.</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you too. [NEWLINE] [NEWLINE] [STARTQ] <mask> I meant by obtaining enjoyment from other aspects of the relationship is something like: I'll give you a back massage<mask> you go down on me. [ENDQ] [NEWLINE] Ahh, that was lost on me.   I assumed something more like "I'm an honest person that you trust and you have trust issues."  A massage shares a lot of commonality with sexual acts. Touching, intimacy, listening intently to please, learning about your partner's body, relaxation, etc.  I can see the mutual benefit and could see<mask> it might work. [NEWLINE] [NEWLINE] <mask> again, it's irrespective to an obligation component.<mask> your partner decides that she doesn't want massages or refuses to have sex in exchange for anything, I can't find a good reason for her to feel that she needs to. [NEWLINE] [NEWLINE] That said I do see her obligation in acknowledging your needs (and *this is a need*).  <mask><mask> I don't believe in her obligation to perform, I do believe in her obligation to allow your needs to be met. Her performance is her option. At least that's all I can logically reason.</s><pad>
Label encoding: <s>Thank you too. [NEWLINE] [NEWLINE] [STARTQ] What I meant by obtaining enjoyment from other aspects of the relationship is something like: I'll give you a back massage if you go down on me. [ENDQ] [NEWLINE] Ahh, that was lost on me.   I assumed something more like "I'm an honest person that you trust and you have trust issues."  A massage shares a lot of commonality with sexual acts. Touching, intimacy, listening intently to please, learning about your partner's body, relaxation, etc.  I can see the mutual benefit and could see how it might work. [NEWLINE] [NEWLINE] Though again, it's irrespective to an obligation component. If your partner decides that she doesn't want massages or refuses to have sex in exchange for anything, I can't find a good reason for her to feel that she needs to. [NEWLINE] [NEWLINE] That said I do see her obligation in acknowledging your needs (and *this is a need*).   So while I don't believe in her obligation to perform, I do believe in her obligation to allow your needs to be met. Her performance is her option. At least that's all I can logically reason.</s><pad>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is being made to feel special just something women are entitled to? [ENDQ] [NEWLINE] Not at all.<mask><mask>, I would say that<mask> a girl didn't make you feel special, then she is probably the wrong one. The way your CMV is phrased--you are looking to (appear to) be Mr Right, not looking for whoever is Ms Right. I mean,<mask> highly could you think of a woman that you PUA manipulated into your bed by treating them like crap? [NEWLINE] [NEWLINE] I'm actually going to stand by the advice to slow your roll. The guys I ended up dating for any length of time I hung out with informally and/or in groups on more than one occasion before I "dated" them. And often times the dates never felt date-y. More a natural extension of an ongoing relationship. Like in the middle of an awesome conversation, the guy said, "Hey, I have to go do X right now,<mask> do you want to grab dinner on Friday?"<mask> I am really enjoying the conversation, I am going to say yes.<mask> it's low pressure in terms of date-i-ness. </s>
Label encoding: <s> [STARTQ] Is being made to feel special just something women are entitled to? [ENDQ] [NEWLINE] Not at all. In fact, I would say that if a girl didn't make you feel special, then she is probably the wrong one. The way your CMV is phrased--you are looking to (appear to) be Mr Right, not looking for whoever is Ms Right. I mean, how highly could you think of a woman that you PUA manipulated into your bed by treating them like crap? [NEWLINE] [NEWLINE] I'm actually going to stand by the advice to slow your roll. The guys I ended up dating for any length of time I hung out with informally and/or in groups on more than one occasion before I "dated" them. And often times the dates never felt date-y. More a natural extension of an ongoing relationship. Like in the middle of an awesome conversation, the guy said, "Hey, I have to go do X right now, but do you want to grab dinner on Friday?" If I am really enjoying the conversation, I am going to say yes. But it's low pressure in terms of date-i-ness. </s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They didn't fall at free fall speeds. [ENDQ] [NEWLINE] **This is incorrect.** Bldg 7 did - In NIST's own words: [NEWLINE] [NEWLINE] [NEWLINE] "The analyses of the video (both the estimation of the instant the roofline began to descend and the calculated velocity and acceleration of a point on the roofline) revealed three distinct stages characterizing the 5.4 seconds of collapse: [NEWLINE] [NEWLINE] Stage 1 (0 to 1.75 seconds): acceleration less than that of gravity (i.e., slower than free fall). [NEWLINE] [NEWLINE] *Stage 2 (1.75 to 4.0 seconds): gravitational acceleration (free fall)* (my italics) [NEWLINE] [NEWLINE] Stage 3 (4.0 to 5.4 seconds): decreased acceleration, again less than that of gravity" [NEWLINE] [NEWLINE] [URL].cfm [NEWLINE] [NEWLINE] I have to say<mask> your entire rebuttal is nicely written and (seemingly) well sourced, the fact that you either lied or were simply ignorant of this fundamental fact throws the entire validity of everything you say into total question. I mean, c'mon NIST admits this quite plainly -<mask> could you miss this? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] They didn't fall at free fall speeds. [ENDQ] [NEWLINE] **This is incorrect.** Bldg 7 did - In NIST's own words: [NEWLINE] [NEWLINE] [NEWLINE] "The analyses of the video (both the estimation of the instant the roofline began to descend and the calculated velocity and acceleration of a point on the roofline) revealed three distinct stages characterizing the 5.4 seconds of collapse: [NEWLINE] [NEWLINE] Stage 1 (0 to 1.75 seconds): acceleration less than that of gravity (i.e., slower than free fall). [NEWLINE] [NEWLINE] *Stage 2 (1.75 to 4.0 seconds): gravitational acceleration (free fall)* (my italics) [NEWLINE] [NEWLINE] Stage 3 (4.0 to 5.4 seconds): decreased acceleration, again less than that of gravity" [NEWLINE] [NEWLINE] [URL].cfm [NEWLINE] [NEWLINE] I have to say while your entire rebuttal is nicely written and (seemingly) well sourced, the fact that you either lied or were simply ignorant of this fundamental fact throws the entire validity of everything you say into total question. I mean, c'mon NIST admits this quite plainly - how could you miss this? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If one calls academics into the debate, such<mask> this statement: [NEWLINE] [STARTQ] Even the industry own journals admit that they are having a hard time keeping up with evolving germs, for example. [ENDQ] [NEWLINE] then one is obliged to provide said sources.<mask> one just says "<mask><mask> this, or I feel this way" then sources are generally unnecessary (<mask> these styles of rhetorical argument are generally disfavoured due to their level of bias [your opinion does not sit on equal footing with my facts]),<mask><mask> you say "I've read journal articles that say this", then you need to find those sources, or otherwise be dismissed<mask> committing the fallacy of "arguing from authority". [NEWLINE] [NEWLINE] <mask> I'm trying to say with all the academic language here is that saying "do a google search" in response to someone's skepticism of your information is not an acceptable response,<mask> the burden relies on YOU, the citer of information, not the questioner to validate your claims (using the royal "you" here). [NEWLINE] [NEWLINE] I hope that helps you understand<mask> this comes up all the time, and<mask> people get upset at this. [NEWLINE] [NEWLINE] cheers!</s>
Label encoding: <s>If one calls academics into the debate, such as this statement: [NEWLINE] [STARTQ] Even the industry own journals admit that they are having a hard time keeping up with evolving germs, for example. [ENDQ] [NEWLINE] then one is obliged to provide said sources. If one just says " I think this, or I feel this way" then sources are generally unnecessary ( but these styles of rhetorical argument are generally disfavoured due to their level of bias [your opinion does not sit on equal footing with my facts]), but if you say "I've read journal articles that say this", then you need to find those sources, or otherwise be dismissed as committing the fallacy of "arguing from authority". [NEWLINE] [NEWLINE] What I'm trying to say with all the academic language here is that saying "do a google search" in response to someone's skepticism of your information is not an acceptable response, because the burden relies on YOU, the citer of information, not the questioner to validate your claims (using the royal "you" here). [NEWLINE] [NEWLINE] I hope that helps you understand why this comes up all the time, and why people get upset at this. [NEWLINE] [NEWLINE] cheers!</s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes you can,<mask> that doesn't do anything to convince me otherwise. [NEWLINE] [NEWLINE] It isn't a mystery that public schools in the US are in poor shape, especially cities. Typing "American Public Schools" into Google doesn't give you results about<mask> great school are doing,<mask> more on<mask> terrible they are. I don't hear from friends in public schools raving about<mask> much freedom they have to work,<mask> quite the contrary. I hear more about people burning out from teaching for shit pay than I do about people lining up to transform the minds in the most downtrodden of schools. [NEWLINE] [NEWLINE] I'm well aware that I am subject to confirmation bias and that I have no data on hand to back up my claims,<mask> at risk of sounding snarky, being told that there are counterviews or that I'm generalizing doesn't do anything to change my view. I've heard teachers are not allowed to touch students even<mask> fights break out<mask> a policy in multiple large districts (DC, Chicago, etc). I haven't heard about a teacher who said "fuck the system" and physically restrained a student from doing something they shouldn't.</s>
Label encoding: <s>Yes you can, but that doesn't do anything to convince me otherwise. [NEWLINE] [NEWLINE] It isn't a mystery that public schools in the US are in poor shape, especially cities. Typing "American Public Schools" into Google doesn't give you results about how great school are doing, but more on how terrible they are. I don't hear from friends in public schools raving about how much freedom they have to work, but quite the contrary. I hear more about people burning out from teaching for shit pay than I do about people lining up to transform the minds in the most downtrodden of schools. [NEWLINE] [NEWLINE] I'm well aware that I am subject to confirmation bias and that I have no data on hand to back up my claims, but at risk of sounding snarky, being told that there are counterviews or that I'm generalizing doesn't do anything to change my view. I've heard teachers are not allowed to touch students even when fights break out as a policy in multiple large districts (DC, Chicago, etc). I haven't heard about a teacher who said "fuck the system" and physically restrained a student from doing something they shouldn't.</s>
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From my experiences in a country, that's not at all the case. There's no racism. Yes, the group isn't find of outsiders, which is a strong point, and a good one. [NEWLINE] [NEWLINE] See, the way I see it, cityfolk don't give a damn about the environment<mask> there are people who's job is cleaning their mess,<mask> they don't know about the damage they're causing. [NEWLINE] [NEWLINE] Whereas in the countryside,<mask> you don't pick it up, no one will. And you get criticized for making the community look bad. In the city no one cares. [NEWLINE] [NEWLINE] <mask> I said values, I meant stuff like getting along with neighbors, caring for people in need and discipline of kids. From<mask> I've seen in the city, there's none of that, kids are kept on literal leashes and no one cares, you're anonymous in the city. [NEWLINE] [NEWLINE] My biggest fear in living in the city is being separated from a complete stranger by nothing more than a dry wall. You hear them, you get used to them and then you start hating them for everything they do. You develop pret peeves.. </s>
Label encoding: <s>From my experiences in a country, that's not at all the case. There's no racism. Yes, the group isn't find of outsiders, which is a strong point, and a good one. [NEWLINE] [NEWLINE] See, the way I see it, cityfolk don't give a damn about the environment because there are people who's job is cleaning their mess, so they don't know about the damage they're causing. [NEWLINE] [NEWLINE] Whereas in the countryside, if you don't pick it up, no one will. And you get criticized for making the community look bad. In the city no one cares. [NEWLINE] [NEWLINE] When I said values, I meant stuff like getting along with neighbors, caring for people in need and discipline of kids. From what I've seen in the city, there's none of that, kids are kept on literal leashes and no one cares, you're anonymous in the city. [NEWLINE] [NEWLINE] My biggest fear in living in the city is being separated from a complete stranger by nothing more than a dry wall. You hear them, you get used to them and then you start hating them for everything they do. You develop pret peeves.. </s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Whoever is in charge of the US nuclear arsenal obtained this position<mask> they will serve US interests first and foremost. He or she should counterattack<mask> at least the act would show that he or she has acted with the interests of the USA in mind.<mask> confronted with weapons of mass destruction, the populus is more concerned that they will even survive the attack. They might even go<mask> far<mask> to rise against the government<mask> it maintained its policy of inaction. The people would not care that the environment is destroyed<mask> they survive; they would first concern themselves about an immediate success (survival) before attending to more long term successes (environmental maintenance). Humans often have a mentality of saving themselves before saving others,<mask> most likely the people would care about attempting for their safety before attending to another's, especially<mask> they do not know the other person.<mask>,<mask> there was nothing we could do to remedy the situation, the most human thing anyone would do is to enact one final act of revenge. Is it not human to channel an entire country's emotion of pain and despair onto an external force<mask> that external force is<mask> caused their emotions in the first place?</s>
Label encoding: <s>Whoever is in charge of the US nuclear arsenal obtained this position because they will serve US interests first and foremost. He or she should counterattack because at least the act would show that he or she has acted with the interests of the USA in mind. When confronted with weapons of mass destruction, the populus is more concerned that they will even survive the attack. They might even go so far as to rise against the government if it maintained its policy of inaction. The people would not care that the environment is destroyed if they survive; they would first concern themselves about an immediate success (survival) before attending to more long term successes (environmental maintenance). Humans often have a mentality of saving themselves before saving others, so most likely the people would care about attempting for their safety before attending to another's, especially if they do not know the other person. Besides, if there was nothing we could do to remedy the situation, the most human thing anyone would do is to enact one final act of revenge. Is it not human to channel an entire country's emotion of pain and despair onto an external force if that external force is what caused their emotions in the first place?</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>your best bet is go to a apple store and play with the phone [NEWLINE] [NEWLINE] yea the home button just take you to the home screen, double click for multitab and triple for accessibility function<mask><mask> [NEWLINE] [NEWLINE] the back function and more options function are designed within the app itself [NEWLINE] [NEWLINE] and slide from bottom of the screen bring up quick access to general stuff like wifi, volume control, etc. and slide from top for notification [NEWLINE] [NEWLINE] one standout example I can think of is holding the home button on the home screen until the app icons start shaking then you can easily delete or move it around and drag it over other app to form a folder. all of that in 1 go. you can't even do the same on android<mask> deleting anything on the home screen is a shortcut and not the actual app<mask> you gonna go to setting [STARTQ] blah blah to get rid of it. [ENDQ] [NEWLINE] I used both platform and apple is<mask> much easier to do<mask> you want to do without being in the way and the app store is just awesome. shit tons of apps. [NEWLINE] [NEWLINE] android just happen to be cheaper and allow to be tweaked a lot like a PC does [NEWLINE] </s><pad>
Label encoding: <s>your best bet is go to a apple store and play with the phone [NEWLINE] [NEWLINE] yea the home button just take you to the home screen, double click for multitab and triple for accessibility function i think [NEWLINE] [NEWLINE] the back function and more options function are designed within the app itself [NEWLINE] [NEWLINE] and slide from bottom of the screen bring up quick access to general stuff like wifi, volume control, etc. and slide from top for notification [NEWLINE] [NEWLINE] one standout example I can think of is holding the home button on the home screen until the app icons start shaking then you can easily delete or move it around and drag it over other app to form a folder. all of that in 1 go. you can't even do the same on android because deleting anything on the home screen is a shortcut and not the actual app so you gonna go to setting [STARTQ] blah blah to get rid of it. [ENDQ] [NEWLINE] I used both platform and apple is so much easier to do what you want to do without being in the way and the app store is just awesome. shit tons of apps. [NEWLINE] [NEWLINE] android just happen to be cheaper and allow to be tweaked a lot like a PC does [NEWLINE] </s><pad>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To me a lot of this boils down to<mask> we are justified in saying we ‘know’ something. We generally use the term more loosely outside of the atheism/theism debate,<mask> I can understand<mask> people sometimes use it this way. [NEWLINE] [NEWLINE] <mask> I am thinking about is<mask> we talk about scientific fact or laws of nature most people generally say that we *know* the laws of gravity exist.<mask> technically we can’t be anymore sure of this than we can of whether or not God exists. We might after all be a brain in a vat, and outside the vat the laws of physics might be fundamentally different from<mask> we perceive them, now this is not very likely,<mask> never the less it is locally possible. Does this mean we should add qualifiers every time we make a statement of fact about the physical world? It seems a bit gratuitous. [NEWLINE] [NEWLINE] I am not a hardline atheist myself, and<mask><mask> it is a valuable distinction to make,<mask> I can forgive people who don’t find joy in technical details to just use the shorthand and say “*we know*” [NEWLINE] </s>
Label encoding: <s>To me a lot of this boils down to when we are justified in saying we ‘know’ something. We generally use the term more loosely outside of the atheism/theism debate, so I can understand why people sometimes use it this way. [NEWLINE] [NEWLINE] What I am thinking about is when we talk about scientific fact or laws of nature most people generally say that we *know* the laws of gravity exist. But technically we can’t be anymore sure of this than we can of whether or not God exists. We might after all be a brain in a vat, and outside the vat the laws of physics might be fundamentally different from how we perceive them, now this is not very likely, but never the less it is locally possible. Does this mean we should add qualifiers every time we make a statement of fact about the physical world? It seems a bit gratuitous. [NEWLINE] [NEWLINE] I am not a hardline atheist myself, and I think it is a valuable distinction to make, but I can forgive people who don’t find joy in technical details to just use the shorthand and say “*we know*” [NEWLINE] </s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1 its advertisement,  a competition they forget after leaving, a ribbon or other reminds them, this means people are more likely to remember. [NEWLINE] [NEWLINE] 2 no one wants to do a lot of effort for nothing, even<mask> it wasn't much by adult standards, psychologically its better to give them a bauble to keep them motivated [NEWLINE] [NEWLINE] 3 mass ordering is cheaper then individual orders, 20 participation trophies will likely set you back less then a single large trophy [NEWLINE] [NEWLINE] 4 only a single person can win, that means that<mask> your not first you might<mask> well quit, this fosters quitters and cheaters [NEWLINE] [NEWLINE] 5 trophys mean nothing. i have half a dozen medals and other assorted items, doesn't matter to me<mask> i placed first or last, its the memories behind the competition,  for example for a race we had a fight with a couple other kids, now normally i'm pretty fast,<mask> one of the guys kicked me in the legs<mask> i was unable to place first,<mask> i did give the guy an excellent punch to the stomach,<mask><mask> i look at that ribbon i remember that fight and the look on his face.</s>
Label encoding: <s>1 its advertisement,  a competition they forget after leaving, a ribbon or other reminds them, this means people are more likely to remember. [NEWLINE] [NEWLINE] 2 no one wants to do a lot of effort for nothing, even if it wasn't much by adult standards, psychologically its better to give them a bauble to keep them motivated [NEWLINE] [NEWLINE] 3 mass ordering is cheaper then individual orders, 20 participation trophies will likely set you back less then a single large trophy [NEWLINE] [NEWLINE] 4 only a single person can win, that means that if your not first you might as well quit, this fosters quitters and cheaters [NEWLINE] [NEWLINE] 5 trophys mean nothing. i have half a dozen medals and other assorted items, doesn't matter to me if i placed first or last, its the memories behind the competition,  for example for a race we had a fight with a couple other kids, now normally i'm pretty fast, but one of the guys kicked me in the legs so i was unable to place first, but i did give the guy an excellent punch to the stomach, so when i look at that ribbon i remember that fight and the look on his face.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh,<mask><mask> with most of<mask> you've said.  Killing animals *and* babies is usually morally wrong,<mask> I don't think it's *inherently* wrong from the perspective of preference utilitarianism. [NEWLINE] [NEWLINE] <mask><mask> that our judgements are markedly skewed in favour of our species<mask> even<mask> a child is not related to you, they would usually grow up to be a useful member of a tribe, and that benefits you and your ability to proliferate your genes. <mask> the visceral disgust and infanticide.  Might just be an evolutionary "just<mask> " story<mask>, who knows? [NEWLINE] [NEWLINE] Still, I can't pretend to be aloof from the conclusions I endorse.  Some of them are pretty sickening,<mask><mask><mask> it's just that reason does not always comport with out intuitions, moral or mathematical.  The monty hall problem always got on my nerves until I finally found an intuitive way of getting it, and even mathematicians wrote angry letters in journals saying the solution must be wrong.  Not that I mean to say that moral matters can be resolved with the certainty of mathematical matters</s>
Label encoding: <s>Oh, I agree with most of what you've said.  Killing animals *and* babies is usually morally wrong, but I don't think it's *inherently* wrong from the perspective of preference utilitarianism. [NEWLINE] [NEWLINE] I think that our judgements are markedly skewed in favour of our species because even if a child is not related to you, they would usually grow up to be a useful member of a tribe, and that benefits you and your ability to proliferate your genes.  Hence the visceral disgust and infanticide.  Might just be an evolutionary "just so " story though, who knows? [NEWLINE] [NEWLINE] Still, I can't pretend to be aloof from the conclusions I endorse.  Some of them are pretty sickening, but I think it's just that reason does not always comport with out intuitions, moral or mathematical.  The monty hall problem always got on my nerves until I finally found an intuitive way of getting it, and even mathematicians wrote angry letters in journals saying the solution must be wrong.  Not that I mean to say that moral matters can be resolved with the certainty of mathematical matters</s>
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm relatively sure that most children above 12 years old in the developed world would know<mask> organ donation is [ENDQ] [NEWLINE] **Stop assuming.** Even<mask> this statement is correct (which it may or may not be), it doesn't mean they know the system is opt-in.<mask>, "most" is not "all." "Most" is 51%. **Even<mask> most do know about the program, should those that do not know about it die<mask> of their lack of knowledge about the opt-in system?** [NEWLINE] [NEWLINE] [STARTQ] <mask> the problem arised that people didn't know that they had to opt-in, that could be easily dealt with. [ENDQ] [NEWLINE] Really?<mask>? By asking them<mask> they're on the operating table? *Everyone* would just say "I didn't know about it," whether they did or not.<mask> you want to go with this you'll need to provide specific and foolproof way to make this determination. And no, you're not putting grandma on the polygraph. [NEWLINE] [NEWLINE] Let's just forget the whole religion point<mask> we are both making some assumptions and we're at an impasse.</s>
Label encoding: <s> [STARTQ] I'm relatively sure that most children above 12 years old in the developed world would know what organ donation is [ENDQ] [NEWLINE] **Stop assuming.** Even if this statement is correct (which it may or may not be), it doesn't mean they know the system is opt-in. Also, "most" is not "all." "Most" is 51%. **Even if most do know about the program, should those that do not know about it die because of their lack of knowledge about the opt-in system?** [NEWLINE] [NEWLINE] [STARTQ] If the problem arised that people didn't know that they had to opt-in, that could be easily dealt with. [ENDQ] [NEWLINE] Really? How? By asking them as they're on the operating table? *Everyone* would just say "I didn't know about it," whether they did or not. If you want to go with this you'll need to provide specific and foolproof way to make this determination. And no, you're not putting grandma on the polygraph. [NEWLINE] [NEWLINE] Let's just forget the whole religion point because we are both making some assumptions and we're at an impasse.</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not going to argue about standard of living, you're probably right about that. [NEWLINE] [NEWLINE] <mask>, regarding freedoms. The US has much more lax laws on defamation, which some regard<mask> being more 'free'. Personally,<mask><mask> the UK (and my country, Australia) goes too far on restricting defamation,<mask> the US doesn't go far enough. There's<mask> their dogmatic support of the right to bear arms, which again is a freedom of sorts. And yes I'm playing devil's advocate here,<mask> I don't believe that these freedoms are necessarily an improvement. [NEWLINE] [NEWLINE] One definite advantage is their fair use laws, which are MUCH better than<mask> you have in the UK with fair dealing. [NEWLINE] [NEWLINE] The recent introduction of porn blocking in the UK is a terrible move backward. Yes you can opt out of it,<mask> something like this should most definitely be opt in. [NEWLINE] [NEWLINE] You<mask> gain, by moving to the US *from the UK*, a *slight* improvement in socio-economic mobility. That doesn't say much,<mask>,<mask> both the US and UK are among the worst in the developed world.</s>
Label encoding: <s>I'm not going to argue about standard of living, you're probably right about that. [NEWLINE] [NEWLINE] However, regarding freedoms. The US has much more lax laws on defamation, which some regard as being more 'free'. Personally, I think the UK (and my country, Australia) goes too far on restricting defamation, but the US doesn't go far enough. There's also their dogmatic support of the right to bear arms, which again is a freedom of sorts. And yes I'm playing devil's advocate here, because I don't believe that these freedoms are necessarily an improvement. [NEWLINE] [NEWLINE] One definite advantage is their fair use laws, which are MUCH better than what you have in the UK with fair dealing. [NEWLINE] [NEWLINE] The recent introduction of porn blocking in the UK is a terrible move backward. Yes you can opt out of it, but something like this should most definitely be opt in. [NEWLINE] [NEWLINE] You also gain, by moving to the US *from the UK*, a *slight* improvement in socio-economic mobility. That doesn't say much, however, as both the US and UK are among the worst in the developed world.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0622, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The difference is that any of the things I mention could be done without the need for criminal prosecution, which I would assume is going to be massively difficult to prosecute in any case given under the current archaic legislation (which I believe is medieval) only one person has actually been convicted. Plus, assuming you were sentencing me,<mask> long do you think would it take in prison for me to become a non-threat? Chances are that the only option is to throw me into prison for the rest of my natural life, which causes its own problems<mask><mask> under current regulations it is rare that a judge can give such a sentence. [NEWLINE] [NEWLINE] I<mask> wasn't talking about being given citizenship too. Rather I was talking about<mask> any foreign fighter would be considered<mask> being no different to a native soldier in terms of their role within that force and<mask> subject to the same protection of that army. [NEWLINE] [NEWLINE] <mask> to your fantasy scenario, credibility of Putin in some circles, and particularly in the West, is very low (<mask> I personally like Putin). Any assertion he makes will face an uphill battle to gain any traction outside of pro-Russian groups. [NEWLINE] </s>
Label encoding: <s>The difference is that any of the things I mention could be done without the need for criminal prosecution, which I would assume is going to be massively difficult to prosecute in any case given under the current archaic legislation (which I believe is medieval) only one person has actually been convicted. Plus, assuming you were sentencing me, how long do you think would it take in prison for me to become a non-threat? Chances are that the only option is to throw me into prison for the rest of my natural life, which causes its own problems given that under current regulations it is rare that a judge can give such a sentence. [NEWLINE] [NEWLINE] I also wasn't talking about being given citizenship too. Rather I was talking about how any foreign fighter would be considered as being no different to a native soldier in terms of their role within that force and therefore subject to the same protection of that army. [NEWLINE] [NEWLINE] As to your fantasy scenario, credibility of Putin in some circles, and particularly in the West, is very low ( although I personally like Putin). Any assertion he makes will face an uphill battle to gain any traction outside of pro-Russian groups. [NEWLINE] </s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0616, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Every single thing that we do has risk associated with it.  With respect to your examples, getting into a head on collision with a drunk driver is bad, this is true,<mask> you accepted the risk of other drivers<mask> you chose to drive,<mask> you engaged in a risky proposition (driving) and were unlucky enough to encounter a bad situation (drunk driver).  Cancer is thought to be due to a combination of genetics and environmental factors,<mask> even rare forms of cancer have to do with you participating in modern society with its associated carcinogens in the air we breathe, the foods we eat and the liquids we drink.  Granted the chance of any of these things is lower than<mask> you engage in things such<mask> diving, sky diving etc,<mask> its more of an issue of the likelihood of something happening, not that driving, or living in a city is completely devoid of risk.  I am not egotistical enough to say<mask> is an appropriate level of risk for people to engage in or not engage in,<mask><mask> not mourn the loss of life and leave it at that without pointing fingers and saying its your own fault?</s>
Label encoding: <s>Every single thing that we do has risk associated with it.  With respect to your examples, getting into a head on collision with a drunk driver is bad, this is true, but you accepted the risk of other drivers when you chose to drive, therefore you engaged in a risky proposition (driving) and were unlucky enough to encounter a bad situation (drunk driver).  Cancer is thought to be due to a combination of genetics and environmental factors, so even rare forms of cancer have to do with you participating in modern society with its associated carcinogens in the air we breathe, the foods we eat and the liquids we drink.  Granted the chance of any of these things is lower than when you engage in things such as diving, sky diving etc, but its more of an issue of the likelihood of something happening, not that driving, or living in a city is completely devoid of risk.  I am not egotistical enough to say what is an appropriate level of risk for people to engage in or not engage in, so why not mourn the loss of life and leave it at that without pointing fingers and saying its your own fault?</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I have a hard time understanding is<mask> women choose to *eat meat* knowing that they are ending a life. It is very hard for me to wrap my head around the fact that women can be<mask> selfish to take a life,<mask> there are *<mask> many more dietary options.* [ENDQ] [NEWLINE] It's very difficult to have your mind changed<mask> you don't make any effort to understand their reasoning.  The people you disagree with are compassionate, decent, intelligent people.  Even<mask> you disagree with them, they've formed their opinions by reason. [NEWLINE] [NEWLINE] [NEWLINE] People who approve of abortions, and those who don't disagree on the definition of "human being". [NEWLINE] [NEWLINE] They do not feel that a collection of human cells living parasitically inside a woman's body is due the same rights<mask> fully developed people. [NEWLINE] [NEWLINE] Devout vegans feel that all animal life is worthy. [NEWLINE] [NEWLINE] _________ [NEWLINE] [NEWLINE] The question I have for you is - *At<mask> point do you, personally, believe that a fertilized egg becomes a person?*  The moment of fertilization or at some point after?  </s>
Label encoding: <s> [STARTQ] What I have a hard time understanding is why women choose to *eat meat* knowing that they are ending a life. It is very hard for me to wrap my head around the fact that women can be so selfish to take a life, when there are * so many more dietary options.* [ENDQ] [NEWLINE] It's very difficult to have your mind changed when you don't make any effort to understand their reasoning.  The people you disagree with are compassionate, decent, intelligent people.  Even if you disagree with them, they've formed their opinions by reason. [NEWLINE] [NEWLINE] [NEWLINE] People who approve of abortions, and those who don't disagree on the definition of "human being". [NEWLINE] [NEWLINE] They do not feel that a collection of human cells living parasitically inside a woman's body is due the same rights as fully developed people. [NEWLINE] [NEWLINE] Devout vegans feel that all animal life is worthy. [NEWLINE] [NEWLINE] _________ [NEWLINE] [NEWLINE] The question I have for you is - *At what point do you, personally, believe that a fertilized egg becomes a person?*  The moment of fertilization or at some point after?  </s>
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I fail to see any negative consequences in any way<mask> a man gets a woman pregnant. Can you find any? I mean<mask> stops a guy from getting a girl pregnant on a Monday, walking away from that situation, and being in a pick-up bar again by Tuesday?   I don't see anything that stops that behavior. [NEWLINE] [NEWLINE] This is an important question.<mask> a taxpayer, I would just be paying for these children who have a lack of resources. I would be paying<mask> that men could have sex with<mask> many women<mask> they wanted<mask> there is no consequences for this behavior. You don't think that men would take advantage of this system. [NEWLINE] [NEWLINE] And<mask>,<mask> you said, the system is unfair for women, wouldn't you just be taking a system that you think is unfair for men and converting that into a system that unfair for women?  Per your view, one party is free<mask> free can be and the other party is forced to make a choice.  That doesn't seem at all fair to me. Is your perspective just creating a new system that would be unfair for one demographic?</s>
Label encoding: <s>I fail to see any negative consequences in any way if a man gets a woman pregnant. Can you find any? I mean what stops a guy from getting a girl pregnant on a Monday, walking away from that situation, and being in a pick-up bar again by Tuesday?   I don't see anything that stops that behavior. [NEWLINE] [NEWLINE] This is an important question. As a taxpayer, I would just be paying for these children who have a lack of resources. I would be paying so that men could have sex with as many women as they wanted because there is no consequences for this behavior. You don't think that men would take advantage of this system. [NEWLINE] [NEWLINE] And if, as you said, the system is unfair for women, wouldn't you just be taking a system that you think is unfair for men and converting that into a system that unfair for women?  Per your view, one party is free as free can be and the other party is forced to make a choice.  That doesn't seem at all fair to me. Is your perspective just creating a new system that would be unfair for one demographic?</s>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well,<mask><mask> all these statistics make a pretty convincing case<mask> to<mask> prostitution *should* be legalized,<mask> right now almost all those prostitutes currently working in San Francisco have no legal recourse<mask> they're raped or abused or forced into the profession. [NEWLINE] [NEWLINE] "Your Honor, I approached him to sell sex<mask> he raped me instead and didn't pay me afterwards. I'm perfectly willing to accept jail time, multiple criminal charges, and significant fines in order to take him down with me." [NEWLINE] [NEWLINE] -Said very few prostitutes working in San Francisco. [NEWLINE] [NEWLINE] No, seriously,<mask><mask> that the horrifying levels of abuse, physical battery, and rape are a problem - a problem that could be better solved by legalizing prostitution and giving them legal protections and protocols to deal with these things, rather than their current illegal status<mask> they have no such answers towards these problems. [NEWLINE] [NEWLINE] <mask><mask><mask> I can tell, keeping prostitution illegal seems to make the problems in America worse, not better. Check your own statistics - 40% sexual violence in the Netherlands sounds a *lot* better than 85% rape in Minneapolis, doesn't it?</s>
Label encoding: <s>Well, I think all these statistics make a pretty convincing case as to why prostitution *should* be legalized, since right now almost all those prostitutes currently working in San Francisco have no legal recourse if they're raped or abused or forced into the profession. [NEWLINE] [NEWLINE] "Your Honor, I approached him to sell sex but he raped me instead and didn't pay me afterwards. I'm perfectly willing to accept jail time, multiple criminal charges, and significant fines in order to take him down with me." [NEWLINE] [NEWLINE] -Said very few prostitutes working in San Francisco. [NEWLINE] [NEWLINE] No, seriously, I agree that the horrifying levels of abuse, physical battery, and rape are a problem - a problem that could be better solved by legalizing prostitution and giving them legal protections and protocols to deal with these things, rather than their current illegal status where they have no such answers towards these problems. [NEWLINE] [NEWLINE] As far as I can tell, keeping prostitution illegal seems to make the problems in America worse, not better. Check your own statistics - 40% sexual violence in the Netherlands sounds a *lot* better than 85% rape in Minneapolis, doesn't it?</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] THE OP DID NOT REFER TO A WEALTH TAX. HE SAID A MASS REDISTRIBUTION OF ALL WEALTH AKIN TO A RESET BUTTON. IT'S IN THE TITLE OF THE POST. THOSE ARE TWO DIFFERENT THINGS. [ENDQ] [NEWLINE] [NEWLINE] I'll just quote the OP: [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> might a one-time mass-redistribution of wealth be enacted? (Gradually? All at once? Through taxes or direct confiscation?) [ENDQ] [NEWLINE] <mask>, you write: [NEWLINE] [NEWLINE] [STARTQ] many (<mask> not all) of those assets would have to be taken away entirely, and the only way to redistribute that wealth would be to monetize it and pump that cash into the basic income. [ENDQ] [NEWLINE] No, that's just silly. <mask> the government were to confiscate assets in the form of stocks, then it could just hold onto the assets, and distribute the dividends<mask> the basic income. [NEWLINE] [NEWLINE] This is known<mask> a sovereign wealth fund. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] It could<mask> simply distribute them<mask> stocks, of course.</s>
Label encoding: <s> [STARTQ] THE OP DID NOT REFER TO A WEALTH TAX. HE SAID A MASS REDISTRIBUTION OF ALL WEALTH AKIN TO A RESET BUTTON. IT'S IN THE TITLE OF THE POST. THOSE ARE TWO DIFFERENT THINGS. [ENDQ] [NEWLINE] [NEWLINE] I'll just quote the OP: [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] How might a one-time mass-redistribution of wealth be enacted? (Gradually? All at once? Through taxes or direct confiscation?) [ENDQ] [NEWLINE] Also, you write: [NEWLINE] [NEWLINE] [STARTQ] many ( if not all) of those assets would have to be taken away entirely, and the only way to redistribute that wealth would be to monetize it and pump that cash into the basic income. [ENDQ] [NEWLINE] No, that's just silly.  If the government were to confiscate assets in the form of stocks, then it could just hold onto the assets, and distribute the dividends as the basic income. [NEWLINE] [NEWLINE] This is known as a sovereign wealth fund. [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] It could also simply distribute them as stocks, of course.</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It would have serious downsides. [NEWLINE] [NEWLINE] The biggest is that in a robbery, you want to instill fear<mask> quickly<mask> possible, do get your victim to give up their money,<mask> you can get away<mask> fast ad possible. [NEWLINE] [NEWLINE] Your purpose is to get the cash, not get caught, and not get killed or injured in the process. [NEWLINE] [NEWLINE] <mask> you go after someone with a super soaker, they're not going to be scared until they think that they're sprayed with a small amount of fluid.  You can't just set them on fire to begin with,<mask> then you don't get their money,<mask> its on fire with them. [NEWLINE] [NEWLINE] Guns people understand.  You put a gun in their face, they worry about being shot, give you the money, and you're gone. [NEWLINE] [NEWLINE] <mask> you sprayed me with lighter fluid, then asked for cash under threat of fire, I'm going to tackle you, and we both burn to death. [NEWLINE] [NEWLINE] Guns just work really well. Its possible to use a super soaker,<mask> guns are just more efficient.</s>
Label encoding: <s>It would have serious downsides. [NEWLINE] [NEWLINE] The biggest is that in a robbery, you want to instill fear as quickly as possible, do get your victim to give up their money, so you can get away as fast ad possible. [NEWLINE] [NEWLINE] Your purpose is to get the cash, not get caught, and not get killed or injured in the process. [NEWLINE] [NEWLINE] If you go after someone with a super soaker, they're not going to be scared until they think that they're sprayed with a small amount of fluid.  You can't just set them on fire to begin with, because then you don't get their money, because its on fire with them. [NEWLINE] [NEWLINE] Guns people understand.  You put a gun in their face, they worry about being shot, give you the money, and you're gone. [NEWLINE] [NEWLINE] If you sprayed me with lighter fluid, then asked for cash under threat of fire, I'm going to tackle you, and we both burn to death. [NEWLINE] [NEWLINE] Guns just work really well. Its possible to use a super soaker, but guns are just more efficient.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've read the gospels. They're not that deep. I mean some things have deep meanings (The feeding of the 5 thousand is a good one, even<mask> Christians insist on making it'magic' rather than an actual meaningful lesson about sharing)<mask> the majority is just dull and/or pointless. [NEWLINE] [NEWLINE] You get stuff like: [NEWLINE] [NEWLINE] John 1 [NEWLINE] [NEWLINE] [STARTQ] 6 There was a man sent from God whose name was John. 7 He came<mask> a witness to testify concerning that light,<mask> that through him all might believe. 8 He himself was not the light; he came only<mask> a witness to the light. [ENDQ] [NEWLINE] 7 and 8 are basically repeating itself... [NEWLINE] [NEWLINE] [STARTQ] 18 No one has ever seen God [ENDQ] [NEWLINE] Contradicts the old testament, and the idea (stated just a few lines above) that the word is God, and Jesus is the word, making Jesus God, and lots of people see him... [NEWLINE] [NEWLINE] We haven't even gotten halfway through the first chapter of John and it's already full of silly phrasing and contradictions. [NEWLINE] [NEWLINE] That's not good literature.</s>
Label encoding: <s>I've read the gospels. They're not that deep. I mean some things have deep meanings (The feeding of the 5 thousand is a good one, even if Christians insist on making it'magic' rather than an actual meaningful lesson about sharing) but the majority is just dull and/or pointless. [NEWLINE] [NEWLINE] You get stuff like: [NEWLINE] [NEWLINE] John 1 [NEWLINE] [NEWLINE] [STARTQ] 6 There was a man sent from God whose name was John. 7 He came as a witness to testify concerning that light, so that through him all might believe. 8 He himself was not the light; he came only as a witness to the light. [ENDQ] [NEWLINE] 7 and 8 are basically repeating itself... [NEWLINE] [NEWLINE] [STARTQ] 18 No one has ever seen God [ENDQ] [NEWLINE] Contradicts the old testament, and the idea (stated just a few lines above) that the word is God, and Jesus is the word, making Jesus God, and lots of people see him... [NEWLINE] [NEWLINE] We haven't even gotten halfway through the first chapter of John and it's already full of silly phrasing and contradictions. [NEWLINE] [NEWLINE] That's not good literature.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The issue with HIV transmission study is that it was flawed. [NEWLINE] Study took place in Africa and they compared muslims (cut) to non muslims (uncut)... Here lies the flaw: muslim countries are much, much more sexually repressive than non muslim countries (men and women there will have very small number of sexual partners during their life) and muslim men have the right to have several spouses diminishing the risk of extra marital sex with a non trusted partener. These 2 reasons are<mask> HIV infection rates are<mask> low.<mask> the non muslim part of Africa is generally more promiscuous<mask> the galoping infection rates. [NEWLINE] Nothing to do with circumcision. [NEWLINE] [NEWLINE] <mask> the authors of the study couldn't figure ou this mistake is beyond me.. almost feel like they're dishonest on purpose...<mask>? [NEWLINE] [NEWLINE] Don't forget that surgeons, anesthesiologists and then the hospital businesses really like the idea that they will perform an operation on every male born (guaranteed, continuous flow of customers), don't forget that they gain good money from these acts.  </s><pad>
Label encoding: <s>The issue with HIV transmission study is that it was flawed. [NEWLINE] Study took place in Africa and they compared muslims (cut) to non muslims (uncut)... Here lies the flaw: muslim countries are much, much more sexually repressive than non muslim countries (men and women there will have very small number of sexual partners during their life) and muslim men have the right to have several spouses diminishing the risk of extra marital sex with a non trusted partener. These 2 reasons are why HIV infection rates are so low. While the non muslim part of Africa is generally more promiscuous hence the galoping infection rates. [NEWLINE] Nothing to do with circumcision. [NEWLINE] [NEWLINE] Why the authors of the study couldn't figure ou this mistake is beyond me.. almost feel like they're dishonest on purpose... why? [NEWLINE] [NEWLINE] Don't forget that surgeons, anesthesiologists and then the hospital businesses really like the idea that they will perform an operation on every male born (guaranteed, continuous flow of customers), don't forget that they gain good money from these acts.  </s><pad>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, I apologize for saying your wife is shameful for her action. I'm a Christian,<mask> you can clearly tell, and<mask> not finding your wife's actions not ideal, I should keep my "negative" thoughts to myself. You seem like a good fella<mask> you could have just cursed me out like some redditters might<mask> religion or their family is brought up.<mask> you didn't.<mask> before I go I want to clarify on the prayer aspect of things. We pray for God's help,<mask> sometimes those prayers don't result in a DIRECT fix. The fix is sometimes indirect. You may meet someone<mask> depressed to help you out. You're sick, God has the medical field available for you. You can't explain to a child that "Oh, pray and your problems will be directly fixed or that every little problem WILL be fixed."<mask> I know you didn't condemn anyone. I'm stating that I personally can't condemn people for their views or actions. I may find it distasteful,<mask> that's it.<mask> anyways have a blessed day. God bless! </s>
Label encoding: <s>First off, I apologize for saying your wife is shameful for her action. I'm a Christian, as you can clearly tell, and despite not finding your wife's actions not ideal, I should keep my "negative" thoughts to myself. You seem like a good fella because you could have just cursed me out like some redditters might if religion or their family is brought up. But you didn't. But before I go I want to clarify on the prayer aspect of things. We pray for God's help, but sometimes those prayers don't result in a DIRECT fix. The fix is sometimes indirect. You may meet someone when depressed to help you out. You're sick, God has the medical field available for you. You can't explain to a child that "Oh, pray and your problems will be directly fixed or that every little problem WILL be fixed." Also I know you didn't condemn anyone. I'm stating that I personally can't condemn people for their views or actions. I may find it distasteful, but that's it. But anyways have a blessed day. God bless! </s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All of these things that you've listed<mask> emblematic of 'black culture' are<mask><mask> common in the cultures of poor people. There's an inter-city underclass of the extremely poor in the US, and the majority of those people are black,<mask> you associate the two. [NEWLINE] [NEWLINE] I'd recommend Elijah Anderson's Code Of The Street, in which he discusses life in these kinds of environments. [NEWLINE] [NEWLINE] <mask><mask><mask>, it's very problematic labelling any of<mask> you're talking about 'black culture',<mask> I guess we're going to have to use that<mask> a shorthand. <mask> I'd argue is that cultures don't form out of conscious choice. They are born out of necessity and shaped by the environments they exist in.  Criminal behaviour is glorified<mask> for the poorest of the poor, it is often the only way to'succeed'. <mask> the police don't bother coming to your neighbourhood, you develop alternate ways of dealing with problems.  The fact is that'mainstream culture' would not survive in the places that 'black culture' has developed.</s><pad>
Label encoding: <s>All of these things that you've listed as emblematic of 'black culture' are in fact common in the cultures of poor people. There's an inter-city underclass of the extremely poor in the US, and the majority of those people are black, hence you associate the two. [NEWLINE] [NEWLINE] I'd recommend Elijah Anderson's Code Of The Street, in which he discusses life in these kinds of environments. [NEWLINE] [NEWLINE] First of all, it's very problematic labelling any of what you're talking about 'black culture', but I guess we're going to have to use that as a shorthand.  What I'd argue is that cultures don't form out of conscious choice. They are born out of necessity and shaped by the environments they exist in.  Criminal behaviour is glorified because for the poorest of the poor, it is often the only way to'succeed'.  When the police don't bother coming to your neighbourhood, you develop alternate ways of dealing with problems.  The fact is that'mainstream culture' would not survive in the places that 'black culture' has developed.</s><pad>
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's the way I see it. Bullies, and it starts young, not only feed on the weakest victim,<mask> they feed off of the "followers"<mask> well- the kids that maybe don't have the conviction needed to stick up for the underdog for fear they will be the next victim,<mask> feel ashamed for it. This policy empowers those kids (the majority) to speak up without fear and do the right thing, knowing that their peers will likely do the same. It's taking the power of the "mob mentality" away from the bully and giving it to the decent kids. [NEWLINE] [NEWLINE] Second, you talk about shifting liability away from the school, etc,<mask> keep in mind, the school is us. It sucks that bullying is a school liability in the first place- it honestly should fall on the parent of the poorly raised child's parents.<mask>, here we are, and the best way to deal with it is to do our best to give our kids the encouragement to not put up with, or be manipulated by bullies, not even in a passive sense. </s>
Label encoding: <s>Here's the way I see it. Bullies, and it starts young, not only feed on the weakest victim, but they feed off of the "followers" as well- the kids that maybe don't have the conviction needed to stick up for the underdog for fear they will be the next victim, but feel ashamed for it. This policy empowers those kids (the majority) to speak up without fear and do the right thing, knowing that their peers will likely do the same. It's taking the power of the "mob mentality" away from the bully and giving it to the decent kids. [NEWLINE] [NEWLINE] Second, you talk about shifting liability away from the school, etc, but keep in mind, the school is us. It sucks that bullying is a school liability in the first place- it honestly should fall on the parent of the poorly raised child's parents. But, here we are, and the best way to deal with it is to do our best to give our kids the encouragement to not put up with, or be manipulated by bullies, not even in a passive sense. </s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My parents held the same view, OP, and we did not have a Santa Clause growing up. Their reasoning was more religious based (<mask> lie about one god-like figure and expect us to then trust that they were speaking the truth about God). Instead, we were taught that Christmas was about giving to those in need. We would pick a family at random and give them toys for the kids. (one time it was a family that had 5 kids running around outside, barefoot, dead of winter, no power on in the house--we bought them a ton of groceries, toys, clothes, and we paid their electric bill. Never met that family, and we just dropped the stuff off on their porch in the dead of night) [NEWLINE] [NEWLINE] To me, that's the meaning of Christmas. [NEWLINE] [NEWLINE] My Christmases had more wonder and "magic" in them than "getting stuff." [NEWLINE] [NEWLINE] Even now,<mask> an adult, I'd rather not spend cash on a tree, and would rather take that money and give it to someone who obviously needs it more than I do. </s>
Label encoding: <s>My parents held the same view, OP, and we did not have a Santa Clause growing up. Their reasoning was more religious based ( why lie about one god-like figure and expect us to then trust that they were speaking the truth about God). Instead, we were taught that Christmas was about giving to those in need. We would pick a family at random and give them toys for the kids. (one time it was a family that had 5 kids running around outside, barefoot, dead of winter, no power on in the house--we bought them a ton of groceries, toys, clothes, and we paid their electric bill. Never met that family, and we just dropped the stuff off on their porch in the dead of night) [NEWLINE] [NEWLINE] To me, that's the meaning of Christmas. [NEWLINE] [NEWLINE] My Christmases had more wonder and "magic" in them than "getting stuff." [NEWLINE] [NEWLINE] Even now, as an adult, I'd rather not spend cash on a tree, and would rather take that money and give it to someone who obviously needs it more than I do. </s>
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is somewhat contradicted by [this study]( [URL].1.9748884.html).<mask> obesity may require more health care expenses<mask> the person is alive,<mask> is noted in your source, obese people die earlier and more quickly, meaning obscenely expensive end-of-life care is not needed for them.<mask>, morbid<mask> it is, it's possible we're actually increasing our healthcare costs - and<mask> our taxes - in the long run by encouraging healthy body weights. [NEWLINE] [NEWLINE] Personally,<mask><mask><mask> you want to encourage a healthy lifestyle, you should do it for the right reason -<mask> you want to see people be happy and healthy.<mask> an aside -<mask> it's not directly related your point - I<mask> think a lot of the'shaming' culture is just people taking advantage of a socially undesirable problem that's impossible to hide<mask> a get-out-of-jail free card to make fun of people with impunity, not truly<mask> they believe they're helping, and that's<mask> likely part of<mask> fuels OP's and other people's opinions on this matter.</s>
Label encoding: <s>This is somewhat contradicted by [this study]( [URL].1.9748884.html). While obesity may require more health care expenses while the person is alive, as is noted in your source, obese people die earlier and more quickly, meaning obscenely expensive end-of-life care is not needed for them. So, morbid as it is, it's possible we're actually increasing our healthcare costs - and therefore our taxes - in the long run by encouraging healthy body weights. [NEWLINE] [NEWLINE] Personally, I think if you want to encourage a healthy lifestyle, you should do it for the right reason - because you want to see people be happy and healthy. As an aside - as it's not directly related your point - I also think a lot of the'shaming' culture is just people taking advantage of a socially undesirable problem that's impossible to hide as a get-out-of-jail free card to make fun of people with impunity, not truly because they believe they're helping, and that's what likely part of what fuels OP's and other people's opinions on this matter.</s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see the periodic table<mask> a way to show that the elements belong together. It shows there is order in the world that can be understood, and it's beautiful. [NEWLINE] [NEWLINE] One of the most amazing features of the periodic table is that the guy who made it has left *holes* in the table. He said "I don't know<mask> it is,<mask> there must be an element I have never seen before that goes here". He has predicted the qualities of these elements before they were discovered. Even<mask> the artificial superheavy elements are not useful in our every day life,<mask> we left them out, there would be holes in the table, never to be filled. [NEWLINE] [NEWLINE] I don't think we should look at the elements in the table one by one. The atomic mass of oxygen alone is not very interesting. The important information is "<mask> different" it is from the nearby elements. The table shows<mask> the nature of matter changes<mask> we keep adding protons. Having all the elements there will validate and strengthen this knowledge,<mask><mask><mask> frequent or<mask> stable an element is.</s>
Label encoding: <s>I see the periodic table as a way to show that the elements belong together. It shows there is order in the world that can be understood, and it's beautiful. [NEWLINE] [NEWLINE] One of the most amazing features of the periodic table is that the guy who made it has left *holes* in the table. He said "I don't know what it is, but there must be an element I have never seen before that goes here". He has predicted the qualities of these elements before they were discovered. Even if the artificial superheavy elements are not useful in our every day life, if we left them out, there would be holes in the table, never to be filled. [NEWLINE] [NEWLINE] I don't think we should look at the elements in the table one by one. The atomic mass of oxygen alone is not very interesting. The important information is " how different" it is from the nearby elements. The table shows how the nature of matter changes as we keep adding protons. Having all the elements there will validate and strengthen this knowledge, regardless of how frequent or how stable an element is.</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My thoughts on this... [NEWLINE] [NEWLINE] 1. Philosophically, I would disagree with putting primates in zoos<mask>,<mask> our closest relatives, they are very intelligent and behave a lot like us to the point that I consider them people. Bonobos are<mask> already not allowed in zoos<mask> they fuck all the time. [NEWLINE] [NEWLINE] 2. Piggybacking off my first point, there are<mask> a lot of others that are intelligent enough to realize that life kinda sucks for them contained in a zoo all their lives. I can't imagine putting my cat or dog in one. [NEWLINE] [NEWLINE] 3. I would only agree with putting intelligent animals in zoos<mask> it is primarily to preserve their species. A zoo is practically the only place<mask> they would be given warmth and comfort. [NEWLINE] [NEWLINE] Of course, the problem with all my points that that,<mask> do we draw the line on intelligence?<mask> do we know<mask> an animal is dissatisfied with his/her life in the zoo? I'm not totally against zoos,<mask> liveliness and health should come first for everyone in them.</s>
Label encoding: <s>My thoughts on this... [NEWLINE] [NEWLINE] 1. Philosophically, I would disagree with putting primates in zoos since, as our closest relatives, they are very intelligent and behave a lot like us to the point that I consider them people. Bonobos are also already not allowed in zoos because they fuck all the time. [NEWLINE] [NEWLINE] 2. Piggybacking off my first point, there are also a lot of others that are intelligent enough to realize that life kinda sucks for them contained in a zoo all their lives. I can't imagine putting my cat or dog in one. [NEWLINE] [NEWLINE] 3. I would only agree with putting intelligent animals in zoos if it is primarily to preserve their species. A zoo is practically the only place where they would be given warmth and comfort. [NEWLINE] [NEWLINE] Of course, the problem with all my points that that, where do we draw the line on intelligence? How do we know if an animal is dissatisfied with his/her life in the zoo? I'm not totally against zoos, but liveliness and health should come first for everyone in them.</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not about "cruel or unusual punishment," it's about constitutional rights, which say that the government [NEWLINE] [NEWLINE] [STARTQ] shall make no law respecting an establishment of religion, or prohibiting the free exercise thereof. [ENDQ] [NEWLINE] Prohibiting certain clothing would be prohibiting the free exercise of religion. [NEWLINE] [NEWLINE] <mask> for this: [NEWLINE] [NEWLINE] [STARTQ] ignoring all religious obligations put all prisoners on equal grounds, which a commendable goal. No special treatment for anyone. [ENDQ] [NEWLINE] This is very much like arguments against gay marriage that say "it's not discrimination - it is treating everyone the same! It's not just gay people who cannot marry a person of the same sex, it's everyone! No special treatment." Laws and policies that impact one demographic disproportionately are discriminatory, and disallowing head scarves or prayer mats only affects Muslims,<mask> a Christian would be free to pray to themselves before dinner. Making rules<mask> that only certain religions (which do no require headgear or beards or prayer mats) would be able to practice<mask> others would be prohibited from it is the opposite of equal.</s>
Label encoding: <s>It's not about "cruel or unusual punishment," it's about constitutional rights, which say that the government [NEWLINE] [NEWLINE] [STARTQ] shall make no law respecting an establishment of religion, or prohibiting the free exercise thereof. [ENDQ] [NEWLINE] Prohibiting certain clothing would be prohibiting the free exercise of religion. [NEWLINE] [NEWLINE] As for this: [NEWLINE] [NEWLINE] [STARTQ] ignoring all religious obligations put all prisoners on equal grounds, which a commendable goal. No special treatment for anyone. [ENDQ] [NEWLINE] This is very much like arguments against gay marriage that say "it's not discrimination - it is treating everyone the same! It's not just gay people who cannot marry a person of the same sex, it's everyone! No special treatment." Laws and policies that impact one demographic disproportionately are discriminatory, and disallowing head scarves or prayer mats only affects Muslims, while a Christian would be free to pray to themselves before dinner. Making rules so that only certain religions (which do no require headgear or beards or prayer mats) would be able to practice but others would be prohibited from it is the opposite of equal.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>without going into scientific and specific details, a problem here is your methodology to statuate<mask> it's healthy or not : [NEWLINE] [NEWLINE] - you "feel like" it's bad. We should not listen to our gut feeling<mask> we can listen to experts and scientific who actually have some knowledge. This is the root of all obscurantism. We feel like our opinion is worth more than the one of experts,<mask> it's just wrong. [NEWLINE] [NEWLINE] - long term research are not the only way to statuate. We can analyze<mask> we know. For example<mask> we know all chemicals involved and can rule them safe, then we should not wait 10years study to rule the mix safe. [NEWLINE] [NEWLINE] - a third misconception is that<mask> you state "Maybe 20 years from now there will be an increase in 'Painful-Watery-Lung-Death' caused by E-Cigs." --&gt;<mask> it's not certain does not mean it's 50/50. MAYBE it will come out bad,<mask> PROBABLY not. And it's important</s>
Label encoding: <s>without going into scientific and specific details, a problem here is your methodology to statuate if it's healthy or not : [NEWLINE] [NEWLINE] - you "feel like" it's bad. We should not listen to our gut feeling when we can listen to experts and scientific who actually have some knowledge. This is the root of all obscurantism. We feel like our opinion is worth more than the one of experts, but it's just wrong. [NEWLINE] [NEWLINE] - long term research are not the only way to statuate. We can analyze what we know. For example if we know all chemicals involved and can rule them safe, then we should not wait 10years study to rule the mix safe. [NEWLINE] [NEWLINE] - a third misconception is that when you state "Maybe 20 years from now there will be an increase in 'Painful-Watery-Lung-Death' caused by E-Cigs." --&gt; because it's not certain does not mean it's 50/50. MAYBE it will come out bad, but PROBABLY not. And it's important</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Again: "No" would've been a perfectly acceptable answer. They could've said "No"<mask> I asked<mask> they were still open. They could've said "No"<mask> I told there it would be a large party.<mask> I told them we didn't want a limited menu, and said we were leaving then, they could've said "Have a nice night". They chose to, not only say "Yes",<mask> to do<mask> over and over again. [ENDQ] [NEWLINE] It sounds like the waitstaff wanted you gone,<mask> were required by the manager to seat and serve you.<mask> again, you didn't break any rules,<mask> basic empathy would generally alert you to the fact that the waitstaff didn't want to be there, and would probably be fired<mask> they asked you to leave.<mask> you're putting people in a position that requires them to choose between being inconvenienced or being fired, then you're being a dick. You just are. You're entitled to be a dick,<mask> that doesn't mean there's nothing wrong with it. </s>
Label encoding: <s> [STARTQ] Again: "No" would've been a perfectly acceptable answer. They could've said "No" when I asked if they were still open. They could've said "No" when I told there it would be a large party. When I told them we didn't want a limited menu, and said we were leaving then, they could've said "Have a nice night". They chose to, not only say "Yes", but to do so over and over again. [ENDQ] [NEWLINE] It sounds like the waitstaff wanted you gone, but were required by the manager to seat and serve you. So again, you didn't break any rules, but basic empathy would generally alert you to the fact that the waitstaff didn't want to be there, and would probably be fired if they asked you to leave. If you're putting people in a position that requires them to choose between being inconvenienced or being fired, then you're being a dick. You just are. You're entitled to be a dick, but that doesn't mean there's nothing wrong with it. </s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>According to Christian doctrine<mask> I was taught it, there's no concept of time in the afterlife. It's just an eternal experience of the presence of and love of God. [NEWLINE] [NEWLINE] Many Christians see all goodness in the world<mask> a shadow of<mask> God will be. Presently we turn from something partially good to something else partially good,<mask> the goods are imperfect.<mask><mask> everything culminates in the perfect good of God, there's no reason to want to turn to something else (e.g. no reason to be bored).<mask>, boredom is just part of the fallen nature of humanity, and after the fallen parts are stripped away and we are given our new bodies, "boredom" will cease to exist. [NEWLINE] [NEWLINE] You imagine humanity<mask> essentially unaltered, still temporal beings, your current self<mask> better and with heavenly resources,<mask> that's not<mask> Christians think heaven will be like. Christian think all negative aspects of humanity vanish in heaven,<mask> those negative aspects were not part of the original Garden of Eden design of humanity. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>According to Christian doctrine as I was taught it, there's no concept of time in the afterlife. It's just an eternal experience of the presence of and love of God. [NEWLINE] [NEWLINE] Many Christians see all goodness in the world as a shadow of what God will be. Presently we turn from something partially good to something else partially good, because the goods are imperfect. But when everything culminates in the perfect good of God, there's no reason to want to turn to something else (e.g. no reason to be bored). Additionally, boredom is just part of the fallen nature of humanity, and after the fallen parts are stripped away and we are given our new bodies, "boredom" will cease to exist. [NEWLINE] [NEWLINE] You imagine humanity as essentially unaltered, still temporal beings, your current self but better and with heavenly resources, but that's not what Christians think heaven will be like. Christian think all negative aspects of humanity vanish in heaven, because those negative aspects were not part of the original Garden of Eden design of humanity. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0481, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's a small thing,<mask> its madness and<mask> we can only fit one lane of people in at a time it seems really inefficient for<mask> many people to be standing around. [ENDQ] [NEWLINE] Inefficient for each individual person, yes.<mask> it's actually more efficient for the airline, and<mask> collectively for all the passengers<mask> well. [NEWLINE] [NEWLINE] Let's say it takes 30 seconds to stand up from your chair, put away your magazine or reading material, take your last swig of water and put that in your backpack, gather all your luggage in your arms, and walk over to the boarding tunnel entry line. [NEWLINE] [NEWLINE] <mask> everybody did that at the very last second they were able, it would add that 30 seconds of time in for every person. [NEWLINE] [NEWLINE] The way it works now is that every person gets up and takes their 30 seconds to gather their stuff and walk over, and then they're all already in line to be able to walk through the tunnel immediately after the passenger ahead of them instead of taking their 30 seconds at that time. </s>
Label encoding: <s> [STARTQ] It's a small thing, but its madness and because we can only fit one lane of people in at a time it seems really inefficient for so many people to be standing around. [ENDQ] [NEWLINE] Inefficient for each individual person, yes. But it's actually more efficient for the airline, and thus collectively for all the passengers as well. [NEWLINE] [NEWLINE] Let's say it takes 30 seconds to stand up from your chair, put away your magazine or reading material, take your last swig of water and put that in your backpack, gather all your luggage in your arms, and walk over to the boarding tunnel entry line. [NEWLINE] [NEWLINE] If everybody did that at the very last second they were able, it would add that 30 seconds of time in for every person. [NEWLINE] [NEWLINE] The way it works now is that every person gets up and takes their 30 seconds to gather their stuff and walk over, and then they're all already in line to be able to walk through the tunnel immediately after the passenger ahead of them instead of taking their 30 seconds at that time. </s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is a distinction between academic feminism and popular feminism, and I am certainly not trying to absolutely encapsulate fifty years of academic thought, particularly<mask><mask> much of it consists of vastly differing opinions. I'm just trying to give a general overview of the beliefs of at least a part of the MRM. [NEWLINE] [NEWLINE] With reference to patriarchy theory, I have read both the term and the theory in feminist literature.<mask> you are correct in that strictly speaking it refers to men in positions of power, I have often seen this extended to "men possess social power". I was simply trying to point out the misconception that extension involves. [NEWLINE] [NEWLINE] Intersectionality originated in the intersection of of Marxist, feminist, critical race and half a dozen other critiques of social order. The point I'm making is not that intersectionality is not feminist<mask> that in a history of power, class has far more influence than gender. Nowhere do I deny that privilege exists, I'm simply criticizing the extent, magnitude and monolithic nature I have occasionally seen it ascribed. </s>
Label encoding: <s>There is a distinction between academic feminism and popular feminism, and I am certainly not trying to absolutely encapsulate fifty years of academic thought, particularly when so much of it consists of vastly differing opinions. I'm just trying to give a general overview of the beliefs of at least a part of the MRM. [NEWLINE] [NEWLINE] With reference to patriarchy theory, I have read both the term and the theory in feminist literature. While you are correct in that strictly speaking it refers to men in positions of power, I have often seen this extended to "men possess social power". I was simply trying to point out the misconception that extension involves. [NEWLINE] [NEWLINE] Intersectionality originated in the intersection of of Marxist, feminist, critical race and half a dozen other critiques of social order. The point I'm making is not that intersectionality is not feminist but that in a history of power, class has far more influence than gender. Nowhere do I deny that privilege exists, I'm simply criticizing the extent, magnitude and monolithic nature I have occasionally seen it ascribed. </s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> do you know that even those bits of suffering are needless?<mask> does some small amount of relative suffering somehow make the universe not worth living in?<mask> there was not even a chance of injuring yourself by accident,<mask> would the point in living be? It wouldn't be called "survival"<mask> we all had a 100% chance of living through life. [NEWLINE] [NEWLINE] How do accidents preclude a benevolent god? For all you know, the<mask> -called "needless" suffering actually makes you a better person and prepares you for some new assignment in the next life. In other words, it offers an opportunity for growth. [NEWLINE] [NEWLINE] I don't know of a single experience in the universe that teaches you about<mask> makes life worth living more than the consequences of suffering -- be it accidental or otherwise. A benevolent god would -- nay, MUST -- allow it to happen. [NEWLINE] [NEWLINE] I completely disagree that evil and suffering proves that god is not real or that he is malevolent. I find that argument to be sophomoric and shortsighted.</s>
Label encoding: <s>How do you know that even those bits of suffering are needless? How does some small amount of relative suffering somehow make the universe not worth living in? If there was not even a chance of injuring yourself by accident, what would the point in living be? It wouldn't be called "survival" if we all had a 100% chance of living through life. [NEWLINE] [NEWLINE] How do accidents preclude a benevolent god? For all you know, the so -called "needless" suffering actually makes you a better person and prepares you for some new assignment in the next life. In other words, it offers an opportunity for growth. [NEWLINE] [NEWLINE] I don't know of a single experience in the universe that teaches you about what makes life worth living more than the consequences of suffering -- be it accidental or otherwise. A benevolent god would -- nay, MUST -- allow it to happen. [NEWLINE] [NEWLINE] I completely disagree that evil and suffering proves that god is not real or that he is malevolent. I find that argument to be sophomoric and shortsighted.</s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>you'll have an easy time reading about the jfk cover-up. [NEWLINE] [NEWLINE] <mask> for mlk (hint: ballistics tests clear james earl ray) [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] twelve jurors reached a unanimous verdict on December 8, 1999 after about an hour of deliberations that Dr. Martin Luther King, Jr. was assassinated<mask><mask><mask> of a conspiracy. In a press statement held the following day in Atlanta, Mrs. Coretta Scott King welcomed the verdict, saying, “There is abundant evidence of a major high level conspiracy in the assassination of my husband, Martin Luther King, Jr. And the civil court's unanimous verdict has validated our belief. I wholeheartedly applaud the verdict of the jury and I feel that justice has been well served in their deliberations. This verdict is not only a great victory for my family,<mask><mask> a great victory for America. It is a great victory for truth itself. It is important to know that this was a SWIFT verdict, delivered after about an hour of jury deliberation.</s>
Label encoding: <s>you'll have an easy time reading about the jfk cover-up. [NEWLINE] [NEWLINE] as for mlk (hint: ballistics tests clear james earl ray) [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] twelve jurors reached a unanimous verdict on December 8, 1999 after about an hour of deliberations that Dr. Martin Luther King, Jr. was assassinated as a result of a conspiracy. In a press statement held the following day in Atlanta, Mrs. Coretta Scott King welcomed the verdict, saying, “There is abundant evidence of a major high level conspiracy in the assassination of my husband, Martin Luther King, Jr. And the civil court's unanimous verdict has validated our belief. I wholeheartedly applaud the verdict of the jury and I feel that justice has been well served in their deliberations. This verdict is not only a great victory for my family, but also a great victory for America. It is a great victory for truth itself. It is important to know that this was a SWIFT verdict, delivered after about an hour of jury deliberation.</s>
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Historically, it's been more common for men to be obligated to go to war under threat of imprisonment, torture, and/or execution. I say that counts<mask> oppression. [ENDQ] [NEWLINE] That is certainly true, and almost all of the warrior elites I listed fought alongside these kinds of soldiers. My point was that for warrior elites risking death in no way indicates that they were oppressed. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the disposability of men comes from evolutionary biology, not sexism. [ENDQ] [NEWLINE] <mask> most of the time<mask> we see a women die on TV or such we don't think anything along the lines of 'oh no, a drop in the birthrate!'.<mask><mask><mask> evolutionary biology is the cause it will be<mask> those socieites with paternalistic attitudes towards women were more likely to survive/thrive due to higher birthrates,<mask> entrenching those attitudes. [NEWLINE] [NEWLINE] [STARTQ] It makes sense for cavemen,<mask> I don't think that should be the standard for modern practices. [ENDQ] [NEWLINE] <mask><mask> we both agree on this.</s>
Label encoding: <s> [STARTQ] Historically, it's been more common for men to be obligated to go to war under threat of imprisonment, torture, and/or execution. I say that counts as oppression. [ENDQ] [NEWLINE] That is certainly true, and almost all of the warrior elites I listed fought alongside these kinds of soldiers. My point was that for warrior elites risking death in no way indicates that they were oppressed. [NEWLINE] [NEWLINE] [STARTQ] I think the disposability of men comes from evolutionary biology, not sexism. [ENDQ] [NEWLINE] But most of the time when we see a women die on TV or such we don't think anything along the lines of 'oh no, a drop in the birthrate!'. I think if evolutionary biology is the cause it will be because those socieites with paternalistic attitudes towards women were more likely to survive/thrive due to higher birthrates, thus entrenching those attitudes. [NEWLINE] [NEWLINE] [STARTQ] It makes sense for cavemen, but I don't think that should be the standard for modern practices. [ENDQ] [NEWLINE] I think we both agree on this.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Ok, you can drive to and from work and in emergencies; that's it. Car accidents kill way too many people to keep letting people drive all over the place whenever they want. [ENDQ] [NEWLINE] <mask> the hell are you trying to argue? Cars are useful in our lives; obesity is not.<mask>, keep the cars, and fight obesity. [NEWLINE] [NEWLINE] [STARTQ] <mask> we should ban these things then<mask> people can choose to not partake in them and allow them to slowly kill them (or quickly with alcohol in some cases).<mask> someone cannot "partake" in obesity whenever they feel like it like you say, then<mask> should we punish people for something that they don't necessarily have control over? [ENDQ] [NEWLINE] The fact that people can choose not to partake in smoking and drinking is a reason to **not** ban those things. Obesity kills people, and those people are stuck with it. The plan is not to punish fat people,<mask> that may end up incidentally happening. The plan is to get rid of the excess weight, and develop healthier habits.</s>
Label encoding: <s> [STARTQ] Ok, you can drive to and from work and in emergencies; that's it. Car accidents kill way too many people to keep letting people drive all over the place whenever they want. [ENDQ] [NEWLINE] What the hell are you trying to argue? Cars are useful in our lives; obesity is not. Therefore, keep the cars, and fight obesity. [NEWLINE] [NEWLINE] [STARTQ] So we should ban these things then because people can choose to not partake in them and allow them to slowly kill them (or quickly with alcohol in some cases). If someone cannot "partake" in obesity whenever they feel like it like you say, then why should we punish people for something that they don't necessarily have control over? [ENDQ] [NEWLINE] The fact that people can choose not to partake in smoking and drinking is a reason to **not** ban those things. Obesity kills people, and those people are stuck with it. The plan is not to punish fat people, although that may end up incidentally happening. The plan is to get rid of the excess weight, and develop healthier habits.</s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I was walking down the road with my friends, saw a turkey, and proceeded to snap its neck, my friends would be pretty horrified even<mask> they ate a turkey sandwich an hour before that. This makes me thing that it's less about not valuing the life of the turkey and more of the disconnect between the idea of the physical turkey and the food on your plate. [ENDQ] [NEWLINE] Or rather, the wastefulness of the act. <mask> you guys were stuck in a forest, starving, and you snapped a turkeys neck and ate him, I doubt they'd fault you.  Killing and eating an animal isn't intrinsically wrong. [NEWLINE] [NEWLINE] [STARTQ] <mask> there's a vast gap between the "intelligence" of a cockroach and that of a chicken. [ENDQ] [NEWLINE] Then clearly there's a point along the spectrum between bugs and chickens that you feel is the boundary between permissible killing and immoral killing.  For many people, this line is at a separate place than<mask> you put it -- in my case, anything less than human.  </s>
Label encoding: <s> [STARTQ] If I was walking down the road with my friends, saw a turkey, and proceeded to snap its neck, my friends would be pretty horrified even if they ate a turkey sandwich an hour before that. This makes me thing that it's less about not valuing the life of the turkey and more of the disconnect between the idea of the physical turkey and the food on your plate. [ENDQ] [NEWLINE] Or rather, the wastefulness of the act.  If you guys were stuck in a forest, starving, and you snapped a turkeys neck and ate him, I doubt they'd fault you.  Killing and eating an animal isn't intrinsically wrong. [NEWLINE] [NEWLINE] [STARTQ] but there's a vast gap between the "intelligence" of a cockroach and that of a chicken. [ENDQ] [NEWLINE] Then clearly there's a point along the spectrum between bugs and chickens that you feel is the boundary between permissible killing and immoral killing.  For many people, this line is at a separate place than where you put it -- in my case, anything less than human.  </s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh man, I<mask> know<mask> you're coming from here. [NEWLINE] [NEWLINE] I'm of European ancestry,<mask> am mostly a mutt, with more German, English and Scots-Irish blood than anything else<mask> plenty of other stuff blended in there. My family moved to New Mexico<mask> I was 2 and I grew up here. My spouse identifies most with the label "Brown" (he's of Mexican descent,<mask><mask> you go back he's a mix of Native and Spanish, probably,<mask> there's totally possibilities of other European blood, African blood, and a ton of different tribes in there). [NEWLINE] [NEWLINE] And<mask> the culture that I identify most strongly with is place-based: the green chile thing, the mañana attitude, walking around not in a huge hurry, smiling at eye contact, luminarias, stuff like that. It's not<mask> my blood comes from,<mask> it's the place I'm in love with, and I feel like there's more choice there,<mask> I get to choose<mask> I live.</s>
Label encoding: <s>Oh man, I so know what you're coming from here. [NEWLINE] [NEWLINE] I'm of European ancestry, but am mostly a mutt, with more German, English and Scots-Irish blood than anything else but plenty of other stuff blended in there. My family moved to New Mexico when I was 2 and I grew up here. My spouse identifies most with the label "Brown" (he's of Mexican descent, so if you go back he's a mix of Native and Spanish, probably, but there's totally possibilities of other European blood, African blood, and a ton of different tribes in there). [NEWLINE] [NEWLINE] And so the culture that I identify most strongly with is place-based: the green chile thing, the mañana attitude, walking around not in a huge hurry, smiling at eye contact, luminarias, stuff like that. It's not where my blood comes from, but it's the place I'm in love with, and I feel like there's more choice there, because I get to choose where I live.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> in the world would you choose to attack someone rather than just leave? [ENDQ] [NEWLINE] <mask> Zimmerman's account of the events is incomplete. [NEWLINE] [NEWLINE] You've hit the core here.  Attacking Zimmerman flies in the face of reason<mask> his version of events are true. [NEWLINE] [NEWLINE] <mask> we're willing to believe them,<mask>, for whatever reason, it makes more sense to believe that Martin did this bizzare thing, this thing<mask>, after I just told you I'd do it, you wondered out loud<mask> on earth I would. [NEWLINE] [NEWLINE] <mask><mask> Zimmerman attempted to detain Martin, suddenly the altercation makes perfect sense. [NEWLINE] [NEWLINE] [STARTQ] the black guy is dead and that's enough for you to assume that the white guy was a murder. [ENDQ] [NEWLINE] Well, it is, actually, unless there's evidence that something else happened.  There is additional evidence, and Zimmerman paints a fairly clear picture. <mask> like you said, it doesn't make sense for Martin to just jump out and attack Zimmerman on his way home. [NEWLINE] </s>
Label encoding: <s> [STARTQ] why in the world would you choose to attack someone rather than just leave? [ENDQ] [NEWLINE] Because Zimmerman's account of the events is incomplete. [NEWLINE] [NEWLINE] You've hit the core here.  Attacking Zimmerman flies in the face of reason if his version of events are true. [NEWLINE] [NEWLINE] But we're willing to believe them, because, for whatever reason, it makes more sense to believe that Martin did this bizzare thing, this thing where, after I just told you I'd do it, you wondered out loud why on earth I would. [NEWLINE] [NEWLINE] But if Zimmerman attempted to detain Martin, suddenly the altercation makes perfect sense. [NEWLINE] [NEWLINE] [STARTQ] the black guy is dead and that's enough for you to assume that the white guy was a murder. [ENDQ] [NEWLINE] Well, it is, actually, unless there's evidence that something else happened.  There is additional evidence, and Zimmerman paints a fairly clear picture.  But like you said, it doesn't make sense for Martin to just jump out and attack Zimmerman on his way home. [NEWLINE] </s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Marriage is very related to the concern of incest simply<mask><mask> you are able to marry incestuously, then it is accepted.<mask> it is accepted, then it will become more populous and more people will do it; this could produce the negative effects, touched on briefly in the original post. [ENDQ] [NEWLINE] <mask> incestuous marriage would only ever be widely accepted by society *after* a hypothetical mainstream acceptance of incestuous sex. Making the focus on the marriage aspect once again irrelevant. [NEWLINE] [NEWLINE] [STARTQ] This may be true,<mask> my post is referring to the potential effects on our society of incest and<mask> the legalization of gay marriage can have an impact on incest becoming accepted. Society is centered around one country, one instance of time, etc.<mask>, you're right.<mask>, that was intended. [ENDQ] [NEWLINE] <mask> then did you include in your argument the assertion that "any sexual relationship outside of marriage between a man and woman can be argued effectively<mask> unnatural"?<mask> does the naturalness, or lack thereof, have to do with anything?</s>
Label encoding: <s> [STARTQ] Marriage is very related to the concern of incest simply because if you are able to marry incestuously, then it is accepted. If it is accepted, then it will become more populous and more people will do it; this could produce the negative effects, touched on briefly in the original post. [ENDQ] [NEWLINE] But incestuous marriage would only ever be widely accepted by society *after* a hypothetical mainstream acceptance of incestuous sex. Making the focus on the marriage aspect once again irrelevant. [NEWLINE] [NEWLINE] [STARTQ] This may be true, but my post is referring to the potential effects on our society of incest and how the legalization of gay marriage can have an impact on incest becoming accepted. Society is centered around one country, one instance of time, etc. So, you're right. However, that was intended. [ENDQ] [NEWLINE] So then did you include in your argument the assertion that "any sexual relationship outside of marriage between a man and woman can be argued effectively as unnatural"? What does the naturalness, or lack thereof, have to do with anything?</s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm been stuck with this debate with myself for months now, and I can't think<mask> is the right moral path.. [NEWLINE] [NEWLINE] I've always thought that killing an animal for the right purpose, food and not fun, is morally right<mask> there animals don't understand death, like a human does. They will have the same wonderful life, in the wild, even<mask> I kill it or not.. [NEWLINE] <mask> for producing animals, in order for food, I<mask> think this is ok<mask><mask><mask> the animals are treated right, and fair. [NEWLINE] [NEWLINE] That's<mask> i'm a vegan right now,<mask> I hate<mask> the main meat sources we have come from horrible people who treat animals with un needed cruelty, or atleast it seems like they do. [NEWLINE] [NEWLINE] Edit:<mask> i'm<mask> asking is, is it wrong to actually eat meat from these factories? People make it seem like it's horrible,<mask> I just don't know. [NEWLINE] [NEWLINE] Any pro and con arguments would help me out a lot, thanks!</s>
Label encoding: <s>I'm been stuck with this debate with myself for months now, and I can't think what is the right moral path.. [NEWLINE] [NEWLINE] I've always thought that killing an animal for the right purpose, food and not fun, is morally right because there animals don't understand death, like a human does. They will have the same wonderful life, in the wild, even if I kill it or not.. [NEWLINE] As for producing animals, in order for food, I also think this is ok as long as the animals are treated right, and fair. [NEWLINE] [NEWLINE] That's why i'm a vegan right now, because I hate how the main meat sources we have come from horrible people who treat animals with un needed cruelty, or atleast it seems like they do. [NEWLINE] [NEWLINE] Edit: What i'm also asking is, is it wrong to actually eat meat from these factories? People make it seem like it's horrible, but I just don't know. [NEWLINE] [NEWLINE] Any pro and con arguments would help me out a lot, thanks!</s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay<mask> I get<mask> you say and<mask><mask>. I found it especially interesting that you mentioned the long history of animal spirits, which<mask> makes sense,<mask> well<mask> the idea of being taken by the Holy Spirit. Howeve<mask> happens<mask> this dillusion does interfere with ones life to a great deal. Say a particularly strong Christian decides that the Holy Spirit has told them they are not to allow their kids to do somthing simple and socially necessary like get an education? Do they then fit the criteria of a person suffering from a debilitating delusion? I ask<mask> I wince read a post by a "deer kin" who was upset that her son had eaten deer meat at a friends house. The son told her he liked the meat and was going to keep eating it. Her reaction was to essentially abandon her young son who was around ten I belive. Would this constitute a delusion? I'm not trying to argue with you, just asking out of curiosity<mask> you seem far better versed in this than I am.</s>
Label encoding: <s>Okay so I get what you say and I agree. I found it especially interesting that you mentioned the long history of animal spirits, which also makes sense, as well as the idea of being taken by the Holy Spirit. Howeve what happens when this dillusion does interfere with ones life to a great deal. Say a particularly strong Christian decides that the Holy Spirit has told them they are not to allow their kids to do somthing simple and socially necessary like get an education? Do they then fit the criteria of a person suffering from a debilitating delusion? I ask because I wince read a post by a "deer kin" who was upset that her son had eaten deer meat at a friends house. The son told her he liked the meat and was going to keep eating it. Her reaction was to essentially abandon her young son who was around ten I belive. Would this constitute a delusion? I'm not trying to argue with you, just asking out of curiosity as you seem far better versed in this than I am.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Heterosexual: potentially sexually attracted to individuals on the opposite end of the gender spectrum. [NEWLINE] [NEWLINE] Homosexual: potentially sexually attracted to individuals on the same end of the gender spectrum. [NEWLINE] [NEWLINE] Asexual: no sexual attraction. [NEWLINE] [NEWLINE] Bisexual: potentially attracted to individuals at both ends of the gender spectrum. [NEWLINE] [NEWLINE] That doesn't cover the entire list.<mask> you acknowledge gender<mask> a spectrum rather than a binary, more terms are necessary to be accurate. [NEWLINE] [NEWLINE] Pansexual: potentially attracted to individuals anywhere on the gender spectrum. [NEWLINE] [NEWLINE] Polysexual: potentially attracted to individuals at multiple points/ranges on the gender spectrum. (This is a superset of bi and pan,<mask> can<mask> include those who are attracted to, say, masculine or gender-neutral individuals,<mask> not feminine individuals.) [NEWLINE] [NEWLINE] None of this is "I want to be different!" It's just about more accurate labels. Hetero/homo/a/bisexual is only sufficient<mask> you think gender is a strict binary.</s>
Label encoding: <s>Heterosexual: potentially sexually attracted to individuals on the opposite end of the gender spectrum. [NEWLINE] [NEWLINE] Homosexual: potentially sexually attracted to individuals on the same end of the gender spectrum. [NEWLINE] [NEWLINE] Asexual: no sexual attraction. [NEWLINE] [NEWLINE] Bisexual: potentially attracted to individuals at both ends of the gender spectrum. [NEWLINE] [NEWLINE] That doesn't cover the entire list. If you acknowledge gender as a spectrum rather than a binary, more terms are necessary to be accurate. [NEWLINE] [NEWLINE] Pansexual: potentially attracted to individuals anywhere on the gender spectrum. [NEWLINE] [NEWLINE] Polysexual: potentially attracted to individuals at multiple points/ranges on the gender spectrum. (This is a superset of bi and pan, but can also include those who are attracted to, say, masculine or gender-neutral individuals, but not feminine individuals.) [NEWLINE] [NEWLINE] None of this is "I want to be different!" It's just about more accurate labels. Hetero/homo/a/bisexual is only sufficient if you think gender is a strict binary.</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can you do a perfect cartwheel? [NEWLINE] [NEWLINE] Can you do a perfect cartwheel your first try? [NEWLINE] [NEWLINE] Can you do a perfect cartwheel<mask> losing blood with adrenalin, fear and  without any time to mentally prepare for doing one? [NEWLINE] [NEWLINE] --------- [NEWLINE] [NEWLINE] Anyone can see themselves doing a cartwheel<mask><mask><mask> they can walk, and its not like the laws of nature of<mask> to balance are hard to figure out;<mask> your first try will most likely be sloppy, even<mask> you give your subconscious plenty of time take estimates about<mask> it will need to do to balance. [NEWLINE] [NEWLINE] Now going into a fight without training, will make you sloppy *at best* its fairly likely you will trip over yourself,<mask> your "muscle memory" for a task in<mask> developed<mask> a 1 year olds first try at walking. [NEWLINE] [NEWLINE] Add in the surprise(I'm assuming you don't know an attack dog is going to jump at you minutes in advance) and the blood lost and it all goes downhill. </s>
Label encoding: <s>Can you do a perfect cartwheel? [NEWLINE] [NEWLINE] Can you do a perfect cartwheel your first try? [NEWLINE] [NEWLINE] Can you do a perfect cartwheel while losing blood with adrenalin, fear and  without any time to mentally prepare for doing one? [NEWLINE] [NEWLINE] --------- [NEWLINE] [NEWLINE] Anyone can see themselves doing a cartwheel so long as they can walk, and its not like the laws of nature of how to balance are hard to figure out; however your first try will most likely be sloppy, even if you give your subconscious plenty of time take estimates about what it will need to do to balance. [NEWLINE] [NEWLINE] Now going into a fight without training, will make you sloppy *at best* its fairly likely you will trip over yourself, as your "muscle memory" for a task in as developed as a 1 year olds first try at walking. [NEWLINE] [NEWLINE] Add in the surprise(I'm assuming you don't know an attack dog is going to jump at you minutes in advance) and the blood lost and it all goes downhill. </s>
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Point of order: I want to be sure you know<mask> a bidet works. It sounds like you might be getting the wrong picture. For one thing, a bidet doesn't (or doesn't have to) "shoot water up your ass." A lot of bidets are just like a faucet in a sink. You run water on your hand, get some soap on there, soap up your ass, then scoop water with your hand to rinse out the shitty soap water, rinse and repeat until the streets are all clean downtown.<mask> it's installed in a home or hotel room, it likely has a hot water hookup<mask> you can avoid the dreaded cold water sphincter spasm. [NEWLINE] [NEWLINE] <mask> you're imagining something like a little drinking fountain you have to squat on, you've got the wrong idea. Using a bidet is more like washing your ass in the sink, only it's conveniently located closer to the ground (and you aren't going to brush your teeth in it later). </s><pad>
Label encoding: <s>Point of order: I want to be sure you know how a bidet works. It sounds like you might be getting the wrong picture. For one thing, a bidet doesn't (or doesn't have to) "shoot water up your ass." A lot of bidets are just like a faucet in a sink. You run water on your hand, get some soap on there, soap up your ass, then scoop water with your hand to rinse out the shitty soap water, rinse and repeat until the streets are all clean downtown. If it's installed in a home or hotel room, it likely has a hot water hookup so you can avoid the dreaded cold water sphincter spasm. [NEWLINE] [NEWLINE] If you're imagining something like a little drinking fountain you have to squat on, you've got the wrong idea. Using a bidet is more like washing your ass in the sink, only it's conveniently located closer to the ground (and you aren't going to brush your teeth in it later). </s><pad>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0821, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've done both the tactical too-much-field-time thing and the strategic am-I-even-in-the-military-anymore thing, which is precisely<mask> I know the military experience is<mask> inconsistent. I've never had a civilian job<mask> easy<mask> my easier years in the military, nor one<mask> hard<mask> my more difficult years. [NEWLINE] [NEWLINE] Some people don't have the luxury of seeing both sides of that coin and spent their entire career on one side or the other. The ones that had the good life and went from a cushy military job to a slightly less cushy civilian one you'll never hear a peep from,<mask> you hear no end of it from those who had to slog through misery for 4+ years straight. That, and popular media, is precisely<mask> the public perception of the military is that it always consists of stressful 12+ hour days of foot patrols<mask><mask> that experience is no more common than the military 9-5 office job with free coffee and snacks.</s>
Label encoding: <s>I've done both the tactical too-much-field-time thing and the strategic am-I-even-in-the-military-anymore thing, which is precisely why I know the military experience is so inconsistent. I've never had a civilian job as easy as my easier years in the military, nor one as hard as my more difficult years. [NEWLINE] [NEWLINE] Some people don't have the luxury of seeing both sides of that coin and spent their entire career on one side or the other. The ones that had the good life and went from a cushy military job to a slightly less cushy civilian one you'll never hear a peep from, but you hear no end of it from those who had to slog through misery for 4+ years straight. That, and popular media, is precisely why the public perception of the military is that it always consists of stressful 12+ hour days of foot patrols even though that experience is no more common than the military 9-5 office job with free coffee and snacks.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Different disorders require different treatment.<mask> others have said, it's easier and/or more effective to fix the body than fix the brain in this case. [NEWLINE] [NEWLINE] Example: I am/was depressed. I was on a bunch of different pills and having regular therapist appointments for a<mask> to try to deal with it<mask> none of it really helped (one of the pills worked well for a<mask> except that it knocked me out<mask> soon<mask> the sun went down and slowed my metabolism to the point I gained 30 lbs in 6 months. Both made it less comfortable for me to maintain the few social contacts I did have which ended up hurting emotionally more than the pill was helping).<mask> finally acted<mask> a true solution was getting into a situation<mask> I had control over my own life instead of having to live the life my family forced on me. This caused a mindset change that made the depression controllable. In this case it was more effective to change the environment than it was to try to medicate and change my brain.</s>
Label encoding: <s>Different disorders require different treatment. As others have said, it's easier and/or more effective to fix the body than fix the brain in this case. [NEWLINE] [NEWLINE] Example: I am/was depressed. I was on a bunch of different pills and having regular therapist appointments for a while to try to deal with it but none of it really helped (one of the pills worked well for a while except that it knocked me out as soon as the sun went down and slowed my metabolism to the point I gained 30 lbs in 6 months. Both made it less comfortable for me to maintain the few social contacts I did have which ended up hurting emotionally more than the pill was helping). What finally acted as a true solution was getting into a situation where I had control over my own life instead of having to live the life my family forced on me. This caused a mindset change that made the depression controllable. In this case it was more effective to change the environment than it was to try to medicate and change my brain.</s>
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But people are almost nothing like cells (except in a boring, biological sense that's not relevant here).  Cells don't have the kinds of traits that we think give an entity moral significance, which are usually things like sentience, or the capacity to suffer, or pleasures/pains, that sort of thing.  We<mask> have the capacity to make decisions and reflect on ourselves and our actions, none of which a solitary cell could do. <mask> I'd initially question the 'humans<mask> cells' comparison. [NEWLINE] [NEWLINE] Beyond that, even<mask> it did work, it seems like you're kind of approaching the Naturalistic Fallacy, in that you're going from "such-and-<mask> thing happens in nature" to "we *ought* to do such-and-<mask> thing. [NEWLINE] [NEWLINE] Finally,<mask> people<mask> individuals aren't important,<mask> the society is, does this not cause conflict due to society being made up of individuals? You'd have to value and not-value individuals.  </s>
Label encoding: <s>But people are almost nothing like cells (except in a boring, biological sense that's not relevant here).  Cells don't have the kinds of traits that we think give an entity moral significance, which are usually things like sentience, or the capacity to suffer, or pleasures/pains, that sort of thing.  We also have the capacity to make decisions and reflect on ourselves and our actions, none of which a solitary cell could do.  So I'd initially question the 'humans as cells' comparison. [NEWLINE] [NEWLINE] Beyond that, even if it did work, it seems like you're kind of approaching the Naturalistic Fallacy, in that you're going from "such-and- so thing happens in nature" to "we *ought* to do such-and- so thing. [NEWLINE] [NEWLINE] Finally, if people as individuals aren't important, but the society is, does this not cause conflict due to society being made up of individuals? You'd have to value and not-value individuals.  </s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's funny that none of the rebuttals here are refuting the holes in the story. They are all just questioning the motive for the government to either perpetuate the attacks or allow the attacks to happen. The insider trading is interesting,<mask> I bet it is more a symptom than the reason. Someone in the know figured they would quietly make some money on the whole ordeal. Simple. The state needed a BIG attack that they could convince the public was perpetrated by terrorists that live in caves. I don't see any issue with motive. States do expensive, destructive, evil things every day. [NEWLINE] [NEWLINE] <mask><mask><mask> people saying there is no way they could have kept it secret, I'd point to the Snowden revelations. These are highly secret spy programs that have existed for decades, involve thousands of people, and span through international governments. We would still be oblivious<mask> it wasn't for Snowden. Sure, the conspiracy theorists pieced together many things that he revealed,<mask> without hard evidence, few normal people were convinced. </s>
Label encoding: <s>It's funny that none of the rebuttals here are refuting the holes in the story. They are all just questioning the motive for the government to either perpetuate the attacks or allow the attacks to happen. The insider trading is interesting, but I bet it is more a symptom than the reason. Someone in the know figured they would quietly make some money on the whole ordeal. Simple. The state needed a BIG attack that they could convince the public was perpetrated by terrorists that live in caves. I don't see any issue with motive. States do expensive, destructive, evil things every day. [NEWLINE] [NEWLINE] As far as people saying there is no way they could have kept it secret, I'd point to the Snowden revelations. These are highly secret spy programs that have existed for decades, involve thousands of people, and span through international governments. We would still be oblivious if it wasn't for Snowden. Sure, the conspiracy theorists pieced together many things that he revealed, but without hard evidence, few normal people were convinced. </s>
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> happens outside of the workplace is no concern of the employer [ENDQ] [NEWLINE] The employer cares about employee performance surely.  Were it the case that vacations are shown to profitably improve productivity, then surely employers would want their employees taking vacations right? [NEWLINE] [NEWLINE] <mask> it were then shown that the most cost effective way to encourage employees to take vacations, were to offer paid vacations, then that would be the go to strategy right? [NEWLINE] [NEWLINE] Then there is the matter of hiring, attracting the best employees.  This is a big deal in my industry,<mask> anyone (who is really good) can work remote<mask> traveling the world.  There have to be some good benefits<mask> you want (the best of) us 9 to 5. [NEWLINE] [NEWLINE] [STARTQ] <mask> should the government 'force' employers to pay for their employees to sit on a beach [ENDQ] [NEWLINE] the government, in this case, is representing the employees. [NEWLINE] [NEWLINE] <mask> you feel that people have the right to collectively bargain,<mask> is wrong with collectively bargaining through government?</s>
Label encoding: <s> [STARTQ] what happens outside of the workplace is no concern of the employer [ENDQ] [NEWLINE] The employer cares about employee performance surely.  Were it the case that vacations are shown to profitably improve productivity, then surely employers would want their employees taking vacations right? [NEWLINE] [NEWLINE] If it were then shown that the most cost effective way to encourage employees to take vacations, were to offer paid vacations, then that would be the go to strategy right? [NEWLINE] [NEWLINE] Then there is the matter of hiring, attracting the best employees.  This is a big deal in my industry, because anyone (who is really good) can work remote while traveling the world.  There have to be some good benefits if you want (the best of) us 9 to 5. [NEWLINE] [NEWLINE] [STARTQ] Why should the government 'force' employers to pay for their employees to sit on a beach [ENDQ] [NEWLINE] the government, in this case, is representing the employees. [NEWLINE] [NEWLINE] if you feel that people have the right to collectively bargain, what is wrong with collectively bargaining through government?</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Be careful with this way of thinking.  Turkeys can talk to each other, they have a rather large vocabulary<mask><mask>.  Pigs are considered VERY smart and CERTAINLY can express and achieve emotional states (<mask> you already mention that you don't eat pigs, kudos).  Plus, I would worry that without the ability to communicate more directly with animals, we are guessing pretty largely about<mask> they are capable of mentally. [NEWLINE] [NEWLINE] Read up on Prairie Dogs and their SUPER complex language system. [NEWLINE] [NEWLINE] Personally<mask><mask> we are treading a dangerous line to pick and chose<mask> life is valuable and<mask> ones are not. <mask><mask> we lack the information and data to make that decision well, and plus I don't agree that life below a certain threshold is discardable.  It's not like we<mask> a species NEED to eat meat to stay alive or even happy and healthy nor do we NEED to eat animals to have enough food to feed the entire planet (<mask><mask> animals hurt that equation).</s>
Label encoding: <s>Be careful with this way of thinking.  Turkeys can talk to each other, they have a rather large vocabulary in fact.  Pigs are considered VERY smart and CERTAINLY can express and achieve emotional states ( but you already mention that you don't eat pigs, kudos).  Plus, I would worry that without the ability to communicate more directly with animals, we are guessing pretty largely about what they are capable of mentally. [NEWLINE] [NEWLINE] Read up on Prairie Dogs and their SUPER complex language system. [NEWLINE] [NEWLINE] Personally I think we are treading a dangerous line to pick and chose what life is valuable and what ones are not.  I think we lack the information and data to make that decision well, and plus I don't agree that life below a certain threshold is discardable.  It's not like we as a species NEED to eat meat to stay alive or even happy and healthy nor do we NEED to eat animals to have enough food to feed the entire planet ( in fact animals hurt that equation).</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You make a very strong argument and,<mask> my view hasn't been completely changed, you have opened my eyes up to situations<mask> reporting sexual assault would not necessarily be beneficial to either party. I'm sorry you had to go through<mask> you did, and I'm glad you were able to deal with it in a way that made you feel more comfortable. [NEWLINE] [NEWLINE] Sexual assault does happen to men, which is the reason<mask> I made my post<mask> gender neutral<mask> possible. I hate that it gets ignored and made light of,<mask> it's an incredibly traumatic event, and no one should feel like they won't be taken seriously<mask> they report it. It's a serious problem that we,<mask> a society, need to take steps to address. Men being sexually assaulted by women is just<mask> damaging to the victim<mask> women being assaulted by men, and the fact that it's seen<mask> otherwise is just wrong. [NEWLINE] [NEWLINE] You have a delta coming your way, I just need to get to a computer first. </s>
Label encoding: <s>You make a very strong argument and, while my view hasn't been completely changed, you have opened my eyes up to situations where reporting sexual assault would not necessarily be beneficial to either party. I'm sorry you had to go through what you did, and I'm glad you were able to deal with it in a way that made you feel more comfortable. [NEWLINE] [NEWLINE] Sexual assault does happen to men, which is the reason why I made my post as gender neutral as possible. I hate that it gets ignored and made light of, because it's an incredibly traumatic event, and no one should feel like they won't be taken seriously if they report it. It's a serious problem that we, as a society, need to take steps to address. Men being sexually assaulted by women is just as damaging to the victim as women being assaulted by men, and the fact that it's seen as otherwise is just wrong. [NEWLINE] [NEWLINE] You have a delta coming your way, I just need to get to a computer first. </s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If Pablo makes more money than his competitor, he probably<mask> 1. spends more (consumes more) than his competitor, and [NEWLINE] 2. owns more capital than his competitor. [NEWLINE] You are right that Pablo shouldn't have to pay more taxes just<mask> he earns more money,<mask><mask><mask> he should have to pay more taxes<mask> he 1. spends more and 2. owns more. Our income tax system is imperfect<mask> it does this indirectly,<mask> it's better than nothing. [NEWLINE] [NEWLINE] 1. Pablo should be taxed more<mask> he is spending more<mask> he is using more of our resources than his competitor. [NEWLINE] [NEWLINE] 2. Pablo should be taxed more<mask> he owns more<mask><mask> you own money or land, you can make more money with basically zero effort or contribution to society. E.g.<mask> you have a million dollars and put it in an index fund, you will make enough money in interest to live on for the rest of your life. Other people essentially become your slaves. Taxes counteract this.</s>
Label encoding: <s>If Pablo makes more money than his competitor, he probably also 1. spends more (consumes more) than his competitor, and [NEWLINE] 2. owns more capital than his competitor. [NEWLINE] You are right that Pablo shouldn't have to pay more taxes just because he earns more money, but I think he should have to pay more taxes if he 1. spends more and 2. owns more. Our income tax system is imperfect because it does this indirectly, but it's better than nothing. [NEWLINE] [NEWLINE] 1. Pablo should be taxed more if he is spending more because he is using more of our resources than his competitor. [NEWLINE] [NEWLINE] 2. Pablo should be taxed more if he owns more because if you own money or land, you can make more money with basically zero effort or contribution to society. E.g. If you have a million dollars and put it in an index fund, you will make enough money in interest to live on for the rest of your life. Other people essentially become your slaves. Taxes counteract this.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hello.<mask><mask> you need to clarify which laws specifically have proved to be damaging. [NEWLINE] [NEWLINE] Yes, the Constitution is difficult to amend,<mask> I see this<mask> a good thing.<mask> free speech and the 4th, 5th amendments could be curtailed in a willy-nilly Senate vote or whatnot, I guarantee you we would have lost a great deal of our rights in and around the 9/11 tragedy<mask> our government legislated the Patriot Act. [NEWLINE] [NEWLINE] It would have been *much easier* for them to legally pass a bill that said "all of our emails are no longer private", or that government officials could raid "any home" in the name of national security.<mask> the Constitution - and its difficulty to change - continues to protect against that. [NEWLINE] [NEWLINE] And<mask> another example,<mask> we do have something that absolutely needs to get changed, and everyone agrees, it gets changed. Slavery is one example of a change. Prohibition (and its repeal) is another). </s>
Label encoding: <s>Hello. I think you need to clarify which laws specifically have proved to be damaging. [NEWLINE] [NEWLINE] Yes, the Constitution is difficult to amend, but I see this as a good thing. If free speech and the 4th, 5th amendments could be curtailed in a willy-nilly Senate vote or whatnot, I guarantee you we would have lost a great deal of our rights in and around the 9/11 tragedy when our government legislated the Patriot Act. [NEWLINE] [NEWLINE] It would have been *much easier* for them to legally pass a bill that said "all of our emails are no longer private", or that government officials could raid "any home" in the name of national security. But the Constitution - and its difficulty to change - continues to protect against that. [NEWLINE] [NEWLINE] And as another example, when we do have something that absolutely needs to get changed, and everyone agrees, it gets changed. Slavery is one example of a change. Prohibition (and its repeal) is another). </s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was reading the other day that even Snowden actually isn't asking for a pardon to return to the USA. I mean, I am sure he would take one,<mask> that isn't<mask> he is asking for. [NEWLINE] [NEWLINE] Whistleblower laws in the US allow for people to break the law by revealing government information, and avoid punishment<mask> they can show they did<mask> for the purpose of the public good. The problem for Snowden is that these laws explicitly exclude intelligence agencies, and that under the Espionage Act he is charged with, he cannot make a public good defense. [NEWLINE] [NEWLINE] Snowden has said repeatedly that he would come back and face trial<mask> he were allowed to defend himself<mask> a whistleblower in court.<mask> the government doesn't even need to pardon him to get him to return from Russia - simply charge him under something less draconian and anachronistic than the Espionage Act, and let him make his defense. (Presumably the case would look a lot like the debate in this thread.)</s>
Label encoding: <s>I was reading the other day that even Snowden actually isn't asking for a pardon to return to the USA. I mean, I am sure he would take one, but that isn't what he is asking for. [NEWLINE] [NEWLINE] Whistleblower laws in the US allow for people to break the law by revealing government information, and avoid punishment if they can show they did so for the purpose of the public good. The problem for Snowden is that these laws explicitly exclude intelligence agencies, and that under the Espionage Act he is charged with, he cannot make a public good defense. [NEWLINE] [NEWLINE] Snowden has said repeatedly that he would come back and face trial if he were allowed to defend himself as a whistleblower in court. So the government doesn't even need to pardon him to get him to return from Russia - simply charge him under something less draconian and anachronistic than the Espionage Act, and let him make his defense. (Presumably the case would look a lot like the debate in this thread.)</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of those successful recruitment of people outside their usual pool of candidates,<mask> often do you think the terrorist organization had to exposes themselves? [NEWLINE] [NEWLINE] There is only a few ways to catch a bad guy during their cycle of crime. [NEWLINE] [NEWLINE] You can't catch them<mask> their are planning in the safety of their own hideouts. [NEWLINE] [NEWLINE] You can't catch them<mask> they are scouting their location (they aren't doing anything wrong.) [NEWLINE] [NEWLINE] You can catch them<mask> they recruit help or purchase tools (depending on the tools.) [NEWLINE] [NEWLINE] You can't catch them<mask> they practice on their own. [NEWLINE] [NEWLINE] You can catch them<mask> they are executing their crime. [NEWLINE] [NEWLINE] Forcing your enemies to change their effective tactics to a less effective one seems like a usual crime-fighting technique to me. [NEWLINE] [NEWLINE] <mask> the statistic of Arab Muslim American involved in terrorism began to drop significant,<mask> White Muslim American suspect began to rise, it would<mask> be reasonable to switch the racial profile to adjust for the change. [NEWLINE] </s>
Label encoding: <s>Of those successful recruitment of people outside their usual pool of candidates, how often do you think the terrorist organization had to exposes themselves? [NEWLINE] [NEWLINE] There is only a few ways to catch a bad guy during their cycle of crime. [NEWLINE] [NEWLINE] You can't catch them while their are planning in the safety of their own hideouts. [NEWLINE] [NEWLINE] You can't catch them while they are scouting their location (they aren't doing anything wrong.) [NEWLINE] [NEWLINE] You can catch them when they recruit help or purchase tools (depending on the tools.) [NEWLINE] [NEWLINE] You can't catch them when they practice on their own. [NEWLINE] [NEWLINE] You can catch them when they are executing their crime. [NEWLINE] [NEWLINE] Forcing your enemies to change their effective tactics to a less effective one seems like a usual crime-fighting technique to me. [NEWLINE] [NEWLINE] If the statistic of Arab Muslim American involved in terrorism began to drop significant, while White Muslim American suspect began to rise, it would also be reasonable to switch the racial profile to adjust for the change. [NEWLINE] </s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Those movements are completely incomparable and its a poor argument to put forth. [NEWLINE] [NEWLINE] Have a look at that first link. That is a sound and completely reasonable concept. [NEWLINE] [NEWLINE] Oh no, the texts available around the subject aren't in-line with<mask> we think is ok. Have a look back through the history of Feminism. Some of the texts that have been used to explore the concepts are abominable. This is very important. Mein Kampf is required reading in all sorts of courses. Not<mask> they want you to follow the philosophy<mask> to consider it and form your own opinion which is<mask> that text served to do for me<mask> I read it. Both groups have toxic elements and some really valuable thinking<mask> well. [NEWLINE] [NEWLINE] The hypocrisy of people is<mask> shits me up the wall. Mysandry and mysoginy are both very real and exist within all gender equality groups. Maintain the same standard for both or you don't help at all. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Those movements are completely incomparable and its a poor argument to put forth. [NEWLINE] [NEWLINE] Have a look at that first link. That is a sound and completely reasonable concept. [NEWLINE] [NEWLINE] Oh no, the texts available around the subject aren't in-line with what we think is ok. Have a look back through the history of Feminism. Some of the texts that have been used to explore the concepts are abominable. This is very important. Mein Kampf is required reading in all sorts of courses. Not because they want you to follow the philosophy but to consider it and form your own opinion which is what that text served to do for me when I read it. Both groups have toxic elements and some really valuable thinking as well. [NEWLINE] [NEWLINE] The hypocrisy of people is what shits me up the wall. Mysandry and mysoginy are both very real and exist within all gender equality groups. Maintain the same standard for both or you don't help at all. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I completely disagree with you that the Constitution is too difficult to amend. Asking for 2/3 legislative approval and ratification by 3/4 of the states seems like a minimum for making changes to a text that defines your nation. [NEWLINE] [NEWLINE] Imagine<mask> altering the constitution simply required a legislative majority. The US virtually being a dichotomy, this means that at any time, whichever party controlled Congress could essentially modify the Constitution<mask> they see fit. Instead of having had 17 amendments in 200 years (the Bill of Rights is basically an addendum) we would have 17 a month... [NEWLINE] [NEWLINE] Edit: Just wanted to add, after reading some of your answers, it seems to me that you are making the fallacy of<mask><mask> the constitution is difficult to amend<mask> some of the changes you personnally think should happen are not happening. To give a concrete example, you talk about gun control,<mask> [polls]( [URL] /) show that the majority of Americans do not want the government to curtail gun ownership. </s>
Label encoding: <s>I completely disagree with you that the Constitution is too difficult to amend. Asking for 2/3 legislative approval and ratification by 3/4 of the states seems like a minimum for making changes to a text that defines your nation. [NEWLINE] [NEWLINE] Imagine if altering the constitution simply required a legislative majority. The US virtually being a dichotomy, this means that at any time, whichever party controlled Congress could essentially modify the Constitution however they see fit. Instead of having had 17 amendments in 200 years (the Bill of Rights is basically an addendum) we would have 17 a month... [NEWLINE] [NEWLINE] Edit: Just wanted to add, after reading some of your answers, it seems to me that you are making the fallacy of assuming that the constitution is difficult to amend because some of the changes you personnally think should happen are not happening. To give a concrete example, you talk about gun control, but [polls]( [URL] /) show that the majority of Americans do not want the government to curtail gun ownership. </s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When I draw a smiley face, I draw the eyes first.<mask>,<mask> I type one, I'll type the eyes first. [NEWLINE] [NEWLINE] It definitely doesn't interfere with my punctuation. :) [NEWLINE] [NEWLINE] And I honestly never thought of typing (: due to the parenthesis being higher than the colon on the keyboard. That's a bit over analyzing, and<mask>, we type many words that aren't flowy, from top to bottom or left to right,<mask> I'll type them anyway. I refuse to type ppl<mask><mask> it's easier than people. [NEWLINE] [NEWLINE] In other honestly<mask>, you're probably right that :) is right handed and (: is left handed.<mask> you said right handers tilt their head to the right,<mask><mask><mask>, you have to tilt your head to the left to read :). [NEWLINE] [NEWLINE] It's like<mask> you draw a stick figure, you'll go from the head down, not feet up. Same with typing a smiley face.</s>
Label encoding: <s>When I draw a smiley face, I draw the eyes first. Hence, when I type one, I'll type the eyes first. [NEWLINE] [NEWLINE] It definitely doesn't interfere with my punctuation. :) [NEWLINE] [NEWLINE] And I honestly never thought of typing (: due to the parenthesis being higher than the colon on the keyboard. That's a bit over analyzing, and besides, we type many words that aren't flowy, from top to bottom or left to right, but I'll type them anyway. I refuse to type ppl even though it's easier than people. [NEWLINE] [NEWLINE] In other honestly though, you're probably right that :) is right handed and (: is left handed. But you said right handers tilt their head to the right, but in fact, you have to tilt your head to the left to read :). [NEWLINE] [NEWLINE] It's like when you draw a stick figure, you'll go from the head down, not feet up. Same with typing a smiley face.</s>
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] nobody should ever have to go out that way<mask> they wish not to [ENDQ] [NEWLINE] Agreed,<mask> that doesn't make suicide a dignified act. There are plenty of things everyone should be allowed to choose that are nonetheless undignified. [NEWLINE] [NEWLINE] [STARTQ] nobody should ever say that someone who makes that decision wasn't dignified in doing<mask> or was "taking the easy way out" by not "enduring pain to make them stronger". That's just fucked up. [ENDQ] [NEWLINE] <mask> is that wrong? It might be fucked up,<mask> sometimes the truth is fucked up. Overcoming hardship is the definition of dignity. [NEWLINE] [NEWLINE] [STARTQ] Maybe it's<mask> those people were still happy, functional human beings until the moment they had the relative luxury of dying quickly. Don't you think the terminally ill deserve that same luxury? [ENDQ] [NEWLINE] <mask><mask> this statement is underpinned by the incorrect assumption that dying quickly somehow preserves who you are. Death destroys everything; a quick death just destroys faster.</s><pad>
Label encoding: <s> [STARTQ] nobody should ever have to go out that way if they wish not to [ENDQ] [NEWLINE] Agreed, but that doesn't make suicide a dignified act. There are plenty of things everyone should be allowed to choose that are nonetheless undignified. [NEWLINE] [NEWLINE] [STARTQ] nobody should ever say that someone who makes that decision wasn't dignified in doing so or was "taking the easy way out" by not "enduring pain to make them stronger". That's just fucked up. [ENDQ] [NEWLINE] Why is that wrong? It might be fucked up, but sometimes the truth is fucked up. Overcoming hardship is the definition of dignity. [NEWLINE] [NEWLINE] [STARTQ] Maybe it's because those people were still happy, functional human beings until the moment they had the relative luxury of dying quickly. Don't you think the terminally ill deserve that same luxury? [ENDQ] [NEWLINE] I think this statement is underpinned by the incorrect assumption that dying quickly somehow preserves who you are. Death destroys everything; a quick death just destroys faster.</s><pad>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All that means is that Christianity adopted a more clever business model. Scientology has managed to accrue a couple billion dollars largely supported by a small group of wealthy members. [NEWLINE] [NEWLINE] Do you know<mask> much the Catholic Church received in revenue last year? Well no one does<mask> they don't have to disclose it.<mask><mask> we do know is that the church's operating budget, just in the US, was over 150 *billion* dollars. The net worth of its total assets exceeds that of the richest corporation on earth **by an order of magnitude**. [NEWLINE] [NEWLINE] And they managed to get themselves their own sovereign country. [NEWLINE] [NEWLINE] <mask> I'd say the only real difference between the two churches is their business model and the fact that the catholics discovered that a pay-<mask> -you-want system is considerably more effective. [NEWLINE] [NEWLINE] <mask> scientology shifted business models tomorrow, they'd avoid all accusations of fraud and<mask> they managed to survive<mask> more than a trend could have their own nation in a century</s>
Label encoding: <s>All that means is that Christianity adopted a more clever business model. Scientology has managed to accrue a couple billion dollars largely supported by a small group of wealthy members. [NEWLINE] [NEWLINE] Do you know how much the Catholic Church received in revenue last year? Well no one does because they don't have to disclose it. But what we do know is that the church's operating budget, just in the US, was over 150 *billion* dollars. The net worth of its total assets exceeds that of the richest corporation on earth **by an order of magnitude**. [NEWLINE] [NEWLINE] And they managed to get themselves their own sovereign country. [NEWLINE] [NEWLINE] So I'd say the only real difference between the two churches is their business model and the fact that the catholics discovered that a pay- what -you-want system is considerably more effective. [NEWLINE] [NEWLINE] If scientology shifted business models tomorrow, they'd avoid all accusations of fraud and if they managed to survive as more than a trend could have their own nation in a century</s>
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0659, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> that<mask> you get all your clothing or electronics from sources that don't contribute to sweatshops or child labor, you earn the right to try to convince me to do the same, including calling me a hypocrite for saying I'm a "people person"<mask> buying that stuff. I admit that, and I am trying to slowly change it by not buying products from certain companies. [NEWLINE] [NEWLINE] <mask><mask> you truly think that any animal suffering is no big deal, then by all means, keep contributing to those industries.<mask> nowadays in our society it's insanely easy to reduce/remove animal products from your diet without feeling like you're eating cardboard and suffering emotionally<mask> of it. [NEWLINE] [NEWLINE] I think everyone needs to be more open to criticism of their own lifestyle<mask> they can step back and see things they haven't been shown or didn't think about before (including myself). Otherwise we all just do the same shit our ancestors did without really stopping to think<mask> there's a better way nowadays.</s>
Label encoding: <s>I think that if you get all your clothing or electronics from sources that don't contribute to sweatshops or child labor, you earn the right to try to convince me to do the same, including calling me a hypocrite for saying I'm a "people person" yet buying that stuff. I admit that, and I am trying to slowly change it by not buying products from certain companies. [NEWLINE] [NEWLINE] So if you truly think that any animal suffering is no big deal, then by all means, keep contributing to those industries. But nowadays in our society it's insanely easy to reduce/remove animal products from your diet without feeling like you're eating cardboard and suffering emotionally because of it. [NEWLINE] [NEWLINE] I think everyone needs to be more open to criticism of their own lifestyle so they can step back and see things they haven't been shown or didn't think about before (including myself). Otherwise we all just do the same shit our ancestors did without really stopping to think if there's a better way nowadays.</s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are ways to vote strategically in range/approval,<mask><mask> you assume that people have some preference between any pair of candidates,<mask> they're much more natural seeming than the ways you vote strategically in IRV, and<mask> much less potentially helpful. [NEWLINE] [NEWLINE] It basically has to do with<mask> you set your cutoff.<mask> your preferences are A [STARTQ] B&gt;C&gt;D, then you could vote for A, or A and B, or A and B and C.<mask> you want to maximize the chance that your ideal candidate wins you should vote for only A.<mask> this is risky,<mask> it could<mask> lead to A and B losing to C or D instead of A losing and B winning. [ENDQ] [NEWLINE] You can<mask> do the equivalent in range voting: vote maximum for A and minimum for all other candidates to maximize the chance that A wins.<mask> again, this is risky,<mask> depending on<mask> other people vote this could lead to a less preferred candidate winning.</s>
Label encoding: <s>There are ways to vote strategically in range/approval, given that you assume that people have some preference between any pair of candidates, but they're much more natural seeming than the ways you vote strategically in IRV, and also much less potentially helpful. [NEWLINE] [NEWLINE] It basically has to do with where you set your cutoff. If your preferences are A [STARTQ] B&gt;C&gt;D, then you could vote for A, or A and B, or A and B and C. If you want to maximize the chance that your ideal candidate wins you should vote for only A. But this is risky, because it could also lead to A and B losing to C or D instead of A losing and B winning. [ENDQ] [NEWLINE] You can also do the equivalent in range voting: vote maximum for A and minimum for all other candidates to maximize the chance that A wins. But again, this is risky, because depending on how other people vote this could lead to a less preferred candidate winning.</s>
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fair point (<mask> a weapon is used -- a dull butter knife can be deadly). [NEWLINE] [NEWLINE] <mask><mask><mask><mask><mask> about "use of force rarely escalating situations" --<mask><mask><mask> we understand "escalation" might a little narrow. [NEWLINE] [NEWLINE] Cops showing up, sirens blaring, that can be enough.  Cops not trained on<mask> to talk to a suspect/perp/agitated individual,<mask> a problem. [NEWLINE] [NEWLINE] Better training is clearly a key component --<mask>, for example, the Sammy Yatim case in Toronto.  The kid approached a guy on the streetcar, and all he had was a BIKE with him, and he said he didn't feel threatened.  A cop appears on the scene, and the kid is shot within SECONDS. <mask><mask>,<mask> you have a gun, you might just use it, and then someone who would otherwise be alive -- and<mask> properly medicated, even thriving -- is gone. [NEWLINE] </s>
Label encoding: <s>Fair point ( how a weapon is used -- a dull butter knife can be deadly). [NEWLINE] [NEWLINE] But I think I disagree about "use of force rarely escalating situations" -- I think how we understand "escalation" might a little narrow. [NEWLINE] [NEWLINE] Cops showing up, sirens blaring, that can be enough.  Cops not trained on how to talk to a suspect/perp/agitated individual, also a problem. [NEWLINE] [NEWLINE] Better training is clearly a key component -- but, for example, the Sammy Yatim case in Toronto.  The kid approached a guy on the streetcar, and all he had was a BIKE with him, and he said he didn't feel threatened.  A cop appears on the scene, and the kid is shot within SECONDS.  I think, if you have a gun, you might just use it, and then someone who would otherwise be alive -- and if properly medicated, even thriving -- is gone. [NEWLINE] </s>
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0963, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0585, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm going to attack this from a different angle than others are. [NEWLINE] [NEWLINE] (Assuming the United States here) [NEWLINE] [NEWLINE] The punishment you proscribe is far too draconian. In most places in the US it is extremely difficult or nearly impossible to live without access to an automobile. Even in many large cities, public transportation is<mask>fy at best and if you want to go anywhere beyond the city limit, you'll need to be able to drive there. [NEWLINE] [NEWLINE] <mask> this punishment basically does is make it impossible for people convicted of DUIs to live in the country or the suburbs. Many will likely be unable to commute to work, take their kids to school, or go to the store to buy groceries without the help of someone else to drive them. This burden is simply too much for most people to bear and it would cause undue harm to many innocent people, namely the families of convicts who may have to uproot and relocate to a place<mask> the convict can function autonomously.</s><pad>
Label encoding: <s>I'm going to attack this from a different angle than others are. [NEWLINE] [NEWLINE] (Assuming the United States here) [NEWLINE] [NEWLINE] The punishment you proscribe is far too draconian. In most places in the US it is extremely difficult or nearly impossible to live without access to an automobile. Even in many large cities, public transportation is iffy at best and if you want to go anywhere beyond the city limit, you'll need to be able to drive there. [NEWLINE] [NEWLINE] What this punishment basically does is make it impossible for people convicted of DUIs to live in the country or the suburbs. Many will likely be unable to commute to work, take their kids to school, or go to the store to buy groceries without the help of someone else to drive them. This burden is simply too much for most people to bear and it would cause undue harm to many innocent people, namely the families of convicts who may have to uproot and relocate to a place where the convict can function autonomously.</s><pad>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A lot of people seemed to be of the opinion that having a working phone with you<mask> driving is just too dangerous. The phone should be put away and turned off to avoid temptation and anyone seen with their phone behind the wheel of a vehicle should be fined/arrested/whatever. [ENDQ] [NEWLINE] You answered your own question.<mask> you are sitting in a parking lot then technically you are not driving,<mask> that is not a violation.<mask>,<mask> you are in traffic or at a light and you get on your phone you are still excepted to drive. There are other cars around you in motion, and shortly you will be in motion.<mask><mask> the light turns green in the middle of your text? Odds are you will roll forward<mask> you continue to type it. [NEWLINE] [NEWLINE] <mask>,<mask><mask> it's kind of pathetic that we are<mask> attached to our phones that we need to check them<mask> waiting at a red light. Do you get that bored that fast?</s>
Label encoding: <s> [STARTQ] A lot of people seemed to be of the opinion that having a working phone with you while driving is just too dangerous. The phone should be put away and turned off to avoid temptation and anyone seen with their phone behind the wheel of a vehicle should be fined/arrested/whatever. [ENDQ] [NEWLINE] You answered your own question. If you are sitting in a parking lot then technically you are not driving, so that is not a violation. However, if you are in traffic or at a light and you get on your phone you are still excepted to drive. There are other cars around you in motion, and shortly you will be in motion. What if the light turns green in the middle of your text? Odds are you will roll forward while you continue to type it. [NEWLINE] [NEWLINE] Also, I think it's kind of pathetic that we are so attached to our phones that we need to check them while waiting at a red light. Do you get that bored that fast?</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Homosexuality doesn't present any threat to our species genetically or otherwise. Incest,<mask>, does pose a genetic threat and this is a generally known fact.<mask> you're in an incestuous relationship and you decide to have kids, you're putting your kids at risk, just<mask><mask> you smoked or drank<mask> pregnant. Nature doesn't favor incest; people are generally hardwired to seek out mates whose genetics compensate for their genetic shortcomings and vice versa. It's actually pretty neat:<mask> a guy smells like a girl's brother or father, she won't be attracted to him<mask> she's instinctively looking for a genetically different mate. I'm very briefly summarizing the phenomenon,<mask> the point is that people are naturally inclined to search for genetically different mates, i.e., not relatives. I'd<mask><mask> people who do pursue incestuous relationships are influenced more by nurture than nature.<mask> not only is incest unnatural, it's harmful, both in the short and long run.</s>
Label encoding: <s>Homosexuality doesn't present any threat to our species genetically or otherwise. Incest, however, does pose a genetic threat and this is a generally known fact. If you're in an incestuous relationship and you decide to have kids, you're putting your kids at risk, just as if you smoked or drank while pregnant. Nature doesn't favor incest; people are generally hardwired to seek out mates whose genetics compensate for their genetic shortcomings and vice versa. It's actually pretty neat: if a guy smells like a girl's brother or father, she won't be attracted to him because she's instinctively looking for a genetically different mate. I'm very briefly summarizing the phenomenon, but the point is that people are naturally inclined to search for genetically different mates, i.e., not relatives. I'd argue that people who do pursue incestuous relationships are influenced more by nurture than nature. So not only is incest unnatural, it's harmful, both in the short and long run.</s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But there is always a choice between pushing the technological envelope, and re-igniting interest in established forms. It would seem that your love is for long-form storytelling for the big screen. Any success you can hope to have will be in the area<mask> you have the most passion. [NEWLINE] [NEWLINE] I for one am not a filmmaker,<mask> long for the expansiveness and depth of films made in the 70's. There *is* a market out there for this. Give people more credit than the studios do. It might pay off. [NEWLINE] [NEWLINE] Edit: I guess<mask> I'm getting at is that big media companies are short-sightedly profit driven --<mask> much<mask> that they put out junk, and foreclose on still-viable media. They often realize this mistake<mask>, and supposedly dead forms sometimes re-emerge.<mask> your interest is in one of these "dying" forms, it would seem that your best chance is in this kind of revival.</s>
Label encoding: <s>But there is always a choice between pushing the technological envelope, and re-igniting interest in established forms. It would seem that your love is for long-form storytelling for the big screen. Any success you can hope to have will be in the area where you have the most passion. [NEWLINE] [NEWLINE] I for one am not a filmmaker, but long for the expansiveness and depth of films made in the 70's. There *is* a market out there for this. Give people more credit than the studios do. It might pay off. [NEWLINE] [NEWLINE] Edit: I guess what I'm getting at is that big media companies are short-sightedly profit driven -- so much so that they put out junk, and foreclose on still-viable media. They often realize this mistake though, and supposedly dead forms sometimes re-emerge. Because your interest is in one of these "dying" forms, it would seem that your best chance is in this kind of revival.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In a sense, you are right,<mask> low wage earners are the ones who produce (mostly) everything in the modern economy,<mask><mask><mask> you aren't taking some important things into consideration. First,<mask> CherrySlurpee pointed out, things need to be invented before they will be created.<mask> nobody was the millionaire who inventing anything we would still be using stone tools. On the flip side,<mask> nobody worked on an assembly line mass producing things then we wouldn't have any nice things either.<mask><mask><mask> I am tempted to say that they are both equally necessary<mask> without either we would not have any of the ammenities of modern life.<mask> changes this<mask>, is looking to the future.<mask> we see things like 3D printing and increasingly capable robots and automation, it is becoming easier and easier to imagine a world<mask> there is very little manual labor. We would still need to invent great things,<mask> could conceivably not need people to produce them.</s>
Label encoding: <s>In a sense, you are right, because low wage earners are the ones who produce (mostly) everything in the modern economy, but I think you aren't taking some important things into consideration. First, as CherrySlurpee pointed out, things need to be invented before they will be created. If nobody was the millionaire who inventing anything we would still be using stone tools. On the flip side, if nobody worked on an assembly line mass producing things then we wouldn't have any nice things either. Because of this I am tempted to say that they are both equally necessary since without either we would not have any of the ammenities of modern life. What changes this though, is looking to the future. As we see things like 3D printing and increasingly capable robots and automation, it is becoming easier and easier to imagine a world where there is very little manual labor. We would still need to invent great things, but could conceivably not need people to produce them.</s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The supreme court ruled that cops need either reasonable suspicion or probable cause to force someone to wait for a k9 unit, and<mask> is that scarring psycological warfare? [NEWLINE] [NEWLINE] Yes, the police aren't 100% perfect?  Should they be more careful, yes, are the small number of mistaken no knocks that have made the press evidence we live in a "police state"? [NEWLINE] [NEWLINE] <mask> you can prove<mask> the money came from you can get whatever was seized back.  Could you get your assets back in a police state? [NEWLINE] [NEWLINE] Actually in around 2008 the NYT broke a story that the NSA had spliced into the core internet fiber.  You can look up the hardware they were using to collect data from those splices and read the manuals.  iirc it was mostly metadata stuff. [NEWLINE] [NEWLINE] Well, yes, they look for pattern, like,<mask> often does X person trade emails with person Y on a watchlist?  That's hardly sinister.</s><pad>
Label encoding: <s>The supreme court ruled that cops need either reasonable suspicion or probable cause to force someone to wait for a k9 unit, and how is that scarring psycological warfare? [NEWLINE] [NEWLINE] Yes, the police aren't 100% perfect?  Should they be more careful, yes, are the small number of mistaken no knocks that have made the press evidence we live in a "police state"? [NEWLINE] [NEWLINE] If you can prove where the money came from you can get whatever was seized back.  Could you get your assets back in a police state? [NEWLINE] [NEWLINE] Actually in around 2008 the NYT broke a story that the NSA had spliced into the core internet fiber.  You can look up the hardware they were using to collect data from those splices and read the manuals.  iirc it was mostly metadata stuff. [NEWLINE] [NEWLINE] Well, yes, they look for pattern, like, how often does X person trade emails with person Y on a watchlist?  That's hardly sinister.</s><pad>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Experiencing things I had never experienced in real life by proxy through great literature actually prepared me mentally and emotionally to deal with the same problems I read about<mask> I encountered them later on in life. [ENDQ] [NEWLINE] Yeah, the proxy effect is probably the best rationale for having "inexperienced" people read these works. [NEWLINE] [NEWLINE] <mask>, the mode of delivery for these life lessons is usually very boring and difficult for a high schooler to get into--read the [first few paragraphs of Ivanhoe]( [URL] ) and tell me<mask> the *average* student wouldn't immediately get frustrated and go to the easily available Sparknotes. I feel that a lot of the older classics use a lot of dense and archaic language that is an immediate turn-off to modern teenagers. [NEWLINE] [NEWLINE] More recent works like Catcher in the Rye and the Crucible, I would agree<mask> they are<mask> accessible<mask> you're gonna get.<mask> are there enough "sexy" classics?</s>
Label encoding: <s> [STARTQ] Experiencing things I had never experienced in real life by proxy through great literature actually prepared me mentally and emotionally to deal with the same problems I read about as I encountered them later on in life. [ENDQ] [NEWLINE] Yeah, the proxy effect is probably the best rationale for having "inexperienced" people read these works. [NEWLINE] [NEWLINE] However, the mode of delivery for these life lessons is usually very boring and difficult for a high schooler to get into--read the [first few paragraphs of Ivanhoe]( [URL] ) and tell me how the *average* student wouldn't immediately get frustrated and go to the easily available Sparknotes. I feel that a lot of the older classics use a lot of dense and archaic language that is an immediate turn-off to modern teenagers. [NEWLINE] [NEWLINE] More recent works like Catcher in the Rye and the Crucible, I would agree because they are as accessible as you're gonna get. But are there enough "sexy" classics?</s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When police officers give a suspect the benefit of the doubt, that police officer will most likely be killed. People don't like<mask> the cops shoot an unarmed man,<mask><mask> the general public fails to take into account is the fact that a very large and strong man can physically overpower the average police officer. The only weapon which a police officer has is his gun, and the police NEVER shoot to kill. They aim for the suspect's legs or arms to incapacitate rather than kill.<mask> most of these decisions are decided<mask> quickly that they simply don't have time to line up a perfect shot. [NEWLINE] [NEWLINE] No cop in the United States of America has ever intended to kill a suspect, they were simply trying to subdue a suspect who was resisting arrest.<mask> is a cop meant to subdue a very large and very strong man who is resisting arrest and who is too big to be physically overpowered? The only way which I can think of is with a gun. </s><pad>
Label encoding: <s>When police officers give a suspect the benefit of the doubt, that police officer will most likely be killed. People don't like when the cops shoot an unarmed man, but what the general public fails to take into account is the fact that a very large and strong man can physically overpower the average police officer. The only weapon which a police officer has is his gun, and the police NEVER shoot to kill. They aim for the suspect's legs or arms to incapacitate rather than kill. But most of these decisions are decided so quickly that they simply don't have time to line up a perfect shot. [NEWLINE] [NEWLINE] No cop in the United States of America has ever intended to kill a suspect, they were simply trying to subdue a suspect who was resisting arrest. How is a cop meant to subdue a very large and very strong man who is resisting arrest and who is too big to be physically overpowered? The only way which I can think of is with a gun. </s><pad>
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The OP made a sweeping, unqualified statement. I don't think it's clear that they were speaking generally,<mask> they didn't use any language in their OP which would indicate this (<mask> ample opportunity - not a single qualifying "generally"/"usually"/"mostly"/"in the majority of cases"). They<mask> tried to invalidate my initial reply suggesting an exception, which implies that they're not expecting to accomodate exceptions. [NEWLINE] [NEWLINE] I don't believe we should be attempting to change their view to "a wireless mouse is (universally) better than a wired mouse",<mask> rather "a wired mouse is not necessarily better than a wireless mouse". The point of CMV isn't usually to convince the OP of the polar opposite of their view,<mask> simply that their view (<mask> stated) does not hold and should be amended in some way.<mask> the concept of a delta.<mask> this isn't the case, the OP should clarify.</s>
Label encoding: <s>The OP made a sweeping, unqualified statement. I don't think it's clear that they were speaking generally, as they didn't use any language in their OP which would indicate this ( despite ample opportunity - not a single qualifying "generally"/"usually"/"mostly"/"in the majority of cases"). They also tried to invalidate my initial reply suggesting an exception, which implies that they're not expecting to accomodate exceptions. [NEWLINE] [NEWLINE] I don't believe we should be attempting to change their view to "a wireless mouse is (universally) better than a wired mouse", but rather "a wired mouse is not necessarily better than a wireless mouse". The point of CMV isn't usually to convince the OP of the polar opposite of their view, but simply that their view ( as stated) does not hold and should be amended in some way. Hence the concept of a delta. If this isn't the case, the OP should clarify.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you are missing the point. The majority of of trade jobs in my area(I'll<mask> dare to say America overall) are already over saturated. I know people who [NEWLINE] [NEWLINE] completed welding school(settled<mask> a minimum wage janitor), [NEWLINE] [NEWLINE] who completed electrician school(working minimum wage at walmart), [NEWLINE] [NEWLINE] who completed the classes for pest control(jobless) [NEWLINE] [NEWLINE] An entire class who complete police academy and received state certification(90% never got a job after putting in 3 semesters) [NEWLINE] [NEWLINE] "Reminds me of a 14 year old girl that thinks her dad owes her a new car just<mask> she showed up to school that day." [NEWLINE] [NEWLINE] <mask> minimum wage is just about showing up? They don't actually work?<mask> you tried to sell something on craigslist for $100 and someone offered you $20 would you accept it? It has nothing to do with entitlement. It is about having a fair wage.</s>
Label encoding: <s>I think you are missing the point. The majority of of trade jobs in my area(I'll also dare to say America overall) are already over saturated. I know people who [NEWLINE] [NEWLINE] completed welding school(settled as a minimum wage janitor), [NEWLINE] [NEWLINE] who completed electrician school(working minimum wage at walmart), [NEWLINE] [NEWLINE] who completed the classes for pest control(jobless) [NEWLINE] [NEWLINE] An entire class who complete police academy and received state certification(90% never got a job after putting in 3 semesters) [NEWLINE] [NEWLINE] "Reminds me of a 14 year old girl that thinks her dad owes her a new car just because she showed up to school that day." [NEWLINE] [NEWLINE] So minimum wage is just about showing up? They don't actually work? If you tried to sell something on craigslist for $100 and someone offered you $20 would you accept it? It has nothing to do with entitlement. It is about having a fair wage.</s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you that the sheer number of parties makes this a difficult undertaking, and I by no means claim this to be a perfect idea. It's just an example of possible improvements (at least in my eyes) that could be made. [NEWLINE] [NEWLINE] You're correct, I suppose my post is less about the compulsory voting aspect (my title probably could have put less emphasis on this) and more on only counting votes for people who can prove at least some degree of political knowledge. My angle is that citizens should have the choice<mask> to whether they wish to research this and have a say, or leave the voting to those who are willing to do<mask>. I remain unconvinced that forcing people to vote is really having an effect on political knowledge on a large scale, and is more solidifying political bias (for every person who looked at Rudd's portrayal in the media and said "this is unfair", there was 5 more people saying "<mask> a c**t").</s>
Label encoding: <s>I agree with you that the sheer number of parties makes this a difficult undertaking, and I by no means claim this to be a perfect idea. It's just an example of possible improvements (at least in my eyes) that could be made. [NEWLINE] [NEWLINE] You're correct, I suppose my post is less about the compulsory voting aspect (my title probably could have put less emphasis on this) and more on only counting votes for people who can prove at least some degree of political knowledge. My angle is that citizens should have the choice as to whether they wish to research this and have a say, or leave the voting to those who are willing to do so. I remain unconvinced that forcing people to vote is really having an effect on political knowledge on a large scale, and is more solidifying political bias (for every person who looked at Rudd's portrayal in the media and said "this is unfair", there was 5 more people saying " what a c**t").</s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks a lot for your explanations. You've earned your ∆ by showing that it probably isn't a good idea to put women through infantry training (altough some data on this would still be great.) [NEWLINE] [NEWLINE] For the second argument: I see the point your making<mask> I don't know<mask> the problem is with it. I never said that standards should be lowered. [NEWLINE] [NEWLINE] Again I want to take Switzerland<mask> an example<mask> I don't know<mask> it works exactly in other countries. Right now about 50% - 60% of all draftees are unfit for military service<mask> in our system this doesn't mean that they have no other duties. Either they can serve in the civil defense or they have to pay compensation. [NEWLINE] [NEWLINE] Even<mask> the quota of those who fail to meet the military's standards is even higher for women (I guess with the current system it'd be around 80%) in my eyes this doesn't exclude them from the other duties.</s>
Label encoding: <s>Thanks a lot for your explanations. You've earned your ∆ by showing that it probably isn't a good idea to put women through infantry training (altough some data on this would still be great.) [NEWLINE] [NEWLINE] For the second argument: I see the point your making but I don't know what the problem is with it. I never said that standards should be lowered. [NEWLINE] [NEWLINE] Again I want to take Switzerland as an example because I don't know how it works exactly in other countries. Right now about 50% - 60% of all draftees are unfit for military service but in our system this doesn't mean that they have no other duties. Either they can serve in the civil defense or they have to pay compensation. [NEWLINE] [NEWLINE] Even if the quota of those who fail to meet the military's standards is even higher for women (I guess with the current system it'd be around 80%) in my eyes this doesn't exclude them from the other duties.</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> rather that the best athletes don't play soccer. [ENDQ] [NEWLINE] Best athletes don't win games. You can be<mask> good an athlete<mask> you want,<mask><mask> you're not good at the sport then you're screwed. There's an American "wonderkid" (or was), Freddy Adu iirc. He was *released* from a Brazilian team<mask> he wasn't good enough technically. That's<mask> the US sorely lacks. He's  been called "American Pelé" (seriously),<mask> was released about 7 or 8 months in. He then trained with Blackpool, a second tier English club who finished 20th out of 24, and wasn't offered a contract<mask> was allowed to continue training. [NEWLINE] [NEWLINE] <mask> a 25 year old previous "wonderkid" can't get a contract at a second tier club then it doesn't matter<mask> athletes you send in, your youth football is nowhere near good enough. Simple<mask> that.</s>
Label encoding: <s> [STARTQ] but rather that the best athletes don't play soccer. [ENDQ] [NEWLINE] Best athletes don't win games. You can be as good an athlete as you want, but if you're not good at the sport then you're screwed. There's an American "wonderkid" (or was), Freddy Adu iirc. He was *released* from a Brazilian team because he wasn't good enough technically. That's what the US sorely lacks. He's  been called "American Pelé" (seriously), yet was released about 7 or 8 months in. He then trained with Blackpool, a second tier English club who finished 20th out of 24, and wasn't offered a contract but was allowed to continue training. [NEWLINE] [NEWLINE] If a 25 year old previous "wonderkid" can't get a contract at a second tier club then it doesn't matter what athletes you send in, your youth football is nowhere near good enough. Simple as that.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0655, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So it is on the owners, management, stock owners to be a moral citizen and pay those people working 40 hours the wage. There is a difference between everyone paying a solid livable wage<mask> they are part of the community and forcing the issue.<mask> you force it on the economy the balance of the economy fights back.<mask><mask> it is just there and has been happening it becomes normal over time. Do you eat at McDonalds? Wendy's? Starbucks? Do you shop at Best Buy? Toys R Us? I have had friends and myself that have all worked at these places, put in 40 hours a week or more for multiple weeks and none of these places paid 15 dollars an hour for a cashier, burger flipper, warehouse stocker, bariesta.<mask> you believe they should make a livable wage at 40 hours a week no matter<mask> then do not frequent these places. The owners and management would quickly change the wage they pay people. </s>
Label encoding: <s>So it is on the owners, management, stock owners to be a moral citizen and pay those people working 40 hours the wage. There is a difference between everyone paying a solid livable wage because they are part of the community and forcing the issue. When you force it on the economy the balance of the economy fights back. But if it is just there and has been happening it becomes normal over time. Do you eat at McDonalds? Wendy's? Starbucks? Do you shop at Best Buy? Toys R Us? I have had friends and myself that have all worked at these places, put in 40 hours a week or more for multiple weeks and none of these places paid 15 dollars an hour for a cashier, burger flipper, warehouse stocker, bariesta. If you believe they should make a livable wage at 40 hours a week no matter what then do not frequent these places. The owners and management would quickly change the wage they pay people. </s>
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I do agree to an extent.<mask><mask><mask> a massive mistake a lot of people make is to not think deeply enough about the time period in which the painting was released<mask> well<mask> looking at<mask> the artist contemporaries were doing at the time. [NEWLINE] [NEWLINE] The first example you provided is a work by Piet Mondrian circa 1930ish. [NEWLINE] <mask> now we are very used to seeing such bold lines and flat colours such<mask> those often used in advertising, the web and print, Work like this was a massive juxtaposition (important wanky art word) to the more realistic styles like impressionism and post-impressionism that came before. [NEWLINE] [NEWLINE] Sometimes ground breaking art can be<mask> simple<mask> looking at<mask> everyone else is doing and then doing the opposite. You don't have to like it<mask> you have to respect some of it for challenging the norms, it is just sometimes you have to look a little deeper to understand<mask> the norms<mask> at the time.</s>
Label encoding: <s>I do agree to an extent. However I think a massive mistake a lot of people make is to not think deeply enough about the time period in which the painting was released as well as looking at what the artist contemporaries were doing at the time. [NEWLINE] [NEWLINE] The first example you provided is a work by Piet Mondrian circa 1930ish. [NEWLINE] While now we are very used to seeing such bold lines and flat colours such as those often used in advertising, the web and print, Work like this was a massive juxtaposition (important wanky art word) to the more realistic styles like impressionism and post-impressionism that came before. [NEWLINE] [NEWLINE] Sometimes ground breaking art can be as simple as looking at what everyone else is doing and then doing the opposite. You don't have to like it but you have to respect some of it for challenging the norms, it is just sometimes you have to look a little deeper to understand what the norms where at the time.</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think killing yourself is very egotistical for a couple reasons. [NEWLINE] [NEWLINE] 1. It assumes that you know<mask> much about your potential future, even in the short term, that you can make a massive irreversible decision for your entire life.<mask><mask> things get way better in five minutes? It's arrogant to say that they certainly won't, and you can probably wait till then,<mask> Camus says: “Should I kill myself, or have a cup of coffee?” [NEWLINE] [NEWLINE] 2. It assumes that your own pain and suffering are more important to alleviate than the good you could do for others. [NEWLINE] [NEWLINE] 3. It destroys the substantial effort and resources others have put into you to get you<mask> you are - even<mask> you've had a shit life, somebody had to feed you unless you grew up feral ;) [NEWLINE] [NEWLINE] [NEWLINE] For these reasons<mask><mask> one who commits suicide is inherently egotistical and should try to broaden their perspectives.</s>
Label encoding: <s>I think killing yourself is very egotistical for a couple reasons. [NEWLINE] [NEWLINE] 1. It assumes that you know so much about your potential future, even in the short term, that you can make a massive irreversible decision for your entire life. What if things get way better in five minutes? It's arrogant to say that they certainly won't, and you can probably wait till then, as Camus says: “Should I kill myself, or have a cup of coffee?” [NEWLINE] [NEWLINE] 2. It assumes that your own pain and suffering are more important to alleviate than the good you could do for others. [NEWLINE] [NEWLINE] 3. It destroys the substantial effort and resources others have put into you to get you where you are - even if you've had a shit life, somebody had to feed you unless you grew up feral ;) [NEWLINE] [NEWLINE] [NEWLINE] For these reasons I think one who commits suicide is inherently egotistical and should try to broaden their perspectives.</s>
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Democracy is flawed, for sure. [NEWLINE] [NEWLINE] Australia<mask> has the major parties (in recent years a split parliament has resulted in independents and the greens party having significant power). [NEWLINE] [NEWLINE] I don't like churches telling people<mask> to vote,<mask> at the same time you really can't/shouldn't stop them. [NEWLINE] [NEWLINE] I maintain that people voting is a good idea. For example, say you have someone very young and naive who just goes along and mirrors their parent's views. I believe that the actual ritual and practice of being involved encourages a better curiosity and understanding. I feel Australia's compulsory voting is a good thing. Would America benefit from it? Hard to say. Either way I couldn't see things them changing. It would require a vote, which would require voluntary voters  to insist on it, which would in a sense be them 'giving up power'. The only reason is<mask> 'bribing' voters got terribly out of control.</s>
Label encoding: <s>Democracy is flawed, for sure. [NEWLINE] [NEWLINE] Australia also has the major parties (in recent years a split parliament has resulted in independents and the greens party having significant power). [NEWLINE] [NEWLINE] I don't like churches telling people how to vote, but at the same time you really can't/shouldn't stop them. [NEWLINE] [NEWLINE] I maintain that people voting is a good idea. For example, say you have someone very young and naive who just goes along and mirrors their parent's views. I believe that the actual ritual and practice of being involved encourages a better curiosity and understanding. I feel Australia's compulsory voting is a good thing. Would America benefit from it? Hard to say. Either way I couldn't see things them changing. It would require a vote, which would require voluntary voters  to insist on it, which would in a sense be them 'giving up power'. The only reason is if 'bribing' voters got terribly out of control.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The special position of athletes is not something that I'm familiar with in a European setting. [ENDQ] [NEWLINE] The extent to which Universities in America seek out athletes is definitely greater than in Europe.<mask> I've definitely seen athletes in Europe (specifically the UK) receive special accommodation (in this case for Rowing at Cambridge). [NEWLINE] [NEWLINE] [STARTQ] Musicians have their own specialized institutions. [ENDQ] [NEWLINE] Most large universities in the US have facilities and programs for the development of the arts. Specialization might be more common in Europe<mask> conservatories and sports academies are fairly rare in the US. [NEWLINE] [NEWLINE] [STARTQ] <mask> sports are not an academics pursuit. [ENDQ] [NEWLINE] I never said they were and actually specifically acknowledge that they are not. I simply state that Universities have, historically (even in Europe) pursued a broad idea of education. I mean you can go all the way back to the Platonic academies and see that they put value on physical fitness and sport. </s>
Label encoding: <s> [STARTQ] The special position of athletes is not something that I'm familiar with in a European setting. [ENDQ] [NEWLINE] The extent to which Universities in America seek out athletes is definitely greater than in Europe. However I've definitely seen athletes in Europe (specifically the UK) receive special accommodation (in this case for Rowing at Cambridge). [NEWLINE] [NEWLINE] [STARTQ] Musicians have their own specialized institutions. [ENDQ] [NEWLINE] Most large universities in the US have facilities and programs for the development of the arts. Specialization might be more common in Europe but conservatories and sports academies are fairly rare in the US. [NEWLINE] [NEWLINE] [STARTQ] But sports are not an academics pursuit. [ENDQ] [NEWLINE] I never said they were and actually specifically acknowledge that they are not. I simply state that Universities have, historically (even in Europe) pursued a broad idea of education. I mean you can go all the way back to the Platonic academies and see that they put value on physical fitness and sport. </s>
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because in this situation the male is not the one who knowingly brought a child into this world. The female is, and only the female.<mask> protests from the male.<mask> the male should get off totaly scott free,<mask> he makes it officialy known that he does not want the baby before the abortion limit has passed, giving enough time to the female to react. [NEWLINE] [NEWLINE] **He/she who wants the baby brought to term should be responsible for it.** [NEWLINE] [NEWLINE] Under normal circumstances this will be both parents. [NEWLINE] [NEWLINE] <mask> sometimes it will be only one of them. Then the other one should get off scott free. This could work even for females, too. For example<mask> the man wants the child and the woman does not, then after the kid is born the male should be solely responsible and the woman should get off scott free. Instead of the only option being abortion for the woman,<mask> it is currently.</s>
Label encoding: <s>Because in this situation the male is not the one who knowingly brought a child into this world. The female is, and only the female. Despite protests from the male. Therefore the male should get off totaly scott free, if he makes it officialy known that he does not want the baby before the abortion limit has passed, giving enough time to the female to react. [NEWLINE] [NEWLINE] **He/she who wants the baby brought to term should be responsible for it.** [NEWLINE] [NEWLINE] Under normal circumstances this will be both parents. [NEWLINE] [NEWLINE] But sometimes it will be only one of them. Then the other one should get off scott free. This could work even for females, too. For example when the man wants the child and the woman does not, then after the kid is born the male should be solely responsible and the woman should get off scott free. Instead of the only option being abortion for the woman, as it is currently.</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> you agree then that it's self-perpetuating. It exists<mask> it exists and nobody is changing it. Homosexuality was once a horrible disease and everyone inflicted with it should be locked up. Did it change overnight? Of course not. Has there been a movement in that subject that has radically changed society's perception? Hell yes. [NEWLINE] [NEWLINE] <mask> it's the way it is<mask> nobody is calling it out on a large scale. [NEWLINE] [NEWLINE] In your own definition: [NEWLINE] [STARTQ] The female equivalent is matriarchy. [ENDQ] So a woman can't be patriarchal,<mask> matriarchal. [NEWLINE] [NEWLINE] I can see<mask> feminism has its place, and it is correct to an extent, it's the feminists who think the whole world is evil<mask> the president isn't female that I can't stand. (The friend I lost was one of those... I tolerated it<mask> I still enjoyed her company and she challenged my views regularly.) [NEWLINE] </s><pad>
Label encoding: <s>So you agree then that it's self-perpetuating. It exists because it exists and nobody is changing it. Homosexuality was once a horrible disease and everyone inflicted with it should be locked up. Did it change overnight? Of course not. Has there been a movement in that subject that has radically changed society's perception? Hell yes. [NEWLINE] [NEWLINE] So it's the way it is because nobody is calling it out on a large scale. [NEWLINE] [NEWLINE] In your own definition: [NEWLINE] [STARTQ] The female equivalent is matriarchy. [ENDQ] So a woman can't be patriarchal, but matriarchal. [NEWLINE] [NEWLINE] I can see where feminism has its place, and it is correct to an extent, it's the feminists who think the whole world is evil because the president isn't female that I can't stand. (The friend I lost was one of those... I tolerated it because I still enjoyed her company and she challenged my views regularly.) [NEWLINE] </s><pad>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Go look at the posts around adviceanimals and you'll find that racism (outside of just random slurs) is actually quite frequently upvoted. [ENDQ] [NEWLINE] Maybe the people in AdviceAnimals are racist, and support or enjoy those views? [NEWLINE] [NEWLINE] [STARTQ] And the point I'm making is<mask><mask> that anti-bigotry is a valid cause for an exception. [ENDQ] [NEWLINE] <mask> else will be considered valid for exception, once you open that door? Should we be able to downvote brigade anyone supporting bankers,<mask> I find those people offensive. I find lots of things offensive that I would like to see wiped out of existence.<mask><mask>, I find it offensive that people downvote things they don't agree with, even<mask> it contributes to the larger discussion. I find that anti-intellectual. [NEWLINE] [NEWLINE] <mask> you disagree with something<mask> of its racist or 'offensive' tones,'report' it, and move on.</s><pad>
Label encoding: <s> [STARTQ] Go look at the posts around adviceanimals and you'll find that racism (outside of just random slurs) is actually quite frequently upvoted. [ENDQ] [NEWLINE] Maybe the people in AdviceAnimals are racist, and support or enjoy those views? [NEWLINE] [NEWLINE] [STARTQ] And the point I'm making is I think that anti-bigotry is a valid cause for an exception. [ENDQ] [NEWLINE] What else will be considered valid for exception, once you open that door? Should we be able to downvote brigade anyone supporting bankers, because I find those people offensive. I find lots of things offensive that I would like to see wiped out of existence. In fact, I find it offensive that people downvote things they don't agree with, even when it contributes to the larger discussion. I find that anti-intellectual. [NEWLINE] [NEWLINE] If you disagree with something because of its racist or 'offensive' tones,'report' it, and move on.</s><pad>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My view will not be changed<mask> you tell me that people just see me<mask> more professional or educated<mask> I'm white,<mask> that has nothing to do with race and has everything to do with the way I present myself. [ENDQ] [NEWLINE] <mask> your view that people do not think of you<mask> more professional or educated<mask> you're white cannot be changed<mask> people do not think differently of you based on race?  That's circular reasoning at the very least. <mask> your view cannot be changed<mask> post here? [NEWLINE] [NEWLINE] [STARTQ] I'm getting a lot of replies citing<mask> ethnic sounding names vs white sounding names affect job interviews. This is a cultural issue, the color of someone's skin has nothing to do with their name. [ENDQ] [NEWLINE] This is incorrect.  The color of your skin absolutely correlates with your name.  There are many more black people named "Tyrone" than there are white people.  It's just a fact.</s>
Label encoding: <s> [STARTQ] My view will not be changed because you tell me that people just see me as more professional or educated because I'm white, because that has nothing to do with race and has everything to do with the way I present myself. [ENDQ] [NEWLINE] So your view that people do not think of you as more professional or educated because you're white cannot be changed because people do not think differently of you based on race?  That's circular reasoning at the very least.  If your view cannot be changed why post here? [NEWLINE] [NEWLINE] [STARTQ] I'm getting a lot of replies citing how ethnic sounding names vs white sounding names affect job interviews. This is a cultural issue, the color of someone's skin has nothing to do with their name. [ENDQ] [NEWLINE] This is incorrect.  The color of your skin absolutely correlates with your name.  There are many more black people named "Tyrone" than there are white people.  It's just a fact.</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A down payment on a second car [ENDQ] [NEWLINE] 6k is more than enough for a car.  Anything more than that and you are just paying for a fancier hunk of metal that is depreciating, unlike a ring.<mask><mask> can you justify that,<mask> not a nice ring? [NEWLINE] [NEWLINE] <mask> for the things that would be better for a marriage than a ring, my fiance and I need none of them.  We both already own houses, cars.  There's very few things that money could buy that would help it.  The only thing I need money for is ER,<mask> a once in a lifetime purchase isn't going to hurt that too much. [NEWLINE] [NEWLINE] [STARTQ] <mask> I am suggesting that it is ridiculous to even want to do<mask>. [ENDQ] [NEWLINE] Originally, you had much stronger wording.  "Morally reprehensible and financially irresponsible." <mask>, it seems your view has changed at least somewhat.</s>
Label encoding: <s> [STARTQ] A down payment on a second car [ENDQ] [NEWLINE] 6k is more than enough for a car.  Anything more than that and you are just paying for a fancier hunk of metal that is depreciating, unlike a ring. So how can you justify that, but not a nice ring? [NEWLINE] [NEWLINE] As for the things that would be better for a marriage than a ring, my fiance and I need none of them.  We both already own houses, cars.  There's very few things that money could buy that would help it.  The only thing I need money for is ER, but a once in a lifetime purchase isn't going to hurt that too much. [NEWLINE] [NEWLINE] [STARTQ] But I am suggesting that it is ridiculous to even want to do so. [ENDQ] [NEWLINE] Originally, you had much stronger wording.  "Morally reprehensible and financially irresponsible."  So, it seems your view has changed at least somewhat.</s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> you're relying on the majority of people to think "this is stupid" for your idea to work, then it's probably a stupid idea. [ENDQ] [NEWLINE] <mask> I must agree with this point in the theoretical sense, I'm<mask> speaking practically. [Much like<mask> a society develops some assholes]( [URL] ) (I couldn't find the later analysis,<mask> it found the population tended to find equilibrium, rather than reach extremes), the above solution is more of a personal response than a position of the ideal situation. It's a work-around to a presently-flawed system. [NEWLINE] [NEWLINE] You've refined my view somewhat, and I appreciate that,<mask> I still think this is an acceptable reaction on a small scale, isn't it? Is there any reason my<mask> and I shouldn't spare our conflicting votes and compromise on a third party,<mask> third parties are<mask> woefully under-represented due to the spoiler effect.</s>
Label encoding: <s> [STARTQ] But if you're relying on the majority of people to think "this is stupid" for your idea to work, then it's probably a stupid idea. [ENDQ] [NEWLINE] While I must agree with this point in the theoretical sense, I'm also speaking practically. [Much like how a society develops some assholes]( [URL] ) (I couldn't find the later analysis, but it found the population tended to find equilibrium, rather than reach extremes), the above solution is more of a personal response than a position of the ideal situation. It's a work-around to a presently-flawed system. [NEWLINE] [NEWLINE] You've refined my view somewhat, and I appreciate that, but I still think this is an acceptable reaction on a small scale, isn't it? Is there any reason my SO and I shouldn't spare our conflicting votes and compromise on a third party, when third parties are so woefully under-represented due to the spoiler effect.</s>
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've read a number of your responses, and<mask><mask> the flaw in your argument is that you're hyper focused on the individual who is not causing harm.<mask><mask> from the individual DUI ticket holder's point of view, you have a defensible point.<mask> that's not<mask> these laws exist. We make DUIs and intent to murder and shooting a gun off illegal<mask> we don't want those events happening in general.<mask> you ask anyone "Do you want to live in a world with or without drunk drivers?" They're always going to say "Oh yeah without."<mask> one individual drunk driver who hasn't hit anyone isn't the portrait of risk, a bunch of drunk people driving really increases the chances something bad will happen.<mask><mask> these laws deter people from these crimes, do you disagree? The point of the laws is to generally prevent these things from happening and punishing people who make the terrible decisions to engage in them. [NEWLINE] </s>
Label encoding: <s>I've read a number of your responses, and I think the flaw in your argument is that you're hyper focused on the individual who is not causing harm. I think from the individual DUI ticket holder's point of view, you have a defensible point. But that's not why these laws exist. We make DUIs and intent to murder and shooting a gun off illegal because we don't want those events happening in general. If you ask anyone "Do you want to live in a world with or without drunk drivers?" They're always going to say "Oh yeah without." While one individual drunk driver who hasn't hit anyone isn't the portrait of risk, a bunch of drunk people driving really increases the chances something bad will happen. I think these laws deter people from these crimes, do you disagree? The point of the laws is to generally prevent these things from happening and punishing people who make the terrible decisions to engage in them. [NEWLINE] </s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your assertion is that it is "wrong",<mask> is it wrong? All of your evidence seems to be **potential** concerns with<mask> the money **could** go. You don't know<mask> the money is going, for all you know, you give it directly to the homeless people, they spend it on clothes and food and you avoid the potential bureaucracy and corruption of charities. Yes, worst case they spend it on meth to inject into their eyeballs,<mask> you can't assume the worst and base your argument off of that assumption. [NEWLINE] [NEWLINE] A more prudent statement is to say: [NEWLINE] [NEWLINE] "I dont believe giving money to people on the street is<mask> generally beneficial<mask> giving to charities." [NEWLINE] [NEWLINE] Do you see the difference? Your statement says it is wrong, this would imply it actively does harm, and<mask> that it does more harm than good. This is something you cannot vouch for.<mask> your statement is incorrect.</s>
Label encoding: <s>Your assertion is that it is "wrong", how is it wrong? All of your evidence seems to be **potential** concerns with where the money **could** go. You don't know where the money is going, for all you know, you give it directly to the homeless people, they spend it on clothes and food and you avoid the potential bureaucracy and corruption of charities. Yes, worst case they spend it on meth to inject into their eyeballs, but you can't assume the worst and base your argument off of that assumption. [NEWLINE] [NEWLINE] A more prudent statement is to say: [NEWLINE] [NEWLINE] "I dont believe giving money to people on the street is as generally beneficial as giving to charities." [NEWLINE] [NEWLINE] Do you see the difference? Your statement says it is wrong, this would imply it actively does harm, and moreover that it does more harm than good. This is something you cannot vouch for. Therefore your statement is incorrect.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I wonder to<mask> degree this has to do with past, not present inequality. [ENDQ] [NEWLINE] That could be very true,<mask> the past does help shape the present.<mask> we still see more men at the top, we are conditioned to think of high-level positions<mask> a man's place. [NEWLINE] [NEWLINE] It is going to take a<mask> to fix, I certainly won't deny that. I just have a problem with the argument that some people here are making that simply<mask> there are now individual protections in place, the problem ceases to exist. Without true, structural change, we really haven't done anything. [NEWLINE] [NEWLINE] <mask>, I'm absolutely not blaming "Men." For the vast majority of men, they have no say whatsoever in<mask> a woman makes--they are just<mask> dis-empowered<mask> anyone else.<mask><mask> the key distinction is that not all men have power,<mask> most of the people who do are men.</s>
Label encoding: <s> [STARTQ] I wonder to what degree this has to do with past, not present inequality. [ENDQ] [NEWLINE] That could be very true, but the past does help shape the present. If we still see more men at the top, we are conditioned to think of high-level positions as a man's place. [NEWLINE] [NEWLINE] It is going to take a while to fix, I certainly won't deny that. I just have a problem with the argument that some people here are making that simply because there are now individual protections in place, the problem ceases to exist. Without true, structural change, we really haven't done anything. [NEWLINE] [NEWLINE] Also, I'm absolutely not blaming "Men." For the vast majority of men, they have no say whatsoever in what a woman makes--they are just as dis-empowered as anyone else. I think the key distinction is that not all men have power, but most of the people who do are men.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ [NEWLINE] [NEWLINE] That's a really good explanation. All too often, literally in every experience I've had with the term, it's been used by one party to dismiss the opinions of another party. e.g. [NEWLINE] [NEWLINE] Me: "It's just down the street, like a 10 minute walk" [NEWLINE] Woman: "you don't understand, you're a man, I might be raped, you don't have to worry about that. That's your male privilege." [NEWLINE] [NEWLINE] I've been raped, the above woman doesn't know<mask> I've been through and I get to deal with laughter<mask> I bring it up. I don't blame her for being ignorant of my past,<mask> I can blame her for entirely dismissing my opinions and insights simply<mask> I have a penis. [NEWLINE] [NEWLINE] That's the problem -<mask> it's used to dismiss someone rather than simply to suggest that they're not considering perspective in their statement.</s>
Label encoding: <s>∆ [NEWLINE] [NEWLINE] That's a really good explanation. All too often, literally in every experience I've had with the term, it's been used by one party to dismiss the opinions of another party. e.g. [NEWLINE] [NEWLINE] Me: "It's just down the street, like a 10 minute walk" [NEWLINE] Woman: "you don't understand, you're a man, I might be raped, you don't have to worry about that. That's your male privilege." [NEWLINE] [NEWLINE] I've been raped, the above woman doesn't know what I've been through and I get to deal with laughter if I bring it up. I don't blame her for being ignorant of my past, but I can blame her for entirely dismissing my opinions and insights simply because I have a penis. [NEWLINE] [NEWLINE] That's the problem - when it's used to dismiss someone rather than simply to suggest that they're not considering perspective in their statement.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Perhaps guys can "look like shit"<mask> they simply don't care? [NEWLINE] <mask> I see more than anything is women going through all these rituals to make other women jealous, or to impress other women. Most guys I know couldn't care less. The injustice and oppression (<mask> it pertains to this) feels to me to be entirely self-inflicted, and self-perpetuating<mask><mask><mask> women feel the need to keep doing<mask>. [NEWLINE] [NEWLINE] <mask> do you think would happen<mask> women simply stopped wearing makeup? The world wouldn't implode. Those who didn't mind wouldn't mind a bit. Those who didn't mind wouldn't find anyone attractive and the situation would resolve itself. [NEWLINE] [NEWLINE] [STARTQ] Don't confuse 'the patriarchy' with'some patriarchal figure who happens to be specific to a given woman.' [ENDQ] [NEWLINE] <mask> you're saying all men in general, rather than a man in her life. Gotcha.</s>
Label encoding: <s>Perhaps guys can "look like shit" because they simply don't care? [NEWLINE] What I see more than anything is women going through all these rituals to make other women jealous, or to impress other women. Most guys I know couldn't care less. The injustice and oppression ( as it pertains to this) feels to me to be entirely self-inflicted, and self-perpetuating so long as women feel the need to keep doing so. [NEWLINE] [NEWLINE] What do you think would happen if women simply stopped wearing makeup? The world wouldn't implode. Those who didn't mind wouldn't mind a bit. Those who didn't mind wouldn't find anyone attractive and the situation would resolve itself. [NEWLINE] [NEWLINE] [STARTQ] Don't confuse 'the patriarchy' with'some patriarchal figure who happens to be specific to a given woman.' [ENDQ] [NEWLINE] So you're saying all men in general, rather than a man in her life. Gotcha.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>With all due respect,<mask> you replace "pill" with "beverage," you can make an argument that soda is exactly<mask> you just described: Something that tastes good<mask> slowly poisons people. The FDA hasn't banned that<mask> at least on some level, we acknowledge that in this country, a person should have the right to put whatever they want in their body,<mask><mask><mask> it doesn't hurt anyone else. Not everyone agrees, and there are obvious exceptions,<mask> on the whole, that's the direction we're moving towards. The move towards legalizing marijuana and ending the larger war on drugs is a product of this. [NEWLINE] [NEWLINE] Tobacco officials don't really need to pay off anyone to keep cigarettes legal. Remember abolition in the 20s? People don't like the idea of the government banning a substance that was previously legal, even<mask> it's poisonous. It's legal right now<mask> it's always been legal.</s>
Label encoding: <s>With all due respect, if you replace "pill" with "beverage," you can make an argument that soda is exactly what you just described: Something that tastes good but slowly poisons people. The FDA hasn't banned that because at least on some level, we acknowledge that in this country, a person should have the right to put whatever they want in their body, so long as it doesn't hurt anyone else. Not everyone agrees, and there are obvious exceptions, but on the whole, that's the direction we're moving towards. The move towards legalizing marijuana and ending the larger war on drugs is a product of this. [NEWLINE] [NEWLINE] Tobacco officials don't really need to pay off anyone to keep cigarettes legal. Remember abolition in the 20s? People don't like the idea of the government banning a substance that was previously legal, even if it's poisonous. It's legal right now because it's always been legal.</s>
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The issue is that there's a power imbalance here, and whenever there's a power imbalance in a relationship, the situation is ripe for coercion and exploitation. <mask> there's a power imbalance, the person of lower "status" might not feel like they're able to set appropriate limits with their partner.  I'm not saying that power imbalances are *always* exploitative, or that exploitative relationships can't happen between people who are in the same age range,<mask> a power imbalance is definitely a risk.  It's not just age differences that can cause a problem - teacher/student, boss/employee, rich/poor, celebrity/normal, etc. are all situations<mask> a little extra care and a little extra discretion are warranted. <mask> it might be legal and understandable for James Franco to seek a 17-year-old for a fling, it shows an appalling lack of judgement.  </s>
Label encoding: <s>The issue is that there's a power imbalance here, and whenever there's a power imbalance in a relationship, the situation is ripe for coercion and exploitation.  When there's a power imbalance, the person of lower "status" might not feel like they're able to set appropriate limits with their partner.  I'm not saying that power imbalances are *always* exploitative, or that exploitative relationships can't happen between people who are in the same age range, but a power imbalance is definitely a risk.  It's not just age differences that can cause a problem - teacher/student, boss/employee, rich/poor, celebrity/normal, etc. are all situations where a little extra care and a little extra discretion are warranted.  While it might be legal and understandable for James Franco to seek a 17-year-old for a fling, it shows an appalling lack of judgement.  </s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0640, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Dude, dude. [NEWLINE] [NEWLINE] Suppose we collected a data set on a thousand of such interactions, controlling for race, gender, age, nationality, gender attitudes in the home, education, income, time of day,<mask> on, miraculously manage to objectively rate the attractiveness of the female, and regress frequency of cat calls on that attractiveness, and find that attractiveness positively affects frequency of cat calls, i.e. more attractive women get more cat calls. [NEWLINE] [NEWLINE] Now<mask>? Should hot women stop being hot? Should they wear burkas in New York City? Should we make it illegal for a woman to travel without a male guardian,<mask> it is in Saudi Arabia? You've just wasted millions of taxpayer money (<mask> that's<mask> much it will cost you to collect anything close to such a dataset) and come up with a moot point that has approximately 0 policy implications. Well done, sir. You're fired.</s>
Label encoding: <s>Dude, dude. [NEWLINE] [NEWLINE] Suppose we collected a data set on a thousand of such interactions, controlling for race, gender, age, nationality, gender attitudes in the home, education, income, time of day, so on, miraculously manage to objectively rate the attractiveness of the female, and regress frequency of cat calls on that attractiveness, and find that attractiveness positively affects frequency of cat calls, i.e. more attractive women get more cat calls. [NEWLINE] [NEWLINE] Now what? Should hot women stop being hot? Should they wear burkas in New York City? Should we make it illegal for a woman to travel without a male guardian, as it is in Saudi Arabia? You've just wasted millions of taxpayer money ( because that's how much it will cost you to collect anything close to such a dataset) and come up with a moot point that has approximately 0 policy implications. Well done, sir. You're fired.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Without it being an understood rule to check for cyclists before opening your door, you'll unsuspectingly force society's cyclists to avoid doing you the favour of staying out of your way (which is undoubtedly<mask> they're doing) and hold up traffic by being in the car lane instead.  Are you prepared to make that concession?  Do you think that based on this principle we'd have a better system?  Or maybe should we put the onus on the folks in the parked cars<mask> that cyclists feel safe enough not to need to take up the majority of the traffic lane? [NEWLINE] [NEWLINE] I cycle every day and must admit that a smart cyclist _will_ take the initiative to ride defensively with respect to door prizing,<mask> we have<mask> much more to lose! <mask> realize that<mask> we ride closer to the parked cars we do<mask> in order to make the drivers' lives easier and not for our own benefit!</s>
Label encoding: <s>Without it being an understood rule to check for cyclists before opening your door, you'll unsuspectingly force society's cyclists to avoid doing you the favour of staying out of your way (which is undoubtedly what they're doing) and hold up traffic by being in the car lane instead.  Are you prepared to make that concession?  Do you think that based on this principle we'd have a better system?  Or maybe should we put the onus on the folks in the parked cars so that cyclists feel safe enough not to need to take up the majority of the traffic lane? [NEWLINE] [NEWLINE] I cycle every day and must admit that a smart cyclist _will_ take the initiative to ride defensively with respect to door prizing, since we have so much more to lose!  But realize that when we ride closer to the parked cars we do so in order to make the drivers' lives easier and not for our own benefit!</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0601, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like your reply,<mask> it seems that the only way to determine<mask> to allocate seats based on physical conditions is self reporting or not.<mask><mask> not at the airlines discretion? For example, instead of checking in online I could contact an airline representative at the airport who would verify I am too tall to fit comfortably and safely into their regular seats and check me in onto a more leg room seat. [NEWLINE] [NEWLINE] [STARTQ] Willingness to pay is a very close approximation of need,<mask> the people who pay for more legroom line up quite well with the people who need more legroom. [ENDQ] [NEWLINE] In my experience this is not the case. In many instances I have seen, extra leg room seats are not only a few extra dollars,<mask> a significant percentage of the original price of the ticket and the people occupying those seats are less obvious in their need for the space and more obvious in their ability to pay for it.</s>
Label encoding: <s>I like your reply, but it seems that the only way to determine how to allocate seats based on physical conditions is self reporting or not. But why not at the airlines discretion? For example, instead of checking in online I could contact an airline representative at the airport who would verify I am too tall to fit comfortably and safely into their regular seats and check me in onto a more leg room seat. [NEWLINE] [NEWLINE] [STARTQ] Willingness to pay is a very close approximation of need, so the people who pay for more legroom line up quite well with the people who need more legroom. [ENDQ] [NEWLINE] In my experience this is not the case. In many instances I have seen, extra leg room seats are not only a few extra dollars, but a significant percentage of the original price of the ticket and the people occupying those seats are less obvious in their need for the space and more obvious in their ability to pay for it.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0623, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Imagine all the types of marine organisms you have eaten: from Salmon to crab to mussels. Somewhere around 80% ( [STARTQ] 90,000 described species) of all marine organisms start their life<mask> plankton in the open ocean. Whether it's a fish or invertebrate, they morph into different stages depending on ocean conditions and food availability. Now imagine<mask> many of those organisms produce a shell or eat something that produces a shell (either at an adult or larval stage). Take for example a pteropod, essentially a free swimming snail, that is trying to form a shell in more acidic water....it's going to put more resources (and<mask> ENERGY) into building a shell which may compromise it's further development. [ENDQ] [NEWLINE] In aquaculture, even a small temperature change can cause a fish to grow up to be deformed<mask> imagine<mask> a small pH change can do. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Imagine all the types of marine organisms you have eaten: from Salmon to crab to mussels. Somewhere around 80% ( [STARTQ] 90,000 described species) of all marine organisms start their life as plankton in the open ocean. Whether it's a fish or invertebrate, they morph into different stages depending on ocean conditions and food availability. Now imagine how many of those organisms produce a shell or eat something that produces a shell (either at an adult or larval stage). Take for example a pteropod, essentially a free swimming snail, that is trying to form a shell in more acidic water....it's going to put more resources (and therefore ENERGY) into building a shell which may compromise it's further development. [ENDQ] [NEWLINE] In aquaculture, even a small temperature change can cause a fish to grow up to be deformed so imagine what a small pH change can do. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have to say, its a little low not giving a "guest" option. [NEWLINE] [NEWLINE] I am gay, living with my gf, neither of us are loud out there people. (<mask> your issue with "putting it in my face" is another topic). [NEWLINE] [NEWLINE] Wedding invitation from her cousin comes around and she only gets an invitation for her... no spot for "guest" or "+1." [NEWLINE] [NEWLINE] We were pretty upset by it. I mean, we are adults. Any other adult gets a guest option on their invitation. [NEWLINE] [NEWLINE] Well come to find out, it wasnt just us. They did this to a lot of people. [NEWLINE] [NEWLINE] You do<mask> you want.<mask> its pretty rude<mask><mask>. We were not the only people upset by this, and there was a bit of drama around it at the wedding (<mask><mask> my gf). </s><pad><pad>
Label encoding: <s>I have to say, its a little low not giving a "guest" option. [NEWLINE] [NEWLINE] I am gay, living with my gf, neither of us are loud out there people. ( although your issue with "putting it in my face" is another topic). [NEWLINE] [NEWLINE] Wedding invitation from her cousin comes around and she only gets an invitation for her... no spot for "guest" or "+1." [NEWLINE] [NEWLINE] We were pretty upset by it. I mean, we are adults. Any other adult gets a guest option on their invitation. [NEWLINE] [NEWLINE] Well come to find out, it wasnt just us. They did this to a lot of people. [NEWLINE] [NEWLINE] You do what you want. But its pretty rude IMO. We were not the only people upset by this, and there was a bit of drama around it at the wedding ( according to my gf). </s><pad><pad>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks! I'm glad I was able to help out in some way. [NEWLINE] [NEWLINE] Truth be told, I remember looking into the inside job theory, which seemed to come out everywhere hours after it happened. Everything was really chaotic and no one was sure was happened,<mask> it was nice to grasp onto something that wrapped up everything in a neat, easy package. Things were even more confusing<mask> Zeitgeist and Loose Change came out. Once I realized that truthers were just cherry picking evidence that confirmed the ideas they already believed in, it was much easier to move past it. [NEWLINE] [NEWLINE] I don't think you're dumb or stupid. It's an idea that's easy to buy into, especially<mask> truthers only provide evidence that agrees with their preconceived ideas. [NEWLINE] [NEWLINE] Oh, and [here's]( [URL] ) a video of the Mythbusters episode I referenced. Sorry about the shitty video opening...</s>
Label encoding: <s>Thanks! I'm glad I was able to help out in some way. [NEWLINE] [NEWLINE] Truth be told, I remember looking into the inside job theory, which seemed to come out everywhere hours after it happened. Everything was really chaotic and no one was sure was happened, so it was nice to grasp onto something that wrapped up everything in a neat, easy package. Things were even more confusing when Zeitgeist and Loose Change came out. Once I realized that truthers were just cherry picking evidence that confirmed the ideas they already believed in, it was much easier to move past it. [NEWLINE] [NEWLINE] I don't think you're dumb or stupid. It's an idea that's easy to buy into, especially when truthers only provide evidence that agrees with their preconceived ideas. [NEWLINE] [NEWLINE] Oh, and [here's]( [URL] ) a video of the Mythbusters episode I referenced. Sorry about the shitty video opening...</s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] literally every ideology says that [ENDQ] [NEWLINE] This is just not true. Policies such<mask> socialism sacrifice individual freedom for protection and government policies that claim to help the majority of people. Libertarians would reject these policies in favor of freedom. [NEWLINE] [NEWLINE] [STARTQ] they generally don't say anything about the actual effectiveness of policies [ENDQ] [NEWLINE] I actually agree with you here. Libertarians are notorious for not really thinking through the practicality of their policies. A great deal of which are idealistic and most likely wouldn't work in the real world.<mask> we really don't have much in the way of tested Libertarian ideas,<mask> it makes it difficult to gauge the practicality. Most people would dismiss Libertarian ideas simply<mask> they are<mask> used to their current situation of being governed. For example most people completely dismiss the notion of a country with very limited government<mask> they are boggled at the idea of a country without a strong government. </s>
Label encoding: <s> [STARTQ] literally every ideology says that [ENDQ] [NEWLINE] This is just not true. Policies such as socialism sacrifice individual freedom for protection and government policies that claim to help the majority of people. Libertarians would reject these policies in favor of freedom. [NEWLINE] [NEWLINE] [STARTQ] they generally don't say anything about the actual effectiveness of policies [ENDQ] [NEWLINE] I actually agree with you here. Libertarians are notorious for not really thinking through the practicality of their policies. A great deal of which are idealistic and most likely wouldn't work in the real world. But we really don't have much in the way of tested Libertarian ideas, so it makes it difficult to gauge the practicality. Most people would dismiss Libertarian ideas simply because they are so used to their current situation of being governed. For example most people completely dismiss the notion of a country with very limited government because they are boggled at the idea of a country without a strong government. </s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's<mask> nonsensical to say that I didn't buy the data<mask> I have a copy of the data that I can copy and manipulate without the control or consent of the person selling me the data. [ENDQ] [NEWLINE] It isn't nonsensical<mask> you consider that human civilization is more than just the universe and its physical laws. You're taking a ludicrously reductionist perspective which ignores societal context - a context which is needed to explain<mask> the thing came to be in the first place. [NEWLINE] [NEWLINE] The rights we give content creators encourage them to produce their work and make it available to the general public. Yes, there will always be some who will make art without demand for anything else.<mask>, such art is only a small amount of the total made, and it leaves out much of our civilizations' greatest works. Furthermore,<mask><mask><mask> art has existed, it has enjoyed very close ties to compensation for that art.</s><pad>
Label encoding: <s> [STARTQ] It's also nonsensical to say that I didn't buy the data when I have a copy of the data that I can copy and manipulate without the control or consent of the person selling me the data. [ENDQ] [NEWLINE] It isn't nonsensical when you consider that human civilization is more than just the universe and its physical laws. You're taking a ludicrously reductionist perspective which ignores societal context - a context which is needed to explain how the thing came to be in the first place. [NEWLINE] [NEWLINE] The rights we give content creators encourage them to produce their work and make it available to the general public. Yes, there will always be some who will make art without demand for anything else. However, such art is only a small amount of the total made, and it leaves out much of our civilizations' greatest works. Furthermore, as long as art has existed, it has enjoyed very close ties to compensation for that art.</s><pad>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So you would ban private schools, and things that involve money education inequality? This would lead to a war on "violations of education equality." The war would include: tutors, after school classes, summer classes, requesting specific teachers, moving near better schools, home schooling/ parental schooling, subject specific schooling, alternative schooling, over seas schooling, grade skipping/ early graduation and scholarships/ internships. These are all things that, usually parents, decide to do different with their child than others, to promote education, and you can't allow parent that have more time off from work<mask> they can afford it, or parents that can offer more money to education, to change their childs education from the state provided education. You can not allow any of these<mask> you wish to eliminate education inequality between rich and poor. Do you really want to jail/fine people for trying to provide these thing to their children?</s>
Label encoding: <s>So you would ban private schools, and things that involve money education inequality? This would lead to a war on "violations of education equality." The war would include: tutors, after school classes, summer classes, requesting specific teachers, moving near better schools, home schooling/ parental schooling, subject specific schooling, alternative schooling, over seas schooling, grade skipping/ early graduation and scholarships/ internships. These are all things that, usually parents, decide to do different with their child than others, to promote education, and you can't allow parent that have more time off from work because they can afford it, or parents that can offer more money to education, to change their childs education from the state provided education. You can not allow any of these if you wish to eliminate education inequality between rich and poor. Do you really want to jail/fine people for trying to provide these thing to their children?</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a difference between not standing in someone's way<mask> they make a bad decision and actively helping them. [NEWLINE] [NEWLINE] <mask> your friend wants to drop out of college<mask> having a decent GPA and no other prospects, do you say, "<mask><mask> that's a bad choice<mask> it's your choice," or do you say, "Here, I'll print out the paperwork for you?"<mask> your friend wants to marry someone you think is an abusive psycho, do you say, "I don't think she's right for you<mask> I hope I'm wrong," or do you say, "Here, you can borrow my grandmother's ring to propose with. I booked a moonlight cruise for you to take her on<mask> you pop the question?" [NEWLINE] [NEWLINE] Same thing for sleeping with someone in a monogamous relationship. You're not just allowing them to make a bad choice, you're facilitating it. </s>
Label encoding: <s>There's a difference between not standing in someone's way as they make a bad decision and actively helping them. [NEWLINE] [NEWLINE] If your friend wants to drop out of college despite having a decent GPA and no other prospects, do you say, " I think that's a bad choice but it's your choice," or do you say, "Here, I'll print out the paperwork for you?" If your friend wants to marry someone you think is an abusive psycho, do you say, "I don't think she's right for you but I hope I'm wrong," or do you say, "Here, you can borrow my grandmother's ring to propose with. I booked a moonlight cruise for you to take her on when you pop the question?" [NEWLINE] [NEWLINE] Same thing for sleeping with someone in a monogamous relationship. You're not just allowing them to make a bad choice, you're facilitating it. </s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Its actually quite common in depression to have to switch medication several times before you find one that catches on,<mask> i would advise you to look into the ones you haven't tried<mask>. [NEWLINE] [NEWLINE] <mask> don't expect to much with counselors,<mask> they can teach you ways to cope and channel and even understand<mask> and<mask> it happens they can't simply talk you out of it<mask> its not a conscious action. [NEWLINE] [NEWLINE] <mask><mask> even small coping mechanisms help in the long run, <mask> they turn into habits, (think of it<mask> driving a car, you no longer have to actively look which gear you shift into you just automatically do<mask> to adapt to the road,) [NEWLINE] [NEWLINE] and don't forget your only 22, even<mask> it takes you another 3 years to get to a place<mask> your happy or at least content with your life, you'd still have most of your live to enjoy [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Its actually quite common in depression to have to switch medication several times before you find one that catches on, so i would advise you to look into the ones you haven't tried yet. [NEWLINE] [NEWLINE] Also don't expect to much with counselors, while they can teach you ways to cope and channel and even understand what and why it happens they can't simply talk you out of it because its not a conscious action. [NEWLINE] [NEWLINE] however even small coping mechanisms help in the long run,  as they turn into habits, (think of it as driving a car, you no longer have to actively look which gear you shift into you just automatically do so to adapt to the road,) [NEWLINE] [NEWLINE] and don't forget your only 22, even if it takes you another 3 years to get to a place where your happy or at least content with your life, you'd still have most of your live to enjoy [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know this, of course democrats are in it for the money. I was just arguing against the argument you just made. I was literally turning the argument<mask> to<mask> you said that profit companies are essentially the only motivating factor in secular communities. Democrats<mask> are not the same<mask> liberals which is<mask> you may be noticing<mask> much hate against the democrats on this website, almost rivaling the vitriol against the republicans. [NEWLINE] [NEWLINE] The United States Constitution is completely secular. The community is not,<mask> the governing rules are secular. The pledge of allegiance is one of the worst arguments of all time, did you know that one nation under god wasn't added until the cold war? (source is religious) [URL] [NEWLINE] [NEWLINE] "Morality" is common place in all animals,<mask> I cannot technically describe morality to you unless a definition of it is given. I can<mask> show you this [URL].</s>
Label encoding: <s>I know this, of course democrats are in it for the money. I was just arguing against the argument you just made. I was literally turning the argument as to how you said that profit companies are essentially the only motivating factor in secular communities. Democrats however are not the same as liberals which is why you may be noticing so much hate against the democrats on this website, almost rivaling the vitriol against the republicans. [NEWLINE] [NEWLINE] The United States Constitution is completely secular. The community is not, but the governing rules are secular. The pledge of allegiance is one of the worst arguments of all time, did you know that one nation under god wasn't added until the cold war? (source is religious) [URL] [NEWLINE] [NEWLINE] "Morality" is common place in all animals, but I cannot technically describe morality to you unless a definition of it is given. I can however show you this [URL].</s>
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry I upset you<mask> much. I value teachers. I want them to be paid better for the job they do.<mask><mask><mask> there are numerous bad teachers out there that are not helping the system, and nothing tends to happen to them, either. [NEWLINE] [NEWLINE] In regards to the administration comment, my friend, who's a teacher, told me that<mask> he was not on good terms with the principal *on a personal level* it was adversely affecting his career at that school.<mask> I was suggesting something like that should not be allowed to take place. [NEWLINE] [NEWLINE] <mask> it's any consolation, you have sort of come to this party a little late, and I have more-or-less CMV. I've asked a few others this already, and I'll ask you, do you have any ideas of<mask> to improve the status quo?<mask> do you feel the major flaws lie?</s>
Label encoding: <s>I'm sorry I upset you so much. I value teachers. I want them to be paid better for the job they do. But I think there are numerous bad teachers out there that are not helping the system, and nothing tends to happen to them, either. [NEWLINE] [NEWLINE] In regards to the administration comment, my friend, who's a teacher, told me that because he was not on good terms with the principal *on a personal level* it was adversely affecting his career at that school. So I was suggesting something like that should not be allowed to take place. [NEWLINE] [NEWLINE] If it's any consolation, you have sort of come to this party a little late, and I have more-or-less CMV. I've asked a few others this already, and I'll ask you, do you have any ideas of how to improve the status quo? Where do you feel the major flaws lie?</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] it's consensus in the field [ENDQ] [NEWLINE] <mask> field?<mask> consensus? Who? [NEWLINE] [NEWLINE] You got nothing to prove me wrong. [NEWLINE] [NEWLINE] I have: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [STARTQ] 18% were raped by a stranger. [ENDQ] [NEWLINE] Which<mask> we are talking about in this instance. Not the 82% that is done by someone we know. [NEWLINE] [NEWLINE] Or I have this: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [STARTQ] Research has found that the vast majority of rapes are planned. Rape is the responsibility of the rapist alone. Women, children and men of every age, physical type and demeanor are raped. **Opportunity is the most important factor determining<mask> a given rapist will rape.** [ENDQ] [NEWLINE] The last line is the place card here. It is a crime of the moment. [NEWLINE] [NEWLINE] Keep telling me I am wrong. Until you bring something to the table, I am more right than you are.</s>
Label encoding: <s> [STARTQ] it's consensus in the field [ENDQ] [NEWLINE] What field? What consensus? Who? [NEWLINE] [NEWLINE] You got nothing to prove me wrong. [NEWLINE] [NEWLINE] I have: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [STARTQ] 18% were raped by a stranger. [ENDQ] [NEWLINE] Which what we are talking about in this instance. Not the 82% that is done by someone we know. [NEWLINE] [NEWLINE] Or I have this: [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] [STARTQ] Research has found that the vast majority of rapes are planned. Rape is the responsibility of the rapist alone. Women, children and men of every age, physical type and demeanor are raped. **Opportunity is the most important factor determining when a given rapist will rape.** [ENDQ] [NEWLINE] The last line is the place card here. It is a crime of the moment. [NEWLINE] [NEWLINE] Keep telling me I am wrong. Until you bring something to the table, I am more right than you are.</s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Keep in mind that for this discussion, OP has stipulated that a fetus is a person. [NEWLINE] [NEWLINE] I don't agree with the assumptions that you are making. Personhood is not defined by level of intelligence, brain function, or development level.<mask> is your definition of personhood,<mask> you've clearly indicated that a non-person is a "blob of cells" that can be terminated at will? [NEWLINE] [NEWLINE] I define personhood<mask> a any living collection of cells that is genetically human is a person. Of course, I realize that many people define it differently.<mask><mask> you're going to do<mask>, you need to have a clear definition that addresses people in a coma, people in a vegetative state, a baby born with only a brain stem, etc. Of course, it is logically coherent to say those aren't people,<mask> most aren't willing to say that. </s>
Label encoding: <s>Keep in mind that for this discussion, OP has stipulated that a fetus is a person. [NEWLINE] [NEWLINE] I don't agree with the assumptions that you are making. Personhood is not defined by level of intelligence, brain function, or development level. What is your definition of personhood, where you've clearly indicated that a non-person is a "blob of cells" that can be terminated at will? [NEWLINE] [NEWLINE] I define personhood as a any living collection of cells that is genetically human is a person. Of course, I realize that many people define it differently. But if you're going to do so, you need to have a clear definition that addresses people in a coma, people in a vegetative state, a baby born with only a brain stem, etc. Of course, it is logically coherent to say those aren't people, but most aren't willing to say that. </s>
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I was pretty clear that was a separate argument.  It really has nothing to do with this at all. [NEWLINE] [NEWLINE] Consent for the purpose of this argument is individual.  The child has not consented to immunizations.  The parent has "legally consented" to such a thing. The child has not consented. [NEWLINE] [NEWLINE] This only exists<mask> a legal construct<mask> such conduct is legally permissible within the system.  To be clear again, this has nothing to do with the debate here.  I am merely recognizing legal consent<mask> valid for the system and immunizations<mask> serving a purpose in society. [NEWLINE] [NEWLINE] Not saying that a child is actually consenting.  They're not on any ethical level doing such a thing.  They are physically violated against their will for the will of the majority.  Such a thing is done legally through the use of implied consent. </s>
Label encoding: <s>I was pretty clear that was a separate argument.  It really has nothing to do with this at all. [NEWLINE] [NEWLINE] Consent for the purpose of this argument is individual.  The child has not consented to immunizations.  The parent has "legally consented" to such a thing. The child has not consented. [NEWLINE] [NEWLINE] This only exists as a legal construct so such conduct is legally permissible within the system.  To be clear again, this has nothing to do with the debate here.  I am merely recognizing legal consent as valid for the system and immunizations as serving a purpose in society. [NEWLINE] [NEWLINE] Not saying that a child is actually consenting.  They're not on any ethical level doing such a thing.  They are physically violated against their will for the will of the majority.  Such a thing is done legally through the use of implied consent. </s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many of those items work in a large city,<mask> everything is relatively close, and lots of people will be going in teh same direction, makign it efficient.<mask> I don't live in a large city. [NEWLINE] [NEWLINE] I'm not going to be able to efficiently take public transportation on my morning routine of: [NEWLINE] - drop off kid 1 at preschool [NEWLINE] - drop off kid 2 at (carpool house)<mask> they will in turn drive kid 2 to school [NEWLINE] - drive to work [NEWLINE] - use my car during the day at work to get to various job sites, go to meetings in other cities, etc [NEWLINE] - drive home and pick up kid 1 from preschool (kid 2 gets home via a different way) [NEWLINE] [NEWLINE] To try and orchestrate that type of morning routine around public transportation would be a nightmare, not to mention it still wouldn't work for my job. </s>
Label encoding: <s>Many of those items work in a large city, where everything is relatively close, and lots of people will be going in teh same direction, makign it efficient. But I don't live in a large city. [NEWLINE] [NEWLINE] I'm not going to be able to efficiently take public transportation on my morning routine of: [NEWLINE] - drop off kid 1 at preschool [NEWLINE] - drop off kid 2 at (carpool house) where they will in turn drive kid 2 to school [NEWLINE] - drive to work [NEWLINE] - use my car during the day at work to get to various job sites, go to meetings in other cities, etc [NEWLINE] - drive home and pick up kid 1 from preschool (kid 2 gets home via a different way) [NEWLINE] [NEWLINE] To try and orchestrate that type of morning routine around public transportation would be a nightmare, not to mention it still wouldn't work for my job. </s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A "nation" is just an abstract. It doesn't exist except<mask> a useful linguistic tool. Every single person in the country is switched out every generation to say otherwise is ignorance. [NEWLINE] [NEWLINE] And<mask> someone feels "troubled" that one day French people won't be primarily white that someone needs to reevaluate their values. Who gives a shit<mask> a bunch of Germans speak Turkish? Are you sad that almost nobody speaks Gaelic in Ireland anymore? Are you mad that the word "algebra" (arabic) is<mask> common in English? [NEWLINE] [NEWLINE] You keep asserting that cultures and populations changing is bad<mask> you have<mask> to explain *<mask> *. Oh people like it the way it is? Well that sucks for them<mask> it's changing no matter<mask>. [NEWLINE] [NEWLINE] <mask> people "kept things the way they were" segregation would still exist and being gay would be a crime.</s>
Label encoding: <s>A "nation" is just an abstract. It doesn't exist except as a useful linguistic tool. Every single person in the country is switched out every generation to say otherwise is ignorance. [NEWLINE] [NEWLINE] And if someone feels "troubled" that one day French people won't be primarily white that someone needs to reevaluate their values. Who gives a shit if a bunch of Germans speak Turkish? Are you sad that almost nobody speaks Gaelic in Ireland anymore? Are you mad that the word "algebra" (arabic) is so common in English? [NEWLINE] [NEWLINE] You keep asserting that cultures and populations changing is bad but you have yet to explain * why *. Oh people like it the way it is? Well that sucks for them because it's changing no matter what. [NEWLINE] [NEWLINE] If people "kept things the way they were" segregation would still exist and being gay would be a crime.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ok well i am going with original spoons and forks vs. original chopsticks. the original chopsticks were the first disposable easy to produce eating instrument. the spoons and forks were harder to craft and not<mask> disposable. this makes chopsticks superior.<mask>? [NEWLINE] [NEWLINE] prions man prions.<mask> bacteria, viruses and other microbiota. you see, you can't wash all of that nasty stuff off of the spoons and forks<mask> easily, and a disposable eating utensil is much better at curbing the spread of disease (which historically is the number 1 killer of humans). [NEWLINE] [NEWLINE] now disposable spoons and forks are made of cheap crappy plastics that are hard to recycle (and unprofitable to do<mask> ).<mask> they end up in landfills and become a part of the environmental problem. most disposable chopsticks are made of wood and are biodegradable. </s><pad>
Label encoding: <s>ok well i am going with original spoons and forks vs. original chopsticks. the original chopsticks were the first disposable easy to produce eating instrument. the spoons and forks were harder to craft and not as disposable. this makes chopsticks superior. why? [NEWLINE] [NEWLINE] prions man prions. also bacteria, viruses and other microbiota. you see, you can't wash all of that nasty stuff off of the spoons and forks so easily, and a disposable eating utensil is much better at curbing the spread of disease (which historically is the number 1 killer of humans). [NEWLINE] [NEWLINE] now disposable spoons and forks are made of cheap crappy plastics that are hard to recycle (and unprofitable to do so ). so they end up in landfills and become a part of the environmental problem. most disposable chopsticks are made of wood and are biodegradable. </s><pad>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] the use of it medically outweights any possible stigma. [ENDQ] [NEWLINE] Really?<mask> do you determine this weight? Do you have any research or statistics, or are you just claiming it<mask> objectively true? [NEWLINE] [NEWLINE] <mask><mask> the [study by the National Center for Transgender Equality and the National Gay and Lesbian Task Force]( [URL].pdf) (pdf), 19% of surveyed trans people have been refused medical care (pg 73), 28% have been verbally harassed in a medical setting (pg 74), and 2% report being physically attacked in a doctor's office (pg 74). [NEWLINE] [NEWLINE] Do you have evidence that more than 2% of trans people are ever in a medical situation<mask> their health or life are endangered by their doctor not knowing their assigned sex?<mask> not, I don't think you get to just *declare*<mask> fact that "any possible stigma" is outweighed.</s>
Label encoding: <s> [STARTQ] the use of it medically outweights any possible stigma. [ENDQ] [NEWLINE] Really? How do you determine this weight? Do you have any research or statistics, or are you just claiming it as objectively true? [NEWLINE] [NEWLINE] According to the [study by the National Center for Transgender Equality and the National Gay and Lesbian Task Force]( [URL].pdf) (pdf), 19% of surveyed trans people have been refused medical care (pg 73), 28% have been verbally harassed in a medical setting (pg 74), and 2% report being physically attacked in a doctor's office (pg 74). [NEWLINE] [NEWLINE] Do you have evidence that more than 2% of trans people are ever in a medical situation where their health or life are endangered by their doctor not knowing their assigned sex? If not, I don't think you get to just *declare* as fact that "any possible stigma" is outweighed.</s>
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The thing is, most people who do those high-profile atrocities aren't people who reason and change their views easily,<mask><mask><mask> view it is. Just look at Anders Behring Breivik - he blew up the government's offices and attacked youth politicians in Norway. His motivation was the belief that the government, via multiculturalism, is undermining Norway and norwegian people's ability to live their lives. [NEWLINE] [NEWLINE] Does this belief hold up to scrutiny by us? No. [NEWLINE] Did it hold up to scrutiny by him? Probably,<mask> he followed through with it. [NEWLINE] [NEWLINE] Point being, his motivations are obviously clear and uncounterable to him - not to us. This is true for all cases of extreme actions,<mask><mask> ; theire motivations are beyond reach of reason,<mask><mask> they were, the perpetrators wouldn't be sure enough of their views to take such action.</s>
Label encoding: <s>The thing is, most people who do those high-profile atrocities aren't people who reason and change their views easily, regardless of what view it is. Just look at Anders Behring Breivik - he blew up the government's offices and attacked youth politicians in Norway. His motivation was the belief that the government, via multiculturalism, is undermining Norway and norwegian people's ability to live their lives. [NEWLINE] [NEWLINE] Does this belief hold up to scrutiny by us? No. [NEWLINE] Did it hold up to scrutiny by him? Probably, because he followed through with it. [NEWLINE] [NEWLINE] Point being, his motivations are obviously clear and uncounterable to him - not to us. This is true for all cases of extreme actions, I think ; theire motivations are beyond reach of reason, because if they were, the perpetrators wouldn't be sure enough of their views to take such action.</s>
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Its all about knowing who you are talking to<mask> discussing these words. For example, often nigger and other offensive words are censored from television,<mask> broadcaster's doesn't often want to offend their audience. They don't know who all is watching (Unless of course the show is meant to be offensive, shocking, or satirical.) [NEWLINE] [NEWLINE] <mask> you know your audience, or who you are talking to, and you know they won't be offended, then there is no reason to censor yourself.<mask> you are speaking to a group<mask> you don't know their sensitivities, then you should reasonably censor yourself. Even<mask> taking offense is illogical, it can only hurt whatever rational discussion you are trying to have. [NEWLINE] [NEWLINE] Its<mask> important to note that actually using offensive words in lieu of something like, "the N word" doesn't actually add any kind of benefit to a discussion.</s>
Label encoding: <s>Its all about knowing who you are talking to when discussing these words. For example, often nigger and other offensive words are censored from television, as broadcaster's doesn't often want to offend their audience. They don't know who all is watching (Unless of course the show is meant to be offensive, shocking, or satirical.) [NEWLINE] [NEWLINE] If you know your audience, or who you are talking to, and you know they won't be offended, then there is no reason to censor yourself. If you are speaking to a group where you don't know their sensitivities, then you should reasonably censor yourself. Even if taking offense is illogical, it can only hurt whatever rational discussion you are trying to have. [NEWLINE] [NEWLINE] Its also important to note that actually using offensive words in lieu of something like, "the N word" doesn't actually add any kind of benefit to a discussion.</s>
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're talking about a group of people who've studied history and looked at<mask> we screwed up with the electromagnetic ether and epicycles. [NEWLINE] [NEWLINE] People have looked into theories about whether we got our explanation of gravity wrong, like [Modified Newtonian Dynamics]( [URL] ).  Subsequent evidence like the [Bullet Cluster]( [URL] ) makes these MOND explanations less consistent with<mask> we observe than dark matter. [NEWLINE] [NEWLINE] [STARTQ] Isn't it vastly more likely that we have misunderstood some basic fact, than that the universe is filled with invisible matter spooky woo-woo energy? [ENDQ] [NEWLINE] It would be<mask> several different ways of measuring astronomical phenomena didn't point to the fact that there appears to be missing matter in the universe. People are actively trying to either find this missing matter, or develop a hypothesis that modifies gravity and is still consistent with our observations of the universe.</s>
Label encoding: <s>You're talking about a group of people who've studied history and looked at how we screwed up with the electromagnetic ether and epicycles. [NEWLINE] [NEWLINE] People have looked into theories about whether we got our explanation of gravity wrong, like [Modified Newtonian Dynamics]( [URL] ).  Subsequent evidence like the [Bullet Cluster]( [URL] ) makes these MOND explanations less consistent with what we observe than dark matter. [NEWLINE] [NEWLINE] [STARTQ] Isn't it vastly more likely that we have misunderstood some basic fact, than that the universe is filled with invisible matter spooky woo-woo energy? [ENDQ] [NEWLINE] It would be if several different ways of measuring astronomical phenomena didn't point to the fact that there appears to be missing matter in the universe. People are actively trying to either find this missing matter, or develop a hypothesis that modifies gravity and is still consistent with our observations of the universe.</s>
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The financial barrier isn't the only thing preventing everyone from having access to healthcare. It may be the largest barrier right now.<mask> imagine<mask> there is only 1 doctor in the world and he is willing to work for free, is everyone "entitled" to his care? [NEWLINE] [NEWLINE] <mask> there is only 1 dose of medication for a particular disease that plagues 10 people, are they all "entitled" to it? [NEWLINE] [NEWLINE] There are many, many things that can prevent a person from the health care that he may need. Not just money. That's<mask> health care isn't a human right. You have the right to your health<mask> you are personally capable of maintaining it, without infringing on the rights of others.<mask> beyond that, it will have to be provided by society, which functions with its own sets of rules. [NEWLINE] [NEWLINE] *edit: spelling and phrasing</s>
Label encoding: <s>The financial barrier isn't the only thing preventing everyone from having access to healthcare. It may be the largest barrier right now. But imagine if there is only 1 doctor in the world and he is willing to work for free, is everyone "entitled" to his care? [NEWLINE] [NEWLINE] If there is only 1 dose of medication for a particular disease that plagues 10 people, are they all "entitled" to it? [NEWLINE] [NEWLINE] There are many, many things that can prevent a person from the health care that he may need. Not just money. That's why health care isn't a human right. You have the right to your health as you are personally capable of maintaining it, without infringing on the rights of others. But beyond that, it will have to be provided by society, which functions with its own sets of rules. [NEWLINE] [NEWLINE] *edit: spelling and phrasing</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The difference being is that it is only within feminism, that these groups maintain legitimacy. Every other fundamental (ie. egalitarianism, capitalism) reject, and discredit those who go against the fundamental beliefs. [NEWLINE] [NEWLINE] <mask> in feminism the radical feminist, is just<mask> legitimate<mask> the Marxist feminist. And for the record I would say the exact same argument<mask> for Christianity.<mask> subsections of a group are just<mask> legitimate<mask> other subsections, then we must judge that group by<mask> makes them common and then we must evaluate those commonalities by<mask> they allow the group to do. [NEWLINE] [NEWLINE] It is important to note that you claim thy have to push a secret agenda for egalitarianism to not be about equality. And that is the key difference,<mask> egalitarianism is about equality they have to push it secretly,<mask> their claims are not legitimate within the theory and they would be discredited. </s>
Label encoding: <s>The difference being is that it is only within feminism, that these groups maintain legitimacy. Every other fundamental (ie. egalitarianism, capitalism) reject, and discredit those who go against the fundamental beliefs. [NEWLINE] [NEWLINE] However in feminism the radical feminist, is just as legitimate as the Marxist feminist. And for the record I would say the exact same argument as for Christianity. If subsections of a group are just as legitimate as other subsections, then we must judge that group by what makes them common and then we must evaluate those commonalities by what they allow the group to do. [NEWLINE] [NEWLINE] It is important to note that you claim thy have to push a secret agenda for egalitarianism to not be about equality. And that is the key difference, because egalitarianism is about equality they have to push it secretly, because their claims are not legitimate within the theory and they would be discredited. </s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I live a suburban neighborhood<mask> I can get away with it. [ENDQ] [NEWLINE] <mask> a fellow 'burb dweller I'm asserting that no, you can't.  It is even worse for them out there<mask><mask>.  Not only is there a higher tendency to have animals that see them<mask> prey,<mask> your neighbors might have dogs that aren't very forgiving.  My dog has mauled a couple of cats that our neighbors let run free in our neighborhood.  No, I don't endorse this behavior<mask> I am<mask> not going to curtail it.  Mostly<mask> the toms spray my wooden fence<mask> it smells like hell out there and they get into the trash can on the patio. [NEWLINE] [NEWLINE] Of course the neighbors freaked (same owners for both cats),<mask> they have little to no recourse<mask> I am properly confining my pets.</s>
Label encoding: <s> [STARTQ] But I live a suburban neighborhood so I can get away with it. [ENDQ] [NEWLINE] As a fellow 'burb dweller I'm asserting that no, you can't.  It is even worse for them out there I think.  Not only is there a higher tendency to have animals that see them as prey, but your neighbors might have dogs that aren't very forgiving.  My dog has mauled a couple of cats that our neighbors let run free in our neighborhood.  No, I don't endorse this behavior but I am also not going to curtail it.  Mostly because the toms spray my wooden fence so it smells like hell out there and they get into the trash can on the patio. [NEWLINE] [NEWLINE] Of course the neighbors freaked (same owners for both cats), but they have little to no recourse because I am properly confining my pets.</s>
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0743, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think he was justified in doing that, no. [NEWLINE] [NEWLINE] One of the things that irked me about Stargate Atlantis, was the degree to which the IOA (the bureaucratic international oversight body for the Stargate Program) disapproved and criticised pretty much everything that the Atlantis crew did. [NEWLINE] [NEWLINE] They apparently did not understand the fact that said crew were a long way from Earth, and<mask><mask><mask> were making rules for<mask> they did things, on the basis of the situation that they were in.  The IOA,<mask><mask><mask><mask>, were judging Atlantis on the basis of their own situation on Earth, and it just didn't work. [NEWLINE] [NEWLINE] <mask><mask><mask> there needs to be a balance.  On the one hand you have protocol which doesn't change at all,<mask> there needs to be room for operational flexibility on the other.</s>
Label encoding: <s>I don't think he was justified in doing that, no. [NEWLINE] [NEWLINE] One of the things that irked me about Stargate Atlantis, was the degree to which the IOA (the bureaucratic international oversight body for the Stargate Program) disapproved and criticised pretty much everything that the Atlantis crew did. [NEWLINE] [NEWLINE] They apparently did not understand the fact that said crew were a long way from Earth, and as a result were making rules for how they did things, on the basis of the situation that they were in.  The IOA, on the other hand, were judging Atlantis on the basis of their own situation on Earth, and it just didn't work. [NEWLINE] [NEWLINE] So I think there needs to be a balance.  On the one hand you have protocol which doesn't change at all, but there needs to be room for operational flexibility on the other.</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let us distinguish between parked and your other two examples. [NEWLINE] [NEWLINE] A parked car cannot and should not move.  Others do not expect it to move.  It is safe to text<mask> parked. [NEWLINE] [NEWLINE] A car at a light or in traffic is expected to move<mask> circumstances change. <mask> you are texting you are less likely to observe those circumstances.  You may well be an unmoving object at a time<mask> other cars expect you to be a moving object.  This can be a hazard. [NEWLINE] [NEWLINE] [STARTQ] the worst that could happen is that I don't notice the light change immediately and the person behind me gets mildly annoyed. [ENDQ] [NEWLINE] No, the worst that could happen is that you don't notice the light change for ten seconds, there was no person behind you (or she has already gone around you), and a car crashes speeding into you.</s>
Label encoding: <s>Let us distinguish between parked and your other two examples. [NEWLINE] [NEWLINE] A parked car cannot and should not move.  Others do not expect it to move.  It is safe to text while parked. [NEWLINE] [NEWLINE] A car at a light or in traffic is expected to move when circumstances change.  If you are texting you are less likely to observe those circumstances.  You may well be an unmoving object at a time when other cars expect you to be a moving object.  This can be a hazard. [NEWLINE] [NEWLINE] [STARTQ] the worst that could happen is that I don't notice the light change immediately and the person behind me gets mildly annoyed. [ENDQ] [NEWLINE] No, the worst that could happen is that you don't notice the light change for ten seconds, there was no person behind you (or she has already gone around you), and a car crashes speeding into you.</s>
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0700, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Frankly, the legal mechanism to enforce this is infeasible. It's more government oversight over personal matters--a proposition that would see stiff opposition from a wide variety of people. Not a chance in hell that it would pass into law. [NEWLINE] [NEWLINE] Really, this debate is mental masterbation. [NEWLINE] [NEWLINE] A smarter and more practical solution is fund research to give men more birth control options--ie a pill that would defertilize sperm. [NEWLINE] [NEWLINE] Men who take such a pill would have substantially more control over their sex lives, which would provide much more peace of mind than an onerous legal framework like this. Abortions would drop, unexpected pregnancies would drop, court cases would be far less common. With more control, the problem becomes way less pronounced. [NEWLINE] [NEWLINE] Prevention over reaction--that's the real world solution for this problem.</s>
Label encoding: <s>Frankly, the legal mechanism to enforce this is infeasible. It's more government oversight over personal matters--a proposition that would see stiff opposition from a wide variety of people. Not a chance in hell that it would pass into law. [NEWLINE] [NEWLINE] Really, this debate is mental masterbation. [NEWLINE] [NEWLINE] A smarter and more practical solution is fund research to give men more birth control options--ie a pill that would defertilize sperm. [NEWLINE] [NEWLINE] Men who take such a pill would have substantially more control over their sex lives, which would provide much more peace of mind than an onerous legal framework like this. Abortions would drop, unexpected pregnancies would drop, court cases would be far less common. With more control, the problem becomes way less pronounced. [NEWLINE] [NEWLINE] Prevention over reaction--that's the real world solution for this problem.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes Lnozell,<mask> you meet the basic criteria for a sociopath (a person who can't feel empathy, has no conscience, and<mask> manifests many other anti-social personality disorder issues, such<mask> being manipulative) and you see,<mask> you said in your own words, NOTHING wrong with you, then you're bad.<mask> a sociopath, you'd be perfectly capable raping and killing someone without feeling bad. To most civilized people, that kind of mental disconnect is something that makes you unfit to live in normal society.<mask> apparently to you your mental incapability to feel empathy isn't a problem at all.<mask> yes, in my humble opinion, you're a bad person. And I honestly think that<mask> everyone was the exact opposite of you (emotional empaths) the world would be a perfect place free from all evil.</s><pad>
Label encoding: <s>Yes Lnozell, if you meet the basic criteria for a sociopath (a person who can't feel empathy, has no conscience, and also manifests many other anti-social personality disorder issues, such as being manipulative) and you see, as you said in your own words, NOTHING wrong with you, then you're bad. As a sociopath, you'd be perfectly capable raping and killing someone without feeling bad. To most civilized people, that kind of mental disconnect is something that makes you unfit to live in normal society. But apparently to you your mental incapability to feel empathy isn't a problem at all. So yes, in my humble opinion, you're a bad person. And I honestly think that if everyone was the exact opposite of you (emotional empaths) the world would be a perfect place free from all evil.</s><pad>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The point you're making is too focused on<mask> we view those donating, I feel.<mask> it comes to charity, it's not a contest between benefactors to appear the most generous,<mask> an attempt to alleviate some problem that is not otherwise being solved. Donating to a soup kitchen is about feeding the hungry, and<mask> the absurdly rich can help feed mountains more than you or I can for a smaller percentage of their income, then they certainly deserve praise. [NEWLINE] [NEWLINE] I don't have sources and specific information with me,<mask> I believe large donations from a significantly few numbers of well off people is<mask> make up a *huge* portion of all charitable donations, period.<mask> charity is a small percentage of the rich's budget, the rich are a large percentage of charities' income, and I feel<mask><mask> you're discounting that contribution.</s>
Label encoding: <s>The point you're making is too focused on how we view those donating, I feel. When it comes to charity, it's not a contest between benefactors to appear the most generous, but an attempt to alleviate some problem that is not otherwise being solved. Donating to a soup kitchen is about feeding the hungry, and if the absurdly rich can help feed mountains more than you or I can for a smaller percentage of their income, then they certainly deserve praise. [NEWLINE] [NEWLINE] I don't have sources and specific information with me, but I believe large donations from a significantly few numbers of well off people is what make up a *huge* portion of all charitable donations, period. While charity is a small percentage of the rich's budget, the rich are a large percentage of charities' income, and I feel as though you're discounting that contribution.</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I was taught that the answer to this argument is that God created light *instantaneously*. Like flipping on a light switch. [ENDQ] [NEWLINE] Suppose we observe a supernova several million light years away.<mask> exactly are we seeing?<mask><mask> standard science, we are seeing light that is millions of years old just reaching us now after traveling a huge distance. [NEWLINE] [NEWLINE] <mask> is your explanation of<mask> it works? Is it just a beam of light that we're seeing that's showing us a star exploding that never actually existed? [NEWLINE] [NEWLINE] We know light doesn't travel instantaneously (without divine intervention). Do you have an explanation of distant starlight that doesn't require constant divine intervention (<mask> opposed to a one-time miracle) or nonexistent stars?<mask><mask>, I'd love to hear it. I don't understand<mask> instantaneous creation solves any problems.</s>
Label encoding: <s> [STARTQ] I was taught that the answer to this argument is that God created light *instantaneously*. Like flipping on a light switch. [ENDQ] [NEWLINE] Suppose we observe a supernova several million light years away. What exactly are we seeing? According to standard science, we are seeing light that is millions of years old just reaching us now after traveling a huge distance. [NEWLINE] [NEWLINE] What is your explanation of how it works? Is it just a beam of light that we're seeing that's showing us a star exploding that never actually existed? [NEWLINE] [NEWLINE] We know light doesn't travel instantaneously (without divine intervention). Do you have an explanation of distant starlight that doesn't require constant divine intervention ( as opposed to a one-time miracle) or nonexistent stars? If so, I'd love to hear it. I don't understand how instantaneous creation solves any problems.</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am not having much luck finding any studies on either side of this issue, and<mask><mask> it's<mask> it's very difficult to do a controlled study using this one variable. [NEWLINE] [NEWLINE] I *am* finding lots of studies about "attribution": women are even more likely to blame the victim than men<mask> they are shown pictures of provocatively dressed and modestly dressed women and told that they are both rape victims. [NEWLINE] [NEWLINE] And,<mask> suspected, I found a study about date rape which confirms our shared view that it's more likely in that scenario...<mask> *only slightly*. The strongest corollary was age of the victim's first sexual experience, oddly. [NEWLINE] [NEWLINE] <mask> I'm still digging for a stranger-rape study that addresses this topic other than the one I cited.  I'll get back to you in a separate reply.</s>
Label encoding: <s>I am not having much luck finding any studies on either side of this issue, and I think it's because it's very difficult to do a controlled study using this one variable. [NEWLINE] [NEWLINE] I *am* finding lots of studies about "attribution": women are even more likely to blame the victim than men if they are shown pictures of provocatively dressed and modestly dressed women and told that they are both rape victims. [NEWLINE] [NEWLINE] And, as suspected, I found a study about date rape which confirms our shared view that it's more likely in that scenario... although *only slightly*. The strongest corollary was age of the victim's first sexual experience, oddly. [NEWLINE] [NEWLINE] But I'm still digging for a stranger-rape study that addresses this topic other than the one I cited.  I'll get back to you in a separate reply.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0472, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0683, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No,<mask> this isn't the Johnson family next door.  It's the interaction of nation states that control the fates of billions of lives. [NEWLINE] [NEWLINE] Ignore the information about domestic spying, leave U.S. spying on foreign nations and the intelligence that foreign nations conduct against each other.  This stuff is exactly the job we've tasked to the intelligence services and one that we all acknowledge requires secrecy. [NEWLINE] [NEWLINE] Are you comfortable in a world<mask> low- to mid-level contractors are free to independently decide whether foreign intelligence operations are compromised? <mask> a low-level contractor can potential disrupt high-level foreign relations? [NEWLINE] [NEWLINE] Make no mistake, Snowden *could* have exercised discretion and blacked out everything not invoking domestic spying.  He didn't, and abdicated that responsibility to journalists who lack security expertise or clearance.  </s>
Label encoding: <s>No, because this isn't the Johnson family next door.  It's the interaction of nation states that control the fates of billions of lives. [NEWLINE] [NEWLINE] Ignore the information about domestic spying, leave U.S. spying on foreign nations and the intelligence that foreign nations conduct against each other.  This stuff is exactly the job we've tasked to the intelligence services and one that we all acknowledge requires secrecy. [NEWLINE] [NEWLINE] Are you comfortable in a world where low- to mid-level contractors are free to independently decide whether foreign intelligence operations are compromised?  Where a low-level contractor can potential disrupt high-level foreign relations? [NEWLINE] [NEWLINE] Make no mistake, Snowden *could* have exercised discretion and blacked out everything not invoking domestic spying.  He didn't, and abdicated that responsibility to journalists who lack security expertise or clearance.  </s>
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Freedom of speech doesn't actually stop me from beating the shit out of you for something you say,<mask>. [ENDQ] [NEWLINE] The First Amendment isn't going to swoop in and keep the government from physically attacking you either.  It protects you from legal repercussion and punishes the government<mask> it attacks you. [NEWLINE] [NEWLINE] [STARTQ] Assault and battery laws do that,<mask> it's never legal to violently attack someone. [ENDQ] [NEWLINE] That's not true. <mask> you are attacking me or something else, I have the right to try to stop you. <mask> you are attacking my property I have the right to attempt to remove or to call the police to remove you, and they would use physical and legal force to do<mask>. [NEWLINE] [NEWLINE] <mask> you saying something that<mask><mask> with is not one of the exceptions.  That's<mask> free speech is.</s>
Label encoding: <s> [STARTQ] Freedom of speech doesn't actually stop me from beating the shit out of you for something you say, though. [ENDQ] [NEWLINE] The First Amendment isn't going to swoop in and keep the government from physically attacking you either.  It protects you from legal repercussion and punishes the government when it attacks you. [NEWLINE] [NEWLINE] [STARTQ] Assault and battery laws do that, because it's never legal to violently attack someone. [ENDQ] [NEWLINE] That's not true.  If you are attacking me or something else, I have the right to try to stop you.  If you are attacking my property I have the right to attempt to remove or to call the police to remove you, and they would use physical and legal force to do so. [NEWLINE] [NEWLINE] But you saying something that I disagree with is not one of the exceptions.  That's what free speech is.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not<mask> it works<mask>. Your post is talking about America, not India. The symbol makes people immediately think of genocide in this country and nobody likes to think about that. That is<mask> people don't like<mask> others wear swastikas,<mask><mask> intention.<mask><mask>, most sane people would totally understand<mask> you explained the reasoning behind it,<mask> it's that gut feel people initially get. [NEWLINE] [NEWLINE] I'm not saying you shouldn't be allowed to wear it,<mask> have some tact. I read WWII books a lot and many of them have swastikas on them. I don't see the problem<mask> it's a history book, not a nazi membership card,<mask> I've learned it bothers other people.<mask> I put the books away<mask> people don't have to see the swastika. It's all about tact.</s>
Label encoding: <s>That's not how it works though. Your post is talking about America, not India. The symbol makes people immediately think of genocide in this country and nobody likes to think about that. That is why people don't like when others wear swastikas, regardless of intention. In fact, most sane people would totally understand if you explained the reasoning behind it, but it's that gut feel people initially get. [NEWLINE] [NEWLINE] I'm not saying you shouldn't be allowed to wear it, but have some tact. I read WWII books a lot and many of them have swastikas on them. I don't see the problem because it's a history book, not a nazi membership card, but I've learned it bothers other people. So I put the books away so people don't have to see the swastika. It's all about tact.</s>
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It isn't particularly important to explain the concept,<mask> I was trying to illustrate that a secure system does not rely on the strength of the password for the strength of a system<mask> users are being served in a non admin capacity. [NEWLINE] [NEWLINE] For example -<mask> you are being served by a bank - other users' passwords shouldn't be public knowledge<mask> yours is hacked. [NEWLINE] [NEWLINE] And with the website - I was mentioning that<mask> was "brute forced" was the password and not the hash (in particular - not the key). There's a crucial distinction. [NEWLINE] [NEWLINE] Hash functions being public are not just the case,<mask><mask> necessary. [NEWLINE] [NEWLINE] It is<mask> stuff like the NSA backdoor ( [URL] ) [NEWLINE] [NEWLINE] is such a big deal. [NEWLINE] [NEWLINE] Hash functions being public mean that researchers can reverse engineer ways to mount an attack on the system.</s>
Label encoding: <s>It isn't particularly important to explain the concept, but I was trying to illustrate that a secure system does not rely on the strength of the password for the strength of a system where users are being served in a non admin capacity. [NEWLINE] [NEWLINE] For example - if you are being served by a bank - other users' passwords shouldn't be public knowledge if yours is hacked. [NEWLINE] [NEWLINE] And with the website - I was mentioning that what was "brute forced" was the password and not the hash (in particular - not the key). There's a crucial distinction. [NEWLINE] [NEWLINE] Hash functions being public are not just the case, but also necessary. [NEWLINE] [NEWLINE] It is why stuff like the NSA backdoor ( [URL] ) [NEWLINE] [NEWLINE] is such a big deal. [NEWLINE] [NEWLINE] Hash functions being public mean that researchers can reverse engineer ways to mount an attack on the system.</s>
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sex is more than just orgasm. I like to have sex with my husband<mask> of the way it makes me feel. The emotions, the closeness, the love, etc.<mask> it's much easier for me to cum with a vibrator. Everything has to line up just right to cum during sex and it's not always easy (some women it is).<mask> does it mean I'm in love with my vibrator? Ugh no! It just means that it's way fucking easier. Our orgasm organ isn't sticking out of our body and easily accessible. The vibration gets it all. It's just a different kind of experience. To try to take that away from your girlfriend<mask> you're insecure about it is selfish. It's easy for you to cum every time you have sex. It's not<mask> easy for a woman. </s>
Label encoding: <s>Sex is more than just orgasm. I like to have sex with my husband because of the way it makes me feel. The emotions, the closeness, the love, etc. But it's much easier for me to cum with a vibrator. Everything has to line up just right to cum during sex and it's not always easy (some women it is). But does it mean I'm in love with my vibrator? Ugh no! It just means that it's way fucking easier. Our orgasm organ isn't sticking out of our body and easily accessible. The vibration gets it all. It's just a different kind of experience. To try to take that away from your girlfriend because you're insecure about it is selfish. It's easy for you to cum every time you have sex. It's not as easy for a woman. </s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do you think that? To me it's *obvious* that every job that can be automated, will be. Technology gets cheaper every year, and soon automation will be less expensive than a worker.<mask><mask> would a company *not* automate? [NEWLINE] [NEWLINE] Remember, companies think about profit and little else.<mask> you have a choice between a worker that is paid $50,000 every year,<mask> only works 40 hours a week, takes breaks, can get tired, can get distracted, goes on vacation, can potentially sue you for injuries, or a robot that costs $50,000 *once*, plus whatever minor maintenance costs down the road,<mask> can work 24/7/365, never sleeps, never takes breaks, never goes on vacation, never complains, never loses focus, which do you think they'll choose?</s>
Label encoding: <s>Why do you think that? To me it's *obvious* that every job that can be automated, will be. Technology gets cheaper every year, and soon automation will be less expensive than a worker. So why would a company *not* automate? [NEWLINE] [NEWLINE] Remember, companies think about profit and little else. If you have a choice between a worker that is paid $50,000 every year, but only works 40 hours a week, takes breaks, can get tired, can get distracted, goes on vacation, can potentially sue you for injuries, or a robot that costs $50,000 *once*, plus whatever minor maintenance costs down the road, but can work 24/7/365, never sleeps, never takes breaks, never goes on vacation, never complains, never loses focus, which do you think they'll choose?</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The cost of a new nuclear plant is ~$108M. For solar, the prices range from 145-260M (25% capacity factor for PV, 20% for thermal).<mask> on-shore wind is admittedly cheaper at only 86M, it can<mask> only provide 34% of the need for the country (compared to 90% for nuclear).<mask> you want off-shore wind, which can provide 37% of the need, it'll cost you 220M. [NEWLINE] [NEWLINE] <mask> for the environmental effects, yes, open-pit mining is bad for the environment. It only affects the region it's in<mask>.<mask> energy is created on-site at wind farms, there's a possibility that the overall effect would be larger<mask> more wind farms would be needed than mining operations. Energy doesn't travel well,<mask> ore does.</s>
Label encoding: <s>The cost of a new nuclear plant is ~$108M. For solar, the prices range from 145-260M (25% capacity factor for PV, 20% for thermal). While on-shore wind is admittedly cheaper at only 86M, it can also only provide 34% of the need for the country (compared to 90% for nuclear). If you want off-shore wind, which can provide 37% of the need, it'll cost you 220M. [NEWLINE] [NEWLINE] As for the environmental effects, yes, open-pit mining is bad for the environment. It only affects the region it's in though. Because energy is created on-site at wind farms, there's a possibility that the overall effect would be larger since more wind farms would be needed than mining operations. Energy doesn't travel well, but ore does.</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Criminals are hardly professional especially in the US...  Many criminals are petty and not exactly confident in<mask> they're doing especially in the case of muggings. And no offense<mask> twenty or thirty times is not a large sample size in the grand scheme of things.<mask> that is beside the point.<mask> I am saying is that having martial arts<mask> a discipline in self defense would help you not just<mask> of the last ditch effort to save your life,<mask><mask><mask> of the ability to react under distress and have a certain level of confidence in<mask> you're doing.<mask> you seem to think 1% or<mask> being a murder isn't a big deal.... That's a huge number<mask> it comes to percentage.<mask> there was an airline and it had 1% of its people die in crashes I would not take that airline. </s>
Label encoding: <s>Criminals are hardly professional especially in the US...  Many criminals are petty and not exactly confident in what they're doing especially in the case of muggings. And no offense but twenty or thirty times is not a large sample size in the grand scheme of things. However that is beside the point. What I am saying is that having martial arts as a discipline in self defense would help you not just because of the last ditch effort to save your life, but also because of the ability to react under distress and have a certain level of confidence in what you're doing. Also you seem to think 1% or so being a murder isn't a big deal.... That's a huge number when it comes to percentage. If there was an airline and it had 1% of its people die in crashes I would not take that airline. </s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See, it's not that people are necessarily disillusioned with leftism that people are supporting Donald Trump. It's more due to the fact that major segments of the Left have basically written off white and male Americans and done everything possible to alienate them. This is<mask> people are by and large fed up with "political correctness". Donald Trump is the result of this, not the cause.<mask> you're the enemy anyway,<mask> bother caring<mask> they think? Not to mention Trump,<mask> being comparatively uncharismatic has basically figured out the Right-wing populist formula of displacing economic issues on to easy scapegoats in a way that the more moderate Republicans simply haven't. [NEWLINE] [NEWLINE] <mask><mask> ; it's not<mask> much that people agree with Donald Trump<mask> much<mask> they don't care<mask> they figure they can't please them anyway.</s>
Label encoding: <s>See, it's not that people are necessarily disillusioned with leftism that people are supporting Donald Trump. It's more due to the fact that major segments of the Left have basically written off white and male Americans and done everything possible to alienate them. This is why people are by and large fed up with "political correctness". Donald Trump is the result of this, not the cause. If you're the enemy anyway, why bother caring what they think? Not to mention Trump, despite being comparatively uncharismatic has basically figured out the Right-wing populist formula of displacing economic issues on to easy scapegoats in a way that the more moderate Republicans simply haven't. [NEWLINE] [NEWLINE] TLDR ; it's not so much that people agree with Donald Trump so much as they don't care when they figure they can't please them anyway.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Feminism is only about women's rights. Not men's.<mask> feminists say they want equality for men and women that means elevating women to be equal to men by focusing in women's rights, not men's. [ENDQ] [NEWLINE] This is exactly<mask> i'm saying,<mask> most that believe that both men and women need elavating in society<mask> they are both not equal in different aspects, would still call themselves feminists<mask> they aren't really. [NEWLINE] [NEWLINE] [STARTQ] The result is equality between men abs women. [ENDQ] [NEWLINE] No, elevating women would not<mask> elevate men to be on par with women, men are less than women in the aspects that they are lower such<mask> paternity rights, college courses, women's only events etc.<mask> feminism would not give equality. It would infact just bring about a matriarchy.</s>
Label encoding: <s> [STARTQ] Feminism is only about women's rights. Not men's. When feminists say they want equality for men and women that means elevating women to be equal to men by focusing in women's rights, not men's. [ENDQ] [NEWLINE] This is exactly what i'm saying, however most that believe that both men and women need elavating in society as they are both not equal in different aspects, would still call themselves feminists when they aren't really. [NEWLINE] [NEWLINE] [STARTQ] The result is equality between men abs women. [ENDQ] [NEWLINE] No, elevating women would not also elevate men to be on par with women, men are less than women in the aspects that they are lower such as paternity rights, college courses, women's only events etc. So feminism would not give equality. It would infact just bring about a matriarchy.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I asked the question to another poster, and I will ask it to you.<mask> a person feels gay marriage is a bad environment for a child, even believing it will cause the child emotional abuse, and a gay couple in their community wants to adopt, should they get involved,<mask> of the poor innocent child? [ENDQ] [NEWLINE] No,<mask> there is no substantial evidence to suggest that same-sex relationships are bad environments to raise children in. [NEWLINE] [NEWLINE] [STARTQ] Alternatively,<mask> a middle eastern couple in your community has a different view of gender roles, one they accept<mask> you define<mask> oppressive or even emotionally abusive, should the community get involved? [ENDQ] [NEWLINE] <mask> the man is being oppressive or emotionally abusive, yes, the authorities should get involved. [NEWLINE] [NEWLINE] I don't see<mask> these have particularly much to do with the topic at hand.</s>
Label encoding: <s> [STARTQ] I asked the question to another poster, and I will ask it to you. If a person feels gay marriage is a bad environment for a child, even believing it will cause the child emotional abuse, and a gay couple in their community wants to adopt, should they get involved, because of the poor innocent child? [ENDQ] [NEWLINE] No, because there is no substantial evidence to suggest that same-sex relationships are bad environments to raise children in. [NEWLINE] [NEWLINE] [STARTQ] Alternatively, if a middle eastern couple in your community has a different view of gender roles, one they accept but you define as oppressive or even emotionally abusive, should the community get involved? [ENDQ] [NEWLINE] If the man is being oppressive or emotionally abusive, yes, the authorities should get involved. [NEWLINE] [NEWLINE] I don't see how these have particularly much to do with the topic at hand.</s>
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0580, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've thought about this before and<mask><mask> that part of it stems from America's ethnic diversity.  In many other countries around the world, countries have a large majority of the population coming from one or maybe two ethnic backgrounds.  I'd<mask> venture to say that the cultural identities are more diverse in the US<mask>... [NEWLINE] [NEWLINE] Due to the sheer size of America, it stands in the political/cultural/economic machine's interest to have a unifying bond across the massive land mass.  I've been all over the US and<mask> you see that American flag it means something to you. <mask> we are abroad and we see that flag it feels familiar, like home. <mask><mask> it is mostly grass roots love of the flag,<mask> it certainly is an icon that helps keep the peace in this great land.  </s><pad>
Label encoding: <s>I've thought about this before and I think that part of it stems from America's ethnic diversity.  In many other countries around the world, countries have a large majority of the population coming from one or maybe two ethnic backgrounds.  I'd also venture to say that the cultural identities are more diverse in the US because... [NEWLINE] [NEWLINE] Due to the sheer size of America, it stands in the political/cultural/economic machine's interest to have a unifying bond across the massive land mass.  I've been all over the US and when you see that American flag it means something to you.  When we are abroad and we see that flag it feels familiar, like home.  I think it is mostly grass roots love of the flag, but it certainly is an icon that helps keep the peace in this great land.  </s><pad>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't vote with any expectation that my vote will change the result,<mask> statistically it won't. You are 100% correct. I<mask> agree that the electoral college and the senate are mal-apportioned (fellow Californian here). Both should be reformed. [NEWLINE] [NEWLINE] I vote<mask> an act of self-expression,<mask> to speak, of my personal values.<mask><mask> of voting<mask> similar to amateur art. Amateurs make art, music, etc. for their own personal satisfaction and perhaps to share with friends. I form my own political opinions, write a blog, and vote for the same reasons. [NEWLINE] [NEWLINE] Just to make sure the metaphor is clear: politicians can be thought of like professional artists. Their "art" is<mask> influences the world,<mask> that doesn't mean the work of amateurs is meaningless.</s>
Label encoding: <s>I don't vote with any expectation that my vote will change the result, because statistically it won't. You are 100% correct. I also agree that the electoral college and the senate are mal-apportioned (fellow Californian here). Both should be reformed. [NEWLINE] [NEWLINE] I vote as an act of self-expression, so to speak, of my personal values. I think of voting as similar to amateur art. Amateurs make art, music, etc. for their own personal satisfaction and perhaps to share with friends. I form my own political opinions, write a blog, and vote for the same reasons. [NEWLINE] [NEWLINE] Just to make sure the metaphor is clear: politicians can be thought of like professional artists. Their "art" is what influences the world, but that doesn't mean the work of amateurs is meaningless.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Scar fucked up the whole kingdom and lead to ecological destruction, his motive was to be king and opportunistically allied with Hyenas who  ate him the moment he was hurt. Scar was tyrannical once he was in power.<mask> you dont trust the end of the story,<mask> trust any of it?<mask> Satan being mentioned in less than 20% of the Bible, you are convinced it's some elaborate propaganda scheme.<mask> exactly did Satan do that was good? Fight his creator and lose? Corrupt two perfect individuals, given freewill<mask> never sinning into beinng in a sinful state characterized by suffering? Tempting humans to be evil? He is no hero, he is a petty rebel to weak to kill the king<mask> instead fucks with his creations to get back at him. Satan is pitiful</s>
Label encoding: <s>Scar fucked up the whole kingdom and lead to ecological destruction, his motive was to be king and opportunistically allied with Hyenas who  ate him the moment he was hurt. Scar was tyrannical once he was in power. If you dont trust the end of the story, why trust any of it? Besides Satan being mentioned in less than 20% of the Bible, you are convinced it's some elaborate propaganda scheme. What exactly did Satan do that was good? Fight his creator and lose? Corrupt two perfect individuals, given freewill but never sinning into beinng in a sinful state characterized by suffering? Tempting humans to be evil? He is no hero, he is a petty rebel to weak to kill the king so instead fucks with his creations to get back at him. Satan is pitiful</s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who is paying for the insurance?<mask> Hobby Lobby is not refusing to provide birth control. They are refusing to provide one type of birth control that goes against their belief system. Refusing blood transfusions is different<mask> there may be no other alternatives. One is taking away one option from a sea of other choices, the other is taking away other options. These two things are not the same.<mask> is it infringing on the legal rights of others? Hobby Lobby is not stopping people from using the drug they are just not paying for something they don't believe is right. Should I have to pay for the KKK<mask> some people want to be in it? My analogy is far-fetched<mask><mask> was yours. A business or an individual should not be forced to pay for a non-governmental institution that they do not believe in. </s>
Label encoding: <s>Who is paying for the insurance? Also Hobby Lobby is not refusing to provide birth control. They are refusing to provide one type of birth control that goes against their belief system. Refusing blood transfusions is different because there may be no other alternatives. One is taking away one option from a sea of other choices, the other is taking away other options. These two things are not the same. How is it infringing on the legal rights of others? Hobby Lobby is not stopping people from using the drug they are just not paying for something they don't believe is right. Should I have to pay for the KKK because some people want to be in it? My analogy is far-fetched but so was yours. A business or an individual should not be forced to pay for a non-governmental institution that they do not believe in. </s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&amp;#8710; [NEWLINE] I am a physicist/mathematician,<mask> that is part of my own bias; I spend many months by myself thinking of strange problems with minimal interaction with people, working in my own environment;<mask>, my job is not test based in the least.<mask><mask> that it is important to think quick on your feet, and acknowledge that memorization is important,<mask> I feel that there is too much emphasis on memorization.<mask><mask> that there should be some form of tests,<mask> I guess I am rather against the current system of having a single exam worth 90% of your mark.  I am not necessarily against tests,<mask> rather<mask> they are administered.  I believe courses should have emphasis on learning the information and applying it;<mask> I do agree a choice would be beneficial.</s><pad>
Label encoding: <s>&amp;#8710; [NEWLINE] I am a physicist/mathematician, so that is part of my own bias; I spend many months by myself thinking of strange problems with minimal interaction with people, working in my own environment; therefore, my job is not test based in the least. I agree that it is important to think quick on your feet, and acknowledge that memorization is important, but I feel that there is too much emphasis on memorization. I agree that there should be some form of tests, but I guess I am rather against the current system of having a single exam worth 90% of your mark.  I am not necessarily against tests, but rather how they are administered.  I believe courses should have emphasis on learning the information and applying it; though I do agree a choice would be beneficial.</s><pad>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Denying someone a life-saving medical procedure to punish them for life choices that are completely unrelated to their medical needs is unethical. [ENDQ] [NEWLINE] You make it sound like we have tons of organs, and that we can save every bodies life, and we are choosing not to. The reality is that the waiting list to receive life-saving organs is a lot larger than the amount of organs available. Don't you think it would be more unethical to give an organ to someone who wouldn't do the same, over giving it to someone who would? [NEWLINE] [NEWLINE] The likelihood of receiving an organ is such that<mask> you do receive one, yes it is a reward.<mask><mask> we should reward those who are contributing in their being organs,<mask> opposed to those who are contributing in making the waiting list longer, and the supply shorter. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Denying someone a life-saving medical procedure to punish them for life choices that are completely unrelated to their medical needs is unethical. [ENDQ] [NEWLINE] You make it sound like we have tons of organs, and that we can save every bodies life, and we are choosing not to. The reality is that the waiting list to receive life-saving organs is a lot larger than the amount of organs available. Don't you think it would be more unethical to give an organ to someone who wouldn't do the same, over giving it to someone who would? [NEWLINE] [NEWLINE] The likelihood of receiving an organ is such that when you do receive one, yes it is a reward. I think we should reward those who are contributing in their being organs, as opposed to those who are contributing in making the waiting list longer, and the supply shorter. [NEWLINE] </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You try to implement a system like this<mask> freeloaders have the option of doing absolutely nothing, and I guarantee you there WILL be rioting in the streets, and I will be one of the rioters. [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] <mask> unemployment drops too high, the same thing will happen, only difference is the economy will be collapsing at the same time.<mask><mask><mask><mask> there is an unlimited supply of jobs to be had, robotics are getting cheaper and cheaper, and<mask><mask><mask><mask> that we can just keep making up jobs at the same pace. [NEWLINE] [NEWLINE] [NEWLINE] <mask> unemployment doesn't drop significantly, I would agree with you, a large working class is never going to agree to pay for freeloaders,<mask> unless a lot of new jobs crop up, the economy can only take<mask> many jobs being lost.</s>
Label encoding: <s> [STARTQ] You try to implement a system like this where freeloaders have the option of doing absolutely nothing, and I guarantee you there WILL be rioting in the streets, and I will be one of the rioters. [ENDQ] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] If unemployment drops too high, the same thing will happen, only difference is the economy will be collapsing at the same time. I do not think there is an unlimited supply of jobs to be had, robotics are getting cheaper and cheaper, and I do not think that we can just keep making up jobs at the same pace. [NEWLINE] [NEWLINE] [NEWLINE] If unemployment doesn't drop significantly, I would agree with you, a large working class is never going to agree to pay for freeloaders, but unless a lot of new jobs crop up, the economy can only take so many jobs being lost.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> I mean is that the United States is an agreement between states.<mask> all states should have equal representation<mask> they are all giving up the same amount of authority to be a part of the national government. [ENDQ] [NEWLINE] You're still describing<mask> the senate came to be, not<mask> it *should still* be this way today. The OP is saying the senate should be abolished, today. Do you defend the way it functions and distorts equal representation of people today? [NEWLINE] [NEWLINE] Your investment model is not applicable. In matters of politics, of deciding<mask> we govern ourselves, we have have equal stake<mask> people.<mask> investing money some people may invest more than others,<mask> of course unequal stakes makes sense. Is there even a historical political philosopher whose views are not dictatorial or fascist who argues against equal representation of people?</s>
Label encoding: <s> [STARTQ] What I mean is that the United States is an agreement between states. Therefore all states should have equal representation because they are all giving up the same amount of authority to be a part of the national government. [ENDQ] [NEWLINE] You're still describing how the senate came to be, not why it *should still* be this way today. The OP is saying the senate should be abolished, today. Do you defend the way it functions and distorts equal representation of people today? [NEWLINE] [NEWLINE] Your investment model is not applicable. In matters of politics, of deciding how we govern ourselves, we have have equal stake as people. When investing money some people may invest more than others, so of course unequal stakes makes sense. Is there even a historical political philosopher whose views are not dictatorial or fascist who argues against equal representation of people?</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Debatable, it depends<mask> your goals are.  I'm not just going to lie about<mask><mask> on a controversial issue<mask> it offends people, I will speak my mind. [NEWLINE] [NEWLINE] <mask> I feel like offending someone is important to bettering them<mask> a person, then perhaps I will. For example, I may try to talk someone out of an unhealthy or dangerous habbit, and they may take offense to me telling them<mask> to do. [NEWLINE] [NEWLINE] <mask> for it being my responsibility,<mask><mask> to a degree, it is. There are consequences for everything. Will<mask> I say kill my friendship with someone? Will<mask> I say be used against me? Will<mask> I say cause someone anxiety or depression?<mask> I can use my intuition to decide<mask>'s best for myself and everyone else, then I will.</s>
Label encoding: <s>Debatable, it depends what your goals are.  I'm not just going to lie about my opinion on a controversial issue because it offends people, I will speak my mind. [NEWLINE] [NEWLINE] If I feel like offending someone is important to bettering them as a person, then perhaps I will. For example, I may try to talk someone out of an unhealthy or dangerous habbit, and they may take offense to me telling them what to do. [NEWLINE] [NEWLINE] As for it being my responsibility, I think to a degree, it is. There are consequences for everything. Will what I say kill my friendship with someone? Will what I say be used against me? Will what I say cause someone anxiety or depression? If I can use my intuition to decide what's best for myself and everyone else, then I will.</s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see it pretty often,<mask> I live in the South.  Racism in general is more tolerated here,<mask> even among the most liberal/progressive people in my city<mask> we'll call out a young person for saying stuff they shouldn't, old people almost always get a pass.  Caveat:<mask> they're in the public view, it changes.  Paula Deen and now this fire chief (from literally the next county over from me, wtf) got all that shit thrown at them<mask><mask> she hadn't been a famous TV personality or a public official, and instead had just been a restaurant owner or some guy at the bait shop, the worst that would've happened would've been some head-shaking facepalms and nobody would've directly called them out for<mask> they said.</s>
Label encoding: <s>I see it pretty often, but I live in the South.  Racism in general is more tolerated here, but even among the most liberal/progressive people in my city where we'll call out a young person for saying stuff they shouldn't, old people almost always get a pass.  Caveat: If they're in the public view, it changes.  Paula Deen and now this fire chief (from literally the next county over from me, wtf) got all that shit thrown at them but if she hadn't been a famous TV personality or a public official, and instead had just been a restaurant owner or some guy at the bait shop, the worst that would've happened would've been some head-shaking facepalms and nobody would've directly called them out for what they said.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask><mask> incentives should be given to vote. After all,<mask><mask><mask> laziness is the best assumption for<mask> someone did not vote then those who are not voting for a purpose are being effectively silenced. By making every effort to vote you are getting a better representation of the thoughts and feelings of those who do vote (<mask> would not have otherwise) and a better understanding of those who are *actively* staying home. [NEWLINE] [NEWLINE] Voting is communication. The multiple different reasons for the same decision (to vote or not to vote) increases the amount of noise in the communication and reduces understanding. It's better to get vague votes<mask> be able to understand intent than it is to get only high quality voters<mask> get the wrong message<mask> high quality voters just happened to be poor representations of the population<mask> a whole.</s>
Label encoding: <s>I would argue that incentives should be given to vote. After all, as long as laziness is the best assumption for why someone did not vote then those who are not voting for a purpose are being effectively silenced. By making every effort to vote you are getting a better representation of the thoughts and feelings of those who do vote ( but would not have otherwise) and a better understanding of those who are *actively* staying home. [NEWLINE] [NEWLINE] Voting is communication. The multiple different reasons for the same decision (to vote or not to vote) increases the amount of noise in the communication and reduces understanding. It's better to get vague votes but be able to understand intent than it is to get only high quality voters but get the wrong message as high quality voters just happened to be poor representations of the population as a whole.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] &gt;<mask><mask> this hypothetical third party is neither receiving the gift, nor giving it<mask> right do they have to tell the giver it's not generous enough? That seems insane. [ENDQ] [NEWLINE] [STARTQ] The thread title says we should stop calling them generous.  The third party could be a radio station reporting on it or someone in an interview or whatever.  Not sure<mask> that's insane. [ENDQ] [NEWLINE] This avoided my question almost in it's entirety, of course any radio station can call them anything they want, that's the nature media. [NEWLINE] [NEWLINE] I'm asking on<mask> authority your hypothetical third party dictates<mask> generosity is. [NEWLINE] [NEWLINE] <mask> it's arbitrary and anyone can define it differently, then the entire point is moot anyway. [NEWLINE] [NEWLINE] <mask> makes your new definition of generosity better than the current one? [NEWLINE] </s>
Label encoding: <s> [STARTQ] &gt; But if this hypothetical third party is neither receiving the gift, nor giving it what right do they have to tell the giver it's not generous enough? That seems insane. [ENDQ] [NEWLINE] [STARTQ] The thread title says we should stop calling them generous.  The third party could be a radio station reporting on it or someone in an interview or whatever.  Not sure why that's insane. [ENDQ] [NEWLINE] This avoided my question almost in it's entirety, of course any radio station can call them anything they want, that's the nature media. [NEWLINE] [NEWLINE] I'm asking on what authority your hypothetical third party dictates what generosity is. [NEWLINE] [NEWLINE] If it's arbitrary and anyone can define it differently, then the entire point is moot anyway. [NEWLINE] [NEWLINE] What makes your new definition of generosity better than the current one? [NEWLINE] </s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Whoa, thank you for catching that! You are entirely correct on the yottabyte measurement. I don't precisely recall the news articles that put that into my brain<mask> I obviously wasn't being careful enough about sources. [NEWLINE] [NEWLINE] [STARTQ] There's two sides to a case, and the defence is going to want to see any potentially relevant video [ENDQ] [NEWLINE] You are right on that<mask> well, and I can see<mask> that could lead to serious legal and storage issues<mask> it comes down to something like trying to use that video to characterize the LEO in front of a jury. Judging the relevancy of any particular data to any particular case will always remain a somewhat controversial decision. [NEWLINE] [NEWLINE] Even<mask> it is impossible today, I still believe it is something we should keep on table<mask> technology someday enables it. </s>
Label encoding: <s>Whoa, thank you for catching that! You are entirely correct on the yottabyte measurement. I don't precisely recall the news articles that put that into my brain but I obviously wasn't being careful enough about sources. [NEWLINE] [NEWLINE] [STARTQ] There's two sides to a case, and the defence is going to want to see any potentially relevant video [ENDQ] [NEWLINE] You are right on that as well, and I can see how that could lead to serious legal and storage issues if it comes down to something like trying to use that video to characterize the LEO in front of a jury. Judging the relevancy of any particular data to any particular case will always remain a somewhat controversial decision. [NEWLINE] [NEWLINE] Even if it is impossible today, I still believe it is something we should keep on table if technology someday enables it. </s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry Dune17k, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Dune17k+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry Dune17k, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=Dune17k+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your property taxes aren't the issue here. The portion of taxes that goes to education is enough to pay the teachers (A pittance), and keep the power on. Everything else is about time, donation, and level of care. [NEWLINE] [NEWLINE] <mask> you *do* have a say at a public school. Public schools are built on the backs of the Teachers willing to work hard and the parents of students willing to work with them. There is a reason PTA and District Councils exist, and its to help interface with parents to create a better school program. [NEWLINE] [NEWLINE] And I understand, I would send my kids to a private school<mask> the public school was bad.<mask><mask> everyone decided to use public schools (In a far off universe very different than this one), then the public schools would *get better*</s>
Label encoding: <s>Your property taxes aren't the issue here. The portion of taxes that goes to education is enough to pay the teachers (A pittance), and keep the power on. Everything else is about time, donation, and level of care. [NEWLINE] [NEWLINE] Because you *do* have a say at a public school. Public schools are built on the backs of the Teachers willing to work hard and the parents of students willing to work with them. There is a reason PTA and District Councils exist, and its to help interface with parents to create a better school program. [NEWLINE] [NEWLINE] And I understand, I would send my kids to a private school if the public school was bad. But if everyone decided to use public schools (In a far off universe very different than this one), then the public schools would *get better*</s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OP, I am afriad you have interpreted these results wrong. [NEWLINE] [NEWLINE] ON AVERAGE, you are correct.  Technically ON AVERAGE black people have lower IQ's than Whites, Asians higher than that, etc. [NEWLINE] That can be translated<mask> a list of average intelligence of races ordered from least to greatest intelligence. [NEWLINE] [NEWLINE] The way you phrase this<mask>, implies that the black race<mask> a whole is inferior.  No, it does not. Neil DeGrasse Tyson is an example. [NEWLINE] These results just mean that there is a higher percentage of low IQ people in the black race than in the white race. [NEWLINE] [NEWLINE] You should NOT treat anyone differently for this.  No one should be judged for race, especially<mask> you might mistake a black genius for an idiot<mask> you are prejudice.</s>
Label encoding: <s>OP, I am afriad you have interpreted these results wrong. [NEWLINE] [NEWLINE] ON AVERAGE, you are correct.  Technically ON AVERAGE black people have lower IQ's than Whites, Asians higher than that, etc. [NEWLINE] That can be translated as a list of average intelligence of races ordered from least to greatest intelligence. [NEWLINE] [NEWLINE] The way you phrase this however, implies that the black race as a whole is inferior.  No, it does not. Neil DeGrasse Tyson is an example. [NEWLINE] These results just mean that there is a higher percentage of low IQ people in the black race than in the white race. [NEWLINE] [NEWLINE] You should NOT treat anyone differently for this.  No one should be judged for race, especially since you might mistake a black genius for an idiot because you are prejudice.</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry hazybluez, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=hazybluez+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry hazybluez, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=hazybluez+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry edmiborn, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=edmiborn+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry edmiborn, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=edmiborn+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Povert is relevant<mask>, in essence, environmentalism is a luxury and poor people can't afford (many) luxuries. Now<mask> I say that environmentalism is a luxury, i DON'T mean the world will be fine without it. I certainly feel it's necessary.<mask> I mean is that an individual or a family will be fine<mask> they don't practice it. They may want to,<mask> they certainly want to feed themselves a lot more. And personally, I don't blame them<mask> they choose a TV over living greener<mask> they have some money to spend. A TV allows them to get away from the shit in their lives, environmentalism doesn't. [NEWLINE] [NEWLINE] All of this disregards the poverty -&gt; worse education -&gt; less awareness chain of reasoning. </s>
Label encoding: <s>Povert is relevant because, in essence, environmentalism is a luxury and poor people can't afford (many) luxuries. Now when I say that environmentalism is a luxury, i DON'T mean the world will be fine without it. I certainly feel it's necessary. What I mean is that an individual or a family will be fine if they don't practice it. They may want to, but they certainly want to feed themselves a lot more. And personally, I don't blame them if they choose a TV over living greener when they have some money to spend. A TV allows them to get away from the shit in their lives, environmentalism doesn't. [NEWLINE] [NEWLINE] All of this disregards the poverty -&gt; worse education -&gt; less awareness chain of reasoning. </s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's true you can't combat unconditional belief with reason.<mask> I feel like this is just a classic westerner complaint. People belive in God are irrational and<mask> useless, is<mask> your basically arguing.<mask> some one who's gone from atheist to agnostic, to now a mix of most regions<mask> you are arguing with is<mask> a religion.<mask> always sunny puts it: have you poured through the data yourself? Check<mask> evolution is true step by step, you havent? Well than you are basing your belief of evolution in faith. [NEWLINE] [NEWLINE] <mask> I'd like to argue at this point is that they're are people with strong beliefs, and some people who don't like to keep an open mind. You are angry at closed minded people, not religious or god beleiving people. </s>
Label encoding: <s>It's true you can't combat unconditional belief with reason. But I feel like this is just a classic westerner complaint. People belive in God are irrational and therefore useless, is what your basically arguing. As some one who's gone from atheist to agnostic, to now a mix of most regions what you are arguing with is also a religion. As always sunny puts it: have you poured through the data yourself? Check how evolution is true step by step, you havent? Well than you are basing your belief of evolution in faith. [NEWLINE] [NEWLINE] What I'd like to argue at this point is that they're are people with strong beliefs, and some people who don't like to keep an open mind. You are angry at closed minded people, not religious or god beleiving people. </s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry MerryWalrus, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view (<mask> minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=MerryWalrus+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Label encoding: <s>Sorry MerryWalrus, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 1\. "Direct responses to a CMV post must challenge at least one aspect of OP’s current view ( however minor), unless they are asking a clarifying question. Arguments in favor of the view OP is willing to change must be restricted to replies to comments." [See the wiki page for more information.]( [URL] #wiki_rule_1) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+1+Post+Appeal&amp;message=MerryWalrus+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s><pad>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've listed several aspects,<mask> does each affect you directly? Do you have kids who'd be in school? Do you need to depend on government healthcare? Does our military policy actually impact your life in some way? Does the broken system factor in to whether you have a nice house or can afford groceries? [NEWLINE] [NEWLINE] Your big-picture viewpoint is useful to talk about,<mask> is a privilege afforded someone who's basic needs are already met.<mask> you don't want to move to the US, it can only be<mask> you believe you're already in a better place. [NEWLINE] [NEWLINE] <mask> you lived in a facist, racists, militarized country<mask> you wouldn't even be able to post something like this without being killed or imprisoned, you'd probably take the US<mask> an alternative. Who wouldn't?</s>
Label encoding: <s>You've listed several aspects, but does each affect you directly? Do you have kids who'd be in school? Do you need to depend on government healthcare? Does our military policy actually impact your life in some way? Does the broken system factor in to whether you have a nice house or can afford groceries? [NEWLINE] [NEWLINE] Your big-picture viewpoint is useful to talk about, but is a privilege afforded someone who's basic needs are already met. If you don't want to move to the US, it can only be because you believe you're already in a better place. [NEWLINE] [NEWLINE] If you lived in a facist, racists, militarized country where you wouldn't even be able to post something like this without being killed or imprisoned, you'd probably take the US as an alternative. Who wouldn't?</s>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can call anything you like free will.  That doesn't mean it has even the slightest resemblance to<mask> other people mean by the term. [NEWLINE] [NEWLINE] Language is a consensual process, yadda yadda etc. etc. [NEWLINE] [NEWLINE] And your use of "decision" in this context is suspect.  Light hitting your retina and travelling down the optical nerve and being processed in your visual cortex isn't your decision; it's an autonomous process that doesn't require your conscious input. <mask> is breathing, and<mask> are automatic thoughts. [NEWLINE] [NEWLINE] Of course, you could just redefine all biological occurences to be "decisions" and preserve a notion of free will in that way,<mask> that would just be perverting language<mask> we agree on the fundamentals. (or do we?)</s>
Label encoding: <s>You can call anything you like free will.  That doesn't mean it has even the slightest resemblance to what other people mean by the term. [NEWLINE] [NEWLINE] Language is a consensual process, yadda yadda etc. etc. [NEWLINE] [NEWLINE] And your use of "decision" in this context is suspect.  Light hitting your retina and travelling down the optical nerve and being processed in your visual cortex isn't your decision; it's an autonomous process that doesn't require your conscious input.  So is breathing, and so are automatic thoughts. [NEWLINE] [NEWLINE] Of course, you could just redefine all biological occurences to be "decisions" and preserve a notion of free will in that way, but that would just be perverting language when we agree on the fundamentals. (or do we?)</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't believe your argument<mask> stated, and I don't imagine you believe it either.<mask> you're desperately understaffed for six months and the only resume that has come in contains a couple of typos, I doubt you'd hurt your business over a Spelling Crusade. [NEWLINE] [NEWLINE] Going further,<mask> the resume of a candidate with 20 years' experience and 4 field-related patents to his name has a handful of spelling errors, would you decline that person in favor of one with a string of two-month bartending jobs and flawless prose? [NEWLINE] [NEWLINE] The question I'd like to posit back to you is: just<mask> damning is a resume flaw in your eyes?<mask> much extra experience/charisma/recommendation would a candidate need to make up for your average grammatical error?</s>
Label encoding: <s>I don't believe your argument as stated, and I don't imagine you believe it either. If you're desperately understaffed for six months and the only resume that has come in contains a couple of typos, I doubt you'd hurt your business over a Spelling Crusade. [NEWLINE] [NEWLINE] Going further, if the resume of a candidate with 20 years' experience and 4 field-related patents to his name has a handful of spelling errors, would you decline that person in favor of one with a string of two-month bartending jobs and flawless prose? [NEWLINE] [NEWLINE] The question I'd like to posit back to you is: just how damning is a resume flaw in your eyes? How much extra experience/charisma/recommendation would a candidate need to make up for your average grammatical error?</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a hard time believing that mental health services would have actually helped this guy. Laymen like to through out "mental illness" over stuff like this without concrete solutions or true understanding of<mask> mental illness even is. [NEWLINE] [NEWLINE] For the record I don't know<mask> type of illness this guy may have - antisocial maybe? Seems like there is a severe lack of empathy going on, which is part of the problem. People with a severe lack of empathy don't tend to gain it back. They tend to end up in prison. Which leads me to -<mask> exactly are docs supposed to do about this? More funding for forensic psych would be nice,<mask> I just don't know<mask> your standard "more funding/less stigma" approach would make a lick of difference with guys like this.  </s>
Label encoding: <s>I have a hard time believing that mental health services would have actually helped this guy. Laymen like to through out "mental illness" over stuff like this without concrete solutions or true understanding of what mental illness even is. [NEWLINE] [NEWLINE] For the record I don't know what type of illness this guy may have - antisocial maybe? Seems like there is a severe lack of empathy going on, which is part of the problem. People with a severe lack of empathy don't tend to gain it back. They tend to end up in prison. Which leads me to - what exactly are docs supposed to do about this? More funding for forensic psych would be nice, but I just don't know if your standard "more funding/less stigma" approach would make a lick of difference with guys like this.  </s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I mostly agree that the civil war scenario is not worth discussing.  I just figured I'd rather local police have the deflated basketball than the military. [NEWLINE] [NEWLINE] [STARTQ] And the polarization of police and citizens is clearly happening now [ENDQ] [NEWLINE] I don't see this getting worse.  I don't see numerous botched raids.  Its always been bad between cops and African Americans,<mask> this has nothing to do with weaponry.  The average cop still looks the same<mask> he did, except he has more non-lethal weapons now. I don't have enough to go on to say, SWAT cops are too incompetent to be slightly better armed than we are. <mask>, would you change your stance<mask> all these well-armed cops had cameras on them? <mask> that's probably only ten years away.</s>
Label encoding: <s>I mostly agree that the civil war scenario is not worth discussing.  I just figured I'd rather local police have the deflated basketball than the military. [NEWLINE] [NEWLINE] [STARTQ] And the polarization of police and citizens is clearly happening now [ENDQ] [NEWLINE] I don't see this getting worse.  I don't see numerous botched raids.  Its always been bad between cops and African Americans, but this has nothing to do with weaponry.  The average cop still looks the same as he did, except he has more non-lethal weapons now. I don't have enough to go on to say, SWAT cops are too incompetent to be slightly better armed than we are.  Also, would you change your stance if all these well-armed cops had cameras on them?  Because that's probably only ten years away.</s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The argument is not that it is wrong to remove someone's rights<mask> they choose to do a certain thing, the argument is that it is wring to remove someone's right's<mask> they chose this thing in particular. [NEWLINE] [NEWLINE] The argument is that this thing,<mask><mask> whether it is a choice or not, should not be something that is oppressed. Whether or not it is a choice, in this particular subject, should not be significant information. This is not true for all situations, like<mask> committing crimes,<mask> it is true here. [NEWLINE] [NEWLINE] Like, you can't remove someone's rights for eating chocolate ice cream today.<mask><mask> they *chose* to do it, choosing to eat chocolate ice cream is not a thing that should warrant getting your rights removed in the first place.</s>
Label encoding: <s>The argument is not that it is wrong to remove someone's rights because they choose to do a certain thing, the argument is that it is wring to remove someone's right's because they chose this thing in particular. [NEWLINE] [NEWLINE] The argument is that this thing, regardless of whether it is a choice or not, should not be something that is oppressed. Whether or not it is a choice, in this particular subject, should not be significant information. This is not true for all situations, like when committing crimes, but it is true here. [NEWLINE] [NEWLINE] Like, you can't remove someone's rights for eating chocolate ice cream today. Even though they *chose* to do it, choosing to eat chocolate ice cream is not a thing that should warrant getting your rights removed in the first place.</s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So your initial counter argument is: [NEWLINE] [NEWLINE] [STARTQ] a small  sample of people who work in the tech industry does [not] explain to me<mask> or<mask> they unbiasedly prefer Apple for specific reasons. [ENDQ] [NEWLINE] You then rebut a follow up comment about<mask> these are significant persons in the tech industry by selecting *one person from the "small sample"*. This is a poor argument. [NEWLINE] [NEWLINE] [STARTQ] None of them explain<mask> it is superior for<mask> they do. [ENDQ] [NEWLINE] <mask> that's not<mask> the question asked. One can *infer* that they use the technology<mask> it provides a superior development experience. Which it absolutely does. Simply having a unix based operating system and access to a proper command line makes development on an apple significantly more pleasant than developing on a PC.</s>
Label encoding: <s>So your initial counter argument is: [NEWLINE] [NEWLINE] [STARTQ] a small  sample of people who work in the tech industry does [not] explain to me why or if they unbiasedly prefer Apple for specific reasons. [ENDQ] [NEWLINE] You then rebut a follow up comment about how these are significant persons in the tech industry by selecting *one person from the "small sample"*. This is a poor argument. [NEWLINE] [NEWLINE] [STARTQ] None of them explain why it is superior for what they do. [ENDQ] [NEWLINE] Because that's not what the question asked. One can *infer* that they use the technology because it provides a superior development experience. Which it absolutely does. Simply having a unix based operating system and access to a proper command line makes development on an apple significantly more pleasant than developing on a PC.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Pacifism is good for changing people's opinions,<mask> terrible for surviving. [ENDQ] [NEWLINE] Then perhaps you should think about it this way: saying you have no respect for pacifists,<mask> you yourself are not a pacifist, is like saying you have no respect for doctors merely<mask> you are a soldier.<mask><mask>, it would be pretty bad<mask> there was no one to resist aggressors.<mask> it's really, really helpful for society to have that one guy who never raises his voice, never gets in a fight, and would rather die than hurt someone else,<mask> everyone trusts that guy. Everyone knows that guy would take a bullet for the good of the team, and<mask><mask> there's a conflict somewhere else, that's who's the best at sorting it out. </s>
Label encoding: <s> [STARTQ] Pacifism is good for changing people's opinions, but terrible for surviving. [ENDQ] [NEWLINE] Then perhaps you should think about it this way: saying you have no respect for pacifists, when you yourself are not a pacifist, is like saying you have no respect for doctors merely because you are a soldier. I agree, it would be pretty bad if there was no one to resist aggressors. But it's really, really helpful for society to have that one guy who never raises his voice, never gets in a fight, and would rather die than hurt someone else, because everyone trusts that guy. Everyone knows that guy would take a bullet for the good of the team, and so when there's a conflict somewhere else, that's who's the best at sorting it out. </s>
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm 30. I don't drink and never have. I have friends, a loving family, a wife, 3 kids. Not drinking never stopped me.<mask> anything, it's given me a unique talking point or ice breaker at social events, and it has made me appear<mask> a man with principles (and rightly<mask> ).<mask>, at events there have been cases<mask> people have been grateful I am not drinking at all,<mask> it made them feel like they didn't need to get drunk<mask> someone else wasn't; made it easier for them to not give in to peer pressure I guess. [NEWLINE] [NEWLINE] You do<mask> you want man. It's your life, and you only got one. Don't for a second think that you're missing out on something by not drinking<mask>.</s>
Label encoding: <s>I'm 30. I don't drink and never have. I have friends, a loving family, a wife, 3 kids. Not drinking never stopped me. If anything, it's given me a unique talking point or ice breaker at social events, and it has made me appear as a man with principles (and rightly so ). Also, at events there have been cases where people have been grateful I am not drinking at all, as it made them feel like they didn't need to get drunk if someone else wasn't; made it easier for them to not give in to peer pressure I guess. [NEWLINE] [NEWLINE] You do what you want man. It's your life, and you only got one. Don't for a second think that you're missing out on something by not drinking though.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My real interest is in whether or not the sacrifices you make to tackle the eventual skill plateau in any specific aspect of your life can justify your time investment. [ENDQ] [NEWLINE] Ask Tom Brady whether he thinks it's worth the time it takes to be Tom Brady. [NEWLINE] [NEWLINE] Do you think Mozart or Beethoven regret devoting their lives to music? [NEWLINE] [NEWLINE] Do you think Einstein would have preferred to dabble more? [NEWLINE] [NEWLINE] It obviously depends on the person,<mask> for those who *want* to be Masters, it's worth it to them.  It isn't to me, which is<mask> I'm not one. [NEWLINE] [NEWLINE] <mask> it's a little silly to ask me whether<mask><mask> Tom Brady spends too much time on football - it's his call, not mine.</s><pad>
Label encoding: <s> [STARTQ] My real interest is in whether or not the sacrifices you make to tackle the eventual skill plateau in any specific aspect of your life can justify your time investment. [ENDQ] [NEWLINE] Ask Tom Brady whether he thinks it's worth the time it takes to be Tom Brady. [NEWLINE] [NEWLINE] Do you think Mozart or Beethoven regret devoting their lives to music? [NEWLINE] [NEWLINE] Do you think Einstein would have preferred to dabble more? [NEWLINE] [NEWLINE] It obviously depends on the person, but for those who *want* to be Masters, it's worth it to them.  It isn't to me, which is why I'm not one. [NEWLINE] [NEWLINE] But it's a little silly to ask me whether I think Tom Brady spends too much time on football - it's his call, not mine.</s><pad>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We fear terrorism more than, say, food poisoning<mask> there are things that we can do<mask> individuals to avoid food poisoning<mask> opposed to being involved in a terrorist attack. You can try to avoid food poisoning by not eating undercooked meat. The only way to avoid a terrorist attack is to never leave your house (and even that isn't foolproof).<mask><mask> statistically people are more likely to get food poisoning than be the victim of a terrorist attack, it is easier to avoid food poisoning than it is to avoid being a victim of a terrorist attack. It's logical to fear something that one has no control over for just that reason; there is very little,<mask> anything at all, that an individual can to do avoid it even<mask> the odds of it happening are extremely slim.</s>
Label encoding: <s>We fear terrorism more than, say, food poisoning because there are things that we can do as individuals to avoid food poisoning as opposed to being involved in a terrorist attack. You can try to avoid food poisoning by not eating undercooked meat. The only way to avoid a terrorist attack is to never leave your house (and even that isn't foolproof). Even though statistically people are more likely to get food poisoning than be the victim of a terrorist attack, it is easier to avoid food poisoning than it is to avoid being a victim of a terrorist attack. It's logical to fear something that one has no control over for just that reason; there is very little, if anything at all, that an individual can to do avoid it even if the odds of it happening are extremely slim.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the word'sexual dimorphism' might sort of be<mask> you're looking for,<mask> I don't think it really applies to very much, considering the amount of variation of<mask>'s considered 'women's work' around the world. Unless you think one culture is more 'natural' than the others, you'd have to blend together all of humanity's gendered labor division and you'd come up with a bit of a mess. [NEWLINE] [NEWLINE] There's nothing 'female' about being a nurse or a teacher or a call-centre employee, for example. Nothing in biology prepares women for those roles. In some places female teachers are extremely rare. In some places farming is traditionally considered women's work and weaving is men's work. And<mask> on and<mask> forth</s>
Label encoding: <s>I think the word'sexual dimorphism' might sort of be what you're looking for, but I don't think it really applies to very much, considering the amount of variation of what's considered 'women's work' around the world. Unless you think one culture is more 'natural' than the others, you'd have to blend together all of humanity's gendered labor division and you'd come up with a bit of a mess. [NEWLINE] [NEWLINE] There's nothing 'female' about being a nurse or a teacher or a call-centre employee, for example. Nothing in biology prepares women for those roles. In some places female teachers are extremely rare. In some places farming is traditionally considered women's work and weaving is men's work. And so on and so forth</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> shouldn't the answers to these value questions be "let the people decide<mask> they are the ones who will bear the costs and outcome<mask> a whole"? [ENDQ] [NEWLINE] Good question.<mask> we decide about many issues in politics that have consequences for different people than the electorate! Take environmental decisions,<mask> our kids (who don't exist<mask> or can't vote) are the ones bearing the consequences. Take animal welfare concerns, on which the animals have no say. Take laws that concern mentally handicapped people, and<mask> on. [NEWLINE] [NEWLINE] I would have thought that experts on value judgment (and, say, weighing information etc) would be better suited to decide than the general public,<mask> that's<mask> is needed in government.<mask> take an amateur<mask> you can have an expert?</s>
Label encoding: <s> [STARTQ] Why shouldn't the answers to these value questions be "let the people decide since they are the ones who will bear the costs and outcome as a whole"? [ENDQ] [NEWLINE] Good question. But we decide about many issues in politics that have consequences for different people than the electorate! Take environmental decisions, where our kids (who don't exist yet or can't vote) are the ones bearing the consequences. Take animal welfare concerns, on which the animals have no say. Take laws that concern mentally handicapped people, and so on. [NEWLINE] [NEWLINE] I would have thought that experts on value judgment (and, say, weighing information etc) would be better suited to decide than the general public, because that's what is needed in government. Why take an amateur when you can have an expert?</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>i'm not op,<mask> none of those shapes are actual perfect circles, triangles, and squares.<mask> the dude said in the article you linked, shapes are geometrical idealizations. the images you linked resemble perfect shapes,<mask> they aren't. none of the sides on the triangles and squares will be perfectly smooth and straight, and obviously the circles won't be perfect either. [NEWLINE] [NEWLINE] <mask> [NEWLINE] [NEWLINE] [STARTQ] <mask> independent creatures will *necessarily* invent<mask>'s fundamentally the same mathematics [ENDQ] [NEWLINE] (italics mine) [NEWLINE] Who's to say they will?<mask> can't they make a base-238482896047396037599573 number system? Imagine that there will be no more intelligent life-forms after humans; would these axioms still be independent?</s>
Label encoding: <s>i'm not op, but none of those shapes are actual perfect circles, triangles, and squares. as the dude said in the article you linked, shapes are geometrical idealizations. the images you linked resemble perfect shapes, but they aren't. none of the sides on the triangles and squares will be perfectly smooth and straight, and obviously the circles won't be perfect either. [NEWLINE] [NEWLINE] also [NEWLINE] [NEWLINE] [STARTQ] If independent creatures will *necessarily* invent what's fundamentally the same mathematics [ENDQ] [NEWLINE] (italics mine) [NEWLINE] Who's to say they will? Why can't they make a base-238482896047396037599573 number system? Imagine that there will be no more intelligent life-forms after humans; would these axioms still be independent?</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**Note:** Your thread has **not** been removed. [NEWLINE] Your post's topic seems to be fairly common on this subreddit.  Similar posts can be found through our [wiki page]( [URL] #link) or via the [search function]( [URL] ;amp;amp;restrict_sr=on). [NEWLINE] [NEWLINE] Regards, the mods of /r/changemyview. [NEWLINE] [NEWLINE] [NEWLINE] *[I am a bot](/r/AutoModerator/comments/q11pu/<mask> _is_automoderator/), and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose?to=%2Fr%2Fchangemyview)<mask> you have any questions or concerns.*</s>
Label encoding: <s>**Note:** Your thread has **not** been removed. [NEWLINE] Your post's topic seems to be fairly common on this subreddit.  Similar posts can be found through our [wiki page]( [URL] #link) or via the [search function]( [URL] ;amp;amp;restrict_sr=on). [NEWLINE] [NEWLINE] Regards, the mods of /r/changemyview. [NEWLINE] [NEWLINE] [NEWLINE] *[I am a bot](/r/AutoModerator/comments/q11pu/ what _is_automoderator/), and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose?to=%2Fr%2Fchangemyview) if you have any questions or concerns.*</s>
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0713, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me just say<mask> a guy thought I was pretty and stared at me for 10 minutes I'd be upset,<mask><mask> he just said I had a nice dress I wouldn't be.  Some guys assume girls will think you're creepy unless you're hot.  Not really.  One time at the eye doctor, I was truing on new frames and discussing them with my mom.  After I picked one and told him<mask> I liked them, his old man (<mask><mask> he was waiting for an appointment) said, "I'd be careful.  You might get too pretty" [NEWLINE] [NEWLINE] Here is the thing.  He wasn't being creepy about it.  Nothing about our chat or his body language implied objectification or anything threatening.  Just kindness.</s>
Label encoding: <s>Let me just say if a guy thought I was pretty and stared at me for 10 minutes I'd be upset, but if he just said I had a nice dress I wouldn't be.  Some guys assume girls will think you're creepy unless you're hot.  Not really.  One time at the eye doctor, I was truing on new frames and discussing them with my mom.  After I picked one and told him how I liked them, his old man ( I think he was waiting for an appointment) said, "I'd be careful.  You might get too pretty" [NEWLINE] [NEWLINE] Here is the thing.  He wasn't being creepy about it.  Nothing about our chat or his body language implied objectification or anything threatening.  Just kindness.</s>
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am not actually bisexual.  I am straight and female.  To tell the truth, I am arguing this from a theoretical perspective,<mask> I am personally disgusted by the idea of sex with the same gender,<mask> see no logical reason to be, unless that disgust was caused by social factors.  I am applying that same principle of attraction to everyone else here. [NEWLINE] [NEWLINE] I should add that, ashamedly, I do not want to be bisexual at all. <mask><mask> of myself<mask> 100% straight.  I do not want this theoretical standard of attraction to apply to me,<mask> I should technically be bisexual, I just can't think up of a reason<mask> it shouldn't.  I really do want to be convinced that I can be 100% straight.</s>
Label encoding: <s>I am not actually bisexual.  I am straight and female.  To tell the truth, I am arguing this from a theoretical perspective, because I am personally disgusted by the idea of sex with the same gender, but see no logical reason to be, unless that disgust was caused by social factors.  I am applying that same principle of attraction to everyone else here. [NEWLINE] [NEWLINE] I should add that, ashamedly, I do not want to be bisexual at all.  I think of myself as 100% straight.  I do not want this theoretical standard of attraction to apply to me, where I should technically be bisexual, I just can't think up of a reason why it shouldn't.  I really do want to be convinced that I can be 100% straight.</s>
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, it's called recomping in weightlifting forums. See [URL] / and leangains. [NEWLINE] [NEWLINE] The basic idea is that<mask> you are already lean, it's not very efficient to go this route. You can't do a 1:1 fat loss:muscle gain. You could cut 1lb/week of fat,<mask> then you'd likely only be gaining small amounts of muscle<mask> working out, or losing muscle<mask> not working out. Any lean, somewhat trained person is better off bulking and cutting. Personally I like the idea of a slow bulk/cut cycle<mask> I'd prefer to minimize the fat gain and I'm OK taking a little longer to gain muscle.<mask> usually you do need to eat a surplus to gain muscle at a reasonable rate.</s>
Label encoding: <s>Yes, it's called recomping in weightlifting forums. See [URL] / and leangains. [NEWLINE] [NEWLINE] The basic idea is that if you are already lean, it's not very efficient to go this route. You can't do a 1:1 fat loss:muscle gain. You could cut 1lb/week of fat, but then you'd likely only be gaining small amounts of muscle if working out, or losing muscle if not working out. Any lean, somewhat trained person is better off bulking and cutting. Personally I like the idea of a slow bulk/cut cycle because I'd prefer to minimize the fat gain and I'm OK taking a little longer to gain muscle. But usually you do need to eat a surplus to gain muscle at a reasonable rate.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The way it's been explained to me is that, even knowing exactly<mask> will happen in the future to any given person, that it doesn't necessarily mean that he intervenes *directly* in human matters. Honestly the way ideas of free will and master plans are interpreted depend heavily on the denomination in question; there are really very few things that *all* denominations agree on universally,<mask><mask> we have<mask> many in the first place. [NEWLINE] [NEWLINE] <mask><mask> that, generally speaking, the explanations you'll get from christians on this matter don't tend to follow logically,<mask> then, there's a reason they call it faith. Even<mask> I don't agree with them, I'd rather live and let live than waste time worrying about the consistency of their beliefs.</s>
Label encoding: <s>The way it's been explained to me is that, even knowing exactly what will happen in the future to any given person, that it doesn't necessarily mean that he intervenes *directly* in human matters. Honestly the way ideas of free will and master plans are interpreted depend heavily on the denomination in question; there are really very few things that *all* denominations agree on universally, hence why we have so many in the first place. [NEWLINE] [NEWLINE] I agree that, generally speaking, the explanations you'll get from christians on this matter don't tend to follow logically, but then, there's a reason they call it faith. Even if I don't agree with them, I'd rather live and let live than waste time worrying about the consistency of their beliefs.</s>
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Different than the person you originally replied to. [NEWLINE] [NEWLINE] [STARTQ] It sounded like you were claiming that six-year-olds would never be exposed to the word "botanist" in a context<mask> the meaning wasn't clear. [ENDQ] [NEWLINE] Perhaps they would or wouldn't. I don't know that it would matter whether a 6 year old would know<mask> a botanist was on first sight or not. By the time it *did* matter that they know<mask> a botanist is, they'd find out. [NEWLINE] [NEWLINE] <mask><mask> the point ButtaBeButtaFree was trying to make is that<mask> your examples may *seem* "easier" they are largely irrelevant<mask> English speakers do not actually have a hard time figuring out<mask> words mean.</s><pad>
Label encoding: <s>Different than the person you originally replied to. [NEWLINE] [NEWLINE] [STARTQ] It sounded like you were claiming that six-year-olds would never be exposed to the word "botanist" in a context where the meaning wasn't clear. [ENDQ] [NEWLINE] Perhaps they would or wouldn't. I don't know that it would matter whether a 6 year old would know what a botanist was on first sight or not. By the time it *did* matter that they know what a botanist is, they'd find out. [NEWLINE] [NEWLINE] I think the point ButtaBeButtaFree was trying to make is that while your examples may *seem* "easier" they are largely irrelevant as English speakers do not actually have a hard time figuring out what words mean.</s><pad>
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0505, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0704, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you were born with only one leg, isn't it reasonable that you'd feel like part of you was missing? [ENDQ] [NEWLINE] No. I wasnt born with it and<mask> could not have miss it. I would maybe be curious<mask> having two limbs would be like<mask> I<mask> couldnt consider myself a two legged person (which is kind of your point, right?). Most people born missing limbs dont notice they dont have them<mask> they never have had them. IE the people that brush their teeth with their feet<mask> they have never had hands. [NEWLINE] [NEWLINE] We are all born with issues and desires. Some of those we can work hard and achive<mask> just randomly deciding our reality isnt the same<mask> everyone around us isnt realistic. </s>
Label encoding: <s> [STARTQ] If you were born with only one leg, isn't it reasonable that you'd feel like part of you was missing? [ENDQ] [NEWLINE] No. I wasnt born with it and therefore could not have miss it. I would maybe be curious what having two limbs would be like but I also couldnt consider myself a two legged person (which is kind of your point, right?). Most people born missing limbs dont notice they dont have them because they never have had them. IE the people that brush their teeth with their feet because they have never had hands. [NEWLINE] [NEWLINE] We are all born with issues and desires. Some of those we can work hard and achive but just randomly deciding our reality isnt the same as everyone around us isnt realistic. </s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0476, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My dining hall had sandwiches (hot an cold), a large salad bar with lots of fruit and vegetables, burgers, hot dogs, american standard hot line (sometimes chicken fried steak, sometimes fried chicken, sometimes meatloaf, etc, often had brown rice<mask> a side<mask> not always), oriental hot line (sometimes chinese, sometimes thai always had white rice and fried rice), an Italian line (pizzas baked in a stone oven, pastas often involving chicken, etc) and a specialty hot line (risottos, steak, fancy roasted chicken dishes, etc<mask> they were in smaller amounts and required tickets that cost an extra dollar), a smoothie line, and a dessert line, and cereal on the walls all day long.</s>
Label encoding: <s>My dining hall had sandwiches (hot an cold), a large salad bar with lots of fruit and vegetables, burgers, hot dogs, american standard hot line (sometimes chicken fried steak, sometimes fried chicken, sometimes meatloaf, etc, often had brown rice as a side but not always), oriental hot line (sometimes chinese, sometimes thai always had white rice and fried rice), an Italian line (pizzas baked in a stone oven, pastas often involving chicken, etc) and a specialty hot line (risottos, steak, fancy roasted chicken dishes, etc but they were in smaller amounts and required tickets that cost an extra dollar), a smoothie line, and a dessert line, and cereal on the walls all day long.</s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>, wouln't mature, responsible parents who are in a bad realtionship agree that they should work something out "for the kids"? Wouldn't the fact that they are raising human lives together be some type of incentive for them to iron out these "rough spots" in the relationship? [ENDQ] [NEWLINE] Yes they would, and they do. [NEWLINE] [NEWLINE] You'll get a ton of responses to this CMV that are essentially "divorce is better than watching your parents fight all the time, and the kids will *know*".  It's bullshit.  Parents who *want* to hide marital stress from their children hide marital stress from their children.  The only challenge is that *both* parents must me equally willing to do that.</s>
Label encoding: <s> [STARTQ] Also, wouln't mature, responsible parents who are in a bad realtionship agree that they should work something out "for the kids"? Wouldn't the fact that they are raising human lives together be some type of incentive for them to iron out these "rough spots" in the relationship? [ENDQ] [NEWLINE] Yes they would, and they do. [NEWLINE] [NEWLINE] You'll get a ton of responses to this CMV that are essentially "divorce is better than watching your parents fight all the time, and the kids will *know*".  It's bullshit.  Parents who *want* to hide marital stress from their children hide marital stress from their children.  The only challenge is that *both* parents must me equally willing to do that.</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0605, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The proportion of patients reporting cannabis withdrawal in recent treatment studies has ranged from 50-95% (Budney &amp; Hughes, 2006). Symptoms commonly experienced include sleep difficulty; decreased appetite and weight loss; irritability; nervousness and anxiety; restlessness; and increased anger and aggression. The majority of symptoms peak between day two and six of abstinence and most return to baseline by day 14. Sleep difficulty, anger/aggression, irritability and physical tension have persisted for three to four weeks in some studies (Budney, Moore, Vandrey, &amp; Hughes, 2003; Kouri &amp; Pope, 2000). Strange dreams failed to return to baseline during a 45-day abstinence study (Budney et al., 2003).</s>
Label encoding: <s>The proportion of patients reporting cannabis withdrawal in recent treatment studies has ranged from 50-95% (Budney &amp; Hughes, 2006). Symptoms commonly experienced include sleep difficulty; decreased appetite and weight loss; irritability; nervousness and anxiety; restlessness; and increased anger and aggression. The majority of symptoms peak between day two and six of abstinence and most return to baseline by day 14. Sleep difficulty, anger/aggression, irritability and physical tension have persisted for three to four weeks in some studies (Budney, Moore, Vandrey, &amp; Hughes, 2003; Kouri &amp; Pope, 2000). Strange dreams failed to return to baseline during a 45-day abstinence study (Budney et al., 2003).</s>
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are you advocating the government be in charge of distribution? <mask><mask> the purity of certain drugs could be an issue.  Heroine needs to be cut down a lot to be safe.  Pure heroine will kill you. <mask> the government is selling me heroine, and decides my political activism is annoying, that I am an enemy, would they not have a very convenient way to kill me?  They know exactly<mask> pure my supply is, just cut it half<mask> much and watch me OD. [NEWLINE] [NEWLINE] <mask>, decriminalization tends to work better that legalization for harder drugs.  It allows dealers and manufacturers to still be prosecuted.  Generally possession would not carry a criminal charge,<mask> might come with a recommendation or court order for rehab.  </s>
Label encoding: <s>Are you advocating the government be in charge of distribution?  If so the purity of certain drugs could be an issue.  Heroine needs to be cut down a lot to be safe.  Pure heroine will kill you.  If the government is selling me heroine, and decides my political activism is annoying, that I am an enemy, would they not have a very convenient way to kill me?  They know exactly how pure my supply is, just cut it half as much and watch me OD. [NEWLINE] [NEWLINE] Also, decriminalization tends to work better that legalization for harder drugs.  It allows dealers and manufacturers to still be prosecuted.  Generally possession would not carry a criminal charge, but might come with a recommendation or court order for rehab.  </s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Keep in mind that Christianity doesn't really contain any tenets on<mask> govts. must act, only<mask> people should act. [ENDQ] [NEWLINE] <mask><mask> democracies didn't really exist during biblical times, I don't think that distinction is relevant. [NEWLINE] [NEWLINE] <mask> I lobby/vote for my government to aid the poor and sick I'm acting in the spirit of Christianity, just<mask> surely<mask><mask> I gave directly. [NEWLINE] [NEWLINE] <mask> abortion, same sex marriage, etc are sins, surely it is the individual that sins, not the government that allows those sins.<mask> is it choosing not to sin<mask> that choice is unavailable? [NEWLINE] [NEWLINE] A Christian shouldn't require sin to be against the law, not should they oppose the government aiding the poor and sick.</s>
Label encoding: <s> [STARTQ] Keep in mind that Christianity doesn't really contain any tenets on how govts. must act, only how people should act. [ENDQ] [NEWLINE] Given that democracies didn't really exist during biblical times, I don't think that distinction is relevant. [NEWLINE] [NEWLINE] If I lobby/vote for my government to aid the poor and sick I'm acting in the spirit of Christianity, just as surely as if I gave directly. [NEWLINE] [NEWLINE] If abortion, same sex marriage, etc are sins, surely it is the individual that sins, not the government that allows those sins. How is it choosing not to sin if that choice is unavailable? [NEWLINE] [NEWLINE] A Christian shouldn't require sin to be against the law, not should they oppose the government aiding the poor and sick.</s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe you committed a logical fallacy.  It helps in a logical debate, in which you present numbers/statistics to support your argument, that you don't carelessly exaggerate the statistic. [NEWLINE] [NEWLINE] You state you don't wish to hire women for the risk of paid pregnancy leave. Lets just pretend she makes an even $10 an hr w/ a 40 hr work week.   Now only 14 of those weeks are paid, the result comes down to about $5600.  This number is much lower than your 2 years of originally claimed pay leave, which equals, $38,400. [NEWLINE] [NEWLINE] You shouldn't embellish facts to help support your argument in a debate, given information greatly influences the argument at hand. </s>
Label encoding: <s>I believe you committed a logical fallacy.  It helps in a logical debate, in which you present numbers/statistics to support your argument, that you don't carelessly exaggerate the statistic. [NEWLINE] [NEWLINE] You state you don't wish to hire women for the risk of paid pregnancy leave. Lets just pretend she makes an even $10 an hr w/ a 40 hr work week.   Now only 14 of those weeks are paid, the result comes down to about $5600.  This number is much lower than your 2 years of originally claimed pay leave, which equals, $38,400. [NEWLINE] [NEWLINE] You shouldn't embellish facts to help support your argument in a debate, given information greatly influences the argument at hand. </s>
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If OP's reasoning is based on harm to a fetus, then other things harming a fetus are absolutely relevant, and enter the debate with OP's reasoning. This is<mask> OP's argument would take the form: [NEWLINE] [NEWLINE] [Some threshold of] Harm to a fetus should be illegal. [NEWLINE] Alcohol and tobacco consumption<mask> pregnant harm a fetus. [NEWLINE] <mask> alcohol and tobacco consumption<mask> pregnant should be illegal. [NEWLINE] [NEWLINE] Discussion about the proper legality other things harming a fetus constitute potential counterexamples to the first premise, which<mask> valid break OP's whole argument.<mask> could that be off-limits, and more importantly,<mask> on earth did you get your understanding of logical fallacies? [NEWLINE] [NEWLINE] I just don't understand<mask> this is confusing.</s>
Label encoding: <s>If OP's reasoning is based on harm to a fetus, then other things harming a fetus are absolutely relevant, and enter the debate with OP's reasoning. This is because OP's argument would take the form: [NEWLINE] [NEWLINE] [Some threshold of] Harm to a fetus should be illegal. [NEWLINE] Alcohol and tobacco consumption while pregnant harm a fetus. [NEWLINE] Therefore alcohol and tobacco consumption while pregnant should be illegal. [NEWLINE] [NEWLINE] Discussion about the proper legality other things harming a fetus constitute potential counterexamples to the first premise, which if valid break OP's whole argument. How could that be off-limits, and more importantly, where on earth did you get your understanding of logical fallacies? [NEWLINE] [NEWLINE] I just don't understand how this is confusing.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sorry you're offended,<mask><mask> was I supposed to know you were being serious<mask> you ask a derpy question? [NEWLINE] [NEWLINE] Dude, I don't know who they do it to. You should ask them. I don't understand<mask> you imply that men or old people can't be attractive either. I don't find her attractive. My point is that they are making a power play to make themselves feel better and disregarding others' space, which is a terrible, shitty thing to do. That's the bottom of it. Implying that her being attractive somehow makes their behavior excusable is derailing and shitty.<mask> they are doing does not need to be analyzed.<mask> they are doing needs to be not done.</s>
Label encoding: <s>I'm sorry you're offended, but how was I supposed to know you were being serious if you ask a derpy question? [NEWLINE] [NEWLINE] Dude, I don't know who they do it to. You should ask them. I don't understand why you imply that men or old people can't be attractive either. I don't find her attractive. My point is that they are making a power play to make themselves feel better and disregarding others' space, which is a terrible, shitty thing to do. That's the bottom of it. Implying that her being attractive somehow makes their behavior excusable is derailing and shitty. What they are doing does not need to be analyzed. What they are doing needs to be not done.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe it comes down to<mask> you value more, the hiker's lives or the bears?  I'm no woodsman,<mask> my understanding of predators is<mask> they attack one human, they will attack another. Most bears do not attack people. Especially those that have been raised in National Parks and that have had lots of human encounters. These attacks are exceedingly rare ultimately. The bears/wild animals that do attack humans are dangerous aberrations. By culling the populations of these aggressive genes, we ultimately ensure the long-term survival of these populations by acclimating them to live in a human-run world. [NEWLINE] [NEWLINE] <mask>, I would say the National Parks have long been altered for our enjoyment. Wildlife preserves much less<mask>.</s>
Label encoding: <s>I believe it comes down to what you value more, the hiker's lives or the bears?  I'm no woodsman, but my understanding of predators is if they attack one human, they will attack another. Most bears do not attack people. Especially those that have been raised in National Parks and that have had lots of human encounters. These attacks are exceedingly rare ultimately. The bears/wild animals that do attack humans are dangerous aberrations. By culling the populations of these aggressive genes, we ultimately ensure the long-term survival of these populations by acclimating them to live in a human-run world. [NEWLINE] [NEWLINE] Also, I would say the National Parks have long been altered for our enjoyment. Wildlife preserves much less so.</s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm gay and I've been attracted to my share of guys<mask> growing up<mask> I never pined over them. I don't think this particular concern is such a big deal. Unrequited love is way more common in the world of fiction than real life<mask><mask> sexual orientation. Most young people will at some point develop a mild crush on a celebrity, teacher, babysitter, or someone they think is out of their league and they will feel unable to act on it.<mask> they'll quickly get over it and probably won't talk to their parents about it anyway. Who gives their teens dating advice other than "respect yourself and use protection" or "don't have sex" or "high school relationships aren't serious" anyway?</s>
Label encoding: <s>I'm gay and I've been attracted to my share of guys while growing up but I never pined over them. I don't think this particular concern is such a big deal. Unrequited love is way more common in the world of fiction than real life regardless of sexual orientation. Most young people will at some point develop a mild crush on a celebrity, teacher, babysitter, or someone they think is out of their league and they will feel unable to act on it. But they'll quickly get over it and probably won't talk to their parents about it anyway. Who gives their teens dating advice other than "respect yourself and use protection" or "don't have sex" or "high school relationships aren't serious" anyway?</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They never used to give vaccines to babies or pregnant women.<mask><mask><mask><mask> that they cause autism,<mask> at the same time, I see no valid reason to give a crap load of chemicals to a child who's immune system has not developed.<mask><mask> at birth vaccinations are more harmful than the pharmaceutical industry lets on (i.e. mental illness rates have soared among children). Mercury is toxic. You shouldn't even eat tuna<mask> you're pregnant. I'm not against vaccination, I'm a nursing student, it's a wonderful thing. I just don't think that we should administer anything that may derail a childs development at that early of an age. Vaccines should be administered after one year of age, at least. </s>
Label encoding: <s>They never used to give vaccines to babies or pregnant women. I do not think that they cause autism, but at the same time, I see no valid reason to give a crap load of chemicals to a child who's immune system has not developed. I think at birth vaccinations are more harmful than the pharmaceutical industry lets on (i.e. mental illness rates have soared among children). Mercury is toxic. You shouldn't even eat tuna while you're pregnant. I'm not against vaccination, I'm a nursing student, it's a wonderful thing. I just don't think that we should administer anything that may derail a childs development at that early of an age. Vaccines should be administered after one year of age, at least. </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're calling Subway a good diet? A sandwich. With questionable ingredients. And a bag of chips. That's your good diet? Hate to break it to you,<mask> there are NO "good" fast food places, and I challenge you to come up with one.<mask><mask>, I feel I might make a cmv about it, just to see. The fact of the matter is for the most part we know nothing about<mask> to take care of ourselves, and most of us have no fucking clue<mask> our body needs from food. Combine this with the over-processed, over-mutated sludge we're under the impression is healthy<mask> it's marketed that way, and you see<mask> we've arrived<mask> we have.</s>
Label encoding: <s>You're calling Subway a good diet? A sandwich. With questionable ingredients. And a bag of chips. That's your good diet? Hate to break it to you, but there are NO "good" fast food places, and I challenge you to come up with one. In fact, I feel I might make a cmv about it, just to see. The fact of the matter is for the most part we know nothing about how to take care of ourselves, and most of us have no fucking clue what our body needs from food. Combine this with the over-processed, over-mutated sludge we're under the impression is healthy because it's marketed that way, and you see why we've arrived where we have.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0586, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Informed consent.  A child is not emotionally mature enough to consent to sex and on top of that the power dynamic will always work out in the benefit of the pediphile.  Sexually attraction can be both genetic and environmentally altered<mask> in that sense all attraction is equally viable,<mask> from a moral perspective a social contract is necessary for sexual activity to be mutual and not be considered rape. [NEWLINE] <mask> we define everything<mask> relative and all desires<mask> equally viable to be acted upon, we must then morally accept the serial killer who can not qualm his desire for flesh. [NEWLINE] [NEWLINE] Objectively speaking, relativism holds true,<mask> subjectivity (i.e. moral reasoning and social order)  are necessary for civil society.</s>
Label encoding: <s>Informed consent.  A child is not emotionally mature enough to consent to sex and on top of that the power dynamic will always work out in the benefit of the pediphile.  Sexually attraction can be both genetic and environmentally altered so in that sense all attraction is equally viable, but from a moral perspective a social contract is necessary for sexual activity to be mutual and not be considered rape. [NEWLINE] If we define everything as relative and all desires as equally viable to be acted upon, we must then morally accept the serial killer who can not qualm his desire for flesh. [NEWLINE] [NEWLINE] Objectively speaking, relativism holds true, but subjectivity (i.e. moral reasoning and social order)  are necessary for civil society.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0665, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think religion is special in that regard.  My beliefs aren't written in a book that I'm supposed to follow.  I totally understand that religions evolve over time (irony in it's finest form),<mask><mask> should people be able to look at it and say, "Well, X doesn't really apply anymore<mask> society has evolved to the point<mask> it's irrelevant," and then look at the subject of gay marriage and be<mask> adamant that God is against it and<mask> it's loosely mentioned in the Bible, I'm going to fight<mask> hard<mask> I can against it? [NEWLINE] [NEWLINE] To me, not only is it hypocrisy,<mask> it's flat out using your religion to mask your prejudice about something that makes you uncomfortable.</s>
Label encoding: <s>I think religion is special in that regard.  My beliefs aren't written in a book that I'm supposed to follow.  I totally understand that religions evolve over time (irony in it's finest form), but why should people be able to look at it and say, "Well, X doesn't really apply anymore because society has evolved to the point where it's irrelevant," and then look at the subject of gay marriage and be so adamant that God is against it and since it's loosely mentioned in the Bible, I'm going to fight as hard as I can against it? [NEWLINE] [NEWLINE] To me, not only is it hypocrisy, but it's flat out using your religion to mask your prejudice about something that makes you uncomfortable.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The timbre of the sax makes it a great background instrument to provide warmth that brass lack. [ENDQ] [NEWLINE] This hits on something that I wasn't able to express in my post: I feel like every tone that comes out of a saxophone is kind of an ugly compromise between woodwind and brass.  That's a feeling I regularly get<mask> I'm able to make it through a decent length of sax music: All these expressions would hit me<mask> much harder<mask> they came from the woodwind or brass specialist. [NEWLINE] [NEWLINE] <mask> it doesn't really change any views I had, you are right about the supporting role question and I appreciate the examples!  There were surprisingly few sax cringes during the big band medley.  </s>
Label encoding: <s> [STARTQ] The timbre of the sax makes it a great background instrument to provide warmth that brass lack. [ENDQ] [NEWLINE] This hits on something that I wasn't able to express in my post: I feel like every tone that comes out of a saxophone is kind of an ugly compromise between woodwind and brass.  That's a feeling I regularly get when I'm able to make it through a decent length of sax music: All these expressions would hit me so much harder if they came from the woodwind or brass specialist. [NEWLINE] [NEWLINE] Although it doesn't really change any views I had, you are right about the supporting role question and I appreciate the examples!  There were surprisingly few sax cringes during the big band medley.  </s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First off, great article and very solid info; I'd quibble that this campaign isn't about explaining rape culture or helping people understand it,<mask> attempting to address the root causes of it. [NEWLINE] [NEWLINE] That said, it's the right kind of data point to show that this type of education could be constructive. &amp;#8710; [NEWLINE] [NEWLINE] <mask>, by itself it isn't very compelling: British Columbia<mask> a whole saw a 5.4% drop in sexual assault and Vancouver, with only 10% of the population and only a 10% drop, can't explain that shift -- it may be that there was a larger, unrelated trend at work (or it may be that similar campaigns were being mounted elsewhere in BC).</s>
Label encoding: <s>First off, great article and very solid info; I'd quibble that this campaign isn't about explaining rape culture or helping people understand it, but attempting to address the root causes of it. [NEWLINE] [NEWLINE] That said, it's the right kind of data point to show that this type of education could be constructive. &amp;#8710; [NEWLINE] [NEWLINE] However, by itself it isn't very compelling: British Columbia as a whole saw a 5.4% drop in sexual assault and Vancouver, with only 10% of the population and only a 10% drop, can't explain that shift -- it may be that there was a larger, unrelated trend at work (or it may be that similar campaigns were being mounted elsewhere in BC).</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your problem is not with *democracy*, it's with *people*. Democracy is designed to represent *all* of the people and give an equal opportunity for *all* of their views to be heard. [NEWLINE] [NEWLINE] Your issue is that there are a lot of people that have views that you find ugly. Maybe you want to think that your country is above those views,<mask> clearly it's not. [NEWLINE] [NEWLINE] Democracy works<mask> nobody gets the power to say one person's opinion disqualifies them from being heard. Think of it this way:<mask> you want to disqualify someone from voting<mask> of their opinion, then *someone else* could disqualify *you* from voting<mask> they didn't like *your* opinion. [NEWLINE] </s><pad>
Label encoding: <s>Your problem is not with *democracy*, it's with *people*. Democracy is designed to represent *all* of the people and give an equal opportunity for *all* of their views to be heard. [NEWLINE] [NEWLINE] Your issue is that there are a lot of people that have views that you find ugly. Maybe you want to think that your country is above those views, but clearly it's not. [NEWLINE] [NEWLINE] Democracy works because nobody gets the power to say one person's opinion disqualifies them from being heard. Think of it this way: if you want to disqualify someone from voting because of their opinion, then *someone else* could disqualify *you* from voting because they didn't like *your* opinion. [NEWLINE] </s><pad>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You would almost have a point<mask> the bills were entirely written by members of Congress<mask> they aren't.  These massive bills are often written largely by lobbyists -- and the congressmen in question have *openly admitted* that they don't understand<mask> is in them. [NEWLINE] [NEWLINE] Multi-thousand page bills - with lots of details - cannot be digested in a couple of days.  And that's all they have before having to vote. [NEWLINE] [NEWLINE] Ignoring that,<mask>, very good bills have been written that are not long. See [Title IX]( [URL].php) -- the landmark sex discrimination bill.  That's the link to the whole thing. [NEWLINE] [NEWLINE] Or do you think that's a bad bill?</s>
Label encoding: <s>You would almost have a point if the bills were entirely written by members of Congress but they aren't.  These massive bills are often written largely by lobbyists -- and the congressmen in question have *openly admitted* that they don't understand what is in them. [NEWLINE] [NEWLINE] Multi-thousand page bills - with lots of details - cannot be digested in a couple of days.  And that's all they have before having to vote. [NEWLINE] [NEWLINE] Ignoring that, though, very good bills have been written that are not long. See [Title IX]( [URL].php) -- the landmark sex discrimination bill.  That's the link to the whole thing. [NEWLINE] [NEWLINE] Or do you think that's a bad bill?</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry ConstantFlux46360, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=ConstantFlux46360+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry ConstantFlux46360, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=ConstantFlux46360+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Someone said that I "must be a woman" and that was deleted for being rude and hostile. I took that<mask> a compliment<mask> apparently the mods think it is an insult to say that someone is a woman. [NEWLINE] [NEWLINE] <mask> the fuck is the above post not considered rude and hostile?<mask> the fuck could anybody in their right mind consider a description of a rectovaginal fistula<mask> pornographic? That is<mask> can happen<mask> someone is penetrated against their will. It's a matter of fact, and people don't want to talk about<mask> it's terrifying. Fistulas don't happen to someone who is "forced to penetrate." It is a huge difference and the reason<mask> rape should be limited to being penetrated. </s>
Label encoding: <s>Someone said that I "must be a woman" and that was deleted for being rude and hostile. I took that as a compliment but apparently the mods think it is an insult to say that someone is a woman. [NEWLINE] [NEWLINE] How the fuck is the above post not considered rude and hostile? How the fuck could anybody in their right mind consider a description of a rectovaginal fistula as pornographic? That is what can happen when someone is penetrated against their will. It's a matter of fact, and people don't want to talk about because it's terrifying. Fistulas don't happen to someone who is "forced to penetrate." It is a huge difference and the reason why rape should be limited to being penetrated. </s>
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If many people (and I truly mean many in a "large percentage of the population" way, not in a "large numbers out of context" way) outside of the government wanted to murder each other and only the existence of armed police served to stop them, would you think that the government should let them alone? Probably not. The concept of government is to safeguard the "rights" of everyone<mask> best<mask> possible, not to cater to certain people. [NEWLINE] [NEWLINE] Like it or not, most people in government are of the quite justifiable position that hard drugs cause far more harm than good and shouldn't be easily available to the public. [NEWLINE] [NEWLINE] EDIT:<mask>, rule 5- no "low effort" comments.</s>
Label encoding: <s>If many people (and I truly mean many in a "large percentage of the population" way, not in a "large numbers out of context" way) outside of the government wanted to murder each other and only the existence of armed police served to stop them, would you think that the government should let them alone? Probably not. The concept of government is to safeguard the "rights" of everyone as best as possible, not to cater to certain people. [NEWLINE] [NEWLINE] Like it or not, most people in government are of the quite justifiable position that hard drugs cause far more harm than good and shouldn't be easily available to the public. [NEWLINE] [NEWLINE] EDIT: Also, rule 5- no "low effort" comments.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You believe that these nice guys are coming from a place of misogynistic entitlement [ENDQ] [NEWLINE] Calling someone a bitch for not wanting to date you for whatever reason is misogynistic entitlement [NEWLINE] [NEWLINE] Calling someone a *WHORE* for not returning your affections and wanting to just be friend IS misogynistic entitlement. [NEWLINE] [NEWLINE] Believing that being "nice" to someone you're attracted to entitles you to first dibs IS misoynistic entitlement. [NEWLINE] [NEWLINE] And no they dehumanize themselves<mask> they show that common decency and friendship are just tools to them. And nope, no sense of victimization here. Re-read my first post. I don't suffer fools and assholes in my social circle. </s>
Label encoding: <s> [STARTQ] You believe that these nice guys are coming from a place of misogynistic entitlement [ENDQ] [NEWLINE] Calling someone a bitch for not wanting to date you for whatever reason is misogynistic entitlement [NEWLINE] [NEWLINE] Calling someone a *WHORE* for not returning your affections and wanting to just be friend IS misogynistic entitlement. [NEWLINE] [NEWLINE] Believing that being "nice" to someone you're attracted to entitles you to first dibs IS misoynistic entitlement. [NEWLINE] [NEWLINE] And no they dehumanize themselves when they show that common decency and friendship are just tools to them. And nope, no sense of victimization here. Re-read my first post. I don't suffer fools and assholes in my social circle. </s>
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0687, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> my vocabulary is non-infinite,<mask> it is a non-inconsequential loss of language, that I may or may not have a ready, proper replacement for. [ENDQ] [NEWLINE] There are no shortage of perfectly strong cuss words to get your point across without denigrating a group of people. <mask> you're not able to articulate your outrage or anger without resorting to using a word like "retard" or "fag", you're just being intellectually lazy and not trying hard enough. [NEWLINE] [NEWLINE] There are *plenty* of strong words out there.  You have a huge canvas and you just need to use it without resorting to the same colors of paint every single time.</s>
Label encoding: <s> [STARTQ] because my vocabulary is non-infinite, so it is a non-inconsequential loss of language, that I may or may not have a ready, proper replacement for. [ENDQ] [NEWLINE] There are no shortage of perfectly strong cuss words to get your point across without denigrating a group of people.  If you're not able to articulate your outrage or anger without resorting to using a word like "retard" or "fag", you're just being intellectually lazy and not trying hard enough. [NEWLINE] [NEWLINE] There are *plenty* of strong words out there.  You have a huge canvas and you just need to use it without resorting to the same colors of paint every single time.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry spank859, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=spank859+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry spank859, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=spank859+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Transitioning works to fix gender dysphoria. Psychology does not. [ENDQ] [NEWLINE] Surgery may - I dont dispute that. <mask> accepting someone for whatever they want to be accept - of course thats going to "solve" the problem.  Accepting someone else's delusion is always going to "solve" the problem. [NEWLINE] [NEWLINE] I dont agree that psychology treatment doesnt help. <mask><mask> we have just gone too quick to agree to cut on people before actually doing the hard work of counselling them. Its hard<mask> the person doesnt want to be counselled.  Its like trying to stop an alcoholic from drinking.  They want to drink,<mask> until they want to change - they wont.</s><pad>
Label encoding: <s> [STARTQ] Transitioning works to fix gender dysphoria. Psychology does not. [ENDQ] [NEWLINE] Surgery may - I dont dispute that.  But accepting someone for whatever they want to be accept - of course thats going to "solve" the problem.  Accepting someone else's delusion is always going to "solve" the problem. [NEWLINE] [NEWLINE] I dont agree that psychology treatment doesnt help.  I think we have just gone too quick to agree to cut on people before actually doing the hard work of counselling them. Its hard because the person doesnt want to be counselled.  Its like trying to stop an alcoholic from drinking.  They want to drink, so until they want to change - they wont.</s><pad>
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Either you're counting wrong and eating much more than you think you are... [NEWLINE] [NEWLINE] Or you're a magical fairy who can create energy out of nothing.  In which case, congratulations on being a perpetual motion machine.  You are truly a goddess among mortals. [NEWLINE] [NEWLINE] <mask>  I would go for counting wrong. [NEWLINE] [NEWLINE] edit: seriously<mask>,<mask> you're absolutely sure you're eating 650 calories a day and still gaining weight, go to the nearest research university and submit yourself for testing.  You are literally a miracle and your body defies known laws of physics.  Humanity needs your genes to be documented  and studied<mask> we can create a new generation of super-humans that are immune to starvation.</s>
Label encoding: <s>Either you're counting wrong and eating much more than you think you are... [NEWLINE] [NEWLINE] Or you're a magical fairy who can create energy out of nothing.  In which case, congratulations on being a perpetual motion machine.  You are truly a goddess among mortals. [NEWLINE] [NEWLINE] But  I would go for counting wrong. [NEWLINE] [NEWLINE] edit: seriously though, if you're absolutely sure you're eating 650 calories a day and still gaining weight, go to the nearest research university and submit yourself for testing.  You are literally a miracle and your body defies known laws of physics.  Humanity needs your genes to be documented  and studied so we can create a new generation of super-humans that are immune to starvation.</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0551, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0756, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Death is the enemy of humanity. Death is the embodiment of evil. Choosing death is siding with death, which means taking a stand against humanity. [ENDQ] [NEWLINE] Death is an integral part of humanity. It's not evil or good, it's just is. Same<mask> life. They are in a balance with each other. You don't choose death,<mask> there is no choice. It's like choosing to exist or not. You don't have the ability to do that. [NEWLINE] [NEWLINE] [STARTQ] In short, choosing death means choosing the easy way out. Dying isn't easy,<mask> it's easier than living with your problems. [ENDQ] [NEWLINE] Being dead is easy. Choosing to die is never easy.</s>
Label encoding: <s> [STARTQ] Death is the enemy of humanity. Death is the embodiment of evil. Choosing death is siding with death, which means taking a stand against humanity. [ENDQ] [NEWLINE] Death is an integral part of humanity. It's not evil or good, it's just is. Same as life. They are in a balance with each other. You don't choose death, because there is no choice. It's like choosing to exist or not. You don't have the ability to do that. [NEWLINE] [NEWLINE] [STARTQ] In short, choosing death means choosing the easy way out. Dying isn't easy, but it's easier than living with your problems. [ENDQ] [NEWLINE] Being dead is easy. Choosing to die is never easy.</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't have a face online, and our mind is naturally wired to percieve things<mask> threatening. (Better to think you see tiger and be wrong, than to think you DON'T see a tiger and be wrong.) [NEWLINE] [NEWLINE] Without the face to empathize/engage with, it's very easy to interpret me<mask> being hostile, or just ignore me. (No snowflake in an avalanch feels responsible.) [NEWLINE] [NEWLINE] <mask><mask> face-to-face discussion is just<mask> different than a huge, anonymous, text-only forum that it's pretty hard to interpret<mask> happens here<mask> something people ACTUALLY wished they could do in real life.  Our medium affects our expression hugely.</s>
Label encoding: <s>I don't have a face online, and our mind is naturally wired to percieve things as threatening. (Better to think you see tiger and be wrong, than to think you DON'T see a tiger and be wrong.) [NEWLINE] [NEWLINE] Without the face to empathize/engage with, it's very easy to interpret me as being hostile, or just ignore me. (No snowflake in an avalanch feels responsible.) [NEWLINE] [NEWLINE] I think face-to-face discussion is just so different than a huge, anonymous, text-only forum that it's pretty hard to interpret what happens here as something people ACTUALLY wished they could do in real life.  Our medium affects our expression hugely.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0531, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I suppose one way I might try to C-Your-V would be to consider political cartoons.  They use similar techniques to comic strips,<mask> are communicating a much different message. [NEWLINE] [NEWLINE] Alternately, consider the graphic novel.  "Watchmen" uses the same medium<mask> "Richie Rich,"<mask> communicates a much different message. [NEWLINE] [NEWLINE] <mask> it "weirds you out," it makes changing your view rather difficult,<mask> we're talking about a basic unconscious "yick" factor, perhaps. <mask> consider the thought<mask> the alternate were true: [NEWLINE] [NEWLINE] "Literature for children should not be expressed in novel form,<mask> that's an adult medium." [NEWLINE] [NEWLINE] Just a thought.</s>
Label encoding: <s>I suppose one way I might try to C-Your-V would be to consider political cartoons.  They use similar techniques to comic strips, but are communicating a much different message. [NEWLINE] [NEWLINE] Alternately, consider the graphic novel.  "Watchmen" uses the same medium as "Richie Rich," but communicates a much different message. [NEWLINE] [NEWLINE] Because it "weirds you out," it makes changing your view rather difficult, because we're talking about a basic unconscious "yick" factor, perhaps.  But consider the thought if the alternate were true: [NEWLINE] [NEWLINE] "Literature for children should not be expressed in novel form, because that's an adult medium." [NEWLINE] [NEWLINE] Just a thought.</s>
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry RichardRogers, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=RichardRogers+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry RichardRogers, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 5\. "No low effort comments. Comments that are only jokes or 'written upvotes', for example. Humor and affirmations of agreement can be contained within more substantial comments." [See the wiki page for more information.]( [URL] #wiki_rule_5) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+5+Post+Appeal&amp;message=RichardRogers+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why should people inherently believe the victim? The "rapist" could be the victim in this case<mask> it wasn't actually rape. We really don't know who the victim is unless we saw it happen or had evidence it was rape. This type of stuff should go through the police. Let the professionals take care of it, and let the students stay out of it. They don't have much business in taking sides in a situation they really know nothing about other than "he said, she said." [NEWLINE] [NEWLINE] [STARTQ] they just inherently don't believe the victim. [ENDQ] [NEWLINE] <mask> they don't know<mask> he/she is a victim. All they have is his/her word at this point. </s>
Label encoding: <s>Why should people inherently believe the victim? The "rapist" could be the victim in this case if it wasn't actually rape. We really don't know who the victim is unless we saw it happen or had evidence it was rape. This type of stuff should go through the police. Let the professionals take care of it, and let the students stay out of it. They don't have much business in taking sides in a situation they really know nothing about other than "he said, she said." [NEWLINE] [NEWLINE] [STARTQ] they just inherently don't believe the victim. [ENDQ] [NEWLINE] Because they don't know if he/she is a victim. All they have is his/her word at this point. </s>
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why anyone would be "proud" of a government for doing<mask> it is designed and *obligated* to do is completely beyond me. Furthermore<mask> the government fails its people and the global community,<mask> you continue to have pride in said system only confuses me more. [NEWLINE] [NEWLINE] The list of governments it has undermined or removed, is terribly long(Bay of Pigs, Mogadishu). The people both abroad and locally it has used and abandoned(think about the veterans of the wars we fight, or the Al-Queda) is similarly long.<mask> anyone has pride in a government which treats people<mask> expendable is not only confusing to me, it's maddening.</s>
Label encoding: <s>Why anyone would be "proud" of a government for doing what it is designed and *obligated* to do is completely beyond me. Furthermore when the government fails its people and the global community, why you continue to have pride in said system only confuses me more. [NEWLINE] [NEWLINE] The list of governments it has undermined or removed, is terribly long(Bay of Pigs, Mogadishu). The people both abroad and locally it has used and abandoned(think about the veterans of the wars we fight, or the Al-Queda) is similarly long. Why anyone has pride in a government which treats people as expendable is not only confusing to me, it's maddening.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Also, we have a fundamental different viewpoint of workers. I would like a more holistic  approach which doesn't treat workers solely<mask> commodities. You seem fine with corporations doing<mask>. [NEWLINE] [NEWLINE] I'm not interested in arguing this any further until you at Ileast read the article I linked to which demonstrates that is possible to have very successful CEOS who don't have $500 million salaries. [NEWLINE] [NEWLINE] Again, I'm talking about the entire market being overvalued. I don't think any CEO is *really worth* $500 million. You seem to not understand this viewpoint and are trying to get me to assume that CEOs are worth $500 million.<mask> I'm checking out of this debate </s>
Label encoding: <s>Also, we have a fundamental different viewpoint of workers. I would like a more holistic  approach which doesn't treat workers solely as commodities. You seem fine with corporations doing so. [NEWLINE] [NEWLINE] I'm not interested in arguing this any further until you at Ileast read the article I linked to which demonstrates that is possible to have very successful CEOS who don't have $500 million salaries. [NEWLINE] [NEWLINE] Again, I'm talking about the entire market being overvalued. I don't think any CEO is *really worth* $500 million. You seem to not understand this viewpoint and are trying to get me to assume that CEOs are worth $500 million. So I'm checking out of this debate </s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a South East Asian, who are not familiar with Confucius, I don't think<mask><mask> with this. Asian staple food is carbs, rice or noodles or other starches.. And we don't eat steak. Fish or chicken can be dealt with spoons (or chopsticks,<mask> we're not talking about the Chinese here) easily.. And beef is often cooked in small pieces<mask> the spices are absorbed better. [NEWLINE] [NEWLINE] <mask> the need to cut something on your plate is really not common. [NEWLINE] [NEWLINE] Edit: to think of it spoons are probably a heritage from the European colonial era..<mask> they had knives too<mask> they don't get adopted nearly<mask> well<mask> spoons. </s>
Label encoding: <s>As a South East Asian, who are not familiar with Confucius, I don't think I agree with this. Asian staple food is carbs, rice or noodles or other starches.. And we don't eat steak. Fish or chicken can be dealt with spoons (or chopsticks, but we're not talking about the Chinese here) easily.. And beef is often cooked in small pieces so the spices are absorbed better. [NEWLINE] [NEWLINE] So the need to cut something on your plate is really not common. [NEWLINE] [NEWLINE] Edit: to think of it spoons are probably a heritage from the European colonial era.. But they had knives too but they don't get adopted nearly as well as spoons. </s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I realize that it is infinitely easier to live<mask> a closeted atheist than to live<mask> a closeted homosexual. <mask>, being gay has far more acceptance than being an atheist.  Neither is a choice,<mask> a person can't force themselves to believe in a god any easier than they can force themselves to think the Earth is the center of the universe. [NEWLINE] [NEWLINE] I don't believe either is an easy thing to be open about in certain areas of the country,<mask> coming out<mask> gay is a lot more accepted than coming out<mask> an atheist. <mask><mask> it ultimately comes down to the fact that homosexuality doesn't inherently make the claim that a person's religion is false.  </s>
Label encoding: <s>I realize that it is infinitely easier to live as a closeted atheist than to live as a closeted homosexual.  However, being gay has far more acceptance than being an atheist.  Neither is a choice, as a person can't force themselves to believe in a god any easier than they can force themselves to think the Earth is the center of the universe. [NEWLINE] [NEWLINE] I don't believe either is an easy thing to be open about in certain areas of the country, but coming out as gay is a lot more accepted than coming out as an atheist.  I think it ultimately comes down to the fact that homosexuality doesn't inherently make the claim that a person's religion is false.  </s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Exercise machines are about performing repetitive tasks and enduring their pain to achieve a long-term goal. [NEWLINE] [NEWLINE] Playing like children stimulates imagination, social skills, humour, even strategising. [NEWLINE] [NEWLINE] Enduring drudgery and pain to achieve some goal can be useful,<mask> doing it all the time, never having fun, is a terrible way to live. This sounds better to me: [NEWLINE] [STARTQ] For true hackers, the boundaries between "play", "work", "science" and "art" all tend to disappear, or to merge into a high-level creative playfulness. [ENDQ] -- [URL] [NEWLINE] [NEWLINE] Note: "Hacker" in this context does not refer to a type of criminal.</s>
Label encoding: <s>Exercise machines are about performing repetitive tasks and enduring their pain to achieve a long-term goal. [NEWLINE] [NEWLINE] Playing like children stimulates imagination, social skills, humour, even strategising. [NEWLINE] [NEWLINE] Enduring drudgery and pain to achieve some goal can be useful, but doing it all the time, never having fun, is a terrible way to live. This sounds better to me: [NEWLINE] [STARTQ] For true hackers, the boundaries between "play", "work", "science" and "art" all tend to disappear, or to merge into a high-level creative playfulness. [ENDQ] -- [URL] [NEWLINE] [NEWLINE] Note: "Hacker" in this context does not refer to a type of criminal.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Careful!  Our society has become<mask> WEAK from the feminism that has infiltrated every bit of our government and corporations that our standards of engineering have declined abysmally.  There's every chance the Sun Lander Project will be infiltrated by feminist engineers who will bring their PC weakness to our manly Sun Lander.  Then they will use their underhanded betrayal of the Sun Lander Project to showcase<mask> True American^TM Manliness is a failure, like they have done<mask> often in the past. [NEWLINE] [NEWLINE] <mask> we want it to succeed we must first scour the feminist influence from every aspect of our society<mask> the True American Sun Lander Project can succeed like it deserves!</s>
Label encoding: <s>Careful!  Our society has become SO WEAK from the feminism that has infiltrated every bit of our government and corporations that our standards of engineering have declined abysmally.  There's every chance the Sun Lander Project will be infiltrated by feminist engineers who will bring their PC weakness to our manly Sun Lander.  Then they will use their underhanded betrayal of the Sun Lander Project to showcase how True American^TM Manliness is a failure, like they have done so often in the past. [NEWLINE] [NEWLINE] If we want it to succeed we must first scour the feminist influence from every aspect of our society so the True American Sun Lander Project can succeed like it deserves!</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0597, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> whatever 51% of the population says is automatically true? [ENDQ] [NEWLINE] (1) Truth is for mathematics and for scientific claims. Political opinions are about<mask> we feel we should live. They<mask> concern themselves with subjective beliefs and desires, not objective truth and facts. [NEWLINE] [NEWLINE] (2) The US is a republican democracy, not a direct democracy. There exist protections that are meant to shield the minority from the tyranny of majority opinions. It is imperfect which is<mask> it took us<mask> long to decide that slavery was immoral. Nevertheless it does work. [NEWLINE] [NEWLINE] [STARTQ] <mask> an absurd premise. [ENDQ] [NEWLINE] Your examples are strawmen<mask> you falsely misrepresent the facts of<mask> the US constitution actually says.</s>
Label encoding: <s> [STARTQ] So whatever 51% of the population says is automatically true? [ENDQ] [NEWLINE] (1) Truth is for mathematics and for scientific claims. Political opinions are about how we feel we should live. They therefore concern themselves with subjective beliefs and desires, not objective truth and facts. [NEWLINE] [NEWLINE] (2) The US is a republican democracy, not a direct democracy. There exist protections that are meant to shield the minority from the tyranny of majority opinions. It is imperfect which is why it took us so long to decide that slavery was immoral. Nevertheless it does work. [NEWLINE] [NEWLINE] [STARTQ] What an absurd premise. [ENDQ] [NEWLINE] Your examples are strawmen because you falsely misrepresent the facts of what the US constitution actually says.</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>We can't really know that there is no way to end all war.  Human civilization has already evolved in many unexpected ways, which earlier cultures would have regarded<mask> impossible.  Nonetheless, it is quite true that at this time, there is a severe lack of political unity in our world, and there is no consensus about the need to end war, terrorism, or political violence of various sorts.  Obviously, it will be tremendously hard to obtain that kind of consensus. <mask> I am trying to argue, is that we cannot give up on that struggle,<mask> in the end, war will destroy us. <mask> we want to survive<mask> a species, we do need world peace.</s>
Label encoding: <s>We can't really know that there is no way to end all war.  Human civilization has already evolved in many unexpected ways, which earlier cultures would have regarded as impossible.  Nonetheless, it is quite true that at this time, there is a severe lack of political unity in our world, and there is no consensus about the need to end war, terrorism, or political violence of various sorts.  Obviously, it will be tremendously hard to obtain that kind of consensus.  What I am trying to argue, is that we cannot give up on that struggle, since in the end, war will destroy us.  If we want to survive as a species, we do need world peace.</s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why can't we have both? [NEWLINE] [NEWLINE] Children who are not being challenged enough in school may grow apathetic towards learning and education. Children who struggle in school may grow apathetic towards learning and education. [NEWLINE] [NEWLINE] <mask> children are different from each other and may require differing methods of teaching, we should strive to provide the best education for each child<mask> per their abilities. [NEWLINE] [NEWLINE] In the school system I grew up in, struggling kids were given extra time with the teacher, and gifted kids were too. Middle of the road kids were given fairly ample time<mask> well, it was the point of separating us out like that<mask> we could have our strengths and learning speeds better suited to our needs.</s>
Label encoding: <s>Why can't we have both? [NEWLINE] [NEWLINE] Children who are not being challenged enough in school may grow apathetic towards learning and education. Children who struggle in school may grow apathetic towards learning and education. [NEWLINE] [NEWLINE] Because children are different from each other and may require differing methods of teaching, we should strive to provide the best education for each child as per their abilities. [NEWLINE] [NEWLINE] In the school system I grew up in, struggling kids were given extra time with the teacher, and gifted kids were too. Middle of the road kids were given fairly ample time as well, it was the point of separating us out like that so we could have our strengths and learning speeds better suited to our needs.</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well,<mask> that were true (and it is in a sense,<mask><mask><mask> a lot of consequentialists that aren't act utilitarians might dispute it), then I would say that rights simply can't be reconciled with utilitarian ethics without bringing in some other concept like justice [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask>,<mask> you believe that this absolute label will prevent more harm (people mistakenly thinking they have a circumstance<mask> they ought to rape) than it will cause (people missing a circumstance<mask> they ought to) then the Utilitarian has a moral obligation to apply the absolute label. [ENDQ] [NEWLINE] This doesn't address the hypothetical from an individual perspective<mask> you're very convinced by your assessment of the consequences</s>
Label encoding: <s>Well, if that were true (and it is in a sense, but I think a lot of consequentialists that aren't act utilitarians might dispute it), then I would say that rights simply can't be reconciled with utilitarian ethics without bringing in some other concept like justice [NEWLINE] [NEWLINE] [STARTQ] On the contrary, if you believe that this absolute label will prevent more harm (people mistakenly thinking they have a circumstance where they ought to rape) than it will cause (people missing a circumstance where they ought to) then the Utilitarian has a moral obligation to apply the absolute label. [ENDQ] [NEWLINE] This doesn't address the hypothetical from an individual perspective if you're very convinced by your assessment of the consequences</s>
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ahh, yes I did not view it in that regard. ∆ Of course making people sad is certainly a negative effect,<mask> wasn't really<mask> I was looking for. The argument behind stereotypical portrayals of women in video games (<mask><mask> my girlfriend) is that contributes to the idea that men are in power, and<mask> directly influences societal problems that affect women due to a patriarchal system. I'm wondering<mask> there is any evidence that there is a link between these portrayals of women and real societal problems (e.g. Gender wage gap, underrepresentation of women in STEM fields) or<mask> this is only seen<mask> a problem "<mask> it hurts peoples' feelings". </s>
Label encoding: <s>Ahh, yes I did not view it in that regard. ∆ Of course making people sad is certainly a negative effect, but wasn't really what I was looking for. The argument behind stereotypical portrayals of women in video games ( according to my girlfriend) is that contributes to the idea that men are in power, and thus directly influences societal problems that affect women due to a patriarchal system. I'm wondering if there is any evidence that there is a link between these portrayals of women and real societal problems (e.g. Gender wage gap, underrepresentation of women in STEM fields) or if this is only seen as a problem " because it hurts peoples' feelings". </s>
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0518, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0749, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] They are charging more and throttling bandwidth<mask> they have a competing service that they want to incentivize you to use in favor of Netflix [ENDQ] [NEWLINE] <mask> they are doing it in order to create an incentive to use their own service, then I believe that Netflix would have ground to sue for Verizon's anti-compete practices. [NEWLINE] [NEWLINE] [STARTQ] Bandwidth is bandwidth. Netflix bandwidth isn't somehow more costly than<mask> I were downloading 12 mbps of music from iTunes, for instance. [ENDQ] [NEWLINE] I am not saying that Netflix is more costly than other services,<mask><mask> I'm not watching Netflix, it doesn't automatically mean that I am going to use another bandwidth hogging service instead.</s>
Label encoding: <s> [STARTQ] They are charging more and throttling bandwidth because they have a competing service that they want to incentivize you to use in favor of Netflix [ENDQ] [NEWLINE] If they are doing it in order to create an incentive to use their own service, then I believe that Netflix would have ground to sue for Verizon's anti-compete practices. [NEWLINE] [NEWLINE] [STARTQ] Bandwidth is bandwidth. Netflix bandwidth isn't somehow more costly than if I were downloading 12 mbps of music from iTunes, for instance. [ENDQ] [NEWLINE] I am not saying that Netflix is more costly than other services, but if I'm not watching Netflix, it doesn't automatically mean that I am going to use another bandwidth hogging service instead.</s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It most assuredly does, this is not even under question.<mask> women have a better success rate under gender-blind review then the study would be out claiming they have been discriminated against,<mask> is it not the same for men? [ENDQ] [NEWLINE] Whether women do better under gender blind review does not yield any insight into whether men do. There is only evidence of gender bias<mask> one gender improves significantly more than the other under gender blind review. The men did not receive a larger bonus than the women under the gender-blind review,<mask> there is no evidence of gender bias against men. The facts<mask> presented do not support the claim that women are favored over men by reviewers. </s>
Label encoding: <s> [STARTQ] It most assuredly does, this is not even under question. If women have a better success rate under gender-blind review then the study would be out claiming they have been discriminated against, why is it not the same for men? [ENDQ] [NEWLINE] Whether women do better under gender blind review does not yield any insight into whether men do. There is only evidence of gender bias if one gender improves significantly more than the other under gender blind review. The men did not receive a larger bonus than the women under the gender-blind review, so there is no evidence of gender bias against men. The facts as presented do not support the claim that women are favored over men by reviewers. </s>
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The convenience of carry makes it useful for self defense. Not only against other people,<mask> it can be used against animals<mask> well. Maybe not a bear,<mask> a snake or rabid dog it would work for. [NEWLINE] [NEWLINE] <mask>, your argument is a little weak. You can look at anything and give it another use. You could say texting's main purpose is to kill people. It is really distracting, and it is very inefficient. You could easily call someone, and be able to see<mask> you are doing. Calls are<mask> much faster than texting.<mask> many people die from calling and driving. Not many. See<mask> easy it is to twist something's purpose? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>The convenience of carry makes it useful for self defense. Not only against other people, but it can be used against animals as well. Maybe not a bear, but a snake or rabid dog it would work for. [NEWLINE] [NEWLINE] Also, your argument is a little weak. You can look at anything and give it another use. You could say texting's main purpose is to kill people. It is really distracting, and it is very inefficient. You could easily call someone, and be able to see what you are doing. Calls are also much faster than texting. How many people die from calling and driving. Not many. See how easy it is to twist something's purpose? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0573, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0873, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think part of it would depend on the person.<mask> you're good at math and have an interest in science you might find getting a degree in physics easier than getting a degree in Russian. The relative difficulty of different degrees<mask> depends on<mask> well you want to do in your courses. Passing an art history course is likely easier than passing a calculus course,<mask> getting an A+ in art history is likely harder than getting an A+ in calculus. Ultimately,<mask> one's goal is just to get a degree, a non-STEM major is likely easier,<mask> for anyone willing to put the work in, both STEM and non-STEM programs can be challenging and rewarding. </s>
Label encoding: <s>I think part of it would depend on the person. If you're good at math and have an interest in science you might find getting a degree in physics easier than getting a degree in Russian. The relative difficulty of different degrees also depends on how well you want to do in your courses. Passing an art history course is likely easier than passing a calculus course, but getting an A+ in art history is likely harder than getting an A+ in calculus. Ultimately, if one's goal is just to get a degree, a non-STEM major is likely easier, but for anyone willing to put the work in, both STEM and non-STEM programs can be challenging and rewarding. </s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[You're right that showing people evidence is a poor way to convince them of an argument.]( [URL] /)<mask>, people who have already made up their mind on racist issues aren't terribly helpful to a civil rights cause in the first place, and sometimes, the best you can hope for is that someone who is still on the fence will hear your argument and be swayed. [NEWLINE] [NEWLINE] <mask> dealing with racists, I tend to use the strategy outlined, just continually ask them to explain their views and watch<mask> they get more and more moderate.<mask> generally speaking, the utility for all these arguments is to sway the people who haven't dug in their feet and doubled down on their bigotry.</s><pad>
Label encoding: <s>[You're right that showing people evidence is a poor way to convince them of an argument.]( [URL] /) However, people who have already made up their mind on racist issues aren't terribly helpful to a civil rights cause in the first place, and sometimes, the best you can hope for is that someone who is still on the fence will hear your argument and be swayed. [NEWLINE] [NEWLINE] When dealing with racists, I tend to use the strategy outlined, just continually ask them to explain their views and watch as they get more and more moderate. But generally speaking, the utility for all these arguments is to sway the people who haven't dug in their feet and doubled down on their bigotry.</s><pad>
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know that the court wants to make a distinction,<mask> I am saying that it is arbitrary and illogical, given the reasoning for upending the existing laws. [NEWLINE] [NEWLINE] <mask> a right is<mask> fundamental that the court has to strip powers from the states, it logically follows that the inconvenience of applying that right consistently isn't a sufficiently compelling reason to make an arbitrary distinction between 2 males and 3 males. [NEWLINE] [NEWLINE] I don't disagree with you that they will likely enforce that distinction. I do claim that anyone who thinks the previous laws infringed on anyone's rights needs to hold that the new laws<mask> infringe in the same way, even<mask> fewer people are affected. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I know that the court wants to make a distinction, but I am saying that it is arbitrary and illogical, given the reasoning for upending the existing laws. [NEWLINE] [NEWLINE] If a right is so fundamental that the court has to strip powers from the states, it logically follows that the inconvenience of applying that right consistently isn't a sufficiently compelling reason to make an arbitrary distinction between 2 males and 3 males. [NEWLINE] [NEWLINE] I don't disagree with you that they will likely enforce that distinction. I do claim that anyone who thinks the previous laws infringed on anyone's rights needs to hold that the new laws also infringe in the same way, even if fewer people are affected. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Cultural appropriation which isn't considered harmful is simply called cultural exchange.<mask> the appropriation is "a show of admiration for them and the cultures whence they come" then it is exchange. It happens all the time and is<mask> fine. [NEWLINE] [NEWLINE] Even some forms of cultural appropriation are, like you said, "at worst benign".<mask> really it's a case by case approach we should be taking and not a question over whether cultural appropriation is/isn't generally fine. [NEWLINE] [NEWLINE] <mask> there are cases<mask> cultural exchange is both inappropriate (making it appropriation) and tangible harmful (making it more than just benign). [NEWLINE] [NEWLINE] <mask> you would like examples feel free to ask. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Cultural appropriation which isn't considered harmful is simply called cultural exchange. If the appropriation is "a show of admiration for them and the cultures whence they come" then it is exchange. It happens all the time and is indeed fine. [NEWLINE] [NEWLINE] Even some forms of cultural appropriation are, like you said, "at worst benign". So really it's a case by case approach we should be taking and not a question over whether cultural appropriation is/isn't generally fine. [NEWLINE] [NEWLINE] So there are cases where cultural exchange is both inappropriate (making it appropriation) and tangible harmful (making it more than just benign). [NEWLINE] [NEWLINE] If you would like examples feel free to ask. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0621, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0648, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am not aware of any country in the EU which prevents multiple citizenships,<mask> you know best, I guess. [NEWLINE] [NEWLINE] And a Green Card just doesn't let you vote in presidential elections,<mask> is still an excellent asset. The important thing to consider here is that you don't know<mask> the world will be in 20 years,<mask> your children are adults,<mask> you should strive to give them the maximum amount of options. Personally, the Eurozone isn't looking too hot, and<mask> the Eurozone collapses, the EU will too, and the lack of an open market will without doubt hurt all European economies,<mask> I would get the sweet fuck out of Europe.</s>
Label encoding: <s>I am not aware of any country in the EU which prevents multiple citizenships, but you know best, I guess. [NEWLINE] [NEWLINE] And a Green Card just doesn't let you vote in presidential elections, but is still an excellent asset. The important thing to consider here is that you don't know how the world will be in 20 years, when your children are adults, so you should strive to give them the maximum amount of options. Personally, the Eurozone isn't looking too hot, and if the Eurozone collapses, the EU will too, and the lack of an open market will without doubt hurt all European economies, so I would get the sweet fuck out of Europe.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry kabraxcis, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=kabraxcis+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry kabraxcis, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=kabraxcis+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry TornadoCreator, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even<mask> the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] <mask> you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=TornadoCreator+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Label encoding: <s>Sorry TornadoCreator, your comment has been removed: [NEWLINE] [NEWLINE] [STARTQ] Comment Rule 2\. "Don't be rude or hostile to other users. Your comment will be removed even if the rest of it is solid." [See the wiki page for more information.]( [URL] #wiki_rule_2) [ENDQ] [NEWLINE] If you would like to appeal, please [message the moderators by clicking this link.]( [URL] ;subject=Removed+Comment+Rule+2+Post+Appeal&amp;message=TornadoCreator+would+like+to+appeal+the+removal+of+[his/her+post]( [URL] \))</s>
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've eaten in restaurants for years. I always tip my server. I always ask for whatever changes to the menu I want, and compromise<mask> they can't oblige exactly. I rarely send anything back,<mask> I rarely get served anything<mask> inedible it calls for replacement. [NEWLINE] [NEWLINE] In other words, just normal day-to-day human interactions, not calling for retaliation on either side. You sound like someone ready to pounce and apply the final solution for<mask> is often just a minor error and easily remedied. [NEWLINE] [NEWLINE] And<mask> you don't care for a particular place, just don't go back. No need to nuke it. [NEWLINE] </s>
Label encoding: <s>I've eaten in restaurants for years. I always tip my server. I always ask for whatever changes to the menu I want, and compromise if they can't oblige exactly. I rarely send anything back, because I rarely get served anything so inedible it calls for replacement. [NEWLINE] [NEWLINE] In other words, just normal day-to-day human interactions, not calling for retaliation on either side. You sound like someone ready to pounce and apply the final solution for what is often just a minor error and easily remedied. [NEWLINE] [NEWLINE] And if you don't care for a particular place, just don't go back. No need to nuke it. [NEWLINE] </s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0514, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And to add on,<mask> I stayed inside to play LoL, [NEWLINE] [NEWLINE] * I probably won't have a girlfriend. [NEWLINE] * Probably won't be making a bunch of new friends in Greek Life [NEWLINE] * Almost certainly won't be networking and be working this awesome job, (which I'm at right now) [NEWLINE] [NEWLINE] Quitting LoL has made my life better in almost every aspect. Hell, I even got better grades! I'm spending an extra 30+ hours experiencing college, building my career, and meeting new people. Now,<mask> I meet someone I don't know, I can actually have a conversation with them about<mask> I do for fun. </s>
Label encoding: <s>And to add on, if I stayed inside to play LoL, [NEWLINE] [NEWLINE] * I probably won't have a girlfriend. [NEWLINE] * Probably won't be making a bunch of new friends in Greek Life [NEWLINE] * Almost certainly won't be networking and be working this awesome job, (which I'm at right now) [NEWLINE] [NEWLINE] Quitting LoL has made my life better in almost every aspect. Hell, I even got better grades! I'm spending an extra 30+ hours experiencing college, building my career, and meeting new people. Now, if I meet someone I don't know, I can actually have a conversation with them about what I do for fun. </s>
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> to respond... Is your overall argument that people should talk more to other people? Sure.<mask> having been a girl in a park,<mask> well<mask> the "diverse" minority in a neighborhood, I don't always want to have an intellectual conversation about my background or views. [NEWLINE] [NEWLINE] That's<mask> groups, clubs and social groups revolving around certain topics address. They provide a platform for discussion<mask> people don't always want to talk. In general, the people who go to those clubs are the same people from your incredibly diverse neighborhood,<mask> by your geographical description you should have no problem finding people of varying backgrounds enjoying the same hobby. </s>
Label encoding: <s>I'm not sure how to respond... Is your overall argument that people should talk more to other people? Sure. But having been a girl in a park, as well as the "diverse" minority in a neighborhood, I don't always want to have an intellectual conversation about my background or views. [NEWLINE] [NEWLINE] That's what groups, clubs and social groups revolving around certain topics address. They provide a platform for discussion because people don't always want to talk. In general, the people who go to those clubs are the same people from your incredibly diverse neighborhood, so by your geographical description you should have no problem finding people of varying backgrounds enjoying the same hobby. </s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But ISIL is *not* and entirely new sounding entity.  The US government is regularly using that name to refer to<mask> most of the rest of the media are calling "ISIS".  The media reports on the government's statements, often with direct quotes,<mask> we often get "ISIL" and "ISIS" used in the same sentence (or at least paragraph) to refer to the same group. [NEWLINE] [NEWLINE] The confusion (<mask> there is any) is due to two names being used concurrently.  The most elegant solution is for one of the two names to be adopted universally. <mask><mask> the better name for that adoption would be "ISIL".</s>
Label encoding: <s>But ISIL is *not* and entirely new sounding entity.  The US government is regularly using that name to refer to what most of the rest of the media are calling "ISIS".  The media reports on the government's statements, often with direct quotes, so we often get "ISIL" and "ISIS" used in the same sentence (or at least paragraph) to refer to the same group. [NEWLINE] [NEWLINE] The confusion ( if there is any) is due to two names being used concurrently.  The most elegant solution is for one of the two names to be adopted universally.  I think the better name for that adoption would be "ISIL".</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If we lived in a world<mask> we could make everyone like or respect everyone else, it **would** be the bully's duty to not harass someone. <mask> we live in a far less than perfect world, people do<mask> they **can** and not<mask> they **should**. [NEWLINE] [NEWLINE] We **should** live in a world free from pain, and hunger, and fear and hatred, and abuse, and<mask> much more....<mask> we don't. We're imperfect people living in an imperfect world. [NEWLINE] [NEWLINE] Is it **really** such a bad thing that I believe we should educate both bully and victim on<mask> to deal with these situations? [NEWLINE] </s>
Label encoding: <s>If we lived in a world where we could make everyone like or respect everyone else, it **would** be the bully's duty to not harass someone.  Since we live in a far less than perfect world, people do what they **can** and not what they **should**. [NEWLINE] [NEWLINE] We **should** live in a world free from pain, and hunger, and fear and hatred, and abuse, and so much more.... But we don't. We're imperfect people living in an imperfect world. [NEWLINE] [NEWLINE] Is it **really** such a bad thing that I believe we should educate both bully and victim on how to deal with these situations? [NEWLINE] </s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0516, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe<mask> people say that to your wife, they're just being polite.  Maybe they pity her for giving up her great career.  You judge mothers for working, and other people judge women for staying home, especially<mask> they had a good career. [NEWLINE] [NEWLINE] And by "can't afford", these people probably mean that they couldn't maintain their pre-parenthood lifestyles without a second income.  And<mask>'s wrong with that?  Not everybody believes you have to give up<mask> much of yourself in order to be a good parent.  Happy people make the best parents, and<mask> working allows you to to be happy, then you should work.</s>
Label encoding: <s>Maybe when people say that to your wife, they're just being polite.  Maybe they pity her for giving up her great career.  You judge mothers for working, and other people judge women for staying home, especially if they had a good career. [NEWLINE] [NEWLINE] And by "can't afford", these people probably mean that they couldn't maintain their pre-parenthood lifestyles without a second income.  And what's wrong with that?  Not everybody believes you have to give up so much of yourself in order to be a good parent.  Happy people make the best parents, and if working allows you to to be happy, then you should work.</s>
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] With analog clocks, no mental math is required. It's all visual. [ENDQ] [NEWLINE] <mask><mask>. More mental math is required<mask> people generally understand<mask> long a minute is, and<mask> long certain actions take. That's<mask> we judge time - not by<mask> far away a minute hand or hour hand (which, coincidentally, occupy the same space on a clock) is from the point on a clock. Maybe you can look at a clock and recognize<mask> much time you have,<mask> only<mask> you've learned to read a clock and know<mask> it represents. It's not all visual; the visual is a representation which must be mentally translated. </s>
Label encoding: <s> [STARTQ] With analog clocks, no mental math is required. It's all visual. [ENDQ] [NEWLINE] I disagree. More mental math is required because people generally understand how long a minute is, and how long certain actions take. That's how we judge time - not by how far away a minute hand or hour hand (which, coincidentally, occupy the same space on a clock) is from the point on a clock. Maybe you can look at a clock and recognize how much time you have, but only because you've learned to read a clock and know what it represents. It's not all visual; the visual is a representation which must be mentally translated. </s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that you're wrong,<mask> like BenIncognito said we can't do anything that is against our nature. Our species is not a solid monument, it evolves and changes constantly,<mask> there isn't one Human Nature that will follow us and haunt our conscience through the ages. The Deus Ex HR had one beautiful ending in which you flooded the whole control-the-world base and leave the humans in charge - not affected by any bias.<mask><mask> that this is<mask> we should all do - express our opinions, even<mask> they are contrary, and in the end entire specie over the course of few hundred years will make it's choice.</s>
Label encoding: <s>I think that you're wrong, because like BenIncognito said we can't do anything that is against our nature. Our species is not a solid monument, it evolves and changes constantly, so there isn't one Human Nature that will follow us and haunt our conscience through the ages. The Deus Ex HR had one beautiful ending in which you flooded the whole control-the-world base and leave the humans in charge - not affected by any bias. I think that this is what we should all do - express our opinions, even if they are contrary, and in the end entire specie over the course of few hundred years will make it's choice.</s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0589, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm talking specifically about the group of unreasonable people who are 'pro-vaccination'. [NEWLINE] [NEWLINE] I am pro-vaccination,<mask><mask> these people are is not radicals in a group,<mask> a bunch of idiots jumping on a band-wagon of hate. They seem incapable of understanding that injecting babies with strains of deadly diseases could potentially be bad for their health. For the most part it isn't in the way it is done,<mask> it's far from perfect. [NEWLINE] [NEWLINE] Jim Carrey seems to be pro-efficient-and-safe-vaccination, not anti-vaccination;<mask> these idiots decided that he's a destructive moron.</s>
Label encoding: <s>I'm talking specifically about the group of unreasonable people who are 'pro-vaccination'. [NEWLINE] [NEWLINE] I am pro-vaccination, but what these people are is not radicals in a group, but a bunch of idiots jumping on a band-wagon of hate. They seem incapable of understanding that injecting babies with strains of deadly diseases could potentially be bad for their health. For the most part it isn't in the way it is done, but it's far from perfect. [NEWLINE] [NEWLINE] Jim Carrey seems to be pro-efficient-and-safe-vaccination, not anti-vaccination; yet these idiots decided that he's a destructive moron.</s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I do,<mask> definitely not with all black people. Mainly just black men who remind me of the way he looked and spoke. Edit: I just saw your edits.<mask><mask> associating a traumatic event with situations and people is pretty normal. I don't think every black man who looks like him is going to assault me, and I certainly don't treat people differently<mask> of it. I always try to come to my own conclusions about people<mask> I get to know them.<mask> sometimes I can't help<mask> be reminded of the event<mask> I see someone dressed the same way, or have similar features, or with the same raspy voice.</s>
Label encoding: <s>I do, but definitely not with all black people. Mainly just black men who remind me of the way he looked and spoke. Edit: I just saw your edits. I think associating a traumatic event with situations and people is pretty normal. I don't think every black man who looks like him is going to assault me, and I certainly don't treat people differently because of it. I always try to come to my own conclusions about people as I get to know them. But sometimes I can't help but be reminded of the event when I see someone dressed the same way, or have similar features, or with the same raspy voice.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are drawing a connection between two unlike things. Neighbors making you wear a dumb hat serves no purpose. Government does, and government cannot function under the ridiculous idea that everyone must explicitly consent in order to be under their rule. Tacit consent is a long standing theory, and all of your complaints essentially regress back to, "<mask> I don't wanna," which is not very compelling. You have natural rights to live<mask> you want. Without a government your natural rights are very challenging to enforce and<mask> they are very poorly protected, you seem to be weighing the, "<mask> I don't wanna," more heavily than having very hard to protect rights.</s>
Label encoding: <s>You are drawing a connection between two unlike things. Neighbors making you wear a dumb hat serves no purpose. Government does, and government cannot function under the ridiculous idea that everyone must explicitly consent in order to be under their rule. Tacit consent is a long standing theory, and all of your complaints essentially regress back to, " But I don't wanna," which is not very compelling. You have natural rights to live where you want. Without a government your natural rights are very challenging to enforce and therefore they are very poorly protected, you seem to be weighing the, " But I don't wanna," more heavily than having very hard to protect rights.</s>
Loss: tensor(0.0576, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your issue is with the term superior or inferior.<mask> certain aspects of a culture may be more criminal, or offensive, this does not make them inherently inferior. Inferior is entirely up to the person viewing it. [NEWLINE] [NEWLINE] <mask> intrinsic numerical value would you assign to each trait of various cultures that you could weigh them against each other and come up with a white culture&gt;black culture answer? For example black culture<mask> promotes physical exercise at least in males. Maybe that is worth +5 points.<mask> that scale is entirely in the eyes of the beholder. Nothing any human does is implicitly inferior or superior than<mask> another does.</s>
Label encoding: <s>I think your issue is with the term superior or inferior. While certain aspects of a culture may be more criminal, or offensive, this does not make them inherently inferior. Inferior is entirely up to the person viewing it. [NEWLINE] [NEWLINE] What intrinsic numerical value would you assign to each trait of various cultures that you could weigh them against each other and come up with a white culture&gt;black culture answer? For example black culture also promotes physical exercise at least in males. Maybe that is worth +5 points. But that scale is entirely in the eyes of the beholder. Nothing any human does is implicitly inferior or superior than what another does.</s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This view does hinge quite strongly on the idea that the best way to convert someone is to preach to them constantly. Most people are aware of the Christian worldview, and probably the things that would make nonbelievers consider it the most are attempts to make it seem healthier, more intelligent, more correct, etc...and none of those would necessarily come from refusing to let up in your proselytizing for a single second. [NEWLINE] [NEWLINE] I'm not saying Christians live in a way consistent with the notion of heaven/hell,<mask> that doesn't mean the most effective method of conversion is to put all their worldly business aside and rant at nonbelievers.</s>
Label encoding: <s>This view does hinge quite strongly on the idea that the best way to convert someone is to preach to them constantly. Most people are aware of the Christian worldview, and probably the things that would make nonbelievers consider it the most are attempts to make it seem healthier, more intelligent, more correct, etc...and none of those would necessarily come from refusing to let up in your proselytizing for a single second. [NEWLINE] [NEWLINE] I'm not saying Christians live in a way consistent with the notion of heaven/hell, but that doesn't mean the most effective method of conversion is to put all their worldly business aside and rant at nonbelievers.</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] many [studies]( [URL] ) find that women are less ambitious on average than men. [ENDQ] [NEWLINE] [NEWLINE] Ok,<mask><mask><mask> a survey with a limited demographic (readers of a news paper) which doesn't even take into account<mask> career people have says that women expects to earn less and has 3% less women wanting to be executives compared to men. This doesn't even indicate something, it's just pure BS. Even the news article who refers to the survey doesn't jump to your conclusion. [NEWLINE] [NEWLINE] Am I missing something here or could you at least link a scientific study supporting your claim instead of a news article which doesn't?</s>
Label encoding: <s> [STARTQ] many [studies]( [URL] ) find that women are less ambitious on average than men. [ENDQ] [NEWLINE] [NEWLINE] Ok, so according to a survey with a limited demographic (readers of a news paper) which doesn't even take into account what career people have says that women expects to earn less and has 3% less women wanting to be executives compared to men. This doesn't even indicate something, it's just pure BS. Even the news article who refers to the survey doesn't jump to your conclusion. [NEWLINE] [NEWLINE] Am I missing something here or could you at least link a scientific study supporting your claim instead of a news article which doesn't?</s>
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0688, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> such a law exists,<mask> your wording makes it seem safe to assume that doctors are prevented from providing services to cash-paying people outside of the universal healthcare system in order to prevent the system from collapsing. [NEWLINE] [NEWLINE] <mask> the choice had to be made between universal healthcare and the ability for doctors to provide services to whomever they want,<mask><mask> that we are more obligated to provide universal healthcare than to allow doctors the ability to serve whoever they want. No one is conscripted into service, doctors choose to be doctors and work hard to become doctors, doctors just don't get to pick and choose their clientele. </s>
Label encoding: <s>I don't know why such a law exists, but your wording makes it seem safe to assume that doctors are prevented from providing services to cash-paying people outside of the universal healthcare system in order to prevent the system from collapsing. [NEWLINE] [NEWLINE] If the choice had to be made between universal healthcare and the ability for doctors to provide services to whomever they want, I think that we are more obligated to provide universal healthcare than to allow doctors the ability to serve whoever they want. No one is conscripted into service, doctors choose to be doctors and work hard to become doctors, doctors just don't get to pick and choose their clientele. </s>
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Are you telling me there is in NO WAY there isn't the smallest bit of projection here? You aren't in denial of anything? You aren't proud of the argument you have presented to me<mask> far? [ENDQ] [NEWLINE] Oh I'm very keen at self-assessment. In certain aspects of my life I notice emotional projections, find myself in denial of certain truths, and am proud of certain arguments. In this case, I can assure you, I am not any of those 3 things (I haven't really provided an argument<mask> much<mask> using ad hominem to call into question your fast-and-loose grasp of psychology).</s>
Label encoding: <s> [STARTQ] Are you telling me there is in NO WAY there isn't the smallest bit of projection here? You aren't in denial of anything? You aren't proud of the argument you have presented to me thus far? [ENDQ] [NEWLINE] Oh I'm very keen at self-assessment. In certain aspects of my life I notice emotional projections, find myself in denial of certain truths, and am proud of certain arguments. In this case, I can assure you, I am not any of those 3 things (I haven't really provided an argument as much as using ad hominem to call into question your fast-and-loose grasp of psychology).</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My art teacher taught me,<mask> many art teachers probably taught others, that art is made and becomes successful mostly for two purposes. To be desired and practical. People buy chairs<mask> they are practical and the more aesthetically pleasing and well made ones are desired. Fan art is very much desired by fans and there for  it's practical and a good way to make money.<mask> people are using their talents to make really good original art instead of fan art they may see a decrease in desire from consumers<mask> their original character or song or<mask> ever it may be is not already actively sought after.<mask> really good fan art will always sell. </s>
Label encoding: <s>My art teacher taught me, as many art teachers probably taught others, that art is made and becomes successful mostly for two purposes. To be desired and practical. People buy chairs because they are practical and the more aesthetically pleasing and well made ones are desired. Fan art is very much desired by fans and there for  it's practical and a good way to make money. If people are using their talents to make really good original art instead of fan art they may see a decrease in desire from consumers because their original character or song or what ever it may be is not already actively sought after. But really good fan art will always sell. </s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I lived most of my entire life in a very wintery area. [NEWLINE] [NEWLINE] <mask><mask> the choice is only between getting a factory farm-raised cow and a wild moose, of course the moose is a better option<mask><mask><mask> amount of suffering. [NEWLINE] [NEWLINE] <mask> this doesn't make getting factory farmed animals magically okay. And<mask> most everyone can get enough food and nutrition from plant sources, this doesn't apply to most of the population. [NEWLINE] [NEWLINE] Bottom line:<mask> we got rid of animal farming and the only meat consumed was from hunting, I would consider that a giant step forward for humans, animals, and the environment.</s>
Label encoding: <s>I lived most of my entire life in a very wintery area. [NEWLINE] [NEWLINE] But if the choice is only between getting a factory farm-raised cow and a wild moose, of course the moose is a better option as far as amount of suffering. [NEWLINE] [NEWLINE] But this doesn't make getting factory farmed animals magically okay. And since most everyone can get enough food and nutrition from plant sources, this doesn't apply to most of the population. [NEWLINE] [NEWLINE] Bottom line: if we got rid of animal farming and the only meat consumed was from hunting, I would consider that a giant step forward for humans, animals, and the environment.</s>
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0667, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Any utilitarian would disagree with you. It's a question of "does this action cause more harm, or more good?" [NEWLINE] [NEWLINE] Masturbating about someone in the way you describe doesn't hurt them in any way.<mask> you don't know can't hurt you. The exception to this rule might be some emotional discomfort<mask> they found out about it,<mask> that's it. It's unlikely to occur and unlikely to cause and serious harm<mask> it does occur. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, such fantasies can enhance an already-pleasurable experience. [NEWLINE] [NEWLINE] In other words, it causes net pleasure,<mask> it's good.</s>
Label encoding: <s>Any utilitarian would disagree with you. It's a question of "does this action cause more harm, or more good?" [NEWLINE] [NEWLINE] Masturbating about someone in the way you describe doesn't hurt them in any way. What you don't know can't hurt you. The exception to this rule might be some emotional discomfort if they found out about it, but that's it. It's unlikely to occur and unlikely to cause and serious harm when it does occur. [NEWLINE] [NEWLINE] On the other hand, such fantasies can enhance an already-pleasurable experience. [NEWLINE] [NEWLINE] In other words, it causes net pleasure, so it's good.</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't see<mask> it matters. We aren't designed for any specific purpose.<mask> people choose monogamy, whatever their brains may be craving deep down, they're doing monogamy.<mask><mask><mask> it doesn't kill us it's a net positive. Or at least neutral. I don't care<mask> people choose to be monogamous or not. We are not designed to be polyamorous any more than designed to be monogamous. We aren't designed to be either. We exist and make choices and those choices determine our survival and procreation. That which survives isn't necessarily right or wrong it's just<mask> survives. </s>
Label encoding: <s>I don't see how it matters. We aren't designed for any specific purpose. If people choose monogamy, whatever their brains may be craving deep down, they're doing monogamy. As long as it doesn't kill us it's a net positive. Or at least neutral. I don't care if people choose to be monogamous or not. We are not designed to be polyamorous any more than designed to be monogamous. We aren't designed to be either. We exist and make choices and those choices determine our survival and procreation. That which survives isn't necessarily right or wrong it's just what survives. </s>
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I really don't think you understand the meaning of 'rape culture'.<mask> a practice is part of your culture, it's often<mask> easy to overlook that you won't even notice<mask> it's happened. Kind of like<mask> racist people don't even realize<mask> they're being racist. [NEWLINE] [NEWLINE] Just like a racist person might complain that black people'really don't have it that hard', you'll have men excusing poor behavior towards women- the way this CMV does- for the same reasons.<mask> it really means,<mask> they say that, is that they've never had to experience the repercussions of these cultural norms for themselves.</s>
Label encoding: <s>I really don't think you understand the meaning of 'rape culture'. When a practice is part of your culture, it's often so easy to overlook that you won't even notice when it's happened. Kind of like how racist people don't even realize when they're being racist. [NEWLINE] [NEWLINE] Just like a racist person might complain that black people'really don't have it that hard', you'll have men excusing poor behavior towards women- the way this CMV does- for the same reasons. What it really means, when they say that, is that they've never had to experience the repercussions of these cultural norms for themselves.</s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> go to a strip club<mask> you can see tits for free on your computer?  There is something about the experience, the sights and smells that makes it worth it. [NEWLINE] [NEWLINE] Why go to a bands live show<mask> you can listen to them at home for cheaper?  The sound quality at home is probably way better and you aren't subject to the sound guy putting the kick to high or the vocals to low.  There is something in the experience. [NEWLINE] [NEWLINE] You say you can replicate the warmth of vinyl with other means.  I would have no idea<mask> to do it or<mask> I could even afford it.  </s>
Label encoding: <s>Why go to a strip club when you can see tits for free on your computer?  There is something about the experience, the sights and smells that makes it worth it. [NEWLINE] [NEWLINE] Why go to a bands live show when you can listen to them at home for cheaper?  The sound quality at home is probably way better and you aren't subject to the sound guy putting the kick to high or the vocals to low.  There is something in the experience. [NEWLINE] [NEWLINE] You say you can replicate the warmth of vinyl with other means.  I would have no idea how to do it or if I could even afford it.  </s>
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe that addiction (psychological or physical) is just something brought about by personal weakness. It seems to me that addiction to something like cigarettes or alcohol is brought upon oneself. You may have cravings for something,<mask> that doesn't mean you have to indulge it. [NEWLINE] [NEWLINE] I understand that something similar to an addiction can occur<mask> used for pain, etc.<mask> that once the problem is gone, there is no reason to continue it. In the case of chronic pain, I don't consider non-abusive use an addiction. [NEWLINE] [NEWLINE] <mask> CMV. I want to hear<mask> you guys have to say about addiction.</s>
Label encoding: <s>I believe that addiction (psychological or physical) is just something brought about by personal weakness. It seems to me that addiction to something like cigarettes or alcohol is brought upon oneself. You may have cravings for something, but that doesn't mean you have to indulge it. [NEWLINE] [NEWLINE] I understand that something similar to an addiction can occur when used for pain, etc. but that once the problem is gone, there is no reason to continue it. In the case of chronic pain, I don't consider non-abusive use an addiction. [NEWLINE] [NEWLINE] So CMV. I want to hear what you guys have to say about addiction.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your whole argument seems to be "these other browsers copied Chrome,<mask> Chrome is no good."  Expand on these "other browsers" which have overtaken Chrome.  I mean<mask> many browsers are there anyway?  Firefox?  Not<mask> it's natively missing right click --&gt; Google image search and Right Click --&gt; translate.  Opera?  Safari?  And don't say IE is bitmo better<mask> it's asking me<mask> I want to disable Microsoft's own browser addons, or<mask> I REALLLLy want to open the Microsoft hotfix that I downloaded FROM Microsoft.com.  </s>
Label encoding: <s>Your whole argument seems to be "these other browsers copied Chrome, so Chrome is no good."  Expand on these "other browsers" which have overtaken Chrome.  I mean how many browsers are there anyway?  Firefox?  Not when it's natively missing right click --&gt; Google image search and Right Click --&gt; translate.  Opera?  Safari?  And don't say IE is bitmo better when it's asking me if I want to disable Microsoft's own browser addons, or if I REALLLLy want to open the Microsoft hotfix that I downloaded FROM Microsoft.com.  </s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's precisely<mask> you're saying. All of your talk about gut bacteria and metabolic changes completely ignores the fact that none of that matters<mask> you modulate your eating habits. [NEWLINE] [NEWLINE] No, not everyone can eat the same thing in the same amounts and be equally healthy. [NEWLINE] [NEWLINE] Whatever drugs you're on with whatever metabolic side effects can be neutralized in terms of their effects on your weight by eating a proper diet, and the same goes for your malfunctioning gut bacteria. [NEWLINE] [NEWLINE] At the end of the day getting fat comes from the food you eat, and you'll only get fat<mask> you eat too much food. </s>
Label encoding: <s>That's precisely what you're saying. All of your talk about gut bacteria and metabolic changes completely ignores the fact that none of that matters if you modulate your eating habits. [NEWLINE] [NEWLINE] No, not everyone can eat the same thing in the same amounts and be equally healthy. [NEWLINE] [NEWLINE] Whatever drugs you're on with whatever metabolic side effects can be neutralized in terms of their effects on your weight by eating a proper diet, and the same goes for your malfunctioning gut bacteria. [NEWLINE] [NEWLINE] At the end of the day getting fat comes from the food you eat, and you'll only get fat if you eat too much food. </s>
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0515, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> you are [seemingly]<mask> vehemently opposed to Sanders... [NEWLINE] [NEWLINE] <mask> more relevantly I'm not sure<mask> you expect from any candidate. You say you want concrete, achievable promises,<mask> every candidate from the past few elections has made many concrete promises that sound achievable - and followed through on almost none of them. [NEWLINE] [NEWLINE] <mask><mask><mask>, Sanders is the only one who understands that even the President can't just make laws by himself. He can't just wish things into existence. [NEWLINE] [NEWLINE] It sounds like you want a dictator who doesn't even claim to care about the working class.</s>
Label encoding: <s>I'm not sure why you are [seemingly] so vehemently opposed to Sanders... [NEWLINE] [NEWLINE] But more relevantly I'm not sure what you expect from any candidate. You say you want concrete, achievable promises, but every candidate from the past few elections has made many concrete promises that sound achievable - and followed through on almost none of them. [NEWLINE] [NEWLINE] In my opinion, Sanders is the only one who understands that even the President can't just make laws by himself. He can't just wish things into existence. [NEWLINE] [NEWLINE] It sounds like you want a dictator who doesn't even claim to care about the working class.</s>
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is really going to depend on<mask> you define natural... Most of those foods at your local farmers market are not natural in the way that a plant growing in the wild would be. They are often grown in poor soil using the aid of man made fertilizers, they are often sprayed with herbicides and insecticides (including organic approved pesticides for organic crops), and they have been either genetically altered or selectively bred for desirable traits. [NEWLINE] [NEWLINE] <mask> you define natural<mask> something<mask> it would be without human intervention, then there is little that is natural about the foods we eat (including organic crops at a local farmer's market).</s>
Label encoding: <s>This is really going to depend on how you define natural... Most of those foods at your local farmers market are not natural in the way that a plant growing in the wild would be. They are often grown in poor soil using the aid of man made fertilizers, they are often sprayed with herbicides and insecticides (including organic approved pesticides for organic crops), and they have been either genetically altered or selectively bred for desirable traits. [NEWLINE] [NEWLINE] If you define natural as something as it would be without human intervention, then there is little that is natural about the foods we eat (including organic crops at a local farmer's market).</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Many countries with ostensible free speech have provisions that forbid the incitement to violence [ENDQ] [NEWLINE] I know. That was my point. [NEWLINE] [NEWLINE] [STARTQ] You can be judged for it by society,<mask><mask><mask><mask> you aren't breaking any laws (ie, you aren't actually in possession or manufacture of child pornography) you'd be well within your rights to convey that and we<mask> a democratic society oughtn't interfere with that freedom [ENDQ] [NEWLINE] My first post was unclear. I meant "convey that child pornography is beautiful" to mean that the person would be showing others child pornography, which is very much a crime.</s>
Label encoding: <s> [STARTQ] Many countries with ostensible free speech have provisions that forbid the incitement to violence [ENDQ] [NEWLINE] I know. That was my point. [NEWLINE] [NEWLINE] [STARTQ] You can be judged for it by society, but so long as you aren't breaking any laws (ie, you aren't actually in possession or manufacture of child pornography) you'd be well within your rights to convey that and we as a democratic society oughtn't interfere with that freedom [ENDQ] [NEWLINE] My first post was unclear. I meant "convey that child pornography is beautiful" to mean that the person would be showing others child pornography, which is very much a crime.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> being proud and humbled isn't mutually exclusive. [NEWLINE] [NEWLINE] For example lets take the Olympic games, you would feel proud to be there,<mask> one of the best athletes from your country.<mask> at the same time you could feel small and insignificant compared to the size of the event and the skills of the opposition. [NEWLINE] [NEWLINE] Or maybe a kid that won some kinda contest of skill to meet the president. He would feel proud to have won the contest,<mask> at the same time he would feel insignificant/small compared to the president. [NEWLINE] [NEWLINE] I think both of the examples above could be described<mask> humbled.</s>
Label encoding: <s>I think being proud and humbled isn't mutually exclusive. [NEWLINE] [NEWLINE] For example lets take the Olympic games, you would feel proud to be there, as one of the best athletes from your country. But at the same time you could feel small and insignificant compared to the size of the event and the skills of the opposition. [NEWLINE] [NEWLINE] Or maybe a kid that won some kinda contest of skill to meet the president. He would feel proud to have won the contest, while at the same time he would feel insignificant/small compared to the president. [NEWLINE] [NEWLINE] I think both of the examples above could be described as humbled.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] wouldn't referring someone<mask> the first whatever to do something simply remind everyone that the difference is still there [ENDQ] [NEWLINE] That's the whole point of it. "First x to do y" announcements both celebrate<mask> far we have come, and serve<mask> a reminder that we have a long way to go. [NEWLINE] [NEWLINE] There was a time<mask> a black president would have been unthinkable. Pointing it out shows that we have progressed far enough to consider them equals,<mask> remind us that it's an uncommon thing and we shouldn't stop progressing. We reach our goal<mask> we stop having to make announcements about a particular group.</s>
Label encoding: <s> [STARTQ] wouldn't referring someone as the first whatever to do something simply remind everyone that the difference is still there [ENDQ] [NEWLINE] That's the whole point of it. "First x to do y" announcements both celebrate how far we have come, and serve as a reminder that we have a long way to go. [NEWLINE] [NEWLINE] There was a time when a black president would have been unthinkable. Pointing it out shows that we have progressed far enough to consider them equals, but remind us that it's an uncommon thing and we shouldn't stop progressing. We reach our goal when we stop having to make announcements about a particular group.</s>
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0696, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually, no, it's nothing like a watch,<mask> a watch is a physical thing that once stolen is no longer viable for commerce. By pirating software, you're not *removing* something that is up for sale.<mask> is it cutting into their profits<mask> they would not have made any money off of me anyway? [NEWLINE] [NEWLINE] <mask>, just<mask> /u/llsaidknockyouout mentioned, you make a good point<mask> you bring up the potential for pirating inspiring the uploaders to do more. I hadn't considered that, and I'll have to think on the implications of that.</s>
Label encoding: <s>Actually, no, it's nothing like a watch, because a watch is a physical thing that once stolen is no longer viable for commerce. By pirating software, you're not *removing* something that is up for sale. How is it cutting into their profits if they would not have made any money off of me anyway? [NEWLINE] [NEWLINE] However, just as /u/llsaidknockyouout mentioned, you make a good point when you bring up the potential for pirating inspiring the uploaders to do more. I hadn't considered that, and I'll have to think on the implications of that.</s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't feel a need for constant connection, personally. In some industries I can absolutely see<mask> this gives a person a competitive edge--making it a necessity to have a smartphone.<mask> this isn't my situation. [NEWLINE] [NEWLINE] Your reasoning makes sense, and is similar to my own line of thought on the subject (<mask> far).<mask> I don't need one professionally, and have access to the internet via the computer to a degree that satisfies me, I'm don't really think I need a smartphone at all. And missing all those snap chat poops is just one more reason not to get one...ha!!</s>
Label encoding: <s>I don't feel a need for constant connection, personally. In some industries I can absolutely see how this gives a person a competitive edge--making it a necessity to have a smartphone. But this isn't my situation. [NEWLINE] [NEWLINE] Your reasoning makes sense, and is similar to my own line of thought on the subject ( thus far). Because I don't need one professionally, and have access to the internet via the computer to a degree that satisfies me, I'm don't really think I need a smartphone at all. And missing all those snap chat poops is just one more reason not to get one...ha!!</s>
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sounds like someone has a reading comprehension problem. [NEWLINE] [NEWLINE] [STARTQ] **<mask> ** you are unable to perform in the bedroom due to porn related issues, yes it is bad. **<mask> ** porn destroys relationships and affects your social life, yes it is bad. [ENDQ] [NEWLINE] [STARTQ] internet and porn has affected **many** men and women negatively [ENDQ] [NEWLINE] [STARTQ] **Addiction is different for every individual, some may not even be affected by it.** [ENDQ] [NEWLINE] [STARTQ] **For me** porn is bad [ENDQ] [NEWLINE] <mask><mask> his second to last paragraph went off the deep end, it didn't really mesh well with the rest. [NEWLINE] </s>
Label encoding: <s>Sounds like someone has a reading comprehension problem. [NEWLINE] [NEWLINE] [STARTQ] ** If ** you are unable to perform in the bedroom due to porn related issues, yes it is bad. ** If ** porn destroys relationships and affects your social life, yes it is bad. [ENDQ] [NEWLINE] [STARTQ] internet and porn has affected **many** men and women negatively [ENDQ] [NEWLINE] [STARTQ] **Addiction is different for every individual, some may not even be affected by it.** [ENDQ] [NEWLINE] [STARTQ] **For me** porn is bad [ENDQ] [NEWLINE] I agree his second to last paragraph went off the deep end, it didn't really mesh well with the rest. [NEWLINE] </s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0709, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your experience is not typical. <mask> people feel that they're being shamed or discriminated against<mask> of their size, most will actually *gain* weight.  Most fat people have tried dieting,<mask> the vast majority of diets fail in the long term... and yo-yo dieting has some nasty physiological (not to mention psychological) side effects.  The effects of fat shaming can have other serious consequences, too - for instance, one thing that I frequently hear from fat people is that they're<mask> tired of being lectured by their doctors that they delay or avoid getting routine medical care.  </s>
Label encoding: <s>Your experience is not typical.  When people feel that they're being shamed or discriminated against because of their size, most will actually *gain* weight.  Most fat people have tried dieting, but the vast majority of diets fail in the long term... and yo-yo dieting has some nasty physiological (not to mention psychological) side effects.  The effects of fat shaming can have other serious consequences, too - for instance, one thing that I frequently hear from fat people is that they're so tired of being lectured by their doctors that they delay or avoid getting routine medical care.  </s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ok, I'll bite. Lets use Super Bowl<mask> an example<mask> you mentioned. The premise of this season is to get first place, no one else is rewarded,<mask> should your kid (even second place) get something? Again, I'm not trying to be cold<mask><mask><mask> children all deserve to be happy I'm just curious<mask> to<mask> someone should earn a trophy for 8th place? in the real world 4th place gets you laid off<mask> you didn't try hard enough. I guess I'm trying to get the mentality of a parent to see<mask> they do this for their kid? </s>
Label encoding: <s>ok, I'll bite. Lets use Super Bowl as an example as you mentioned. The premise of this season is to get first place, no one else is rewarded, why should your kid (even second place) get something? Again, I'm not trying to be cold as I think children all deserve to be happy I'm just curious as to why someone should earn a trophy for 8th place? in the real world 4th place gets you laid off because you didn't try hard enough. I guess I'm trying to get the mentality of a parent to see why they do this for their kid? </s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ASICs already make up the greater portion of the network. Most GPU miners have switched to mining Litecoin or similar. [NEWLINE] [NEWLINE] Current manufacturers include [Butterfly Labs \(with a poor delivery track-record\)]( [URL] /), [KNC Miner]( [URL] /), [Avalon]( [URL].com/), [BitMine]( [URL] /), [BitFury]( [URL] /), [TerraHash]( [URL] /) and [ASICMiner \(who only really release their products on Bitcointalk\)]( [URL].php?topic=282867.0).</s>
Label encoding: <s>ASICs already make up the greater portion of the network. Most GPU miners have switched to mining Litecoin or similar. [NEWLINE] [NEWLINE] Current manufacturers include [Butterfly Labs \(with a poor delivery track-record\)]( [URL] /), [KNC Miner]( [URL] /), [Avalon]( [URL].com/), [BitMine]( [URL] /), [BitFury]( [URL] /), [TerraHash]( [URL] /) and [ASICMiner \(who only really release their products on Bitcointalk\)]( [URL].php?topic=282867.0).</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, libertarian socialism, which right now is the more popular socialist ideology, is having a socialist economic system (everyone owns everything in eli5 terms) without a strong state.  Sometimes there is no state at all, in which case it is anarchism. [NEWLINE] [NEWLINE] Power is not centralized into one entity (the state)<mask> it is spread around the people or groups of people. (representatives usually) [NEWLINE] [NEWLINE] <mask> libertarian socialism and regular right libertarianism have in common is their view of the state.  Both sides see the state<mask> something that should be minimized or done away with completely.  </s>
Label encoding: <s>No, libertarian socialism, which right now is the more popular socialist ideology, is having a socialist economic system (everyone owns everything in eli5 terms) without a strong state.  Sometimes there is no state at all, in which case it is anarchism. [NEWLINE] [NEWLINE] Power is not centralized into one entity (the state) but it is spread around the people or groups of people. (representatives usually) [NEWLINE] [NEWLINE] What libertarian socialism and regular right libertarianism have in common is their view of the state.  Both sides see the state as something that should be minimized or done away with completely.  </s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How is volunteering incompatible with narcissism?  There are plenty of narcissistic [voluntourists]( [URL].aspx) helping out in all sorts of worthy endeavors.  A selfie picking up trash can be far smugger than a selfie dancing and celebrating. [NEWLINE] [NEWLINE] By all means, volunteer - good deeds are good.  And by all means, be humble and modest. <mask> I don't think that volunteering inherently creates modesty.  Finding a way to modestly show solidarity is a very tricky needle to thread.  It's certainly compatible with volunteering,<mask> I don't see the two<mask> linked.</s>
Label encoding: <s>How is volunteering incompatible with narcissism?  There are plenty of narcissistic [voluntourists]( [URL].aspx) helping out in all sorts of worthy endeavors.  A selfie picking up trash can be far smugger than a selfie dancing and celebrating. [NEWLINE] [NEWLINE] By all means, volunteer - good deeds are good.  And by all means, be humble and modest.  But I don't think that volunteering inherently creates modesty.  Finding a way to modestly show solidarity is a very tricky needle to thread.  It's certainly compatible with volunteering, but I don't see the two as linked.</s>
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0566, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Nope, I didn't say that. [NEWLINE] [NEWLINE] In a society<mask> different groups have markedly different characteristics, it would be quite useful. In a modern melting pot, there is very little that you can determine for race. [NEWLINE] [NEWLINE] Gender is a different story.<mask> I'm buying a little boy clothes, I'd probably buy something different than<mask> I were buying a little girl clothes. Men and women are socialized differently, and it would be nonsensical to claim that this doesn't effect their preferences. And I'm certainly not going to ask the little boy<mask> he would prefer shorts or a skirt.</s>
Label encoding: <s>Nope, I didn't say that. [NEWLINE] [NEWLINE] In a society where different groups have markedly different characteristics, it would be quite useful. In a modern melting pot, there is very little that you can determine for race. [NEWLINE] [NEWLINE] Gender is a different story. If I'm buying a little boy clothes, I'd probably buy something different than if I were buying a little girl clothes. Men and women are socialized differently, and it would be nonsensical to claim that this doesn't effect their preferences. And I'm certainly not going to ask the little boy if he would prefer shorts or a skirt.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's about superiority. They want their language to be the superior one, or maybe they just want to boast and take other people down. I'm not saying<mask> they do is super effective,<mask> it is an expression of their cultural association. [NEWLINE] [NEWLINE] Edit:   I guess I was challenging that grammar nazism is inherently stopping change,<mask> its development is a change in itself and reflects<mask> people currently view education and culture. <mask> there is a resurgence on conservative language, maybe it is<mask> there is a resurgence of people who associate with conservative language, which is still a change. </s>
Label encoding: <s>It's about superiority. They want their language to be the superior one, or maybe they just want to boast and take other people down. I'm not saying what they do is super effective, but it is an expression of their cultural association. [NEWLINE] [NEWLINE] Edit:   I guess I was challenging that grammar nazism is inherently stopping change, since its development is a change in itself and reflects how people currently view education and culture.  If there is a resurgence on conservative language, maybe it is because there is a resurgence of people who associate with conservative language, which is still a change. </s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0584, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But we do this all the time with other stuff. Just had a job interview? We believe we'll get the job<mask> that's<mask> we *hope* will happen. Relative diagnosed with cancer? We believe they'll be okay (or at least relatively pain-free)<mask> we *hope* they'll be okay. [NEWLINE] [NEWLINE] There is nothing wrong with choosing to believe that which you hope is true<mask> you can't reach a definitive answer. [NEWLINE] [NEWLINE] For you this realization may have been<mask> you lost your faith. For me it was the point<mask> I re-gained mine.</s>
Label encoding: <s>But we do this all the time with other stuff. Just had a job interview? We believe we'll get the job because that's what we *hope* will happen. Relative diagnosed with cancer? We believe they'll be okay (or at least relatively pain-free) because we *hope* they'll be okay. [NEWLINE] [NEWLINE] There is nothing wrong with choosing to believe that which you hope is true when you can't reach a definitive answer. [NEWLINE] [NEWLINE] For you this realization may have been where you lost your faith. For me it was the point where I re-gained mine.</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Direct democracy facilitated by or done over the Internet would weaken national security. By performing legally-binding elections over the Internet, we open legal system up to hackers of all sorts: from script kiddies to corporations to other countries' intelligence services. The technology to execute elections online just isn't there<mask>, and it won't come anytime soon. The requirements (both legal and technical) are incredibly strict and often contradictory; for instance,<mask> do we perform voter authentication online<mask> ensuring a secret ballot, all the<mask> allowing for post-election audits and defending against voter coercion and vote-buying?</s>
Label encoding: <s>Direct democracy facilitated by or done over the Internet would weaken national security. By performing legally-binding elections over the Internet, we open legal system up to hackers of all sorts: from script kiddies to corporations to other countries' intelligence services. The technology to execute elections online just isn't there yet, and it won't come anytime soon. The requirements (both legal and technical) are incredibly strict and often contradictory; for instance, how do we perform voter authentication online while ensuring a secret ballot, all the while allowing for post-election audits and defending against voter coercion and vote-buying?</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Irrational thinking doesn't make people insane. [NEWLINE] [NEWLINE] Insane is like, something literally went wrong in the brain at some point and now you have people who would be dysfunctional without or without religion in the world. [NEWLINE] [NEWLINE] It's true that religion can be a clinging point for the insane to rally to,<mask> religion itself is not indicitive of mental illness. Irrationality is actually a biologically evolved mechanism. It's the whole reason we can conceptualize things like gravity, black holes, quantum mechanics, particle physics, etc.<mask> irrationality is merely abstract thinking based on an incorrect premise.</s>
Label encoding: <s>Irrational thinking doesn't make people insane. [NEWLINE] [NEWLINE] Insane is like, something literally went wrong in the brain at some point and now you have people who would be dysfunctional without or without religion in the world. [NEWLINE] [NEWLINE] It's true that religion can be a clinging point for the insane to rally to, but religion itself is not indicitive of mental illness. Irrationality is actually a biologically evolved mechanism. It's the whole reason we can conceptualize things like gravity, black holes, quantum mechanics, particle physics, etc. Because irrationality is merely abstract thinking based on an incorrect premise.</s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>to answer your questions, even<mask> milk productions slows or stops, or teeth come in, or you want to transition or supplement, consultants can help you with these minor problems. having seen my wife come upon and surpass all of these challenges with ease, it's not<mask> difficult<mask> our culture has led you to believe. [NEWLINE] [NEWLINE] full disclosure: i am only *almost* certain of all of these things<mask> of the consults we've had. i wouldn't put it past my wife to keep feeding until she was more than a year old,<mask><mask> we've already started weening her.</s>
Label encoding: <s>to answer your questions, even if milk productions slows or stops, or teeth come in, or you want to transition or supplement, consultants can help you with these minor problems. having seen my wife come upon and surpass all of these challenges with ease, it's not as difficult as our culture has led you to believe. [NEWLINE] [NEWLINE] full disclosure: i am only *almost* certain of all of these things because of the consults we've had. i wouldn't put it past my wife to keep feeding until she was more than a year old, even though we've already started weening her.</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Obama calls it obamacare.  MSNBC calls it obamacare.  Everyone calls it obamacare supporters and opponents alike.  I<mask> don't substitute eight letter words that everyone knows the meaning of for a 22 letter series of words that most people won't know<mask> it means. [NEWLINE] [NEWLINE] It depends on<mask> you think the problem is. <mask> the problem is OP's weight being my business then Obamacare makes it worse. <mask> the problem is providing OP affordable healthcare then yes. <mask> this is a CMV about whether OP's weight is my business<mask><mask> the problem is the former.</s>
Label encoding: <s>Obama calls it obamacare.  MSNBC calls it obamacare.  Everyone calls it obamacare supporters and opponents alike.  I also don't substitute eight letter words that everyone knows the meaning of for a 22 letter series of words that most people won't know what it means. [NEWLINE] [NEWLINE] It depends on what you think the problem is.  If the problem is OP's weight being my business then Obamacare makes it worse.  If the problem is providing OP affordable healthcare then yes.  As this is a CMV about whether OP's weight is my business I think the problem is the former.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's not just the wealthy, it's everyone in the top 3/4 of the system. [NEWLINE] [NEWLINE] That's most people. [NEWLINE] [NEWLINE] Mom and pop store that was someones dream to open? Nope! Government's now bitch! [NEWLINE] [NEWLINE] House that your grandfather built? That's the governments now! [NEWLINE] [NEWLINE] Hoo lee shit, you bothered to save up any money<mask> a nest egg? TOUGH COOKIES IT'S THE GOVERNMENTS NOW! [NEWLINE] [NEWLINE] Seriously,<mask> something like that happened, it would just fuck over the middle class the most.</s>
Label encoding: <s>It's not just the wealthy, it's everyone in the top 3/4 of the system. [NEWLINE] [NEWLINE] That's most people. [NEWLINE] [NEWLINE] Mom and pop store that was someones dream to open? Nope! Government's now bitch! [NEWLINE] [NEWLINE] House that your grandfather built? That's the governments now! [NEWLINE] [NEWLINE] Hoo lee shit, you bothered to save up any money as a nest egg? TOUGH COOKIES IT'S THE GOVERNMENTS NOW! [NEWLINE] [NEWLINE] Seriously, if something like that happened, it would just fuck over the middle class the most.</s>
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>See this is<mask> you don't need to bring economics into it<mask>. This is a real argument that race matters<mask><mask> socioeconomic class. My point was just that a lot of people seem to be arguing purely or mostly based on economic disparity.<mask> most of those arguments you could switch around the race and the situation wouldn't have changed. The OP was clearly asking<mask> the benefits/detriments are to certain races specifically everything else being equal, and people are ignoring that question. You're actually answering the OP's real question here, unlike BlackSuperSonic apwas above. </s>
Label encoding: <s>See this is why you don't need to bring economics into it though. This is a real argument that race matters regardless of socioeconomic class. My point was just that a lot of people seem to be arguing purely or mostly based on economic disparity. But most of those arguments you could switch around the race and the situation wouldn't have changed. The OP was clearly asking what the benefits/detriments are to certain races specifically everything else being equal, and people are ignoring that question. You're actually answering the OP's real question here, unlike BlackSuperSonic apwas above. </s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0739, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0546, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think singling out sherpas ignores the ubiquitous nature of dangerous jobs in the third world. I'm not trying to bring up a straw man,<mask> paying others to do work that we first-worlders see<mask> too dangerous is very common in the developing world.<mask><mask>, I would say that the entire economy of the developing world depends on those sorts of jobs. Is the joy your smart phone gives you worth it<mask> you consider the hardships of those who made it? The (current) choice for many of these people is often either dangerous work or no work at all. </s>
Label encoding: <s>I think singling out sherpas ignores the ubiquitous nature of dangerous jobs in the third world. I'm not trying to bring up a straw man, but paying others to do work that we first-worlders see as too dangerous is very common in the developing world. In fact, I would say that the entire economy of the developing world depends on those sorts of jobs. Is the joy your smart phone gives you worth it when you consider the hardships of those who made it? The (current) choice for many of these people is often either dangerous work or no work at all. </s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You're right. That's<mask> it's important to celebrate the success of this sort of desalination technology. It's already here, and it's going to be widely adopted.<mask> you take a look, you'll see that desalination requires large amounts of energy. That's<mask> it's all linked up together. We need energy, huge amounts of it, to develop and innovate.<mask> it is happening. [ENDQ] [STARTQ] [ENDQ] [NEWLINE] That is wonderful news. I suppose this is the same for other resources? Either way, you did change my view. [NEWLINE] [NEWLINE] ∆</s>
Label encoding: <s> [STARTQ] You're right. That's why it's important to celebrate the success of this sort of desalination technology. It's already here, and it's going to be widely adopted. If you take a look, you'll see that desalination requires large amounts of energy. That's why it's all linked up together. We need energy, huge amounts of it, to develop and innovate. But it is happening. [ENDQ] [STARTQ] [ENDQ] [NEWLINE] That is wonderful news. I suppose this is the same for other resources? Either way, you did change my view. [NEWLINE] [NEWLINE] ∆</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0499, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I didn't say that they were. [ENDQ] [NEWLINE] You didn't say that most criminals are black? You sure?<mask> it looks like that is<mask> you said. [NEWLINE] [STARTQ] Most criminals are black, [ENDQ] [NEWLINE] Am I missing something? [NEWLINE] [NEWLINE] [STARTQ] <mask> we know someone is a criminal<mask> don't know<mask> race they are we should guess black. [ENDQ] [NEWLINE] No, we should guess they are **white**. Even<mask> a disproportionate percentage of black people are criminals, black people are still only 13% of the US population. There are far more white criminals than black criminals.</s>
Label encoding: <s> [STARTQ] I didn't say that they were. [ENDQ] [NEWLINE] You didn't say that most criminals are black? You sure? Because it looks like that is what you said. [NEWLINE] [STARTQ] Most criminals are black, [ENDQ] [NEWLINE] Am I missing something? [NEWLINE] [NEWLINE] [STARTQ] If we know someone is a criminal but don't know what race they are we should guess black. [ENDQ] [NEWLINE] No, we should guess they are **white**. Even if a disproportionate percentage of black people are criminals, black people are still only 13% of the US population. There are far more white criminals than black criminals.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's more of a legal fight than a financial one. Transitioning treatments are standard medical procedures advocated for by the AMA and the APA.<mask><mask>, they're usually very similar to procedures used for treating various cisgender medical conditions. It's just that currently insurance companies often exclude coverage of the exact same procedures<mask> used for transition reasons. [NEWLINE] [NEWLINE] <mask> trans folks really need legally is a bar on such insurance company discrimination. These are valid medical procedures and should be paid for by insurance. We don't need charity; we just need insurance companies to stop discriminating against us. </s>
Label encoding: <s>It's more of a legal fight than a financial one. Transitioning treatments are standard medical procedures advocated for by the AMA and the APA. In fact, they're usually very similar to procedures used for treating various cisgender medical conditions. It's just that currently insurance companies often exclude coverage of the exact same procedures when used for transition reasons. [NEWLINE] [NEWLINE] What trans folks really need legally is a bar on such insurance company discrimination. These are valid medical procedures and should be paid for by insurance. We don't need charity; we just need insurance companies to stop discriminating against us. </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How are you supposed to measure that? I mean, there's a big difference between "I'm more interesting in subject X than subject Y" which leads to me putting a great deal more weight on this rather than that and "I don't want to know about Y" which leads to discounting that topic completely. [NEWLINE] [NEWLINE] Frankly, people have different values and will not have the same reactions you do. This is a good thing. There are plenty of serious issue that you undervalue. By having a very wide variety of interests we cover everything, more or less.</s>
Label encoding: <s>How are you supposed to measure that? I mean, there's a big difference between "I'm more interesting in subject X than subject Y" which leads to me putting a great deal more weight on this rather than that and "I don't want to know about Y" which leads to discounting that topic completely. [NEWLINE] [NEWLINE] Frankly, people have different values and will not have the same reactions you do. This is a good thing. There are plenty of serious issue that you undervalue. By having a very wide variety of interests we cover everything, more or less.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Good for you, the problem is society is unwilling to properly dis-incentivize bad driving.<mask> of now<mask> you cause a fatal accident unwillingly you rarely go to jail. I've argued that drunk driving should be legal<mask><mask><mask> you hold people accountable for causing accidents (statistically very few people beyond the drunk driver and sometimes their passenger are killed in drunk crashes)<mask> for the same reason people destroy me on reddit.<mask> would you do<mask> a person choose to drive 150 mph and caused a fatal accident and they survived, vehicular manslaughter, 10 years???</s>
Label encoding: <s>Good for you, the problem is society is unwilling to properly dis-incentivize bad driving. As of now if you cause a fatal accident unwillingly you rarely go to jail. I've argued that drunk driving should be legal as long as you hold people accountable for causing accidents (statistically very few people beyond the drunk driver and sometimes their passenger are killed in drunk crashes) but for the same reason people destroy me on reddit. What would you do if a person choose to drive 150 mph and caused a fatal accident and they survived, vehicular manslaughter, 10 years???</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not going to change your opinion. Honestly, who cares? You're not hating on gay people, you're not actively opposing them, you're just not a fan of the culture that some of them choose to partake in. They just happen to be the ones people remember, every group of people have the ones that stick out. Hating Westboro doesn't mean you hate Christianity. Hating one one black person doesn't mean you hate all black people. I'm all for gay rights, I just don't think this is an opinion you should worry about changing.</s>
Label encoding: <s>I'm not going to change your opinion. Honestly, who cares? You're not hating on gay people, you're not actively opposing them, you're just not a fan of the culture that some of them choose to partake in. They just happen to be the ones people remember, every group of people have the ones that stick out. Hating Westboro doesn't mean you hate Christianity. Hating one one black person doesn't mean you hate all black people. I'm all for gay rights, I just don't think this is an opinion you should worry about changing.</s>
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To be honest, I didn't really want to concentrate on the minorities within this question. Head scratching about<mask> to classify certain people is somewhat irrelevant to the point I *really* wanted to make here. [NEWLINE] [NEWLINE] The point I wanted to make is that<mask> a friend does something well like play the piano, they shouldn't be congratulated for having talent,<mask> that assumes too much. Saying that they have talent assumes that they were born above everyone else and are especially cut out for this activity. Rather we should congratulate them for hard work, not<mask><mask> they are special.</s><pad>
Label encoding: <s>To be honest, I didn't really want to concentrate on the minorities within this question. Head scratching about how to classify certain people is somewhat irrelevant to the point I *really* wanted to make here. [NEWLINE] [NEWLINE] The point I wanted to make is that when a friend does something well like play the piano, they shouldn't be congratulated for having talent, since that assumes too much. Saying that they have talent assumes that they were born above everyone else and are especially cut out for this activity. Rather we should congratulate them for hard work, not assuming that they are special.</s><pad>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The goals of this society would be to improve our science and technology<mask> that our society could expand to off planet. [NEWLINE] [NEWLINE] Cells initially cooperated<mask> that enhanced their survival. Eventually this cooperation expanded to the point that the collective could sacrifice individual cells for the good of the collective. Human society has not<mask> developed<mask><mask><mask> organisms to recognize that the collective is more important than the individual. [NEWLINE] [NEWLINE] Many parents who refuse to abort a defective fetus<mask><mask> it will have an enjoyable life.<mask> the purpose of an individual's life is not enjoyment<mask> contribution to society.</s>
Label encoding: <s>The goals of this society would be to improve our science and technology so that our society could expand to off planet. [NEWLINE] [NEWLINE] Cells initially cooperated because that enhanced their survival. Eventually this cooperation expanded to the point that the collective could sacrifice individual cells for the good of the collective. Human society has not yet developed as far as organisms to recognize that the collective is more important than the individual. [NEWLINE] [NEWLINE] Many parents who refuse to abort a defective fetus argue that it will have an enjoyable life. But the purpose of an individual's life is not enjoyment but contribution to society.</s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL] [NEWLINE] [NEWLINE] Disenfranchising certain groups on certain issues is not the way to go. Should only immigrants be allowed to vote on legislation regarding immigration control? After all, native born citizens never had to go through that.<mask><mask>, you can see most issues in a democracy has to do with  minority group of some sort, weather it be the poor, immigrants, racial and ethnic minorities(affirmative action), disproportionate power held by the richest few, the list goes on. Should only those who are immediately affected by certain legislation be allowed to vote om them?</s>
Label encoding: <s> [URL] [NEWLINE] [NEWLINE] Disenfranchising certain groups on certain issues is not the way to go. Should only immigrants be allowed to vote on legislation regarding immigration control? After all, native born citizens never had to go through that. In fact, you can see most issues in a democracy has to do with  minority group of some sort, weather it be the poor, immigrants, racial and ethnic minorities(affirmative action), disproportionate power held by the richest few, the list goes on. Should only those who are immediately affected by certain legislation be allowed to vote om them?</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Treason" in the US is a dated term used in the Constitution to describe<mask> is effectively an act of war or aid to the enemy. It's a general term.<mask><mask><mask><mask> you can convict someone of "treason" in the United States anymore. [NEWLINE] [NEWLINE] In any case, I may be wrong about its implementation,<mask> in any case the use of "treason" in the US is for acts that jeopardize the security of the government and the country, not simply exercise of speech against it (<mask> it was known in other governments).</s>
Label encoding: <s>"Treason" in the US is a dated term used in the Constitution to describe what is effectively an act of war or aid to the enemy. It's a general term. I do not think you can convict someone of "treason" in the United States anymore. [NEWLINE] [NEWLINE] In any case, I may be wrong about its implementation, but in any case the use of "treason" in the US is for acts that jeopardize the security of the government and the country, not simply exercise of speech against it ( as it was known in other governments).</s>
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0510, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Obviously I'm biased,<mask> are you. Premier is a better pure producer,<mask> Kanye is a top-tier producer<mask> well<mask> a platinum selling rapper.<mask><mask><mask> I know Diddy is the only other person who can say that and Kanye is definitely better on the rap front. [ENDQ] [NEWLINE] [NEWLINE] Of the top of my head, DOOM comes to mind<mask> a top tier producer and a contender for GOAT flows.<mask> has Kanye done to match the artistry of Special Herbs? Or the insane grimy creativity of Mm.. FOOD?</s><pad>
Label encoding: <s> [STARTQ] Obviously I'm biased, as are you. Premier is a better pure producer, but Kanye is a top-tier producer as well as a platinum selling rapper. As far as I know Diddy is the only other person who can say that and Kanye is definitely better on the rap front. [ENDQ] [NEWLINE] [NEWLINE] Of the top of my head, DOOM comes to mind as a top tier producer and a contender for GOAT flows. What has Kanye done to match the artistry of Special Herbs? Or the insane grimy creativity of Mm.. FOOD?</s><pad>
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I take sentient means<mask> it means colloquially : responsive to or conscious of sense impressions. [NEWLINE] [NEWLINE] <mask> I weren't a person it wouldn't matter to me<mask> I wouldn't be able to know<mask>'s going on!<mask><mask> you want to dig Socratically I'll bite and admit that my value for persons is grounded in consciousness and at the very least every creature capable (either at the moment or one that has the potential for consciousness ie a human fetus after the first trimester)  of human level consciousness. [NEWLINE] |edit:a letter|</s>
Label encoding: <s>I take sentient means what it means colloquially : responsive to or conscious of sense impressions. [NEWLINE] [NEWLINE] If I weren't a person it wouldn't matter to me because I wouldn't be able to know what's going on! But since you want to dig Socratically I'll bite and admit that my value for persons is grounded in consciousness and at the very least every creature capable (either at the moment or one that has the potential for consciousness ie a human fetus after the first trimester)  of human level consciousness. [NEWLINE] |edit:a letter|</s>
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your right. 9 months of being inconvenienced in a way your biology was designed from the ground up to handle is not the same<mask> being a slave for 18 years. And emotional trauma? You are being extremely short sighted<mask> you think the mother's emotional trauma automatically trumps the emotional trauma of lost agency and being forced against your will.<mask> hay, there is a reason men commit suicide<mask> much more then women. [Just ask Derrick Miller]( [URL] ) (and fyi, that a feminist source,<mask> no point looking for a pro male bias).</s>
Label encoding: <s>Your right. 9 months of being inconvenienced in a way your biology was designed from the ground up to handle is not the same as being a slave for 18 years. And emotional trauma? You are being extremely short sighted if you think the mother's emotional trauma automatically trumps the emotional trauma of lost agency and being forced against your will. But hay, there is a reason men commit suicide so much more then women. [Just ask Derrick Miller]( [URL] ) (and fyi, that a feminist source, so no point looking for a pro male bias).</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] That just makes home ownership a very risky form of saving [ENDQ] [NEWLINE] Historically, home prices are very stable.  One of the reasons no one saw the widespread housing downturn coming is that nationwide housing price reductions in the US were almost unheard of. [NEWLINE] [NEWLINE] You<mask> don't take money out of the house for hip replacements.  That is<mask> your savings is for.  You live in the home until it isn't feasible, then you trade down to a cheaper home with lower maintenance costs or use the proceeds from the sale to finance a rental unit.</s>
Label encoding: <s> [STARTQ] That just makes home ownership a very risky form of saving [ENDQ] [NEWLINE] Historically, home prices are very stable.  One of the reasons no one saw the widespread housing downturn coming is that nationwide housing price reductions in the US were almost unheard of. [NEWLINE] [NEWLINE] You also don't take money out of the house for hip replacements.  That is what your savings is for.  You live in the home until it isn't feasible, then you trade down to a cheaper home with lower maintenance costs or use the proceeds from the sale to finance a rental unit.</s>
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] ighting is outlawed in the European leagues,<mask> no one outside Don Cherry would<mask><mask> it's<mask> European players lack the toughness to fight. [ENDQ] [NEWLINE] name one European goon in the NHL. [NEWLINE] [NEWLINE] I've never once heard that it creates a "sluggish" game. I have,<mask>, heard tons of people saying its<mask> women can't handle the physical aspect of the game. [NEWLINE] [NEWLINE] Different players grow up in different cultures, its<mask> the Russians were all finesse and never checked and<mask> we saw such a clash<mask> Canada played them</s>
Label encoding: <s> [STARTQ] ighting is outlawed in the European leagues, but no one outside Don Cherry would argue that it's because European players lack the toughness to fight. [ENDQ] [NEWLINE] name one European goon in the NHL. [NEWLINE] [NEWLINE] I've never once heard that it creates a "sluggish" game. I have, however, heard tons of people saying its because women can't handle the physical aspect of the game. [NEWLINE] [NEWLINE] Different players grow up in different cultures, its why the Russians were all finesse and never checked and why we saw such a clash when Canada played them</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks, I never thought of the 3rd point before actually. Thinking about it a bit more I feel I'm less worried about people calling me out on saying stupid things and more concerned about getting a non-response. [NEWLINE] [NEWLINE] Sometimes two people will be having a conversation that I feel<mask><mask> I can't jump into for whatever reason.<mask> I interject with a statement they might give a one word response<mask><mask> they acknowledge<mask> I said out of courtesy,<mask> then turn back to each other and continue their conversation<mask><mask> I was interrupting. </s>
Label encoding: <s>Thanks, I never thought of the 3rd point before actually. Thinking about it a bit more I feel I'm less worried about people calling me out on saying stupid things and more concerned about getting a non-response. [NEWLINE] [NEWLINE] Sometimes two people will be having a conversation that I feel as though I can't jump into for whatever reason. If I interject with a statement they might give a one word response as if they acknowledge what I said out of courtesy, but then turn back to each other and continue their conversation as if I was interrupting. </s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1.<mask> you call laziness is actually just enhanced productivity.<mask> you read your sentence with a different tone it's a positive. "Now we don't even need to jot things down or even type them into our smart phones at the very least[!]" [NEWLINE] [NEWLINE] 2. We already rely on technology 100%. Our society would collapse without it. [NEWLINE] [NEWLINE] 3. You already have an Amazon account, this doesn't invade your privacy any more than shopping on amazon. Making purchases isn't a private event, it involves a second party. </s>
Label encoding: <s>1. What you call laziness is actually just enhanced productivity. If you read your sentence with a different tone it's a positive. "Now we don't even need to jot things down or even type them into our smart phones at the very least[!]" [NEWLINE] [NEWLINE] 2. We already rely on technology 100%. Our society would collapse without it. [NEWLINE] [NEWLINE] 3. You already have an Amazon account, this doesn't invade your privacy any more than shopping on amazon. Making purchases isn't a private event, it involves a second party. </s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can't see<mask> the road widens just before the intersection? [NEWLINE] [NEWLINE] <mask> that's a very generous 2 lane road. Many in my area are<mask> narrow that with a car parallel parked along them, two normal sized cars can't pass one another (one has to tuck in and wait for the other to pass)...basically a lane and a half. [NEWLINE] [NEWLINE] Usually in the intersections, it's a tricky enough maneuver just to negotiate the turn in anything larger than a small SUV...<mask> you had to maneuver beyond that, forget it.</s>
Label encoding: <s>You can't see how the road widens just before the intersection? [NEWLINE] [NEWLINE] Also that's a very generous 2 lane road. Many in my area are so narrow that with a car parallel parked along them, two normal sized cars can't pass one another (one has to tuck in and wait for the other to pass)...basically a lane and a half. [NEWLINE] [NEWLINE] Usually in the intersections, it's a tricky enough maneuver just to negotiate the turn in anything larger than a small SUV... if you had to maneuver beyond that, forget it.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not refuting that<mask> you say is 100% true for you.<mask> I know *plenty* of people for whom Christmas is racked with pain and suffering. Loss and issues and departures from ideal outcomes are felt most strongly around that period. There's an aching dysphoria for a huge segment of the population through the christmas season<mask> they confront that their lives don't live up to the Hallmark card version. I don't know<mask> old you are,<mask> give it a few years, and you'll see<mask> I mean.</s><pad>
Label encoding: <s>I'm not refuting that what you say is 100% true for you. But I know *plenty* of people for whom Christmas is racked with pain and suffering. Loss and issues and departures from ideal outcomes are felt most strongly around that period. There's an aching dysphoria for a huge segment of the population through the christmas season when they confront that their lives don't live up to the Hallmark card version. I don't know how old you are, but give it a few years, and you'll see what I mean.</s><pad>
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No, I'm not. [NEWLINE] [NEWLINE] [STARTQ] In 2009, the average income of the top 1% was $960,000 with a minimum income of $343,927. In 2007, the richest 1% of the American population owned 34.6% of the country's total wealth [ENDQ] [NEWLINE] [URL] %25#Economic_context [NEWLINE] [NEWLINE] Doctors, lawyers, and small business owners do not (generally) make $350k a year.  They *absolutely* don't make enough to bring the average up to $960k a year.</s>
Label encoding: <s>No, I'm not. [NEWLINE] [NEWLINE] [STARTQ] In 2009, the average income of the top 1% was $960,000 with a minimum income of $343,927. In 2007, the richest 1% of the American population owned 34.6% of the country's total wealth [ENDQ] [NEWLINE] [URL] %25#Economic_context [NEWLINE] [NEWLINE] Doctors, lawyers, and small business owners do not (generally) make $350k a year.  They *absolutely* don't make enough to bring the average up to $960k a year.</s>
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask> you're proposing then is that pedophile be renamed to Gold Star Pedophile,<mask> convicted child molesters are just a normal pedophile? That still runs into the problem of demonizing pedophiles, does it not? [NEWLINE] [NEWLINE] I get<mask> you're saying, that society's attitudes towards pedophiles need to change.<mask> awarding a gold star to a pedophile for doing<mask> he's supposed to do doesn't sound helpful. Should we recognize and award a gold star to blacks who get a job or take care of their kids? </s>
Label encoding: <s>So what you're proposing then is that pedophile be renamed to Gold Star Pedophile, while convicted child molesters are just a normal pedophile? That still runs into the problem of demonizing pedophiles, does it not? [NEWLINE] [NEWLINE] I get what you're saying, that society's attitudes towards pedophiles need to change. But awarding a gold star to a pedophile for doing what he's supposed to do doesn't sound helpful. Should we recognize and award a gold star to blacks who get a job or take care of their kids? </s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So, they don't get to choose whether to give it for free,<mask> you do? [NEWLINE] [NEWLINE] Who said "huge profit". <mask> about any profit? <mask> you are an author or a studio-type musician you won't be able to do it full time<mask> you don't have a way to make money off of it. [NEWLINE] [NEWLINE] You're right, there will not be "no" incentive,<mask> there's going to be a lot more artists working office jobs to pay the rent<mask> you refuse to pay for<mask> you consume.</s>
Label encoding: <s>So, they don't get to choose whether to give it for free, but you do? [NEWLINE] [NEWLINE] Who said "huge profit".  How about any profit?  If you are an author or a studio-type musician you won't be able to do it full time if you don't have a way to make money off of it. [NEWLINE] [NEWLINE] You're right, there will not be "no" incentive, but there's going to be a lot more artists working office jobs to pay the rent because you refuse to pay for what you consume.</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a chance someone could recover from emotional or physical trauma. They have a chance at happiness, a chance to make a new life, and a chance to get over whatever it was that happened to them. Even<mask> they don't get over it entirely, they still have a pretty good chance of being happy, even<mask> it's only for a few moments a day. [NEWLINE] [NEWLINE] Being murdered<mask><mask><mask><mask> is permanent. You have no shot at ever being happy or having a good life or recovering. You're dead. That's it.</s>
Label encoding: <s>There's a chance someone could recover from emotional or physical trauma. They have a chance at happiness, a chance to make a new life, and a chance to get over whatever it was that happened to them. Even if they don't get over it entirely, they still have a pretty good chance of being happy, even if it's only for a few moments a day. [NEWLINE] [NEWLINE] Being murdered on the other hand is permanent. You have no shot at ever being happy or having a good life or recovering. You're dead. That's it.</s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0554, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is no "magic number" of scholarships that a college football team needs.  It is simply a number set by NCAA rule. <mask> the NCAA said that from now on the max number of football scholarships a school could have would be 60 do you think that schools would stop playing football? <mask><mask> the number were 20 or 50?  Or,<mask><mask> the scholarships could be divided up between players<mask> in most sports?  The Ivy league does not have athletic scholarships at all.  Scholarships are based on need and academics. [NEWLINE] </s>
Label encoding: <s>There is no "magic number" of scholarships that a college football team needs.  It is simply a number set by NCAA rule.  If the NCAA said that from now on the max number of football scholarships a school could have would be 60 do you think that schools would stop playing football?  What if the number were 20 or 50?  Or, what if the scholarships could be divided up between players as in most sports?  The Ivy league does not have athletic scholarships at all.  Scholarships are based on need and academics. [NEWLINE] </s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are some cases<mask> it makes a lot of sense.  Take Viagara for example.  Lots of people were suffering from ED which they either thought was normal (ie not a medical condition which could be treated with medicine), or were embarrassed about.  Either way they would never bring it up with their doctor without Viagara having been heavily advertised.  You could make similar cases for things like allergies, fibromayalgia, acid reflux etc. <mask> those are allowed, it's tough to draw the line with<mask> is not.</s>
Label encoding: <s>There are some cases where it makes a lot of sense.  Take Viagara for example.  Lots of people were suffering from ED which they either thought was normal (ie not a medical condition which could be treated with medicine), or were embarrassed about.  Either way they would never bring it up with their doctor without Viagara having been heavily advertised.  You could make similar cases for things like allergies, fibromayalgia, acid reflux etc.  If those are allowed, it's tough to draw the line with what is not.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] without at least a fair amount of popular support [ENDQ] [NEWLINE] The *Final Solution* was never publicly acknowledged by the Nazi Government during the war; German Citizens Nazi or not virtually all had no idea<mask> was taking place, they<mask><mask> genuinely shocked and disgusted<mask> the rest of the world. Now the public removal of Jewish people was no secrete to the German people,<mask> the systematic genocide was kept hush hush; there is a reason<mask> only SS<mask> allowed to mange the camps, and they<mask> constructed away from populated areas.</s>
Label encoding: <s> [STARTQ] without at least a fair amount of popular support [ENDQ] [NEWLINE] The *Final Solution* was never publicly acknowledged by the Nazi Government during the war; German Citizens Nazi or not virtually all had no idea what was taking place, they where as genuinely shocked and disgusted as the rest of the world. Now the public removal of Jewish people was no secrete to the German people, but the systematic genocide was kept hush hush; there is a reason why only SS where allowed to mange the camps, and they where constructed away from populated areas.</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Two major points: [NEWLINE] [NEWLINE] First,<mask><mask> you misunderstand the goal of a government "giving away land for free".  The goal is not to get rid of it, the goal is to encourage population growth, expansion, and development. [NEWLINE] [NEWLINE] Second, the federal governemnt doesn't collect property taxes.  Property taxes are usually a state or county/town imposed tax.  These taxes pay for things like roads, schools, fire and police departments,<mask> well<mask> other emergency services, which you likely use living in the area.</s>
Label encoding: <s>Two major points: [NEWLINE] [NEWLINE] First, I think you misunderstand the goal of a government "giving away land for free".  The goal is not to get rid of it, the goal is to encourage population growth, expansion, and development. [NEWLINE] [NEWLINE] Second, the federal governemnt doesn't collect property taxes.  Property taxes are usually a state or county/town imposed tax.  These taxes pay for things like roads, schools, fire and police departments, as well as other emergency services, which you likely use living in the area.</s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some people want to make a change in the way we live, and it drives them. I'm content with just avoiding things I don't like,<mask> some people want to have complete power over those things. Being the leader would give it to them, at least to some degree. To be the top dog over a couple hundred million people has got to be exhilarating in its own way, and<mask> the stress associated is probably quite large the ability you now have over the way people are allowed to live is probably enough to balance that out...</s>
Label encoding: <s>Some people want to make a change in the way we live, and it drives them. I'm content with just avoiding things I don't like, but some people want to have complete power over those things. Being the leader would give it to them, at least to some degree. To be the top dog over a couple hundred million people has got to be exhilarating in its own way, and though the stress associated is probably quite large the ability you now have over the way people are allowed to live is probably enough to balance that out...</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] ok well that is the whole basis of democracy<mask> maybe that should be your CMV, seems like you really buried the lead on this one. [ENDQ] [NEWLINE] <mask> I sell my vote to the highest bidder, do you think my vote should count? I'd be surprised<mask> anyone does. There's definitely a line that divides votes that should count and those that shouldn't. I believe uninformed votes are no better than purchased votes. [NEWLINE] [NEWLINE] My sympathies to the woman in the article,<mask> she is a outlier of the system.</s>
Label encoding: <s> [STARTQ] ok well that is the whole basis of democracy so maybe that should be your CMV, seems like you really buried the lead on this one. [ENDQ] [NEWLINE] If I sell my vote to the highest bidder, do you think my vote should count? I'd be surprised if anyone does. There's definitely a line that divides votes that should count and those that shouldn't. I believe uninformed votes are no better than purchased votes. [NEWLINE] [NEWLINE] My sympathies to the woman in the article, but she is a outlier of the system.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry probably wasn't necessary for me to put organic in there<mask> I'd say it's seen<mask> a general rule that organic farms have to be free range to be classed<mask> organic. Organic food wasn't made specifically for welfare reasons,<mask> you can't have organic factory farms, it just<mask> happens that the criteria that make it sustainable<mask> fit the criteria of free range farms( or at least ones that aren't intensive farms). That might not be true of all cases and<mask> you find anything to the contrary please do post it.</s>
Label encoding: <s>Sorry probably wasn't necessary for me to put organic in there but I'd say it's seen as a general rule that organic farms have to be free range to be classed as organic. Organic food wasn't made specifically for welfare reasons, but you can't have organic factory farms, it just so happens that the criteria that make it sustainable also fit the criteria of free range farms( or at least ones that aren't intensive farms). That might not be true of all cases and if you find anything to the contrary please do post it.</s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>True. I can see<mask> you read it that way. I said it in a facetious way<mask> I was assuming you were going into a'society imposed things are all bad' route, and i get kinda tired of those. And you know<mask> they say about assumptions! My apologies. [NEWLINE] [NEWLINE] <mask> of course, I would rather a happy person do things to add value to themselves (Even arbitrary value) rather than do those things mask massive issues of self-esteem. And I generally credit most people with doing the former.</s><pad>
Label encoding: <s>True. I can see how you read it that way. I said it in a facetious way because I was assuming you were going into a'society imposed things are all bad' route, and i get kinda tired of those. And you know what they say about assumptions! My apologies. [NEWLINE] [NEWLINE] But of course, I would rather a happy person do things to add value to themselves (Even arbitrary value) rather than do those things mask massive issues of self-esteem. And I generally credit most people with doing the former.</s><pad>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0619, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In my experience urban and suburban people are far more likely to have strong (sometimes mystical) feelings about nature and conservation.  People who live and work with nature type things on a day to day basis tend to see them in much more utilitarian ways. [NEWLINE] [NEWLINE] I am not sure which social values you are interested in exactly.  Rural areas do tend to be more socially conservative simply<mask> they tend to be less socially diverse. <mask>, rural areas that regularly cycle their population from a diverse pool of residents can be quite socially dynamic.</s>
Label encoding: <s>In my experience urban and suburban people are far more likely to have strong (sometimes mystical) feelings about nature and conservation.  People who live and work with nature type things on a day to day basis tend to see them in much more utilitarian ways. [NEWLINE] [NEWLINE] I am not sure which social values you are interested in exactly.  Rural areas do tend to be more socially conservative simply because they tend to be less socially diverse.  However, rural areas that regularly cycle their population from a diverse pool of residents can be quite socially dynamic.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you 100%.<mask> I wonder<mask> the OP really means the "concept" of good grammar rather than our tolerance for accepting it. For example,<mask> we all agree that "ain't" _is_ a word, then common use of "ain't" isn't bad grammar. [NEWLINE] [NEWLINE] And of course there's the whole idea of context switching.<mask><mask> any context were accepted anywhere? [NEWLINE] [NEWLINE] I still still cringe at grammar that's not mine.<mask> I wonder<mask> the OP has a good point? [NEWLINE] </s>
Label encoding: <s>I agree with you 100%. However I wonder if the OP really means the "concept" of good grammar rather than our tolerance for accepting it. For example, if we all agree that "ain't" _is_ a word, then common use of "ain't" isn't bad grammar. [NEWLINE] [NEWLINE] And of course there's the whole idea of context switching. What if any context were accepted anywhere? [NEWLINE] [NEWLINE] I still still cringe at grammar that's not mine. But I wonder if the OP has a good point? [NEWLINE] </s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When you're dead you won't have the ability to have a sore jaw. You won't feel sad or hurt or pain or pleasure. You won't feel anything. You won't exist. There is no you anymore.<mask> it isn't like someone in a candy store. It's like no one in a candy store experiencing nothing, having no regret for<mask> sore their jaw was<mask> they were alive and no memory of eating candy in the first place. [NEWLINE] [NEWLINE] edit: can't believe I typed your instead of you're...</s>
Label encoding: <s>When you're dead you won't have the ability to have a sore jaw. You won't feel sad or hurt or pain or pleasure. You won't feel anything. You won't exist. There is no you anymore. So it isn't like someone in a candy store. It's like no one in a candy store experiencing nothing, having no regret for how sore their jaw was while they were alive and no memory of eating candy in the first place. [NEWLINE] [NEWLINE] edit: can't believe I typed your instead of you're...</s>
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Vibration is a cool feature,<mask> is it enough? In my mind, a vibration in the controller doesn't really translate into an action on the screen. It's a little bit of haptic feedback (the way my phone vibrates a little<mask> I touch certain objects),<mask> for things like explosions? I feel like that would be more immersion-breaking than enhancing. Something huge blows up and all I get is a little wrist-wiggle? I don't know<mask> that would really help my enjoyment of the game.</s>
Label encoding: <s>Vibration is a cool feature, but is it enough? In my mind, a vibration in the controller doesn't really translate into an action on the screen. It's a little bit of haptic feedback (the way my phone vibrates a little when I touch certain objects), but for things like explosions? I feel like that would be more immersion-breaking than enhancing. Something huge blows up and all I get is a little wrist-wiggle? I don't know if that would really help my enjoyment of the game.</s>
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>In cases<mask> the father actually seeks custody, he has a very good chance of getting at least a joint arrangement.  These days,<mask><mask> the biggest concern for judges deciding primary custody is to keep the kids in their current school district to minimize the disruption to their lives (or at least that was definitely the case during my uncle's divorce, in which he was awarded primary custody<mask> he chose to live in the same area<mask> his ex-wife chose to move in with her boyfriend on the other side of town).  </s>
Label encoding: <s>In cases where the father actually seeks custody, he has a very good chance of getting at least a joint arrangement.  These days, I think the biggest concern for judges deciding primary custody is to keep the kids in their current school district to minimize the disruption to their lives (or at least that was definitely the case during my uncle's divorce, in which he was awarded primary custody because he chose to live in the same area while his ex-wife chose to move in with her boyfriend on the other side of town).  </s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [NEWLINE] [STARTQ] Women are only not financially responsible<mask> there is no named father.<mask> the father isn't named, he's not on the hook for child support either,<mask> it doesn't matter. [ENDQ] [NEWLINE] <mask> it does matter. [NEWLINE] [NEWLINE] Due to biology, and biased laws, a woman may do this to avoid parental obligations. She may deliberately not name the father, even<mask> she knows who it is. [NEWLINE] [NEWLINE] A man cannot do this. [NEWLINE] [NEWLINE] Which in practice leads to this: [NEWLINE] [NEWLINE] [URL].csp [NEWLINE] </s>
Label encoding: <s> [NEWLINE] [STARTQ] Women are only not financially responsible if there is no named father. If the father isn't named, he's not on the hook for child support either, so it doesn't matter. [ENDQ] [NEWLINE] But it does matter. [NEWLINE] [NEWLINE] Due to biology, and biased laws, a woman may do this to avoid parental obligations. She may deliberately not name the father, even if she knows who it is. [NEWLINE] [NEWLINE] A man cannot do this. [NEWLINE] [NEWLINE] Which in practice leads to this: [NEWLINE] [NEWLINE] [URL].csp [NEWLINE] </s>
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, for starters<mask> you have enough money you can and probably do hire people to scope out the best people to donate to for whatever desires you have.  Beyond that I specifically stated that I valued sacrifices of time and effort above money. <mask>, my OP wasn't "Millionaires and Billionaires **can't** be generous" just that those who donate<mask> little that they wouldn't notice<mask> their accountant hadn't told them shouldn't be called generous.  Re that point: see my edit to the OP.</s>
Label encoding: <s>Well, for starters if you have enough money you can and probably do hire people to scope out the best people to donate to for whatever desires you have.  Beyond that I specifically stated that I valued sacrifices of time and effort above money.  Lastly, my OP wasn't "Millionaires and Billionaires **can't** be generous" just that those who donate so little that they wouldn't notice if their accountant hadn't told them shouldn't be called generous.  Re that point: see my edit to the OP.</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a scientist, I can't understate the fact that this is such a pervasive and common semantics problem for science and the public view. It's not your fault at all. Scientists try to speak very specifically and accurately, then the rest of the world runs around saying, "The Theory of Evolution" is "just a theory" and<mask> forth. Doubters latch on  to the word 'theory' to discount the certainty of it, not aware that the word means something different to the people who coined it.</s>
Label encoding: <s>As a scientist, I can't understate the fact that this is such a pervasive and common semantics problem for science and the public view. It's not your fault at all. Scientists try to speak very specifically and accurately, then the rest of the world runs around saying, "The Theory of Evolution" is "just a theory" and so forth. Doubters latch on  to the word 'theory' to discount the certainty of it, not aware that the word means something different to the people who coined it.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Feminism doesn't equal the abolition of feeling feminine.  Feminism is,<mask><mask> to many things,<mask> women can be comfortable being<mask> they want, and that includes being and feeling feminine.  To be honest, it isn't very pro-feminism of you to be<mask> judgemental about a woman wanting to wear dresses, corsets, and get all "trussed" up.  "Feeling feminine" is subjective, and can be whatever makes a woman feel more connected to herself.  </s>
Label encoding: <s>Feminism doesn't equal the abolition of feeling feminine.  Feminism is, in addition to many things, so women can be comfortable being however they want, and that includes being and feeling feminine.  To be honest, it isn't very pro-feminism of you to be so judgemental about a woman wanting to wear dresses, corsets, and get all "trussed" up.  "Feeling feminine" is subjective, and can be whatever makes a woman feel more connected to herself.  </s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A bit off topic,<mask> some back of the envelope estimates tell me that radioactive waste becomes<mask> radioactive<mask> naturally found uranium ore in 70-400 years. After that point, you can literally dump it back into the mine (assuming solid waste), and the place will already be less dangerous than it was before.<mask> you shield it at least a bit and seal off the place, it is now pretty much 100% safe. [NEWLINE] [NEWLINE] <mask> you just need to store dangerous waste for a few centuries max and you're set. </s>
Label encoding: <s>A bit off topic, but some back of the envelope estimates tell me that radioactive waste becomes as radioactive as naturally found uranium ore in 70-400 years. After that point, you can literally dump it back into the mine (assuming solid waste), and the place will already be less dangerous than it was before. If you shield it at least a bit and seal off the place, it is now pretty much 100% safe. [NEWLINE] [NEWLINE] So you just need to store dangerous waste for a few centuries max and you're set. </s>
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To your point about humanity being created from nothing being proven false. To me this is one of the biggest reasons for believing in a supernatural being.<mask> one believes that the entire universe was created by a Big Bang of subatomic particles the logical next question is<mask> did the subatomic particles come from? The only thing that makes<mask> is that something was created from nothing. To me it is more reasonable to believe that a supernatural being that has always existed created the universe than to believe the same thing about subatomic particles. </s>
Label encoding: <s>To your point about humanity being created from nothing being proven false. To me this is one of the biggest reasons for believing in a supernatural being. If one believes that the entire universe was created by a Big Bang of subatomic particles the logical next question is where did the subatomic particles come from? The only thing that makes since is that something was created from nothing. To me it is more reasonable to believe that a supernatural being that has always existed created the universe than to believe the same thing about subatomic particles. </s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>∆ Thanks, this helps. This post and another is helping me see<mask>'s going on here. [NEWLINE] [NEWLINE] I worked in a "family" restaurant for most of my serving career,<mask> my expectations of other parents/kids behavior was that of an "upscale" restaurant<mask> of my own experience growing up. [NEWLINE] [NEWLINE] Now I'm starting to wonder<mask> none of this is explained to new employees that are hired. "This is a family restaurant, are you comfortable with children..." for example would have been nice.</s>
Label encoding: <s>∆ Thanks, this helps. This post and another is helping me see what's going on here. [NEWLINE] [NEWLINE] I worked in a "family" restaurant for most of my serving career, but my expectations of other parents/kids behavior was that of an "upscale" restaurant because of my own experience growing up. [NEWLINE] [NEWLINE] Now I'm starting to wonder why none of this is explained to new employees that are hired. "This is a family restaurant, are you comfortable with children..." for example would have been nice.</s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is merely a factor, and<mask><mask> the analogy was too simple. It isn't a binary evaluation of the individual. It falls on the gradient of appeal/attraction. I'm saying this counting against them isn't unreasonable for an individuals preference. Yes,<mask> you develop more about the individual a better picture comes up.<mask>,<mask> the only distinguishing difference is sexual history, I have no issue with anyone preferring to make the decision based on that. They can be for a promiscuous woman or against. </s>
Label encoding: <s>It is merely a factor, and I think the analogy was too simple. It isn't a binary evaluation of the individual. It falls on the gradient of appeal/attraction. I'm saying this counting against them isn't unreasonable for an individuals preference. Yes, as you develop more about the individual a better picture comes up. But, if the only distinguishing difference is sexual history, I have no issue with anyone preferring to make the decision based on that. They can be for a promiscuous woman or against. </s>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why do we explore space? [NEWLINE] [NEWLINE] <mask> we look at the question from a purely objective, there is no reason. After all, there is no practical use for knowing the properties of galaxies thousands of ligtyears away.<mask> we pour millions of dollars into space programs. [NEWLINE] [NEWLINE] We don't explore space for practical purposes, we do it to satisfy our curiosity. [NEWLINE] [NEWLINE] It's the same with philosophy. We don't do it for any practical reason - we do it to satisfy our curiosity.   </s><pad>
Label encoding: <s>Why do we explore space? [NEWLINE] [NEWLINE] If we look at the question from a purely objective, there is no reason. After all, there is no practical use for knowing the properties of galaxies thousands of ligtyears away. yet we pour millions of dollars into space programs. [NEWLINE] [NEWLINE] We don't explore space for practical purposes, we do it to satisfy our curiosity. [NEWLINE] [NEWLINE] It's the same with philosophy. We don't do it for any practical reason - we do it to satisfy our curiosity.   </s><pad>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1) That there is not a lot of variance is one of the main reasons I brought up this question. Ok, lets automate cooking, and leave cashiers to facilitate customer interaction and provide a service. [NEWLINE] [NEWLINE] 2) This is completely separate topic for debate, which is much bigger. Some people totally deny this idea that job creation should be responsibility of a business.<mask><mask> automation will sooner or later kill great deal of unskilled labor,<mask> we better start thinking now.<mask> again this is another discussion. </s>
Label encoding: <s>1) That there is not a lot of variance is one of the main reasons I brought up this question. Ok, lets automate cooking, and leave cashiers to facilitate customer interaction and provide a service. [NEWLINE] [NEWLINE] 2) This is completely separate topic for debate, which is much bigger. Some people totally deny this idea that job creation should be responsibility of a business. I think automation will sooner or later kill great deal of unskilled labor, so we better start thinking now. But again this is another discussion. </s>
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] (hunger, sex, survival) [ENDQ] [NEWLINE] Hunger: fasting for religious reasons (or to a lesser degree, going on a diet) [NEWLINE] [NEWLINE] Sex: celibate monks [NEWLINE] [NEWLINE] Survival: suicide (seppuku) [NEWLINE] [NEWLINE] [NEWLINE] Sure, all three examples are uncommon,<mask> my point is just that we don't necessarily *have* to act in accordance with our evolutionary goals. Neither does an AI, especially<mask> --<mask> you said yourself -- it can modify it's own code. [NEWLINE] </s>
Label encoding: <s> [STARTQ] (hunger, sex, survival) [ENDQ] [NEWLINE] Hunger: fasting for religious reasons (or to a lesser degree, going on a diet) [NEWLINE] [NEWLINE] Sex: celibate monks [NEWLINE] [NEWLINE] Survival: suicide (seppuku) [NEWLINE] [NEWLINE] [NEWLINE] Sure, all three examples are uncommon, but my point is just that we don't necessarily *have* to act in accordance with our evolutionary goals. Neither does an AI, especially since -- as you said yourself -- it can modify it's own code. [NEWLINE] </s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Times change,<mask> you are aware. We went back on prohibition and there is no reason we can't go back on any other law. I simply have a libertarian view on government, and<mask> I find it wrong for the government to come in and say "no, you have to serve this person." Forcing racists to serve minorities doesn't help get rid of racism at all, it simply takes away a freedom from the business owner. We should be using education to fight racism, not nanny-state laws.</s>
Label encoding: <s>Times change, as you are aware. We went back on prohibition and there is no reason we can't go back on any other law. I simply have a libertarian view on government, and so I find it wrong for the government to come in and say "no, you have to serve this person." Forcing racists to serve minorities doesn't help get rid of racism at all, it simply takes away a freedom from the business owner. We should be using education to fight racism, not nanny-state laws.</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Older players naturally had more time to practice,<mask> age segregation makes sence - to even the playing field. [ENDQ] [NEWLINE] This is wrong. The current world champion is only 23, and the world number 3 (<mask><mask> 3? Im referring to Caruana). The theory is that<mask> your brain is really in optimal shape from 20-30, it is ideal that you have already learned everything and became a strong player earlier,<mask> you can be at your peak at the same age your brain is. </s><pad>
Label encoding: <s> [STARTQ] Older players naturally had more time to practice, so age segregation makes sence - to even the playing field. [ENDQ] [NEWLINE] This is wrong. The current world champion is only 23, and the world number 3 ( I think 3? Im referring to Caruana). The theory is that since your brain is really in optimal shape from 20-30, it is ideal that you have already learned everything and became a strong player earlier, so you can be at your peak at the same age your brain is. </s><pad>
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That is my view yes<mask> I was intending to discuss Islam specifically<mask> it is currently in the UK media.<mask><mask> I have said that my problem is with certain Islamist sects and not Islam<mask> a whole and especially not Muslims, people (including you) assume I am ignorant and making a generalization. I assume people (like yourself) are probably doing this subconsciously.<mask>, I'm not even talking about terrorism. Rather, more extreme views which are held by certain sects. [NEWLINE] Edit:Grammar</s>
Label encoding: <s>That is my view yes but I was intending to discuss Islam specifically because it is currently in the UK media. Even though I have said that my problem is with certain Islamist sects and not Islam as a whole and especially not Muslims, people (including you) assume I am ignorant and making a generalization. I assume people (like yourself) are probably doing this subconsciously. Also, I'm not even talking about terrorism. Rather, more extreme views which are held by certain sects. [NEWLINE] Edit:Grammar</s>
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is not a corollary<mask><mask>'s in tension with the child's well being changes from abortion (bodily autonomy) to child support (full rights to discretionary spending). You may be able to<mask><mask> still the latter tension wins out against the child's well being (it doesn't),<mask> that doesn't change the fact that you're not making a corollary, you're making a separate argument about the relation of the child's well being to its parent's rights to their discretionary spending.</s>
Label encoding: <s>It is not a corollary because what's in tension with the child's well being changes from abortion (bodily autonomy) to child support (full rights to discretionary spending). You may be able to argue that still the latter tension wins out against the child's well being (it doesn't), but that doesn't change the fact that you're not making a corollary, you're making a separate argument about the relation of the child's well being to its parent's rights to their discretionary spending.</s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0594, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know the difference between vegetarian and vegan. I didn't use a slash<mask> I was confused. [NEWLINE] [NEWLINE] People can be vegan for health reasons or philosophical reasons or just<mask>. Even<mask> they are vegan for philosophical reasons, there's no one philosophy that all philosophical vegans follow.<mask> someone is a vegan<mask> not for philosophical reasons, there's no reason they won't wear leather. Or maybe they are philosophical,<mask> their philosophy doesn't forbid wearing leather. There isn't just one type of vegan.</s>
Label encoding: <s>I know the difference between vegetarian and vegan. I didn't use a slash because I was confused. [NEWLINE] [NEWLINE] People can be vegan for health reasons or philosophical reasons or just because. Even if they are vegan for philosophical reasons, there's no one philosophy that all philosophical vegans follow. If someone is a vegan but not for philosophical reasons, there's no reason they won't wear leather. Or maybe they are philosophical, but their philosophy doesn't forbid wearing leather. There isn't just one type of vegan.</s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Their academic fields don't really focus on *<mask> * you have satisfying sex<mask>, and with studies that do there isn't really any clear consensus and the field is quite young [NEWLINE] [NEWLINE] <mask><mask> can you really teach that *isn't* positions and techniques<mask> we already teach the biological side of things and the emotional side of things? Like<mask> would a class on '<mask> to have good sex between two men' actually work without touching on that topic and not covering things that are already taught? [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Their academic fields don't really focus on * how * you have satisfying sex though, and with studies that do there isn't really any clear consensus and the field is quite young [NEWLINE] [NEWLINE] But what can you really teach that *isn't* positions and techniques when we already teach the biological side of things and the emotional side of things? Like how would a class on'How to have good sex between two men' actually work without touching on that topic and not covering things that are already taught? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am personally afraid of clowns, and<mask> mascots. I am afraid of them coming up to me and being social, attracting the attention of all passerby's. Mascots more-<mask><mask> you usually can't tell which way they are looking and I feel<mask><mask> their eyes are always on me.<mask> i'm not afraid of them in the sense that people are afraid of the boogieman,<mask> more<mask> of the inevitable awkard social interaction and attention that they bring.</s>
Label encoding: <s>I am personally afraid of clowns, and also mascots. I am afraid of them coming up to me and being social, attracting the attention of all passerby's. Mascots more- so because you usually can't tell which way they are looking and I feel as though their eyes are always on me. So i'm not afraid of them in the sense that people are afraid of the boogieman, but more so of the inevitable awkard social interaction and attention that they bring.</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes.<mask> I'm not discussing the pros and cons of communism and consumerism. My point is simply that these people have been kicked like dogs repeatedly and of course they will hate you for it. [NEWLINE] [NEWLINE] The fact is it is VERY HARD to lift your self out of poverty, and it is intentionally kept this way.<mask> we are spoon fed this "pull yourself up by your bootstraps" attitude and expect that everyone should be able to do this. It's not the case. [NEWLINE] </s>
Label encoding: <s>Yes. But I'm not discussing the pros and cons of communism and consumerism. My point is simply that these people have been kicked like dogs repeatedly and of course they will hate you for it. [NEWLINE] [NEWLINE] The fact is it is VERY HARD to lift your self out of poverty, and it is intentionally kept this way. Yet we are spoon fed this "pull yourself up by your bootstraps" attitude and expect that everyone should be able to do this. It's not the case. [NEWLINE] </s>
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would say reasonably sure - not that I can put a percentage to it. [NEWLINE] [NEWLINE] [STARTQ] a whole lot of things can go wrong fast, and it's hard to be sure of much. [ENDQ] [NEWLINE] <mask> it is hard to be sure of much then *killing is unjustified*.<mask> you are not reasonably sure that<mask> you are about to do isn't murdering an innocent then you are only justified in non-lethal means. [NEWLINE] [NEWLINE] You can temporarily incapacitate someone, you cannot temporarily kill someone.</s>
Label encoding: <s>I would say reasonably sure - not that I can put a percentage to it. [NEWLINE] [NEWLINE] [STARTQ] a whole lot of things can go wrong fast, and it's hard to be sure of much. [ENDQ] [NEWLINE] If it is hard to be sure of much then *killing is unjustified*. If you are not reasonably sure that what you are about to do isn't murdering an innocent then you are only justified in non-lethal means. [NEWLINE] [NEWLINE] You can temporarily incapacitate someone, you cannot temporarily kill someone.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not a good comparison. The people who drafted the US constitution didn't claim infallibility, or claimed the divine right to rule. [NEWLINE] [NEWLINE] They are the reflection of the will of the people of the time. Not the will of an omnipotent and omniscient god being projected to the people. [NEWLINE] [NEWLINE] That is, unless you're willing to contend that the laws and rules of the Catholic church are the product of social and economic forces like any other government system. Then I would agree with you.</s>
Label encoding: <s>Not a good comparison. The people who drafted the US constitution didn't claim infallibility, or claimed the divine right to rule. [NEWLINE] [NEWLINE] They are the reflection of the will of the people of the time. Not the will of an omnipotent and omniscient god being projected to the people. [NEWLINE] [NEWLINE] That is, unless you're willing to contend that the laws and rules of the Catholic church are the product of social and economic forces like any other government system. Then I would agree with you.</s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Agree with most of your post,<mask> this: [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> Iran is a freaking democracy (almost,<mask> not fully) [ENDQ] [NEWLINE] Reaaaaally not... like, not even close.  The President is a figurehead: the Ayatollah has the ultimate power, and appoints everything from presidential candidates to military officers. [NEWLINE] [NEWLINE] It's a theocracy, pure and simple, and the violent shutdown of the Green Movement is recent proof of<mask> happens<mask> real democracy is attempted.</s>
Label encoding: <s>Agree with most of your post, but this: [NEWLINE] [NEWLINE] [STARTQ] Even though Iran is a freaking democracy (almost, but not fully) [ENDQ] [NEWLINE] Reaaaaally not... like, not even close.  The President is a figurehead: the Ayatollah has the ultimate power, and appoints everything from presidential candidates to military officers. [NEWLINE] [NEWLINE] It's a theocracy, pure and simple, and the violent shutdown of the Green Movement is recent proof of what happens when real democracy is attempted.</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>School, with the exception of vocational programs, is about preparing you for life, not about preparing you for a job.  The ability to appreciate the arts allows you to experience and enjoy a wider variety of the pleasures of this world. [NEWLINE] [NEWLINE] The bottom line is asking yourself<mask> is the goal of life.  Is it to get a job and make money?  Is it to experience the richness of this world every way we can?  Is it something else, or something in between?  </s>
Label encoding: <s>School, with the exception of vocational programs, is about preparing you for life, not about preparing you for a job.  The ability to appreciate the arts allows you to experience and enjoy a wider variety of the pleasures of this world. [NEWLINE] [NEWLINE] The bottom line is asking yourself what is the goal of life.  Is it to get a job and make money?  Is it to experience the richness of this world every way we can?  Is it something else, or something in between?  </s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not *no* actors,<mask> perhaps not enough actors.  It is one thing to consider a disability matching the character's<mask> a useful boost; it is another to treat it<mask> the only factor. <mask> I can find an able-bodied actor who happens to be a much better singer, actor, coworker than the best disabled candidate (which is more likely simply<mask> a larger population will have a greater probability of finding someone at the extreme ends of things), I won't turn them down.</s>
Label encoding: <s>Not *no* actors, but perhaps not enough actors.  It is one thing to consider a disability matching the character's as a useful boost; it is another to treat it as the only factor.  If I can find an able-bodied actor who happens to be a much better singer, actor, coworker than the best disabled candidate (which is more likely simply because a larger population will have a greater probability of finding someone at the extreme ends of things), I won't turn them down.</s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course,<mask> I wouldn't be upset at a disinterested third party who told me that the person whom I thought was one of my best friends was actually telling people, "I'm just pretending to be his friend<mask> he's rich and spends money on me." [NEWLINE] [NEWLINE] Sometimes it's not just harmless behind-their-backs talk.  And in NO cases would I be upset at the person who told. [NEWLINE] [NEWLINE] <mask> are you trying to say again?  It doesn't make sense.</s>
Label encoding: <s>Of course, but I wouldn't be upset at a disinterested third party who told me that the person whom I thought was one of my best friends was actually telling people, "I'm just pretending to be his friend because he's rich and spends money on me." [NEWLINE] [NEWLINE] Sometimes it's not just harmless behind-their-backs talk.  And in NO cases would I be upset at the person who told. [NEWLINE] [NEWLINE] What are you trying to say again?  It doesn't make sense.</s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0419, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What are you talking about? Vendors use Bitpay and Coinbase<mask> services, just the same<mask> Visa and Paypal. It doesn't really matter<mask> the backend works<mask><mask><mask> the vendor gets the money, correct? There are significant advantages from the vendor's standpoint, e.g. transaction fees are much less, no chargebacks, etc. I somewhat agree with the OP, I see bitcoin more<mask> a compliment to cash than an alternative,<mask> there is certainly a niche for it.</s>
Label encoding: <s>What are you talking about? Vendors use Bitpay and Coinbase as services, just the same as Visa and Paypal. It doesn't really matter how the backend works as long as the vendor gets the money, correct? There are significant advantages from the vendor's standpoint, e.g. transaction fees are much less, no chargebacks, etc. I somewhat agree with the OP, I see bitcoin more as a compliment to cash than an alternative, but there is certainly a niche for it.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a tangent to your last point, I would like to point out that porn is a major driving force behind technology. It helps commercialize new tech and brings the potential for big profit into it. Things like VHS, DVD, and the internet have all been greatly helped by porn. These things would still exist without porn<mask> their adaptation would have been slower. In the future, porn is probably going to drive the creation of virtual reality and provide a clear monetary incentive for developing that technology.</s>
Label encoding: <s>As a tangent to your last point, I would like to point out that porn is a major driving force behind technology. It helps commercialize new tech and brings the potential for big profit into it. Things like VHS, DVD, and the internet have all been greatly helped by porn. These things would still exist without porn but their adaptation would have been slower. In the future, porn is probably going to drive the creation of virtual reality and provide a clear monetary incentive for developing that technology.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Woz did all the real work.  He changed the industry, not Jobs.  Jobs was an excellent salesman, and asshole,<mask> he was good at<mask> he did. [NEWLINE] [NEWLINE] Pixar to me was his greatest accomplishment.  To know<mask> important Jobs was to Apple, just look at the periods without him. [NEWLINE] [NEWLINE] Apple had to borrow money from Microsoft<mask> Jobs returned.  Now they leave Cook in charge and his solution to everything seems to be: to sue. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Woz did all the real work.  He changed the industry, not Jobs.  Jobs was an excellent salesman, and asshole, but he was good at what he did. [NEWLINE] [NEWLINE] Pixar to me was his greatest accomplishment.  To know how important Jobs was to Apple, just look at the periods without him. [NEWLINE] [NEWLINE] Apple had to borrow money from Microsoft when Jobs returned.  Now they leave Cook in charge and his solution to everything seems to be: to sue. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The NSFW isn't just about things that will get fired<mask> you look at them at work. It<mask> tells you<mask> type of content to expect. You might not want to read a sexually explicit story, and the title isn't always enough to tell. Having a NSFW tag fixes this problem. Plus you can<mask> use RES to completely filter out all NSFW posts,<mask> you don't have to see them at all. I know this is the way my parent's use Reddit.</s>
Label encoding: <s>The NSFW isn't just about things that will get fired if you look at them at work. It also tells you what type of content to expect. You might not want to read a sexually explicit story, and the title isn't always enough to tell. Having a NSFW tag fixes this problem. Plus you can also use RES to completely filter out all NSFW posts, so you don't have to see them at all. I know this is the way my parent's use Reddit.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You would need enough of those people to not be able to make a jury.  I bet the poll had a large amount of people who did not respond<mask> they didn't know the details, didn't know who he was, or didn't care to know.  All you need is those people to make a jury. [NEWLINE] [NEWLINE] <mask>, the idea that 55% of people polled thinks he did the right thing is not nearly indicative of<mask> a society thinks (that is hardly a majority).</s>
Label encoding: <s>You would need enough of those people to not be able to make a jury.  I bet the poll had a large amount of people who did not respond because they didn't know the details, didn't know who he was, or didn't care to know.  All you need is those people to make a jury. [NEWLINE] [NEWLINE] Also, the idea that 55% of people polled thinks he did the right thing is not nearly indicative of what a society thinks (that is hardly a majority).</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm arguing for both, mainly that private opinions held privately not intended for the public should not be able to get you fired [NEWLINE] [NEWLINE] and yes it does [NEWLINE] [NEWLINE] <mask> mark zuckerberg one day posted "i sure hate niggers" on the header of facebook it would be understandable<mask> he got fired [NEWLINE] [NEWLINE] <mask> he told his wife "I often wonder<mask> blacks commit more crime than whites" in the privacy of his own home, he should not be fired<mask> that were to get leaked</s><pad>
Label encoding: <s>I'm arguing for both, mainly that private opinions held privately not intended for the public should not be able to get you fired [NEWLINE] [NEWLINE] and yes it does [NEWLINE] [NEWLINE] if mark zuckerberg one day posted "i sure hate niggers" on the header of facebook it would be understandable if he got fired [NEWLINE] [NEWLINE] if he told his wife "I often wonder why blacks commit more crime than whites" in the privacy of his own home, he should not be fired if that were to get leaked</s><pad>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0523, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It's really almost coincidence that white countries developed first [ENDQ] [NEWLINE] It's a tad nitpicky,<mask> to say that white countries developing first is a tad fallacious.  Western Europe was the first part of the world to industrialize<mask> it had the right circumstances for it (access to resources, geographical position, cultural development, etc.). <mask> you ran a simulation ten thousand times with only minute differences in situation, Europe would come out on top in the vast majority of those. [NEWLINE] </s>
Label encoding: <s> [STARTQ] It's really almost coincidence that white countries developed first [ENDQ] [NEWLINE] It's a tad nitpicky, but to say that white countries developing first is a tad fallacious.  Western Europe was the first part of the world to industrialize because it had the right circumstances for it (access to resources, geographical position, cultural development, etc.).  If you ran a simulation ten thousand times with only minute differences in situation, Europe would come out on top in the vast majority of those. [NEWLINE] </s>
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0649, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't really understand, are you saying that you think one species can become two or more,<mask> that a species cannot cross into a different class or phylum?<mask><mask>,<mask> do you think those new classes come from? I mean, there weren't birds before the Jurassic period, and there are many fossils which indicate transitional stages between dinosaur and bird, such<mask> Archaeopteryx. Surely it's a reasonably rational inference given the evidence that birds may have descended from dinosaurs?</s>
Label encoding: <s>I don't really understand, are you saying that you think one species can become two or more, but that a species cannot cross into a different class or phylum? If so, where do you think those new classes come from? I mean, there weren't birds before the Jurassic period, and there are many fossils which indicate transitional stages between dinosaur and bird, such as Archaeopteryx. Surely it's a reasonably rational inference given the evidence that birds may have descended from dinosaurs?</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Externalities should be charged to products and incorporated into their price.<mask> does this affect the price of labor? [NEWLINE] [NEWLINE] <mask> evidence is there that a better mosquito net has greater value than a cure for baldness? The value of these two products is the total amount of sales.<mask> the mosquito nets may save some lives, the value of these lives is low. [NEWLINE] [NEWLINE] <mask> are the poor not able to afford to pay for labor? Perhaps<mask> the poor are not valuable to society.</s>
Label encoding: <s>Externalities should be charged to products and incorporated into their price. How does this affect the price of labor? [NEWLINE] [NEWLINE] What evidence is there that a better mosquito net has greater value than a cure for baldness? The value of these two products is the total amount of sales. While the mosquito nets may save some lives, the value of these lives is low. [NEWLINE] [NEWLINE] Why are the poor not able to afford to pay for labor? Perhaps because the poor are not valuable to society.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0549, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0540, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like you a lot. You're nice. [NEWLINE] [NEWLINE] [STARTQ] we will probably find that there are certain constellations of personality traits and interests and<mask> on which tend to go hand in hand and form an intelligible gestalt of "masculinity" or "femininity". [ENDQ] [NEWLINE] I find that difficult to believe. Couldn't it be more likely that any such gestalts are self-fulfilling prophecies sustained by a social expectation that they're true?</s>
Label encoding: <s>I like you a lot. You're nice. [NEWLINE] [NEWLINE] [STARTQ] we will probably find that there are certain constellations of personality traits and interests and so on which tend to go hand in hand and form an intelligible gestalt of "masculinity" or "femininity". [ENDQ] [NEWLINE] I find that difficult to believe. Couldn't it be more likely that any such gestalts are self-fulfilling prophecies sustained by a social expectation that they're true?</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Football *is* boring unless you have a natural interest in it. [ENDQ] [NEWLINE] Disagree. [NEWLINE] [NEWLINE] Even<mask> you're looking at the game from a disinterested perspective, it still is wonderful and exhilarating to see those beautiful passes, the skilfully executed crosses and last moment finishers. [NEWLINE] Even people who have no interest in football start going crazy during the World Cup and<mask> not, not<mask> of emotional investment<mask><mask> they're suddenly *in* the game.</s>
Label encoding: <s> [STARTQ] Football *is* boring unless you have a natural interest in it. [ENDQ] [NEWLINE] Disagree. [NEWLINE] [NEWLINE] Even if you're looking at the game from a disinterested perspective, it still is wonderful and exhilarating to see those beautiful passes, the skilfully executed crosses and last moment finishers. [NEWLINE] Even people who have no interest in football start going crazy during the World Cup and what not, not because of emotional investment but because they're suddenly *in* the game.</s>
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No one here seems to be mentioning Russia. Russia had already invaded the Japanese empire before we dropped the bomb. Not only would this have put them in bombing range during the cold war (which could be argued was seen<mask> inevitable by Truman and his cabinet)<mask> Russia does not take prisoners.<mask> they would have ran through Japan and killed every civilian in their path. We had to end the war to A. ensure that Russia would not invade the US and B. to save lives.</s>
Label encoding: <s>No one here seems to be mentioning Russia. Russia had already invaded the Japanese empire before we dropped the bomb. Not only would this have put them in bombing range during the cold war (which could be argued was seen as inevitable by Truman and his cabinet) but Russia does not take prisoners. So they would have ran through Japan and killed every civilian in their path. We had to end the war to A. ensure that Russia would not invade the US and B. to save lives.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Hi from CERN. [ENDQ] [NEWLINE] lol [NEWLINE] [NEWLINE] [STARTQ] Please feel free to tell me<mask> you think I'm wrong about matter. [ENDQ] [NEWLINE] I didn't say you were wrong about matter. I said that you were wrong about logic.<mask> I mean...you are wrong about matter, too,<mask> that's not the point I was making. [NEWLINE] [NEWLINE] Anyway, your argument is missing several premises and<mask> of that your conclusion doesn't follow from the premises you do have. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Hi from CERN. [ENDQ] [NEWLINE] lol [NEWLINE] [NEWLINE] [STARTQ] Please feel free to tell me why you think I'm wrong about matter. [ENDQ] [NEWLINE] I didn't say you were wrong about matter. I said that you were wrong about logic. But I mean...you are wrong about matter, too, but that's not the point I was making. [NEWLINE] [NEWLINE] Anyway, your argument is missing several premises and because of that your conclusion doesn't follow from the premises you do have. [NEWLINE] </s>
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And<mask> you totally ignored her use of the word "queer".  Which a generation (or two) ago was a horrible hurtful hateful word. [NEWLINE] [NEWLINE] <mask> LGBT activists first started using the word "queer" to self-describe (back before it was LGBTQ) the older generation of activists objected on much the same grounds you are objecting today - hate-speech can't be reappropriated. <mask><mask> "queer" shows, it can be.</s>
Label encoding: <s>And yet you totally ignored her use of the word "queer".  Which a generation (or two) ago was a horrible hurtful hateful word. [NEWLINE] [NEWLINE] When LGBT activists first started using the word "queer" to self-describe (back before it was LGBTQ) the older generation of activists objected on much the same grounds you are objecting today - hate-speech can't be reappropriated.  But as "queer" shows, it can be.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] in the context of the definition of harassment that means that there must be some coordinated effort between the members of such “group”. [ENDQ] [NEWLINE] Under the current definition, this is vague.  You could be right<mask> this is not conclusive. <mask> is being proposed is that this is made clearer. [NEWLINE] [NEWLINE] And<mask><mask> it's best to call it "street harassment"<mask> we don't confuse it with e-mails to Justin Bieber, ok?  </s>
Label encoding: <s> [STARTQ] in the context of the definition of harassment that means that there must be some coordinated effort between the members of such “group”. [ENDQ] [NEWLINE] Under the current definition, this is vague.  You could be right but this is not conclusive.  What is being proposed is that this is made clearer. [NEWLINE] [NEWLINE] And I think it's best to call it "street harassment" so we don't confuse it with e-mails to Justin Bieber, ok?  </s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Right<mask><mask> with that. Morals vary with cultures and time. 300 years ago we had few qualms with slavery. Before that, we were trying and executing people we suspected to be witches based on discrimination and little proof. [NEWLINE] [NEWLINE] Locke's theory of natural rights advanced morality and we now recognize these rights to be true and valid, at least in most developed nations. I hope that<mask> time progresses, other nations and cultures become more tolerant and accepting<mask> well. </s>
Label encoding: <s>Right I agree with that. Morals vary with cultures and time. 300 years ago we had few qualms with slavery. Before that, we were trying and executing people we suspected to be witches based on discrimination and little proof. [NEWLINE] [NEWLINE] Locke's theory of natural rights advanced morality and we now recognize these rights to be true and valid, at least in most developed nations. I hope that as time progresses, other nations and cultures become more tolerant and accepting as well. </s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My first child I was there for everything, I cherish all that time.  My second son, the last child I will ever had I went back to work<mask> he was  six months.  I missed his first word,<mask> he crawled by himself, his first step, first time he feed himself, first patty cake, his first everything.  Wish I could go back.  I wish it more then anything, maybe even more for his sake then my own.</s><pad>
Label encoding: <s>My first child I was there for everything, I cherish all that time.  My second son, the last child I will ever had I went back to work when he was  six months.  I missed his first word, when he crawled by himself, his first step, first time he feed himself, first patty cake, his first everything.  Wish I could go back.  I wish it more then anything, maybe even more for his sake then my own.</s><pad>
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0606, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have a different perspective.  The act of them going to the party is not, in isolation, wrong.  Just like them having sex with you is not, in isolation, wrong.  It may be wrong involving breaking promises to someone,<mask><mask><mask> that's between the people involved.  The act itself is not wrong<mask>. [NEWLINE] [NEWLINE] <mask> opposed to being the getaway driver of a bank robbery.  The act of robbing a bank itself is wrong.</s>
Label encoding: <s>I have a different perspective.  The act of them going to the party is not, in isolation, wrong.  Just like them having sex with you is not, in isolation, wrong.  It may be wrong involving breaking promises to someone, but IMO that's between the people involved.  The act itself is not wrong though. [NEWLINE] [NEWLINE] As opposed to being the getaway driver of a bank robbery.  The act of robbing a bank itself is wrong.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0651, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is great, thanks.<mask> I'm stuck on this point: [NEWLINE] [NEWLINE] [STARTQ] non-synesthetes were each paired with a synesthete, and told to memorize the synesthete's association between 6 colors and the numbers 1-6 [ENDQ] [NEWLINE] Shouldn't the non-synesthetes generate their own associations? Otherwise the study is basically comparing whether someone who invented a pattern can remember the pattern better than someone they teach it to.</s><pad>
Label encoding: <s>This is great, thanks. However I'm stuck on this point: [NEWLINE] [NEWLINE] [STARTQ] non-synesthetes were each paired with a synesthete, and told to memorize the synesthete's association between 6 colors and the numbers 1-6 [ENDQ] [NEWLINE] Shouldn't the non-synesthetes generate their own associations? Otherwise the study is basically comparing whether someone who invented a pattern can remember the pattern better than someone they teach it to.</s><pad>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0533, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. Very few people,<mask> any, buy and sell used food. (You seem to have misunderstood my last point) [NEWLINE] [NEWLINE] 2. You actually *can* give someone else your money, and ask them to pay for you,<mask> you trust that person enough. It just<mask> happens that you can't always just walk past the cashier and out the door with a bag full of groceries. [NEWLINE] [NEWLINE] 3. Groceries are physical. Movies are digital.</s>
Label encoding: <s>1. Very few people, if any, buy and sell used food. (You seem to have misunderstood my last point) [NEWLINE] [NEWLINE] 2. You actually *can* give someone else your money, and ask them to pay for you, if you trust that person enough. It just so happens that you can't always just walk past the cashier and out the door with a bag full of groceries. [NEWLINE] [NEWLINE] 3. Groceries are physical. Movies are digital.</s>
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0638, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is the reason<mask> almost all police chiefs are for stricter gun laws.<mask> it were illegal to possess a gun in public there would be a 100% conviction rate on any crime<mask> the criminal is carrying a gun. Issues like witness fear and intimidation, or lack of evidence would not matter anymore. It would be much easier to take criminals off the street. The illegal guns would over the course of time get confiscated and gun companies would go out of business. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>This is the reason why almost all police chiefs are for stricter gun laws. If it were illegal to possess a gun in public there would be a 100% conviction rate on any crime where the criminal is carrying a gun. Issues like witness fear and intimidation, or lack of evidence would not matter anymore. It would be much easier to take criminals off the street. The illegal guns would over the course of time get confiscated and gun companies would go out of business. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0654, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A true hero would bite the bullet before loading it up and straight up shooting the Joker down. [ENDQ] [NEWLINE] I reject this premise entirely. You can justify the action to yourself,<mask> it is a dellusion to call it heroic. It's morally hypocritical to commit against someone the act you condemn them for committing. Period. This makes it inherently unheroic. Batman is under no moral obligation to beat a mentally ill man to death. The idea is absurd.</s>
Label encoding: <s> [STARTQ] A true hero would bite the bullet before loading it up and straight up shooting the Joker down. [ENDQ] [NEWLINE] I reject this premise entirely. You can justify the action to yourself, but it is a dellusion to call it heroic. It's morally hypocritical to commit against someone the act you condemn them for committing. Period. This makes it inherently unheroic. Batman is under no moral obligation to beat a mentally ill man to death. The idea is absurd.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>True, opening up foreign investment accounts might be a good idea. [NEWLINE] [NEWLINE] I would still be worried about my ability to get to those assets in time.<mask> I am hungry today, having 50K in a Chinese account would not help me immediately. [NEWLINE] [NEWLINE] This is a good argument,<mask> it does not dissuade from buying commodities I can store at home. Instead, I would withdraw ANOTHER 10% of my retirement account and invest it in overseas accounts. [NEWLINE] </s>
Label encoding: <s>True, opening up foreign investment accounts might be a good idea. [NEWLINE] [NEWLINE] I would still be worried about my ability to get to those assets in time. If I am hungry today, having 50K in a Chinese account would not help me immediately. [NEWLINE] [NEWLINE] This is a good argument, but it does not dissuade from buying commodities I can store at home. Instead, I would withdraw ANOTHER 10% of my retirement account and invest it in overseas accounts. [NEWLINE] </s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0457, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I sympathize with you,<mask> don't ever assume someone is skinny<mask> they didn't work for it. Even<mask> they didn't have to put in<mask> much effort to lose weight, that's<mask> they made healthy lifestyle decisions in the first place. [NEWLINE] [NEWLINE] Skinny-hating is a slippery slope to being accepting of your weight,<mask><mask> you don't like people who are fit,<mask> will you ever the motivated to become one of those people yourself?</s>
Label encoding: <s>I sympathize with you, but don't ever assume someone is skinny because they didn't work for it. Even if they didn't have to put in as much effort to lose weight, that's because they made healthy lifestyle decisions in the first place. [NEWLINE] [NEWLINE] Skinny-hating is a slippery slope to being accepting of your weight, because if you don't like people who are fit, how will you ever the motivated to become one of those people yourself?</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Where does simply drawing religious figures weigh in<mask>?<mask><mask> generally with punching up vs punching down, (<mask> some christian's will be quick often to tell you they are a  persecuted minority) - sometimes mocking a position a minority takes still seems like it can be valid,<mask> that idea is generally innocuous to mock (some of the examples here are outright horrible, like burning down temples, attacking people and racism) and ridiculous to hold. Think about Scientology.... </s>
Label encoding: <s>Where does simply drawing religious figures weigh in though? I agree generally with punching up vs punching down, ( although some christian's will be quick often to tell you they are a  persecuted minority) - sometimes mocking a position a minority takes still seems like it can be valid, if that idea is generally innocuous to mock (some of the examples here are outright horrible, like burning down temples, attacking people and racism) and ridiculous to hold. Think about Scientology.... </s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of art is subjective by its nature.<mask> it should be put<mask><mask> art provokes thought or evokes emotion IN YOU then it's good art to you.<mask> it does the same to a lot of people, it's good to a lot of people. [NEWLINE] [NEWLINE] Just<mask> something is technically good in art it's not great art. It has to have thought and emotion.<mask> thought and emotion is lent by the viewer not the artist. </s>
Label encoding: <s>A lot of art is subjective by its nature. So it should be put as if art provokes thought or evokes emotion IN YOU then it's good art to you. If it does the same to a lot of people, it's good to a lot of people. [NEWLINE] [NEWLINE] Just because something is technically good in art it's not great art. It has to have thought and emotion. Also thought and emotion is lent by the viewer not the artist. </s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As I understand it, a $20 minimum wage would increase unemployment by 3-4%. Hardly a disaster. And Section 3 of my scheme is meant to create a feedback mechanism: it progressively punishes business owners and employers by taking money out of their pockets to pay unemployment benefits. Section 4, GDP adjustment, effectively gives *everyone* an incentive to increase capital production and economic activity.<mask> there are large-scale mechanisms designed to both mitigate and discourage unemployment.</s>
Label encoding: <s>As I understand it, a $20 minimum wage would increase unemployment by 3-4%. Hardly a disaster. And Section 3 of my scheme is meant to create a feedback mechanism: it progressively punishes business owners and employers by taking money out of their pockets to pay unemployment benefits. Section 4, GDP adjustment, effectively gives *everyone* an incentive to increase capital production and economic activity. So there are large-scale mechanisms designed to both mitigate and discourage unemployment.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0487, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0519, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0877, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're sort of right,<mask> not entirely correct. [NEWLINE] [NEWLINE] Let's say I have $1 trillion. By the time I have that kind of money, there is no possible way I don't own a bunch of companies.<mask>,<mask> am I going to circulate that money? Between my own franchises. [NEWLINE] [NEWLINE] <mask>,<mask> you don't work for me, you don't work.<mask> I'm an asshole, your life is going to suck.</s>
Label encoding: <s>You're sort of right, but not entirely correct. [NEWLINE] [NEWLINE] Let's say I have $1 trillion. By the time I have that kind of money, there is no possible way I don't own a bunch of companies. So, how am I going to circulate that money? Between my own franchises. [NEWLINE] [NEWLINE] So, if you don't work for me, you don't work. If I'm an asshole, your life is going to suck.</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I will be the primary earner next year.  I will be earning about 80%, and my fiance 20%. [NEWLINE] [NEWLINE] And that is,<mask><mask>,<mask> I feel<mask> strongly about this.  I don't believe that I'm entitled to more of the pot simply<mask> I will be working<mask> my fiance will be going to grad school.  We'll both be working hard - it's simply that he will not be financially rewarded for it.</s>
Label encoding: <s>I will be the primary earner next year.  I will be earning about 80%, and my fiance 20%. [NEWLINE] [NEWLINE] And that is, in fact, why I feel so strongly about this.  I don't believe that I'm entitled to more of the pot simply because I will be working while my fiance will be going to grad school.  We'll both be working hard - it's simply that he will not be financially rewarded for it.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm surprised it was that easy to convince you. I don't even think Tucker covered the point I find most convincing: [NEWLINE] [NEWLINE] 4) You knowing things about the world and sharing them helps spread that knowledge. You can be responsible for educating more people, and<mask> it's something that can be actionable, or something even a few people feel passionate about, learning, discussing, and passing on information and news is a major way that is facilitated.</s>
Label encoding: <s>I'm surprised it was that easy to convince you. I don't even think Tucker covered the point I find most convincing: [NEWLINE] [NEWLINE] 4) You knowing things about the world and sharing them helps spread that knowledge. You can be responsible for educating more people, and if it's something that can be actionable, or something even a few people feel passionate about, learning, discussing, and passing on information and news is a major way that is facilitated.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I had a 94 Geo Metro. At the time, my dad ran a salvage yard and brought home a Barracuda muffler that essentially made the car sound like a dirt bike. Either way,<mask> it comes to mufflers,<mask> was stated above, any shop that has a decent welder can make a muffler that's either too big or small fit with a reduction/expansion price, then weld in the hangers and it's done</s>
Label encoding: <s>I had a 94 Geo Metro. At the time, my dad ran a salvage yard and brought home a Barracuda muffler that essentially made the car sound like a dirt bike. Either way, when it comes to mufflers, as was stated above, any shop that has a decent welder can make a muffler that's either too big or small fit with a reduction/expansion price, then weld in the hangers and it's done</s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>it is an issue, the thing is being the caring type isn't always initially interesting to some women, maybe most, i dunno of any studies that have got any data on this<mask> that's<mask> it seems like to me, anecdotally.<mask><mask> there's a distinction between being a dick and being alpha. it's harder to put into words<mask><mask><mask> the underlying principle is confidence and a sort of selfishness that isn't too extreme. </s>
Label encoding: <s>it is an issue, the thing is being the caring type isn't always initially interesting to some women, maybe most, i dunno of any studies that have got any data on this but that's what it seems like to me, anecdotally. i think there's a distinction between being a dick and being alpha. it's harder to put into words but i think the underlying principle is confidence and a sort of selfishness that isn't too extreme. </s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well, wealth redistribution isn't that difficult, it just takes political will, which requires long-term popular support.<mask> the political will exists, it can be done in a variety of ways. More progressive taxation, subsidies, direct payments, increased minimum wage, etc. These are all controversial programs in that not everyone believes they are effective<mask> of negative unintended consequences (inflation, reduced profitability of business, etc.).<mask> they unquestionably redistribute wealth.</s>
Label encoding: <s>Well, wealth redistribution isn't that difficult, it just takes political will, which requires long-term popular support. When the political will exists, it can be done in a variety of ways. More progressive taxation, subsidies, direct payments, increased minimum wage, etc. These are all controversial programs in that not everyone believes they are effective because of negative unintended consequences (inflation, reduced profitability of business, etc.). But they unquestionably redistribute wealth.</s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think things are made more difficult for people due to cultural reasons. For example, a woman who is older vs. a man who is older. Men tend to "increase in mate value" with age,<mask> women "decrease in mate value".<mask>, it's harder for older women... this seems like an unfair result of cultural preferences.<mask><mask> for a lot of these "nice guys" that cultural bias is unfair to them.</s>
Label encoding: <s>I think things are made more difficult for people due to cultural reasons. For example, a woman who is older vs. a man who is older. Men tend to "increase in mate value" with age, while women "decrease in mate value". So, it's harder for older women... this seems like an unfair result of cultural preferences. I think for a lot of these "nice guys" that cultural bias is unfair to them.</s>
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0738, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All of us are a collection of traits, some common and some less<mask>.  No single trait defines any of us, and without any single one of them we would be different people. [NEWLINE] [NEWLINE] Your foot fetish may be uncommon,<mask> it's completely normal for you.  It's an attraction that's part of who you are. [NEWLINE] [NEWLINE] There is no concrete objective normality. <mask> there were none of us would completely fit it.</s>
Label encoding: <s>All of us are a collection of traits, some common and some less so.  No single trait defines any of us, and without any single one of them we would be different people. [NEWLINE] [NEWLINE] Your foot fetish may be uncommon, but it's completely normal for you.  It's an attraction that's part of who you are. [NEWLINE] [NEWLINE] There is no concrete objective normality.  If there were none of us would completely fit it.</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0884, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right; I was misinformed. [NEWLINE] [STARTQ] In the real world, one bullet usually will not magically cause an attacker to fall over dead. [ENDQ] [NEWLINE] Slow down and read more carefully. My whole point is that the officer didn't need to make him magically fall over dead. He could've tazed or maced the guy, even shot him in a limb or the stomach, instead of unloading a clip on an unarmed man.</s>
Label encoding: <s>You're right; I was misinformed. [NEWLINE] [STARTQ] In the real world, one bullet usually will not magically cause an attacker to fall over dead. [ENDQ] [NEWLINE] Slow down and read more carefully. My whole point is that the officer didn't need to make him magically fall over dead. He could've tazed or maced the guy, even shot him in a limb or the stomach, instead of unloading a clip on an unarmed man.</s>
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] By giving him deus ex tech and massive plot favoritism to the point<mask> he's always going to be the most smartest man in the room. [ENDQ] [NEWLINE] <mask> is that really plot favoritism<mask> that's just his power? That's like saying its plot favoritism that The Flash is always going to be the fastest person in the room. That's just<mask> he is and<mask> he's<mask> he's at in the first place.</s>
Label encoding: <s> [STARTQ] By giving him deus ex tech and massive plot favoritism to the point where he's always going to be the most smartest man in the room. [ENDQ] [NEWLINE] But is that really plot favoritism if that's just his power? That's like saying its plot favoritism that The Flash is always going to be the fastest person in the room. That's just what he is and why he's where he's at in the first place.</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So only men are biologically capable of being a good President of the United States of America? [NEWLINE] [NEWLINE] [STARTQ] <mask> I see there are getting to be more and more stay-at-home dads and more women assuming men's positions in the workforce and at home. [ENDQ] [NEWLINE] <mask> this is currently increasing, that is strong evidence that we haven't reached equality by any measure (even<mask> you assume there will always be some inequality due to biological differences).</s>
Label encoding: <s>So only men are biologically capable of being a good President of the United States of America? [NEWLINE] [NEWLINE] [STARTQ] because I see there are getting to be more and more stay-at-home dads and more women assuming men's positions in the workforce and at home. [ENDQ] [NEWLINE] If this is currently increasing, that is strong evidence that we haven't reached equality by any measure (even if you assume there will always be some inequality due to biological differences).</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 0° - Really cold [ENDQ] [STARTQ] 50° -<mask> -<mask> [ENDQ] [STARTQ] 100° - Really hot [ENDQ] [NEWLINE] I'd never heard this, and I still have no idea<mask> "really cold" 0°F is. (And I assume it's completely subjective to wind, humidity, clothing and personal preferences) [NEWLINE] [NEWLINE] At 0°C, Ice happens. That is a fairly significant event in nature, and should be easily identifiable. </s>
Label encoding: <s> [STARTQ] 0° - Really cold [ENDQ] [STARTQ] 50° - So - so [ENDQ] [STARTQ] 100° - Really hot [ENDQ] [NEWLINE] I'd never heard this, and I still have no idea how "really cold" 0°F is. (And I assume it's completely subjective to wind, humidity, clothing and personal preferences) [NEWLINE] [NEWLINE] At 0°C, Ice happens. That is a fairly significant event in nature, and should be easily identifiable. </s>
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What? There's always an alternative agenda for every major movement. Everyone's going to use the first theory that says they're right to try to prove<mask> they already supported. With something<mask> major<mask> Keynesian economics it's definitely happened. You'd have to go out of your way to prove that *everyone* citing Keynesian theory got there through proper economics and aren't just choosing a theory they agreed with. The opposite claim is ridiculous.</s>
Label encoding: <s>What? There's always an alternative agenda for every major movement. Everyone's going to use the first theory that says they're right to try to prove what they already supported. With something as major as Keynesian economics it's definitely happened. You'd have to go out of your way to prove that *everyone* citing Keynesian theory got there through proper economics and aren't just choosing a theory they agreed with. The opposite claim is ridiculous.</s>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I always found the word "creepy" to be near meaningless. Not necessarily<mask> it literally means nothing,<mask><mask> it implies something is scary or fear inducing without legitimizing the claim. There is no implication of any kind of danger or evidence of<mask> the alleged creepy act is<mask> creepy. It's just one person's interpretation of behavior. [NEWLINE] [NEWLINE] Not to say nothing can be creepy. Seems everything is getting labelled<mask> nowadays<mask>.</s>
Label encoding: <s>I always found the word "creepy" to be near meaningless. Not necessarily because it literally means nothing, but because it implies something is scary or fear inducing without legitimizing the claim. There is no implication of any kind of danger or evidence of why the alleged creepy act is indeed creepy. It's just one person's interpretation of behavior. [NEWLINE] [NEWLINE] Not to say nothing can be creepy. Seems everything is getting labelled so nowadays though.</s>
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> not focus on putting a woman on modern US currency, and more importantly, a person who merits such an honor in her own right, rather than a civil rights "leader" who was shrouded in scandal or a racist whose policies in South America and the West effectively amounted to genocide? [ENDQ] [NEWLINE] <mask> we're ignoring the Susan B Anthony and Sacagawea Dollar coins here right?<mask> we do have women on some of our currency.</s>
Label encoding: <s> [STARTQ] Why not focus on putting a woman on modern US currency, and more importantly, a person who merits such an honor in her own right, rather than a civil rights "leader" who was shrouded in scandal or a racist whose policies in South America and the West effectively amounted to genocide? [ENDQ] [NEWLINE] So we're ignoring the Susan B Anthony and Sacagawea Dollar coins here right? Because we do have women on some of our currency.</s>
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0840, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"reason" should be in quotes, or change the phrase  to "triumph of technicalities over reason". [NEWLINE] [NEWLINE] [NEWLINE] Your arguments place faith in the word "legal" and in the credibility that the word imparts on actions,<mask><mask> it were sacrosanct. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] a[...]person has to avoid fallacies. [ENDQ] [NEWLINE] not<mask> it's at the cost of justice. (in this case)</s>
Label encoding: <s>"reason" should be in quotes, or change the phrase  to "triumph of technicalities over reason". [NEWLINE] [NEWLINE] [NEWLINE] Your arguments place faith in the word "legal" and in the credibility that the word imparts on actions, as if it were sacrosanct. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] a[...]person has to avoid fallacies. [ENDQ] [NEWLINE] not if it's at the cost of justice. (in this case)</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0599, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Consent is permission for something to happen or agreement to do something. I can agree to sex without agreeing to parenthood. I do not see<mask> that is a difficult concept.<mask> related, they are two separate tasks that people do exclusively. [NEWLINE] [NEWLINE] Equality is the state of being equal, esp. in status, rights, and opportunities. The rights and opportunities in this situation are not equal, this is *not* equality.</s>
Label encoding: <s>Consent is permission for something to happen or agreement to do something. I can agree to sex without agreeing to parenthood. I do not see why that is a difficult concept. Though related, they are two separate tasks that people do exclusively. [NEWLINE] [NEWLINE] Equality is the state of being equal, esp. in status, rights, and opportunities. The rights and opportunities in this situation are not equal, this is *not* equality.</s>
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well said!<mask><mask> we agree that the OP is taking too much of an active role in other's decisions. The idea that you would be a stronger person<mask> you were to acquire confidence naturally is valid,<mask> that would be the OP's choice to make, not to shame others into believing. You can present your case to others who drink,<mask> do not expect them to be strongly affected by your judgement<mask> it is their decision.</s>
Label encoding: <s>Well said! I think we agree that the OP is taking too much of an active role in other's decisions. The idea that you would be a stronger person if you were to acquire confidence naturally is valid, however that would be the OP's choice to make, not to shame others into believing. You can present your case to others who drink, but do not expect them to be strongly affected by your judgement when it is their decision.</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Now the question is could it be equally possible that cognitive skills are equally disparate.  I wonder<mask> there has been any comparison of say, male chess grandmasters against their female counter parts, or even a comparison of Jeopardy contestants. [NEWLINE] [NEWLINE] Off the top of my head all of highest rank Jeopardy winners have been men; same for Who Wants to be a Millionaire? contestants who've won the grand prize. </s>
Label encoding: <s>Now the question is could it be equally possible that cognitive skills are equally disparate.  I wonder if there has been any comparison of say, male chess grandmasters against their female counter parts, or even a comparison of Jeopardy contestants. [NEWLINE] [NEWLINE] Off the top of my head all of highest rank Jeopardy winners have been men; same for Who Wants to be a Millionaire? contestants who've won the grand prize. </s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What about situations<mask> a woman is on a specific kind of birth control<mask> she cannot expect a regular period, and it fails,<mask> she does not find out until after the 8 week window? [NEWLINE] [NEWLINE] <mask>, your idea that pregnancy is apparent by the tenth week is kind of misinformed. All bodies are different and some women will have no signs of pregnancy<mask> -<mask> -ever until they start showing around week 15-16.</s>
Label encoding: <s>What about situations where a woman is on a specific kind of birth control where she cannot expect a regular period, and it fails, so she does not find out until after the 8 week window? [NEWLINE] [NEWLINE] Also, your idea that pregnancy is apparent by the tenth week is kind of misinformed. All bodies are different and some women will have no signs of pregnancy what - so -ever until they start showing around week 15-16.</s>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree society has a huge condescension problem with young people. We view them<mask> mature enough to make huge life decisions such<mask> joining the military, marriage or being tried<mask> an adult in court<mask> we don't respect their decision to have sex? An absolutely harmless act that doesn't have to have any effect on their lives and suddenly we don't think they can handle the deep difficult question of "do I want to get fucked?"</s>
Label encoding: <s>I agree society has a huge condescension problem with young people. We view them as mature enough to make huge life decisions such as joining the military, marriage or being tried as an adult in court but we don't respect their decision to have sex? An absolutely harmless act that doesn't have to have any effect on their lives and suddenly we don't think they can handle the deep difficult question of "do I want to get fucked?"</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Even<mask> the animal is free range. You are still murdering him to get meat, no animal would ever consent to be murdered. [NEWLINE] [NEWLINE] With consensual harm-free bestiality, the animal remains alive and is not harmed and can enjoy or find the activity to be neutral. Which is way better than being murdered. [NEWLINE] [NEWLINE] <mask><mask> : Religious opinions are irrelevant in non-religious debates. This is a non-religious debate.</s>
Label encoding: <s>Even if the animal is free range. You are still murdering him to get meat, no animal would ever consent to be murdered. [NEWLINE] [NEWLINE] With consensual harm-free bestiality, the animal remains alive and is not harmed and can enjoy or find the activity to be neutral. Which is way better than being murdered. [NEWLINE] [NEWLINE] IMHO : Religious opinions are irrelevant in non-religious debates. This is a non-religious debate.</s>
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You dont know<mask> it is exciting to watch the best teams in the world compete with each other to determine who is the best? Isnt that pretty much the point of any sport? [NEWLINE] [NEWLINE] I know nothing about Dota and dont follow it,<mask> even I could list 10 dota teams atm. I guess somebody following the scene knows more. And<mask> are 10 teams to follow to few? Thats a perfectly fine number.</s>
Label encoding: <s>You dont know how it is exciting to watch the best teams in the world compete with each other to determine who is the best? Isnt that pretty much the point of any sport? [NEWLINE] [NEWLINE] I know nothing about Dota and dont follow it, but even I could list 10 dota teams atm. I guess somebody following the scene knows more. And why are 10 teams to follow to few? Thats a perfectly fine number.</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To simplify, left wing views support the population taking on some of the living costs of poorer citizens.<mask> a wealthy right-wing person may be selfish.<mask> a poor right wing person, who would have all the vantage of a left-wing system, cannot be characterized<mask> "selfish." A poor left-wing person would be selfish,<mask> they support an ideology that would effectively take wealth from others and give it to them.</s>
Label encoding: <s>To simplify, left wing views support the population taking on some of the living costs of poorer citizens. So a wealthy right-wing person may be selfish. But a poor right wing person, who would have all the vantage of a left-wing system, cannot be characterized as "selfish." A poor left-wing person would be selfish, as they support an ideology that would effectively take wealth from others and give it to them.</s>
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Very true. ∆ [NEWLINE] I do admit there are some exceptions to my statement. They happened at a time before that particular branch of art had come about,<mask>. [NEWLINE] The painting now has more historical significance,<mask><mask> makes it art was the fact that people questioned it<mask> it was something entirely new, rather than the questioning you do yourself, comparing it to the things created in the art style that sprung away from it.</s>
Label encoding: <s>Very true. ∆ [NEWLINE] I do admit there are some exceptions to my statement. They happened at a time before that particular branch of art had come about, though. [NEWLINE] The painting now has more historical significance, because what makes it art was the fact that people questioned it when it was something entirely new, rather than the questioning you do yourself, comparing it to the things created in the art style that sprung away from it.</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Most spontaneous abortions happen before any swelling occurs (first trimester). [ENDQ] [NEWLINE] Swelling happens immediately during pregnancy and detectable with a physical examination, and a first trimester spontaneous abortion causing<mask> much pain that speaking is inhibited would be exceedingly rare.<mask>, nothing can be done to stop a first trimester spontaneous abortion anyway.<mask> there is hemorrhaging, the course of treatment is a transfusion<mask><mask> the biological sex.</s>
Label encoding: <s> [STARTQ] Most spontaneous abortions happen before any swelling occurs (first trimester). [ENDQ] [NEWLINE] Swelling happens immediately during pregnancy and detectable with a physical examination, and a first trimester spontaneous abortion causing so much pain that speaking is inhibited would be exceedingly rare. Also, nothing can be done to stop a first trimester spontaneous abortion anyway. If there is hemorrhaging, the course of treatment is a transfusion regardless of the biological sex.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0496, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0614, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ive talked to people who leaned heavily conservative and were very reasonable.<mask><mask><mask><mask>, ive talked to plenty of extremely unreasonable people, both liberal and conservative, and it definitely falls on both sides of the aisle. I will concede<mask>, that in my experience the most reasonable people ive met leaned left to some degree. The Trump number doesnt quite mean<mask> people purport it to mean,<mask> it is upsetting.</s>
Label encoding: <s>Ive talked to people who leaned heavily conservative and were very reasonable. On the other hand, ive talked to plenty of extremely unreasonable people, both liberal and conservative, and it definitely falls on both sides of the aisle. I will concede though, that in my experience the most reasonable people ive met leaned left to some degree. The Trump number doesnt quite mean what people purport it to mean, but it is upsetting.</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0550, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, that's the thing, supply and demand,<mask> being an artist was just about practice, the supply and demand would balance wouldn't it? People would became artists until it's no longer worth it. Like in any other job,<mask> artists would earn an average wage.<mask> famous artists are paid that much it's exactly<mask> they do something not everyone *can* do, even with the same amount of practice.</s>
Label encoding: <s>Yes, that's the thing, supply and demand, if being an artist was just about practice, the supply and demand would balance wouldn't it? People would became artists until it's no longer worth it. Like in any other job, so artists would earn an average wage. If famous artists are paid that much it's exactly because they do something not everyone *can* do, even with the same amount of practice.</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This delta is currently disallowed<mask> your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for<mask> /u/Zygomatico changed your view.<mask> you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>This delta is currently disallowed as your comment contains either no or little text ([comment rule 4]( [URL] #wiki_rule_4)). Please include an explanation for how /u/Zygomatico changed your view. If you edit this in, replying to my comment will make me rescan yours. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>By name dropping Quantum mechanics you try to give your argument scientific credibility.<mask><mask> QM is not applicable to the soul. Let me try using biology instead: we don't know<mask> turtles lay their eggs at a certain full moon,<mask> there are mysteries,<mask> you may exist more than you are.<mask><mask> a lot of the other stuff you wrote was pretty insightful,<mask> the pseudoscience thing makes me cringe...</s>
Label encoding: <s>By name dropping Quantum mechanics you try to give your argument scientific credibility. In fact QM is not applicable to the soul. Let me try using biology instead: we don't know why turtles lay their eggs at a certain full moon, thus there are mysteries, thus you may exist more than you are. I think a lot of the other stuff you wrote was pretty insightful, but the pseudoscience thing makes me cringe...</s>
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Personally, I find charity to be a disgusting necessity of this economic system. [NEWLINE] [NEWLINE] <mask> you probably control capital, or will<mask> your daddy lets you, think about<mask> you can give a hand *up*, instead of a hand *out*. [NEWLINE] [NEWLINE] [NEWLINE] And realize your family might have wealth<mask> they had a hand in minimizing expenses which, in this day and age, is wages to the lower and middle class. </s>
Label encoding: <s>Personally, I find charity to be a disgusting necessity of this economic system. [NEWLINE] [NEWLINE] Since you probably control capital, or will when your daddy lets you, think about how you can give a hand *up*, instead of a hand *out*. [NEWLINE] [NEWLINE] [NEWLINE] And realize your family might have wealth because they had a hand in minimizing expenses which, in this day and age, is wages to the lower and middle class. </s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some people are simply mentally weak and can't handle the abuse the Gordon delivers in his programs. Nobody is signing up expecting this to be a criticism free program, it's something they agreed to put themselves through. Ultimately, the fault lies with those who committed suicide for not knowing their limits. Ramsey and his crew aren't mind readers,<mask> you say you can handle, they're probably going to believe you.  </s>
Label encoding: <s>Some people are simply mentally weak and can't handle the abuse the Gordon delivers in his programs. Nobody is signing up expecting this to be a criticism free program, it's something they agreed to put themselves through. Ultimately, the fault lies with those who committed suicide for not knowing their limits. Ramsey and his crew aren't mind readers, if you say you can handle, they're probably going to believe you.  </s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Similarly, people who are avid fans of the TV show Family Guy may jokingly mispronounce words that begin "wh" to sound like "hw" (<mask> you're not familiar with this meme, search youtube for "family guy cool whip"). [ENDQ] [NEWLINE] Actually, that was the original pronunciation of "wh",<mask> the spelling. It's still common in some parts of the U.S. and Ireland.</s>
Label encoding: <s> [STARTQ] Similarly, people who are avid fans of the TV show Family Guy may jokingly mispronounce words that begin "wh" to sound like "hw" ( if you're not familiar with this meme, search youtube for "family guy cool whip"). [ENDQ] [NEWLINE] Actually, that was the original pronunciation of "wh", hence the spelling. It's still common in some parts of the U.S. and Ireland.</s>
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't honestly know.  I would guess that black and Hispanic/Latino people would have fewer opportunities to get out of that situation, given the statistics on gang membership, poverty, and race. [NEWLINE] [NEWLINE] I'm not saying that race, poverty, and culture are the same issue.  I *am* saying that they have a close enough correlation to all be taken into account<mask> addressing the issue. </s>
Label encoding: <s>I don't honestly know.  I would guess that black and Hispanic/Latino people would have fewer opportunities to get out of that situation, given the statistics on gang membership, poverty, and race. [NEWLINE] [NEWLINE] I'm not saying that race, poverty, and culture are the same issue.  I *am* saying that they have a close enough correlation to all be taken into account when addressing the issue. </s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The definition of civil disobedience is:  the refusal to comply with certain laws or to pay taxes and fines,<mask> a peaceful form of political protest. [NEWLINE] [NEWLINE] She didn't hurt anyone nor did she cause any damage to property.  All she did was rather aggressively promote her beliefs. <mask> you may not agree with the words she used, the mere fact that she only used words classifies this<mask> civil disobedience.</s>
Label encoding: <s>The definition of civil disobedience is:  the refusal to comply with certain laws or to pay taxes and fines, as a peaceful form of political protest. [NEWLINE] [NEWLINE] She didn't hurt anyone nor did she cause any damage to property.  All she did was rather aggressively promote her beliefs.  While you may not agree with the words she used, the mere fact that she only used words classifies this as civil disobedience.</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> IAmAN00bie is saying is that correlation isn't causation. "This is the way it seems to be" isn't an argument. Your not providing proof of causation. [ENDQ] [NEWLINE] <mask>? She's trying to prove that correlation isn't causation by... providing a bunch of things that correlate with her hypothesis (patriarchy theory) in order to declare causation (patriarchy theory)? [NEWLINE] [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] What IAmAN00bie is saying is that correlation isn't causation. "This is the way it seems to be" isn't an argument. Your not providing proof of causation. [ENDQ] [NEWLINE] What? She's trying to prove that correlation isn't causation by... providing a bunch of things that correlate with her hypothesis (patriarchy theory) in order to declare causation (patriarchy theory)? [NEWLINE] [NEWLINE] </s><pad>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are already plenty of universities<mask> greek organizations own their own houses. The charity they do certainly does not dry up, nor would it have any reason to. You should expect that they would do more in order to keep their neighbors happy. [NEWLINE] [NEWLINE] <mask> for rape allegations, ignoring your clear bias, whether a rape takes place on campus or not,<mask> it is between students it is still a campus issue.</s>
Label encoding: <s>There are already plenty of universities where greek organizations own their own houses. The charity they do certainly does not dry up, nor would it have any reason to. You should expect that they would do more in order to keep their neighbors happy. [NEWLINE] [NEWLINE] As for rape allegations, ignoring your clear bias, whether a rape takes place on campus or not, if it is between students it is still a campus issue.</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes,<mask><mask>. [NEWLINE] [NEWLINE] <mask> that's a big deal to the parents, then perhaps they should try to get a job<mask> they would not be potentially required to work on Christmas? [NEWLINE] [NEWLINE] My argument is that the requirements of your job do not change just<mask> you made a baby. <mask> you cannot or will not meet the requirements of your job anymore, then maybe you should find a different job.</s>
Label encoding: <s>Yes, I agree. [NEWLINE] [NEWLINE] If that's a big deal to the parents, then perhaps they should try to get a job where they would not be potentially required to work on Christmas? [NEWLINE] [NEWLINE] My argument is that the requirements of your job do not change just because you made a baby.  If you cannot or will not meet the requirements of your job anymore, then maybe you should find a different job.</s>
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>meh. still probably ok. :) Detroit has one of the few profitable city wide thermal (steam) systems, and one of the best water systems. It serves a metropolitan area of around 4 million people. People just like to take pot-shots at it,<mask><mask> you are talking about infrastructure, you are talking about entire metropolitan areas usually. Of which Detroit is one of the biggest in the country. </s>
Label encoding: <s>meh. still probably ok. :) Detroit has one of the few profitable city wide thermal (steam) systems, and one of the best water systems. It serves a metropolitan area of around 4 million people. People just like to take pot-shots at it, but when you are talking about infrastructure, you are talking about entire metropolitan areas usually. Of which Detroit is one of the biggest in the country. </s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I have never in my entire life been inconvenienced by a loud vehicle. [ENDQ] [NEWLINE] Then you're lucky. [NEWLINE] [NEWLINE] I don't think the OP is talking about a modified exhaust with a throaty growl that still fits the legal limits of noise. [NEWLINE] [NEWLINE] People who set off alarms in the neighborhood<mask> they drive by (like straight pipes on a Harley) are inconsiderate assholes, period.</s>
Label encoding: <s> [STARTQ] I have never in my entire life been inconvenienced by a loud vehicle. [ENDQ] [NEWLINE] Then you're lucky. [NEWLINE] [NEWLINE] I don't think the OP is talking about a modified exhaust with a throaty growl that still fits the legal limits of noise. [NEWLINE] [NEWLINE] People who set off alarms in the neighborhood as they drive by (like straight pipes on a Harley) are inconsiderate assholes, period.</s>
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem is that the way the law is worded implies that only penetration of a "vagina, anus or mouth" is rape, i.e. the act of penetration is the rape.  This implies that being forced to penetrate would not fall under the definition. [NEWLINE] [NEWLINE] <mask> [there's a mentality among some that men cannot be raped] ( [URL] ), the implication in that law becomes more important.</s>
Label encoding: <s>The problem is that the way the law is worded implies that only penetration of a "vagina, anus or mouth" is rape, i.e. the act of penetration is the rape.  This implies that being forced to penetrate would not fall under the definition. [NEWLINE] [NEWLINE] When [there's a mentality among some that men cannot be raped] ( [URL] ), the implication in that law becomes more important.</s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am aware of this movement and that is<mask> I posted this. I'm<mask> arguing for a legally binding agreement that accepting government funding would require publishing in an open access journal. I'm looking for opinions about<mask> such a requirement would be a bad idea. [NEWLINE] [NEWLINE] <mask> I say "free* I mean freely available. I am very aware of the fact that it doesn't mean research will cost less.</s>
Label encoding: <s>I am aware of this movement and that is why I posted this. I'm also arguing for a legally binding agreement that accepting government funding would require publishing in an open access journal. I'm looking for opinions about why such a requirement would be a bad idea. [NEWLINE] [NEWLINE] When I say "free* I mean freely available. I am very aware of the fact that it doesn't mean research will cost less.</s>
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Babies<mask> don't have an *explicit* desire TO live.<mask> people in vegetative state that doesn't have a chance of recovery. [NEWLINE] [NEWLINE] My grandfather had more than one stroke, and lived the last ~~minutes~~ years of his life with a pretty advanced Alzheimer. He couldn't speak and his cognitive abilities were almost nonexistent. Would killing him be<mask> okay<mask> killing an animal?</s>
Label encoding: <s>Babies also don't have an *explicit* desire TO live. Also people in vegetative state that doesn't have a chance of recovery. [NEWLINE] [NEWLINE] My grandfather had more than one stroke, and lived the last ~~minutes~~ years of his life with a pretty advanced Alzheimer. He couldn't speak and his cognitive abilities were almost nonexistent. Would killing him be as okay as killing an animal?</s>
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I would<mask><mask> death is'more violence' than injury, and that killing and injuring more is<mask>'more violence'.<mask> the group in China had attacked with guns, the death toll would undoubtedly be higher, and the number of injuries would probably be higher<mask> well. [NEWLINE] [NEWLINE] [STARTQ] The violence is already there, it's just easier with a better weapon. [ENDQ] [NEWLINE] And<mask> you have more violence.</s>
Label encoding: <s>I would argue that death is'more violence' than injury, and that killing and injuring more is also'more violence'. If the group in China had attacked with guns, the death toll would undoubtedly be higher, and the number of injuries would probably be higher as well. [NEWLINE] [NEWLINE] [STARTQ] The violence is already there, it's just easier with a better weapon. [ENDQ] [NEWLINE] And therefore you have more violence.</s>
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Because it is YOURS and it is your right to decide<mask> to do with it.<mask> close to death must someone be before they are no longer allowed to do with their property<mask> they wish? Who gets vested with the power to make that call? Are they elected or appointed? [NEWLINE] [NEWLINE] The perfect society you envision is one<mask> property rights do not exist, which means the state controls all property. </s>
Label encoding: <s>Because it is YOURS and it is your right to decide what to do with it. How close to death must someone be before they are no longer allowed to do with their property as they wish? Who gets vested with the power to make that call? Are they elected or appointed? [NEWLINE] [NEWLINE] The perfect society you envision is one where property rights do not exist, which means the state controls all property. </s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I can<mask> bully you for being an anti-Semite [ENDQ] [NEWLINE] <mask> I'm actually being anti-Semitic then I'm not sure this is actually bullying. I get that it's a behaviour in general,<mask> I'm not convinced that in response to a strong negative that it's the same thing. Keep in mind I'm saying I'm not convinced, I could go either way on this.</s>
Label encoding: <s> [STARTQ] I can also bully you for being an anti-Semite [ENDQ] [NEWLINE] If I'm actually being anti-Semitic then I'm not sure this is actually bullying. I get that it's a behaviour in general, but I'm not convinced that in response to a strong negative that it's the same thing. Keep in mind I'm saying I'm not convinced, I could go either way on this.</s>
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think a better way would be to make the pattern of [NEWLINE] [NEWLINE] digit word digit word digit word digit word [NEWLINE] [NEWLINE] like 6destroy6illustrious7carbohydrates0kangaroo. That is 64 times more secure(with a 2500 word dictionary) than a compltely random 8 character password and basically only requires you to memorize a 4 digit number and a mnemonic. </s>
Label encoding: <s>I think a better way would be to make the pattern of [NEWLINE] [NEWLINE] digit word digit word digit word digit word [NEWLINE] [NEWLINE] like 6destroy6illustrious7carbohydrates0kangaroo. That is 64 times more secure(with a 2500 word dictionary) than a compltely random 8 character password and basically only requires you to memorize a 4 digit number and a mnemonic. </s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it depends on your perspective.<mask><mask> BLM lost support from the people who think the two women are representative of the entire movement.<mask>. In a weird way<mask><mask> this was exactly<mask> Sanders needed.<mask><mask>, he owes these two women big time. He's been polling horrifically and now he's got national coverage. People are probably looking up his history in response to this. </s>
Label encoding: <s>I think it depends on your perspective. I think BLM lost support from the people who think the two women are representative of the entire movement. BUT. In a weird way I think this was exactly what Sanders needed. In fact, he owes these two women big time. He's been polling horrifically and now he's got national coverage. People are probably looking up his history in response to this. </s>
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is not about the view of the  accused at all. I am not talking about a counter suit. I am talking about a separate investigation ( for which most of the work would already be done) into the truthfulness and credibility of the accuser. The accused is not a part of this process beyond their testimony<mask> to<mask> happened. They do not get to turn around and accuse the accuser.</s>
Label encoding: <s>It is not about the view of the  accused at all. I am not talking about a counter suit. I am talking about a separate investigation ( for which most of the work would already be done) into the truthfulness and credibility of the accuser. The accused is not a part of this process beyond their testimony as to what happened. They do not get to turn around and accuse the accuser.</s>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I actually make exactly this point elsewhere in the thread.  In the end<mask><mask> we just have to use common sense, and let the over-sensitive whinge and cry.  At least, that's<mask> I typically do in my day-to-day life. [NEWLINE] [NEWLINE] (*I feel like I've made almost every point at some point in this thread.  It's huge!*)</s>
Label encoding: <s>I actually make exactly this point elsewhere in the thread.  In the end I think we just have to use common sense, and let the over-sensitive whinge and cry.  At least, that's what I typically do in my day-to-day life. [NEWLINE] [NEWLINE] (*I feel like I've made almost every point at some point in this thread.  It's huge!*)</s>
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0490, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0544, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course censure is going on. The website is provided for user for free,<mask> it's still a business and they need to protect their interest. Is Reddit even remotely *heavily censored*? I don't think<mask>. [NEWLINE] [NEWLINE] <mask> sure, me stopping you from drawing dicks in my front lawn with gasoline can constitute "censure"<mask> you have a victim complex.</s>
Label encoding: <s>Of course censure is going on. The website is provided for user for free, but it's still a business and they need to protect their interest. Is Reddit even remotely *heavily censored*? I don't think so. [NEWLINE] [NEWLINE] But sure, me stopping you from drawing dicks in my front lawn with gasoline can constitute "censure" if you have a victim complex.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0812, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0556, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0754, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I wonder<mask> this would really play out. Would men who sever these ties just never look back, never want to see or speak to the child?<mask> about the guy's parents (especially for young guys)? Do the get to pick up the mantle he puts down<mask> they want to? [NEWLINE] [NEWLINE] <mask> happens<mask> a guy who signed away his parental rights wants to start seeing that kid? </s><pad>
Label encoding: <s>I wonder how this would really play out. Would men who sever these ties just never look back, never want to see or speak to the child? How about the guy's parents (especially for young guys)? Do the get to pick up the mantle he puts down if they want to? [NEWLINE] [NEWLINE] What happens when a guy who signed away his parental rights wants to start seeing that kid? </s><pad>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Conception is always a possible outcome of sex no matter<mask> much protection/care both parties take to avoid it.  All birth control comes with a chance of failure.  Part of consensual sex is that both parties acknowledge and accept the fact that it could result in pregnancy. [NEWLINE] <mask> you roll the dice and lose, you can't forfeit your debts<mask> you overlooked one of the possible outcomes.</s>
Label encoding: <s>Conception is always a possible outcome of sex no matter how much protection/care both parties take to avoid it.  All birth control comes with a chance of failure.  Part of consensual sex is that both parties acknowledge and accept the fact that it could result in pregnancy. [NEWLINE] When you roll the dice and lose, you can't forfeit your debts because you overlooked one of the possible outcomes.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually, it could be pretty cheap. Yucca Mountain in Nevada was supposed to be the geological waste repository, and one of the reasons it was chosen is that there are lots of ways to get the fuel there by train. Construction had already started<mask> it was shut down, and it is currently the most studied site in the nation due to<mask> was going to be it's nuclear importance.</s><pad>
Label encoding: <s>Actually, it could be pretty cheap. Yucca Mountain in Nevada was supposed to be the geological waste repository, and one of the reasons it was chosen is that there are lots of ways to get the fuel there by train. Construction had already started when it was shut down, and it is currently the most studied site in the nation due to what was going to be it's nuclear importance.</s><pad>
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] "It effects me not at all,<mask> I have no need to form an opinion of any kind." [ENDQ] [NEWLINE] Slight caveat:<mask> he does want kids and has a female friend who would otherwise be an eligible partner, this would concern him only<mask><mask> it would make her ineligible.<mask> otherwise, spot on. I just have a compulsive need to nit-pick sometimes.</s>
Label encoding: <s> [STARTQ] "It effects me not at all, therefore I have no need to form an opinion of any kind." [ENDQ] [NEWLINE] Slight caveat: if he does want kids and has a female friend who would otherwise be an eligible partner, this would concern him only insofar as it would make her ineligible. But otherwise, spot on. I just have a compulsive need to nit-pick sometimes.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>How would respond to a program that would provide family assistance for up to five years<mask> after five years the assissttee has the option to stop receiving payments or to be permanently sterilized.  The benefit being that we are not forcing anyone to get sterilized<mask> we can ensure that they will not have any more children<mask> providing for the children that they already do have. </s>
Label encoding: <s>How would respond to a program that would provide family assistance for up to five years but after five years the assissttee has the option to stop receiving payments or to be permanently sterilized.  The benefit being that we are not forcing anyone to get sterilized but we can ensure that they will not have any more children while providing for the children that they already do have. </s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yea I did mean a federal ban. You raise a very good point about the slippery slope nature of bans, and I got to say I dont have a response to that. It is still discomforting to see the Confederate flag flown in a public / government sphere,<mask> I can understand the reason behind not taking action against it. I'm not totally sold<mask> I am swaying.</s>
Label encoding: <s>Yea I did mean a federal ban. You raise a very good point about the slippery slope nature of bans, and I got to say I dont have a response to that. It is still discomforting to see the Confederate flag flown in a public / government sphere, but I can understand the reason behind not taking action against it. I'm not totally sold but I am swaying.</s>
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A good test on your end would be to point out the spelling mistake and see<mask> the person handles it. That actually tests this person's ability to think on the fly and cover for a flaw, a very good skill for a salesperson to have. [ENDQ] [NEWLINE] OP might not have noticed,<mask><mask> I was ever hiring someone, I'd take that idea into consideration. Thanks!</s>
Label encoding: <s> [STARTQ] A good test on your end would be to point out the spelling mistake and see how the person handles it. That actually tests this person's ability to think on the fly and cover for a flaw, a very good skill for a salesperson to have. [ENDQ] [NEWLINE] OP might not have noticed, but if I was ever hiring someone, I'd take that idea into consideration. Thanks!</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well no. Not Pavlov's conditioned responses.<mask> science from the same field. Psychology, economics, and game theory are the sciences that should be referenced<mask> drafting laws. Laws and electoral systems should be designed<mask> that "gaming the system" to get the most out of it doesn't end up leading you into morally questionable acts, or acts that go against the spirit of the law.</s><pad>
Label encoding: <s>Well no. Not Pavlov's conditioned responses. But science from the same field. Psychology, economics, and game theory are the sciences that should be referenced when drafting laws. Laws and electoral systems should be designed so that "gaming the system" to get the most out of it doesn't end up leading you into morally questionable acts, or acts that go against the spirit of the law.</s><pad>
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0607, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I was more debating the "uncomfortable making demands of their employer" part. It wasn't<mask> much discomfort,<mask> it was "I will be fired (or alternately  just given no hours) after this<mask> they don't have much invested in me and don't have time for this"... ya know? [NEWLINE] [NEWLINE] I'm<mask> guessing there's a law protecting </s><pad><pad>
Label encoding: <s>Well I was more debating the "uncomfortable making demands of their employer" part. It wasn't so much discomfort, as it was "I will be fired (or alternately  just given no hours) after this because they don't have much invested in me and don't have time for this"... ya know? [NEWLINE] [NEWLINE] I'm also guessing there's a law protecting </s><pad><pad>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've always understood that I could tell certain things about a person by their outward appearance,<mask> I've never assumed I could determine whether or not I'd like a person based solely on the way they look. [NEWLINE] [NEWLINE] Sure I can make reasonable guesses and apply myself<mask>,<mask> I always find myself constantly surprised by the people I'd never expect. Try new things and all that.</s>
Label encoding: <s>I've always understood that I could tell certain things about a person by their outward appearance, but I've never assumed I could determine whether or not I'd like a person based solely on the way they look. [NEWLINE] [NEWLINE] Sure I can make reasonable guesses and apply myself accordingly, but I always find myself constantly surprised by the people I'd never expect. Try new things and all that.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What do you think you are doing,<mask> you let a HEAVILY corporately influenced government write regulations, other than giving control to those corporations? [NEWLINE] [NEWLINE] "A chance" to affect change through government?  A chance?  You opposition does not rely on chance, they directly affect change in government, to their favor. <mask> does a chance compare to the sure thing?</s>
Label encoding: <s>What do you think you are doing, when you let a HEAVILY corporately influenced government write regulations, other than giving control to those corporations? [NEWLINE] [NEWLINE] "A chance" to affect change through government?  A chance?  You opposition does not rely on chance, they directly affect change in government, to their favor.  How does a chance compare to the sure thing?</s>
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Public schools aren't a guarantee of an object education. You are just mad that your parents drop<mask> much money to send you to a Christian school<mask> they had a few bad teachers. [NEWLINE] [NEWLINE] Don't you think having a nice home and parents that love you is a good deal<mask> the only cost is having to put up with a few short prayers a day and a bible class?</s>
Label encoding: <s>Public schools aren't a guarantee of an object education. You are just mad that your parents drop so much money to send you to a Christian school where they had a few bad teachers. [NEWLINE] [NEWLINE] Don't you think having a nice home and parents that love you is a good deal if the only cost is having to put up with a few short prayers a day and a bible class?</s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just<mask> a person does not wish they were aborted does not mean that society would not be better off<mask> they had been aborted. [NEWLINE] [NEWLINE] Most of the people in prison do not wish they had been aborted,<mask> society certainly would be better<mask> they had been aborted. [NEWLINE] [NEWLINE] From [URL] [NEWLINE] [NEWLINE] &gt; Arrow<mask> said 80 percent of the prison population once was in foster care</s>
Label encoding: <s>Just because a person does not wish they were aborted does not mean that society would not be better off if they had been aborted. [NEWLINE] [NEWLINE] Most of the people in prison do not wish they had been aborted, but society certainly would be better if they had been aborted. [NEWLINE] [NEWLINE] From [URL] [NEWLINE] [NEWLINE] &gt; Arrow also said 80 percent of the prison population once was in foster care</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like this one. You said<mask> a lot of other people said,<mask> you said it really well.<mask><mask> it was actually your last paragraph at the bottom that swayed me. Sometimes it's easy to get caught up in<mask> seems fair or unfair to one personally,<mask> that evaluation of fairness is directly contradicted<mask> you look at the big picture. [NEWLINE] [NEWLINE] [NEWLINE] ∆ </s>
Label encoding: <s>I like this one. You said what a lot of other people said, but you said it really well. I think it was actually your last paragraph at the bottom that swayed me. Sometimes it's easy to get caught up in what seems fair or unfair to one personally, when that evaluation of fairness is directly contradicted when you look at the big picture. [NEWLINE] [NEWLINE] [NEWLINE] ∆ </s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0495, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From<mask> I have read, women make less money primarily<mask> they are employed in lower paying jobs, work less hours or take time off to raise children. The problem is a cultural one not an economic one. From an early age girls and boys are taught very different values. This leads to  choosing different fields of employment or taking a role<mask> a mother rather than a worker.</s>
Label encoding: <s>From what I have read, women make less money primarily because they are employed in lower paying jobs, work less hours or take time off to raise children. The problem is a cultural one not an economic one. From an early age girls and boys are taught very different values. This leads to  choosing different fields of employment or taking a role as a mother rather than a worker.</s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree there are nuances to everything.  Perhaps my point wasn't clear - that bill, which we both agree served a good purpose, and at least did quite a good job in serving its purpose - is tiny.  It's something like 15-20 pages long.  This is a far cry from a multi-thousand page bill.  That was my point.</s>
Label encoding: <s>I agree there are nuances to everything.  Perhaps my point wasn't clear - that bill, which we both agree served a good purpose, and at least did quite a good job in serving its purpose - is tiny.  It's something like 15-20 pages long.  This is a far cry from a multi-thousand page bill.  That was my point.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask><mask>, I don't buy HAES. I believe that anyone of any size can be healthier at that size.<mask> I start jogging once a week, I probably won't lose weight,<mask> I'll be a teeny-tiny bit healthier and that's good. [ENDQ] [NEWLINE] That's<mask> HAES means. <mask> don't you buy about it?</s>
Label encoding: <s> [STARTQ] First of all, I don't buy HAES. I believe that anyone of any size can be healthier at that size. If I start jogging once a week, I probably won't lose weight, but I'll be a teeny-tiny bit healthier and that's good. [ENDQ] [NEWLINE] That's what HAES means.  What don't you buy about it?</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Many schools (at least in my area) are moving to a more consistent year-round school year. They shortened summer break by about half and redistributed the time into a fall break, more 3 day weekends, days off for less important holidays, etc. This (is intended to) prevent students from just forgetting everything over the summer, allowing schooling to be more continuous. </s>
Label encoding: <s>Many schools (at least in my area) are moving to a more consistent year-round school year. They shortened summer break by about half and redistributed the time into a fall break, more 3 day weekends, days off for less important holidays, etc. This (is intended to) prevent students from just forgetting everything over the summer, allowing schooling to be more continuous. </s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If anything I would think multiple private charities would be less efficient, especially<mask> it comes to rooting out fraud. <mask> one denies you just go exploit another, or even several. [NEWLINE] Anyway, wouldn't this allow the government to spend even more on the military? <mask> you think welfare is good you should want it increased at the expense of the bad things like war.</s>
Label encoding: <s>If anything I would think multiple private charities would be less efficient, especially when it comes to rooting out fraud.  If one denies you just go exploit another, or even several. [NEWLINE] Anyway, wouldn't this allow the government to spend even more on the military?  If you think welfare is good you should want it increased at the expense of the bad things like war.</s>
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Based on<mask> I've learned in my political science classes many years ago, the way the UK handles this (among other countries--and please correct me<mask> I'm wrong) is with public financing of elections.  Each candidate who meets a minimum threshold of support gets free, equal airtime paid for by the government. <mask><mask> outside contributions are disallowed.  </s>
Label encoding: <s>Based on what I've learned in my political science classes many years ago, the way the UK handles this (among other countries--and please correct me if I'm wrong) is with public financing of elections.  Each candidate who meets a minimum threshold of support gets free, equal airtime paid for by the government.  I think outside contributions are disallowed.  </s>
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] personhood is defined<mask> [ENDQ] [NEWLINE] Defined by whom? [NEWLINE] [NEWLINE] I am sorry,<mask> you are going to have a losing proposition<mask> you wish to claim that a rat is a person. [NEWLINE] [NEWLINE] Beyond that, you have not explained<mask> the destruction of property is not violence.<mask> I smash your car with a baseball bat, is that an act of violence?</s>
Label encoding: <s> [STARTQ] personhood is defined as [ENDQ] [NEWLINE] Defined by whom? [NEWLINE] [NEWLINE] I am sorry, but you are going to have a losing proposition if you wish to claim that a rat is a person. [NEWLINE] [NEWLINE] Beyond that, you have not explained how the destruction of property is not violence. If I smash your car with a baseball bat, is that an act of violence?</s>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not all political systems are equal and democracy is more than just having a vote. Subsidiarity and manner of representation are important.The UK practices a very top heavy, centralized style of Government (especially in England). It<mask> has a voting system that encourages majority Governments with not necessarily much vote share (the last Labour Government had only 35% of the vote). [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Not all political systems are equal and democracy is more than just having a vote. Subsidiarity and manner of representation are important.The UK practices a very top heavy, centralized style of Government (especially in England). It also has a voting system that encourages majority Governments with not necessarily much vote share (the last Labour Government had only 35% of the vote). [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Rule#2, Dude. [NEWLINE] [NEWLINE] No need to make a comparison using someone's reddit activity in a way that can be construed<mask> an insult. Even<mask> you didn't mean it<mask> one you should be aware that it comes off<mask> hostile any time you pass personal judgement on a stranger. [NEWLINE] [NEWLINE] Save it for subreddits<mask> being rude isn't against the rules. </s>
Label encoding: <s>Rule#2, Dude. [NEWLINE] [NEWLINE] No need to make a comparison using someone's reddit activity in a way that can be construed as an insult. Even if you didn't mean it as one you should be aware that it comes off as hostile any time you pass personal judgement on a stranger. [NEWLINE] [NEWLINE] Save it for subreddits where being rude isn't against the rules. </s>
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Then take away the "asexual" way of thinking about it. [NEWLINE] [NEWLINE] Let's say you,<mask> a parent, could have the option to guarantee that your child would get married and have children. [NEWLINE] [NEWLINE] Do you think all parents would take that option? Or would some "let go" and believe that perhaps some people are better destined for a meaningful single life?</s>
Label encoding: <s>Then take away the "asexual" way of thinking about it. [NEWLINE] [NEWLINE] Let's say you, as a parent, could have the option to guarantee that your child would get married and have children. [NEWLINE] [NEWLINE] Do you think all parents would take that option? Or would some "let go" and believe that perhaps some people are better destined for a meaningful single life?</s>
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for the personal experience in your reply.<mask><mask> that everyone has a job to do and no company will work without a good team. this is such a difficult argument<mask> we are all<mask> used to this pyramid structure<mask> the people at the top gets paid the most. Do you think that the Company would collapse<mask> you had to leave or perhaps<mask> the CEO quit?</s>
Label encoding: <s>Thanks for the personal experience in your reply. I think that everyone has a job to do and no company will work without a good team. this is such a difficult argument because we are all so used to this pyramid structure where the people at the top gets paid the most. Do you think that the Company would collapse if you had to leave or perhaps if the CEO quit?</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Do you have anything to show for these mandatory arrest laws that show that cops arrest men, not<mask> they are the primary aggressor,<mask><mask> of discrimination?<mask> not then don't bother bringing up the point unless you have facts to back it up. That just shows that men are arrested<mask> they call the police and that a large percent of times no one is arrested.</s>
Label encoding: <s>Do you have anything to show for these mandatory arrest laws that show that cops arrest men, not because they are the primary aggressor, but because of discrimination? If not then don't bother bringing up the point unless you have facts to back it up. That just shows that men are arrested when they call the police and that a large percent of times no one is arrested.</s>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0528, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is the best I could find: [URL].full [NEWLINE] [NEWLINE] It doesn't deal directly with accidentally killing a family member<mask> you think its a home invader<mask> instead the more general case of increased risk of death<mask> guns are in the home. I'm on mobile right now<mask><mask> I get a chance I'll try to find a more directly relevant source. </s>
Label encoding: <s>This is the best I could find: [URL].full [NEWLINE] [NEWLINE] It doesn't deal directly with accidentally killing a family member when you think its a home invader but instead the more general case of increased risk of death when guns are in the home. I'm on mobile right now so when I get a chance I'll try to find a more directly relevant source. </s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well you didn't change my view,<mask><mask><mask> you have the best answer. I really want to award you the delta,<mask> I'm not sure<mask> I can. [NEWLINE] [NEWLINE] I guess you helped me recognize that there isn't a one true king of operating systems, which would be a change of view. [NEWLINE] [NEWLINE] edit: ∆ Hope that works. </s>
Label encoding: <s>Well you didn't change my view, but I think you have the best answer. I really want to award you the delta, but I'm not sure if I can. [NEWLINE] [NEWLINE] I guess you helped me recognize that there isn't a one true king of operating systems, which would be a change of view. [NEWLINE] [NEWLINE] edit: ∆ Hope that works. </s>
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My aunt is doing that. She has one biological daughter and one permanent foster son. Both kids are happy and loved equally, too. [NEWLINE] [NEWLINE] <mask><mask><mask><mask><mask> it is an implausible scenario,<mask> a least with foster children, please be aware that the biological parents are not out of the picture and that you might have to work with them to a degree.</s>
Label encoding: <s>My aunt is doing that. She has one biological daughter and one permanent foster son. Both kids are happy and loved equally, too. [NEWLINE] [NEWLINE] So I do not think it is an implausible scenario, but a least with foster children, please be aware that the biological parents are not out of the picture and that you might have to work with them to a degree.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My point was not that hedonism could lead to negative (by hedonism) consequences,<mask> that hedonism--which values only pleasure--says that there is basically nothing wrong with keeping someone<mask> your slave<mask><mask><mask> you make sure that they are constantly in a state of primal pleasure.  It essentially gives not value to freedom or bodily autonomy.</s>
Label encoding: <s>My point was not that hedonism could lead to negative (by hedonism) consequences, but that hedonism--which values only pleasure--says that there is basically nothing wrong with keeping someone as your slave as long as you make sure that they are constantly in a state of primal pleasure.  It essentially gives not value to freedom or bodily autonomy.</s>
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Surely it depends on the values of the group? Some groups are very inclusive and to be socially popular in those groups you have to be nice to outsiders who aren't in your group. [NEWLINE] [NEWLINE] For example "I am proud of being British,<mask> we british people are<mask> nice to people outside of our group." is a statement you could make.</s>
Label encoding: <s>Surely it depends on the values of the group? Some groups are very inclusive and to be socially popular in those groups you have to be nice to outsiders who aren't in your group. [NEWLINE] [NEWLINE] For example "I am proud of being British, because we british people are so nice to people outside of our group." is a statement you could make.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe the focus on just meat-products is a reflection of a vegans own personal affection towards animals [ENDQ] [NEWLINE] I believe that your pointing out migrant works who pick vegetables is a reflection of your own biases against veganism.  There are migrant workers milking cows and slaughtering chickens<mask> well,<mask> they don't seem to factor into your thought process.</s>
Label encoding: <s> [STARTQ] I believe the focus on just meat-products is a reflection of a vegans own personal affection towards animals [ENDQ] [NEWLINE] I believe that your pointing out migrant works who pick vegetables is a reflection of your own biases against veganism.  There are migrant workers milking cows and slaughtering chickens as well, but they don't seem to factor into your thought process.</s>
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.6726e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0582, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0706, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0454, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] a search term picked up from the day you were experimenting with fetish porn may be suggestion to a script of<mask> interests you [ENDQ] [NEWLINE] And the more ads become targeted, the more obvious it will be to people who see the ads you get<mask> you have been looking at previously.  And imagine<mask> this becomes prevalent off the Internet and in the real world.</s>
Label encoding: <s> [STARTQ] a search term picked up from the day you were experimenting with fetish porn may be suggestion to a script of what interests you [ENDQ] [NEWLINE] And the more ads become targeted, the more obvious it will be to people who see the ads you get what you have been looking at previously.  And imagine when this becomes prevalent off the Internet and in the real world.</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0734, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Would you consider having food to eat is a human right? Or clean drinking water?<mask> these are<mask> services people provide, the bakers that bake the ingredients that farmers farm and the employees at water treatment plants making sure the water coming out of your taps are clean. They're<mask> not doing it for free<mask> it's not really forcing is it? </s><pad>
Label encoding: <s>Would you consider having food to eat is a human right? Or clean drinking water? Because these are also services people provide, the bakers that bake the ingredients that farmers farm and the employees at water treatment plants making sure the water coming out of your taps are clean. They're also not doing it for free so it's not really forcing is it? </s><pad>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Zelda II isn't really an RPG. [NEWLINE] [NEWLINE] And I don't play Pokemon,<mask> it is a handheld game that doesn't change much from game to game. [NEWLINE] [NEWLINE] My point is, you couldn't have a game like Dark Souls twenty years ago. [NEWLINE] [NEWLINE] Chrono Trigger has real-time battles<mask> they're basically still random encounters.</s>
Label encoding: <s>Zelda II isn't really an RPG. [NEWLINE] [NEWLINE] And I don't play Pokemon, but it is a handheld game that doesn't change much from game to game. [NEWLINE] [NEWLINE] My point is, you couldn't have a game like Dark Souls twenty years ago. [NEWLINE] [NEWLINE] Chrono Trigger has real-time battles but they're basically still random encounters.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.1062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But it is a delusion, I'd argue delusions are harmful by their nature<mask> they corrupt the decision making process leading to harmful behaviour. [NEWLINE] [NEWLINE] People without delusions can make decisions that cost lives like deciding to drive home after drinking.<mask> a completely sane person can decide on such harmful behaviour then delusions being added to the mix cant be dismissed<mask> neutral effect.</s>
Label encoding: <s>But it is a delusion, I'd argue delusions are harmful by their nature as they corrupt the decision making process leading to harmful behaviour. [NEWLINE] [NEWLINE] People without delusions can make decisions that cost lives like deciding to drive home after drinking. If a completely sane person can decide on such harmful behaviour then delusions being added to the mix cant be dismissed as neutral effect.</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>an object's cost can be computed.  the raw materials and manufacturing that assembled it cost X, the labor cost Y, shipping and packaging Z, marketing W. [NEWLINE] [NEWLINE] <mask><mask> much is it worth? <mask> much something costs to produce isn't a measure of<mask> it's worth. it's worth whatever people are willing to pay. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>an object's cost can be computed.  the raw materials and manufacturing that assembled it cost X, the labor cost Y, shipping and packaging Z, marketing W. [NEWLINE] [NEWLINE] So how much is it worth?  how much something costs to produce isn't a measure of what it's worth. it's worth whatever people are willing to pay. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A sensitivity to religious and or culture practice certainly makes sense and I'm ashamed to say I hadn't even considered that angle. Thanks for the contribution! [NEWLINE] [NEWLINE] EDIT: To further clarify this makes sense at institution (university) who is trying to attract the best and brightest minds<mask><mask> religion and culture. [NEWLINE] [NEWLINE] &amp;#8710;</s>
Label encoding: <s>A sensitivity to religious and or culture practice certainly makes sense and I'm ashamed to say I hadn't even considered that angle. Thanks for the contribution! [NEWLINE] [NEWLINE] EDIT: To further clarify this makes sense at institution (university) who is trying to attract the best and brightest minds regardless of religion and culture. [NEWLINE] [NEWLINE] &amp;#8710;</s>
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0590, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are worried you will wake up one morning with no ability to buy food for a single day? Do you have an emergency fund in cash in a savings account that is FDIC insured? I'm not sure<mask> you (I'm assuming American) would need more than that between the first sign of economic trouble and being able to access foreign investments. </s>
Label encoding: <s>You are worried you will wake up one morning with no ability to buy food for a single day? Do you have an emergency fund in cash in a savings account that is FDIC insured? I'm not sure why you (I'm assuming American) would need more than that between the first sign of economic trouble and being able to access foreign investments. </s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, that's true,<mask> at the core of Christianity is that they believe in a god and that there was a messiah who taught the message of that god. The details do vary enormously,<mask> there is something of substance upon which their beleif system is built,<mask> ''atheism'' is simply a lack of belief in any gods.</s>
Label encoding: <s>Yes, that's true, but at the core of Christianity is that they believe in a god and that there was a messiah who taught the message of that god. The details do vary enormously, but there is something of substance upon which their beleif system is built, while ''atheism'' is simply a lack of belief in any gods.</s>
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It will become corrupt. Politicians are very good these days at figuring out who will vote for them based on answers to questions. It's impossible to determine whether the questions are loaded without access to the original polling research. This research is kept confidential by the polling firms for their customers.<mask> it would be practically impossible to catch them manipulating the test.</s>
Label encoding: <s>It will become corrupt. Politicians are very good these days at figuring out who will vote for them based on answers to questions. It's impossible to determine whether the questions are loaded without access to the original polling research. This research is kept confidential by the polling firms for their customers. So it would be practically impossible to catch them manipulating the test.</s>
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0547, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually, he did have that option<mask> he used it.<mask> Zimmerman was on the phone, he had lost track of<mask> Martin was. Martin hid himself from Zimmerman and then chose to reveal himself again. No reasonable person in fear of their lives (which is<mask>'s required to use force) would come out of hiding to their pursuer. </s>
Label encoding: <s>Actually, he did have that option because he used it. While Zimmerman was on the phone, he had lost track of where Martin was. Martin hid himself from Zimmerman and then chose to reveal himself again. No reasonable person in fear of their lives (which is what's required to use force) would come out of hiding to their pursuer. </s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One important difference is China has compulsory service whereas the US does not.  For that reason china has a large number of military personnel who are not career soldiers and are not needed to maintain readiness.  That being said<mask><mask> that the us military should reduce its funding and put more money into public works.  That<mask> is a different issue.</s>
Label encoding: <s>One important difference is China has compulsory service whereas the US does not.  For that reason china has a large number of military personnel who are not career soldiers and are not needed to maintain readiness.  That being said I think that the us military should reduce its funding and put more money into public works.  That however is a different issue.</s>
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.0735e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.8447e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.2006e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is mostly only rookies who fall<mask> clipped in on their road bike.<mask> you have to kind of twist your foot around to get it to disengage and<mask> you aren't used to it or are really tired you might fall.<mask> the consequence of falling is usually just embarrassment<mask><mask> there is a car next to you it can be scary</s>
Label encoding: <s>It is mostly only rookies who fall when clipped in on their road bike. But you have to kind of twist your foot around to get it to disengage and if you aren't used to it or are really tired you might fall. Also the consequence of falling is usually just embarrassment but if there is a car next to you it can be scary</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0477, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No one from a third world country should immigrate to first world countries. These countries happen to be mostly white, and white people are pretty racist.<mask><mask>, from Eastern Europe to White countries shouldn't be an issue,<mask><mask> your skin is not a shade of white, you're pretty much fucked in the ass in a white country.</s>
Label encoding: <s>No one from a third world country should immigrate to first world countries. These countries happen to be mostly white, and white people are pretty racist. I think, from Eastern Europe to White countries shouldn't be an issue, but if your skin is not a shade of white, you're pretty much fucked in the ass in a white country.</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] rights are still infringed. [ENDQ] [NEWLINE] sure, never denied that (<mask> my reference to violinist expierement). i'm saying that very few people would agree with the violinist<mask> they accepted the personhood claim due to different rights being more fundamental [NEWLINE] [NEWLINE] i mean sure all laws dictate our personal affairs by this argument. </s><pad>
Label encoding: <s> [STARTQ] rights are still infringed. [ENDQ] [NEWLINE] sure, never denied that ( hence my reference to violinist expierement). i'm saying that very few people would agree with the violinist if they accepted the personhood claim due to different rights being more fundamental [NEWLINE] [NEWLINE] i mean sure all laws dictate our personal affairs by this argument. </s><pad>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.4586e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0564, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It is to these guys. [NEWLINE] [NEWLINE] [URL].com/dictionary/racism [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] You are right. I was not trying to change your view. I was trying to say that your view was not<mask> you said it was. [NEWLINE] [NEWLINE] <mask> you mean no foul, choose your words more carefully next time. </s>
Label encoding: <s>It is to these guys. [NEWLINE] [NEWLINE] [URL].com/dictionary/racism [NEWLINE] [NEWLINE] [URL] [NEWLINE] [NEWLINE] You are right. I was not trying to change your view. I was trying to say that your view was not what you said it was. [NEWLINE] [NEWLINE] If you mean no foul, choose your words more carefully next time. </s>
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If that is true, then you can<mask> say that: [NEWLINE] Eating Pork and Christianity is incompatible [NEWLINE] Getting Tattoos and Christianity is incompatible [NEWLINE] Eating Shrimp and Christianity is incompatible [NEWLINE]... [NEWLINE] [NEWLINE] The list goes ON and ON. [NEWLINE] [NEWLINE] Not that many christians have any problem doing all of the above<mask> still cursing homosexuality.</s>
Label encoding: <s>If that is true, then you can also say that: [NEWLINE] Eating Pork and Christianity is incompatible [NEWLINE] Getting Tattoos and Christianity is incompatible [NEWLINE] Eating Shrimp and Christianity is incompatible [NEWLINE]... [NEWLINE] [NEWLINE] The list goes ON and ON. [NEWLINE] [NEWLINE] Not that many christians have any problem doing all of the above while still cursing homosexuality.</s>
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Yeah, especially<mask> a lot of the early sciences came out of church monks. [ENDQ] [NEWLINE] Who I would argue were basically atheists. [NEWLINE] [NEWLINE] [STARTQ] It's a false dichotomy propagated by r/atheism. [ENDQ] [NEWLINE] That was then, this is now. In 2015, believing in magic isn't a rational position. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Yeah, especially since a lot of the early sciences came out of church monks. [ENDQ] [NEWLINE] Who I would argue were basically atheists. [NEWLINE] [NEWLINE] [STARTQ] It's a false dichotomy propagated by r/atheism. [ENDQ] [NEWLINE] That was then, this is now. In 2015, believing in magic isn't a rational position. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>P.C. sucks, it either makes us all conforming little pussies or make us blind/oblivious to the real suffering out there in the world. The truth needs to be told, one way or another, there is nothing to gain from a sanitized society, it's just denying/lying about who we are.</s>
Label encoding: <s>P.C. sucks, it either makes us all conforming little pussies or make us blind/oblivious to the real suffering out there in the world. The truth needs to be told, one way or another, there is nothing to gain from a sanitized society, it's just denying/lying about who we are.</s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The website is about discussion. [ENDQ] [NEWLINE] I actually have to disagree with you on this. The root of this website was not discussion! It is a news aggregator site that relied (and still relies) on social dynamic for it's content. [NEWLINE] [NEWLINE] The discussion about the news was a (maybe even welcomed) side effect.</s>
Label encoding: <s> [STARTQ] The website is about discussion. [ENDQ] [NEWLINE] I actually have to disagree with you on this. The root of this website was not discussion! It is a news aggregator site that relied (and still relies) on social dynamic for it's content. [NEWLINE] [NEWLINE] The discussion about the news was a (maybe even welcomed) side effect.</s>
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Immortality experienced<mask> an ordinary human being wouldn't be boring anyway. [NEWLINE] [NEWLINE] Our brains wouldn't be able retain a billion years of memories or perhaps even comprehend a billion years.  You would be able to do the same thing over and over and it would not be boring<mask> you wouldn't remember the last time you did it.</s>
Label encoding: <s>Immortality experienced as an ordinary human being wouldn't be boring anyway. [NEWLINE] [NEWLINE] Our brains wouldn't be able retain a billion years of memories or perhaps even comprehend a billion years.  You would be able to do the same thing over and over and it would not be boring because you wouldn't remember the last time you did it.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>.07%... [NEWLINE] [NEWLINE] You  are expecting me to believe that every time 10,000 toddlers spend a minute walking through the parking lot with their parents, 7 of them ddon't make it back alive? [NEWLINE] [NEWLINE] This is<mask> I'm taking about. Your perception of the danger is orders of magnitude higher than the danger.</s>
Label encoding: <s>.07%... [NEWLINE] [NEWLINE] You  are expecting me to believe that every time 10,000 toddlers spend a minute walking through the parking lot with their parents, 7 of them ddon't make it back alive? [NEWLINE] [NEWLINE] This is what I'm taking about. Your perception of the danger is orders of magnitude higher than the danger.</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think it might be a little narrow to insist that any evidence be purely race based.  Culture has a lot to do with identification, and a lot of that is intrinsically linked to skin color.  Identity doesn't form in a cultural vacuum, and I believe it's unreasonable to think that bigotry could possibly stand unrelated to culture.</s>
Label encoding: <s>I think it might be a little narrow to insist that any evidence be purely race based.  Culture has a lot to do with identification, and a lot of that is intrinsically linked to skin color.  Identity doesn't form in a cultural vacuum, and I believe it's unreasonable to think that bigotry could possibly stand unrelated to culture.</s>
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(9.7800e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.9554e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is<mask> true. In my house for example we have Toy story 3 on Blu Ray,<mask> a dvd quality version ripped on the media server, and sometimes access to it via my Netflix subscription. Guess which is the least used option! [NEWLINE] [NEWLINE] Menus suck. Way to disincentive a legal purchase Hollywood. </s>
Label encoding: <s>This is so true. In my house for example we have Toy story 3 on Blu Ray, also a dvd quality version ripped on the media server, and sometimes access to it via my Netflix subscription. Guess which is the least used option! [NEWLINE] [NEWLINE] Menus suck. Way to disincentive a legal purchase Hollywood. </s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I like the way you've taken time to look at the evidence. You've found some good stuff! [NEWLINE] [NEWLINE] Isn't this the older study that more recent studies refer to? (it's 50 years old).<mask><mask> the differences it found (in only male drivers, not female) haven't been found in later studies?</s>
Label encoding: <s>I like the way you've taken time to look at the evidence. You've found some good stuff! [NEWLINE] [NEWLINE] Isn't this the older study that more recent studies refer to? (it's 50 years old). I think the differences it found (in only male drivers, not female) haven't been found in later studies?</s>
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with it.  My measure is work done, not people employed.  One person operating 5 robots is doing a lot more work than 3 people not operating robots. <mask> assuming the retiring person is doing some nonzero amount of productive work, there will be less work done by their retiring than<mask> they didn't.</s>
Label encoding: <s>I agree with it.  My measure is work done, not people employed.  One person operating 5 robots is doing a lot more work than 3 people not operating robots.  But assuming the retiring person is doing some nonzero amount of productive work, there will be less work done by their retiring than if they didn't.</s>
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Lentils, cooked spinach and other dark greens (<mask> you can't absorb<mask> much from raw), tofu, quinoa, cereal, etc. [NEWLINE] [NEWLINE] A lot of oranges and other fruits<mask> vitamin C is supposed to help you metabolize iron. [NEWLINE] [NEWLINE] I can't remember everything, it's been about seven years.</s>
Label encoding: <s>Lentils, cooked spinach and other dark greens ( as you can't absorb as much from raw), tofu, quinoa, cereal, etc. [NEWLINE] [NEWLINE] A lot of oranges and other fruits because vitamin C is supposed to help you metabolize iron. [NEWLINE] [NEWLINE] I can't remember everything, it's been about seven years.</s>
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That seems either untrue or non-falsifiable.  Some people do make changes to their behavior and turn their life around.  That directly refutes your statement. <mask> the response is that the fact that they changed means they had a different nature to begin with, then<mask> could your statement ever be false?  </s>
Label encoding: <s>That seems either untrue or non-falsifiable.  Some people do make changes to their behavior and turn their life around.  That directly refutes your statement.  If the response is that the fact that they changed means they had a different nature to begin with, then how could your statement ever be false?  </s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Actually<mask><mask> with the first line. You treat this hip hop<mask> a religion and you people have this religious fervor over the nonsense that Kanye spews out.<mask> would Kanye West know about racial injustice<mask><mask> he was caught with "30 rocks" he wouldn't get anytime at all<mask> he has afluenza. </s>
Label encoding: <s>Actually I agree with the first line. You treat this hip hop as a religion and you people have this religious fervor over the nonsense that Kanye spews out. What would Kanye West know about racial injustice when if he was caught with "30 rocks" he wouldn't get anytime at all because he has afluenza. </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's too late for gun control. I am of the belief that it would be better<mask> nobody had guns in America.<mask>, those who think that gun control measures would have any significant effect on crime are simply delusional simply<mask> gun culture has been<mask> engrained in our society and too many are already in circulation. </s>
Label encoding: <s>It's too late for gun control. I am of the belief that it would be better if nobody had guns in America. However, those who think that gun control measures would have any significant effect on crime are simply delusional simply because gun culture has been so engrained in our society and too many are already in circulation. </s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] All factors being the same it boils down to who applies and<mask> skills and experience are needed for the job. [ENDQ] [NEWLINE] Yes. <mask> all factors were the same, that's<mask> it would come down to. <mask> women didn't carry the added cost of maternity, that's<mask> it would come down to.</s>
Label encoding: <s> [STARTQ] All factors being the same it boils down to who applies and what skills and experience are needed for the job. [ENDQ] [NEWLINE] Yes.  If all factors were the same, that's what it would come down to.  If women didn't carry the added cost of maternity, that's what it would come down to.</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People don't get SRS to conform or to have people view them differently. They get it<mask> they can feel whole. Look the concept of being trans and the dysphoria and discomfort that goes with it is difficult to understand without being trans,<mask> seriously,  they do it to feel whole and comfortable with their body.</s>
Label encoding: <s>People don't get SRS to conform or to have people view them differently. They get it so they can feel whole. Look the concept of being trans and the dysphoria and discomfort that goes with it is difficult to understand without being trans, but seriously,  they do it to feel whole and comfortable with their body.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I agree with you, oldness does usually bring wisdom and mental stability. [NEWLINE] [NEWLINE] I can see<mask> a 18 or 25 year-old (or anyone, for that matter) appreciates those qualities in a partner<mask><mask> aren't men usually seeking for a wise and mentally stable (i.e. older) partner?</s>
Label encoding: <s>I agree with you, oldness does usually bring wisdom and mental stability. [NEWLINE] [NEWLINE] I can see how a 18 or 25 year-old (or anyone, for that matter) appreciates those qualities in a partner but why aren't men usually seeking for a wise and mentally stable (i.e. older) partner?</s>
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What's wrong with rappers? [NEWLINE] [NEWLINE] Just kidding.  I still don't think that's a "good reason" to kill someone.  I guess we just disagree? [NEWLINE] [NEWLINE] Edit:  Rappers and Rapists are not the same thing.  One is a joke, and the other is no laughing matter.</s>
Label encoding: <s>What's wrong with rappers? [NEWLINE] [NEWLINE] Just kidding.  I still don't think that's a "good reason" to kill someone.  I guess we just disagree? [NEWLINE] [NEWLINE] Edit:  Rappers and Rapists are not the same thing.  One is a joke, and the other is no laughing matter.</s>
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> the deposit laws are for New York,<mask><mask> I live there is a deposit on various single use bottles/cans aside from alcohol.<mask> I go to the store and buy a 12 pack of soda or a case of bottled water I pay a deposit on all of those bottles/cans.</s>
Label encoding: <s>I don't know what the deposit laws are for New York, but where I live there is a deposit on various single use bottles/cans aside from alcohol. If I go to the store and buy a 12 pack of soda or a case of bottled water I pay a deposit on all of those bottles/cans.</s>
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>eh, i don't think "we can't prove we lack evidence for x existing" works in this way:<mask> we can't trust our senses than that's just more evidence that we lack evidence for x existing. Uncertainty is evidence we lack evidence not evidence we lack evidence that we lack evidence of stuff. </s>
Label encoding: <s>eh, i don't think "we can't prove we lack evidence for x existing" works in this way: if we can't trust our senses than that's just more evidence that we lack evidence for x existing. Uncertainty is evidence we lack evidence not evidence we lack evidence that we lack evidence of stuff. </s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Might I ask<mask> this citizenship-test consists of? [NEWLINE] [NEWLINE] In my high-school class (the highest level, system is much like German one) we all took the Dutch version, and I believe 2 of us passed it. [NEWLINE] [NEWLINE] Making a test about a country that isn't absolute shit is very difficult.</s>
Label encoding: <s>Might I ask what this citizenship-test consists of? [NEWLINE] [NEWLINE] In my high-school class (the highest level, system is much like German one) we all took the Dutch version, and I believe 2 of us passed it. [NEWLINE] [NEWLINE] Making a test about a country that isn't absolute shit is very difficult.</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There isn't a difference, most people focus on rape culture<mask> it applies to woman<mask> that doesn't mean it is a gendered phenomenon. [NEWLINE] [NEWLINE] My whole point is that its<mask> real that people openly joke about prison rape.<mask> we didn't have rape culture it wouldn't be culturally appropriate to do that.</s>
Label encoding: <s>There isn't a difference, most people focus on rape culture as it applies to woman but that doesn't mean it is a gendered phenomenon. [NEWLINE] [NEWLINE] My whole point is that its so real that people openly joke about prison rape. If we didn't have rape culture it wouldn't be culturally appropriate to do that.</s>
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0770, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0430, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You missed the one "main beef" that OP pointed out that doesn't apply to necklaces, bracelets, earrings, etc. [NEWLINE] [NEWLINE] &gt; My main beef with fashion glasses is that they take the form of a functional piece of equipment and remove the utility<mask> don't add anything new.</s>
Label encoding: <s>You missed the one "main beef" that OP pointed out that doesn't apply to necklaces, bracelets, earrings, etc. [NEWLINE] [NEWLINE] &gt; My main beef with fashion glasses is that they take the form of a functional piece of equipment and remove the utility but don't add anything new.</s>
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0579, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0458, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What makes affirmative action better than racial(or sectarian or sexual) discrimination<mask> the end result is the same? One demographic is actively discriminated against due to a feature inherent to them - their sexuality, skin colour, parentage w/e. It's the same thing just towards a majority<mask> opposed to a minority.</s>
Label encoding: <s>What makes affirmative action better than racial(or sectarian or sexual) discrimination if the end result is the same? One demographic is actively discriminated against due to a feature inherent to them - their sexuality, skin colour, parentage w/e. It's the same thing just towards a majority as opposed to a minority.</s>
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some awkwardly worded sentences are going to happen,<mask> resumes are typically written in English instead of your native language. The same thing<mask> applies to academic journals etc. even at the top levels. Rejecting the most (over) competent people in the world<mask> of a mistake or two is just petty. </s>
Label encoding: <s>Some awkwardly worded sentences are going to happen, because resumes are typically written in English instead of your native language. The same thing also applies to academic journals etc. even at the top levels. Rejecting the most (over) competent people in the world because of a mistake or two is just petty. </s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't even know<mask> I my local football team has a womens team.<mask> you're fighting against guys who are (were :() playing in the Premier League, literally the best football league in the whole world, then it's a really tough ask to get noticed, and have people follow you instead.</s>
Label encoding: <s>I don't even know if I my local football team has a womens team. When you're fighting against guys who are (were :() playing in the Premier League, literally the best football league in the whole world, then it's a really tough ask to get noticed, and have people follow you instead.</s>
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you don't add anything to society, do you deserve anything back? That isn't my view<mask> Devil's advocate is important here. Your morality is dictating your policy beliefs, which is fine,<mask> realize others have different views, and appeals to logic can't always change a person's moral ideals.</s>
Label encoding: <s>If you don't add anything to society, do you deserve anything back? That isn't my view but Devil's advocate is important here. Your morality is dictating your policy beliefs, which is fine, but realize others have different views, and appeals to logic can't always change a person's moral ideals.</s>
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>From a semantic point of view it would make absolutely no difference,<mask> the A in atheism already stands for Anti. I know that the patron saint of Atheism, Christopher Hitchens used to seperate the two, and I get his point,<mask> from a linguistic point of view it's the same word.</s><pad>
Label encoding: <s>From a semantic point of view it would make absolutely no difference, since the A in atheism already stands for Anti. I know that the patron saint of Atheism, Christopher Hitchens used to seperate the two, and I get his point, but from a linguistic point of view it's the same word.</s><pad>
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Almost all modern democracies have done a better job of protecting peoples freedoms that other modern governments. It<mask> seems to me that in most respects modern democracies have grown freer over time, and that many of the strides towards freedom taht took place during the course of the 20th century came from democratic movements.</s>
Label encoding: <s>Almost all modern democracies have done a better job of protecting peoples freedoms that other modern governments. It also seems to me that in most respects modern democracies have grown freer over time, and that many of the strides towards freedom taht took place during the course of the 20th century came from democratic movements.</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have read your other posts. You seem to be going with the assumption that everyone who pirates wouldn't have paid<mask> they didn't pirate?<mask><mask>,<mask> someone truly would have paid nothing, then there is no immorality in pirating something.<mask> there are many who would have paid. </s>
Label encoding: <s>I have read your other posts. You seem to be going with the assumption that everyone who pirates wouldn't have paid if they didn't pirate? I agree, if someone truly would have paid nothing, then there is no immorality in pirating something. But there are many who would have paid. </s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a paramedic I come into contact with addicts frequently. The problem is that for those who have nothing<mask> the high, they have nothing to lose except that high. Until we<mask> a society find a way to empower these people to have something<mask> the high, that is<mask> they will seek out. </s>
Label encoding: <s>As a paramedic I come into contact with addicts frequently. The problem is that for those who have nothing but the high, they have nothing to lose except that high. Until we as a society find a way to empower these people to have something besides the high, that is what they will seek out. </s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've visited. It's definitely a great tourist destination.<mask> I find it hard to believe that it's the best place to live in. [NEWLINE] [NEWLINE] There just seems to be a huge lack of support in comparison to other first-world countries that people don't know about in countries with that support system.</s>
Label encoding: <s>I've visited. It's definitely a great tourist destination. But I find it hard to believe that it's the best place to live in. [NEWLINE] [NEWLINE] There just seems to be a huge lack of support in comparison to other first-world countries that people don't know about in countries with that support system.</s>
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0578, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is very true.<mask> really,<mask> you think about it, "white" doesn't really make sense either. "White" encompasses many people of various shades and tones, very few who are actually white. [NEWLINE] Maybe whites should be called "People of Lighter Color"<mask> it's more accurate.</s>
Label encoding: <s>This is very true. But really, if you think about it, "white" doesn't really make sense either. "White" encompasses many people of various shades and tones, very few who are actually white. [NEWLINE] Maybe whites should be called "People of Lighter Color" so it's more accurate.</s>
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fine, I concede that there's a vanishingly small, negligible chance he will get busted watching netflix. Your scenario is incredibly unlikely<mask> it<mask> is not literally impossible to be busted<mask> sitting at home watching TV. [NEWLINE] [NEWLINE] <mask>'s the delta bot? You totally tore apart my argument just now.</s>
Label encoding: <s>Fine, I concede that there's a vanishingly small, negligible chance he will get busted watching netflix. Your scenario is incredibly unlikely but it indeed is not literally impossible to be busted while sitting at home watching TV. [NEWLINE] [NEWLINE] Where's the delta bot? You totally tore apart my argument just now.</s>
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While your statement of "<mask><mask> that it is very possible..." is a very good one, I find myself thinking it is insufficient to my statement of belief.<mask> I am looking for is an intermediate between "believe" and "I consider it very possible,"<mask> such a term does not exist.</s>
Label encoding: <s>While your statement of " I think that it is very possible..." is a very good one, I find myself thinking it is insufficient to my statement of belief. What I am looking for is an intermediate between "believe" and "I consider it very possible," but such a term does not exist.</s>
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Things like cancer, schizophrenia, depression are not transient. They very much<mask> have a permanent effect on the body of the individual for the rest of their lives. Genetic predispositions don't guarantee you're going to get it,<mask> it does make it damn hard to turn off once it's on.</s>
Label encoding: <s>Things like cancer, schizophrenia, depression are not transient. They very much so have a permanent effect on the body of the individual for the rest of their lives. Genetic predispositions don't guarantee you're going to get it, but it does make it damn hard to turn off once it's on.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So<mask><mask> you're arguing that legitimacy is the difference between corporations and government?<mask> that can't be<mask> you hold that the government doesn't have the legitimacy to use force, it just does. I<mask><mask><mask> well, except non-government<mask> illegitimately use force - and often. </s>
Label encoding: <s>So in fact you're arguing that legitimacy is the difference between corporations and government? But that can't be because you hold that the government doesn't have the legitimacy to use force, it just does. I argue that as well, except non-government also illegitimately use force - and often. </s>
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] difference between hate speech and more generally offensive [ENDQ] [NEWLINE] Such<mask>? [NEWLINE] [NEWLINE] [STARTQ] I feel that **sometimes** CMV is just an outlet to broadcast firmly held racist views. [ENDQ] [NEWLINE] you made no refernce to "sometimes" in your orginal post; has your view changed? </s>
Label encoding: <s> [STARTQ] difference between hate speech and more generally offensive [ENDQ] [NEWLINE] Such as? [NEWLINE] [NEWLINE] [STARTQ] I feel that **sometimes** CMV is just an outlet to broadcast firmly held racist views. [ENDQ] [NEWLINE] you made no refernce to "sometimes" in your orginal post; has your view changed? </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0539, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>reporting requires active mod work<mask> downvotes allow community moderation without forcing mods to wade through everything especially<mask> a comment is bad<mask> mods might not really want to remove it. [NEWLINE] [NEWLINE] the problem comes in<mask> downvotes aren't used in the limited manner the rediquette asks you to use it in.</s>
Label encoding: <s>reporting requires active mod work while downvotes allow community moderation without forcing mods to wade through everything especially if a comment is bad but mods might not really want to remove it. [NEWLINE] [NEWLINE] the problem comes in when downvotes aren't used in the limited manner the rediquette asks you to use it in.</s>
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you want children, sometimes you're not going to get precisely the child you want. That's life. [ENDQ] [NEWLINE] <mask> that's *not* life for women. Only for men. Women get to keep their options open and decide precisely which fetus/child they would like to keep. </s><pad>
Label encoding: <s> [STARTQ] If you want children, sometimes you're not going to get precisely the child you want. That's life. [ENDQ] [NEWLINE] But that's *not* life for women. Only for men. Women get to keep their options open and decide precisely which fetus/child they would like to keep. </s><pad>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0723, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0548, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0595, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's interesting. I looked it up and you're right (outside of NJ I guess). That being said I find it a bit weird that our military doesn't use them<mask> it's legal for private citizens? Regardless, few would<mask><mask> the private citizen could stand up to predator drones.</s>
Label encoding: <s>That's interesting. I looked it up and you're right (outside of NJ I guess). That being said I find it a bit weird that our military doesn't use them but it's legal for private citizens? Regardless, few would argue that the private citizen could stand up to predator drones.</s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I suppose this could boil down to a philosophical question -<mask> makes the person I am now, me?  The funny thing is, our bodies recycle cells at such a rate that within 7-10 years, pretty much every cell in your body would be different to the ones you have now.</s>
Label encoding: <s>I suppose this could boil down to a philosophical question - what makes the person I am now, me?  The funny thing is, our bodies recycle cells at such a rate that within 7-10 years, pretty much every cell in your body would be different to the ones you have now.</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0489, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Imagine, for a moment, that you discovered that human beings are eaten upon our death, by some form of supernatural entity. [NEWLINE] [NEWLINE] We are<mask> created by those same entities (conception being artificially encouraged by them). [NEWLINE] [NEWLINE] Would you rather not have been born than live in that world?</s>
Label encoding: <s>Imagine, for a moment, that you discovered that human beings are eaten upon our death, by some form of supernatural entity. [NEWLINE] [NEWLINE] We are also created by those same entities (conception being artificially encouraged by them). [NEWLINE] [NEWLINE] Would you rather not have been born than live in that world?</s>
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/McKoijion. ^[[History](/r/changemyview/wiki/user/McKoijion)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/McKoijion. ^[[History](/r/changemyview/wiki/user/McKoijion)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm a transsexual, I'm not speaking for the umbrella definition of transgender.  For someone born male to have "Sex: F" on our legal documentation requires undergoing medical treatment, which typically means at least hormone therapy,<mask> not surgery (it depends on the local laws).  </s>
Label encoding: <s>I'm a transsexual, I'm not speaking for the umbrella definition of transgender.  For someone born male to have "Sex: F" on our legal documentation requires undergoing medical treatment, which typically means at least hormone therapy, if not surgery (it depends on the local laws).  </s>
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Raintee97. ^[[History](/r/changemyview/wiki/user/Raintee97)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Raintee97. ^[[History](/r/changemyview/wiki/user/Raintee97)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0567, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm highly skeptical that all life on Earth would end<mask> all insects disappeared. I don't see<mask> extremophile bacteria or organisms living around hydrothermal vents would all go extinct without insects. Life on Earth existed before insects existed,<mask> it is possible for life to exist without insects.</s>
Label encoding: <s>I'm highly skeptical that all life on Earth would end if all insects disappeared. I don't see how extremophile bacteria or organisms living around hydrothermal vents would all go extinct without insects. Life on Earth existed before insects existed, so it is possible for life to exist without insects.</s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/maxpenny42. ^[[History](/r/changemyview/wiki/user/maxpenny42)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/maxpenny42. ^[[History](/r/changemyview/wiki/user/maxpenny42)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0646, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That's not<mask> he was saying. He was pointing out that several people here have changed views and<mask><mask> he was trying to imply that this was evidence against your view. I'm not saying<mask><mask> with this idea<mask> that's<mask> I believe that comment is getting at. </s>
Label encoding: <s>That's not what he was saying. He was pointing out that several people here have changed views and I think he was trying to imply that this was evidence against your view. I'm not saying I agree with this idea but that's what I believe that comment is getting at. </s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You assume that's the reason the society formed that way, and that only one group was complicit in that arrangement. [NEWLINE] [NEWLINE] Given<mask> we see in matriarchal societies(e.g. a lot of the same), claims that these things are due to patriarchy are dubious.</s>
Label encoding: <s>You assume that's the reason the society formed that way, and that only one group was complicit in that arrangement. [NEWLINE] [NEWLINE] Given what we see in matriarchal societies(e.g. a lot of the same), claims that these things are due to patriarchy are dubious.</s>
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0534, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ah, the history of dog bans.  I missed that day of social studies.  I realize that at different times of history, different places have been hostile to specific breeds.  German Shepherds were a symbol of oppressive police.<mask> and<mask> were they banned?  </s>
Label encoding: <s>Ah, the history of dog bans.  I missed that day of social studies.  I realize that at different times of history, different places have been hostile to specific breeds.  German Shepherds were a symbol of oppressive police. When and where were they banned?  </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>*You* said that, in response to "Most religious people". Clearly /u/Wolf_Dancing is referring to<mask> religious people tend to believe--<mask> you want to argue about the biblical text specifically then you probably shouldn't have engaged with him in the first place. </s>
Label encoding: <s>*You* said that, in response to "Most religious people". Clearly /u/Wolf_Dancing is referring to what religious people tend to believe-- if you want to argue about the biblical text specifically then you probably shouldn't have engaged with him in the first place. </s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I am asserting that it was a war primarily for profit.  wealthy individuals saw an opportunity, took advantage of our hurt and fervor, and made bank.  to us, the american people, it is about terrorism,<mask> to the people making decisions it is about money.</s>
Label encoding: <s>I am asserting that it was a war primarily for profit.  wealthy individuals saw an opportunity, took advantage of our hurt and fervor, and made bank.  to us, the american people, it is about terrorism, but to the people making decisions it is about money.</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No one would hear about the bad policies of said companies,<mask> nbc is owned by GE. Abc is owned by Disney, and cbs is owned by Viacom. [NEWLINE] [NEWLINE] There would be no npr...<mask> well it's funded by the state. </s>
Label encoding: <s>No one would hear about the bad policies of said companies, because nbc is owned by GE. Abc is owned by Disney, and cbs is owned by Viacom. [NEWLINE] [NEWLINE] There would be no npr... Because well it's funded by the state. </s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well emergency medical services and taxes (the skills in ops post)  CAN be taught. [NEWLINE] [NEWLINE] I'm not suggesting we stop teaching critical thinking,<mask> sometimes you need outside help to learn some skills and it makes sense that the government would provide that help to a basic extent </s>
Label encoding: <s>Well emergency medical services and taxes (the skills in ops post)  CAN be taught. [NEWLINE] [NEWLINE] I'm not suggesting we stop teaching critical thinking, but sometimes you need outside help to learn some skills and it makes sense that the government would provide that help to a basic extent </s>
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a number of huge differences<mask>. Young children become moral adults. Young children have moral adult parents who will never lose their empathy towards their children. [NEWLINE] [NEWLINE] Finally and most importantly, the goal here is to benefit humanity, being immoral to young children defeats the whole point.</s>
Label encoding: <s>There are a number of huge differences though. Young children become moral adults. Young children have moral adult parents who will never lose their empathy towards their children. [NEWLINE] [NEWLINE] Finally and most importantly, the goal here is to benefit humanity, being immoral to young children defeats the whole point.</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Evsie. ^[[History](/r/changemyview/wiki/user/Evsie)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Evsie. ^[[History](/r/changemyview/wiki/user/Evsie)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ehhh... I can see<mask> such a one-size-fits-all approach can backfire, escpecially with manic patients or ones that are hesitant to reveal information. It may be more knowledgeable than a human,<mask> situational awareness and experience will *always* trump that.</s>
Label encoding: <s>Ehhh... I can see how such a one-size-fits-all approach can backfire, escpecially with manic patients or ones that are hesitant to reveal information. It may be more knowledgeable than a human, but situational awareness and experience will *always* trump that.</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Trump quote during the debate: [NEWLINE] &gt;And the Mexican government is much smarter, much sharper, much more cunning. And they send the bad ones over<mask> they don’t want to pay for them. They don’t want to take care of them.</s>
Label encoding: <s>Trump quote during the debate: [NEWLINE] &gt;And the Mexican government is much smarter, much sharper, much more cunning. And they send the bad ones over because they don’t want to pay for them. They don’t want to take care of them.</s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Of course. Simply saying "female" wouldn't be accurate for everything either, that's not<mask> I'm suggesting. My point is just that the idea of sex<mask> some immutable single binary doesn't work<mask> dealing with trans issues (and really any intersex issues in general).</s>
Label encoding: <s>Of course. Simply saying "female" wouldn't be accurate for everything either, that's not what I'm suggesting. My point is just that the idea of sex as some immutable single binary doesn't work when dealing with trans issues (and really any intersex issues in general).</s>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.7969e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0542, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0536, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0482, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0620, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0613, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think your formula would only work<mask> you think intelligence is a fixed trait that cannot be changed. I just don't think that's the case. Intelligence is fluid and can be improved through persistent effort. I don't think one trait is more important than the other.</s>
Label encoding: <s>I think your formula would only work if you think intelligence is a fixed trait that cannot be changed. I just don't think that's the case. Intelligence is fluid and can be improved through persistent effort. I don't think one trait is more important than the other.</s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You'd imagine women, who always sit, would be bitten more than men,<mask> men are significantly more likely to be bitten by toilet spiders!  Black widows are particularly fond of attacking men's genitals.  [Source]( [URL] /?page=1)  </s>
Label encoding: <s>You'd imagine women, who always sit, would be bitten more than men, but men are significantly more likely to be bitten by toilet spiders!  Black widows are particularly fond of attacking men's genitals.  [Source]( [URL] /?page=1)  </s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Anotherchallenger. ^[[History](/r/changemyview/wiki/user/Anotherchallenger)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Anotherchallenger. ^[[History](/r/changemyview/wiki/user/Anotherchallenger)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Cooper720. ^[[History](/r/changemyview/wiki/user/Cooper720)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Cooper720. ^[[History](/r/changemyview/wiki/user/Cooper720)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/Pensky. ^[[History](/r/changemyview/wiki/user/Pensky)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/Pensky. ^[[History](/r/changemyview/wiki/user/Pensky)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/sylban. ^[[History](/r/changemyview/wiki/user/sylban)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/sylban. ^[[History](/r/changemyview/wiki/user/sylban)] [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][/r/DeltaBot]</s>
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Semantics mostly. The Federal government has the right to TAX people for things they don't have a choice about. They do not have the right to fine them for things they don't have a choice about. [NEWLINE] [NEWLINE] That's<mask> I understand it anyway.</s>
Label encoding: <s>Semantics mostly. The Federal government has the right to TAX people for things they don't have a choice about. They do not have the right to fine them for things they don't have a choice about. [NEWLINE] [NEWLINE] That's how I understand it anyway.</s>
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the crime point varies greatly. I don't disbelieve your personal experience with<mask> you have lived and grime,<mask> I live 10-15 away from the downtown of a very large city and crime here is one of the lowest in the country. </s>
Label encoding: <s>I think the crime point varies greatly. I don't disbelieve your personal experience with where you have lived and grime, but I live 10-15 away from the downtown of a very large city and crime here is one of the lowest in the country. </s>
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I live in Texas, originally from Iowa, and have several coworkers from California.  It should be noted that I am not opposed to hops per se, just that there are more hopped (and heavily hopped) varieties than there are others, limiting choice. [NEWLINE] </s>
Label encoding: <s>I live in Texas, originally from Iowa, and have several coworkers from California.  It should be noted that I am not opposed to hops per se, just that there are more hopped (and heavily hopped) varieties than there are others, limiting choice. [NEWLINE] </s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not a Christian, I was just explaining the ideal of Christian faith,<mask> I will say this: wouldn't a world with free will require the capability for evil?<mask> you can't choose to do wrong, then you don't have free will.</s>
Label encoding: <s>I'm not a Christian, I was just explaining the ideal of Christian faith, but I will say this: wouldn't a world with free will require the capability for evil? If you can't choose to do wrong, then you don't have free will.</s>
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0466, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It seems to me like gay people are treated incredibly well. They are generally accepted. Homophobes<mask><mask><mask><mask> seem to get a lot of hate. Somebody makes a comment about not supporting gay marriage and suddenly the whole nation is at their throat. </s>
Label encoding: <s>It seems to me like gay people are treated incredibly well. They are generally accepted. Homophobes on the other hand seem to get a lot of hate. Somebody makes a comment about not supporting gay marriage and suddenly the whole nation is at their throat. </s>
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] To my mind they remain legal cause people are free to do with their bodies whatever they want. [ENDQ] [NEWLINE] The problem isn't that they can damage themselves, it's that they can damage or traumatise others<mask> they're involved in a crash. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] To my mind they remain legal cause people are free to do with their bodies whatever they want. [ENDQ] [NEWLINE] The problem isn't that they can damage themselves, it's that they can damage or traumatise others if they're involved in a crash. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0511, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0561, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a Californian this is a terrible system we have. Many props like Prop 13 are terrible NIMBY ask hand outs to the elderly population that discourage modernization and development. [NEWLINE] [NEWLINE] Its better to make the elected officials accountable than to let the mob enact justice</s>
Label encoding: <s>As a Californian this is a terrible system we have. Many props like Prop 13 are terrible NIMBY ask hand outs to the elderly population that discourage modernization and development. [NEWLINE] [NEWLINE] Its better to make the elected officials accountable than to let the mob enact justice</s>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0527, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You can think about it in terms of realism versus surrealism. The stricter you keep realism, the more rules you have to follow the less scenarios is possible within the confines of that universe. Like<mask> Batman have a lesser width of options than The Mask.</s>
Label encoding: <s>You can think about it in terms of realism versus surrealism. The stricter you keep realism, the more rules you have to follow the less scenarios is possible within the confines of that universe. Like how Batman have a lesser width of options than The Mask.</s>
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can I see that rant<mask> well? I've never posted a CMV,<mask><mask> I did, especially for Steve Jobs is "<mask><mask> that everyone who achieved a large amount of success must have been cutthroat/an ass at some point"</s>
Label encoding: <s>Can I see that rant as well? I've never posted a CMV, but if I did, especially for Steve Jobs is " I think that everyone who achieved a large amount of success must have been cutthroat/an ass at some point"</s>
Loss: tensor(0.0560, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This sounds like a much simpler and cleaner solution than straight up limiting people's ability to procreate. [NEWLINE] [NEWLINE] <mask>, people largely limit their own procreation<mask> given access to birth control and abortion. We can basically have our cake and eat it too.</s><pad>
Label encoding: <s>This sounds like a much simpler and cleaner solution than straight up limiting people's ability to procreate. [NEWLINE] [NEWLINE] Besides, people largely limit their own procreation when given access to birth control and abortion. We can basically have our cake and eat it too.</s><pad>
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, it would probably be pointless<mask> all I had was a sniper rifle and no other powers in the situation. The whole thing was a minor point and not that well thought out - I changed my view on it somewhat and gave someone else a delta.</s>
Label encoding: <s>Yes, it would probably be pointless if all I had was a sniper rifle and no other powers in the situation. The whole thing was a minor point and not that well thought out - I changed my view on it somewhat and gave someone else a delta.</s>
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't know<mask> coontown is or<mask> those words mean,<mask> banning a sub without legitimacy means that potentially this could happen to any sub<mask> the admins feel like it. An admin doesn't like movies? r/movies is banned.</s>
Label encoding: <s>I don't know what coontown is or what those words mean, but banning a sub without legitimacy means that potentially this could happen to any sub if the admins feel like it. An admin doesn't like movies? r/movies is banned.</s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0488, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0434, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Please read my comments carefully: Nobody I stated that this was a good idea. It's not at all. [NEWLINE] [NEWLINE] I just wanted to point out that<mask> OP suggests may be considered the status quo already in countries which are considered to be democratic.</s>
Label encoding: <s>Please read my comments carefully: Nobody I stated that this was a good idea. It's not at all. [NEWLINE] [NEWLINE] I just wanted to point out that what OP suggests may be considered the status quo already in countries which are considered to be democratic.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0588, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So a person would end up being charged with murder<mask><mask> he's completely innocent of the crime,<mask> someone screwed him over, knowing he would take the rap for<mask> the actual criminal did? Yeah. Seems like a good idea to me. </s>
Label encoding: <s>So a person would end up being charged with murder even though he's completely innocent of the crime, but someone screwed him over, knowing he would take the rap for what the actual criminal did? Yeah. Seems like a good idea to me. </s>
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think Op specified "early" death which could mean drastically premature. I can swallow that<mask> "not working out",<mask> I'd find it odd that OP has never come into contact with a single couple who remained married and alive into old age.</s>
Label encoding: <s>I think Op specified "early" death which could mean drastically premature. I can swallow that as "not working out", but I'd find it odd that OP has never come into contact with a single couple who remained married and alive into old age.</s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No. [NEWLINE] [NEWLINE] I am saying that by your standards it is irrational- not by mine.<mask> I ask again, do you think it makes sense for children to be scared<mask> walking alone at night? Or should they<mask> not be scared?</s>
Label encoding: <s>No. [NEWLINE] [NEWLINE] I am saying that by your standards it is irrational- not by mine. So I ask again, do you think it makes sense for children to be scared when walking alone at night? Or should they also not be scared?</s>
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's just the reality that they want not to have children. [ENDQ] [NEWLINE] <mask> you have a survey that specifically indicates that I'd be interested in reading it.  otherwise, you are way out of line speaking for them in this hypothetical.</s>
Label encoding: <s> [STARTQ] There's just the reality that they want not to have children. [ENDQ] [NEWLINE] If you have a survey that specifically indicates that I'd be interested in reading it.  otherwise, you are way out of line speaking for them in this hypothetical.</s>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You should know full well driving a car is not the most risky thing I do. Riding a motorcycle is the most risky thing I do. "Probably" doesn't enter into it,<mask> I've already stated I ride a motorcycle. </s>
Label encoding: <s>You should know full well driving a car is not the most risky thing I do. Riding a motorcycle is the most risky thing I do. "Probably" doesn't enter into it, as I've already stated I ride a motorcycle. </s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It seems that it should be: "There is no benefit to me personally to be a vegetarian" [ENDQ] [NEWLINE] <mask> egoism is the correct account of morality, then there is no difference between that title and the title of this thread.</s>
Label encoding: <s> [STARTQ] It seems that it should be: "There is no benefit to me personally to be a vegetarian" [ENDQ] [NEWLINE] If egoism is the correct account of morality, then there is no difference between that title and the title of this thread.</s>
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>To add on, the obvious solution (not hiring smokers) just flat out isn't available in many places. [29 states]( [URL] /) have laws in place that prevent employers from refusing to hire a person<mask> that person smokes. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>To add on, the obvious solution (not hiring smokers) just flat out isn't available in many places. [29 states]( [URL] /) have laws in place that prevent employers from refusing to hire a person because that person smokes. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>All the best to you on that journey.  Meta-ethics are the hardest thing in the world (one of the only places<mask> I ever made use of my metaphysics training -- blergh),<mask><mask> the most rewarding.</s><pad>
Label encoding: <s>All the best to you on that journey.  Meta-ethics are the hardest thing in the world (one of the only places where I ever made use of my metaphysics training -- blergh), but also the most rewarding.</s><pad>
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0629, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0517, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>IIRC, all of the major religions offer some sort of salvation/post death happiness/whatever to non-believers<mask> they lead lives that accidentally or not, follow<mask> that religion considers to be good/holy/whatever.</s>
Label encoding: <s>IIRC, all of the major religions offer some sort of salvation/post death happiness/whatever to non-believers if they lead lives that accidentally or not, follow what that religion considers to be good/holy/whatever.</s>
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I know the title is that and Orwell was pissed off<mask> people called it<mask> I called it,<mask> everyone knew<mask> I meant<mask> most people just use the shorthand version nowadays, just like lotr for lord of the rings.</s>
Label encoding: <s>I know the title is that and Orwell was pissed off when people called it what I called it, but everyone knew what I meant as most people just use the shorthand version nowadays, just like lotr for lord of the rings.</s>
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If I wake up in the middle of the night to someone robbing me, I'm not gonna ask them<mask> their kids are starving. They pose a threat, and I'm gonna defend my property and girlfriend<mask> I have to</s>
Label encoding: <s>If I wake up in the middle of the night to someone robbing me, I'm not gonna ask them if their kids are starving. They pose a threat, and I'm gonna defend my property and girlfriend however I have to</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0587, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0568, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Taking away all semblance of any free markets does that. [ENDQ] [NEWLINE] That is<mask> the OP proposes with respect to healthcare, and<mask> the comment I responded to is proposing for clothing, housing, food, water and energy.</s>
Label encoding: <s> [STARTQ] Taking away all semblance of any free markets does that. [ENDQ] [NEWLINE] That is what the OP proposes with respect to healthcare, and what the comment I responded to is proposing for clothing, housing, food, water and energy.</s>
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0452, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Before you posted this, did you ever think<mask> don't we already do this? Some animals don't lend themselves to farming, these animals breed slow, require massive space and feed, and are difficult to reproduce in captivity.</s>
Label encoding: <s>Before you posted this, did you ever think why don't we already do this? Some animals don't lend themselves to farming, these animals breed slow, require massive space and feed, and are difficult to reproduce in captivity.</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0512, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Humans existed before shelter, surplus nutrition, formal education, and medical knowledge.  Someone provided these things. [NEWLINE] [NEWLINE] You are not saying we have a right to them, you are saying they should be provided. </s>
Label encoding: <s>Humans existed before shelter, surplus nutrition, formal education, and medical knowledge.  Someone provided these things. [NEWLINE] [NEWLINE] You are not saying we have a right to them, you are saying they should be provided. </s>
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0608, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Completely innocent? Being irresponsible with<mask> you keep your firearm to the point that someone can just take it is not innocent at all. And the actual criminal still gets charged<mask> he always would have. That is obvious.</s>
Label encoding: <s>Completely innocent? Being irresponsible with how you keep your firearm to the point that someone can just take it is not innocent at all. And the actual criminal still gets charged as he always would have. That is obvious.</s>
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for your answer. I understand that someone equally proficient in both languages would not have to mentally translate to the other. More specifically, I was curious<mask> to which alphabet was more efficient at conveying information. I</s>
Label encoding: <s>Thanks for your answer. I understand that someone equally proficient in both languages would not have to mentally translate to the other. More specifically, I was curious as to which alphabet was more efficient at conveying information. I</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That was the point of the shotgun weddings study: to try and look at people who did not plan marriage in advance. It found 90% of the marriage premium remained. [NEWLINE] [NEWLINE] *Edit: fixed percentage sign*</s><pad>
Label encoding: <s>That was the point of the shotgun weddings study: to try and look at people who did not plan marriage in advance. It found 90% of the marriage premium remained. [NEWLINE] [NEWLINE] *Edit: fixed percentage sign*</s><pad>
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But a popular opinion.<mask> I alone had the opinion the awards would lose credibility with me and only me. Are there not a lot of people who are now beginning to wonder<mask> LD is not getting an Oscar?</s>
Label encoding: <s>But a popular opinion. If I alone had the opinion the awards would lose credibility with me and only me. Are there not a lot of people who are now beginning to wonder why LD is not getting an Oscar?</s>
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0876, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0473, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0631, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry,<mask> my claims don't seam extraordinary to me, maybe it is<mask> I've been reading about and following the social work of a specific sex professionals association,<mask> their side is more natural to me.</s>
Label encoding: <s>Sorry, but my claims don't seam extraordinary to me, maybe it is because I've been reading about and following the social work of a specific sex professionals association, so their side is more natural to me.</s>
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you're missing the judgement part. It's not<mask> much "don't think you're better than anyone"<mask> "don't do anything that makes someone else think you're better than them".</s>
Label encoding: <s>I think you're missing the judgement part. It's not so much "don't think you're better than anyone" as "don't do anything that makes someone else think you're better than them".</s>
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I call my parents by their first name out of habit. My parents never told me otherwise. [NEWLINE] Their argument is, that thats<mask> everyone in the family calls them. And<mask> not their own children?</s>
Label encoding: <s>I call my parents by their first name out of habit. My parents never told me otherwise. [NEWLINE] Their argument is, that thats what everyone in the family calls them. And why not their own children?</s>
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If the "fundies" are ok with just calling something different. Then<mask> not make the government name "marriage" and let the fundies refer to the marriage of gay couples<mask> "civil unions"?</s><pad>
Label encoding: <s>If the "fundies" are ok with just calling something different. Then why not make the government name "marriage" and let the fundies refer to the marriage of gay couples as "civil unions"?</s><pad>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0474, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Keep in mind<mask> that ammunition goes bad after awhile. There's nothing wrong with having &lt; 1,000 rounds for<mask> shtf,<mask> more than that may be less than a great idea.</s>
Label encoding: <s>Keep in mind though that ammunition goes bad after awhile. There's nothing wrong with having &lt; 1,000 rounds for when shtf, but more than that may be less than a great idea.</s>
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The polar bears [seem to be doing just fine.]( [URL] /),<mask><mask> you want to make an emotional argument it's easier to focus on [this one bear that starved]( [URL] ).</s>
Label encoding: <s>The polar bears [seem to be doing just fine.]( [URL] /), but if you want to make an emotional argument it's easier to focus on [this one bear that starved]( [URL] ).</s>
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure<mask> I'm reading this right...<mask> are you saying that most cultures around the world view spiders<mask> our friends?  I know it's<mask> the point...<mask> fuck spiders.</s>
Label encoding: <s>I'm not sure if I'm reading this right... But are you saying that most cultures around the world view spiders as our friends?  I know it's besides the point... But fuck spiders.</s>
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I feel like a lot of that would only apply<mask> you were white.<mask> you were another ethnicity I feel like racism still would have been quite a hinderance to you living the "American Dream"</s>
Label encoding: <s>I feel like a lot of that would only apply if you were white. If you were another ethnicity I feel like racism still would have been quite a hinderance to you living the "American Dream"</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What<mask> someone with a loud chronic breathing problem wanted to go see a movie (i.e., they wheezed/coughed loudly and frequently)?  Or someone with Tourette's?</s>
Label encoding: <s>What if someone with a loud chronic breathing problem wanted to go see a movie (i.e., they wheezed/coughed loudly and frequently)?  Or someone with Tourette's?</s>
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not true. I had a place to live for a year. I have less than I would have had<mask> I'd paid that money into a mortgage,<mask> I don't have *nothing*.</s>
Label encoding: <s>Not true. I had a place to live for a year. I have less than I would have had if I'd paid that money into a mortgage, but I don't have *nothing*.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Can't or won't? This is the point of the argument.<mask> he can't, he's not omnipotent.<mask> he won't, he's not omnibenevolent. </s>
Label encoding: <s>Can't or won't? This is the point of the argument. If he can't, he's not omnipotent. If he won't, he's not omnibenevolent. </s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It should be on the right side of your screen, under awarding deltas, it's like a five digit code.  I'm on a iPhone and can't open the screen. </s>
Label encoding: <s>It should be on the right side of your screen, under awarding deltas, it's like a five digit code.  I'm on a iPhone and can't open the screen. </s>
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Clarification: are you arguing that businesses should be required **by law** to accommodate well-behaved dogs, or just that more businesses should accommodate them voluntarily? These are radically different propositions.</s>
Label encoding: <s>Clarification: are you arguing that businesses should be required **by law** to accommodate well-behaved dogs, or just that more businesses should accommodate them voluntarily? These are radically different propositions.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OP, I've agreed with pretty much everything you've said in this thread, right down to the eating babies is okay<mask> no one cared bit. It's kind of creepy.  </s>
Label encoding: <s>OP, I've agreed with pretty much everything you've said in this thread, right down to the eating babies is okay if no one cared bit. It's kind of creepy.  </s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Confirmed: 1 delta awarded to /u/DerekReinbold. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Label encoding: <s>Confirmed: 1 delta awarded to /u/DerekReinbold. [NEWLINE] [NEWLINE] ^[[Wiki]( [URL] )][[Code]( [URL] )][[Subreddit]( [URL] /)]</s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Assuming the toilet is compatible with your pipes (I know next to nothing about plumbing) a more efficient toilet would pay for itself, and then start to save you money, wouldn't it?</s>
Label encoding: <s>Assuming the toilet is compatible with your pipes (I know next to nothing about plumbing) a more efficient toilet would pay for itself, and then start to save you money, wouldn't it?</s>
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just wondering. Did you continue to drink/get drunk with them in order to have a good time? Or was it just that one night that set the atmosphere for every other subsequent interaction?</s>
Label encoding: <s>Just wondering. Did you continue to drink/get drunk with them in order to have a good time? Or was it just that one night that set the atmosphere for every other subsequent interaction?</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>our government is not a solid single organism, look at congress today they can't ever agree on shit, you think these people could band together to form an evil cabal to shape future events?</s>
Label encoding: <s>our government is not a solid single organism, look at congress today they can't ever agree on shit, you think these people could band together to form an evil cabal to shape future events?</s>
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] TL;DR: X is better than Y at saving money at the barber shop. [ENDQ] [NEWLINE] Nice summary. I haven't visited a barber in the past 5 years. </s>
Label encoding: <s> [STARTQ] TL;DR: X is better than Y at saving money at the barber shop. [ENDQ] [NEWLINE] Nice summary. I haven't visited a barber in the past 5 years. </s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>ok,<mask> instead of racial minority, I could say muslims, or atheists, or people pro-gay rights, or feminists... Is that more acceptable to openly disparage?</s>
Label encoding: <s>ok, so instead of racial minority, I could say muslims, or atheists, or people pro-gay rights, or feminists... Is that more acceptable to openly disparage?</s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"Comments that are only jokes or "written upvotes", for example." [NEWLINE] [NEWLINE] Making witty comments to prove<mask> cool you are to other redditors will get your comments removed here.</s>
Label encoding: <s>"Comments that are only jokes or "written upvotes", for example." [NEWLINE] [NEWLINE] Making witty comments to prove how cool you are to other redditors will get your comments removed here.</s>
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> You already have the knowledge that the dog will eat the stake.<mask> it should be pretty obvious that no, this is not a trained dog. [NEWLINE] [NEWLINE] Would you punish it? </s>
Label encoding: <s> You already have the knowledge that the dog will eat the stake. Therefore it should be pretty obvious that no, this is not a trained dog. [NEWLINE] [NEWLINE] Would you punish it? </s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Farmer Joe has a hard enough time of making ends meet, hes a farmer.<mask><mask> guarantees someone is willing to deliver Joes supplies? The service provided seems hardly profitable. </s>
Label encoding: <s>Farmer Joe has a hard enough time of making ends meet, hes a farmer. Also what guarantees someone is willing to deliver Joes supplies? The service provided seems hardly profitable. </s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think that's sort of the point /u/brynleypearlstone was trying to make: there are more Catholics outside of Europe<mask> the bureaucracy has remained eurocentric.</s>
Label encoding: <s>I think that's sort of the point /u/brynleypearlstone was trying to make: there are more Catholics outside of Europe yet the bureaucracy has remained eurocentric.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Some of them wish that.<mask> many don't. It's wrong either way. [NEWLINE] [NEWLINE] One of the main points here are that the goal doesn't always justify the means  </s>
Label encoding: <s>Some of them wish that. While many don't. It's wrong either way. [NEWLINE] [NEWLINE] One of the main points here are that the goal doesn't always justify the means  </s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you are a responsible successful person at 25 and you have the recourses to run for president, you may not have the recourses or the correct political climate to run at 35</s>
Label encoding: <s>If you are a responsible successful person at 25 and you have the recourses to run for president, you may not have the recourses or the correct political climate to run at 35</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] **<mask><mask><mask> :** a child is not a full citizen. [ENDQ] [NEWLINE] <mask><mask> can't caretakers (a nanny or teacher for example) hit children?</s>
Label encoding: <s> [STARTQ] ** tldr :** a child is not a full citizen. [ENDQ] [NEWLINE] So why can't caretakers (a nanny or teacher for example) hit children?</s>
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Eh, not 100% sure.<mask> I do know that most companies that don't rely on the State are much better at providing good services with reduced risk of harming people. </s>
Label encoding: <s>Eh, not 100% sure. But I do know that most companies that don't rely on the State are much better at providing good services with reduced risk of harming people. </s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you're a salesmen you shouldn't be discussing any personal beliefs with anyone, in a professional environment, unless they mirror the clients beliefs/interests. Too risky otherwise.</s>
Label encoding: <s>If you're a salesmen you shouldn't be discussing any personal beliefs with anyone, in a professional environment, unless they mirror the clients beliefs/interests. Too risky otherwise.</s>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As demonstrated in the story of Eden, lying is not inherent to humans<mask> is a result of embarrassment or desire.  A soul that is one with God would have neither. </s>
Label encoding: <s>As demonstrated in the story of Eden, lying is not inherent to humans but is a result of embarrassment or desire.  A soul that is one with God would have neither. </s>
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ah, I took your opinion to mean "We'll create an accidental non-human AI before we deliberately create an AI that emulates humans". Was this not<mask> you believe?</s>
Label encoding: <s>Ah, I took your opinion to mean "We'll create an accidental non-human AI before we deliberately create an AI that emulates humans". Was this not what you believe?</s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is a difference in my view. I work, I earn money, part of that is taken to support the society I live in. Inheritance is a different beast altogether.</s>
Label encoding: <s>There is a difference in my view. I work, I earn money, part of that is taken to support the society I live in. Inheritance is a different beast altogether.</s>
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0432, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Eh, just not the way I see it.  To me, both candidates are pretty crappy<mask><mask> vote for either? <mask><mask> with each an equal amount. </s><pad>
Label encoding: <s>Eh, just not the way I see it.  To me, both candidates are pretty crappy so why vote for either?  I disagree with each an equal amount. </s><pad>
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry, it's just a known saying. Would you rather I said, "Two wrongs don't make a right." You're the one putting Jesus into things. :)</s>
Label encoding: <s>Sorry, it's just a known saying. Would you rather I said, "Two wrongs don't make a right." You're the one putting Jesus into things. :)</s>
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0569, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0437, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Is the vegan restaurant owner discriminating against the meat eater?  Should the owner be forced to serve meat? [NEWLINE] [NEWLINE] Should the PC store be forced to sell Apple products?</s>
Label encoding: <s>Is the vegan restaurant owner discriminating against the meat eater?  Should the owner be forced to serve meat? [NEWLINE] [NEWLINE] Should the PC store be forced to sell Apple products?</s>
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0513, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0559, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe he wasn't actually guilty, and a trial would have proven that had it not been rigged<mask> well. America decided who to blame and we stuck with it.</s>
Label encoding: <s>Maybe he wasn't actually guilty, and a trial would have proven that had it not been rigged as well. America decided who to blame and we stuck with it.</s>
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Its 76~% I can't remember the exact figure,<mask> it was around that. There was an infograghic awhile Ago comparing different websites demographics.</s>
Label encoding: <s>Its 76~% I can't remember the exact figure, but it was around that. There was an infograghic awhile Ago comparing different websites demographics.</s>
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The two young men were running at him<mask><mask> they were about to attack him!!! [NEWLINE] [NEWLINE] Tell me you didn't see that part or<mask> you interpret it differently</s><pad><pad>
Label encoding: <s>The two young men were running at him as if they were about to attack him!!! [NEWLINE] [NEWLINE] Tell me you didn't see that part or how you interpret it differently</s><pad><pad>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0460, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Sorry, until there's any sort of oversight on the ways in which money intended for child support is spent, it's functionally spousal support<mask> well.</s>
Label encoding: <s>Sorry, until there's any sort of oversight on the ways in which money intended for child support is spent, it's functionally spousal support as well.</s>
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not gay<mask> I've been to some LGBT events, and to some "gay bars" and never felt like I shouldn't be there.  </s>
Label encoding: <s>I'm not gay but I've been to some LGBT events, and to some "gay bars" and never felt like I shouldn't be there.  </s>
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Have you seen age of ultron? Its definitely fiction,<mask> the concept of a fully AI weapons system will inevitably be a reality. That could easily turn.</s>
Label encoding: <s>Have you seen age of ultron? Its definitely fiction, but the concept of a fully AI weapons system will inevitably be a reality. That could easily turn.</s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0352, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The penalty for treason is life imprisonment or death. <mask> do we need to toss in stripping citizenship<mask> well?  Killing them isn't enough?</s>
Label encoding: <s>The penalty for treason is life imprisonment or death.  Why do we need to toss in stripping citizenship as well?  Killing them isn't enough?</s>
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0448, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0893, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your biggest mistake is thinking that it can't be both. Affirmative action is discrimination,<mask> that doesn't mean it's wrong.</s>
Label encoding: <s>Your biggest mistake is thinking that it can't be both. Affirmative action is discrimination, but that doesn't mean it's wrong.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hmm, it seems the majority take this approach. And it's a preference. I don't think this will quite convince me<mask>. </s>
Label encoding: <s>Hmm, it seems the majority take this approach. And it's a preference. I don't think this will quite convince me yet. </s>
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0500, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Silver is OK. Read OP, it's not limited to gold. [NEWLINE] [NEWLINE] Annuities gonna be worthless, in a collapse scenario.</s>
Label encoding: <s>Silver is OK. Read OP, it's not limited to gold. [NEWLINE] [NEWLINE] Annuities gonna be worthless, in a collapse scenario.</s>
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Well you are incorrect with regard to point 3. [ENDQ] [NEWLINE] Nope, [adoption requires the consent of both parents]( [URL] ).</s>
Label encoding: <s> [STARTQ] Well you are incorrect with regard to point 3. [ENDQ] [NEWLINE] Nope, [adoption requires the consent of both parents]( [URL] ).</s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think that unpredictability is a real concept.<mask> would you make something unpredictable? You could only make it more complex.</s>
Label encoding: <s>I don't think that unpredictability is a real concept. How would you make something unpredictable? You could only make it more complex.</s>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0442, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I honestly disagree<mask> the situations that arise from straight people frequenting gay bars would be very different than gay people frequenting straight bars.</s>
Label encoding: <s>I honestly disagree because the situations that arise from straight people frequenting gay bars would be very different than gay people frequenting straight bars.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>He wasn't even acting<mask> a soldier<mask> he was killed, he was a symbolic guard for security and respect at a memorial.</s>
Label encoding: <s>He wasn't even acting as a soldier when he was killed, he was a symbolic guard for security and respect at a memorial.</s>
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0491, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Seriously, it's not like<mask> soon<mask> a UBI is in place the dollar goes the way of the Zimbabwe dollar. </s>
Label encoding: <s>Seriously, it's not like as soon as a UBI is in place the dollar goes the way of the Zimbabwe dollar. </s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Perhaps you can consider prologues<mask> something to whet your intellectual appetite before you sink your teeth into the main story. </s>
Label encoding: <s>Perhaps you can consider prologues as something to whet your intellectual appetite before you sink your teeth into the main story. </s>
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure. I guess I am saying you should try to use non-lethal force or attempt to escape the situation.</s>
Label encoding: <s>I'm not sure. I guess I am saying you should try to use non-lethal force or attempt to escape the situation.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0750, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But you're describing a problem that wouldn't exist<mask> Batman didn't share space with the rest of the JLA.</s>
Label encoding: <s>But you're describing a problem that wouldn't exist if Batman didn't share space with the rest of the JLA.</s>
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>as i mentioned earlier..<mask> was silverstein in the know about this whole inside job and cover up? like really...</s>
Label encoding: <s>as i mentioned earlier.. why was silverstein in the know about this whole inside job and cover up? like really...</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well then are you in understanding of other faiths? It sounds like your "Achristian" rather than atheist?</s>
Label encoding: <s>Well then are you in understanding of other faiths? It sounds like your "Achristian" rather than atheist?</s>
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm right there with you in the biology field.  We are every bit<mask> good<mask> a mediocre dude.</s>
Label encoding: <s>I'm right there with you in the biology field.  We are every bit as good as a mediocre dude.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>anyone who claims to have the knowledge about the creation of the universe one way or another is equally terrible </s><pad>
Label encoding: <s>anyone who claims to have the knowledge about the creation of the universe one way or another is equally terrible </s><pad>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the whole thing is a big joke actually.<mask> not, may God have mercy on us all.</s>
Label encoding: <s>I think the whole thing is a big joke actually. If not, may God have mercy on us all.</s>
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well I am currently leaning toward a vocational school.<mask> is the benefit  in going towards a university?</s>
Label encoding: <s>Well I am currently leaning toward a vocational school. What is the benefit  in going towards a university?</s>
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>how am i being rude, i'm telling this guy that he's being a nazi apologist.</s>
Label encoding: <s>how am i being rude, i'm telling this guy that he's being a nazi apologist.</s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Did you just stop reading after the first sentence? Your post is not in conflict with mine. </s><pad>
Label encoding: <s>Did you just stop reading after the first sentence? Your post is not in conflict with mine. </s><pad>
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0506, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't really have anything else to add to<mask> you wrote,<mask> good post bro.</s>
Label encoding: <s>I don't really have anything else to add to what you wrote, so good post bro.</s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0563, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So you agree that some universities are flawed, and I happen to be going to one.</s>
Label encoding: <s>So you agree that some universities are flawed, and I happen to be going to one.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Why flamboyant?<mask> not, say, simply holding hands with his boyfriend?</s>
Label encoding: <s>Why flamboyant? Why not, say, simply holding hands with his boyfriend?</s>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Personally I am supporting her and Taylor by buying one of her shirts. They deserve happiness</s>
Label encoding: <s>Personally I am supporting her and Taylor by buying one of her shirts. They deserve happiness</s>
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks.  That comment really made a meaningful contribution to this debate didn't it?</s>
Label encoding: <s>Thanks.  That comment really made a meaningful contribution to this debate didn't it?</s>
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm gonna be honest, I actually thought you were talking about pokemon before I clicked</s>
Label encoding: <s>I'm gonna be honest, I actually thought you were talking about pokemon before I clicked</s>
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ya, kicking nations out of the UN defeats the entire point of the UN</s><pad>
Label encoding: <s>Ya, kicking nations out of the UN defeats the entire point of the UN</s><pad>
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0498, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Are we not? Isn't that government in and of itself? Taxes?</s>
Label encoding: <s>Are we not? Isn't that government in and of itself? Taxes?</s>
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Then Malazan is probably epic fantasy, not high fantasy.</s>
Label encoding: <s>Then Malazan is probably epic fantasy, not high fantasy.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, the counterjerk there should be in support of you</s>
Label encoding: <s>Yes, the counterjerk there should be in support of you</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>lol, I can't believe you gave this guy a delta.</s>
Label encoding: <s>lol, I can't believe you gave this guy a delta.</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Could the flash just pull him in to the speed force?</s>
Label encoding: <s>Could the flash just pull him in to the speed force?</s>
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Maybe they expect people not to be assholes about it?</s>
Label encoding: <s>Maybe they expect people not to be assholes about it?</s>
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Whose mind are you trying to change here? :-)</s>
Label encoding: <s>Whose mind are you trying to change here? :-)</s>
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [URL].jpg [NEWLINE] [NEWLINE] this image comes to mind.</s>
Label encoding: <s> [URL].jpg [NEWLINE] [NEWLINE] this image comes to mind.</s>
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Exactly. Nothing is scarier than the unknown. </s><pad>
Label encoding: <s>Exactly. Nothing is scarier than the unknown. </s><pad>
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0915, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Dear DeltaBot, please rescan my comment.</s>
Label encoding: <s>Dear DeltaBot, please rescan my comment.</s>
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.0040e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.4757e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Am 19, can confirm I know nothing </s>
Label encoding: <s>Am 19, can confirm I know nothing </s>
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>And it's compulsory for some reason. </s>
Label encoding: <s>And it's compulsory for some reason. </s>
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.3468e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You gotta roll it back, brah.</s>
Label encoding: <s>You gotta roll it back, brah.</s>
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(8.6712e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The constitution has changed<mask> then.</s>
Label encoding: <s>The constitution has changed since then.</s>
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Definitely a consideration.</s>
Label encoding: <s>Definitely a consideration.</s>
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.9916e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.1391e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3626e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4883e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.5842e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.9723e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.4263e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.7090e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4781e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.1937e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.5770e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6814e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4381e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.3256e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3927e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.7710e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.1924e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.2541e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.5890e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4921e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6772e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3274e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6721e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.3746e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.2965e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4150e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.1336e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.6351e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8880e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3027e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4327e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.1749e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.6256e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.2662e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.8977e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[deleted]</s>
Label encoding: <s>[deleted]</s>
Loss: tensor(3.2952e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.1997e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3858e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.9875e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.5284e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.5159e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.6258e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3632e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.0897e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(5.5112e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.4777e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.2611e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.2499e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(6.5830e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.0691e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.3018e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.1440e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(2.8690e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0001, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8362e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3159e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(2.8266e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(7.7485e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7429e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(4.3077e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0004, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.1829e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8213e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.4024e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.0259e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.7167e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.3613e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8690e-05, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(3.8527e-05, device='cuda:0', grad_fn=<NllLossBackward>)
--------------EPOCH 5-------------
Test Accuracy: tensor(0.7027, device='cuda:0')
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Saving model at iteration: 0
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0367, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0583, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0385, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Integration has a number of aspects, and I suspect the debate will become complicated<mask> we don't analyze them and distinguish between different types of integration and the justifications for each. [NEWLINE] [NEWLINE] [STARTQ]...<mask> today their ancestors chastise Mexicans and Arabs for not learning English... [ENDQ] [NEWLINE] Yes. Nations need a certain amount of *coordination* to flourish.  For example, an Australian or Brit who immigrates to the US has to give up his proud cultural tradition of driving on the left.  This isn't a judgment about which side of the street is a *better* side to drive on. It's purely a relative judgment. <mask> you drive on the left side of the highway in Sydney, nothing unusual happens;<mask> you drive on the left side of the highway in NYC, you're doing to have a head on collision. It's beneficial to a nation to have everyone driving on the same side of the road.  That's<mask> *coordination* is. [NEWLINE] [NEWLINE] That applies to language<mask> well.  One of the strengths of America, historically, is that we've had one national language that everyone can communicate in.  We've always had large immigrant communities,<mask> never had a problem with large segments of the population being excluded from American society, or rivalry between different linguistic blocs, or the public sphere fragmenting into two different discussions.  Getting everyone to speak English is just good coordination. [NEWLINE] [NEWLINE] There are other kinds of coordination, too - different types of behavior and norms that are fine<mask> everyone follows them,<mask> bad<mask> they undermine<mask> everyone else is trying to accomplish with their behavior. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the idea that immigrants need to integrate into the culture of their host nations stems from racism, or at the very least a feeling that their culture is somehow superior. [ENDQ] [NEWLINE] Yes! It does stem from a feeling of cultural superiority.<mask> you are moving from Country A to Country B, then<mask><mask> we need to assume one of the following two things is going on: [NEWLINE] [NEWLINE] 1. Everything was fine<mask> you were living in A,<mask> you love the culture B<mask> much that you decide to put up with the difficulty of the move. [NEWLINE] [NEWLINE] 2. Life in A was getting to be extremely shitty in one way or another,<mask><mask> your indifference, contempt, or outright loathing for the culture of B, you decided to move to B. [NEWLINE] [NEWLINE] People in situation -1- obviously don't need to be encouraged to integrate.  People in situation -2-,<mask>
Label encoding: <s>Integration has a number of aspects, and I suspect the debate will become complicated if we don't analyze them and distinguish between different types of integration and the justifications for each. [NEWLINE] [NEWLINE] [STARTQ]... yet today their ancestors chastise Mexicans and Arabs for not learning English... [ENDQ] [NEWLINE] Yes. Nations need a certain amount of *coordination* to flourish.  For example, an Australian or Brit who immigrates to the US has to give up his proud cultural tradition of driving on the left.  This isn't a judgment about which side of the street is a *better* side to drive on. It's purely a relative judgment.  If you drive on the left side of the highway in Sydney, nothing unusual happens; if you drive on the left side of the highway in NYC, you're doing to have a head on collision. It's beneficial to a nation to have everyone driving on the same side of the road.  That's what *coordination* is. [NEWLINE] [NEWLINE] That applies to language as well.  One of the strengths of America, historically, is that we've had one national language that everyone can communicate in.  We've always had large immigrant communities, but never had a problem with large segments of the population being excluded from American society, or rivalry between different linguistic blocs, or the public sphere fragmenting into two different discussions.  Getting everyone to speak English is just good coordination. [NEWLINE] [NEWLINE] There are other kinds of coordination, too - different types of behavior and norms that are fine if everyone follows them, but bad if they undermine what everyone else is trying to accomplish with their behavior. [NEWLINE] [NEWLINE] [STARTQ] I think the idea that immigrants need to integrate into the culture of their host nations stems from racism, or at the very least a feeling that their culture is somehow superior. [ENDQ] [NEWLINE] Yes! It does stem from a feeling of cultural superiority. If you are moving from Country A to Country B, then I think we need to assume one of the following two things is going on: [NEWLINE] [NEWLINE] 1. Everything was fine when you were living in A, but you love the culture B so much that you decide to put up with the difficulty of the move. [NEWLINE] [NEWLINE] 2. Life in A was getting to be extremely shitty in one way or another, so despite your indifference, contempt, or outright loathing for the culture of B, you decided to move to B. [NEWLINE] [NEWLINE] People in situation -1- obviously don't need to be encouraged to integrate.  People in situation -2-, on
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I debated whether to respond to you or simply to post this to /r/badeconomics and be done with it,<mask> decided on the former. [NEWLINE] [NEWLINE] You have 2 separate views here: 1,<mask> you have some misunderstanding about<mask> it means to be in economic crisis,<mask> you come up with pearls like "The European crisis is the hoax of the century", and 2,<mask> you are saying that you disagree with the economic indicators we use to measure economic wellbeing (I am talking about your assertion that the EU is doing great based on things like gender equality and HDI). I will start with CMV1. [NEWLINE] [NEWLINE] "Economic crisis" is a bit of a hand-wavy word,<mask><mask> it's getting at is the general conditions of a recession, which has a specific definition: it means negative GDP growth for two consecutive quarters or more. [<mask> we can see here] ( [URL].GDP.MKTP.KD.ZG/countries/EU?display=graph), the EU experienced a double-dip recession starting in 2008 and is currently in recover from the second dip, which began in 2011 - which is exactly<mask> you would have been told in any mainstream media outlet.<mask> a recession is a sufficient condition for calling a situation an economic crisis,<mask><mask> we can agree here.<mask>, on top of that, it is extremely true that the unemployment situation in the South is very bad,<mask> is the fallout from the social security obligations which you like<mask> much. The latter two issues are examples of the EU's structural economic crisis, which is the specific thing everybody's worried about. It differs from your run of the mill recession<mask> it is caused by the economic system not working, and can only be fixed by reengineering that system. [NEWLINE] [NEWLINE] Next, you're making the common American mistake of<mask><mask> the EU is a country. Which it's not. We must further make a distinction between the EU and the Eurozone (which countries like the UK or Sweden don't belong to), and the Schengen area. The EU crisis, debt or otherwise, disproportionately affects the Eurozone,<mask> the price of the Euro essentially reflects everything every Euro country does. Finally,<mask> of the way EU legislation works, each country is basically free to implement whatever economic policies they want, which makes a whole lot of variation in economic performance across the board, which you<mask> will have heard in any mainstream media outlet: some EU countries are doing pretty well through the crisis,<mask>
Label encoding: <s>I debated whether to respond to you or simply to post this to /r/badeconomics and be done with it, but decided on the former. [NEWLINE] [NEWLINE] You have 2 separate views here: 1, where you have some misunderstanding about what it means to be in economic crisis, so you come up with pearls like "The European crisis is the hoax of the century", and 2, where you are saying that you disagree with the economic indicators we use to measure economic wellbeing (I am talking about your assertion that the EU is doing great based on things like gender equality and HDI). I will start with CMV1. [NEWLINE] [NEWLINE] "Economic crisis" is a bit of a hand-wavy word, but what it's getting at is the general conditions of a recession, which has a specific definition: it means negative GDP growth for two consecutive quarters or more. [ As we can see here] ( [URL].GDP.MKTP.KD.ZG/countries/EU?display=graph), the EU experienced a double-dip recession starting in 2008 and is currently in recover from the second dip, which began in 2011 - which is exactly what you would have been told in any mainstream media outlet. Since a recession is a sufficient condition for calling a situation an economic crisis, I think we can agree here. But, on top of that, it is extremely true that the unemployment situation in the South is very bad, as is the fallout from the social security obligations which you like so much. The latter two issues are examples of the EU's structural economic crisis, which is the specific thing everybody's worried about. It differs from your run of the mill recession because it is caused by the economic system not working, and can only be fixed by reengineering that system. [NEWLINE] [NEWLINE] Next, you're making the common American mistake of assuming that the EU is a country. Which it's not. We must further make a distinction between the EU and the Eurozone (which countries like the UK or Sweden don't belong to), and the Schengen area. The EU crisis, debt or otherwise, disproportionately affects the Eurozone, because the price of the Euro essentially reflects everything every Euro country does. Finally, because of the way EU legislation works, each country is basically free to implement whatever economic policies they want, which makes a whole lot of variation in economic performance across the board, which you also will have heard in any mainstream media outlet: some EU countries are doing pretty well through the crisis, but
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] 1. A united Empire would be a stronger bulwark against Dominion aggression.<mask>, the Empire is generally good for Skyrim. [ENDQ] [NEWLINE] <mask> clearly the Empire is not united<mask> it appeared to be. <mask> it were this dissent for it wouldn't have started in the first place.   There is only one party to blame<mask> it comes to this scenario and that is leadership. [NEWLINE] [NEWLINE] [STARTQ] This should be fairly obvious.<mask> Skyrim was to secede from the Empire, then High Rock would be cut off from Cyrodiil, and could very well secede in turn. Hammerfell was already kicked out of the Empire, and whatever is left of Morrowind that has not been destroyed by the eruption of Mt. Vvardenfell lies under the de facto control of the Kingdom of Argonia.<mask><mask><mask><mask> a successful Stormcloak rebellion, the Dominion would be able to destroy the nations of men piecemeal. [ENDQ] [NEWLINE] <mask> at-least this would actually involve fighting instead of having a puppet government.  Wouldn't you rather go down fighting than submit to this?  Just try to imagine this in a real world terms imagine the government of the nation you are in is overthrown by a foreign power and has occupational forces (which you see all over Skyrim) in your lands  with your new government giving into any and all demands.  Would you really side with that? [NEWLINE] [NEWLINE] [STARTQ] Of course, a quick and decisive victory for the rebels is not quite optimal for the Thalmor; they would prefer that both sides bleed one another out (and don't forget that Ulfric is a Thalmor asset of a sort).<mask><mask> they cannot have that, I'm sure that they would prefer to take on a bunch of smaller human kingdoms than one larger human Empire. Of course, for any of those who actually like the Dominion, this is a wonderful reason to join the Stormcloaks. [ENDQ] [NEWLINE] With the puppet government installed and under their thumb the rebellion only stands to hurt them in the end. <mask> they wanted to murder all humans they would have simply continued their crusade<mask> they were clearly winning to begin with.   Instead the human races are being beaten into submission and it's sicking that other humans are taking part in it. [NEWLINE] [NEWLINE] [STARTQ] There is<mask> the matter of trade and mutual benefit. Jarl Balgruuf says<mask> much. Skyrim gets to benefit from the rich lands to the south, and in exchange Cyrodiil gets the protection of some
Label encoding: <s> [STARTQ] 1. A united Empire would be a stronger bulwark against Dominion aggression. Also, the Empire is generally good for Skyrim. [ENDQ] [NEWLINE] But clearly the Empire is not united as it appeared to be.  If it were this dissent for it wouldn't have started in the first place.   There is only one party to blame when it comes to this scenario and that is leadership. [NEWLINE] [NEWLINE] [STARTQ] This should be fairly obvious. If Skyrim was to secede from the Empire, then High Rock would be cut off from Cyrodiil, and could very well secede in turn. Hammerfell was already kicked out of the Empire, and whatever is left of Morrowind that has not been destroyed by the eruption of Mt. Vvardenfell lies under the de facto control of the Kingdom of Argonia. In the event of a successful Stormcloak rebellion, the Dominion would be able to destroy the nations of men piecemeal. [ENDQ] [NEWLINE] But at-least this would actually involve fighting instead of having a puppet government.  Wouldn't you rather go down fighting than submit to this?  Just try to imagine this in a real world terms imagine the government of the nation you are in is overthrown by a foreign power and has occupational forces (which you see all over Skyrim) in your lands  with your new government giving into any and all demands.  Would you really side with that? [NEWLINE] [NEWLINE] [STARTQ] Of course, a quick and decisive victory for the rebels is not quite optimal for the Thalmor; they would prefer that both sides bleed one another out (and don't forget that Ulfric is a Thalmor asset of a sort). But if they cannot have that, I'm sure that they would prefer to take on a bunch of smaller human kingdoms than one larger human Empire. Of course, for any of those who actually like the Dominion, this is a wonderful reason to join the Stormcloaks. [ENDQ] [NEWLINE] With the puppet government installed and under their thumb the rebellion only stands to hurt them in the end.  If they wanted to murder all humans they would have simply continued their crusade as they were clearly winning to begin with.   Instead the human races are being beaten into submission and it's sicking that other humans are taking part in it. [NEWLINE] [NEWLINE] [STARTQ] There is also the matter of trade and mutual benefit. Jarl Balgruuf says as much. Skyrim gets to benefit from the rich lands to the south, and in exchange Cyrodiil gets the protection of some
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Again, thank you for your helpful responses.  I've done some real thinking, thanks in large part to your critiques, and think I've changed some of the assumptions I hold which have led to the view in this thread.  I have not changed my particular view about, yes, the objective lack of quality that,<mask> not essential to the Trap genre, is nearly universal. <mask> I have changed is the idea that this matters.  I'll do my best to explain myself. [NEWLINE] [NEWLINE] First, (1)I draw a sharp distinction between "quality" and "enjoyment".  One can enjoy poor quality music.  A couple of my mistakes were thinking that (2) there is any way to put that respectfully without sounding pejorative or like I have a stick<mask> far up my ass that it's coming out my mouth and (3) that a lot of others share this opinion<mask> an assumption to<mask> they look at music.  One of my all-time favorite bands whom I love dearly has a history of poor quality music.  I love their lyrics and I love the "sound" that the band creates,<mask> only recently have they proved themselves<mask> being good technical musicians.  On the one hand, I really enjoy this band; on the other, I don't consider them extraordinarily skilled musically.  I hold both those views without cognitive dissonance.  Not many people do,<mask>.  They equate aesthetic pleasure with aesthetic quality. [NEWLINE] [NEWLINE] Another one of my assumptions that has been debated in this thread is that (4) we can make objective judgments about music.  This view has not changed.  This is not to say that music is wholly objective.  Furthermore, the fact that music has strong subjective elements does not negate those objective elements.  I don't enjoy a lot of music that I respect<mask> having quality in creativity, originality, complexity, structure, etc.  Some of it I can't stand.  I still think they're quality compositions and performances.  Is there an objective qualitative difference between "Bohemian Rhapsody" and "Shake It Off"?  Of course.  Can someone find more enjoyment in "Shake It Off" than "Bohemian Rhapsody"?  Of course.  Just<mask> the criterion for making such a claim isn't immediately obvious doesn't mean we should discount the claim.  Just<mask> "music is an entirely objective matter" is false, "music is an entirely subjective manner" is false.
Label encoding: <s>Again, thank you for your helpful responses.  I've done some real thinking, thanks in large part to your critiques, and think I've changed some of the assumptions I hold which have led to the view in this thread.  I have not changed my particular view about, yes, the objective lack of quality that, if not essential to the Trap genre, is nearly universal.  What I have changed is the idea that this matters.  I'll do my best to explain myself. [NEWLINE] [NEWLINE] First, (1)I draw a sharp distinction between "quality" and "enjoyment".  One can enjoy poor quality music.  A couple of my mistakes were thinking that (2) there is any way to put that respectfully without sounding pejorative or like I have a stick so far up my ass that it's coming out my mouth and (3) that a lot of others share this opinion as an assumption to how they look at music.  One of my all-time favorite bands whom I love dearly has a history of poor quality music.  I love their lyrics and I love the "sound" that the band creates, but only recently have they proved themselves as being good technical musicians.  On the one hand, I really enjoy this band; on the other, I don't consider them extraordinarily skilled musically.  I hold both those views without cognitive dissonance.  Not many people do, though.  They equate aesthetic pleasure with aesthetic quality. [NEWLINE] [NEWLINE] Another one of my assumptions that has been debated in this thread is that (4) we can make objective judgments about music.  This view has not changed.  This is not to say that music is wholly objective.  Furthermore, the fact that music has strong subjective elements does not negate those objective elements.  I don't enjoy a lot of music that I respect as having quality in creativity, originality, complexity, structure, etc.  Some of it I can't stand.  I still think they're quality compositions and performances.  Is there an objective qualitative difference between "Bohemian Rhapsody" and "Shake It Off"?  Of course.  Can someone find more enjoyment in "Shake It Off" than "Bohemian Rhapsody"?  Of course.  Just because the criterion for making such a claim isn't immediately obvious doesn't mean we should discount the claim.  Just as "music is an entirely objective matter" is false, "music is an entirely subjective manner" is false.
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0508, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you are leaving out some pretty important points<mask> it comes to<mask> efficient a human can be. [NEWLINE] [NEWLINE] [STARTQ] we use an average of 100 watts<mask> a computer today uses up to 1 kilowatt....It is actually MUCH MUCH easier to power a person than a robot. Even considering a 90% loss in energy<mask> making food organic life is still more efficient. [ENDQ] [NEWLINE] Well yeah,<mask> you directly compare the two I can see<mask> a human can be more efficient than certain machines.<mask> that doesn't take into account all the energy needed just to sustain a human. [This]( [URL].jpg) is<mask> big the ISS is, it can hold up to ten people at a time. You need something literally the size of a football field<mask> astronauts can live in space up to five months. We are constantly shipping up materials<mask> they can live, it is no<mask> close to being efficient enough to sustain space traversing organisms. Biological organisms need lots of resources to survive.<mask> you want to build a self-sustaining spaceship it would have to be gigantic. And<mask> many organisms would you have to bring to keep the population alive? We kinda already have a gigantic spaceship, it's called earth. Using us<mask> an example, we need a ship 24,901 miles around just to keep on living. And it's not a 100% closed loop system either, eventually we will run out of resources. Whether it be 100 years or 1 billion it will happen. [NEWLINE] [NEWLINE] [STARTQ] They could hit something, some part of the ship could fail, things happen. Then the robot would need to be repaired<mask> most damage done to people is self repaired and easily supplemented. and don't forget these are incredibly advanced organisms built from scratch to be very resilient [ENDQ] [NEWLINE] <mask> would a robot need it's own ship? It would be the ship.  Take [satellites]( [URL].jpg) for example, they are kinda like little robots. Same with the [Mars rover]( [URL].jpg/1920px-PIA15279_3rovers-stand_D2011_1215_D521.jpg). These vehicles are much more resilient than biological organisms, that's<mask> we send them into space instead of humans.<mask>, look back at the satellite picture. Note<mask> each iteration is exponentially smaller than the last. We have to rely on a little thing called evolution in order to change whereas an artificial machine can be built smaller, more efficient, faster, better, etc. at whatever speed we can create them
Label encoding: <s>I think you are leaving out some pretty important points when it comes to how efficient a human can be. [NEWLINE] [NEWLINE] [STARTQ] we use an average of 100 watts while a computer today uses up to 1 kilowatt....It is actually MUCH MUCH easier to power a person than a robot. Even considering a 90% loss in energy when making food organic life is still more efficient. [ENDQ] [NEWLINE] Well yeah, when you directly compare the two I can see how a human can be more efficient than certain machines. But that doesn't take into account all the energy needed just to sustain a human. [This]( [URL].jpg) is how big the ISS is, it can hold up to ten people at a time. You need something literally the size of a football field so astronauts can live in space up to five months. We are constantly shipping up materials so they can live, it is no where close to being efficient enough to sustain space traversing organisms. Biological organisms need lots of resources to survive. If you want to build a self-sustaining spaceship it would have to be gigantic. And how many organisms would you have to bring to keep the population alive? We kinda already have a gigantic spaceship, it's called earth. Using us as an example, we need a ship 24,901 miles around just to keep on living. And it's not a 100% closed loop system either, eventually we will run out of resources. Whether it be 100 years or 1 billion it will happen. [NEWLINE] [NEWLINE] [STARTQ] They could hit something, some part of the ship could fail, things happen. Then the robot would need to be repaired while most damage done to people is self repaired and easily supplemented. and don't forget these are incredibly advanced organisms built from scratch to be very resilient [ENDQ] [NEWLINE] Why would a robot need it's own ship? It would be the ship.  Take [satellites]( [URL].jpg) for example, they are kinda like little robots. Same with the [Mars rover]( [URL].jpg/1920px-PIA15279_3rovers-stand_D2011_1215_D521.jpg). These vehicles are much more resilient than biological organisms, that's why we send them into space instead of humans. Also, look back at the satellite picture. Note how each iteration is exponentially smaller than the last. We have to rely on a little thing called evolution in order to change whereas an artificial machine can be built smaller, more efficient, faster, better, etc. at whatever speed we can create them
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> such, it is not voting for a republican or a democrat that makes you more partisan,<mask> thinking of yourself<mask> a republican or democrat (or any label, really). [ENDQ] [NEWLINE] <mask><mask> with this.<mask>, by voting for one, many people become inclined to think of themselves that way. Participation in a group (i.e. voting for its representatives) is a huge step towards thinking of yourself<mask> part of that group. Beyond that, those groups arise<mask> a natural, almost inevitable consequence of voting--only two groups can reasonably compete<mask> there is a single spot on the line, and the two loudest groups will be the competitors. [NEWLINE] [NEWLINE] [STARTQ] in both cases, you're treating people<mask> their labels, rather than<mask> people. [ENDQ] [NEWLINE] These people actively present themselves<mask> labels. I'll use Mitt Romney<mask> an example here. From<mask> I've seen of his actions, I support many of them. During the presidential election,<mask>, he was essentially forced to pander to certain groups--there are things that,<mask> a Republican vying for the votes of extreme Republicans, he *had* to say.<mask>... yes, you are voting for the individual,<mask> with the way things are set up, those individuals have to tie themselves to groups. [NEWLINE] [NEWLINE] It isn't possible for a Republican or Democrat to hold the same values that I do and to express those values in a way that will get them elected to a high position in our country's current political atmosphere. This is not prejudice. It is a result of our system of government. [NEWLINE] [NEWLINE] [STARTQ] This is true,<mask> that doesn't mean you shouldn't vote, it means you should not join a party.<mask><mask><mask> you continue to actively think of yourself<mask> independent, you will be able to resist that trend. [ENDQ] [NEWLINE] That holds for me.<mask>, [59%]( [URL].aspx) of voters actively think of themselves<mask> either Democratic or Republican. Political parties are a direct result of a voting system like that of the United States; the majority of voters identify with these parties.<mask><mask><mask>, the ideals of these groups are<mask> push the country's thought, whether or not I join. [NEWLINE] [NEWLINE] [STARTQ] Wait, you're rational enough to see that he masses don't think,<mask> you're just going to sit back and let them run the elections? [ENDQ] [NEWLINE] The masses far, far outnumber the substance-based voters. Public opinion, including politics, is run largely on soundbites and "flavor-of-the
Label encoding: <s> [STARTQ] As such, it is not voting for a republican or a democrat that makes you more partisan, but thinking of yourself as a republican or democrat (or any label, really). [ENDQ] [NEWLINE] I agree with this. However, by voting for one, many people become inclined to think of themselves that way. Participation in a group (i.e. voting for its representatives) is a huge step towards thinking of yourself as part of that group. Beyond that, those groups arise as a natural, almost inevitable consequence of voting--only two groups can reasonably compete when there is a single spot on the line, and the two loudest groups will be the competitors. [NEWLINE] [NEWLINE] [STARTQ] in both cases, you're treating people as their labels, rather than as people. [ENDQ] [NEWLINE] These people actively present themselves as labels. I'll use Mitt Romney as an example here. From what I've seen of his actions, I support many of them. During the presidential election, however, he was essentially forced to pander to certain groups--there are things that, as a Republican vying for the votes of extreme Republicans, he *had* to say. So... yes, you are voting for the individual, but with the way things are set up, those individuals have to tie themselves to groups. [NEWLINE] [NEWLINE] It isn't possible for a Republican or Democrat to hold the same values that I do and to express those values in a way that will get them elected to a high position in our country's current political atmosphere. This is not prejudice. It is a result of our system of government. [NEWLINE] [NEWLINE] [STARTQ] This is true, but that doesn't mean you shouldn't vote, it means you should not join a party. So long as you continue to actively think of yourself as independent, you will be able to resist that trend. [ENDQ] [NEWLINE] That holds for me. Meanwhile, [59%]( [URL].aspx) of voters actively think of themselves as either Democratic or Republican. Political parties are a direct result of a voting system like that of the United States; the majority of voters identify with these parties. Because of this, the ideals of these groups are what push the country's thought, whether or not I join. [NEWLINE] [NEWLINE] [STARTQ] Wait, you're rational enough to see that he masses don't think, yet you're just going to sit back and let them run the elections? [ENDQ] [NEWLINE] The masses far, far outnumber the substance-based voters. Public opinion, including politics, is run largely on soundbites and "flavor-of-the
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>**TL;DR: In order for the American people to have a significant role with the current government and structure, we would all have to become significantly more informed than we currently are.** [NEWLINE] [NEWLINE] The American political system can be criticized on many levels. For example, many feel that the America's FPTP system is to blame,<mask> it reduces electoral competition<mask> per Duverger's law. Others feel that a more powerful criticism is that single-member districts allow legislators to engage in partisan gerrymandering, diluting the voter strength of the opposition party. Some blame draconinan Voter ID laws meant to disenfranchise certain demographic groups. [NEWLINE] [NEWLINE] Some blame America's campaign finance system, asserting that unrestricted donations from individuals and corporations corrupts the system and makes legislators beholden to their donors and backers rather than to their constituents. Some assert that the aforementioned problems are compounded by issues in the political economy: neoliberal economic policies that allowed the rich to take a higher share of national income and wealth enabled them to game the political system in their favor. [NEWLINE] [NEWLINE] <mask>, many political scientists look at<mask> governance is affected by both actors and institutions. The institutional design of the U.S. federal government necessities compromise;<mask>,<mask> the political parties are unwilling to work together (<mask> is the case now), the gears of government grind to a halt.<mask> per the conditional party government theory,<mask> the parties are internally homogeneous and polarized, efficiency increases in the House due to positive and negative gate-keeping by the Speaker,<mask> efficiency is worsened in the Senate due to the filibuster (unless one party has 60+ members).<mask><mask><mask>, during periods of divided government, little gets done. Furthermore, the constitutional design of the U.S. and the system of federalism could lead to political dysfunction. [NEWLINE] [NEWLINE] <mask> all of these are arguably problems (depending on your viewpoint), I feel that above all else, the main problem is a dangerously ill-informed populace. Most voters don't have any meaningful knowledge of American history, world history, political theory (knowing about the principles upon which America is based), economics, business, science, sociology, foreign policy, etc. I'm not saying that it's practical or expected for people to become experts in all of these fields,<mask> they need some basic knowledge<mask> that they can think for themselves to some degree. [NEWLINE] [NEWLINE] <mask>, people don't spend enough time critically thinking about the various issues in politics (both issues that affect them personally and issues that affect America at large), and don
Label encoding: <s>**TL;DR: In order for the American people to have a significant role with the current government and structure, we would all have to become significantly more informed than we currently are.** [NEWLINE] [NEWLINE] The American political system can be criticized on many levels. For example, many feel that the America's FPTP system is to blame, as it reduces electoral competition as per Duverger's law. Others feel that a more powerful criticism is that single-member districts allow legislators to engage in partisan gerrymandering, diluting the voter strength of the opposition party. Some blame draconinan Voter ID laws meant to disenfranchise certain demographic groups. [NEWLINE] [NEWLINE] Some blame America's campaign finance system, asserting that unrestricted donations from individuals and corporations corrupts the system and makes legislators beholden to their donors and backers rather than to their constituents. Some assert that the aforementioned problems are compounded by issues in the political economy: neoliberal economic policies that allowed the rich to take a higher share of national income and wealth enabled them to game the political system in their favor. [NEWLINE] [NEWLINE] Moreover, many political scientists look at how governance is affected by both actors and institutions. The institutional design of the U.S. federal government necessities compromise; however, when the political parties are unwilling to work together ( as is the case now), the gears of government grind to a halt. As per the conditional party government theory, when the parties are internally homogeneous and polarized, efficiency increases in the House due to positive and negative gate-keeping by the Speaker, but efficiency is worsened in the Senate due to the filibuster (unless one party has 60+ members). As a result, during periods of divided government, little gets done. Furthermore, the constitutional design of the U.S. and the system of federalism could lead to political dysfunction. [NEWLINE] [NEWLINE] While all of these are arguably problems (depending on your viewpoint), I feel that above all else, the main problem is a dangerously ill-informed populace. Most voters don't have any meaningful knowledge of American history, world history, political theory (knowing about the principles upon which America is based), economics, business, science, sociology, foreign policy, etc. I'm not saying that it's practical or expected for people to become experts in all of these fields, but they need some basic knowledge so that they can think for themselves to some degree. [NEWLINE] [NEWLINE] Moreover, people don't spend enough time critically thinking about the various issues in politics (both issues that affect them personally and issues that affect America at large), and don
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The problem here is the simple question "<mask> is talent," that that is definitely hard to answer.<mask> : [NEWLINE] [NEWLINE] 1)<mask> practice alone will take you very far, excellence is something that practice alone will not provide. Everyone knows, for example, the extraordinary dedication to practice that guitarist Andres Segovia had during his lifetime. He is reported to have spent hours a day just doing scales and fingering exercises.<mask>, during his lifetime there were thousands of classical guitarists the world over who worked just<mask> hard<mask> Segovia, and<mask> they were undoubtedly extremely good, they were neither<mask> good nor<mask> innovative<mask> Segovia was. Segovia literally changed the way people played the guitar at the level of basic techniques. Something that rarely happens for established instruments. [NEWLINE] [NEWLINE] The same thing can be seen in many fields of endeavor. For example, there are without question huge numbers of young computer programmers who work diligently trying to build their own open source projects. Git is filled with them. Many of them work very hard at their programming.<mask> the level of quality difference between their code and the code of, say, a young Linux Torvalds is huge. [NEWLINE] [NEWLINE] People who become very good practice.<mask> of the people who practice some become much, much better than others. Even<mask> they have the same teachers and follow the same path. [NEWLINE] [NEWLINE] 2) Former world chess champion Garry Kasparov was famous for his work ethic. He would spend upwards of nearly all of his waking hours preparing for his next match or studying chess. He set a new bar for hard work. His comment on the question was that "the ability to work hard is<mask> a talent." He contended that it was<mask> he enjoyed<mask> he was doing, actually gaining pleasure from the act of deliberate practice, that he was able to work that hard in the first place.  This is a part of the question that seems to be missed by many people. [NEWLINE] [NEWLINE] There is no doubt that different people enjoy different activities. I, for example, tried to play football<mask> a young person. I practiced daily, I had a great coach, I wanted to be good.<mask><mask> it came right down to it, I hated getting hit. I hated the practice itself. I didn't find it an enjoyable activity. I was an adequate player, and perhaps had I stayed with it I might have become good<mask> my lack of enjoyment.<mask> there was simply no getting around the fact that I was miserable 6 months of the year<mask> I
Label encoding: <s>The problem here is the simple question " what is talent," that that is definitely hard to answer. However : [NEWLINE] [NEWLINE] 1) While practice alone will take you very far, excellence is something that practice alone will not provide. Everyone knows, for example, the extraordinary dedication to practice that guitarist Andres Segovia had during his lifetime. He is reported to have spent hours a day just doing scales and fingering exercises. However, during his lifetime there were thousands of classical guitarists the world over who worked just as hard as Segovia, and while they were undoubtedly extremely good, they were neither as good nor as innovative as Segovia was. Segovia literally changed the way people played the guitar at the level of basic techniques. Something that rarely happens for established instruments. [NEWLINE] [NEWLINE] The same thing can be seen in many fields of endeavor. For example, there are without question huge numbers of young computer programmers who work diligently trying to build their own open source projects. Git is filled with them. Many of them work very hard at their programming. But the level of quality difference between their code and the code of, say, a young Linux Torvalds is huge. [NEWLINE] [NEWLINE] People who become very good practice. But of the people who practice some become much, much better than others. Even when they have the same teachers and follow the same path. [NEWLINE] [NEWLINE] 2) Former world chess champion Garry Kasparov was famous for his work ethic. He would spend upwards of nearly all of his waking hours preparing for his next match or studying chess. He set a new bar for hard work. His comment on the question was that "the ability to work hard is also a talent." He contended that it was because he enjoyed what he was doing, actually gaining pleasure from the act of deliberate practice, that he was able to work that hard in the first place.  This is a part of the question that seems to be missed by many people. [NEWLINE] [NEWLINE] There is no doubt that different people enjoy different activities. I, for example, tried to play football as a young person. I practiced daily, I had a great coach, I wanted to be good. But when it came right down to it, I hated getting hit. I hated the practice itself. I didn't find it an enjoyable activity. I was an adequate player, and perhaps had I stayed with it I might have become good despite my lack of enjoyment. But there was simply no getting around the fact that I was miserable 6 months of the year because I
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do inform me [ENDQ] [NEWLINE] I already did.  Twice now,<mask><mask>.  The verses you refer to in Leviticus are within a certain context in Jewish history<mask> we do not reside.  These were commands concerning cultural separation from Canaanites and Phoenicians including commands not to do some of the things they did.  Those commands are some of more than 600 that are contradictory and have open to interpretation<mask> before the time of Christ. <mask><mask> the prohibition on homosexuality is dubiously justified *only* in the Old Testament and frankly makes very little intuitive sense given the words of Christ; I'm inclined to think it either isn't wrong or is at least not important.  The latter is the most I would possibly concede given the evidence, and that's being generous. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> happened to the rich man in the parable of Lazarus? [ENDQ] [NEWLINE] First, the term used there is likely *hades* which would probably have been a translation from *sheol* or *gehenna*.  Jewish audience and all that. <mask><mask> the term used was *sheol*, we can look at the specific reference to Abraham and compare it to the Jewish concept of the Bosom of Abraham, which was essentially the place for the righteous in *sheol*.  In this context, all it means is being dissevered from God's presence.  For an audience that desired communion with God, that likely would have been a torment in and of itself. [NEWLINE] [NEWLINE] [STARTQ] Or the parable of the fishermen on the docks? [ENDQ] [NEWLINE] This one is *gehenna*.  Once again, note the way the story changes<mask> this is about positive and constructive belief, rather than divine coercion.  The former makes more sense to me and is more in tune with the entirety of Christ's message. [NEWLINE] [NEWLINE] [STARTQ] Or in Revelations? [ENDQ] [NEWLINE] Interestingly enough, the Book of Revelation is not eschatological (discussing the end of the world) in my view.  It's an allegory concerning the conflict between early Christians and Romans; specifically Christian apostasy for convenience and a veiled mystical threat against Rome.  That's<mask> in a generally reasonable and narrative New Testament we come along and find this proto-Lord of the Rings spectacle.  In other words, I don't consider Revelation to be a wholly legitimate part of the Bible.  You have to remember that no divine authority established the Biblical canon and it's likely that some dicey stuff got through the filter. [NEWLINE] [NEWLINE] [STARTQ] I
Label encoding: <s> [STARTQ] Do inform me [ENDQ] [NEWLINE] I already did.  Twice now, in fact.  The verses you refer to in Leviticus are within a certain context in Jewish history where we do not reside.  These were commands concerning cultural separation from Canaanites and Phoenicians including commands not to do some of the things they did.  Those commands are some of more than 600 that are contradictory and have open to interpretation since before the time of Christ.  Given that the prohibition on homosexuality is dubiously justified *only* in the Old Testament and frankly makes very little intuitive sense given the words of Christ; I'm inclined to think it either isn't wrong or is at least not important.  The latter is the most I would possibly concede given the evidence, and that's being generous. [NEWLINE] [NEWLINE] [STARTQ] So what happened to the rich man in the parable of Lazarus? [ENDQ] [NEWLINE] First, the term used there is likely *hades* which would probably have been a translation from *sheol* or *gehenna*.  Jewish audience and all that.  So if the term used was *sheol*, we can look at the specific reference to Abraham and compare it to the Jewish concept of the Bosom of Abraham, which was essentially the place for the righteous in *sheol*.  In this context, all it means is being dissevered from God's presence.  For an audience that desired communion with God, that likely would have been a torment in and of itself. [NEWLINE] [NEWLINE] [STARTQ] Or the parable of the fishermen on the docks? [ENDQ] [NEWLINE] This one is *gehenna*.  Once again, note the way the story changes when this is about positive and constructive belief, rather than divine coercion.  The former makes more sense to me and is more in tune with the entirety of Christ's message. [NEWLINE] [NEWLINE] [STARTQ] Or in Revelations? [ENDQ] [NEWLINE] Interestingly enough, the Book of Revelation is not eschatological (discussing the end of the world) in my view.  It's an allegory concerning the conflict between early Christians and Romans; specifically Christian apostasy for convenience and a veiled mystical threat against Rome.  That's why in a generally reasonable and narrative New Testament we come along and find this proto-Lord of the Rings spectacle.  In other words, I don't consider Revelation to be a wholly legitimate part of the Bible.  You have to remember that no divine authority established the Biblical canon and it's likely that some dicey stuff got through the filter. [NEWLINE] [NEWLINE] [STARTQ] I
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Just to get this out of the way first, I'm not debating<mask> the current law is, or<mask> people should obey the law. I'm talking about<mask> *should* be the case. [NEWLINE] [NEWLINE] Here's my logic: [NEWLINE] [NEWLINE] 1. Roads exist for the purpose of people getting around, *not* specifically for cars. The whole idea of "jaywalking" was a PR campaign put on by auto-makers. [Here's an article]( [URL] ) that explains a bit about the PR campaign towards "roads are for cars". [NEWLINE] 2. People in cities (myself included), use a bicycle<mask> their primary commuter vehicle and means to get around the city. [NEWLINE] 3. ~~Safety: Sometimes road conditions aren't safe for cyclists.<mask> a car hits a bike, things are generally much much worse for the bike,<mask> cyclists need the ability to ride safely.<mask> there's a pothole, they need to be able to swerve, or to stop in a place<mask> they have visibility, etc.~~ EDIT: removing this. Multiple people have pointed out that it is legal to carefully avoid hazards in the road,<mask> this isn't really relevant and is distracting from my main point, which is #4 [NEWLINE] 4. Convenience is critical. Coming to a full stop at a stop sign, staying behind a stopped car (e.g. not lane splitting), and not being able to ride through crosswalks seriously impede the flow of bicycle traffic and make it take longer to get anywhere. The only reason we can ever justify driving 60+ mph on a highway is for speed and convenience, and it matters just<mask> much for cyclists<mask> it does for drivers.<mask><mask>, for cyclists, at least they are only likely to kill themselves in an accident,<mask> highway speed vehicles can take out other people. [NEWLINE] 5. More people biking is better for public health and the environment, and<mask> should be encouraged, and following every car law discourages this [NEWLINE] 6. This is the weakest of all of the arguments,<mask> people already aren't following some of these laws,<mask> they deem the laws to be overly harsh. Having a set of laws more tailored for bicycles, or just granting certain exceptions to them, might encourage people to follow some of the laws that are really important, like passing on the left<mask> a car may be turning right. [NEWLINE] [NEWLINE] I realize this is a bit of a cop-out,<mask> I'm not specifying exactly<mask><mask><mask> the laws should be, or<mask> they
Label encoding: <s>Just to get this out of the way first, I'm not debating what the current law is, or if people should obey the law. I'm talking about what *should* be the case. [NEWLINE] [NEWLINE] Here's my logic: [NEWLINE] [NEWLINE] 1. Roads exist for the purpose of people getting around, *not* specifically for cars. The whole idea of "jaywalking" was a PR campaign put on by auto-makers. [Here's an article]( [URL] ) that explains a bit about the PR campaign towards "roads are for cars". [NEWLINE] 2. People in cities (myself included), use a bicycle as their primary commuter vehicle and means to get around the city. [NEWLINE] 3. ~~Safety: Sometimes road conditions aren't safe for cyclists. If a car hits a bike, things are generally much much worse for the bike, so cyclists need the ability to ride safely. If there's a pothole, they need to be able to swerve, or to stop in a place where they have visibility, etc.~~ EDIT: removing this. Multiple people have pointed out that it is legal to carefully avoid hazards in the road, so this isn't really relevant and is distracting from my main point, which is #4 [NEWLINE] 4. Convenience is critical. Coming to a full stop at a stop sign, staying behind a stopped car (e.g. not lane splitting), and not being able to ride through crosswalks seriously impede the flow of bicycle traffic and make it take longer to get anywhere. The only reason we can ever justify driving 60+ mph on a highway is for speed and convenience, and it matters just as much for cyclists as it does for drivers. In fact, for cyclists, at least they are only likely to kill themselves in an accident, where highway speed vehicles can take out other people. [NEWLINE] 5. More people biking is better for public health and the environment, and therefore should be encouraged, and following every car law discourages this [NEWLINE] 6. This is the weakest of all of the arguments, but people already aren't following some of these laws, because they deem the laws to be overly harsh. Having a set of laws more tailored for bicycles, or just granting certain exceptions to them, might encourage people to follow some of the laws that are really important, like passing on the left when a car may be turning right. [NEWLINE] [NEWLINE] I realize this is a bit of a cop-out, but I'm not specifying exactly what I think the laws should be, or how they
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Is the only way to prevent "causing" your fatal accident to not drive at all? [ENDQ] [NEWLINE] Car accidents, IIRC are one of the leading causes of death.<mask> yes. You should know that risk, and are partly to blame<mask> it happens. I don't drive. I know the risks and choose not to.<mask> I ride with someone else (rarely), I know that there's a significantly higher chance of death, and it scares the shit out of me. I ensure I'm with a good driver, driving at a safe time/location, and<mask> on to minimize those risks. [NEWLINE] [NEWLINE] Many of my friends<mask> avoid freeways/highways,<mask> those are<mask> high danger areas. They use GPS to find safer alternative roads. [NEWLINE] [NEWLINE] I'd say it's more situational awareness and safety, than object-based. Know the risks. [NEWLINE] [NEWLINE] [STARTQ] Is it reasonable to cover up<mask><mask> you were in the middle east just<mask> a man won't rape you? [ENDQ] [NEWLINE] Absolutely.<mask> uncovered women get raped often,<mask> covered women don't. And I was a woman in that area, I'd *choose* to cover up to be safer. Just like<mask> I choose to put on a jacket to stay warm and prevent colds. Yes,<mask> the disease wasn't there I wouldn't get sick.<mask> there's no use in blaming the disease and instead should minimize my risk of getting sick. Don't stay out in the cold, be sure to put on a jacket, don't hang around sick people, etc. [NEWLINE] [NEWLINE] [STARTQ] I'm not sure you can separate cause and blame. Your sentiment is essentially "<mask> you hadn't been doing x, then y would not have happened". You can dress it up by saying it's cause and not blame,<mask> it is blame. [ENDQ] [NEWLINE] Blame implies it's personal. It's not. It's not *you*, it's<mask> you did. Which are two distinct things. *<mask> * you did the things that increased risk is not really a concern. It could be choice, forced, whatever.<mask> you did, and<mask> it increased risk. Blame<mask> implies that one person is at fault. They aren't. There are multiple reasons for<mask> something happened. Had either person not been there, it wouldn't have happened. [NEWLINE] [NEWLINE] [STARTQ] Wearing a revealing outfit is a dangerous activity? [ENDQ] [NEWLINE] Statistics say no, and that rape is commonly done by someone you know. In this case, be wary of
Label encoding: <s> [STARTQ] Is the only way to prevent "causing" your fatal accident to not drive at all? [ENDQ] [NEWLINE] Car accidents, IIRC are one of the leading causes of death. So yes. You should know that risk, and are partly to blame if it happens. I don't drive. I know the risks and choose not to. When I ride with someone else (rarely), I know that there's a significantly higher chance of death, and it scares the shit out of me. I ensure I'm with a good driver, driving at a safe time/location, and so on to minimize those risks. [NEWLINE] [NEWLINE] Many of my friends also avoid freeways/highways, since those are also high danger areas. They use GPS to find safer alternative roads. [NEWLINE] [NEWLINE] I'd say it's more situational awareness and safety, than object-based. Know the risks. [NEWLINE] [NEWLINE] [STARTQ] Is it reasonable to cover up as if you were in the middle east just so a man won't rape you? [ENDQ] [NEWLINE] Absolutely. If uncovered women get raped often, while covered women don't. And I was a woman in that area, I'd *choose* to cover up to be safer. Just like how I choose to put on a jacket to stay warm and prevent colds. Yes, if the disease wasn't there I wouldn't get sick. But there's no use in blaming the disease and instead should minimize my risk of getting sick. Don't stay out in the cold, be sure to put on a jacket, don't hang around sick people, etc. [NEWLINE] [NEWLINE] [STARTQ] I'm not sure you can separate cause and blame. Your sentiment is essentially " if you hadn't been doing x, then y would not have happened". You can dress it up by saying it's cause and not blame, but it is blame. [ENDQ] [NEWLINE] Blame implies it's personal. It's not. It's not *you*, it's what you did. Which are two distinct things. * Why * you did the things that increased risk is not really a concern. It could be choice, forced, whatever. But you did, and so it increased risk. Blame also implies that one person is at fault. They aren't. There are multiple reasons for why something happened. Had either person not been there, it wouldn't have happened. [NEWLINE] [NEWLINE] [STARTQ] Wearing a revealing outfit is a dangerous activity? [ENDQ] [NEWLINE] Statistics say no, and that rape is commonly done by someone you know. In this case, be wary of
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Alright,<mask> I'm a typical American consumer. I have a smartphone; it's not a high end one<mask> it lets me browse the internet<mask> I'm out and about and I have a few apps for it. [NEWLINE] [NEWLINE] I was lucky enough to get to try the Google Glass at a friend's who was part of the group who got them early.<mask> frankly, I don't see the appeal of glass over, say, a regular smartphone like the one I have now. From<mask> I can tell, the smartphone I have is superior to Glass in every way. [NEWLINE] [NEWLINE] PRICE [NEWLINE] [NEWLINE] For one, Google Glass is $1,500. Compare that to just $300 for the 32GB Iphone 5s. I understand that it's brand new and the price will eventually come down,<mask> for now, that's way too much. [NEWLINE] [NEWLINE] SPECS [NEWLINE] [NEWLINE] * 5 Megapixel camera for Glass, compared to 8 Megapixel for the Iphone. [NEWLINE] * 720p video camera on Glass, compared to 1080p for the Iphone. [NEWLINE] * 12 GB data storage capacity for Glass, compared to 16, 32, or 64 GB for the Iphone. This doesn't take cloud storage into consideration,<mask>. [NEWLINE] [NEWLINE] CONVENIENCE [NEWLINE] [NEWLINE] One of the big things that I hear touted about Glass is convenience over a traditional smartphone, thereby justifying the price.<mask> let's break that down a bit. [NEWLINE] [NEWLINE] * I can't play games on Glass very easily, such<mask> Tetris or Angry Birds that I can easily play on my phone with the touch screen. These are games that sort of require a touch screen in order to play effectively. [NEWLINE] * The voice feature is nice, and is very responsive,<mask><mask><mask> I want to send a text to someone and don't want everyone around me to know<mask> I'm saying?<mask><mask> I want to visit a website<mask> not have everyone around me know? I could use the touchpad on the side,<mask> it's not<mask> quick<mask> the keypad on my phone. Even<mask> I got used to it, it's still probably still not<mask> quick. [NEWLINE] * Glass isn't a phone in and of itself, functioning more like a handsfree device.<mask> this means that I'll have to carry around my regular phone<mask><mask> to Glass.<mask> that means<mask> another device<mask><mask> to my phone that I have to charge every night and find a place to keep.<mask> I can't keep it in my pocket like my phone,<mask> it's
Label encoding: <s>Alright, so I'm a typical American consumer. I have a smartphone; it's not a high end one but it lets me browse the internet when I'm out and about and I have a few apps for it. [NEWLINE] [NEWLINE] I was lucky enough to get to try the Google Glass at a friend's who was part of the group who got them early. But frankly, I don't see the appeal of glass over, say, a regular smartphone like the one I have now. From what I can tell, the smartphone I have is superior to Glass in every way. [NEWLINE] [NEWLINE] PRICE [NEWLINE] [NEWLINE] For one, Google Glass is $1,500. Compare that to just $300 for the 32GB Iphone 5s. I understand that it's brand new and the price will eventually come down, but for now, that's way too much. [NEWLINE] [NEWLINE] SPECS [NEWLINE] [NEWLINE] * 5 Megapixel camera for Glass, compared to 8 Megapixel for the Iphone. [NEWLINE] * 720p video camera on Glass, compared to 1080p for the Iphone. [NEWLINE] * 12 GB data storage capacity for Glass, compared to 16, 32, or 64 GB for the Iphone. This doesn't take cloud storage into consideration, though. [NEWLINE] [NEWLINE] CONVENIENCE [NEWLINE] [NEWLINE] One of the big things that I hear touted about Glass is convenience over a traditional smartphone, thereby justifying the price. But let's break that down a bit. [NEWLINE] [NEWLINE] * I can't play games on Glass very easily, such as Tetris or Angry Birds that I can easily play on my phone with the touch screen. These are games that sort of require a touch screen in order to play effectively. [NEWLINE] * The voice feature is nice, and is very responsive, but what if I want to send a text to someone and don't want everyone around me to know what I'm saying? What if I want to visit a website but not have everyone around me know? I could use the touchpad on the side, but it's not as quick as the keypad on my phone. Even if I got used to it, it's still probably still not as quick. [NEWLINE] * Glass isn't a phone in and of itself, functioning more like a handsfree device. So this means that I'll have to carry around my regular phone in addition to Glass. So that means yet another device in addition to my phone that I have to charge every night and find a place to keep. But I can't keep it in my pocket like my phone, since it's
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm not talking about a straight guy getting turned down. I'm talking about a love quarrel<mask> the girl is torn between whether a or b is the best option and the guy decides to make the decision for her by eliminating the competition. I'm not saying it would be normal.<mask> it only takes one instance to put everyone else on edge. [ENDQ] [NEWLINE] <mask> is that any different than my example. It deals with people who are angry and turn to murder to get their frustrations in check. Unrequitted love can lead to murder just<mask> much<mask> competition.<mask> you go into abnormal situations, you can't pick and choose which ones you like. They are both equally valid. And equally unlikely. [NEWLINE] [NEWLINE] [STARTQ] It's a scientific field. You obviously don't understand Infantry. They exist to close with and engage the enemies of the United States. Their job is to make other people die before they themselves die in very cut and dry meaning. They don't do combat patrols with the hope of discovering the cure for cancer. [ENDQ] [NEWLINE] You obviously don't understand that military is a job,<mask> people take part not<mask> they are violent,<mask> for a million of reasons. People join for a host of reasons, they are from a poor background, deals to get out of jail,<mask> of education help, benefits. It's a job that involves shooting, yes,<mask><mask> does police work. Do policemen all over the world shoot each other to get rid of competition to other female cops? Think<mask> ridiculous your idea is.<mask> that is reason enough to eliminate women out of the military, they shouldn't be in any jobs! [NEWLINE] [NEWLINE] [STARTQ] They don't do combat patrols with the hope of discovering the cure for cancer. [ENDQ] [NEWLINE] You have no idea<mask> dirty being a scientist is. Last time I checked, I experiment on animals. At least in the army you kill and whoever you kill is dead. A lot of experiments require generating massive amounts of animals who are sick (to mimic disease), who you hurt (to mimic injury) or that you disect<mask> alive (to keep tissues fresh). You inject, and cut, and poison, and watch them die. Science is not for the faint of heart, or for saints. It's dirty and fucked up. The fact that all is done for a good cause does not take away the fact that you are doing it. [NEWLINE] [NEWLINE] [STARTQ] I'm saying that people will be people. There is a fine line between love and hate. It's dangerous enough to be on the front
Label encoding: <s> [STARTQ] I'm not talking about a straight guy getting turned down. I'm talking about a love quarrel where the girl is torn between whether a or b is the best option and the guy decides to make the decision for her by eliminating the competition. I'm not saying it would be normal. But it only takes one instance to put everyone else on edge. [ENDQ] [NEWLINE] How is that any different than my example. It deals with people who are angry and turn to murder to get their frustrations in check. Unrequitted love can lead to murder just as much as competition. If you go into abnormal situations, you can't pick and choose which ones you like. They are both equally valid. And equally unlikely. [NEWLINE] [NEWLINE] [STARTQ] It's a scientific field. You obviously don't understand Infantry. They exist to close with and engage the enemies of the United States. Their job is to make other people die before they themselves die in very cut and dry meaning. They don't do combat patrols with the hope of discovering the cure for cancer. [ENDQ] [NEWLINE] You obviously don't understand that military is a job, where people take part not because they are violent, but for a million of reasons. People join for a host of reasons, they are from a poor background, deals to get out of jail, because of education help, benefits. It's a job that involves shooting, yes, but so does police work. Do policemen all over the world shoot each other to get rid of competition to other female cops? Think how ridiculous your idea is. If that is reason enough to eliminate women out of the military, they shouldn't be in any jobs! [NEWLINE] [NEWLINE] [STARTQ] They don't do combat patrols with the hope of discovering the cure for cancer. [ENDQ] [NEWLINE] You have no idea how dirty being a scientist is. Last time I checked, I experiment on animals. At least in the army you kill and whoever you kill is dead. A lot of experiments require generating massive amounts of animals who are sick (to mimic disease), who you hurt (to mimic injury) or that you disect while alive (to keep tissues fresh). You inject, and cut, and poison, and watch them die. Science is not for the faint of heart, or for saints. It's dirty and fucked up. The fact that all is done for a good cause does not take away the fact that you are doing it. [NEWLINE] [NEWLINE] [STARTQ] I'm saying that people will be people. There is a fine line between love and hate. It's dangerous enough to be on the front
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It makes perfect sense<mask> to<mask> you want done. I'll pick some points out from<mask> you said to expand on it; [NEWLINE] [NEWLINE] [STARTQ] make them less stressful, less like lessons in a classroom and more like a bonding and growing experience, much more like a spiritual retreat; [ENDQ] [NEWLINE] You must understand that the bonding that occurs in the military are more than the average person will ever experience even<mask> they go to a bond building event. [NEWLINE] [NEWLINE] [STARTQ] like a spiritual retreat [ENDQ] [NEWLINE] Each military branch has it's own sub-branch within it designed to deal with only the spiritual. The chaplains are fantastic counselors even for those without a faith. They help out the most<mask> they know you on a more personal level,<mask><mask> can be removed from your life at the same time. [NEWLINE] [NEWLINE] [STARTQ] have them do things<mask> a group, [ENDQ] [NEWLINE] That's 99.99% of the Army, and I would say for the rest of the military. I'd have to say that the problem is the withdraw of that<mask> military members go into the civilian workforce<mask> group anything hardly happens.<mask> often to you do things with 10-30+ people that you know very well and have close relationships with? I'm happy being with my fiance and best friend (one in the same),<mask> some don't have that. Military is very close nit. [NEWLINE] [NEWLINE] [STARTQ] And I would argue the main problem with the services that are available is two-fold: (1) they're not mandatory, [ENDQ] [NEWLINE] ACAP is mandatory,<mask> mental health wasn't<mask> I was leaving. That's something they should do. That's something I really hope is actually going to happen/happening. [NEWLINE] [NEWLINE] [STARTQ] (2) there is a lingering social stigma relating to mental health and recovery. I have been actively working through depression for almost a year now, talking with a psychoanalyst, and working on some CBT, I am very acutely aware of this stigma. [ENDQ] [NEWLINE] shit sucks, doesn't it? Now imagine having those 10-30+ close relationships and 2-5 think you're broken and may snap at the wrong moment and pull the trigger in the wrong way. It was getting better on my way out,<mask> on my way out it wasn't a combat arms unit. I'm sure it's still happening, and it's a shame. It will end up being something like racism. Only time and new blood will make things like that lessen. [NEWLINE] [NEWLINE] [STARTQ] <mask>,<mask><mask> that<mask> the army, the organization
Label encoding: <s>It makes perfect sense as to what you want done. I'll pick some points out from what you said to expand on it; [NEWLINE] [NEWLINE] [STARTQ] make them less stressful, less like lessons in a classroom and more like a bonding and growing experience, much more like a spiritual retreat; [ENDQ] [NEWLINE] You must understand that the bonding that occurs in the military are more than the average person will ever experience even if they go to a bond building event. [NEWLINE] [NEWLINE] [STARTQ] like a spiritual retreat [ENDQ] [NEWLINE] Each military branch has it's own sub-branch within it designed to deal with only the spiritual. The chaplains are fantastic counselors even for those without a faith. They help out the most since they know you on a more personal level, but also can be removed from your life at the same time. [NEWLINE] [NEWLINE] [STARTQ] have them do things as a group, [ENDQ] [NEWLINE] That's 99.99% of the Army, and I would say for the rest of the military. I'd have to say that the problem is the withdraw of that when military members go into the civilian workforce where group anything hardly happens. How often to you do things with 10-30+ people that you know very well and have close relationships with? I'm happy being with my fiance and best friend (one in the same), but some don't have that. Military is very close nit. [NEWLINE] [NEWLINE] [STARTQ] And I would argue the main problem with the services that are available is two-fold: (1) they're not mandatory, [ENDQ] [NEWLINE] ACAP is mandatory, but mental health wasn't when I was leaving. That's something they should do. That's something I really hope is actually going to happen/happening. [NEWLINE] [NEWLINE] [STARTQ] (2) there is a lingering social stigma relating to mental health and recovery. I have been actively working through depression for almost a year now, talking with a psychoanalyst, and working on some CBT, I am very acutely aware of this stigma. [ENDQ] [NEWLINE] shit sucks, doesn't it? Now imagine having those 10-30+ close relationships and 2-5 think you're broken and may snap at the wrong moment and pull the trigger in the wrong way. It was getting better on my way out, but on my way out it wasn't a combat arms unit. I'm sure it's still happening, and it's a shame. It will end up being something like racism. Only time and new blood will make things like that lessen. [NEWLINE] [NEWLINE] [STARTQ] BUT, I think that if the army, the organization
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0562, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0494, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First, ∆ for you. I mentioned earlier<mask> the guide made me more paranoid about being around men than I was before<mask> of my attack. Particularly, PUA *freak* me out<mask> I felt they really don't care-that's<mask> some PUA come off<mask>.  It's very reassuring to find<mask> respectful both you and /u/alucard4571 act, definitely on this subject, [NEWLINE] [NEWLINE] I'm going to give him a alucard4571 anyway,<mask> he did change my mind,<mask> before I had already was aware there were good parts. I probably should have stated it in my first post. [NEWLINE] [NEWLINE] [STARTQ] : common PUA advice says "make it hard for her to say no." After reading your post, I will always do my best to make it easy for her to say no. [ENDQ] [NEWLINE] Exactly. I feel, giving a woman an object to say *yes* would<mask> help the guy too. Having the same attention directed back boosts confidence.<mask> is important,<mask>,<mask> she yes to one thing, it doesn't mean she will say yes to *everything* you do. That's<mask> asking her "<mask> does this feel" and "<mask> else would you like" is a good way to ask for consent in a heated moment. [NEWLINE] [NEWLINE] [STARTQ] In PUA parlance, "abundance" refers to the idea that there are plenty of beautiful women in the world and plenty of nights to go out [ENDQ] [NEWLINE] There is an abundance of woman out there and many nights to spend. The problem is, "woman" is a very vague term. All gender does is sort people who have different reproductive body parts.  Women aren't like apples on an apple tree-you can't just pluck them off the tree like any other apple. There are plenty of women out there who want to have sex, who want to have ONS, who want FWB, who enjoy healthily having more than one sex partner. There are plenty of women out there who enjoy giving BJ, who enjoy receiving oral, who enjoy having her breasts played with, who enjoy anal. The problem is, not every woman does, with any of these things. Even<mask> a girl wants a ONS, she might not enjoy giving BJs. She might just want to make out with you, she might just want to trade oral, she might only want vaginal sex-ect. There is an abundance of woman out there,<mask> clumping them all together would be harmful to
Label encoding: <s>First, ∆ for you. I mentioned earlier how the guide made me more paranoid about being around men than I was before because of my attack. Particularly, PUA *freak* me out because I felt they really don't care-that's what some PUA come off as.  It's very reassuring to find how respectful both you and /u/alucard4571 act, definitely on this subject, [NEWLINE] [NEWLINE] I'm going to give him a alucard4571 anyway, because he did change my mind, but before I had already was aware there were good parts. I probably should have stated it in my first post. [NEWLINE] [NEWLINE] [STARTQ] : common PUA advice says "make it hard for her to say no." After reading your post, I will always do my best to make it easy for her to say no. [ENDQ] [NEWLINE] Exactly. I feel, giving a woman an object to say *yes* would also help the guy too. Having the same attention directed back boosts confidence. What is important, though, if she yes to one thing, it doesn't mean she will say yes to *everything* you do. That's why asking her " How does this feel" and " what else would you like" is a good way to ask for consent in a heated moment. [NEWLINE] [NEWLINE] [STARTQ] In PUA parlance, "abundance" refers to the idea that there are plenty of beautiful women in the world and plenty of nights to go out [ENDQ] [NEWLINE] There is an abundance of woman out there and many nights to spend. The problem is, "woman" is a very vague term. All gender does is sort people who have different reproductive body parts.  Women aren't like apples on an apple tree-you can't just pluck them off the tree like any other apple. There are plenty of women out there who want to have sex, who want to have ONS, who want FWB, who enjoy healthily having more than one sex partner. There are plenty of women out there who enjoy giving BJ, who enjoy receiving oral, who enjoy having her breasts played with, who enjoy anal. The problem is, not every woman does, with any of these things. Even if a girl wants a ONS, she might not enjoy giving BJs. She might just want to make out with you, she might just want to trade oral, she might only want vaginal sex-ect. There is an abundance of woman out there, but clumping them all together would be harmful to
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Bananas are simply good items to show scale.<mask>?<mask> bananas are something everyone knows about<mask> big it is. They are common. Their sizes are pretty engraved in our memories and it easily allows us to grasp the size of the object the banana is comparatively accompanying. [NEWLINE] [NEWLINE] [STARTQ] 1) a bannana is not a uniform size. There are large bananas and small bananas [ENDQ] [NEWLINE] This is willfully deceptive. The size of most things fluctuates to a degree,<mask> to say that there are large and small bananas focuses<mask>ley on outliers.<mask>'s more truthful than your assertion is that most bananas are about the same size. Unless reddit collectively decided upon a standardized object, the banana is unbeat. [NEWLINE] [NEWLINE] [STARTQ] 2) The curve. Bananas are not usually straight. This makes judging their length difficult. It can<mask> mean that it is hard to show the banana very close to the object you are scaling against. [ENDQ] [NEWLINE] This does not render making comparative judgements difficult. I would say that the curve has little to no impact on such observations. You represent an argument that depicts redditors making very precise measurements of such random objects, making the need for a scaling object to be very exact<mask> that is simply not the case.<mask><mask> a banana has a curve doesn't mean you can't get closer to it. You need to explain that. [NEWLINE] [NEWLINE] [STARTQ] 3) Bananas are relatively small, this makes them bad for judging the scale of larger objects. This is especially bad<mask> considered in conjunction with point 1,<mask> any difference between the percieved size of a banana and the actual size will be enhanced for an object many times the size of the banana. [ENDQ] [NEWLINE] Yes bananas are small so it does make it difficult to decipher between one and a very large object.<mask> an object is to too large for a banana to be shown effectively, I would be willing to claim that the need for a scaling object would be void. I feel a straw-man fallacy in this point.<mask> I was showing a large building, I would not include any additional object that's purpose is to show scale. [NEWLINE] [NEWLINE] Furthermore,<mask> we follow with your argument's premise, the small deviation in banana size would be even more negligible<mask> compared to a very large object.<mask> this point wouldn't compliment point 1, your 3rd point is only negatable<mask><mask><mask> fact. For example<mask> I was showing a very large object say 300x bigger than your standard banana then include two bananas in the picture for scale; you would
Label encoding: <s>Bananas are simply good items to show scale. Why? Because bananas are something everyone knows about how big it is. They are common. Their sizes are pretty engraved in our memories and it easily allows us to grasp the size of the object the banana is comparatively accompanying. [NEWLINE] [NEWLINE] [STARTQ] 1) a bannana is not a uniform size. There are large bananas and small bananas [ENDQ] [NEWLINE] This is willfully deceptive. The size of most things fluctuates to a degree, but to say that there are large and small bananas focuses soley on outliers. What's more truthful than your assertion is that most bananas are about the same size. Unless reddit collectively decided upon a standardized object, the banana is unbeat. [NEWLINE] [NEWLINE] [STARTQ] 2) The curve. Bananas are not usually straight. This makes judging their length difficult. It can also mean that it is hard to show the banana very close to the object you are scaling against. [ENDQ] [NEWLINE] This does not render making comparative judgements difficult. I would say that the curve has little to no impact on such observations. You represent an argument that depicts redditors making very precise measurements of such random objects, making the need for a scaling object to be very exact when that is simply not the case. Also because a banana has a curve doesn't mean you can't get closer to it. You need to explain that. [NEWLINE] [NEWLINE] [STARTQ] 3) Bananas are relatively small, this makes them bad for judging the scale of larger objects. This is especially bad when considered in conjunction with point 1, as any difference between the percieved size of a banana and the actual size will be enhanced for an object many times the size of the banana. [ENDQ] [NEWLINE] Yes bananas are small so it does make it difficult to decipher between one and a very large object. If an object is to too large for a banana to be shown effectively, I would be willing to claim that the need for a scaling object would be void. I feel a straw-man fallacy in this point. If I was showing a large building, I would not include any additional object that's purpose is to show scale. [NEWLINE] [NEWLINE] Furthermore, if we follow with your argument's premise, the small deviation in banana size would be even more negligible when compared to a very large object. So this point wouldn't compliment point 1, your 3rd point is only negatable because of this fact. For example if I was showing a very large object say 300x bigger than your standard banana then include two bananas in the picture for scale; you would
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you look at the numbers, there is a higher percentage of men in STEM related fields than women. [ENDQ] [NEWLINE] This is correct. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> we're saying about men and women being equal<mask> it comes to the brain is true, then there should be no disparity. [ENDQ] [NEWLINE] <mask><mask> this may be a fundamental area of disagreement. <mask> women and men are generally equally capable, the effects of different hormones on our minds and bodies result in men and women generally having different interests and abilities.  Obviously, this wouldn't apply to each individual,<mask> rather<mask> a group it does. [NEWLINE] [NEWLINE] [STARTQ] There's simply a societal pressure to not pursue some of these fields. [ENDQ] [NEWLINE] I'm not sure it's<mask> much of a societal pressure<mask> much<mask> an individual preference<mask>.  We're not talking about something that women have never been a part of before such<mask> a hypothetical first female NFL quarterback,<mask> instead an industry<mask> women have been involved.  Everyone in technology work knows about [Grace Hopper]( [URL] ) who is one of the most important figures,<mask> not the most important, in programming. <mask> you want stuff that younger girls would know,<mask> many grew up with geeky girls like Velma from Scooby Doo in the 80's and 90's, or Penny from Inspector Gadget, and now<mask> an adult Abby on NCIS?  The list can go on and on.  I really don't think there is a societal stigma against geeky women at this point in time.  There may be individual or pocket examples of it,<mask> I doubt it is the majority. [NEWLINE] [NEWLINE] [STARTQ] This article talks about<mask> under 20% of Computer Science degrees are going to women. [ENDQ] [NEWLINE] <mask> the same article states that women are getting 60% of degrees,<mask> we don't see anyone other than the MRA crowd whining about<mask> men aren't getting properly educated anymore. <mask> women are getting more degrees, that clearly means they have the option to get a CS degree<mask> they want to<mask> they are going to college in full force. [NEWLINE] [NEWLINE] [STARTQ] One of the keys to fixing that is getting those numbers up artificially (through grants, job spots, etc.)<mask> much<mask> possible,<mask> young, impressionable children can have an idol in that certain field. [ENDQ] [NEWLINE] For the reasons I stated up a bit, we've already got this. [NEWLINE] [NEWLINE] [STARTQ] This strategy will bring us overall better scientists, computer scientists, engineers, psysicists, etc. over time which
Label encoding: <s> [STARTQ] When you look at the numbers, there is a higher percentage of men in STEM related fields than women. [ENDQ] [NEWLINE] This is correct. [NEWLINE] [NEWLINE] [STARTQ] If what we're saying about men and women being equal when it comes to the brain is true, then there should be no disparity. [ENDQ] [NEWLINE] I think this may be a fundamental area of disagreement.  While women and men are generally equally capable, the effects of different hormones on our minds and bodies result in men and women generally having different interests and abilities.  Obviously, this wouldn't apply to each individual, but rather as a group it does. [NEWLINE] [NEWLINE] [STARTQ] There's simply a societal pressure to not pursue some of these fields. [ENDQ] [NEWLINE] I'm not sure it's as much of a societal pressure as much as an individual preference though.  We're not talking about something that women have never been a part of before such as a hypothetical first female NFL quarterback, but instead an industry where women have been involved.  Everyone in technology work knows about [Grace Hopper]( [URL] ) who is one of the most important figures, if not the most important, in programming.  If you want stuff that younger girls would know, how many grew up with geeky girls like Velma from Scooby Doo in the 80's and 90's, or Penny from Inspector Gadget, and now as an adult Abby on NCIS?  The list can go on and on.  I really don't think there is a societal stigma against geeky women at this point in time.  There may be individual or pocket examples of it, but I doubt it is the majority. [NEWLINE] [NEWLINE] [STARTQ] This article talks about how under 20% of Computer Science degrees are going to women. [ENDQ] [NEWLINE] Yet the same article states that women are getting 60% of degrees, yet we don't see anyone other than the MRA crowd whining about how men aren't getting properly educated anymore.  If women are getting more degrees, that clearly means they have the option to get a CS degree if they want to since they are going to college in full force. [NEWLINE] [NEWLINE] [STARTQ] One of the keys to fixing that is getting those numbers up artificially (through grants, job spots, etc.) as much as possible, so young, impressionable children can have an idol in that certain field. [ENDQ] [NEWLINE] For the reasons I stated up a bit, we've already got this. [NEWLINE] [NEWLINE] [STARTQ] This strategy will bring us overall better scientists, computer scientists, engineers, psysicists, etc. over time which
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Not to be mean,<mask> I don't think you have any idea<mask> you're talking about. [ENDQ] [NEWLINE] Well aren't you condescending. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask><mask>, you seem to talk about capitalism or a capitalist system<mask><mask> its something that we<mask> a society have officially implemented. Like at some instant, we weren't capitalist. Then we all decided to start capitalism. We say we have capitalism or a capitalist system<mask> generally we enforce private property rights, more specifically privately owned means of production. Owners generally invest in capital goods to grow to increase productivity and there's a lot of trading going on. [ENDQ] [NEWLINE] I am aware of all of that,<mask> should have been apparent had you read my comments closely rather than leaping to conclusions. [NEWLINE] [NEWLINE] [STARTQ] The thing is<mask>, people do all that stuff, invest and trade, to better themselves and their condition. That includes selling labor. [ENDQ] [NEWLINE] I am aware of that too.<mask><mask> I explicitly stated that in my comments. [NEWLINE] [NEWLINE] [STARTQ] Machines are all about increasing productivity. [ENDQ] [NEWLINE] Currently.<mask> of course in the short term productivity gains can be at the expense of labor. In the long term "creative destruction" should in theory replace jobs that are lost.<mask> the central premise of OP is that there is some hypothetical point at which *all* human labor can be done by some machine equivalent. I.e., there are no jobs left in existence that could be done by a human that can't<mask> be done by a machine, robot, AI or<mask> have you. In that case the question is one of the comparative costs of labor. That is, can humans compete with machines at the same task in a labor market. In purely hypothetical terms we can imagine a scenario<mask> labor demands are sufficiently met by AI that the opportunity cost of doing other work is<mask> low that the minimum input of food is greater than the minimum input for a machine in the form of raw materials and energy, robbing humans of even the barest Comparative Advantage. At that point, in a capitalist society, capital would be more efficiently allocated to the procurement of machines than it would be to the procurement of human labor. In this scenario, human labor would not be used unless additional robots could simply not be obtained. This is the premise of OP, and I am discussing this situation treating it<mask><mask> that premise comes true,<mask> that was the hypothetical under discussion. [NEWLINE] [NEWLINE] [STARTQ] Just like machines make us more productive at a job, they<mask> make us more productive at home.
Label encoding: <s> [STARTQ] Not to be mean, but I don't think you have any idea what you're talking about. [ENDQ] [NEWLINE] Well aren't you condescending. [NEWLINE] [NEWLINE] [STARTQ] First of all, you seem to talk about capitalism or a capitalist system as if its something that we as a society have officially implemented. Like at some instant, we weren't capitalist. Then we all decided to start capitalism. We say we have capitalism or a capitalist system because generally we enforce private property rights, more specifically privately owned means of production. Owners generally invest in capital goods to grow to increase productivity and there's a lot of trading going on. [ENDQ] [NEWLINE] I am aware of all of that, as should have been apparent had you read my comments closely rather than leaping to conclusions. [NEWLINE] [NEWLINE] [STARTQ] The thing is though, people do all that stuff, invest and trade, to better themselves and their condition. That includes selling labor. [ENDQ] [NEWLINE] I am aware of that too. In fact I explicitly stated that in my comments. [NEWLINE] [NEWLINE] [STARTQ] Machines are all about increasing productivity. [ENDQ] [NEWLINE] Currently. But of course in the short term productivity gains can be at the expense of labor. In the long term "creative destruction" should in theory replace jobs that are lost. However the central premise of OP is that there is some hypothetical point at which *all* human labor can be done by some machine equivalent. I.e., there are no jobs left in existence that could be done by a human that can't also be done by a machine, robot, AI or what have you. In that case the question is one of the comparative costs of labor. That is, can humans compete with machines at the same task in a labor market. In purely hypothetical terms we can imagine a scenario where labor demands are sufficiently met by AI that the opportunity cost of doing other work is so low that the minimum input of food is greater than the minimum input for a machine in the form of raw materials and energy, robbing humans of even the barest Comparative Advantage. At that point, in a capitalist society, capital would be more efficiently allocated to the procurement of machines than it would be to the procurement of human labor. In this scenario, human labor would not be used unless additional robots could simply not be obtained. This is the premise of OP, and I am discussing this situation treating it as if that premise comes true, since that was the hypothetical under discussion. [NEWLINE] [NEWLINE] [STARTQ] Just like machines make us more productive at a job, they also make us more productive at home.
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>No problem. And obviously students who do a lot of different things<mask> deserve recognition,<mask><mask><mask> they are generally in a different category than student athletes or serious devotees to a single activity, for 3 reasons. [NEWLINE] [NEWLINE] 1) Doing a lot of activities at a lower level of commitment is unlikely to add up to the hours or strain of a really rigorous athletic program, half-time job, etc. I was lucky enough to go to multiple high schools, and at one of them we had 90 minute morning practice, two hour afternoon practice (<mask> I'd be at school 6-6, basically), plus a several hour Saturday practice, and<mask> we had competitions it was all-day,<mask> I was comfortably at 21-25 hours just from sports. This was a team that had a *modest* shot at state. Around that I volunteered an hour a week and did advanced classes (I would not say I am representative), and I was stretched pretty fucking thin. [NEWLINE] [NEWLINE] At another school I did the same sport,<mask> varsity level, which only had 90 minute after school practices, joined a bunch of clubs, played an instrument and altogether never put more than 15 hours a week outside of school into those activities. It was a lot of things and I felt very involved,<mask> it honestly didn't match the stress of the first school and<mask> there was a conflict or I was behind in work I could skip a club meeting with no consequences. *I had much more ability to scale back my activities<mask> they hurt my grades, by dropping one club or activity for example.* In a serious athletic program it is all or nothing, coaches don't want a player who only comes to practice four times a week. Someone who does several activities and has poor grades maybe demonstrates poor prioritizing, whereas someone who loves X<mask> has to spend at least 25 hours a week to do it is more understandable. Club meetings were often during school anyway, and highschool clubs are generally really easy. The athletic programs that are rigorous enough to actually prepare you to compete at a D1 level really do almost always take up more time and effort than participating in a lot of clubs and less rigorous varsity programs. [NEWLINE] [NEWLINE] Doing the math, 5x1.5 hour practices, 3x1 hour club meetings (usually during school or practice), and 4 hours of volunteering or other club activity *every week* is still &lt;15 hours, much of which is socializing, and I'd be surprised<mask> most people were<mask> involved<mask>
Label encoding: <s>No problem. And obviously students who do a lot of different things also deserve recognition, but I think they are generally in a different category than student athletes or serious devotees to a single activity, for 3 reasons. [NEWLINE] [NEWLINE] 1) Doing a lot of activities at a lower level of commitment is unlikely to add up to the hours or strain of a really rigorous athletic program, half-time job, etc. I was lucky enough to go to multiple high schools, and at one of them we had 90 minute morning practice, two hour afternoon practice ( so I'd be at school 6-6, basically), plus a several hour Saturday practice, and when we had competitions it was all-day, so I was comfortably at 21-25 hours just from sports. This was a team that had a *modest* shot at state. Around that I volunteered an hour a week and did advanced classes (I would not say I am representative), and I was stretched pretty fucking thin. [NEWLINE] [NEWLINE] At another school I did the same sport, also varsity level, which only had 90 minute after school practices, joined a bunch of clubs, played an instrument and altogether never put more than 15 hours a week outside of school into those activities. It was a lot of things and I felt very involved, but it honestly didn't match the stress of the first school and if there was a conflict or I was behind in work I could skip a club meeting with no consequences. *I had much more ability to scale back my activities if they hurt my grades, by dropping one club or activity for example.* In a serious athletic program it is all or nothing, coaches don't want a player who only comes to practice four times a week. Someone who does several activities and has poor grades maybe demonstrates poor prioritizing, whereas someone who loves X but has to spend at least 25 hours a week to do it is more understandable. Club meetings were often during school anyway, and highschool clubs are generally really easy. The athletic programs that are rigorous enough to actually prepare you to compete at a D1 level really do almost always take up more time and effort than participating in a lot of clubs and less rigorous varsity programs. [NEWLINE] [NEWLINE] Doing the math, 5x1.5 hour practices, 3x1 hour club meetings (usually during school or practice), and 4 hours of volunteering or other club activity *every week* is still &lt;15 hours, much of which is socializing, and I'd be surprised if most people were as involved as
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Interestingly enough, they did exactly this in the U.K. about thirty or<mask> years ago.  It all seems logical and practical and like it should work.   Then they decided that legally, ownership was satisfied<mask> you could own something at all. <mask> handguns were turned in in huge numbers and you were allowed to keep rifles and shotguns.  Then it was shotguns only.  Then you could only keep them<mask> you had a proven use for them like on a farm or similar. [NEWLINE] [NEWLINE] Thirty years later and the inventers of most of our modern arms and the people who helped us the most in WWII are completely castrated.  Firearms are essentially gone and crime is out of control. [NEWLINE] [NEWLINE] <mask><mask> was the critical difference? <mask> was pointed out, it's that your argument pre-supposes that firearms are a priviledge and not a right.  You go into the creation of the list with a view that logically will lead you down this path. <mask> let's get into each statement/idea. <mask> some are good, and some are not.  And this isn't really abot<mask> you set up the argument<mask> much<mask> the individual points. [NEWLINE] [NEWLINE] 1 - This sounds great. <mask> this is alrady<mask> happens with concealed carry permits.  A smarter approach would be to make a nationwide standard and database for CCW permits (it's a hodge-podge of conflicting laws currently and some states accept others, some do not), <mask>, make a concerted effort to promote such permit ownership to our youths (say, training leading up to a full permit at adulthood) and<mask> part of self-defense programs and<mask> on.  We need more firearms in the hands of good, trained citizens. [NEWLINE] [NEWLINE] 2 - Absolutely.  We do this in California already to buy any firearm. <mask> the issue again is every state has its own rules and laws.  There needs to be one database and one standard for all states.  I know I'm a big proponent of States rights,<mask> in many ways, the patchwork of laws creates more headache that it's worth<mask> we're trying to tackle a natiowide problem. [NEWLINE] [NEWLINE] 3 - Unfortunately this is completely unworkable. <mask><mask> you can reload your own ammo for pennies a round, and hundreds of millions of rounds are already in private hands, it's never going to happen. [NEWLINE] [NEWLINE] 4 - Correct.  People will just load their own or drive across town
Label encoding: <s>Interestingly enough, they did exactly this in the U.K. about thirty or so years ago.  It all seems logical and practical and like it should work.   Then they decided that legally, ownership was satisfied if you could own something at all.  So handguns were turned in in huge numbers and you were allowed to keep rifles and shotguns.  Then it was shotguns only.  Then you could only keep them if you had a proven use for them like on a farm or similar. [NEWLINE] [NEWLINE] Thirty years later and the inventers of most of our modern arms and the people who helped us the most in WWII are completely castrated.  Firearms are essentially gone and crime is out of control. [NEWLINE] [NEWLINE] So what was the critical difference?  As was pointed out, it's that your argument pre-supposes that firearms are a priviledge and not a right.  You go into the creation of the list with a view that logically will lead you down this path.  But let's get into each statement/idea.  Because some are good, and some are not.  And this isn't really abot how you set up the argument so much as the individual points. [NEWLINE] [NEWLINE] 1 - This sounds great.  But this is alrady what happens with concealed carry permits.  A smarter approach would be to make a nationwide standard and database for CCW permits (it's a hodge-podge of conflicting laws currently and some states accept others, some do not),  Also, make a concerted effort to promote such permit ownership to our youths (say, training leading up to a full permit at adulthood) and as part of self-defense programs and so on.  We need more firearms in the hands of good, trained citizens. [NEWLINE] [NEWLINE] 2 - Absolutely.  We do this in California already to buy any firearm.  But the issue again is every state has its own rules and laws.  There needs to be one database and one standard for all states.  I know I'm a big proponent of States rights, but in many ways, the patchwork of laws creates more headache that it's worth when we're trying to tackle a natiowide problem. [NEWLINE] [NEWLINE] 3 - Unfortunately this is completely unworkable.  Given that you can reload your own ammo for pennies a round, and hundreds of millions of rounds are already in private hands, it's never going to happen. [NEWLINE] [NEWLINE] 4 - Correct.  People will just load their own or drive across town
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0535, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0468, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, this is certainly a problem and I'm not actually sure<mask> to proceed other than to explain some basis of my arche, which I haven't thought about for a<mask>. That admittedly should've been my first step in challenging my intellectual honesty before creating this thread. [NEWLINE] [NEWLINE] I believe this was my thought process<mask> I first adopted this system of belief: A common moral system is that "the needs of the many outweigh the needs of the few."<mask><mask>, this is almost universally accepted in just about any "valid" system of morality, and with the possible exception of egoism, all of these systems of morality are considered flat-out evil.<mask>, I move that it is worthwhile to base your moral code upon this one arche of the masses being more important than those who control them. In essence, all of these moralities have at their base the goal of maximizing good, or making an attempt at doing<mask>. To me, adopting utilitarianism isn't,<mask>, a simple, "I'm taking<mask> we all agree on,"<mask> rather, "<mask><mask> happiness is the imperative, means are irrelevant compared to ends." [NEWLINE] [NEWLINE] Most moral philosophies concern themselves with<mask> means should be taken to reach the moral imperative.<mask>, a Utilitarian simply cuts out the middle man, and says that all that matters is that the ends justify the means. [NEWLINE] [NEWLINE] A perfect example of this is murder. Pacifism dictates that life is sacred (<mask> the only source of happiness), and<mask> that ending another's life is always an evil act. A Utilitarian disagrees, saying that even<mask> one accepts the point that life is all that matters<mask> it is all that generates utility, ending one life to save two is always the moral action to take. [NEWLINE] [NEWLINE] <mask>, utilitarianism is simply the state of being simultaneously a hedonist and a consequentialist,<mask> I'll quickly argue them both separately. [NEWLINE] [NEWLINE] Hedonism is essentially an argument against materialism. There is nothing that can be considered important,<mask> happiness in and of itself. ignoring consequentialism for a moment, and just focusing on the ends<mask><mask> whether or not means are an issue, one must accept that any objective measure of wealth is simply a way to gain more happiness. Any sort of concept that a moral system can be based on, liberty, family, personal gain, equality, all boil down to creating happiness,<mask> there's no reason to argue it by proxy, and<mask> only the happiness itself should be considered. [NEWLINE] [NEWLINE] <mask> for consequential
Label encoding: <s>Yes, this is certainly a problem and I'm not actually sure how to proceed other than to explain some basis of my arche, which I haven't thought about for a while. That admittedly should've been my first step in challenging my intellectual honesty before creating this thread. [NEWLINE] [NEWLINE] I believe this was my thought process when I first adopted this system of belief: A common moral system is that "the needs of the many outweigh the needs of the few." In fact, this is almost universally accepted in just about any "valid" system of morality, and with the possible exception of egoism, all of these systems of morality are considered flat-out evil. Therefore, I move that it is worthwhile to base your moral code upon this one arche of the masses being more important than those who control them. In essence, all of these moralities have at their base the goal of maximizing good, or making an attempt at doing so. To me, adopting utilitarianism isn't, however, a simple, "I'm taking what we all agree on," but rather, " given that happiness is the imperative, means are irrelevant compared to ends." [NEWLINE] [NEWLINE] Most moral philosophies concern themselves with what means should be taken to reach the moral imperative. However, a Utilitarian simply cuts out the middle man, and says that all that matters is that the ends justify the means. [NEWLINE] [NEWLINE] A perfect example of this is murder. Pacifism dictates that life is sacred ( as the only source of happiness), and therefore that ending another's life is always an evil act. A Utilitarian disagrees, saying that even if one accepts the point that life is all that matters as it is all that generates utility, ending one life to save two is always the moral action to take. [NEWLINE] [NEWLINE] Therefore, utilitarianism is simply the state of being simultaneously a hedonist and a consequentialist, so I'll quickly argue them both separately. [NEWLINE] [NEWLINE] Hedonism is essentially an argument against materialism. There is nothing that can be considered important, besides happiness in and of itself. ignoring consequentialism for a moment, and just focusing on the ends regardless of whether or not means are an issue, one must accept that any objective measure of wealth is simply a way to gain more happiness. Any sort of concept that a moral system can be based on, liberty, family, personal gain, equality, all boil down to creating happiness, so there's no reason to argue it by proxy, and therefore only the happiness itself should be considered. [NEWLINE] [NEWLINE] As for consequential
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No they wouldn't.<mask> they wanted to they would. [ENDQ] [NEWLINE] This post clearly shows a misunderstanding of international politics. Countries that want to and do above all else, threaten the security of their regional neighbors. [NEWLINE] [NEWLINE] International politics is rarely legislative (in the sense that it is at the domestic level). Once ideologies rise to the executive functions of a state, they take on a much different role. Security is on the minds of states, which continue to feel the threat of external pressure. It may not seem that way,<mask> it is. In places<mask> sovereignty is in question (and by *in question* I mean not objectively understood or confirmed), powers that take interest draw attention from the international community. [NEWLINE] [NEWLINE] Think for a second:<mask> do you feel about Russia's actions in the Crimea?<mask> do you consider Putin? Is he a rational actor? [NEWLINE] [NEWLINE] We can look all around the globe and find similar problems: Remember the Senkaku Islands? Consider for a minute that Japan doesn't recognize North Korea, and that Taiwan (The Republic of China) lays claim to the lands under the control of the People's Republic of China. *China Airlines* files you to Taiwan,<mask> *Air China* is your direct access to Beijing. [NEWLINE] [NEWLINE] Militaries in modern democracies operate pretty much independently of the legislative agenda. We can turn to France post Charlie Hebdo. Like an inflammatory response, the French military has come to the surface, exposing itself in public places to provide a blanket level of security. Likewise, in the aftermath of 9/11 the response in the U.S. wasn't platoons in public places,<mask> rather targeted engagements in Afghanistan and the later full blown invasion of Iraq. [NEWLINE] [NEWLINE] Understanding<mask> and<mask> the U.S.'s military predominance exists<mask> it does today requires a little bit of history. The global nature probably stems from international security agreements post WWII. Its technological power (and a the bulk of new expenditures), most likely emerged<mask> the amped up byproduct of cold war tensions. [NEWLINE] [NEWLINE] It costs a lot less to run things on a global scale than many people think. An ineffectively operated military will be poor<mask> it comes to managing and quelling conflict. We should assume<mask>, that the standing military is efficient enough to remain effective. Take this for example: F16 fighter jets can make it to Pyongyang, North Korea in 3 minutes from Osan air base. These jets are<mask> in range for Beijing, Shanghai, Tokyo, and Vladivostok
Label encoding: <s> [STARTQ] No they wouldn't. If they wanted to they would. [ENDQ] [NEWLINE] This post clearly shows a misunderstanding of international politics. Countries that want to and do above all else, threaten the security of their regional neighbors. [NEWLINE] [NEWLINE] International politics is rarely legislative (in the sense that it is at the domestic level). Once ideologies rise to the executive functions of a state, they take on a much different role. Security is on the minds of states, which continue to feel the threat of external pressure. It may not seem that way, but it is. In places where sovereignty is in question (and by *in question* I mean not objectively understood or confirmed), powers that take interest draw attention from the international community. [NEWLINE] [NEWLINE] Think for a second: How do you feel about Russia's actions in the Crimea? How do you consider Putin? Is he a rational actor? [NEWLINE] [NEWLINE] We can look all around the globe and find similar problems: Remember the Senkaku Islands? Consider for a minute that Japan doesn't recognize North Korea, and that Taiwan (The Republic of China) lays claim to the lands under the control of the People's Republic of China. *China Airlines* files you to Taiwan, while *Air China* is your direct access to Beijing. [NEWLINE] [NEWLINE] Militaries in modern democracies operate pretty much independently of the legislative agenda. We can turn to France post Charlie Hebdo. Like an inflammatory response, the French military has come to the surface, exposing itself in public places to provide a blanket level of security. Likewise, in the aftermath of 9/11 the response in the U.S. wasn't platoons in public places, but rather targeted engagements in Afghanistan and the later full blown invasion of Iraq. [NEWLINE] [NEWLINE] Understanding how and why the U.S.'s military predominance exists as it does today requires a little bit of history. The global nature probably stems from international security agreements post WWII. Its technological power (and a the bulk of new expenditures), most likely emerged as the amped up byproduct of cold war tensions. [NEWLINE] [NEWLINE] It costs a lot less to run things on a global scale than many people think. An ineffectively operated military will be poor when it comes to managing and quelling conflict. We should assume therefore, that the standing military is efficient enough to remain effective. Take this for example: F16 fighter jets can make it to Pyongyang, North Korea in 3 minutes from Osan air base. These jets are also in range for Beijing, Shanghai, Tokyo, and Vladivostok
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0485, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0465, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you probably shouldn't own a dog - except maybe you are blind, an old-school shepherd or train them for finding people after avalanches or earthquakes. I have several reasons to holdthis view - here they are, sorted by relevance (from least to most relevant). [NEWLINE] [NEWLINE] * They eat poo. [NEWLINE] * They are annoying. Many people have really strong opinions against smoking in public,<mask> it is upsetting them and<mask> laws have been past in many countries that limit/prohibit smoking in public transport, public buildings, etc.<mask>, it is not uncommon to be leg-humped by a dog in the bus, or step into dog-poo in the park, which I find really annoying. This seems to be regarded<mask> unproblematic, some pepole even get annoyed<mask> you do not want to touch their animal companion with questionable hygiene standards.* You may say that (contrary to smoking) these are merely inconveniences,<mask><mask><mask>... [NEWLINE] *...Some dogs are plain dangerous. Have a look at the ["Fatal dog attacks"]( [URL] ) wiki entry. Granted, there are MUCH more people dying from cars, cigaretts, cancer (and that's only deadly stuff with a C), etc.<mask><mask><mask><mask> even one person would be too much. There are a lot of young children on the list<mask> well. [NEWLINE] * My rant<mask> far may have given you the opinion that I just hate animals.<mask>, the opposite is true - I love them!<mask><mask><mask> we shouldn't keep them for our amusement or<mask> we fancy a walk once or twice a day. [NEWLINE] * They eat. In a world<mask> people are starving this is<mask><mask><mask> morally not justifiable. [NEWLINE] * Most importantly: They produce greenhouse gases - and quite a lot of them! [The co2-"paw print" of a big dog that gets fed mostly meat may even be bigger than the emissions caused by an SUV.]( [URL] /) [NEWLINE] [NEWLINE] <mask> I am convinced that you probably shouldn't own a dog. I will not be convinced by single examples<mask> a dog was useful,<mask> by reasons<mask> the overall benefits of private dog-owenership surpass the negetive effects I have listed above. Please CMV! [NEWLINE] [NEWLINE] (*)The fanatic dog-loving may be an issue particular to Germany. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] **EDIT: Deltas have been awarded to the argument that dogs "sage lives" by providing companionship, and helping with mental<mask> well
Label encoding: <s>I think you probably shouldn't own a dog - except maybe you are blind, an old-school shepherd or train them for finding people after avalanches or earthquakes. I have several reasons to holdthis view - here they are, sorted by relevance (from least to most relevant). [NEWLINE] [NEWLINE] * They eat poo. [NEWLINE] * They are annoying. Many people have really strong opinions against smoking in public, because it is upsetting them and accordingly laws have been past in many countries that limit/prohibit smoking in public transport, public buildings, etc. However, it is not uncommon to be leg-humped by a dog in the bus, or step into dog-poo in the park, which I find really annoying. This seems to be regarded as unproblematic, some pepole even get annoyed if you do not want to touch their animal companion with questionable hygiene standards.* You may say that (contrary to smoking) these are merely inconveniences, but in fact... [NEWLINE] *...Some dogs are plain dangerous. Have a look at the ["Fatal dog attacks"]( [URL] ) wiki entry. Granted, there are MUCH more people dying from cars, cigaretts, cancer (and that's only deadly stuff with a C), etc. but in my opinion even one person would be too much. There are a lot of young children on the list as well. [NEWLINE] * My rant so far may have given you the opinion that I just hate animals. However, the opposite is true - I love them! Therefore I think we shouldn't keep them for our amusement or because we fancy a walk once or twice a day. [NEWLINE] * They eat. In a world where people are starving this is in my opinion morally not justifiable. [NEWLINE] * Most importantly: They produce greenhouse gases - and quite a lot of them! [The co2-"paw print" of a big dog that gets fed mostly meat may even be bigger than the emissions caused by an SUV.]( [URL] /) [NEWLINE] [NEWLINE] Therefore I am convinced that you probably shouldn't own a dog. I will not be convinced by single examples where a dog was useful, but by reasons why the overall benefits of private dog-owenership surpass the negetive effects I have listed above. Please CMV! [NEWLINE] [NEWLINE] (*)The fanatic dog-loving may be an issue particular to Germany. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] **EDIT: Deltas have been awarded to the argument that dogs "sage lives" by providing companionship, and helping with mental as well
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Okay, will do. I don't feel qualified enough for heavy-duty programming internships<mask> (I don't really know<mask> I should feel qualified actually,)<mask> maybe I can make coffee runs. [ENDQ] [NEWLINE] Don't sell yourself short. Go out there and apply for all of it. Do not be one of those coffee getters.<mask> you really want to be a programmer, you need a programming internship. A coffee internship will get you nothing.<mask> matters is good experiences that you can talk about in an interview or write about in your resume. Recruiters don't expect interns to be perfect programmers. Many businesses are well aware that internships are a necessary part of education, and they want to do their part to help out a little bit, and maybe get little bit of work out of the intern too<mask> they're at it. [NEWLINE] [NEWLINE] <mask>, it's understandable<mask> you can't get an internship. Shit can be competitive these days.<mask> you don't get one, join one of those computer science clubs or engineering clubs or any sort of club that is making/building applications and other stuff. [NEWLINE] [NEWLINE] [NEWLINE] Don't waste a semester getting no experience. Hop in<mask> soon<mask> you can. Try to find those student projects and hop in. Honestly, these student projects could be even more important than your coursework. Taking some class will net you 2 words on your resume. Working on a student project will net you a nice little paragraph.<mask>, these little student projects are often done with the cooperation of helpful professors or TA's. These will be good connections to have. Don't want to build anything? Then just aim for a leadership position in some CS related organization. Can't find anything cool in CS? Look around in engineering. They always need programmers for their robots or airplanes. [NEWLINE] [NEWLINE] [NEWLINE] Finally,<mask> year of computer science are you? I didn't do CS (engineering),<mask> the freshman year was the worst. I had to take all these math professors that didn't give a fuck. Much of freshman year is structured to literally weed out the "undesirable" students. The public university I went to had 50+% failure rates in many of these classes. (Then again, these 50% failures never bothered to ask me for help during my office hours and neglected to show up at the TA session. You're paying for these resources. Use them!) I have the privilege of being both a student and a TA<mask> I have a little bit of insight in the process. Many of these
Label encoding: <s> [STARTQ] Okay, will do. I don't feel qualified enough for heavy-duty programming internships yet (I don't really know when I should feel qualified actually,) but maybe I can make coffee runs. [ENDQ] [NEWLINE] Don't sell yourself short. Go out there and apply for all of it. Do not be one of those coffee getters. If you really want to be a programmer, you need a programming internship. A coffee internship will get you nothing. What matters is good experiences that you can talk about in an interview or write about in your resume. Recruiters don't expect interns to be perfect programmers. Many businesses are well aware that internships are a necessary part of education, and they want to do their part to help out a little bit, and maybe get little bit of work out of the intern too while they're at it. [NEWLINE] [NEWLINE] Also, it's understandable if you can't get an internship. Shit can be competitive these days. If you don't get one, join one of those computer science clubs or engineering clubs or any sort of club that is making/building applications and other stuff. [NEWLINE] [NEWLINE] [NEWLINE] Don't waste a semester getting no experience. Hop in as soon as you can. Try to find those student projects and hop in. Honestly, these student projects could be even more important than your coursework. Taking some class will net you 2 words on your resume. Working on a student project will net you a nice little paragraph. Moreover, these little student projects are often done with the cooperation of helpful professors or TA's. These will be good connections to have. Don't want to build anything? Then just aim for a leadership position in some CS related organization. Can't find anything cool in CS? Look around in engineering. They always need programmers for their robots or airplanes. [NEWLINE] [NEWLINE] [NEWLINE] Finally, what year of computer science are you? I didn't do CS (engineering), but the freshman year was the worst. I had to take all these math professors that didn't give a fuck. Much of freshman year is structured to literally weed out the "undesirable" students. The public university I went to had 50+% failure rates in many of these classes. (Then again, these 50% failures never bothered to ask me for help during my office hours and neglected to show up at the TA session. You're paying for these resources. Use them!) I have the privilege of being both a student and a TA so I have a little bit of insight in the process. Many of these
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>What spending do you feel is unnecessary? [NEWLINE] [NEWLINE] Here is a pie chart of all discresionary spending (2009 data): [NEWLINE] [NEWLINE] [URL].jpg [NEWLINE] [NEWLINE] Here is all spending (2012): [NEWLINE] [NEWLINE] [URL].png [NEWLINE] [NEWLINE] <mask><mask> are you cutting? [NEWLINE] [NEWLINE] Welfare and entitlement is a popular place.<mask>, keep in mind that for every welfare dollar spent, the government collects somewhere over $0.60 in tax revenue. People who get welfare spend the money immediately in the economy, and that generates a lot of economic activity.<mask> it's very difficult to actually close a budget by cutting safety-net programs. [NEWLINE] [NEWLINE] Next up is Social Security and Medicare. These programs are funded through FICA taxes. They would both be fine<mask> for some reason we didn't exempt all income over $118,500 from FICA taxes. Even with no changes, Social Security could continue to pay out 75% of current benefits forever.<mask>, CMS and medicare does all the actual hard work in health care. They establish the billing codes, decides<mask> is reimbursed at<mask> rate, etc. All the private insurers basically say "we pay<mask> medicare pays" for everything. They do this<mask> running a 3% overhead. For the purpose of comparison, private health insurers run 20-25% overhead. Any shift of healthcare from CMS to the private industry would me a massive loss of efficiency. Again, you can<mask> things,<mask> that doesn't mean you actually help yourself out in the long run. [NEWLINE] [NEWLINE] You can rail against the US Dept of Education or the EPA,<mask> the reality is that completely eliminating these programs are 1. not going to close the budget deficit, and 2. incur substantial societal costs. [NEWLINE] [NEWLINE] Conservatives seem to see taxation and spending<mask> inherently bad the for economy. This is simply not accurate. Money gathered by the government is injected right back into the economy. Money "wasted" on bureaucrats is money that is paid,<mask> a salary, to consumers who then go out and use it to buy goods and services. This is<mask> European countries are able to sustain tax rates that are significantly higher than the US and still have strong economic growth. Taxes are not bad<mask><mask><mask> long<mask> they are spent back into the economy. Here is a graph of GPD per capita: [NEWLINE] [NEWLINE] [URL].png [NEWLINE] [NEWLINE] <mask> you can see, there is not a huge difference in economic growth rate between the US and Europe,<mask> the differences in taxation and spending policies. [NEWLINE] [NEWLINE] In regards to
Label encoding: <s>What spending do you feel is unnecessary? [NEWLINE] [NEWLINE] Here is a pie chart of all discresionary spending (2009 data): [NEWLINE] [NEWLINE] [URL].jpg [NEWLINE] [NEWLINE] Here is all spending (2012): [NEWLINE] [NEWLINE] [URL].png [NEWLINE] [NEWLINE] So where are you cutting? [NEWLINE] [NEWLINE] Welfare and entitlement is a popular place. However, keep in mind that for every welfare dollar spent, the government collects somewhere over $0.60 in tax revenue. People who get welfare spend the money immediately in the economy, and that generates a lot of economic activity. So it's very difficult to actually close a budget by cutting safety-net programs. [NEWLINE] [NEWLINE] Next up is Social Security and Medicare. These programs are funded through FICA taxes. They would both be fine if for some reason we didn't exempt all income over $118,500 from FICA taxes. Even with no changes, Social Security could continue to pay out 75% of current benefits forever. Meanwhile, CMS and medicare does all the actual hard work in health care. They establish the billing codes, decides what is reimbursed at what rate, etc. All the private insurers basically say "we pay what medicare pays" for everything. They do this while running a 3% overhead. For the purpose of comparison, private health insurers run 20-25% overhead. Any shift of healthcare from CMS to the private industry would me a massive loss of efficiency. Again, you can but things, but that doesn't mean you actually help yourself out in the long run. [NEWLINE] [NEWLINE] You can rail against the US Dept of Education or the EPA, but the reality is that completely eliminating these programs are 1. not going to close the budget deficit, and 2. incur substantial societal costs. [NEWLINE] [NEWLINE] Conservatives seem to see taxation and spending as inherently bad the for economy. This is simply not accurate. Money gathered by the government is injected right back into the economy. Money "wasted" on bureaucrats is money that is paid, as a salary, to consumers who then go out and use it to buy goods and services. This is why European countries are able to sustain tax rates that are significantly higher than the US and still have strong economic growth. Taxes are not bad as long as long as they are spent back into the economy. Here is a graph of GPD per capita: [NEWLINE] [NEWLINE] [URL].png [NEWLINE] [NEWLINE] As you can see, there is not a huge difference in economic growth rate between the US and Europe, despite the differences in taxation and spending policies. [NEWLINE] [NEWLINE] In regards to
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>L: "Peanuts shouldn't be mandatory to eat!" [NEWLINE] [NEWLINE] R: "Well... umm... ok then." [NEWLINE] [NEWLINE] L: "Wait, you agree?" [NEWLINE] [NEWLINE] R: "Yeah, of course! There are hazards with peanuts! Nobody's really going to say that 'Mandatory Peanut Consumption' would be a great idea; everybody knows that *people can freaking die from consuming peanuts*." [NEWLINE] [NEWLINE] L: "...I'm confused." [NEWLINE] [NEWLINE] R: "<mask>? Peanuts are a food. Some people are allergic to-" [NEWLINE] [NEWLINE] L: "No, it isn't with that. It's with the argument itself." [NEWLINE] [NEWLINE] [R waits for L to clarify. L sighs, slightly facepalms] [NEWLINE] [NEWLINE] L: "The idea that *something* should be mandatory for an individual to 'consume'. Like vaccines." [NEWLINE] [NEWLINE] R: "Like vaccines." [NEWLINE] [NEWLINE] L: "Like vaccines!" [NEWLINE] [NEWLINE] R: "Like vaccines!?" [NEWLINE] [NEWLINE] L: "Sure! Vaccines are exactly like peanuts. Sort of." [NEWLINE] [NEWLINE] R: "I really hope that vaccines aren't exactly like peanuts,<mask> I get<mask> you're saying." [NEWLINE] [NEWLINE] L: "Exactly. And... I know you support vaccines. Vaccines for everyone. [Narrows eyes] *Whether they like it or not*." [NEWLINE] [NEWLINE] R: "Well...not exactly." [NEWLINE] [NEWLINE] L: "<mask> you said-" [NEWLINE] [NEWLINE] R: "I phrase things badly. I don't really view vaccines in the same way<mask> peanuts; generally, vaccines are harmless.<mask><mask>, vaccines are typically *extremely* beneficial." [NEWLINE] [NEWLINE] L: "<mask> not in all cases." [NEWLINE] [NEWLINE] R: "No, not in all cases! I was never arguing that!" [NEWLINE] [NEWLINE] L: "<mask> -" [NEWLINE] [NEWLINE] R: "Look, I advocate for people to *always* use condoms<mask> they want to prevent children. That doesn't mean that I'm advocating for people with latex allergies to use a latex condom." [NEWLINE] [NEWLINE] L: "Prevent children from<mask>?" [NEWLINE] [NEWLINE] R: "<mask>?" [NEWLINE] [NEWLINE] L: "...Nevermind." [NEWLINE] [NEWLINE] [L is silent. R continues.] [NEWLINE] [NEWLINE] R: "Well,<mask> I say that 'everyone should be vaccinated', I'm saying that everyone that *can* benefit from vaccines *should* benefit from vaccines. Those people that think vaccines will give their children autism are preventing their children from reaping the benefits of a
Label encoding: <s>L: "Peanuts shouldn't be mandatory to eat!" [NEWLINE] [NEWLINE] R: "Well... umm... ok then." [NEWLINE] [NEWLINE] L: "Wait, you agree?" [NEWLINE] [NEWLINE] R: "Yeah, of course! There are hazards with peanuts! Nobody's really going to say that 'Mandatory Peanut Consumption' would be a great idea; everybody knows that *people can freaking die from consuming peanuts*." [NEWLINE] [NEWLINE] L: "...I'm confused." [NEWLINE] [NEWLINE] R: " How? Peanuts are a food. Some people are allergic to-" [NEWLINE] [NEWLINE] L: "No, it isn't with that. It's with the argument itself." [NEWLINE] [NEWLINE] [R waits for L to clarify. L sighs, slightly facepalms] [NEWLINE] [NEWLINE] L: "The idea that *something* should be mandatory for an individual to 'consume'. Like vaccines." [NEWLINE] [NEWLINE] R: "Like vaccines." [NEWLINE] [NEWLINE] L: "Like vaccines!" [NEWLINE] [NEWLINE] R: "Like vaccines!?" [NEWLINE] [NEWLINE] L: "Sure! Vaccines are exactly like peanuts. Sort of." [NEWLINE] [NEWLINE] R: "I really hope that vaccines aren't exactly like peanuts, but I get what you're saying." [NEWLINE] [NEWLINE] L: "Exactly. And... I know you support vaccines. Vaccines for everyone. [Narrows eyes] *Whether they like it or not*." [NEWLINE] [NEWLINE] R: "Well...not exactly." [NEWLINE] [NEWLINE] L: " But you said-" [NEWLINE] [NEWLINE] R: "I phrase things badly. I don't really view vaccines in the same way as peanuts; generally, vaccines are harmless. In fact, vaccines are typically *extremely* beneficial." [NEWLINE] [NEWLINE] L: " But not in all cases." [NEWLINE] [NEWLINE] R: "No, not in all cases! I was never arguing that!" [NEWLINE] [NEWLINE] L: " But -" [NEWLINE] [NEWLINE] R: "Look, I advocate for people to *always* use condoms when they want to prevent children. That doesn't mean that I'm advocating for people with latex allergies to use a latex condom." [NEWLINE] [NEWLINE] L: "Prevent children from what?" [NEWLINE] [NEWLINE] R: " What?" [NEWLINE] [NEWLINE] L: "...Nevermind." [NEWLINE] [NEWLINE] [L is silent. R continues.] [NEWLINE] [NEWLINE] R: "Well, when I say that 'everyone should be vaccinated', I'm saying that everyone that *can* benefit from vaccines *should* benefit from vaccines. Those people that think vaccines will give their children autism are preventing their children from reaping the benefits of a
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> both those statements have an equal value of being true... [ENDQ] [NEWLINE] This.  This is the fallacy you're looking for.  It's not at all true that two mutually exclusive propositions have an equal chance of being true. [NEWLINE] [NEWLINE] I'll give you an example.  Let's suppose you and I are having a conversation, and I tell you that there is a tribe of invisible elves living underneath my floorboards with whom I communicate telepathically. <mask> would it take to convince you I was telling the truth? <mask> you try to come up with any empirical means for discovering the elves' existence, I'd just say they are intangible and invisible and undetectable, in other words you have to take my word for it.  Is there an equal chance that I am either right and telling the truth or lying to you?  Of course not. [NEWLINE] [NEWLINE] **<mask> is most likely to be true is whatever has the most evidence for it.** [NEWLINE] [NEWLINE] [STARTQ] <mask> happens after that, is the church plays on anticipation.<mask> we die, we go to Heaven<mask> we are blissfully happy and reunited with our loved ones.<mask> we die, nothing happens. [ENDQ] [NEWLINE] This is a classic false dichotomy.  In other words, this is setting up the choice<mask><mask> it's only between two options<mask><mask><mask> there are many more options. [NEWLINE] [NEWLINE] <mask> you're basically saying is Christianity offers a choice between two options: [NEWLINE] [NEWLINE] 1. You believe in God, accept Christianity, and go to heaven after death. [NEWLINE] 2. You reject belief in God, deny Christianity, and go to hell after death. [NEWLINE] [NEWLINE] This is<mask> is known<mask> [Pascal's Wager]( [URL]'s_Wager), and it's a logical fallacy.  For one,<mask> about all the other religions,<mask> merely Christianity? <mask> about Islam? <mask> about Hinduism? <mask> do we know which one is right?  The choice isn't between Christianity and atheism, it's between Christianity, atheism, and about a thousand other religions. <mask> Christianity turns out to be wrong, and Islam is true, you're still going to Hell<mask><mask> you chose religion. [NEWLINE] [NEWLINE] [STARTQ] <mask>, is the church converting massive amounts of people weekly... [ENDQ] [NEWLINE] This is<mask> simply not true. <mask> is your source for this?  Both irreligion/atheism (lack of religion) and Islamic beliefs are growing faster than Christianity. <mask><mask>, in some countries, Christianity is
Label encoding: <s> [STARTQ] Even though both those statements have an equal value of being true... [ENDQ] [NEWLINE] This.  This is the fallacy you're looking for.  It's not at all true that two mutually exclusive propositions have an equal chance of being true. [NEWLINE] [NEWLINE] I'll give you an example.  Let's suppose you and I are having a conversation, and I tell you that there is a tribe of invisible elves living underneath my floorboards with whom I communicate telepathically.  What would it take to convince you I was telling the truth?  If you try to come up with any empirical means for discovering the elves' existence, I'd just say they are intangible and invisible and undetectable, in other words you have to take my word for it.  Is there an equal chance that I am either right and telling the truth or lying to you?  Of course not. [NEWLINE] [NEWLINE] ** What is most likely to be true is whatever has the most evidence for it.** [NEWLINE] [NEWLINE] [STARTQ] What happens after that, is the church plays on anticipation. When we die, we go to Heaven where we are blissfully happy and reunited with our loved ones. When we die, nothing happens. [ENDQ] [NEWLINE] This is a classic false dichotomy.  In other words, this is setting up the choice as if it's only between two options when in fact there are many more options. [NEWLINE] [NEWLINE] What you're basically saying is Christianity offers a choice between two options: [NEWLINE] [NEWLINE] 1. You believe in God, accept Christianity, and go to heaven after death. [NEWLINE] 2. You reject belief in God, deny Christianity, and go to hell after death. [NEWLINE] [NEWLINE] This is what is known as [Pascal's Wager]( [URL]'s_Wager), and it's a logical fallacy.  For one, what about all the other religions, why merely Christianity?  What about Islam?  What about Hinduism?  How do we know which one is right?  The choice isn't between Christianity and atheism, it's between Christianity, atheism, and about a thousand other religions.  If Christianity turns out to be wrong, and Islam is true, you're still going to Hell even though you chose religion. [NEWLINE] [NEWLINE] [STARTQ] So, is the church converting massive amounts of people weekly... [ENDQ] [NEWLINE] This is also simply not true.  Where is your source for this?  Both irreligion/atheism (lack of religion) and Islamic beliefs are growing faster than Christianity.  In fact, in some countries, Christianity is
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> china, a country which is known for oppression, censorship, lack of freedom, democracy, etc, is on the rise to become the next global superpower in terms of economics - and possibly influence. [ENDQ] [NEWLINE] <mask> does China erode the positive outlook on the world. [NEWLINE] [NEWLINE] It is just a different economic system (which really isn't all that different now<mask> you consider it).<mask><mask>, China is both liberalizing, and advancing. Don't get me wrong, they still have a long way to go,<mask> for a one party state, they have been on a steady (assuming economic actions),<mask> the Sino-Soviet split which happened in 1960. [NEWLINE] [NEWLINE] [STARTQ] Europe's economy is slowly crumbling and with it many its advancements to society (such<mask> its highly developed social security ). [ENDQ] [NEWLINE] Europe's problems stem from political, economic, geographic, and cultural divisions within the continent. To be completely honest, it really isn't its own separate continent,<mask> its geographic nomenclature is an artifact of classical antiquity. It is considered one politically,<mask> acceptance of the demarcation makes less sense. [NEWLINE] [NEWLINE] The problem with Europe may stem from it being is<mask> vast (not by today's standards), that political, cultural, and economic identities for the various regions developed beyond strict adherence to the present-day integration. Take the United States for example, and its counterpart, the pre-Constitution Articles of Confederation (which faced a different colonial problem: that it was too weak for warfare). [NEWLINE] [NEWLINE] European fragmentation, whether wary nationalists in Germany and Greece, or the strong Anti-Euro block in the U.K., have their own justifications for returning to self-rule. The U.S. economy, (and China's) for example, are just<mask> dependent on weak states<mask> they are on the strong ones.<mask> I can see<mask> the responsibility-for-the-other thing doesn't sit well with some in Europe. [NEWLINE] [NEWLINE] [STARTQ] Governments are spying on us more than ever. [ENDQ] [NEWLINE] <mask> you are from a first world country, the powers the government has against you have been steadily curtailed. Over the long trajectory time scale, surveillance may have certainly increased,<mask> the ability for your government to kill you (without questions),<mask> the international community sits by and respects sovereignty has certainly reduced considerably. [NEWLINE] [NEWLINE] [STARTQ] they have the ability to manipulate and oppress us on a scale never seen before [ENDQ] [NEWLINE] <mask> you are from a representative democracy, become informed (at the level of policy and
Label encoding: <s> [STARTQ] Meanwhile china, a country which is known for oppression, censorship, lack of freedom, democracy, etc, is on the rise to become the next global superpower in terms of economics - and possibly influence. [ENDQ] [NEWLINE] Why does China erode the positive outlook on the world. [NEWLINE] [NEWLINE] It is just a different economic system (which really isn't all that different now if you consider it). In fact, China is both liberalizing, and advancing. Don't get me wrong, they still have a long way to go, but for a one party state, they have been on a steady (assuming economic actions), since the Sino-Soviet split which happened in 1960. [NEWLINE] [NEWLINE] [STARTQ] Europe's economy is slowly crumbling and with it many its advancements to society (such as its highly developed social security ). [ENDQ] [NEWLINE] Europe's problems stem from political, economic, geographic, and cultural divisions within the continent. To be completely honest, it really isn't its own separate continent, as its geographic nomenclature is an artifact of classical antiquity. It is considered one politically, but acceptance of the demarcation makes less sense. [NEWLINE] [NEWLINE] The problem with Europe may stem from it being is so vast (not by today's standards), that political, cultural, and economic identities for the various regions developed beyond strict adherence to the present-day integration. Take the United States for example, and its counterpart, the pre-Constitution Articles of Confederation (which faced a different colonial problem: that it was too weak for warfare). [NEWLINE] [NEWLINE] European fragmentation, whether wary nationalists in Germany and Greece, or the strong Anti-Euro block in the U.K., have their own justifications for returning to self-rule. The U.S. economy, (and China's) for example, are just as dependent on weak states as they are on the strong ones. So I can see how the responsibility-for-the-other thing doesn't sit well with some in Europe. [NEWLINE] [NEWLINE] [STARTQ] Governments are spying on us more than ever. [ENDQ] [NEWLINE] If you are from a first world country, the powers the government has against you have been steadily curtailed. Over the long trajectory time scale, surveillance may have certainly increased, but the ability for your government to kill you (without questions), while the international community sits by and respects sovereignty has certainly reduced considerably. [NEWLINE] [NEWLINE] [STARTQ] they have the ability to manipulate and oppress us on a scale never seen before [ENDQ] [NEWLINE] If you are from a representative democracy, become informed (at the level of policy and
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"I truly hate them and you should too." [NEWLINE] [NEWLINE] See, that's<mask> you get weird. First off, you should never hate someone for their limitations. [NEWLINE] [NEWLINE] <mask>,<mask> originally trying to interpret the reason for your post, I was wondering<mask> you were perhaps autistic or were raised in a less affluent area of the world or something. Or,<mask> you were just one of those people that insists that everyone should interpret<mask> you say in exactly the same way<mask> you think you meant for it to be interpreted. Sadly, it appears the latter is correct. [NEWLINE] [NEWLINE] <mask> you wish to be a sane, socially successful, adult, then you always need to determine<mask> you wish to accomplish before you speak. [NEWLINE] [NEWLINE] Americans especially seem to have this fundamental issue of "needing to express themselves all the time." (I’m actually curious to know<mask> you “read” that sentence<mask> me stating, “You are an American.” A claim that clearly does not appear in the previous sentence,<mask> this is simply me stating an observation). [NEWLINE] [NEWLINE] <mask> you don't want people getting insulted, then make sure you pick your phrasing well. [NEWLINE] [NEWLINE] Being able to determine a goal and then using the best means necessary to accomplish that goal are two very important skills. Both of which you, people like you, and the very people you're lamenting all lack. [NEWLINE] [NEWLINE] <mask> a SJW asked me to change their view of<mask>, "Everyone should speak exactly the way I want them to," (which is exactly the same demand that you're making), then I'd give them the same advice. [NEWLINE] [NEWLINE] With the current state of society, there are different people in it. Different people do things differently. In order for society to have even the most basic chance of functioning, then almost everyone in that society has to obey at least a few sets or rules in order for it to work. [NEWLINE] [NEWLINE] <mask> one of your governing mantras is, "Freedom of Speech," then you can be neither offended by speech, nor demand that others speak or accept your expression in any particular way. You must be willing to deal with the “realness” of others,<mask> being adult enough to make sure that you present yourself less in a “knee jerk reaction” and more in a “deliberately well-thought out” manner. [NEWLINE] [NEWLINE] <mask>, again, the best way to be successful is to determine your goal and then to execute it with an awareness of
Label encoding: <s>"I truly hate them and you should too." [NEWLINE] [NEWLINE] See, that's where you get weird. First off, you should never hate someone for their limitations. [NEWLINE] [NEWLINE] Secondly, when originally trying to interpret the reason for your post, I was wondering if you were perhaps autistic or were raised in a less affluent area of the world or something. Or, if you were just one of those people that insists that everyone should interpret what you say in exactly the same way as you think you meant for it to be interpreted. Sadly, it appears the latter is correct. [NEWLINE] [NEWLINE] If you wish to be a sane, socially successful, adult, then you always need to determine what you wish to accomplish before you speak. [NEWLINE] [NEWLINE] Americans especially seem to have this fundamental issue of "needing to express themselves all the time." (I’m actually curious to know if you “read” that sentence as me stating, “You are an American.” A claim that clearly does not appear in the previous sentence, as this is simply me stating an observation). [NEWLINE] [NEWLINE] If you don't want people getting insulted, then make sure you pick your phrasing well. [NEWLINE] [NEWLINE] Being able to determine a goal and then using the best means necessary to accomplish that goal are two very important skills. Both of which you, people like you, and the very people you're lamenting all lack. [NEWLINE] [NEWLINE] If a SJW asked me to change their view of why, "Everyone should speak exactly the way I want them to," (which is exactly the same demand that you're making), then I'd give them the same advice. [NEWLINE] [NEWLINE] With the current state of society, there are different people in it. Different people do things differently. In order for society to have even the most basic chance of functioning, then almost everyone in that society has to obey at least a few sets or rules in order for it to work. [NEWLINE] [NEWLINE] If one of your governing mantras is, "Freedom of Speech," then you can be neither offended by speech, nor demand that others speak or accept your expression in any particular way. You must be willing to deal with the “realness” of others, while being adult enough to make sure that you present yourself less in a “knee jerk reaction” and more in a “deliberately well-thought out” manner. [NEWLINE] [NEWLINE] So, again, the best way to be successful is to determine your goal and then to execute it with an awareness of
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You say that this is "physically impossible"<mask> you didn't actually state any physics to support your argument. [NEWLINE] [NEWLINE] To start off, an electric motor is simply a coil of wire that rotates a ferrous metal utilizing a rotating magnetic field. It's already been established that straight electric motors and electric stepper motors are highly efficient<mask> of the direct conversion of electric power to kinetic force<mask> defined by [Lorentz' law]( [URL] #Force_on_a_current-carrying_wire)<mask> opposed to the conversion of gasoline<mask> a potential power source to kinetic force. You can<mask><mask> gasoline produces a bigger "bang"<mask> the chemical reaction is highly inefficient<mask> a large amount of energy is lost in heat and light.<mask>, the main loss of efficiency in an electric motor is that of internal resistance in the electric coil which results in heat (<mask> significantly less than exploding hydrocarbons). [NEWLINE] [NEWLINE] [STARTQ] In order to function, a hybrid car must have a series of very heavy batteries, a very heavy electric motor and a series of other heavy electric components (eg systems to recover lost energy during breaking). [ENDQ] [NEWLINE] You are once again lacking a source for any of this. The components for an electric motor are quite simple and much lighter than a combustion engine<mask> it does not have to withstand the physical explosion that hydrocarbons produce.<mask> stated above, it's a coil of wire and a ferrous metal bar to rotate. It can be scaled up,<mask> to drive a car at a low speed, there's no need for it to be very big. Think something the scale of a drill press or a lathe<mask> that's the kind of motor at work here. The mechanism to recover energy from the system is even smaller. An electric generator is literally an electric motor in reverse. In the case of a hybrid car, it uses the exact same hardware with a small circuit and a couple of transistors to drive the most efficient method of recharging the power supply. [NEWLINE] [NEWLINE] The last of the components, the power supply, is an interesting study. [Lithium-ion batteries]( [URL] ) development is becoming surprisingly sophisticated and I highly suggest looking into it. It's been shown that drawing and recharging Li-ion batteries at specific frequencies produces varying efficiency curves (and you bet you're ass the car companies aim for the best efficiency). The batteries are the same size the standard car batteries, making their weight a non-issue.<mask><mask>, [super-capacitor]( [URL] )
Label encoding: <s>You say that this is "physically impossible" but you didn't actually state any physics to support your argument. [NEWLINE] [NEWLINE] To start off, an electric motor is simply a coil of wire that rotates a ferrous metal utilizing a rotating magnetic field. It's already been established that straight electric motors and electric stepper motors are highly efficient because of the direct conversion of electric power to kinetic force as defined by [Lorentz' law]( [URL] #Force_on_a_current-carrying_wire) as opposed to the conversion of gasoline as a potential power source to kinetic force. You can argue that gasoline produces a bigger "bang" but the chemical reaction is highly inefficient since a large amount of energy is lost in heat and light. Meanwhile, the main loss of efficiency in an electric motor is that of internal resistance in the electric coil which results in heat ( though significantly less than exploding hydrocarbons). [NEWLINE] [NEWLINE] [STARTQ] In order to function, a hybrid car must have a series of very heavy batteries, a very heavy electric motor and a series of other heavy electric components (eg systems to recover lost energy during breaking). [ENDQ] [NEWLINE] You are once again lacking a source for any of this. The components for an electric motor are quite simple and much lighter than a combustion engine since it does not have to withstand the physical explosion that hydrocarbons produce. As stated above, it's a coil of wire and a ferrous metal bar to rotate. It can be scaled up, but to drive a car at a low speed, there's no need for it to be very big. Think something the scale of a drill press or a lathe because that's the kind of motor at work here. The mechanism to recover energy from the system is even smaller. An electric generator is literally an electric motor in reverse. In the case of a hybrid car, it uses the exact same hardware with a small circuit and a couple of transistors to drive the most efficient method of recharging the power supply. [NEWLINE] [NEWLINE] The last of the components, the power supply, is an interesting study. [Lithium-ion batteries]( [URL] ) development is becoming surprisingly sophisticated and I highly suggest looking into it. It's been shown that drawing and recharging Li-ion batteries at specific frequencies produces varying efficiency curves (and you bet you're ass the car companies aim for the best efficiency). The batteries are the same size the standard car batteries, making their weight a non-issue. In addition, [super-capacitor]( [URL] )
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> that the best way for you to change your mind is to reflect on<mask> you made up your mind about this in the first place. [NEWLINE] [NEWLINE] <mask> do you find foolish about each behavior and<mask> is it foolish? [NEWLINE] [NEWLINE] Rainbow colors - Many groups adopt symbols and colors<mask> a part of their group identity. It's not always easy to tell<mask> someone is gay, which can make dating and finding other gay friends kind of tricky. A rainbow bracelet at a bar or a bumper sticker on a car are signals to help us find each other. Now<mask> you're talking about someone wearing full rainbow gear outside of a pride event, I can see<mask> that would be foolish,<mask> only<mask> foolish<mask> someone wearing green head to toe<mask> it's not St. Patrick's day. [NEWLINE] [NEWLINE] Lisps - Technically it's not a lisp,<mask><mask> it's called is sort of beside the point. Like a rainbow bracelet, a lisp is another way to give someone a social cue about your identity.<mask><mask> part of the reason the 'lisp' is<mask> prevalent is<mask> it's associated with gay people in the media.<mask><mask> a lot of kids have grown up with lisping men<mask> an example of<mask> to be gay. Another reason people do it is that we all tend to naturally mimic the vocal styles of people we like. Ever heard of "vocal fry"? It's that guttural growling sound that many young women (college age and younger) use<mask> they speak. Vocal fry first showed up in music (think Brittany Spear's Hit Me Baby)<mask> vocal fry is now extremely common among young women. Being a bit older and not part of that social group, I find vocal fry really irritating<mask> I associate it with teenagers (who I often dislike in general). I wonder<mask> something similar could be true for you. Do you find it foolish<mask> gay people have a 'lisp', or do you find 'lisping' foolish<mask> gay people do it? I recognize that there's nothing inherently foolish or wrong with vocal fry. It just irritates me<mask> I associate it with teens. [NEWLINE] [NEWLINE] Leather - Okay, I'll admit it, I find it foolish<mask> **anyone** walks around in leather outside of the sort of kinky settings<mask> it belongs. I live in a **very** gay area and I have **never** seen a gay person walking around in leather fetish gear except at pride events.<mask> people wear wacky clothes to celebrations
Label encoding: <s>I think that the best way for you to change your mind is to reflect on how you made up your mind about this in the first place. [NEWLINE] [NEWLINE] What do you find foolish about each behavior and why is it foolish? [NEWLINE] [NEWLINE] Rainbow colors - Many groups adopt symbols and colors as a part of their group identity. It's not always easy to tell when someone is gay, which can make dating and finding other gay friends kind of tricky. A rainbow bracelet at a bar or a bumper sticker on a car are signals to help us find each other. Now if you're talking about someone wearing full rainbow gear outside of a pride event, I can see how that would be foolish, but only as foolish as someone wearing green head to toe when it's not St. Patrick's day. [NEWLINE] [NEWLINE] Lisps - Technically it's not a lisp, but what it's called is sort of beside the point. Like a rainbow bracelet, a lisp is another way to give someone a social cue about your identity. I think part of the reason the 'lisp' is so prevalent is because it's associated with gay people in the media. I think a lot of kids have grown up with lisping men as an example of how to be gay. Another reason people do it is that we all tend to naturally mimic the vocal styles of people we like. Ever heard of "vocal fry"? It's that guttural growling sound that many young women (college age and younger) use when they speak. Vocal fry first showed up in music (think Brittany Spear's Hit Me Baby) but vocal fry is now extremely common among young women. Being a bit older and not part of that social group, I find vocal fry really irritating because I associate it with teenagers (who I often dislike in general). I wonder if something similar could be true for you. Do you find it foolish when gay people have a 'lisp', or do you find 'lisping' foolish because gay people do it? I recognize that there's nothing inherently foolish or wrong with vocal fry. It just irritates me because I associate it with teens. [NEWLINE] [NEWLINE] Leather - Okay, I'll admit it, I find it foolish when **anyone** walks around in leather outside of the sort of kinky settings where it belongs. I live in a **very** gay area and I have **never** seen a gay person walking around in leather fetish gear except at pride events. But people wear wacky clothes to celebrations
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0497, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Isn't your entire post based on your personal experience?<mask> we take that away from your argument here,<mask> do you really have left? [NEWLINE] [NEWLINE] <mask> seriously. Experience tends to have a tempering effect on information that can often be difficult to quantify. Just<mask> you've arrived at your own point of view on experience (through your experience), events change the way we can view identical information. [NEWLINE] [NEWLINE] <mask> someone researches loss for instance, it may be easy to develop an academic view of the subject,<mask> that knowledge does not confer<mask> the actual experience of loss is like, or<mask> difficult it can be to deal with. In the same way that knowing<mask> to lose weight doesn't often translate into success with losing weight. [NEWLINE] [NEWLINE] The greatest benefit my own experience has conferred on me is a great awareness of<mask> it's like to be wrong. I've been wrong about a lot. It has gradually made me less and less certain over the years and<mask> I'm a lot less concerned than I once was about whether or not I'm right. It<mask> tends to make me more open to being wrong,<mask><mask> someone points out a flaw in my reasoning I'm quicker to give up my old point of view in favor of seeing things differently. [NEWLINE] [NEWLINE] A study conducted by Richard West at James Madison University found that contrary to<mask> one might expect, people who scored very high on SATs or other standardized methods of evaluating cognitive or collegiate ability had a surprising vulnerability to certain kinds of test questions, such<mask> : [NEWLINE] [NEWLINE] A bat and ball cost a dollar and ten cents. The bat costs a dollar more than the ball.<mask> much does the ball cost? [NEWLINE] [NEWLINE] <mask> our initial thought might be that the ball costs ten cents, this is actually incorrect. The correct answer is that the bat costs a dollar and five cents and that the ball costs five cents. [NEWLINE] [NEWLINE] Intelligent people, Richard West found, were more prone to relying on heuristics (or mental short cuts) to solve problems and that this made them MORE likely than the average person to be vulnerable to certain cognitive biases. [NEWLINE] [NEWLINE] In some instances, it is the very clarity with which we see that world that is our biggest barrier to seeing it accurately. This is<mask> experience is most helpful. It's not just having the book knowledge that is important. It's having the relevant life experience that brings that knowledge into its full three dimensions. [NEWLINE] [NEWLINE] <mask> your own arguments have developed, you've become frustrated with the ability of others, who often may not argue
Label encoding: <s>Isn't your entire post based on your personal experience? If we take that away from your argument here, what do you really have left? [NEWLINE] [NEWLINE] But seriously. Experience tends to have a tempering effect on information that can often be difficult to quantify. Just as you've arrived at your own point of view on experience (through your experience), events change the way we can view identical information. [NEWLINE] [NEWLINE] If someone researches loss for instance, it may be easy to develop an academic view of the subject, but that knowledge does not confer what the actual experience of loss is like, or how difficult it can be to deal with. In the same way that knowing how to lose weight doesn't often translate into success with losing weight. [NEWLINE] [NEWLINE] The greatest benefit my own experience has conferred on me is a great awareness of what it's like to be wrong. I've been wrong about a lot. It has gradually made me less and less certain over the years and so I'm a lot less concerned than I once was about whether or not I'm right. It also tends to make me more open to being wrong, so if someone points out a flaw in my reasoning I'm quicker to give up my old point of view in favor of seeing things differently. [NEWLINE] [NEWLINE] A study conducted by Richard West at James Madison University found that contrary to what one might expect, people who scored very high on SATs or other standardized methods of evaluating cognitive or collegiate ability had a surprising vulnerability to certain kinds of test questions, such as : [NEWLINE] [NEWLINE] A bat and ball cost a dollar and ten cents. The bat costs a dollar more than the ball. How much does the ball cost? [NEWLINE] [NEWLINE] While our initial thought might be that the ball costs ten cents, this is actually incorrect. The correct answer is that the bat costs a dollar and five cents and that the ball costs five cents. [NEWLINE] [NEWLINE] Intelligent people, Richard West found, were more prone to relying on heuristics (or mental short cuts) to solve problems and that this made them MORE likely than the average person to be vulnerable to certain cognitive biases. [NEWLINE] [NEWLINE] In some instances, it is the very clarity with which we see that world that is our biggest barrier to seeing it accurately. This is where experience is most helpful. It's not just having the book knowledge that is important. It's having the relevant life experience that brings that knowledge into its full three dimensions. [NEWLINE] [NEWLINE] As your own arguments have developed, you've become frustrated with the ability of others, who often may not argue
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Respectfully,<mask> someone who does a lot of work in behavioral economics,<mask><mask> this response is absolutely wrong<mask> it flies in the face of real world understanding of<mask> the brain works and<mask> people make decisions about their lives. [NEWLINE] [NEWLINE] **All** human behavior **is** a response to incentives and disincentives, either intrinsic or extrinsic.  Everything you do, from the food you eat to the spouse you marry, the religion you believe to the clothes you wear, is the result of decisions you make based upon your own, personal opportunity cost.  Even<mask> you are irrational, you're still doing it<mask> you are responding to your needs and wants *in the moment* rather than your needs and wants *long-term*.  The idea this reality is somehow exclusive to capitalism is a fallacy.  True, capitalism's primary feature is that, with proper regulation, it has historically led to the highest standard of living for the most people<mask> it harnesses this ugly aspect of human behavior in the marketplace leading to greater overall price efficiency and resource allocation,<mask> incentive/disincentives apply in non-capitalistic systems<mask> well. [NEWLINE] [NEWLINE] <mask> you see someone you want to avoid in the grocery store and turn around before you even have time to think about it, that's<mask> you are doing.  There was nothing capitalistic about that process, you just instantaneously calculated the personal incentives and disincentives, weighed by emotional happiness, time investment, personal comfort and appearance, social ramifications, etc. of running into someone in that moment. [NEWLINE] [NEWLINE] <mask> confronted with evidence the Earth was not flat, those who attacked Galileo were responding to incentives and disincentives.  Whether they realized it or not, they weighed the significant harm such a fact would cause to their authority and social capital upon being proven incorrect about a tenant of their teachings; some might have wanted to avoid the emotional pain of cognitive dissonance, i.e., "<mask> this is wrong,<mask> else might be wrong?"; some might have simply been petty and enjoyed punishing<mask> they viewed<mask> impertinence.  There was nothing capitalistic about it. [NEWLINE] [NEWLINE] Defining religion<mask> a product with a feature list may seem uncouth,<mask> in basic neurological terms, that its *exactly*<mask> it is.  On a primal, often subconscious level, people are always weighing the incentives and disincentives to make the decision they think works best for them, even<mask> they've never consciously thought about it or been able to articulate it
Label encoding: <s>Respectfully, as someone who does a lot of work in behavioral economics, I think this response is absolutely wrong because it flies in the face of real world understanding of how the brain works and how people make decisions about their lives. [NEWLINE] [NEWLINE] **All** human behavior **is** a response to incentives and disincentives, either intrinsic or extrinsic.  Everything you do, from the food you eat to the spouse you marry, the religion you believe to the clothes you wear, is the result of decisions you make based upon your own, personal opportunity cost.  Even when you are irrational, you're still doing it as you are responding to your needs and wants *in the moment* rather than your needs and wants *long-term*.  The idea this reality is somehow exclusive to capitalism is a fallacy.  True, capitalism's primary feature is that, with proper regulation, it has historically led to the highest standard of living for the most people because it harnesses this ugly aspect of human behavior in the marketplace leading to greater overall price efficiency and resource allocation, but incentive/disincentives apply in non-capitalistic systems as well. [NEWLINE] [NEWLINE] When you see someone you want to avoid in the grocery store and turn around before you even have time to think about it, that's what you are doing.  There was nothing capitalistic about that process, you just instantaneously calculated the personal incentives and disincentives, weighed by emotional happiness, time investment, personal comfort and appearance, social ramifications, etc. of running into someone in that moment. [NEWLINE] [NEWLINE] When confronted with evidence the Earth was not flat, those who attacked Galileo were responding to incentives and disincentives.  Whether they realized it or not, they weighed the significant harm such a fact would cause to their authority and social capital upon being proven incorrect about a tenant of their teachings; some might have wanted to avoid the emotional pain of cognitive dissonance, i.e., " if this is wrong, what else might be wrong?"; some might have simply been petty and enjoyed punishing what they viewed as impertinence.  There was nothing capitalistic about it. [NEWLINE] [NEWLINE] Defining religion as a product with a feature list may seem uncouth, but in basic neurological terms, that its *exactly* what it is.  On a primal, often subconscious level, people are always weighing the incentives and disincentives to make the decision they think works best for them, even if they've never consciously thought about it or been able to articulate it
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'd have to disagree that the world was better off<mask> the Soviet Union existed and acted<mask> a counterweight to the US in geopolitics. [NEWLINE] [NEWLINE] <mask> the Iraq War (est. 150,000 to 600,000 dead) and Afghanistan/War on Terror (est. 50,000 in Afghanistan + 5-20,000 worldwide) are the major controversial subjects of the post-Cold War and post-9/11 world... [NEWLINE] [NEWLINE] They all simply pale in comparison to the battle over ideology between the West and East that manifested itself in the form of proxy wars, regime changes, and other controversies from the period of 1945 to 1991. [NEWLINE] [NEWLINE] Here's a sampling: [NEWLINE] **Proxy Wars** [NEWLINE] [NEWLINE] * Greek Civil War (1946 - 1949) - Greece (with US + UK support) vs. Communist insurgents - 150,000 killed [NEWLINE] * Arab Israeli Conflict (main phase 1948 - 1973) - Israel (with US support) vs. Arab nations (with USSR support) - around 110,000 killed [NEWLINE] * Korean War (1950 - 1953) - North Korea (with USSR + China support) vs. South Korea + United States + UN support - 3 million+ killed [NEWLINE] * Cuban Revolution (1953 - 1959) - Communist insurgents vs. Batista government (with US support) - 5,000 killed [NEWLINE] * Vietnam War (1955 - 1975) - North Vietnam (supported primarily by USSR and China) vs. South Vietnam (supported primarily by United States) - 3-4 million killed [NEWLINE] * Nicaraguan Revolution and Contra War (1960s to 1990) - Somonza government (with US aid) vs. FSLN (supported by Soviet Union and Cuba) - 40,000 killed [NEWLINE] * Angolan Civil War (1975 - 2002) - MPLA aided by the USSR, Vietnam, and Cuba vs. UNITA and FNLA (aided by the US, China, and South Africa) - 500,000 killed [NEWLINE] * Soviet War in Afghanistan (1979 - 1989) - Soviet Union vs. Afghan rebels (with Pakistan + United States + China support) - 1 million killed [NEWLINE] [NEWLINE] Not wars,<mask> still prominent changes aided either covertly or overtly: [NEWLINE] [NEWLINE] **Regime Changes / Interventions** [NEWLINE] [NEWLINE] * 1948 - Czechoslovakia coup - USSR installs communist government [NEWLINE] * 1953 - Iran's Mossadegh, Prime Minister appointed by the Shah, overthrown after reducing Shah's power, nationalizing British oil, and turning towards USSR - Shah's power restored [NEWLINE] * 1954
Label encoding: <s>I'd have to disagree that the world was better off when the Soviet Union existed and acted as a counterweight to the US in geopolitics. [NEWLINE] [NEWLINE] While the Iraq War (est. 150,000 to 600,000 dead) and Afghanistan/War on Terror (est. 50,000 in Afghanistan + 5-20,000 worldwide) are the major controversial subjects of the post-Cold War and post-9/11 world... [NEWLINE] [NEWLINE] They all simply pale in comparison to the battle over ideology between the West and East that manifested itself in the form of proxy wars, regime changes, and other controversies from the period of 1945 to 1991. [NEWLINE] [NEWLINE] Here's a sampling: [NEWLINE] **Proxy Wars** [NEWLINE] [NEWLINE] * Greek Civil War (1946 - 1949) - Greece (with US + UK support) vs. Communist insurgents - 150,000 killed [NEWLINE] * Arab Israeli Conflict (main phase 1948 - 1973) - Israel (with US support) vs. Arab nations (with USSR support) - around 110,000 killed [NEWLINE] * Korean War (1950 - 1953) - North Korea (with USSR + China support) vs. South Korea + United States + UN support - 3 million+ killed [NEWLINE] * Cuban Revolution (1953 - 1959) - Communist insurgents vs. Batista government (with US support) - 5,000 killed [NEWLINE] * Vietnam War (1955 - 1975) - North Vietnam (supported primarily by USSR and China) vs. South Vietnam (supported primarily by United States) - 3-4 million killed [NEWLINE] * Nicaraguan Revolution and Contra War (1960s to 1990) - Somonza government (with US aid) vs. FSLN (supported by Soviet Union and Cuba) - 40,000 killed [NEWLINE] * Angolan Civil War (1975 - 2002) - MPLA aided by the USSR, Vietnam, and Cuba vs. UNITA and FNLA (aided by the US, China, and South Africa) - 500,000 killed [NEWLINE] * Soviet War in Afghanistan (1979 - 1989) - Soviet Union vs. Afghan rebels (with Pakistan + United States + China support) - 1 million killed [NEWLINE] [NEWLINE] Not wars, but still prominent changes aided either covertly or overtly: [NEWLINE] [NEWLINE] **Regime Changes / Interventions** [NEWLINE] [NEWLINE] * 1948 - Czechoslovakia coup - USSR installs communist government [NEWLINE] * 1953 - Iran's Mossadegh, Prime Minister appointed by the Shah, overthrown after reducing Shah's power, nationalizing British oil, and turning towards USSR - Shah's power restored [NEWLINE] * 1954
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0456, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Clinton was able to convence a huge number of the masses that in his case a BJ was not sex. Ask any kid<mask> a BJ given by their Mom to someone not their Dad is sex and you will probably get a different answer than<mask> Clinton propagandized us into accepting. [ENDQ] [NEWLINE] A BJ is not the same<mask> sex. Both are sexual acts,<mask> are not the same thing. You can complain that society has changing opinions about the morality of the act,<mask> that has very little to do with Clinton, and a lot to do with the fact that most of the previous reasons for limiting sexuality no longer apply (we have birth control now, for example). And last I checked, he was impeached,<mask> clearly he wasn't that successful at "propagandiz[ing] us into accepting". [NEWLINE] [NEWLINE] It was an issue that should have been handled by Bill and Hillary, not the American people. It has no bearing on his ability to head the executive branch,<mask> it is none of our business. [NEWLINE] [NEWLINE] I don't care that the president got a blowjob that might or might not cause an unhappy marriage or divorce. I care about whether the president starts an unjustified war that costs the lives of hundreds of thousands of people, many of whom are Americans. [NEWLINE] [NEWLINE] [STARTQ] she is attempting to nullify freedoms of the press, and speech. [ENDQ] [NEWLINE] Umm, no. She is expressing that she doesn't want them to use those words to describe the campaign. *They are still free to use them anyway.* This is in no way infringing on their rights to freedom of speech or press. I assume you wouldn't like being called an asshole.<mask> you request that I not call you one (which I'm not), it is in no way an infringement of my freedom of speech. [NEWLINE] [NEWLINE] [STARTQ] lets say that the vote was fraudulent due to control of the press, mind control through propaganda or Oprah/Whoopi, I would then hope the EC was pure enough to take the action of changing their votes. [ENDQ] [NEWLINE] We have already determined that there was no control of the press, at least not by the candidate. The rest of your sentence boils down to "<mask> someone I don't like happens to be charismatic and good at garnering support, the EC should invalidate the wishes of millions of other Americans." You have provided no criteria for determining<mask> this would be the case. You even went<mask> far<mask> to mention two charismatic individuals, saying that<mask> they exercise their *right to free
Label encoding: <s> [STARTQ] Clinton was able to convence a huge number of the masses that in his case a BJ was not sex. Ask any kid if a BJ given by their Mom to someone not their Dad is sex and you will probably get a different answer than what Clinton propagandized us into accepting. [ENDQ] [NEWLINE] A BJ is not the same as sex. Both are sexual acts, but are not the same thing. You can complain that society has changing opinions about the morality of the act, but that has very little to do with Clinton, and a lot to do with the fact that most of the previous reasons for limiting sexuality no longer apply (we have birth control now, for example). And last I checked, he was impeached, so clearly he wasn't that successful at "propagandiz[ing] us into accepting". [NEWLINE] [NEWLINE] It was an issue that should have been handled by Bill and Hillary, not the American people. It has no bearing on his ability to head the executive branch, so it is none of our business. [NEWLINE] [NEWLINE] I don't care that the president got a blowjob that might or might not cause an unhappy marriage or divorce. I care about whether the president starts an unjustified war that costs the lives of hundreds of thousands of people, many of whom are Americans. [NEWLINE] [NEWLINE] [STARTQ] she is attempting to nullify freedoms of the press, and speech. [ENDQ] [NEWLINE] Umm, no. She is expressing that she doesn't want them to use those words to describe the campaign. *They are still free to use them anyway.* This is in no way infringing on their rights to freedom of speech or press. I assume you wouldn't like being called an asshole. If you request that I not call you one (which I'm not), it is in no way an infringement of my freedom of speech. [NEWLINE] [NEWLINE] [STARTQ] lets say that the vote was fraudulent due to control of the press, mind control through propaganda or Oprah/Whoopi, I would then hope the EC was pure enough to take the action of changing their votes. [ENDQ] [NEWLINE] We have already determined that there was no control of the press, at least not by the candidate. The rest of your sentence boils down to " If someone I don't like happens to be charismatic and good at garnering support, the EC should invalidate the wishes of millions of other Americans." You have provided no criteria for determining when this would be the case. You even went so far as to mention two charismatic individuals, saying that if they exercise their *right to free
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> they can afford the 30 year debt<mask> can't afford the 1 year debt. It's not arbitrary<mask> there's a key difference. [ENDQ] [NEWLINE] <mask>,<mask> I mentioned, one can extend the terms of the 1 year debt to 30 years. The debt burdens are now the same. Opting for full forgiveness<mask> a less disruptive solution exists seems utterly arbitrary. [NEWLINE] [NEWLINE] [STARTQ] A radical government would presumably ram it through after getting into power. [ENDQ] [NEWLINE] From a feasibility perspective your proposal is entirely fanciful. It is far more likely that reformist policies survive evisceration by special interests than a radical government actually coming into power and enacting your policy. [NEWLINE] [NEWLINE] [STARTQ] I read the paper. It had a single cite from a study in Guatamala that I saw that addressed my actual query, the rest was unrelated stuff. [ENDQ] [NEWLINE] The rest explained<mask> they believed that lack of access to credit interferes with career mobility, which interferes with economic mobility, which interferes with social mobility. This is precisely<mask> you were getting at. [NEWLINE] [NEWLINE] [STARTQ] <mask> it did note the key thing that determines social mobility was education. In supporting social mobility I wouldn't look to debt, I'd look more to increasing education funding. [ENDQ] [NEWLINE] That's a separate issue.<mask> funding for education, medicine, small businesses, etc was all handled by the government without question then yes you would have far less need for credit. Of course that's not the case and<mask> policies that hurt people's ability to obtain credit *in isolation* are extremely detrimental to people's social mobility. [NEWLINE] [NEWLINE] [STARTQ] <mask> I've repeatedly stated, I am not proposing banning federal loans. Most student loans are federal loans. They'd have more rules of course. [ENDQ] [NEWLINE] Your proposal doesn't ban loans. It just makes them un-enforceable in a wider variety of scenarios. This natural reduces the market for credit making it more difficult for people to borrow. [NEWLINE] [NEWLINE] [STARTQ] For a quality business people would probably still make loans. [ENDQ] [NEWLINE] Again,<mask> you make a large number of loans uncollectable you will necessarily make it more difficult for *everyone* to get loans. Not only do lenders have to factor in bankruptcy into their calculation they<mask> have to factor in whatever threshold your set at<mask> the loan is magically scrubbed.<mask><mask> loans are<mask> terrible<mask> you say<mask> would you care<mask> the quality business people get them? [NEWLINE] [NEWLINE] [STARTQ] Most profits angel investors make is from the rare super successes anyway who can afford to pay
Label encoding: <s> [STARTQ] Because they can afford the 30 year debt but can't afford the 1 year debt. It's not arbitrary because there's a key difference. [ENDQ] [NEWLINE] But, as I mentioned, one can extend the terms of the 1 year debt to 30 years. The debt burdens are now the same. Opting for full forgiveness when a less disruptive solution exists seems utterly arbitrary. [NEWLINE] [NEWLINE] [STARTQ] A radical government would presumably ram it through after getting into power. [ENDQ] [NEWLINE] From a feasibility perspective your proposal is entirely fanciful. It is far more likely that reformist policies survive evisceration by special interests than a radical government actually coming into power and enacting your policy. [NEWLINE] [NEWLINE] [STARTQ] I read the paper. It had a single cite from a study in Guatamala that I saw that addressed my actual query, the rest was unrelated stuff. [ENDQ] [NEWLINE] The rest explained how they believed that lack of access to credit interferes with career mobility, which interferes with economic mobility, which interferes with social mobility. This is precisely what you were getting at. [NEWLINE] [NEWLINE] [STARTQ] Though it did note the key thing that determines social mobility was education. In supporting social mobility I wouldn't look to debt, I'd look more to increasing education funding. [ENDQ] [NEWLINE] That's a separate issue. If funding for education, medicine, small businesses, etc was all handled by the government without question then yes you would have far less need for credit. Of course that's not the case and therefore policies that hurt people's ability to obtain credit *in isolation* are extremely detrimental to people's social mobility. [NEWLINE] [NEWLINE] [STARTQ] As I've repeatedly stated, I am not proposing banning federal loans. Most student loans are federal loans. They'd have more rules of course. [ENDQ] [NEWLINE] Your proposal doesn't ban loans. It just makes them un-enforceable in a wider variety of scenarios. This natural reduces the market for credit making it more difficult for people to borrow. [NEWLINE] [NEWLINE] [STARTQ] For a quality business people would probably still make loans. [ENDQ] [NEWLINE] Again, if you make a large number of loans uncollectable you will necessarily make it more difficult for *everyone* to get loans. Not only do lenders have to factor in bankruptcy into their calculation they also have to factor in whatever threshold your set at where the loan is magically scrubbed. Also if loans are as terrible as you say why would you care if the quality business people get them? [NEWLINE] [NEWLINE] [STARTQ] Most profits angel investors make is from the rare super successes anyway who can afford to pay
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0401, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>People forget that pro-life is more than being against abortions. [Some] Pro-lifers believe capital punishment is murder, war is murder, and abortion is murder. [NEWLINE] [NEWLINE] <mask> murder is considered morally wrong, to question one would question another.<mask> we "allow" abortion we have to "allow" capital punishment (some will<mask><mask> capital punishment is not murder<mask> [murder constitutes an unlawful killing<mask> the punishment for a crime is lawfully deemed just]( [URL] +murder&amp;ie=utf-8&amp;oe=utf-8),<mask> a fetus is not guilty of any crimes). This is a question about morality and<mask> we are lenient on one issue and not the other we would be hypocritical. Essentially,<mask><mask> we define our morality should be sound and unwavering. I cannot change<mask> I define my morality on a case by case. I either believe and practice, or believe and am a hypocrite (i.e.<mask> it was morally wrong to wear red,<mask> crimson is acceptable I am a hypocrite).<mask><mask> there are worse things than death and I already pay taxes to house people in prison (I<mask> volunteered at prisons). [NEWLINE] [NEWLINE] <mask>, at<mask> point do we consider it alive? Beginning [Week 5]( [URL] ) the fetus begins to grow brain cells.<mask><mask> about taking a morning-after-pill? Before 5 Weeks is it the same<mask> picking a scab<mask> it has no brain cells? [At ~20 weeks a doctor has to use anesthesia<mask> operating on a baby (i.e. numb pain)]( [URL] /). The baby can survive outside the mother by 22 weeks. Hopefully before 22 weeks a parent can decide whether they are fit to keep the child.<mask> anything, by 22 weeks<mask> it can survive on life support and by the taxes of citizens it is no different than the prisoner I pay for to keep alive.<mask> this argument is not the point. Many people do adopt premature babies,<mask> even<mask> they don't and the child ends up in the system they offer the exact same support<mask> they would for a prisoner by paying taxes. I would love to adopt children some day,<mask> I am of sound mind (PTSD &amp; DID prevent this) and<mask> I have the financial responsibility,<mask> this cannot be imposed.<mask> I am forced into this situation I am in no better position to take care of the child than the parent who gave them up. Should I obligate others to volunteer at prisons or adopt a
Label encoding: <s>People forget that pro-life is more than being against abortions. [Some] Pro-lifers believe capital punishment is murder, war is murder, and abortion is murder. [NEWLINE] [NEWLINE] If murder is considered morally wrong, to question one would question another. If we "allow" abortion we have to "allow" capital punishment (some will argue that capital punishment is not murder because [murder constitutes an unlawful killing where the punishment for a crime is lawfully deemed just]( [URL] +murder&amp;ie=utf-8&amp;oe=utf-8), but a fetus is not guilty of any crimes). This is a question about morality and if we are lenient on one issue and not the other we would be hypocritical. Essentially, if how we define our morality should be sound and unwavering. I cannot change how I define my morality on a case by case. I either believe and practice, or believe and am a hypocrite (i.e. if it was morally wrong to wear red, but crimson is acceptable I am a hypocrite). I think there are worse things than death and I already pay taxes to house people in prison (I also volunteered at prisons). [NEWLINE] [NEWLINE] So, at what point do we consider it alive? Beginning [Week 5]( [URL] ) the fetus begins to grow brain cells. So what about taking a morning-after-pill? Before 5 Weeks is it the same as picking a scab since it has no brain cells? [At ~20 weeks a doctor has to use anesthesia when operating on a baby (i.e. numb pain)]( [URL] /). The baby can survive outside the mother by 22 weeks. Hopefully before 22 weeks a parent can decide whether they are fit to keep the child. If anything, by 22 weeks when it can survive on life support and by the taxes of citizens it is no different than the prisoner I pay for to keep alive. But this argument is not the point. Many people do adopt premature babies, but even if they don't and the child ends up in the system they offer the exact same support as they would for a prisoner by paying taxes. I would love to adopt children some day, when I am of sound mind (PTSD &amp; DID prevent this) and when I have the financial responsibility, but this cannot be imposed. If I am forced into this situation I am in no better position to take care of the child than the parent who gave them up. Should I obligate others to volunteer at prisons or adopt a
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You're right, this is a hornet's nest.  It's never a good idea to just take isolated quotes out of their context. [NEWLINE] [NEWLINE] [STARTQ] “I feel that ‘man-hating’ is an honourable and viable *political* act, that the oppressed have a right to class-hatred against the *class* that is oppressing them.” – Robin Morgan, Ms. Magazine Editor [ENDQ] [NEWLINE] This is from 1970.  I've added italics into the quote<mask> they should be.  On the surface, this appears to be a statement about hating men,<mask> taken in context<mask> it was written,<mask> she is really saying is she hates patriarchy and the men who uphold it.  That's the class she is talking about.  In the context of the times and the specific state of affairs she is talking about, she is writing in response to the extreme sexism of the times that was ongoing within the Leftist movement and the failure of the movement to live up to its revolutionary ideas.  She is writing with passion and exasperation.  You can't begin to understand the context of<mask> she is saying without understanding<mask> was going on<mask> it was written.  Her use of language seems extreme<mask> this is the style people used back then.  It is not<mask> it appears to be. [NEWLINE] [NEWLINE] [STARTQ] “To call a man an animal is to flatter him; he’s a machine, a walking dildo.” -– Valerie Solanas [ENDQ] [NEWLINE] Valerie Solanas said worse things than that.  She probably did hate men,<mask> she was a paranoid schizophrenic, and her ideas stemmed from the extreme sexism of the time (the 60s).  Look at wikipedia for more info. [NEWLINE] [NEWLINE] [STARTQ] “I want to see a man beaten to a bloody pulp with a high-heel shoved in his mouth, like an apple in the mouth of a pig.” — Andrea Dworkin [ENDQ] [NEWLINE] This is taken from fiction, it's something a character who was raped says. Please read this: [URL],<mask> it<mask> addresses some of the other quotes/people on your list. <mask>, this interview clarifies Dworkin's thoughts more. [URL]. [NEWLINE] Andrea Dworkin was far from a man-hater. [NEWLINE] [NEWLINE] [STARTQ] “Rape is nothing more or less than a conscious process of intimidation by which all men keep all women in a state of fear” — Susan Brownm
Label encoding: <s>You're right, this is a hornet's nest.  It's never a good idea to just take isolated quotes out of their context. [NEWLINE] [NEWLINE] [STARTQ] “I feel that ‘man-hating’ is an honourable and viable *political* act, that the oppressed have a right to class-hatred against the *class* that is oppressing them.” – Robin Morgan, Ms. Magazine Editor [ENDQ] [NEWLINE] This is from 1970.  I've added italics into the quote where they should be.  On the surface, this appears to be a statement about hating men, but taken in context where it was written, what she is really saying is she hates patriarchy and the men who uphold it.  That's the class she is talking about.  In the context of the times and the specific state of affairs she is talking about, she is writing in response to the extreme sexism of the times that was ongoing within the Leftist movement and the failure of the movement to live up to its revolutionary ideas.  She is writing with passion and exasperation.  You can't begin to understand the context of what she is saying without understanding what was going on when it was written.  Her use of language seems extreme but this is the style people used back then.  It is not what it appears to be. [NEWLINE] [NEWLINE] [STARTQ] “To call a man an animal is to flatter him; he’s a machine, a walking dildo.” -– Valerie Solanas [ENDQ] [NEWLINE] Valerie Solanas said worse things than that.  She probably did hate men, but she was a paranoid schizophrenic, and her ideas stemmed from the extreme sexism of the time (the 60s).  Look at wikipedia for more info. [NEWLINE] [NEWLINE] [STARTQ] “I want to see a man beaten to a bloody pulp with a high-heel shoved in his mouth, like an apple in the mouth of a pig.” — Andrea Dworkin [ENDQ] [NEWLINE] This is taken from fiction, it's something a character who was raped says. Please read this: [URL], as it also addresses some of the other quotes/people on your list.  Also, this interview clarifies Dworkin's thoughts more. [URL]. [NEWLINE] Andrea Dworkin was far from a man-hater. [NEWLINE] [NEWLINE] [STARTQ] “Rape is nothing more or less than a conscious process of intimidation by which all men keep all women in a state of fear” — Susan Brownm
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think inequality is an inevitablity in an organized society like ours, at least to a certain extent.<mask> would you spend the time, effort, and money to go to medical school<mask> you could make the same salary working the checkout line at a grocery store? We need to provide adequate incentive for people to want to take up certain careers. Some equality is even a good thing,<mask> it gives people a reason to work hard to advance their careers, innovate in their field, and try to better society<mask> they can better their own lives. [NEWLINE] [NEWLINE] <mask>, a system of gross inequality isn't fair. Those with power and wealth do things to increase their power and wealth, that kind of hegemony is arguably inevitable, and at the very least it's a persistent feature of societies with widespread inequality. [NEWLINE] [NEWLINE] A system isn't fair<mask> the inequality is<mask> widespread that people on the bottom range of incomes aren't able to support themselves and the top range of incomes are earning more than they're worth. Is a CEO's job really<mask> much more important and difficult than a rank-and-file employee that they deserve to make a thousand times more money? Or is that simply those at the top seeking to preserve and enhance their own power on the backs of those who truly work hard? Surely they deserve to earn enough to attract quality candidates to the job,<mask> do they really deserve<mask> much that it puts the employees that keep the company moving under persistent financial pressure? [NEWLINE] [NEWLINE] [STARTQ] it's achieved through a fair system and legal means [ENDQ] [NEWLINE] I wouldn't call it fair, necessarily. Economic mobility isn't<mask> high<mask> we'd like to think, and the majority of people remain in the social class their parents were in. The kids of those at the top are more likely to get jobs at the top, meaning they've likely never experienced work at minimum wage and have no idea<mask> life is like without tons of money. There's a disconnect between the rich and the poor in part due to a lack of understanding, and that disconnect leads to an insensitivity towards the poor from the rich. They think they got<mask> they were purely due to hard work, and think that the poor are there purely<mask> of laziness, and neither of those things are true (at least most of the time, obviously there are exceptions). <mask> you're led to believe that you earned your money in a fair system, you're going to do everything you can within the confines of that system to make more money, not understanding that societal barriers often prevent
Label encoding: <s>I think inequality is an inevitablity in an organized society like ours, at least to a certain extent. Why would you spend the time, effort, and money to go to medical school if you could make the same salary working the checkout line at a grocery store? We need to provide adequate incentive for people to want to take up certain careers. Some equality is even a good thing, as it gives people a reason to work hard to advance their careers, innovate in their field, and try to better society so they can better their own lives. [NEWLINE] [NEWLINE] However, a system of gross inequality isn't fair. Those with power and wealth do things to increase their power and wealth, that kind of hegemony is arguably inevitable, and at the very least it's a persistent feature of societies with widespread inequality. [NEWLINE] [NEWLINE] A system isn't fair if the inequality is so widespread that people on the bottom range of incomes aren't able to support themselves and the top range of incomes are earning more than they're worth. Is a CEO's job really so much more important and difficult than a rank-and-file employee that they deserve to make a thousand times more money? Or is that simply those at the top seeking to preserve and enhance their own power on the backs of those who truly work hard? Surely they deserve to earn enough to attract quality candidates to the job, but do they really deserve so much that it puts the employees that keep the company moving under persistent financial pressure? [NEWLINE] [NEWLINE] [STARTQ] it's achieved through a fair system and legal means [ENDQ] [NEWLINE] I wouldn't call it fair, necessarily. Economic mobility isn't as high as we'd like to think, and the majority of people remain in the social class their parents were in. The kids of those at the top are more likely to get jobs at the top, meaning they've likely never experienced work at minimum wage and have no idea what life is like without tons of money. There's a disconnect between the rich and the poor in part due to a lack of understanding, and that disconnect leads to an insensitivity towards the poor from the rich. They think they got where they were purely due to hard work, and think that the poor are there purely because of laziness, and neither of those things are true (at least most of the time, obviously there are exceptions).  If you're led to believe that you earned your money in a fair system, you're going to do everything you can within the confines of that system to make more money, not understanding that societal barriers often prevent
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hi, this is a common response I hear a lot and I would like to change your mind a bit. I use to have the same views<mask> you until I fell into poverty myself and had to use foodstamps for over a year. [NEWLINE] [NEWLINE] Let us pretend that you are in first grade math class. I tell you it is time to take a test. All you know is first grade math. You know nothing about chemistry. Nothing at all.<mask> I came up to you and gave you a test on chemistry,<mask> would you do? You could guess, and maybe a few times you would get lucky and answer a question correctly,<mask> most likely you would look at me like I was insane and tell me that you do not know anything about chemistry, and you will only take this test<mask> it has to do with first grade math. [NEWLINE] [NEWLINE] <mask> no one ever taught you math above a first grade level,<mask> would you learn second grade math and above? You might get lucky and find someone who would teach you. Or maybe you would teach yourself somehow.<mask> most likely, you simply would not bother with any level higher<mask> you would not have the insight or reasoning to go higher, nor would you have the resources to learn even<mask> you wanted to. [NEWLINE] [NEWLINE] Food nutrition is a lot like that for the poor and uneducated. It is not that they aren't capable of learning about healthy food and<mask> to use it, it is that they were never taught<mask> and they do not see their current habits<mask> bad. Their diets are<mask> they were raised on, and yes it is not good,<mask><mask> you were never taught about healthy eating and<mask> to read nutrition levels, it is not hard to see<mask><mask> many people eat badly. [NEWLINE] [NEWLINE] Food companies know this and capitalize on it. They will label their foods<mask> "WHOLE GRAIN" or "100% JUICE, NO SUGAR ADDED!" to imply that the food is healthy. People like you and I, we know<mask> to read labels and we know that just<mask> something is made with whole grain or has no additional sugar added, it does not mean it is healthy.<mask> try to imagine that chemistry example again. You can't understand the questions<mask> you were never taught<mask>. A lot of poor people cannot understand their nutrition labels for the same reason. [NEWLINE] [NEWLINE] This is only part of the reason. There are many other factors.<mask> you have no reliable ride, buying healthy food does cost more<mask> you would have
Label encoding: <s>Hi, this is a common response I hear a lot and I would like to change your mind a bit. I use to have the same views as you until I fell into poverty myself and had to use foodstamps for over a year. [NEWLINE] [NEWLINE] Let us pretend that you are in first grade math class. I tell you it is time to take a test. All you know is first grade math. You know nothing about chemistry. Nothing at all. If I came up to you and gave you a test on chemistry, what would you do? You could guess, and maybe a few times you would get lucky and answer a question correctly, but most likely you would look at me like I was insane and tell me that you do not know anything about chemistry, and you will only take this test if it has to do with first grade math. [NEWLINE] [NEWLINE] If no one ever taught you math above a first grade level, how would you learn second grade math and above? You might get lucky and find someone who would teach you. Or maybe you would teach yourself somehow. But most likely, you simply would not bother with any level higher because you would not have the insight or reasoning to go higher, nor would you have the resources to learn even if you wanted to. [NEWLINE] [NEWLINE] Food nutrition is a lot like that for the poor and uneducated. It is not that they aren't capable of learning about healthy food and how to use it, it is that they were never taught how and they do not see their current habits as bad. Their diets are what they were raised on, and yes it is not good, but if you were never taught about healthy eating and how to read nutrition levels, it is not hard to see why so many people eat badly. [NEWLINE] [NEWLINE] Food companies know this and capitalize on it. They will label their foods as "WHOLE GRAIN" or "100% JUICE, NO SUGAR ADDED!" to imply that the food is healthy. People like you and I, we know how to read labels and we know that just because something is made with whole grain or has no additional sugar added, it does not mean it is healthy. But try to imagine that chemistry example again. You can't understand the questions because you were never taught how. A lot of poor people cannot understand their nutrition labels for the same reason. [NEWLINE] [NEWLINE] This is only part of the reason. There are many other factors. If you have no reliable ride, buying healthy food does cost more because you would have
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>So I've always kind of thought that it was ridiculous that my hometowns line of defense against fires was a bunch of volunteers. I know that they do receive training<mask> knowing some of them personally they really take it all<mask> some joke- not saying that all volunteers do -<mask><mask><mask> that<mask> your going to be in charge of saving someones life and are going to risk your own life you should be a paid professional. [NEWLINE] [NEWLINE] Other than the argument that paying firefighters is too expensive for the state<mask> the heck else do all these small towns rely on locals to just figure it out themselves. [NEWLINE] [NEWLINE] **Edit:** [NEWLINE] [NEWLINE] **Please see thndrchld and Mine's conversation below** [NEWLINE] [NEWLINE] [STARTQ] And<mask> far a<mask> I said by taking it<mask> a joke I know a kid that joined<mask> he wanted to shack-up in a fire engine, An ex of mine's grandfather was the long running lead volunteer fighter and often joke that<mask> a N****'s house caught fire he'd let them burn- these are the things that went on in my small home town. [ENDQ] [STARTQ] And that is one of the main reason that i feel this should be handled by professionals, not just and guy off the street with his own agenda.-<mask> one more question-<mask> you were volunteering with either of the guys I mentioned<mask> would you do? and<mask> could you have those people removed? [ENDQ] [NEWLINE] **celeritas365 made a good point** [NEWLINE] [NEWLINE] [STARTQ] <mask> about mixed? In my town we have some professionals and some volunteers. Some of the volunteers are more for support and can't go into a burning building. The stuff that happens outside is still important.<mask> pay a full firefighter<mask> you don't need all of his or her skills?<mask>, much of<mask> fire departments do is checking for fires<mask> a house alarm goes off and doing small<mask> important things like chemically treating gasoline spills from auto accidents<mask> they don't catch fire.<mask> important, I don't see<mask> a team of ALL professional firefighters need to do these tasks. [ENDQ] [NEWLINE] <mask><mask> maybe thats<mask> i mean more- There should be professionals at all depts, no ALL volunteer depts should exist.- that<mask> solves the money issue with for example 1 pro for every 5 volunteers. [NEWLINE] [NEWLINE] **Ada1629<mask> had a good point on my thoughts** [NEWLINE] [STARTQ] <mask>, I feel uncomfortable with people working for free - it becomes suspect to me that something more nefarious is going on [ENDQ] [STARTQ] I<mask> question<mask> stringent the rules are
Label encoding: <s>So I've always kind of thought that it was ridiculous that my hometowns line of defense against fires was a bunch of volunteers. I know that they do receive training but knowing some of them personally they really take it all as some joke- not saying that all volunteers do - but I think that if your going to be in charge of saving someones life and are going to risk your own life you should be a paid professional. [NEWLINE] [NEWLINE] Other than the argument that paying firefighters is too expensive for the state why the heck else do all these small towns rely on locals to just figure it out themselves. [NEWLINE] [NEWLINE] **Edit:** [NEWLINE] [NEWLINE] **Please see thndrchld and Mine's conversation below** [NEWLINE] [NEWLINE] [STARTQ] And as far a what I said by taking it as a joke I know a kid that joined because he wanted to shack-up in a fire engine, An ex of mine's grandfather was the long running lead volunteer fighter and often joke that if a N****'s house caught fire he'd let them burn- these are the things that went on in my small home town. [ENDQ] [STARTQ] And that is one of the main reason that i feel this should be handled by professionals, not just and guy off the street with his own agenda.- so one more question- if you were volunteering with either of the guys I mentioned what would you do? and how could you have those people removed? [ENDQ] [NEWLINE] **celeritas365 made a good point** [NEWLINE] [NEWLINE] [STARTQ] What about mixed? In my town we have some professionals and some volunteers. Some of the volunteers are more for support and can't go into a burning building. The stuff that happens outside is still important. Why pay a full firefighter when you don't need all of his or her skills? Also, much of what fire departments do is checking for fires when a house alarm goes off and doing small yet important things like chemically treating gasoline spills from auto accidents so they don't catch fire. While important, I don't see why a team of ALL professional firefighters need to do these tasks. [ENDQ] [NEWLINE] I think maybe thats what i mean more- There should be professionals at all depts, no ALL volunteer depts should exist.- that also solves the money issue with for example 1 pro for every 5 volunteers. [NEWLINE] [NEWLINE] **Ada1629 Also had a good point on my thoughts** [NEWLINE] [STARTQ] However, I feel uncomfortable with people working for free - it becomes suspect to me that something more nefarious is going on [ENDQ] [STARTQ] I also question how stringent the rules are
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0424, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Alright, then you do take some of the bible literally,<mask> not other parts? Do you accept that homosexuality is an abomination and that homosexuals should be killed? Do you accept that slavery is an acceptable practice and that slave masters may kill their slaves? [ENDQ] [NEWLINE] [STARTQ] I have to ask these questions<mask> I have no idea<mask> you draw the line of literalism. [ENDQ] [NEWLINE] <mask> excellent questions. I've given a lot of thought about homosexuality, given the recent focus of it in the national media. My short answer is that I'm not sure really<mask> I stand. The longer version is that<mask><mask> homosexuality is a sin (i.e. that it's wrong),<mask> I don't have any problem with gay people getting married from a legal standpoint. I certainly don't think they should be killed. I have a good number of gay friends, and they're almost all pretty awesome people.<mask> they sin...<mask> do we all. Humans like to put degrees of 'badness' on things,<mask> the Bible is pretty clear in this regard<mask><mask>. A sin is a sin is a sin. Is them being gay any worse that me stealing that candy bar<mask> I was 5? Nope. We're all imperfect, that's just<mask> it is. [NEWLINE] [NEWLINE] <mask> for slaves, I don't think that slavery is an acceptable practice. It might have been at one time, culturally,<mask> it's certainly not now. Is it acceptable from a faith-based view, and was it ever? I don't think<mask>. From my understanding of<mask> the Bible tries to guide us to do, we should care about people. Forcing someone into slavery is inarguable - its not caring about people. Sure, there's probably a line there (*<mask> much* should you care),<mask> on a fundamental level, I don't think anyone could construe slavery<mask> caring about other people. [NEWLINE] [NEWLINE] That's part of<mask> I try to do,<mask> I do fail at it pretty often. I try and care about people, and I mean genuinely care about them. I'm not trying to go all hippie on you or anything,<mask> I genuinely hope that you're having a good day, that you're safe and comfortable, and that maybe by asking me these questions you might gain a little insight into<mask><mask><mask> Christian faith ought to be. I don't think you'll suddenly drop everything and believe<mask> I believe,<mask> I do hope that you understand at least a little of<mask> I'm coming from. I
Label encoding: <s> [STARTQ] Alright, then you do take some of the bible literally, but not other parts? Do you accept that homosexuality is an abomination and that homosexuals should be killed? Do you accept that slavery is an acceptable practice and that slave masters may kill their slaves? [ENDQ] [NEWLINE] [STARTQ] I have to ask these questions because I have no idea where you draw the line of literalism. [ENDQ] [NEWLINE] Also excellent questions. I've given a lot of thought about homosexuality, given the recent focus of it in the national media. My short answer is that I'm not sure really where I stand. The longer version is that I think homosexuality is a sin (i.e. that it's wrong), but I don't have any problem with gay people getting married from a legal standpoint. I certainly don't think they should be killed. I have a good number of gay friends, and they're almost all pretty awesome people. So they sin... as do we all. Humans like to put degrees of 'badness' on things, but the Bible is pretty clear in this regard I think. A sin is a sin is a sin. Is them being gay any worse that me stealing that candy bar when I was 5? Nope. We're all imperfect, that's just how it is. [NEWLINE] [NEWLINE] As for slaves, I don't think that slavery is an acceptable practice. It might have been at one time, culturally, but it's certainly not now. Is it acceptable from a faith-based view, and was it ever? I don't think so. From my understanding of what the Bible tries to guide us to do, we should care about people. Forcing someone into slavery is inarguable - its not caring about people. Sure, there's probably a line there (* How much* should you care), but on a fundamental level, I don't think anyone could construe slavery as caring about other people. [NEWLINE] [NEWLINE] That's part of what I try to do, though I do fail at it pretty often. I try and care about people, and I mean genuinely care about them. I'm not trying to go all hippie on you or anything, but I genuinely hope that you're having a good day, that you're safe and comfortable, and that maybe by asking me these questions you might gain a little insight into where I think Christian faith ought to be. I don't think you'll suddenly drop everything and believe what I believe, but I do hope that you understand at least a little of where I'm coming from. I
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you. It does mean a lot. [NEWLINE] [NEWLINE] I do get the math is unbelievable. That is<mask> I keep getting other people to check my numbers and math, and watching to see I'm not missing something. Maybe I forgot to add some spice in my calculations or something. Or maybe my measuring cups were all off, and I'd mistakenly believed the 1 cup was a 1/4 cup. Maybe I'm sleep eating. Something. Anything. I would be<mask> happy<mask> it were that simple. Time and again,<mask>, it is not. I even changed up<mask> I get my calories, thinking<mask> I kept the supposed calorie amount the same, and count was a miscalculation, changing up the food would change the results somehow. Tried between 500-800/day of the following, each for at least a few months: all salads, homemade Mexican flavored meat and veggies, fish and veggie stir fry, small meals of whatever I made my husband, hell, I even tried 500 calories per day of nothing<mask> junk snack food for one month. Same result every time: either no gain, or 2 lbs a month. It's been a<mask><mask> I gained 5 lbs in a month, and that could have been scale issues,<mask> I do definitely seem to be gaining an impossible 2 lbs most months, and no one can find the error. [NEWLINE] [NEWLINE] I don't understand<mask> it's possible.<mask> I finally convince someone I am not crazy or lying, there are two inevitable responses. [NEWLINE] [NEWLINE] * "Have you tried eating more?" Yes, I've tried, and I throw up, or gain weight faster. [NEWLINE] [NEWLINE] * "It must be<mask> you went through<mask> a kid, it fucked up your metabolism." I was starved, abused, and drugged<mask> being starved. I could believe this one, sorta,<mask> the physics weren't<mask> weird. It contributed to every other health complaint I've had in the past few years,<mask> naturally, that is everyone's assumption. I'm not sure the idea is scientifically sound,<mask>,<mask> I said earlier, I've heard that starvation gaining is just a myth/legend. I've<mask> heard it's just<mask> insanely rare and difficult to produce it might<mask> well be legend, and it could be possible. For now this one is 'inconclusive<mask> unlikely' [NEWLINE] [NEWLINE] I am currently trying Bento Boxes,<mask> it is an easy format to eat tasty and healthy food, and the portions/ratios are easy to monitor
Label encoding: <s>Thank you. It does mean a lot. [NEWLINE] [NEWLINE] I do get the math is unbelievable. That is why I keep getting other people to check my numbers and math, and watching to see I'm not missing something. Maybe I forgot to add some spice in my calculations or something. Or maybe my measuring cups were all off, and I'd mistakenly believed the 1 cup was a 1/4 cup. Maybe I'm sleep eating. Something. Anything. I would be so happy if it were that simple. Time and again, though, it is not. I even changed up how I get my calories, thinking if I kept the supposed calorie amount the same, and count was a miscalculation, changing up the food would change the results somehow. Tried between 500-800/day of the following, each for at least a few months: all salads, homemade Mexican flavored meat and veggies, fish and veggie stir fry, small meals of whatever I made my husband, hell, I even tried 500 calories per day of nothing but junk snack food for one month. Same result every time: either no gain, or 2 lbs a month. It's been a while since I gained 5 lbs in a month, and that could have been scale issues, but I do definitely seem to be gaining an impossible 2 lbs most months, and no one can find the error. [NEWLINE] [NEWLINE] I don't understand how it's possible. When I finally convince someone I am not crazy or lying, there are two inevitable responses. [NEWLINE] [NEWLINE] * "Have you tried eating more?" Yes, I've tried, and I throw up, or gain weight faster. [NEWLINE] [NEWLINE] * "It must be what you went through as a kid, it fucked up your metabolism." I was starved, abused, and drugged while being starved. I could believe this one, sorta, if the physics weren't so weird. It contributed to every other health complaint I've had in the past few years, so naturally, that is everyone's assumption. I'm not sure the idea is scientifically sound, though, as I said earlier, I've heard that starvation gaining is just a myth/legend. I've also heard it's just so insanely rare and difficult to produce it might as well be legend, and it could be possible. For now this one is 'inconclusive but unlikely' [NEWLINE] [NEWLINE] I am currently trying Bento Boxes, as it is an easy format to eat tasty and healthy food, and the portions/ratios are easy to monitor
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The obvious example here would be the previous presidential election, fought primarily between Obama, Romney, and Ron Paul. [ENDQ] [NEWLINE] No, it wasn't. You are clearly completely ignorant of Ron Paul's campaign.<mask> made him more notable than typical candidates with Libertarian policies is that he got a lot of momentum within the Republican party. He was a Republican member of the house of Representatives, and a candidate in the Republican primaries.<mask> he lost the Republican primaries, he dropped out of the race. At the general election he wasn't on the ballot. [NEWLINE] [NEWLINE] I voted for Ron Paul in the 2008 and 2012 primaries.<mask> the general elections came around, I voted for Bob Barr and Gary Johnson in 2008 &amp; 2012 respectively. I knew that neither would win,<mask> I wasn't about to vote for either mainstream candidate, and I feel this is more useful than not showing up at the polls. [NEWLINE] [NEWLINE] For most of the things I care about, the major parties are almost indistinguishable.<mask><mask> with Republicans on legislating their religious values, and Democrats for their nanny-state tendencies,<mask> those are small enough parts of the big picture they're not going to push me one way or the other, given all the ways the major parties are the same.<mask><mask> the foreign policy is pretty terrible. We're the policemen of the world, whether they want it or not, whether our people want it or not. Our monetary policy is pretty bad too: the Fed gets to print money, and Congress won't even investigate whether their policies are in American interests. We've essentially given the right to print legal tender to private banks. Our fiscal policies are equally terrible. Even the Paul Ryan plan, which was lambasted by the left for being way too conservative, ran major deficits for the next twenty years. And that's to say nothing of corruption issues,<mask> getting financing for the next campaign supersedes constituency interests. [NEWLINE] [NEWLINE] <mask> I don't like the mainstream candidates,<mask> obviously the third party candidates don't have a snowball's chance in hell.<mask> are my options? I could pick which of the fringe issues are important to me and cast a vote for someone I generally don't like to support an issue I don't think is a big deal. I could skip the voting booth altogether and save myself some trouble. Or I could vote for someone I generally agree with, even knowing they won't win. [NEWLINE] [NEWLINE] I have always chosen the last option, for a couple of reasons. First, ballot access and campaign finance rules favor
Label encoding: <s> [STARTQ] The obvious example here would be the previous presidential election, fought primarily between Obama, Romney, and Ron Paul. [ENDQ] [NEWLINE] No, it wasn't. You are clearly completely ignorant of Ron Paul's campaign. What made him more notable than typical candidates with Libertarian policies is that he got a lot of momentum within the Republican party. He was a Republican member of the house of Representatives, and a candidate in the Republican primaries. When he lost the Republican primaries, he dropped out of the race. At the general election he wasn't on the ballot. [NEWLINE] [NEWLINE] I voted for Ron Paul in the 2008 and 2012 primaries. When the general elections came around, I voted for Bob Barr and Gary Johnson in 2008 &amp; 2012 respectively. I knew that neither would win, but I wasn't about to vote for either mainstream candidate, and I feel this is more useful than not showing up at the polls. [NEWLINE] [NEWLINE] For most of the things I care about, the major parties are almost indistinguishable. I disagree with Republicans on legislating their religious values, and Democrats for their nanny-state tendencies, but those are small enough parts of the big picture they're not going to push me one way or the other, given all the ways the major parties are the same. I think the foreign policy is pretty terrible. We're the policemen of the world, whether they want it or not, whether our people want it or not. Our monetary policy is pretty bad too: the Fed gets to print money, and Congress won't even investigate whether their policies are in American interests. We've essentially given the right to print legal tender to private banks. Our fiscal policies are equally terrible. Even the Paul Ryan plan, which was lambasted by the left for being way too conservative, ran major deficits for the next twenty years. And that's to say nothing of corruption issues, where getting financing for the next campaign supersedes constituency interests. [NEWLINE] [NEWLINE] So I don't like the mainstream candidates, but obviously the third party candidates don't have a snowball's chance in hell. What are my options? I could pick which of the fringe issues are important to me and cast a vote for someone I generally don't like to support an issue I don't think is a big deal. I could skip the voting booth altogether and save myself some trouble. Or I could vote for someone I generally agree with, even knowing they won't win. [NEWLINE] [NEWLINE] I have always chosen the last option, for a couple of reasons. First, ballot access and campaign finance rules favor
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay, lets start with your arguments, which are really more talking points. To eliminate the crux of your first item, animal welfare, I'd point out that<mask> this is an issue, especially<mask> we see clearly egregious mistreatment of animals, it has no impact on your question. The fact that people can mistreat animals in a farm setting should have no direct impact on whether you eat meat in general. [NEWLINE] [NEWLINE] I'm not sure<mask> you're aware of them,<mask><mask> called 'puppy mills' are centers<mask> they mistreat dogs and other animals you might get<mask> a pet. That argument would be exactly the same to saying that we shouldn't have pet dogs<mask> it's possible some people mistreated some animals in order to breed, raise and sell them. [NEWLINE] [NEWLINE] Next, environmental impact and efficiency.<mask> we accept that it is possible for an animal to be bred properly, live comfortably, and later slaughtered and<mask>chered in a humane way -- something a lion or other carnivorous or omnivorous animal certainly wouldn't consider -- then<mask> are we arguing exactly? Animals walk the earth and use resources<mask> they do. An animal takes up the same amount of resources in a season whether it is in a farm or in the wild. Should we say that those animals should never exist<mask> that the earth saves resources? Or, put another way,<mask>'s our goal here, to simply save resources of an inanimate object, or to live in a sustainable way? The earth has a certain sustainable amount of resources, and given sustainable farming practices, these can be continually used for everyone to survive, animals, plants and humans included. The earth has sustained carnivores for millions of years,<mask> we can see that eating meat in general has a long history of sustainability. [NEWLINE] [NEWLINE] <mask>, you may believe that a vegetarian diet is no less healthy than an omnivorous diet,<mask> that's not always true. There is no way of getting around the fact that humans were born omnivores - any argument to the contrary is simply incorrect. Some people can eat only a certain subset of the food we were designed to digest and get by fine, others have serious issues<mask> they move to this diet. I'd be happy to link you to several articles mentioning health issues depending on a person and the veg and fruits they eat<mask> you'd like,<mask> you likely know very well that<mask> you're not careful you can end up short on certain vitamins and minerals. There's a reason for this. Certain people
Label encoding: <s>Okay, lets start with your arguments, which are really more talking points. To eliminate the crux of your first item, animal welfare, I'd point out that while this is an issue, especially when we see clearly egregious mistreatment of animals, it has no impact on your question. The fact that people can mistreat animals in a farm setting should have no direct impact on whether you eat meat in general. [NEWLINE] [NEWLINE] I'm not sure if you're aware of them, but so called 'puppy mills' are centers where they mistreat dogs and other animals you might get as a pet. That argument would be exactly the same to saying that we shouldn't have pet dogs because it's possible some people mistreated some animals in order to breed, raise and sell them. [NEWLINE] [NEWLINE] Next, environmental impact and efficiency. If we accept that it is possible for an animal to be bred properly, live comfortably, and later slaughtered and butchered in a humane way -- something a lion or other carnivorous or omnivorous animal certainly wouldn't consider -- then what are we arguing exactly? Animals walk the earth and use resources as they do. An animal takes up the same amount of resources in a season whether it is in a farm or in the wild. Should we say that those animals should never exist so that the earth saves resources? Or, put another way, what's our goal here, to simply save resources of an inanimate object, or to live in a sustainable way? The earth has a certain sustainable amount of resources, and given sustainable farming practices, these can be continually used for everyone to survive, animals, plants and humans included. The earth has sustained carnivores for millions of years, so we can see that eating meat in general has a long history of sustainability. [NEWLINE] [NEWLINE] Lastly, you may believe that a vegetarian diet is no less healthy than an omnivorous diet, but that's not always true. There is no way of getting around the fact that humans were born omnivores - any argument to the contrary is simply incorrect. Some people can eat only a certain subset of the food we were designed to digest and get by fine, others have serious issues when they move to this diet. I'd be happy to link you to several articles mentioning health issues depending on a person and the veg and fruits they eat if you'd like, but you likely know very well that if you're not careful you can end up short on certain vitamins and minerals. There's a reason for this. Certain people
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. By your hazing argument, we shouldn't just ban greek organizations. We should ban hazing organizations. That includes acapella groups, comedy troupes, the Quidditch teams, marching band, and even, at my school, the student-led campus tour groups. [Even Princeton's Eating Clubs have been accused of rape.]( [URL] ) None of these groups including frats officially haze. I'm not saying that hazing is okay,<mask> I don't think a blanket ban on Greeks is a solution for hazing. [NEWLINE] [NEWLINE] 2. You may<mask><mask> these groups provide a function or service to the community,<mask> I would say that utility is no greater than the community service and philanthropy provided by greeks. "I personally believe this can be accomplished without the Greek system in place."<mask><mask>? Can you tell me specifically of another way that hundreds of otherwise disinterested or minimally motivated men and women will be pushed into philanthropy and service? Fyi, most greek organizations have  required minimum service hours or philanthropic activities, not to mention that they know that doing good work improves their reputation. May not be the purest of intentions,<mask> the results are the same. [NEWLINE] [NEWLINE] 2. You emphasize frats, and in doing<mask> you are marginalizing large numbers of greek organizations that do not fall under your generalizations. For example most sororities do not systematically promote sexual harassment or rape. There<mask> exist multicultural Greeks, which<mask><mask> to service and philanthropy are often dedicated to hosting leadership workshops, racial discussion seminars, cultural showcases, and scholarships for minority members of the community. I would consider all of these programs enhance the experience of the college community<mask> they offer learning opportunities. [NEWLINE] [NEWLINE] 3. You seek evidence that Greek life enhances the experience for non-greeks. I would consider it unreasonable for that to be a criterion for whether a student organization should exist. Does the knitting club benefit others? Does the Association of Engineers enhance student life for all? Or are you saying that criterion should only be applied<mask> weighing the potential harm of an organization versus the potential benefit. I must say the potential harm of a greek organization is very hard to measure, and I would be against a University banning a frat arbitrarily, without any evidence. Decisions must not be based on rumors,<mask> bad the rumors may be (see: Rolling Stone article on the UVa rape, accused frat Pi Psi, accusations revealed to be false). [NEWLINE] [NEWLINE] 4
Label encoding: <s>1. By your hazing argument, we shouldn't just ban greek organizations. We should ban hazing organizations. That includes acapella groups, comedy troupes, the Quidditch teams, marching band, and even, at my school, the student-led campus tour groups. [Even Princeton's Eating Clubs have been accused of rape.]( [URL] ) None of these groups including frats officially haze. I'm not saying that hazing is okay, but I don't think a blanket ban on Greeks is a solution for hazing. [NEWLINE] [NEWLINE] 2. You may argue that these groups provide a function or service to the community, but I would say that utility is no greater than the community service and philanthropy provided by greeks. "I personally believe this can be accomplished without the Greek system in place." How so? Can you tell me specifically of another way that hundreds of otherwise disinterested or minimally motivated men and women will be pushed into philanthropy and service? Fyi, most greek organizations have  required minimum service hours or philanthropic activities, not to mention that they know that doing good work improves their reputation. May not be the purest of intentions, but the results are the same. [NEWLINE] [NEWLINE] 2. You emphasize frats, and in doing so you are marginalizing large numbers of greek organizations that do not fall under your generalizations. For example most sororities do not systematically promote sexual harassment or rape. There also exist multicultural Greeks, which in addition to service and philanthropy are often dedicated to hosting leadership workshops, racial discussion seminars, cultural showcases, and scholarships for minority members of the community. I would consider all of these programs enhance the experience of the college community because they offer learning opportunities. [NEWLINE] [NEWLINE] 3. You seek evidence that Greek life enhances the experience for non-greeks. I would consider it unreasonable for that to be a criterion for whether a student organization should exist. Does the knitting club benefit others? Does the Association of Engineers enhance student life for all? Or are you saying that criterion should only be applied when weighing the potential harm of an organization versus the potential benefit. I must say the potential harm of a greek organization is very hard to measure, and I would be against a University banning a frat arbitrarily, without any evidence. Decisions must not be based on rumors, however bad the rumors may be (see: Rolling Stone article on the UVa rape, accused frat Pi Psi, accusations revealed to be false). [NEWLINE] [NEWLINE] 4
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0504, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This brings up another point.  Your OP is tacitly blaming all Gazans for their role in electing Hamas. <mask> about that Gazans didn't vote for Hamas or voted for Hamas for the governance promises, not the rhetoric? <mask><mask> they voted for Hamas<mask> Hamas was the only party that made an attractive argument in providing social services? [NEWLINE] You can be sure there are some in this category who have died from Israeli actions<mask> 2006.  There would<mask> be collateral deaths from Gazans who are politically apathetic, who didn't vote for anyone.  Not to mention the innocents, the women and children that have become collateral damage. [NEWLINE] [NEWLINE] Even<mask> you accept the idea that those who voted for Hamas deserve it, the ones that didn't support them would not,<mask> are simply unlucky for being born in Gaza. [NEWLINE] [NEWLINE] For an US analogy, no-one voted for Bush<mask> they knew the armed forces will end up in Afghanistan and Iraq for a decade, along with the decline of American influence world-wide.  Even the ones that voted for Bush may not have supported the resultant Neoconservative foreign policy<mask> for domestic issues. <mask> well, the 9/11 victims may not have supported American actions in the Middle-East that would be one principle cause of the terrorist attack. [NEWLINE] [NEWLINE] <mask> you argue<mask>, the Gazans that deserve the Israel attacks are the ones that voted for Hamas specifically for the rhetoric about Israel, your argument may have more merit. <mask> this would be abstracting the whole conflict into a 2-dimensional black/white debate.  For many reasons, this is not a good approach for objectivity,<mask><mask> each side is portrayed.  Here's one perspective on<mask> Hamas won in Gaza in 2006: [NEWLINE] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] This is an **extremely** good read, and prophetic in describing the situation today. [NEWLINE] [NEWLINE] <mask><mask> ; Hamas, prior to the blockade, had a successful charity and social welfare program that helped Gazans and presented an attractive alternative to the Palestinian Authority, seen<mask> corrupt and ineffective. [NEWLINE] [NEWLINE] [STARTQ] People who voted for Hamas emphasize not only the heroic acts of its combatants,<mask><mask> its reputation for clean conduct, modesty, and honesty, which have been pointedly contrasted with the corruption of the Palestinian Authority. Many of its followers do not subscribe to religious fundamentalism,<mask> rather support the organization due to its pragmatic approach characterized by support for the short-term objective of a Palestinian state in the West Bank, Gaza Strip and East Jerusalem,
Label encoding: <s>This brings up another point.  Your OP is tacitly blaming all Gazans for their role in electing Hamas.  What about that Gazans didn't vote for Hamas or voted for Hamas for the governance promises, not the rhetoric?  What if they voted for Hamas because Hamas was the only party that made an attractive argument in providing social services? [NEWLINE] You can be sure there are some in this category who have died from Israeli actions since 2006.  There would also be collateral deaths from Gazans who are politically apathetic, who didn't vote for anyone.  Not to mention the innocents, the women and children that have become collateral damage. [NEWLINE] [NEWLINE] Even if you accept the idea that those who voted for Hamas deserve it, the ones that didn't support them would not, but are simply unlucky for being born in Gaza. [NEWLINE] [NEWLINE] For an US analogy, no-one voted for Bush because they knew the armed forces will end up in Afghanistan and Iraq for a decade, along with the decline of American influence world-wide.  Even the ones that voted for Bush may not have supported the resultant Neoconservative foreign policy but for domestic issues.  As well, the 9/11 victims may not have supported American actions in the Middle-East that would be one principle cause of the terrorist attack. [NEWLINE] [NEWLINE] If you argue however, the Gazans that deserve the Israel attacks are the ones that voted for Hamas specifically for the rhetoric about Israel, your argument may have more merit.  But this would be abstracting the whole conflict into a 2-dimensional black/white debate.  For many reasons, this is not a good approach for objectivity, despite how each side is portrayed.  Here's one perspective on why Hamas won in Gaza in 2006: [NEWLINE] [NEWLINE] [URL] / [NEWLINE] [NEWLINE] This is an **extremely** good read, and prophetic in describing the situation today. [NEWLINE] [NEWLINE] TLDR ; Hamas, prior to the blockade, had a successful charity and social welfare program that helped Gazans and presented an attractive alternative to the Palestinian Authority, seen as corrupt and ineffective. [NEWLINE] [NEWLINE] [STARTQ] People who voted for Hamas emphasize not only the heroic acts of its combatants, but also its reputation for clean conduct, modesty, and honesty, which have been pointedly contrasted with the corruption of the Palestinian Authority. Many of its followers do not subscribe to religious fundamentalism, but rather support the organization due to its pragmatic approach characterized by support for the short-term objective of a Palestinian state in the West Bank, Gaza Strip and East Jerusalem,
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> do you gain from studying philosophy that could not be gained from thoughtful introspection? [ENDQ] [NEWLINE] Ask yourself the same question about math. Sure, with knowledge of basic numbers, we could theoretically use those numbers to derive all the theorems of geometry and calculus, reinvent fields like engineering, and create whole new mathematical principles simply by thinking about them,<mask> math is a completely human invention.<mask>, this is obviously foolish<mask> countless geniuses have already made those discoveries; it makes much more sense to learn about<mask>'s already been discovered, even<mask> (and especially<mask> ) you hope to make new discoveries yourself. And to assume that a person who knows very little about math could come up with those theories is,<mask> strictly theoretically true, practically ridiculous. [NEWLINE] [NEWLINE] The same goes for philosophy. Sure, you COULD use<mask> you know about the world to independently come up with Wittgenstein's *Tractatus*,<mask> I bet you won't,<mask> to do<mask> requires a monumental amount of existing philosophical knowledge that is impossible to figure out in a lifetime without some prior reference.<mask> for the reason alone that it is virtually impossible to reach high levels of philosophy without first understanding the foundations, studying philosophy is a worthwhile endeavor<mask> it gives the student access to knowledge that would have been highly arcane to them had they not studied the subject. [NEWLINE] [NEWLINE] <mask> your main argument seems to be that studying philosophy is a waste of time<mask> it has no practical use. For one thing, the assumption that the only reason to study something is to major in that field is highly incorrect. Many people including myself leave college and never once use their major in the real world. Some use college<mask> an opportunity to train for professional skills, others see it<mask> a way to grow themselves<mask> people and learn more about the world in general before getting a job and no longer having the opportunity. Both of these are worthwhile notions, and the very fact that people choose to study philosophy<mask> opposed to more "useful" majors is proof that its existence is highly desirable at least to the people who are interested in it. [NEWLINE] [NEWLINE] The biggest takeaway I get from your post is that you did not study philosophy, and you have somewhat of a misunderstanding of<mask> it is. Many people who have never studied the subject believe to learn philosophy is to sit around shooting the shit about hypothetical questions and hackneyed moral quandaries. In actuality, my philosophical education (I minored in it<mask> took<mask> many classes in it<mask> a major) involved tons of rigorous
Label encoding: <s> [STARTQ] What do you gain from studying philosophy that could not be gained from thoughtful introspection? [ENDQ] [NEWLINE] Ask yourself the same question about math. Sure, with knowledge of basic numbers, we could theoretically use those numbers to derive all the theorems of geometry and calculus, reinvent fields like engineering, and create whole new mathematical principles simply by thinking about them, because math is a completely human invention. However, this is obviously foolish because countless geniuses have already made those discoveries; it makes much more sense to learn about what's already been discovered, even if (and especially if ) you hope to make new discoveries yourself. And to assume that a person who knows very little about math could come up with those theories is, while strictly theoretically true, practically ridiculous. [NEWLINE] [NEWLINE] The same goes for philosophy. Sure, you COULD use what you know about the world to independently come up with Wittgenstein's *Tractatus*, but I bet you won't, because to do so requires a monumental amount of existing philosophical knowledge that is impossible to figure out in a lifetime without some prior reference. So for the reason alone that it is virtually impossible to reach high levels of philosophy without first understanding the foundations, studying philosophy is a worthwhile endeavor because it gives the student access to knowledge that would have been highly arcane to them had they not studied the subject. [NEWLINE] [NEWLINE] But your main argument seems to be that studying philosophy is a waste of time because it has no practical use. For one thing, the assumption that the only reason to study something is to major in that field is highly incorrect. Many people including myself leave college and never once use their major in the real world. Some use college as an opportunity to train for professional skills, others see it as a way to grow themselves as people and learn more about the world in general before getting a job and no longer having the opportunity. Both of these are worthwhile notions, and the very fact that people choose to study philosophy as opposed to more "useful" majors is proof that its existence is highly desirable at least to the people who are interested in it. [NEWLINE] [NEWLINE] The biggest takeaway I get from your post is that you did not study philosophy, and you have somewhat of a misunderstanding of what it is. Many people who have never studied the subject believe to learn philosophy is to sit around shooting the shit about hypothetical questions and hackneyed moral quandaries. In actuality, my philosophical education (I minored in it but took as many classes in it as a major) involved tons of rigorous
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Hurting people never made anyone apart from sadists feel good. Even<mask> revenge feels justified and righteous in the moment, you will likely eventually come to regret inflicting that much pain on another human, and it certainly won't make you feel better. [ENDQ] [NEWLINE] Have you asked everyone? I'm able to admit that many (<mask> not a majority) of people would likely not benefit from or be able to harm another person in such a way,<mask> it's<mask> very likely some would. I don't think it's wrong to take some pleasure in showing a monster<mask> it feels like to be helpless. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> to get to the heart of this, we need to talk about<mask> you hope to gain from this. The stated goal of the criminal justice system in developed nations is to deter crime and rehabilitate criminals. Whether or not it does this effectively is up for debate,<mask> torture doesn't do that.<mask> should I show any loyalty or respect to a justice system that is just going to inflict pain to try and keep me in line? Whether or not I'm ever sentenced to torture,<mask> should I respect them<mask> a moral authority at all?<mask> should I follow any law of this society?<mask> police and judges can hand out torturing<mask> sentences,<mask> on earth shouldn't I shoot them on sight? [ENDQ] [NEWLINE] I'm aware of supposed function of the justice system,<mask> I don't believe people who have ruined lives deserve rehabilitation.<mask> I stated in previous comments,<mask> should Joe get to receive treatment and reconnect with society<mask> Joe killed someone who will never get to experience life, and their families will be forever scarred?  I don't believe that's justice, and I don't believe Joe should have that right. We cater to the needs of people like Joe with taxpayer money every day<mask> the Joes of the world have done nothing<mask> cause harm. [NEWLINE] [NEWLINE] <mask> someone needs the fear of punishment (torture, jail-time, a slap on the wrist, etc) to avoid causing lasting harm to another individual, then that person likely shouldn't be a member of civilized society to begin with. There's<mask> no obligation to respect anyone<mask> a moral authority<mask><mask><mask> their stance is rational and sound.<mask> the laws have a logical reason for being in place, we tow the line<mask><mask> personal opinion on the matter. [NEWLINE] [NEWLINE] [STARTQ] I understand that people like to dehumanize criminals,<mask> no man, no matter<mask> heinous his crime, was born a monster. Monsters are made by
Label encoding: <s> [STARTQ] Hurting people never made anyone apart from sadists feel good. Even if revenge feels justified and righteous in the moment, you will likely eventually come to regret inflicting that much pain on another human, and it certainly won't make you feel better. [ENDQ] [NEWLINE] Have you asked everyone? I'm able to admit that many ( if not a majority) of people would likely not benefit from or be able to harm another person in such a way, but it's also very likely some would. I don't think it's wrong to take some pleasure in showing a monster what it feels like to be helpless. [NEWLINE] [NEWLINE] [STARTQ] I think to get to the heart of this, we need to talk about what you hope to gain from this. The stated goal of the criminal justice system in developed nations is to deter crime and rehabilitate criminals. Whether or not it does this effectively is up for debate, but torture doesn't do that. Why should I show any loyalty or respect to a justice system that is just going to inflict pain to try and keep me in line? Whether or not I'm ever sentenced to torture, why should I respect them as a moral authority at all? Why should I follow any law of this society? If police and judges can hand out torturing as sentences, why on earth shouldn't I shoot them on sight? [ENDQ] [NEWLINE] I'm aware of supposed function of the justice system, but I don't believe people who have ruined lives deserve rehabilitation. As I stated in previous comments, why should Joe get to receive treatment and reconnect with society if Joe killed someone who will never get to experience life, and their families will be forever scarred?  I don't believe that's justice, and I don't believe Joe should have that right. We cater to the needs of people like Joe with taxpayer money every day when the Joes of the world have done nothing but cause harm. [NEWLINE] [NEWLINE] If someone needs the fear of punishment (torture, jail-time, a slap on the wrist, etc) to avoid causing lasting harm to another individual, then that person likely shouldn't be a member of civilized society to begin with. There's also no obligation to respect anyone as a moral authority as long as their stance is rational and sound. If the laws have a logical reason for being in place, we tow the line regardless of personal opinion on the matter. [NEWLINE] [NEWLINE] [STARTQ] I understand that people like to dehumanize criminals, but no man, no matter how heinous his crime, was born a monster. Monsters are made by
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My view will not be changed<mask> you tell me that people just see me<mask> more professional or educated<mask> I'm white [ENDQ] [NEWLINE] That's the thing... they probably don't, like you said.  White people typically are not judged positively or negatively exclusively<mask> of their race.  We are the default.  People see white skin and immediately start looking at other things, like clothing, to form opinions.  That is<mask> we are privileged in terms of our race. [NEWLINE] [NEWLINE] <mask>, for a POC, their skin color *does* tend to affect someone's opinion of their character.  I won't go into listing the offensive stereotypes,<mask><mask><mask> most of us are aware of them. [NEWLINE] [NEWLINE] <mask> for sources, [here]( [URL] ) is a study that showed racial bias in hiring based on ethnic-sounding names. [NEWLINE] [NEWLINE] <mask> numerous scholarships available for minority students (which, btw, is intended to help level the playing field<mask> of racial disparities in income status): [NEWLINE] [NEWLINE] [STARTQ] [Black and Hispanic students are dramatically underrepresented in the most selective colleges, [ENDQ] even after controlling for family income. The probability of enrolling in a highly selective [NEWLINE] college is five times greater for white students than black students. Even after controlling for [NEWLINE] income, white students are two to three times<mask> likely<mask> black students to gain admission to [NEWLINE] highly selective colleges. These racial disparities appear to have grown in the last 30 years. [NEWLINE] <mask> the racial disparity in selective college admissions persists even after controlling for [NEWLINE] income, income-based admissions practices will not eliminate the racial disparities]( [URL] %20income%20%26%20selective%20college%20enrollment%20august%203%202012.pdf) [NEWLINE] [NEWLINE] Going back to the income disparities, [the median income for white families in 2009 was 62k, compared to 38k for black families, 75k for Asian/PI families, and 39k for Hispanic families]( [URL].pdf). [NEWLINE] [NEWLINE] Edit: [NEWLINE] [NEWLINE] [STARTQ] Growing up a gangbanger lifestyle is not a race issue, it's a culture issue [ENDQ] [NEWLINE] It's actually both. <mask> black people are more likely to be born into poor economic conditions, they are<mask> more likely to fall into the "gangbanger" (*cough*) lifestyle.  You're splitting hairs by trying to separate the two. [NEWLINE] [NEWLINE] Edit 2: [NEWLINE] [NEWLINE] [STARTQ] I'm getting a lot of replies citing<mask> ethnic sounding names vs white sounding names affect job interviews. This is a cultural issue
Label encoding: <s> [STARTQ] My view will not be changed because you tell me that people just see me as more professional or educated because I'm white [ENDQ] [NEWLINE] That's the thing... they probably don't, like you said.  White people typically are not judged positively or negatively exclusively because of their race.  We are the default.  People see white skin and immediately start looking at other things, like clothing, to form opinions.  That is why we are privileged in terms of our race. [NEWLINE] [NEWLINE] However, for a POC, their skin color *does* tend to affect someone's opinion of their character.  I won't go into listing the offensive stereotypes, but I think most of us are aware of them. [NEWLINE] [NEWLINE] As for sources, [here]( [URL] ) is a study that showed racial bias in hiring based on ethnic-sounding names. [NEWLINE] [NEWLINE] Despite numerous scholarships available for minority students (which, btw, is intended to help level the playing field because of racial disparities in income status): [NEWLINE] [NEWLINE] [STARTQ] [Black and Hispanic students are dramatically underrepresented in the most selective colleges, [ENDQ] even after controlling for family income. The probability of enrolling in a highly selective [NEWLINE] college is five times greater for white students than black students. Even after controlling for [NEWLINE] income, white students are two to three times as likely as black students to gain admission to [NEWLINE] highly selective colleges. These racial disparities appear to have grown in the last 30 years. [NEWLINE] Because the racial disparity in selective college admissions persists even after controlling for [NEWLINE] income, income-based admissions practices will not eliminate the racial disparities]( [URL] %20income%20%26%20selective%20college%20enrollment%20august%203%202012.pdf) [NEWLINE] [NEWLINE] Going back to the income disparities, [the median income for white families in 2009 was 62k, compared to 38k for black families, 75k for Asian/PI families, and 39k for Hispanic families]( [URL].pdf). [NEWLINE] [NEWLINE] Edit: [NEWLINE] [NEWLINE] [STARTQ] Growing up a gangbanger lifestyle is not a race issue, it's a culture issue [ENDQ] [NEWLINE] It's actually both.  Since black people are more likely to be born into poor economic conditions, they are also more likely to fall into the "gangbanger" (*cough*) lifestyle.  You're splitting hairs by trying to separate the two. [NEWLINE] [NEWLINE] Edit 2: [NEWLINE] [NEWLINE] [STARTQ] I'm getting a lot of replies citing how ethnic sounding names vs white sounding names affect job interviews. This is a cultural issue
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0827, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Firstly, thank you for the reply. Too often people are only here to confirm their own beliefs. [NEWLINE] [NEWLINE] [STARTQ] My argument is that<mask><mask><mask> feminists and the MRM are separate, they are going to be working against each other instead of together, and that's a problem. [ENDQ] [NEWLINE] They are not working against each other, they are highlighting different issues. An equivalent would be to say that Rohingya people of Myanmar, the most persecuted people in the world, shouldn't protest for themselves<mask> the anti racism movement is already well know. The same could be said for any small persecuted peoples, such<mask> the Kurds. [NEWLINE] [NEWLINE] [STARTQ] My point is that they're fighting a common enemy and should stop getting waylaid by fighting each other. [ENDQ] [NEWLINE] Who or<mask> is the common enemy? Surely it's our current manifestation of our society's culture? Is there only one way to fight an idea<mask> big<mask> that? During the anti apartheid movement, the biggest political movements involved were the African National Congress, the South African Communist Party, and the Pan African Congress,<mask> well<mask> many religious and social organisations. Should they amalgamate into one organisation<mask> there could be in-fighting, or<mask> one is more famous than another?<mask> cannot other organisations exist to fight a common cause. [NEWLINE] [NEWLINE] [STARTQ] <mask> my point was that men's issues were too divisive and shouldn't be talked about, this would be a fair argument.<mask> it isn't. It's that we absolutely should be talking about them--<mask> feminists and gender egalitarians. [ENDQ] [NEWLINE] <mask> are we not allowed to talk about them<mask> feminists, men's rights activists, and gender egalitarians? Just<mask> one movement is already known<mask> a household name, should no others exist? [NEWLINE] [NEWLINE] [STARTQ] We can talk about men's issues in the context of sexism generally/feminism or under the umbrella of gender egalitarianism, without the us vs. them mentality that the MRM has inspired. [ENDQ] [NEWLINE] Surely this attitude is the only thing creating the us vs them mentality. "We can talk about these things,<mask> they can't".<mask> you were arguing for using only the gender egalitarianism movement I could understand,<mask><mask> can a movement built to represent women in an oppressive society be better at talking about men's oppression under the same oppressive society than one built to represent men? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> feminism is the best candidate for becoming a truly gender egalitarian movement,<mask> it already has mainstream sociopolitical support whereas the MRM is dismissed<mask> a joke and
Label encoding: <s>Firstly, thank you for the reply. Too often people are only here to confirm their own beliefs. [NEWLINE] [NEWLINE] [STARTQ] My argument is that as long as feminists and the MRM are separate, they are going to be working against each other instead of together, and that's a problem. [ENDQ] [NEWLINE] They are not working against each other, they are highlighting different issues. An equivalent would be to say that Rohingya people of Myanmar, the most persecuted people in the world, shouldn't protest for themselves because the anti racism movement is already well know. The same could be said for any small persecuted peoples, such as the Kurds. [NEWLINE] [NEWLINE] [STARTQ] My point is that they're fighting a common enemy and should stop getting waylaid by fighting each other. [ENDQ] [NEWLINE] Who or what is the common enemy? Surely it's our current manifestation of our society's culture? Is there only one way to fight an idea as big as that? During the anti apartheid movement, the biggest political movements involved were the African National Congress, the South African Communist Party, and the Pan African Congress, as well as many religious and social organisations. Should they amalgamate into one organisation because there could be in-fighting, or because one is more famous than another? Why cannot other organisations exist to fight a common cause. [NEWLINE] [NEWLINE] [STARTQ] If my point was that men's issues were too divisive and shouldn't be talked about, this would be a fair argument. But it isn't. It's that we absolutely should be talking about them-- as feminists and gender egalitarians. [ENDQ] [NEWLINE] Why are we not allowed to talk about them as feminists, men's rights activists, and gender egalitarians? Just because one movement is already known as a household name, should no others exist? [NEWLINE] [NEWLINE] [STARTQ] We can talk about men's issues in the context of sexism generally/feminism or under the umbrella of gender egalitarianism, without the us vs. them mentality that the MRM has inspired. [ENDQ] [NEWLINE] Surely this attitude is the only thing creating the us vs them mentality. "We can talk about these things, but they can't". If you were arguing for using only the gender egalitarianism movement I could understand, but why can a movement built to represent women in an oppressive society be better at talking about men's oppression under the same oppressive society than one built to represent men? [NEWLINE] [NEWLINE] [STARTQ] I think feminism is the best candidate for becoming a truly gender egalitarian movement, because it already has mainstream sociopolitical support whereas the MRM is dismissed as a joke and
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0381, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] My non-Bitcoin deposits are guaranteed by the government,<mask> they are 100% safe. [ENDQ] [NEWLINE] <mask><mask> safe is the value of the money in which your deposits are denominated? [NEWLINE] [NEWLINE] You just can't have your cake, and eat it too. You can't avoid all risks, you can just choose them. Gold, and Bitcoin, derive their value from the population's consensus of their value<mask> currency. Long term,<mask> gold or Bitcoin is actually in widespread use<mask> currency, this value will be stable. And I mean very stable,<mask> in, nearly identical prices over the course of decades. (That's<mask> it was with gold.) [NEWLINE] [NEWLINE] <mask> you have a government guaranteeing your deposit, you trade one risk for another - the currency can now be debased, and can lose most of its value. This is either a rare risk or a not-<mask> -rare risk, depending on the country you are in. It hasn't happened in the US,<mask>, unless you count the slow devaluation of the USD. Large and fast devaluations have happened in other places, many times. [NEWLINE] [NEWLINE] <mask>'s the interest you get on your FDIC insured deposit? Is it actually higher than inflation? Chances are, it's lower than inflation,<mask> you're trading a small risk of outright loss for a guaranteed small loss over time.<mask> this tradeoff can<mask> be achieved with private deposit insurance. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> does it work, exactly? I give the bank 10BTC to store, and then the bank does<mask>? They only have 10BTC,<mask> do they lend out 100BTC? [ENDQ] [NEWLINE] It works the same way<mask> with USD. Say you give the bank 100,000 USD or BTC or gold, whichever. Suppose the bank has a 5% fractional reserve (<mask><mask> this was the threshold commonly used by banks before the financial crisis,<mask> not even lower, not sure).<mask> they store 5,000 in reserve, and they lend 95,000. The person who gets the loan puts it in a bank. Suppose they put it in the same bank (doesn't make a difference, either this person does, or some other). The bank now has another 95,000 in deposits,<mask> they store 4,750, and loan 90,250. The process continues indefinitely until in equilibrium, the bank approaches a total of 1/(1-0.05) amount in deposits, e.g. 2 million, and 100,000 in reserves (the 5%).
Label encoding: <s> [STARTQ] My non-Bitcoin deposits are guaranteed by the government, so they are 100% safe. [ENDQ] [NEWLINE] But how safe is the value of the money in which your deposits are denominated? [NEWLINE] [NEWLINE] You just can't have your cake, and eat it too. You can't avoid all risks, you can just choose them. Gold, and Bitcoin, derive their value from the population's consensus of their value as currency. Long term, if gold or Bitcoin is actually in widespread use as currency, this value will be stable. And I mean very stable, as in, nearly identical prices over the course of decades. (That's how it was with gold.) [NEWLINE] [NEWLINE] When you have a government guaranteeing your deposit, you trade one risk for another - the currency can now be debased, and can lose most of its value. This is either a rare risk or a not- so -rare risk, depending on the country you are in. It hasn't happened in the US, yet, unless you count the slow devaluation of the USD. Large and fast devaluations have happened in other places, many times. [NEWLINE] [NEWLINE] What's the interest you get on your FDIC insured deposit? Is it actually higher than inflation? Chances are, it's lower than inflation, so you're trading a small risk of outright loss for a guaranteed small loss over time. But this tradeoff can also be achieved with private deposit insurance. [NEWLINE] [NEWLINE] [STARTQ] So how does it work, exactly? I give the bank 10BTC to store, and then the bank does what? They only have 10BTC, how do they lend out 100BTC? [ENDQ] [NEWLINE] It works the same way as with USD. Say you give the bank 100,000 USD or BTC or gold, whichever. Suppose the bank has a 5% fractional reserve ( I think this was the threshold commonly used by banks before the financial crisis, if not even lower, not sure). So they store 5,000 in reserve, and they lend 95,000. The person who gets the loan puts it in a bank. Suppose they put it in the same bank (doesn't make a difference, either this person does, or some other). The bank now has another 95,000 in deposits, so they store 4,750, and loan 90,250. The process continues indefinitely until in equilibrium, the bank approaches a total of 1/(1-0.05) amount in deposits, e.g. 2 million, and 100,000 in reserves (the 5%).
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0443, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1.  Drunk Driving [NEWLINE] [NEWLINE] **A good drunk driver is better than a bad sober driver. Women have worse reaction times than men, we shouldn't disallow them to drive, should we?** [NEWLINE] [NEWLINE] 2.  Firing a weapon in public. [NEWLINE] [NEWLINE] **Of course this has victims. First off, the bullet will come down,<mask> that's not even the immediate issue. Guns are loud, really loud, and could inflict damage to people's ears** [NEWLINE] [NEWLINE] 3.  Practicing medicine without a license. [NEWLINE] [NEWLINE] **Depends on<mask> the doctor makes this clear or not.<mask> he was not, this is deceit,<mask> you knew the risk, you knew the risk, just like with any surgery.** [NEWLINE] [NEWLINE] [NEWLINE] 4.  Selling defective products<mask> the defect is admitted two somewhere in 200 pages of a EULA. [NEWLINE] [NEWLINE] **Would<mask> be considered deceit. I doubt any courts in a voluntarist society would uphold that contract.** [NEWLINE] [NEWLINE] 5.  Placing a landmine on a city street. [NEWLINE] [NEWLINE] **Your actions will result in the possible deaths multiple people, of course this is a crime with victims.** [NEWLINE] [NEWLINE] 6.  Any sort of zoning violation. [NEWLINE] [NEWLINE] **Falls under property rights.<mask> you own the property, do<mask> you will within that property. ** [NEWLINE] [NEWLINE] 7.  Dumping toxic waste. [NEWLINE] [NEWLINE] **This one is more tricky. It depends on<mask> you dump it.<mask> it's not owned, you are not agressing against anyone,<mask> people would probably consider you a dick and avoid your products or service** [NEWLINE] [NEWLINE] 8.  Grave robbing. [NEWLINE] [NEWLINE] [NEWLINE] **Again, depends on<mask> the grave is, and<mask> you consider corpses private property. It's sort of like stealing a watch with sentimental value. It's not worth anything down there in the dirt, and you're not really hurting anyone<mask> you steal it,<mask> you're still a dick<mask> you do.** [NEWLINE] [NEWLINE] [NEWLINE] 9.  Tresspassing. [NEWLINE] [NEWLINE] **It depends on<mask> the people trespassing do<mask> they're trespassing. You're breaching someone's right to privacy, and their private property,<mask> in that sense you're absolutely breaking the law.** [NEWLINE] [NEWLINE] 10. Peeping toms. [NEWLINE] [NEWLINE] **I'm pretty sure the whole "reasonable expectancy of privacy" thing would apply in a voluntarist society, just like it does in our current society. That being said, I have a hard time justifying using
Label encoding: <s>1.  Drunk Driving [NEWLINE] [NEWLINE] **A good drunk driver is better than a bad sober driver. Women have worse reaction times than men, we shouldn't disallow them to drive, should we?** [NEWLINE] [NEWLINE] 2.  Firing a weapon in public. [NEWLINE] [NEWLINE] **Of course this has victims. First off, the bullet will come down, but that's not even the immediate issue. Guns are loud, really loud, and could inflict damage to people's ears** [NEWLINE] [NEWLINE] 3.  Practicing medicine without a license. [NEWLINE] [NEWLINE] **Depends on if the doctor makes this clear or not. If he was not, this is deceit, if you knew the risk, you knew the risk, just like with any surgery.** [NEWLINE] [NEWLINE] [NEWLINE] 4.  Selling defective products when the defect is admitted two somewhere in 200 pages of a EULA. [NEWLINE] [NEWLINE] **Would also be considered deceit. I doubt any courts in a voluntarist society would uphold that contract.** [NEWLINE] [NEWLINE] 5.  Placing a landmine on a city street. [NEWLINE] [NEWLINE] **Your actions will result in the possible deaths multiple people, of course this is a crime with victims.** [NEWLINE] [NEWLINE] 6.  Any sort of zoning violation. [NEWLINE] [NEWLINE] **Falls under property rights. If you own the property, do what you will within that property. ** [NEWLINE] [NEWLINE] 7.  Dumping toxic waste. [NEWLINE] [NEWLINE] **This one is more tricky. It depends on where you dump it. If it's not owned, you are not agressing against anyone, although people would probably consider you a dick and avoid your products or service** [NEWLINE] [NEWLINE] 8.  Grave robbing. [NEWLINE] [NEWLINE] [NEWLINE] **Again, depends on where the grave is, and if you consider corpses private property. It's sort of like stealing a watch with sentimental value. It's not worth anything down there in the dirt, and you're not really hurting anyone if you steal it, but you're still a dick if you do.** [NEWLINE] [NEWLINE] [NEWLINE] 9.  Tresspassing. [NEWLINE] [NEWLINE] **It depends on what the people trespassing do while they're trespassing. You're breaching someone's right to privacy, and their private property, so in that sense you're absolutely breaking the law.** [NEWLINE] [NEWLINE] 10. Peeping toms. [NEWLINE] [NEWLINE] **I'm pretty sure the whole "reasonable expectancy of privacy" thing would apply in a voluntarist society, just like it does in our current society. That being said, I have a hard time justifying using
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0373, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are couple of problems with this: [NEWLINE] [NEWLINE] * No Deviant Candidates: The vast majority of people deviate from social and moral norms in various ways. This system would tend select for incredibly bland people (<mask> it were democratic) - any deviation from social norms (hates sports), or moral norms (has open relationship with wife) would be  disadvantages over more "normal" candidates. This would result in incredibly non-representative candidates - legitimately "normal" people are a tiny minority in our society, and selecting for people who didn't understand deviations from the norm might be a disaster. [NEWLINE] [NEWLINE] * Compromising Privacy of Family and Friends: The privacy of the candidate's family and friends would be compromised - communications have (at least) two sides. Even<mask> it is reasonable to ask the candidate to renounce her privacy, is it fair to allow the candidate make that decision for everyone she has ever communicated with? Imagine your brother runs for office - under this system, every e-mail that you have ever sent him becomes public without your consent. I don't think it is reasonable to subject the family and friends of political figures to that - particularly<mask> they would get no choice in the matter. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on Family: This is related to the norms point above,<mask> candidates with a difficult family life (e.g. a racist uncle, abusive parent, or sister with mental illness) would probably be at a huge disadvantage in elections - they would be put in a position<mask> they would likely be judged based the errant communication of others. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on History: Like above,<mask> from the perspective of a candidate who went through a difficult time in her life or who was a huge jerk<mask> she was young. This hypothetical candidate would be at an extreme disadvantage to someone who grew up in an easy situation and never had anything bad happen to them. [NEWLINE] [NEWLINE] * Renouncing Future Privacy: Positions of power are not for life - imagine someone wanted to serve in parliament/congress for a term. All deviations (e.g. furry, brony, etc.) and family history (e.g. racist uncle, huge fights between family members, etc.) would be exposed forever. Even<mask> they actually managed to get elected, this would be in the public record every time they applied for a job or went out on a date - for the rest of their life. [NEWLINE] [NEWLINE] * Electing Deceitful/Cautious People:
Label encoding: <s>There are couple of problems with this: [NEWLINE] [NEWLINE] * No Deviant Candidates: The vast majority of people deviate from social and moral norms in various ways. This system would tend select for incredibly bland people ( if it were democratic) - any deviation from social norms (hates sports), or moral norms (has open relationship with wife) would be  disadvantages over more "normal" candidates. This would result in incredibly non-representative candidates - legitimately "normal" people are a tiny minority in our society, and selecting for people who didn't understand deviations from the norm might be a disaster. [NEWLINE] [NEWLINE] * Compromising Privacy of Family and Friends: The privacy of the candidate's family and friends would be compromised - communications have (at least) two sides. Even if it is reasonable to ask the candidate to renounce her privacy, is it fair to allow the candidate make that decision for everyone she has ever communicated with? Imagine your brother runs for office - under this system, every e-mail that you have ever sent him becomes public without your consent. I don't think it is reasonable to subject the family and friends of political figures to that - particularly when they would get no choice in the matter. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on Family: This is related to the norms point above, but candidates with a difficult family life (e.g. a racist uncle, abusive parent, or sister with mental illness) would probably be at a huge disadvantage in elections - they would be put in a position where they would likely be judged based the errant communication of others. [NEWLINE] [NEWLINE] * Disqualifying Candidates Based on History: Like above, but from the perspective of a candidate who went through a difficult time in her life or who was a huge jerk when she was young. This hypothetical candidate would be at an extreme disadvantage to someone who grew up in an easy situation and never had anything bad happen to them. [NEWLINE] [NEWLINE] * Renouncing Future Privacy: Positions of power are not for life - imagine someone wanted to serve in parliament/congress for a term. All deviations (e.g. furry, brony, etc.) and family history (e.g. racist uncle, huge fights between family members, etc.) would be exposed forever. Even if they actually managed to get elected, this would be in the public record every time they applied for a job or went out on a date - for the rest of their life. [NEWLINE] [NEWLINE] * Electing Deceitful/Cautious People:
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I came here with an argument,<mask> frankly, everyone else here has expressed my own views with far more finesse than I could.<mask> I'll just leave an anecdote here that might help illustrate<mask> others have pointed out. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] A good friend of mine in high school transitioned just after graduation.<mask> I first met him he identified<mask> a lesbian woman, and was always rocking the most stunning suits. Wingtips, bowler hats, pocket watches - his speciality was Victorian,<mask> he had some great ensembles from the 20s-50s<mask> well. For special occasions, he would wear eyeliner, sometimes lipstick. A few of us would have regular sleepovers and get into his wardrobe, which was a real pleasure to behold, and he would have no problem rocking heels, diamonds, etc.<mask> he tended towards masculinity, exploring femininity was clearly no issue. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] <mask> he told us that he was transitioning, I naively expected him to abandon the lipstick and eyeliner to embrace more masculine alternatives,<mask> he didn't give a fuck and continued to dress and express himself in exactly the same way he did<mask> a teenager. This annoyed some of my friends, particularly one who, upon seeing him walk into a girl's bathroom, turned angrily to me and snapped "He can't have both!" (He told me later that the girl's bathroom was actually more of a safety issue than a gender issue,<mask> he was less likely to get his ass kicked in the ladies'). I'm pleased to say that he does have "both",<mask> we're going to look at this<mask> a binary problem - he's had some surgery,<mask> not *all* surgery, and has reasons for this beyond pain and expense.<mask> he now happily takes his testosterone and identifies<mask> male, he expresses himself in some ways that could be called masculine, and many ways that would be called something else entirely. Most importantly, he is the most happy and comfortable that I have ever known him to be which, in the end, is<mask> matters. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] I bring this up to emphasize<mask>,<mask> there are undoubtedly many trans people who express their gender identity in very traditional ways - lots of makeup and an obsession with fashion, or muscle shirts and a love of sports - there is<mask> a large contingent of trans people who are comfortable somewhere in between. **This is a similar,<mask> not identical, state of affairs cis people**.<mask> you point out, feministic tendencies do not impinge upon your
Label encoding: <s>I came here with an argument, but frankly, everyone else here has expressed my own views with far more finesse than I could. So I'll just leave an anecdote here that might help illustrate what others have pointed out. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] A good friend of mine in high school transitioned just after graduation. When I first met him he identified as a lesbian woman, and was always rocking the most stunning suits. Wingtips, bowler hats, pocket watches - his speciality was Victorian, but he had some great ensembles from the 20s-50s as well. For special occasions, he would wear eyeliner, sometimes lipstick. A few of us would have regular sleepovers and get into his wardrobe, which was a real pleasure to behold, and he would have no problem rocking heels, diamonds, etc. Though he tended towards masculinity, exploring femininity was clearly no issue. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] When he told us that he was transitioning, I naively expected him to abandon the lipstick and eyeliner to embrace more masculine alternatives, but he didn't give a fuck and continued to dress and express himself in exactly the same way he did as a teenager. This annoyed some of my friends, particularly one who, upon seeing him walk into a girl's bathroom, turned angrily to me and snapped "He can't have both!" (He told me later that the girl's bathroom was actually more of a safety issue than a gender issue, as he was less likely to get his ass kicked in the ladies'). I'm pleased to say that he does have "both", if we're going to look at this as a binary problem - he's had some surgery, but not *all* surgery, and has reasons for this beyond pain and expense. While he now happily takes his testosterone and identifies as male, he expresses himself in some ways that could be called masculine, and many ways that would be called something else entirely. Most importantly, he is the most happy and comfortable that I have ever known him to be which, in the end, is what matters. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] I bring this up to emphasize how, while there are undoubtedly many trans people who express their gender identity in very traditional ways - lots of makeup and an obsession with fashion, or muscle shirts and a love of sports - there is also a large contingent of trans people who are comfortable somewhere in between. **This is a similar, though not identical, state of affairs cis people**. As you point out, feministic tendencies do not impinge upon your
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's a lot of assertions here that are just completely incorrect,<mask> I'm gonna break down some specific parts,<mask><mask><mask> there's a big flaw in almost all of your logic: [NEWLINE] [NEWLINE] **Just<mask> it is easier to start rapping doesn't mean being a good rapper takes any less skill.**<mask><mask> a lot of your viewpoints are biased in the sense that rapping may be easier to pick up just<mask> it is unlike singing or playing an instrument in that relatively mediocre rappers are not unlistenable in the same way that bad singers are. Yes, there are some bad rappers in the mainstream. There are<mask> bad songs in [plenty of other genres,]( [URL] )<mask> it makes no sense invalidating the entire genre<mask> of it. [NEWLINE] [NEWLINE] [STARTQ] Anyone is capable over saying words over a beat. [ENDQ] [NEWLINE] This seems to be pretty much your entire argument, essentially. [NEWLINE] [NEWLINE] Most good rappers put in effort to<mask> they say the words to make it compliment the beat and convey a message. I'm gonna be cliche and user Kendrick Lamar for most of my examples, just<mask> he's one of the most vocally diverse rappers. [NEWLINE] [NEWLINE] [In this song, he elongates his words and uses a raspy voice to make it sound like he's wailing,<mask> it's a heart-wrenching part of the album]( [URL] ) [NEWLINE] [NEWLINE] [In this song, he changes his pitch to sound more like an adolescent]( [URL] ) [NEWLINE] [NEWLINE] Those were some extreme examples,<mask> there's<mask> more subtle ones that are still very important [NEWLINE] [NEWLINE] [In this song, he uses a deeper pitch than usually to give off braggadccio/cocky vibes,<mask> he's meant to be freestyling in the car with his friends]( [URL] ) [NEWLINE] [NEWLINE] [<mask><mask>, this song uses a more relaxed tone<mask> he's'made it' at this point in the album]( [URL] ) [NEWLINE] [NEWLINE] All of this are important changes to his voice that are difficult to do in the first place, and even harder to do<mask> simultaneously staying on beat and not sounding too forced. [NEWLINE] [NEWLINE] [STARTQ] Some rappers don't even try to make sense, or be clever [ENDQ] [NEWLINE] Then that individual would be a bad rapper. A lot of other genres have untalented artists, too. [NEWLINE] [NEWLINE] [STARTQ] I can't believe rappers are allowed to win grammies and different music awards. Poets should be allowed to win grammies too then. [ENDQ] [NEWLINE] No,<mask> poetry
Label encoding: <s>There's a lot of assertions here that are just completely incorrect, so I'm gonna break down some specific parts, but I think there's a big flaw in almost all of your logic: [NEWLINE] [NEWLINE] **Just because it is easier to start rapping doesn't mean being a good rapper takes any less skill.** I think a lot of your viewpoints are biased in the sense that rapping may be easier to pick up just because it is unlike singing or playing an instrument in that relatively mediocre rappers are not unlistenable in the same way that bad singers are. Yes, there are some bad rappers in the mainstream. There are also bad songs in [plenty of other genres,]( [URL] ) but it makes no sense invalidating the entire genre because of it. [NEWLINE] [NEWLINE] [STARTQ] Anyone is capable over saying words over a beat. [ENDQ] [NEWLINE] This seems to be pretty much your entire argument, essentially. [NEWLINE] [NEWLINE] Most good rappers put in effort to how they say the words to make it compliment the beat and convey a message. I'm gonna be cliche and user Kendrick Lamar for most of my examples, just because he's one of the most vocally diverse rappers. [NEWLINE] [NEWLINE] [In this song, he elongates his words and uses a raspy voice to make it sound like he's wailing, since it's a heart-wrenching part of the album]( [URL] ) [NEWLINE] [NEWLINE] [In this song, he changes his pitch to sound more like an adolescent]( [URL] ) [NEWLINE] [NEWLINE] Those were some extreme examples, but there's also more subtle ones that are still very important [NEWLINE] [NEWLINE] [In this song, he uses a deeper pitch than usually to give off braggadccio/cocky vibes, since he's meant to be freestyling in the car with his friends]( [URL] ) [NEWLINE] [NEWLINE] [ In contrast, this song uses a more relaxed tone since he's'made it' at this point in the album]( [URL] ) [NEWLINE] [NEWLINE] All of this are important changes to his voice that are difficult to do in the first place, and even harder to do while simultaneously staying on beat and not sounding too forced. [NEWLINE] [NEWLINE] [STARTQ] Some rappers don't even try to make sense, or be clever [ENDQ] [NEWLINE] Then that individual would be a bad rapper. A lot of other genres have untalented artists, too. [NEWLINE] [NEWLINE] [STARTQ] I can't believe rappers are allowed to win grammies and different music awards. Poets should be allowed to win grammies too then. [ENDQ] [NEWLINE] No, because poetry
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's some truth to<mask> you're saying<mask> it's more nuanced than that. [NEWLINE] [STARTQ] You're all affected by the same economy [ENDQ] That's actually not true.  Sure, there's one overall national economy<mask> there is such diversity of business that different regions have completely different economies.  Take the example of Detroit and Houston.  Detroit is on the level of a 3rd world country<mask> being a powerhouse in the past due to manufacturing.  Houston is booming due to energy companies doing<mask> much business right now.  San Jose is doing great<mask> of technology companies.  Lexington Kentucky is...well I have no idea actually. [NEWLINE] [STARTQ] you all likely have a walmart near you [ENDQ] This is a very recent thing, and doesn't hold true for everyone.  I actually do have one near me,<mask> I was probably an adult before I ever set foot in a Walmart.  We like to bash them<mask> they have an aggressive expansion policy and they're one of the few stores that are mostly present nationwide.  The same applies to McDonalds.  Those exist in other countries too<mask>.  In any case, monolithic retail is something that has happened mostly in the past few decades.  I don't know that I had even heard of Walmart until the 90's. [NEWLINE] [STARTQ] you all are affected by the same federal laws [ENDQ] This is true,<mask><mask> the federal government of the U.S. isn't<mask> strong<mask> European federal governments.  Look at<mask> the states in the U.S. are doing completely different things on<mask> should be national issues like gay marriage, marijuana, etc. [NEWLINE] [STARTQ] All the cities are governed top down (federal, state, regional, county, local). [ENDQ] <mask> not<mask> sexy in the media, state and local governments are<mask> most of the laws are passed.  I wouldn't say that cities are governed from the top down, it's mostly from the bottom up with exceptions in key areas (and<mask> those key areas are may vary from one state to the next.) [NEWLINE] [STARTQ] Your city likely has some degree of urban sprawl. [ENDQ] Not everyone lives in cities.  Most redditors from the U.S. are probably going to be found in cities,<mask> the majority of the U.S. is not urban.  The metropolitan area I live in, for example, probably has about twice<mask> many people<mask> the entire state of Oklahoma or Kansas.  Urban sprawl is mainly a problem<mask> the U.S. has shifted too much to an
Label encoding: <s>There's some truth to what you're saying but it's more nuanced than that. [NEWLINE] [STARTQ] You're all affected by the same economy [ENDQ] That's actually not true.  Sure, there's one overall national economy but there is such diversity of business that different regions have completely different economies.  Take the example of Detroit and Houston.  Detroit is on the level of a 3rd world country despite being a powerhouse in the past due to manufacturing.  Houston is booming due to energy companies doing so much business right now.  San Jose is doing great because of technology companies.  Lexington Kentucky is...well I have no idea actually. [NEWLINE] [STARTQ] you all likely have a walmart near you [ENDQ] This is a very recent thing, and doesn't hold true for everyone.  I actually do have one near me, but I was probably an adult before I ever set foot in a Walmart.  We like to bash them because they have an aggressive expansion policy and they're one of the few stores that are mostly present nationwide.  The same applies to McDonalds.  Those exist in other countries too though.  In any case, monolithic retail is something that has happened mostly in the past few decades.  I don't know that I had even heard of Walmart until the 90's. [NEWLINE] [STARTQ] you all are affected by the same federal laws [ENDQ] This is true, but also the federal government of the U.S. isn't as strong as European federal governments.  Look at how the states in the U.S. are doing completely different things on what should be national issues like gay marriage, marijuana, etc. [NEWLINE] [STARTQ] All the cities are governed top down (federal, state, regional, county, local). [ENDQ] While not as sexy in the media, state and local governments are where most of the laws are passed.  I wouldn't say that cities are governed from the top down, it's mostly from the bottom up with exceptions in key areas (and what those key areas are may vary from one state to the next.) [NEWLINE] [STARTQ] Your city likely has some degree of urban sprawl. [ENDQ] Not everyone lives in cities.  Most redditors from the U.S. are probably going to be found in cities, but the majority of the U.S. is not urban.  The metropolitan area I live in, for example, probably has about twice as many people as the entire state of Oklahoma or Kansas.  Urban sprawl is mainly a problem because the U.S. has shifted too much to an
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0502, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0557, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've always thought of science<mask> the study of something "that's already there," which is<mask> it is possible to do experiments on it or research it<mask> an observer.  In the physical sciences, you study something that's already there (physics, biology) and do experiments on things "that already exist" to further your knowledge of it.  One commenter brought up softer social sciences, in which case I'd say quantitative or physical experimentation is replaced by qualitative research<mask> an observer, and specifically the research of *something that's already there* in order to discover knowledge that is currently unknown. [NEWLINE] [NEWLINE] My definition of engineering agrees with yours, which is the creation of something, anything really.  It's the study of<mask> to create something and<mask> to create those things better.  Science applied, I suppose.  Research in engineering is like studying and working on the creation of *something that is<mask> to exist*, which is different from my definition of research in science in the paragraph above. [NEWLINE] [NEWLINE] <mask> CS doesn't depend on the computer hardware,<mask> you say.  You proposed names like "______ Mathematics"<mask> I feel that CS isn't specific to its mathematics, just like biology isn't specific to just statistics<mask><mask> there is biostatistics (i.e. there's a lot more to CS than just mathematics, and there's a lot more to biology than just its statistics).  CS focuses on something, which<mask><mask> is information and computation.  I propose something like "Information Engineering" and "Information Science" or "Computational Engineering and "Computational Science".  "Information or Computational Science" is the theoretical stuff that goes on in CS.  "Information or Computational Engineering" is studying and developing better techniques to manage and manipulate data - kind of the CS stuff that a layman would imagine is CS.  Only<mask> you get in to the hardware (or software) should the word "Computer" be used in the term. <mask> you can have "Computer Engineering"<mask>, then, "Computer Science" gets awkward.  It's like "bicycle science" or "telescope science", which is really either engineering or<mask> you're actually talking about the science behind it, it'd be physics (mechanics for bicycles and optics and EM waves for telescopes). [NEWLINE] [NEWLINE] Back to my point about using the word "mathematics" - all physical sciences + economics uses mathematics,<mask> the field is more than that.  Which is<mask> you
Label encoding: <s>I've always thought of science as the study of something "that's already there," which is why it is possible to do experiments on it or research it as an observer.  In the physical sciences, you study something that's already there (physics, biology) and do experiments on things "that already exist" to further your knowledge of it.  One commenter brought up softer social sciences, in which case I'd say quantitative or physical experimentation is replaced by qualitative research as an observer, and specifically the research of *something that's already there* in order to discover knowledge that is currently unknown. [NEWLINE] [NEWLINE] My definition of engineering agrees with yours, which is the creation of something, anything really.  It's the study of how to create something and how to create those things better.  Science applied, I suppose.  Research in engineering is like studying and working on the creation of *something that is yet to exist*, which is different from my definition of research in science in the paragraph above. [NEWLINE] [NEWLINE] So CS doesn't depend on the computer hardware, as you say.  You proposed names like "______ Mathematics" but I feel that CS isn't specific to its mathematics, just like biology isn't specific to just statistics even though there is biostatistics (i.e. there's a lot more to CS than just mathematics, and there's a lot more to biology than just its statistics).  CS focuses on something, which I think is information and computation.  I propose something like "Information Engineering" and "Information Science" or "Computational Engineering and "Computational Science".  "Information or Computational Science" is the theoretical stuff that goes on in CS.  "Information or Computational Engineering" is studying and developing better techniques to manage and manipulate data - kind of the CS stuff that a layman would imagine is CS.  Only when you get in to the hardware (or software) should the word "Computer" be used in the term.  So you can have "Computer Engineering" but, then, "Computer Science" gets awkward.  It's like "bicycle science" or "telescope science", which is really either engineering or if you're actually talking about the science behind it, it'd be physics (mechanics for bicycles and optics and EM waves for telescopes). [NEWLINE] [NEWLINE] Back to my point about using the word "mathematics" - all physical sciences + economics uses mathematics, but the field is more than that.  Which is why you
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0455, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0570, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0483, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I believe "Thou shalt not kill" is a fundamental rule in life,<mask><mask> I'm not that religious i just believe killing another person is the most heinous act a person can commit,<mask> soldiers go out to other countries and are paid to take lives<mask> they return<mask> heroes. Now i know there are other jobs to do in the army, not just killing<mask> the key part for most of the soldiers is to kill. [ENDQ] [NEWLINE] The concept of "Thou shalt not kill" is<mask>... misinterpreted.  It's technically, "Thou shalt not murder",<mask> murder is an unlawful killing. [NEWLINE] [NEWLINE] War... is lawful by nature.  The reasons for war are always just in the two or more parties involved.  For the US, sending troops overseas is done to kill people who intend to do harm to us in our homes... for those they're killing, they likewise respond that they are fighting to repel the US from their lands or that we are being punished for another transgression (stationing troops near their holy lands or supporting Israel). [NEWLINE] [NEWLINE] Soldiers, by definition, have no choice to kill.  For them, its typically a binary solution: Kill or Be Killed. <mask> we could win wars with pool noodles alone, we would,<mask> we can't.  To expand on the lack of choice: soldiers are<mask> tools of the State.  They cannot choose<mask> they are deployed and<mask>.  That is the nature of a soldier. [NEWLINE] [NEWLINE] [STARTQ] <mask> are they any different to murderers who kill over jealousy or greed?<mask> are they shown<mask> much respect for killing in 'the name of their country'? [ENDQ] [NEWLINE] Simply put:<mask> they are not acting out of those petty needs.  Soldiers rarely want to be in some shithole far from home<mask> their wife is probably sucking their neighbors dick.  They typically signed up<mask> they were either: idealistic, in poor financial situations, or couldn't do anything else.  For them, a few years in uniform with housing and food is better than the alternatives of poverty they faced at home. [NEWLINE] [NEWLINE] Soldiers<mask> kill in order to avoid being killed or to protect their countrymen. <mask> they find out terrorists have a chemical weapon, they will try and stop them before that weapon can be delivered.  The terrorists will fight back, and people will die.  There is no other choice. [NEWLINE] [NEWLINE] [STARTQ] It is<mask><mask> that taking lives, for whatever reason, is wrong. Soldiers don't deserve respect
Label encoding: <s> [STARTQ] I believe "Thou shalt not kill" is a fundamental rule in life, even though I'm not that religious i just believe killing another person is the most heinous act a person can commit, yet soldiers go out to other countries and are paid to take lives but they return as heroes. Now i know there are other jobs to do in the army, not just killing but the key part for most of the soldiers is to kill. [ENDQ] [NEWLINE] The concept of "Thou shalt not kill" is also... misinterpreted.  It's technically, "Thou shalt not murder", where murder is an unlawful killing. [NEWLINE] [NEWLINE] War... is lawful by nature.  The reasons for war are always just in the two or more parties involved.  For the US, sending troops overseas is done to kill people who intend to do harm to us in our homes... for those they're killing, they likewise respond that they are fighting to repel the US from their lands or that we are being punished for another transgression (stationing troops near their holy lands or supporting Israel). [NEWLINE] [NEWLINE] Soldiers, by definition, have no choice to kill.  For them, its typically a binary solution: Kill or Be Killed.  If we could win wars with pool noodles alone, we would, but we can't.  To expand on the lack of choice: soldiers are also tools of the State.  They cannot choose where they are deployed and why.  That is the nature of a soldier. [NEWLINE] [NEWLINE] [STARTQ] How are they any different to murderers who kill over jealousy or greed? Why are they shown so much respect for killing in 'the name of their country'? [ENDQ] [NEWLINE] Simply put: because they are not acting out of those petty needs.  Soldiers rarely want to be in some shithole far from home where their wife is probably sucking their neighbors dick.  They typically signed up because they were either: idealistic, in poor financial situations, or couldn't do anything else.  For them, a few years in uniform with housing and food is better than the alternatives of poverty they faced at home. [NEWLINE] [NEWLINE] Soldiers also kill in order to avoid being killed or to protect their countrymen.  If they find out terrorists have a chemical weapon, they will try and stop them before that weapon can be delivered.  The terrorists will fight back, and people will die.  There is no other choice. [NEWLINE] [NEWLINE] [STARTQ] It is my opinion that taking lives, for whatever reason, is wrong. Soldiers don't deserve respect
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0003, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not sure this will be helpful,<mask> here is some information that might be helpful in considering your position on antidepressants.  For background, I have experience both with research and treatment using Cognitive Behavioral Therapy for depression, I've worked with numerous depressed clients, and I have known people in a personal context who have taken antidepressants for various periods of time (anywhere from a year to several years). [NEWLINE] [NEWLINE] First,<mask> you are considering psychotherapy, there is good evidence to suggest that for depression specifically, combined psychotherapy and antidepressants give a boost in effect than either separately (this isn't the case with all disorders,<mask> it appears to be the case with depression). <mask> you need me to look up the specific citations, I can,<mask> there are four studies off the top of my head that have seen this effect, separately headed up by Blackburn, Murphy, Hollon, and Keller, all of which compared the improvements of clients who had psychotherapy only, antidepressants only, or combined treatment. <mask> an aside, hopefully for psychotherapy you consider an evidence-based treatment like CBT, Interpersonal Psychotherapy, etc. [NEWLINE] [NEWLINE] Second,<mask> depression feels like reality, in CBT, one major goal of the clinician is to help the client view situations more accurately and realistically (NOT "look on the bright side of things").  This is helpful to clients<mask> they are NOT viewing many situations realistically or accurately--their viewpoints are typically skewed by their focus on the negative side of things or a difficulty identifying/weighing positive or neutral elements or interpretations of a situation,<mask> helping their thinking become more accurate, realistic, and balanced results in their thinking becoming more positive. [NEWLINE] [NEWLINE] Third,<mask> you have been depressed for a long time, it's possible you have lost some perspective about<mask> you used to think and react to things<mask> you weren't depressed. <mask> clients have not been depressed for<mask> long and can remember<mask> it was like to not be depressed, they are better able to see<mask> differently they are thinking and reacting to things than they used to. [NEWLINE] [NEWLINE] Fourth, and finally,<mask> people who take antidepressants can react differently to them, many folks I know personally and have worked with professionally have described taking antidepressants in the following ways: "I feel like myself again," "I feel less overwhelmed by things, more optimistic like I used to," "Sometimes I wonder<mask> I need the medication<mask> I feel<mask> normal...<mask> I know the medications are helping me achieve that." (These are all basically
Label encoding: <s>I'm not sure this will be helpful, but here is some information that might be helpful in considering your position on antidepressants.  For background, I have experience both with research and treatment using Cognitive Behavioral Therapy for depression, I've worked with numerous depressed clients, and I have known people in a personal context who have taken antidepressants for various periods of time (anywhere from a year to several years). [NEWLINE] [NEWLINE] First, since you are considering psychotherapy, there is good evidence to suggest that for depression specifically, combined psychotherapy and antidepressants give a boost in effect than either separately (this isn't the case with all disorders, but it appears to be the case with depression).  If you need me to look up the specific citations, I can, but there are four studies off the top of my head that have seen this effect, separately headed up by Blackburn, Murphy, Hollon, and Keller, all of which compared the improvements of clients who had psychotherapy only, antidepressants only, or combined treatment.  As an aside, hopefully for psychotherapy you consider an evidence-based treatment like CBT, Interpersonal Psychotherapy, etc. [NEWLINE] [NEWLINE] Second, while depression feels like reality, in CBT, one major goal of the clinician is to help the client view situations more accurately and realistically (NOT "look on the bright side of things").  This is helpful to clients because they are NOT viewing many situations realistically or accurately--their viewpoints are typically skewed by their focus on the negative side of things or a difficulty identifying/weighing positive or neutral elements or interpretations of a situation, so helping their thinking become more accurate, realistic, and balanced results in their thinking becoming more positive. [NEWLINE] [NEWLINE] Third, because you have been depressed for a long time, it's possible you have lost some perspective about how you used to think and react to things when you weren't depressed.  When clients have not been depressed for as long and can remember what it was like to not be depressed, they are better able to see how differently they are thinking and reacting to things than they used to. [NEWLINE] [NEWLINE] Fourth, and finally, while people who take antidepressants can react differently to them, many folks I know personally and have worked with professionally have described taking antidepressants in the following ways: "I feel like myself again," "I feel less overwhelmed by things, more optimistic like I used to," "Sometimes I wonder if I need the medication because I feel so normal... but I know the medications are helping me achieve that." (These are all basically
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0377, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0441, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0400, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay I keep reading a ton of miss information. [NEWLINE] First off there's a pretty big difference between a service animal and an emotional support animal. [NEWLINE] [NEWLINE] Service animal is (typically) a dog that is trained to perform specific tasks, alert, guide, and generally assist a person with a disability. They are given the right to accompany their handler to any PUBLIC building<mask> well<mask> living in housing that has a no pet policy. [NEWLINE] There's a huge disagreement about the training these animals should receive and identification/registration they should have. 1) these animals do not have to be registered through any data base. None of the data bases out there that offer any kind of formal registration are connected with the ADA or any other government agency. These companies sell badges and "registration" that has no legal standing. 2) there are tests such<mask> "access tests" and such that can be performed and will help with defending oneself against being sued for having a "fake" service dog,<mask>, these are not legally necessary. 3) service dogs can be trained by the owner, a dog trainer, or an organization that trains and places service dogs. I've met quite a few owner trained dogs who are incredibly well trained and do a great job<mask> valid service dogs. [NEWLINE] [NEWLINE] Emotional support animals are dogs, cats, lizards, ect. that help allow a person with a disability to fully use and enjoy their place of living. These animals have a much more limited job. They ARE allowed under the Fair Housing Act to live with a person with a disability even in a "no pets" building. 1) they are NOT allowed in public buildings and are restricted to the home and other animal friendly locations.2) they do not have to have special training or perform specific tasks. Their job is to help comfort and support the person with a disability. 3) you must have a doctors note basically prescribing the dog<mask> a medical necessity for a disability in order to have an emotional support animal. [NEWLINE] [NEWLINE] For both of these it is illegal to deny a person with a disability to have either a service animal or an ESA. To deny a person with a disability of this support system is considered discrimination. [NEWLINE] The animals must not be a hazard or threat to anyone around, it is the handler/owners job to keep the animals under control. [NEWLINE] No deposit specifically for the animal is allowed to be collected<mask><mask> there is damage reimbursement can be sought. [NEWLINE] [NEWLINE] A service dog you find in public buildings must always be trained for specific tasks to help
Label encoding: <s>Okay I keep reading a ton of miss information. [NEWLINE] First off there's a pretty big difference between a service animal and an emotional support animal. [NEWLINE] [NEWLINE] Service animal is (typically) a dog that is trained to perform specific tasks, alert, guide, and generally assist a person with a disability. They are given the right to accompany their handler to any PUBLIC building as well as living in housing that has a no pet policy. [NEWLINE] There's a huge disagreement about the training these animals should receive and identification/registration they should have. 1) these animals do not have to be registered through any data base. None of the data bases out there that offer any kind of formal registration are connected with the ADA or any other government agency. These companies sell badges and "registration" that has no legal standing. 2) there are tests such as "access tests" and such that can be performed and will help with defending oneself against being sued for having a "fake" service dog, however, these are not legally necessary. 3) service dogs can be trained by the owner, a dog trainer, or an organization that trains and places service dogs. I've met quite a few owner trained dogs who are incredibly well trained and do a great job as valid service dogs. [NEWLINE] [NEWLINE] Emotional support animals are dogs, cats, lizards, ect. that help allow a person with a disability to fully use and enjoy their place of living. These animals have a much more limited job. They ARE allowed under the Fair Housing Act to live with a person with a disability even in a "no pets" building. 1) they are NOT allowed in public buildings and are restricted to the home and other animal friendly locations.2) they do not have to have special training or perform specific tasks. Their job is to help comfort and support the person with a disability. 3) you must have a doctors note basically prescribing the dog as a medical necessity for a disability in order to have an emotional support animal. [NEWLINE] [NEWLINE] For both of these it is illegal to deny a person with a disability to have either a service animal or an ESA. To deny a person with a disability of this support system is considered discrimination. [NEWLINE] The animals must not be a hazard or threat to anyone around, it is the handler/owners job to keep the animals under control. [NEWLINE] No deposit specifically for the animal is allowed to be collected however if there is damage reimbursement can be sought. [NEWLINE] [NEWLINE] A service dog you find in public buildings must always be trained for specific tasks to help
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you could<mask><mask> laws put in place by the constitution are now being used in ways that were never intended by the founding fathers. [NEWLINE] [NEWLINE] Take for example the corporations in 1776 vs corporations now.  In 1776, corporations typically only operated in the state they were created.  This was largely due to technology at the time.<mask>, they had the foresight that someone may want to do business across state lines.  In those particular cases, the federal government recognized that there needed to be an overseer to determine<mask> that corporation was allowed to do, particularly<mask> laws in each state differ.  For that reason, they created the commerce clause. [NEWLINE] [NEWLINE] Is the commerce clause inherently bad?  Not particularly,<mask><mask> then several very large controversial laws have been passed under the guise of the commerce clause.  In that same light, several of those laws aren't inherently bad or even debated anymore.  For example, the anti-segregation laws were all passed under this clause and has<mask> been generally agreed on by everyone. <mask>,<mask> that line gets fuzzier is<mask> you run into problems.  Obamacare was passed under the commerce clause<mask> (more or less) citizens were afforded health care in some states and not in others.  Is that related to interstate commerce?  Well, healthcare is inter-state,<mask> in that regard, possibly,<mask> does that afford the right to the federal government to use this clause to do<mask>?  Obama thinks<mask>, and<mask> of last week<mask> does the Supreme court. [NEWLINE] [NEWLINE] <mask> about drugs?  Marijuana is legal in several states that have nothing to do with inter-state commerce.  Technically those corporations are breaking federal law<mask> of the war on drugs (<mask> passed<mask> of the commerce clause).  Those particular companies don't do inter-state business,<mask><mask> they interact with corporations that do interstate business, it makes operating very difficult (i.e. basically none of those companies can have bank accounts<mask> federal law prohibits banks from accepting known drug money). <mask> now,<mask><mask> the government isn't going after those businesses, they still are having a terribly difficult time with finances (especially with taxes)<mask><mask><mask> law. [NEWLINE] [NEWLINE] Now with all this in mind, I'm not here to say those are bad things. I actually support them. My point in all this is that the interpretation of the commerce clause is now<mask> limits the federal government's power.  And all of this came out of a tiny portion of the constitution that
Label encoding: <s>I think you could argue that laws put in place by the constitution are now being used in ways that were never intended by the founding fathers. [NEWLINE] [NEWLINE] Take for example the corporations in 1776 vs corporations now.  In 1776, corporations typically only operated in the state they were created.  This was largely due to technology at the time. However, they had the foresight that someone may want to do business across state lines.  In those particular cases, the federal government recognized that there needed to be an overseer to determine what that corporation was allowed to do, particularly when laws in each state differ.  For that reason, they created the commerce clause. [NEWLINE] [NEWLINE] Is the commerce clause inherently bad?  Not particularly, but since then several very large controversial laws have been passed under the guise of the commerce clause.  In that same light, several of those laws aren't inherently bad or even debated anymore.  For example, the anti-segregation laws were all passed under this clause and has since been generally agreed on by everyone.  However, when that line gets fuzzier is when you run into problems.  Obamacare was passed under the commerce clause because (more or less) citizens were afforded health care in some states and not in others.  Is that related to interstate commerce?  Well, healthcare is inter-state, so in that regard, possibly, but does that afford the right to the federal government to use this clause to do so?  Obama thinks so, and as of last week so does the Supreme court. [NEWLINE] [NEWLINE] What about drugs?  Marijuana is legal in several states that have nothing to do with inter-state commerce.  Technically those corporations are breaking federal law because of the war on drugs ( also passed because of the commerce clause).  Those particular companies don't do inter-state business, but because they interact with corporations that do interstate business, it makes operating very difficult (i.e. basically none of those companies can have bank accounts because federal law prohibits banks from accepting known drug money).  So now, even though the government isn't going after those businesses, they still are having a terribly difficult time with finances (especially with taxes) because of this law. [NEWLINE] [NEWLINE] Now with all this in mind, I'm not here to say those are bad things. I actually support them. My point in all this is that the interpretation of the commerce clause is now what limits the federal government's power.  And all of this came out of a tiny portion of the constitution that
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0718, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0403, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0593, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Islam was founded with war. Christians converted the Roman Empire from the bottom up, missionaries converted the British Isles and the Germanic tribes. The Scandinavians, Rus, and some of the Germanic tribes were forcefully converted,<mask> the leaders were typically converted first, unlike the Islamic expansion were it was an invasion and islamification. [ENDQ] [NEWLINE] I would say rather that Islam simply turned to violence to spread its influence much earlier than Christianity did,<mask> it was during the lifetime of its founder. Christians gained lots of converts in coastal communities,<mask> the religion didn't really take off until Constantine made it the official religion of Rome, backing the faith with state power for the first time. After that, the stories of the two religions really aren't that different with respect to spreading by violence. The Germanic tribes were mainly converted by force, especially in the East. The Teutonic Knights for example fought local pagans in destructive wars that killed a huge portion of the native populations. And the Christianity in the British Isles<mask> benefitted from state support. not to say there weren't many missionaries,<mask><mask> they tend to go hand in hand. (much later during the spanish colonization of the americas, the church was described<mask> 'the cross in the shadow of the sword' or something like that). [NEWLINE] [NEWLINE] <mask><mask>, one interesting feature of the Islamic expansion is that the original Muslims were often rather reluctant to allow large numbers of people to convert,<mask> non-muslims were discriminated against in the form of higher taxes. and being exclusively arabs, there was a common view among early muslims that islam was a religion for the arabs, in the way that Judaism was for Jews. and islam followed the same pattern you mention in which local elites convert and the population gradually follows.<mask><mask> muslims were the minority in the islamic empire for some years after the initial invasions. christians predominated in the levant and egypt, and Zoroastrians and others in Persia. [NEWLINE] [NEWLINE] 'this contradicts everything I have ever read. Shariah is based off of the Quran and other scriptures (Primarily the Hadith) and has been followed<mask> Muhamed.' [NEWLINE] yeah, it definitely is based on the scholars' readings of the quran, and there are a lot of specific laws in it. i'm just saying that<mask> a formal, elaborated body of law, shariah didn't exist until many years after muhammads death. Neither did the quran,<mask>
Label encoding: <s> [STARTQ] Islam was founded with war. Christians converted the Roman Empire from the bottom up, missionaries converted the British Isles and the Germanic tribes. The Scandinavians, Rus, and some of the Germanic tribes were forcefully converted, but the leaders were typically converted first, unlike the Islamic expansion were it was an invasion and islamification. [ENDQ] [NEWLINE] I would say rather that Islam simply turned to violence to spread its influence much earlier than Christianity did, since it was during the lifetime of its founder. Christians gained lots of converts in coastal communities, but the religion didn't really take off until Constantine made it the official religion of Rome, backing the faith with state power for the first time. After that, the stories of the two religions really aren't that different with respect to spreading by violence. The Germanic tribes were mainly converted by force, especially in the East. The Teutonic Knights for example fought local pagans in destructive wars that killed a huge portion of the native populations. And the Christianity in the British Isles also benefitted from state support. not to say there weren't many missionaries, though but they tend to go hand in hand. (much later during the spanish colonization of the americas, the church was described as 'the cross in the shadow of the sword' or something like that). [NEWLINE] [NEWLINE] In fact, one interesting feature of the Islamic expansion is that the original Muslims were often rather reluctant to allow large numbers of people to convert, because non-muslims were discriminated against in the form of higher taxes. and being exclusively arabs, there was a common view among early muslims that islam was a religion for the arabs, in the way that Judaism was for Jews. and islam followed the same pattern you mention in which local elites convert and the population gradually follows. In fact muslims were the minority in the islamic empire for some years after the initial invasions. christians predominated in the levant and egypt, and Zoroastrians and others in Persia. [NEWLINE] [NEWLINE] 'this contradicts everything I have ever read. Shariah is based off of the Quran and other scriptures (Primarily the Hadith) and has been followed since Muhamed.' [NEWLINE] yeah, it definitely is based on the scholars' readings of the quran, and there are a lot of specific laws in it. i'm just saying that as a formal, elaborated body of law, shariah didn't exist until many years after muhammads death. Neither did the quran, in
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0529, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0462, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0394, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0453, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0602, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The profit margin for betting on a 'top' team is extremely small. I'm not sure exactly<mask> it works,<mask> I'm positive that there's rules in place to stop you from getting free money by simply betting on all the top teams. [NEWLINE] [NEWLINE] In essence the more risk you take on, the more profit you make from the bet. Which is somewhat similar to<mask> a lot of financial vehicles work. One example is the 2004 deregulation of CDOs and CDS. [NEWLINE] [NEWLINE] CDO is basically a casserole of different level risk mortgages,<mask> you buy a share of the profits of mortgage payments from 3 levels of homeowners. You get a little low risk mortgages, more medium risk pieces, and an amount of garbage mortgages,<mask> it's estimated that the owner is highly likely to default on their payments. The risk is further heightened,<mask><mask> someone of the three groups makes a payment, it goes towards covering the lowest risk mortgages first, then the second level, then the crap level. Needless to say those were issued by Goldman Sachs, under-priced at a nice profit margin. Those became known<mask> the Abbacus deal. [NEWLINE] [NEWLINE] CDS or credit default swaps are a sort of insurrance one bank may buy from another bank to offset the risk of bank A's mortgages defaulting. The stipulation is that bank A pays bank B a monthly premium, and receives a payment back<mask> bank A's mortgage payers default. Before 04 it was only possible to buy that kind of insurance<mask> you actually owned the insured stock. In 04 Alan Greenspan (Randian believer in the concept that markets should be allowed to govern themselves) made it<mask> you can buy insurance on stocks you don't even own. [NEWLINE] [NEWLINE] Which resulted in Goldman selling Abbacus at a profit,<mask> insuring it,<mask><mask> they didn't own it after it's sold. Imagine being sold a car,<mask> the dealer receives a large insurance payment<mask> it faults and kills you. Or your doctor, taking out a life insurance on you... (<mask> Morbidity and Mortality insurances are common, those are more like a last resort at securing the doctor's life for a few years, after he's already lost his license to practice) [NEWLINE] [NEWLINE] And we all know<mask> that ended. Goldman got paid once for selling the shitty stocks, twice for financial consulting that lead to the sales of it, and a third time,<mask> they defaulted.<mask> of CDS,<mask> the other banks received the bailout, they owed a
Label encoding: <s>The profit margin for betting on a 'top' team is extremely small. I'm not sure exactly how it works, but I'm positive that there's rules in place to stop you from getting free money by simply betting on all the top teams. [NEWLINE] [NEWLINE] In essence the more risk you take on, the more profit you make from the bet. Which is somewhat similar to how a lot of financial vehicles work. One example is the 2004 deregulation of CDOs and CDS. [NEWLINE] [NEWLINE] CDO is basically a casserole of different level risk mortgages, where you buy a share of the profits of mortgage payments from 3 levels of homeowners. You get a little low risk mortgages, more medium risk pieces, and an amount of garbage mortgages, where it's estimated that the owner is highly likely to default on their payments. The risk is further heightened, because when someone of the three groups makes a payment, it goes towards covering the lowest risk mortgages first, then the second level, then the crap level. Needless to say those were issued by Goldman Sachs, under-priced at a nice profit margin. Those became known as the Abbacus deal. [NEWLINE] [NEWLINE] CDS or credit default swaps are a sort of insurrance one bank may buy from another bank to offset the risk of bank A's mortgages defaulting. The stipulation is that bank A pays bank B a monthly premium, and receives a payment back if bank A's mortgage payers default. Before 04 it was only possible to buy that kind of insurance if you actually owned the insured stock. In 04 Alan Greenspan (Randian believer in the concept that markets should be allowed to govern themselves) made it so you can buy insurance on stocks you don't even own. [NEWLINE] [NEWLINE] Which resulted in Goldman selling Abbacus at a profit, while insuring it, even though they didn't own it after it's sold. Imagine being sold a car, where the dealer receives a large insurance payment if it faults and kills you. Or your doctor, taking out a life insurance on you... ( although Morbidity and Mortality insurances are common, those are more like a last resort at securing the doctor's life for a few years, after he's already lost his license to practice) [NEWLINE] [NEWLINE] And we all know how that ended. Goldman got paid once for selling the shitty stocks, twice for financial consulting that lead to the sales of it, and a third time, when they defaulted. Because of CDS, when the other banks received the bailout, they owed a
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I see a number of issues with you argument, I'll try to list them here: [NEWLINE] [NEWLINE] [STARTQ] <mask> public transport such<mask> railways are expanded and given priority over road vehicles in infrastructure, then we can achieve the goals that the automotive industry is trying to achieve very quickly. [ENDQ] [NEWLINE] Public transport is already given priority over automobiles - think bus lanes in most metropolitan areas, railway crossings, tram lines, city center tax for automobiles, etc. [NEWLINE] [NEWLINE] <mask> for R&amp;D, train companies will invest in trains and auto companies will invest in cars. [NEWLINE] [NEWLINE] [STARTQ] We should divert the resources that we spend on making smarter cars (finances and human capital) to making smarter, faster, safer, public transport and expand the rail network.<mask> we just expand the rail network then people will automatically shift from driving a car to riding on a train<mask> their preferred choice of daily commute. [ENDQ] [NEWLINE] Unfortunately you cannot dictate<mask> a private business spends it's money. The need for automobiles will always exist,<mask> public transportation can not cover the whole road network.<mask> this need exists, it is logical someone will invest into bettering the technology. [NEWLINE] [NEWLINE] [STARTQ] Environmental impact- A vast majority of cars are running on fuels that emit pollutants in the environment. [ENDQ] [NEWLINE] For now. The technology for electric vehicles is already here and rapid progress is made in the field of energy storage.<mask>, public transportation does not mean pollution free. A train carrying 20 passengers on a not-<mask> -popular route can pollute more than 4 cars (depending on the energy source). [NEWLINE] [NEWLINE] [STARTQ] Safety- Road injury was one of the top 10 causes of death in the world.<mask><mask> WHO, road injury took lives of 1.3 Million people in the last decade. [ENDQ] [NEWLINE] This is actually a reason for safer cars/self driving cars than for using public transportation. [NEWLINE] [NEWLINE] [STARTQ] Maintenance – a car requires frequent upkeep and maintenance of its parts and components. [ENDQ] [NEWLINE] <mask> do trains, trams, buses, etc. [NEWLINE] [NEWLINE] [STARTQ] Energy efficiency – A car owner takes a Ton of metal with him just to get from point A to B. Fuel efficiency of trains is superior to that of cars. [ENDQ] [NEWLINE] Again,<mask><mask><mask> the train is filled to some capacity. [NEWLINE] [NEWLINE] [STARTQ] Fuel Efficiency. [ENDQ] [NEWLINE] Same<mask> above. [NEWLINE] [NEWLINE] [STARTQ] Skill Transfer- People who graduate in engineering and were thinking of joining an automotive manufacturer can just<mask> easily join a locomotive manufacturer. [ENDQ] [NEWLINE] Yes and no. An engineer can adapt or retrain,
Label encoding: <s>I see a number of issues with you argument, I'll try to list them here: [NEWLINE] [NEWLINE] [STARTQ] If public transport such as railways are expanded and given priority over road vehicles in infrastructure, then we can achieve the goals that the automotive industry is trying to achieve very quickly. [ENDQ] [NEWLINE] Public transport is already given priority over automobiles - think bus lanes in most metropolitan areas, railway crossings, tram lines, city center tax for automobiles, etc. [NEWLINE] [NEWLINE] As for R&amp;D, train companies will invest in trains and auto companies will invest in cars. [NEWLINE] [NEWLINE] [STARTQ] We should divert the resources that we spend on making smarter cars (finances and human capital) to making smarter, faster, safer, public transport and expand the rail network. If we just expand the rail network then people will automatically shift from driving a car to riding on a train as their preferred choice of daily commute. [ENDQ] [NEWLINE] Unfortunately you cannot dictate how a private business spends it's money. The need for automobiles will always exist, as public transportation can not cover the whole road network. If this need exists, it is logical someone will invest into bettering the technology. [NEWLINE] [NEWLINE] [STARTQ] Environmental impact- A vast majority of cars are running on fuels that emit pollutants in the environment. [ENDQ] [NEWLINE] For now. The technology for electric vehicles is already here and rapid progress is made in the field of energy storage. Also, public transportation does not mean pollution free. A train carrying 20 passengers on a not- so -popular route can pollute more than 4 cars (depending on the energy source). [NEWLINE] [NEWLINE] [STARTQ] Safety- Road injury was one of the top 10 causes of death in the world. According to WHO, road injury took lives of 1.3 Million people in the last decade. [ENDQ] [NEWLINE] This is actually a reason for safer cars/self driving cars than for using public transportation. [NEWLINE] [NEWLINE] [STARTQ] Maintenance – a car requires frequent upkeep and maintenance of its parts and components. [ENDQ] [NEWLINE] As do trains, trams, buses, etc. [NEWLINE] [NEWLINE] [STARTQ] Energy efficiency – A car owner takes a Ton of metal with him just to get from point A to B. Fuel efficiency of trains is superior to that of cars. [ENDQ] [NEWLINE] Again, as long as the train is filled to some capacity. [NEWLINE] [NEWLINE] [STARTQ] Fuel Efficiency. [ENDQ] [NEWLINE] Same as above. [NEWLINE] [NEWLINE] [STARTQ] Skill Transfer- People who graduate in engineering and were thinking of joining an automotive manufacturer can just as easily join a locomotive manufacturer. [ENDQ] [NEWLINE] Yes and no. An engineer can adapt or retrain,
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0670, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0427, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hey dude. I am an environmental engineer. My degree was in civil engineering, and I am just finishing up a doctorate in sustainable engineering. I currently work in a carpet company called Interface. The company is amazing. The whole firm is teeming with extremely environmentally conscious people who deeply care about doing the right thing. CEO and shop floor workers alike. The company makes a lot of money for its shareholders by being the best, most ethical environmental choice among our competitors.<mask> you know<mask>? The biggest impact we can have is through influencing others. We push our suppliers to be more sustainable, to find new recycling techniques or to investigate alternative materials. We<mask> encourage our customers to send back their products at the end of their life. We have an incredible programme called Net-Works which is a sustainability home run... Check it out [here.]( [URL] ). It's fucking inspirational. [NEWLINE] [NEWLINE] My point is that there are truly companies out there that are willing to do the right thing. Not all companies are laggards. There are courageous CEOs and companies with visionary and progressive CSER programmes that are making a real, tangible difference by disrupting the market and showing that there is a better way to operate. Unilever. Patagonia. Nike. Puma. Caterpillar. Marks and Spencer. Desso. They are part of a movement. [NEWLINE] [NEWLINE] In the example of Interface, we have changed the whole carpet industry. In 1995, we were the only ones found doing it. Now every carpet company has a similarly radical and exciting environmental programme,<mask> they realised that otherwise they could risk get left behind by the marketplace. [NEWLINE] [NEWLINE] My advice is this: take the well paid job. Get a few years under your belt. Broaden your resume and really develop yourself in<mask> many ways<mask> you can. Take every opportunity to attend conferences, take courses and fully embrace being 'the environmental guy' at your work. Leon about compliance Maybe even set up a 'green team'  in your office which meets every few months to brainstorm about ways to save money through reducing waste or saving energy. Make yourself valuable and attractive. [NEWLINE] [NEWLINE] All kinds of companies need people like us. The key is to have diversity of skills. You need to be a committed leader<mask><mask> you need to be competent, reliable and wise -  that's<mask> it means to be an engineer. In the future, there will be many more examples of companies like mine,<mask> they can only get there through perseverance and dedication from people like you and me
Label encoding: <s>Hey dude. I am an environmental engineer. My degree was in civil engineering, and I am just finishing up a doctorate in sustainable engineering. I currently work in a carpet company called Interface. The company is amazing. The whole firm is teeming with extremely environmentally conscious people who deeply care about doing the right thing. CEO and shop floor workers alike. The company makes a lot of money for its shareholders by being the best, most ethical environmental choice among our competitors. But you know what? The biggest impact we can have is through influencing others. We push our suppliers to be more sustainable, to find new recycling techniques or to investigate alternative materials. We also encourage our customers to send back their products at the end of their life. We have an incredible programme called Net-Works which is a sustainability home run... Check it out [here.]( [URL] ). It's fucking inspirational. [NEWLINE] [NEWLINE] My point is that there are truly companies out there that are willing to do the right thing. Not all companies are laggards. There are courageous CEOs and companies with visionary and progressive CSER programmes that are making a real, tangible difference by disrupting the market and showing that there is a better way to operate. Unilever. Patagonia. Nike. Puma. Caterpillar. Marks and Spencer. Desso. They are part of a movement. [NEWLINE] [NEWLINE] In the example of Interface, we have changed the whole carpet industry. In 1995, we were the only ones found doing it. Now every carpet company has a similarly radical and exciting environmental programme, because they realised that otherwise they could risk get left behind by the marketplace. [NEWLINE] [NEWLINE] My advice is this: take the well paid job. Get a few years under your belt. Broaden your resume and really develop yourself in as many ways as you can. Take every opportunity to attend conferences, take courses and fully embrace being 'the environmental guy' at your work. Leon about compliance Maybe even set up a 'green team'  in your office which meets every few months to brainstorm about ways to save money through reducing waste or saving energy. Make yourself valuable and attractive. [NEWLINE] [NEWLINE] All kinds of companies need people like us. The key is to have diversity of skills. You need to be a committed leader but also you need to be competent, reliable and wise -  that's what it means to be an engineer. In the future, there will be many more examples of companies like mine, but they can only get there through perseverance and dedication from people like you and me
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There's something that is missing from this conversation--the value of college is not purely financial.  I'm going to speak in generalizations here (obviously YMMV based on the individual school, major, etc.) [NEWLINE] [NEWLINE] **<mask> to Brain: Advanced Level** College is a place<mask> you learn to think critically, analyze material at a high level, and put out a significant quantity of your own work.  That work is then scrutinized by experts in the field to which it pertains, returned to you, and often tweaked and improved, back and forth, for some time.  This experience is relatively unique to a college/university setting.  To me, that academic discourse involved some of the the most challenging, engaging, and important conversations that I have ever had in my life and influence me on a daily basis, nearly a decade later. [NEWLINE] [NEWLINE] **Networking and socialization:** One of the significant advantages (which will not manifest immediately) of college is the social interaction and networking that occur there.  Your peers in your field will, eventually, be the people in charge of that field, and<mask>, will be the people with jobs to offer or serve<mask> references/recommendations.  This is often part of the long-term appeal of fraternities and sororities,<mask> well<mask> a huge part of the Ivy League culture.  People are inclined to recommend and hire people they know or who share similar affiliations. <mask>, college is a place<mask> many people meet lifelong friends and significant others.  Obviously, it is not the only place to meet these types of people,<mask> it sets up a culture<mask> those relationships flourish much more easily--living situations, on-campus social events, study groups, etc. facilitate meeting new people and having ongoing interactions. [NEWLINE] [NEWLINE] **Side-by-side, it's better to have one** Yes, a lot of grads are having a tough time getting work.  In my experience,<mask> (and take this with a grain of salt<mask> it is anecdotal), every company I have worked for,<mask> faced with the option of two equally qualified candidates<mask> one was college-educated and one was not, selected the former. [NEWLINE] [NEWLINE] [NEWLINE] The debt is horrific--don't get me wrong. <mask><mask> you prepare properly (start by checking out /r/personalfinance for some planning advice), it is escapable.  For me, the greatest value of my college experience was in the friends and connections I
Label encoding: <s>There's something that is missing from this conversation--the value of college is not purely financial.  I'm going to speak in generalizations here (obviously YMMV based on the individual school, major, etc.) [NEWLINE] [NEWLINE] ** How to Brain: Advanced Level** College is a place where you learn to think critically, analyze material at a high level, and put out a significant quantity of your own work.  That work is then scrutinized by experts in the field to which it pertains, returned to you, and often tweaked and improved, back and forth, for some time.  This experience is relatively unique to a college/university setting.  To me, that academic discourse involved some of the the most challenging, engaging, and important conversations that I have ever had in my life and influence me on a daily basis, nearly a decade later. [NEWLINE] [NEWLINE] **Networking and socialization:** One of the significant advantages (which will not manifest immediately) of college is the social interaction and networking that occur there.  Your peers in your field will, eventually, be the people in charge of that field, and consequently, will be the people with jobs to offer or serve as references/recommendations.  This is often part of the long-term appeal of fraternities and sororities, as well as a huge part of the Ivy League culture.  People are inclined to recommend and hire people they know or who share similar affiliations.  Additionally, college is a place where many people meet lifelong friends and significant others.  Obviously, it is not the only place to meet these types of people, but it sets up a culture where those relationships flourish much more easily--living situations, on-campus social events, study groups, etc. facilitate meeting new people and having ongoing interactions. [NEWLINE] [NEWLINE] **Side-by-side, it's better to have one** Yes, a lot of grads are having a tough time getting work.  In my experience, however (and take this with a grain of salt because it is anecdotal), every company I have worked for, when faced with the option of two equally qualified candidates where one was college-educated and one was not, selected the former. [NEWLINE] [NEWLINE] [NEWLINE] The debt is horrific--don't get me wrong.  But if you prepare properly (start by checking out /r/personalfinance for some planning advice), it is escapable.  For me, the greatest value of my college experience was in the friends and connections I
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0415, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Let me introduce you to presence operations. [NEWLINE] [NEWLINE] The most benign presence operation sounds almost like a game.  "Glowstick tossing". <mask> you do is you go around neighborhoods and toss glowsticks into houses<mask> that<mask> the inhabitants wake up they see the glowsticks,<mask> they're reminded that government agents passed their house overnight, and could do basically anything they wanted. [NEWLINE] [NEWLINE] <mask> that's not going to paralyze the average joe with fear of big brother, is it, [let's crank that weak sauce up a notch.]( [URL].ign.com/futurama/5/51/Bender.jpg) [NEWLINE] [NEWLINE] "Mapping" sounds innocuous too.  It's a little more complicated.  You go around after midnight.  You get in your uniform, get all your guns and stuff<mask> you look intimidating, machine guns and stuff.  You surround a house like a military operation, draw it's windows, doors, fences,<mask> on, then you knock on the door, wake everyone up, then you take the ID of everyone in the house, get their cell phones, draw the rooms in the house, search the rooms, cabinets, bags, toss the place, ask to see all the male children, take pictures of everyone in the house, and<mask> on. <mask> then, the important thing is that you destroy all the drawings you made, and delete all the photos you took. <mask> that was a pretense. <mask> you recorded those things, then that would be that,<mask> the important thing is that you "map" the house again, and again, and again, knowing that it often causes psychological problems for the children.  That's the point.  It's psychological warfare. [NEWLINE] [NEWLINE] Then there are fake arrests.  The first step is to pick a random person, and then vet them.  You have to make sure they're innocent.  You can't fake arrest someone<mask> they're wanted for something,<mask> you have to make sure that they're not wanted for anything.  Then you stage a raid on the house, and arrest them.  Again, you do this after midnight.  You grill them for a few hours and then release them. [NEWLINE] [NEWLINE] Then there's stuff like demolishing the houses of innocent people, or taking their land.  Taking away their civic rights.  Discriminating against them.  Treating them like second class citizens. [NEWLINE] [NEWLINE] Is that<mask> it's like in the US? [NEWLINE] [NEWLINE]
Label encoding: <s>Let me introduce you to presence operations. [NEWLINE] [NEWLINE] The most benign presence operation sounds almost like a game.  "Glowstick tossing".  What you do is you go around neighborhoods and toss glowsticks into houses so that when the inhabitants wake up they see the glowsticks, so they're reminded that government agents passed their house overnight, and could do basically anything they wanted. [NEWLINE] [NEWLINE] But that's not going to paralyze the average joe with fear of big brother, is it, [let's crank that weak sauce up a notch.]( [URL].ign.com/futurama/5/51/Bender.jpg) [NEWLINE] [NEWLINE] "Mapping" sounds innocuous too.  It's a little more complicated.  You go around after midnight.  You get in your uniform, get all your guns and stuff so you look intimidating, machine guns and stuff.  You surround a house like a military operation, draw it's windows, doors, fences, so on, then you knock on the door, wake everyone up, then you take the ID of everyone in the house, get their cell phones, draw the rooms in the house, search the rooms, cabinets, bags, toss the place, ask to see all the male children, take pictures of everyone in the house, and so on.  But then, the important thing is that you destroy all the drawings you made, and delete all the photos you took.  Because that was a pretense.  If you recorded those things, then that would be that, but the important thing is that you "map" the house again, and again, and again, knowing that it often causes psychological problems for the children.  That's the point.  It's psychological warfare. [NEWLINE] [NEWLINE] Then there are fake arrests.  The first step is to pick a random person, and then vet them.  You have to make sure they're innocent.  You can't fake arrest someone if they're wanted for something, so you have to make sure that they're not wanted for anything.  Then you stage a raid on the house, and arrest them.  Again, you do this after midnight.  You grill them for a few hours and then release them. [NEWLINE] [NEWLINE] Then there's stuff like demolishing the houses of innocent people, or taking their land.  Taking away their civic rights.  Discriminating against them.  Treating them like second class citizens. [NEWLINE] [NEWLINE] Is that what it's like in the US? [NEWLINE] [NEWLINE]
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0390, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0374, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask><mask> there's a bias involved here,<mask> not the one that you think. [NEWLINE] [NEWLINE] We have pretty much agreed<mask> a society that the correct standard to use in criminal cases is "beyond a reasonable doubt". [NEWLINE] [NEWLINE] We do this<mask> without it, no one could ever be convicted of anything,<mask> there is always doubt, even<mask> it might be unreasonable. It's a tradeoff between the safety of many innocent people vs. the safety of a very small number of innocent people. [NEWLINE] [NEWLINE] The definition of "reasonable doubt" is usually phrased<mask> : [NEWLINE] [NEWLINE] [STARTQ] "Reasonable doubt is not mere possible doubt. "It is that state of the case which, after the entire comparison and consideration of all the evidence leaves the minds of the jurors in that condition that they canot say they feel an abiding conviction to a moral certainty of the truth of the charge." [ENDQ] [NEWLINE] I think that you're right that death-penalty opponents have a lower chance of convicting someone of a capital charge. [NEWLINE] [NEWLINE] <mask>,<mask><mask> that's<mask> they would use a standard of evidence higher than "beyond a reasonable doubt".<mask> of the worry that the criminal might be put to death, they would require certainty beyond *any* doubt, which isn't a standard that we can possibly use and still protect society against murderers. [NEWLINE] [NEWLINE] The question isn't who has a higher or lower standard of saying someone is guilty. It's "who will properly use the standard of guilt that we,<mask> a society, have decided is appropriate". Anyone that has a strong moral reason not to use that standard of guilt isn't appropriate for a trial. [NEWLINE] [NEWLINE] Some kind of sadist who is all gung ho to kill people would be likely to use *less* than a standard of "beyond a reasonable doubt" in a capital trial, and they should be excluded. [NEWLINE] [NEWLINE] Someone who is morally opposed to killing anyone for any reason would be likely to use *more* than a standard of "beyond a reasonable doubt" and *<mask> * should be excluded. [NEWLINE] [NEWLINE] Both should be excluded. [NEWLINE] [NEWLINE] Either that, or we should just change the standard to "beyond all doubt" and never convict anyone. [NEWLINE] [NEWLINE] Or even better,<mask><mask><mask>, get rid of the death penalty,<mask> it really does require a moral certainty that's far beyond "reasonable doubt". [NEWLINE] [NEWLINE] <mask><mask><mask><mask> we have the death penalty, and that uniform standard of guilt, we can't allow anyone on the jury that
Label encoding: <s>I think there's a bias involved here, but not the one that you think. [NEWLINE] [NEWLINE] We have pretty much agreed as a society that the correct standard to use in criminal cases is "beyond a reasonable doubt". [NEWLINE] [NEWLINE] We do this because without it, no one could ever be convicted of anything, because there is always doubt, even if it might be unreasonable. It's a tradeoff between the safety of many innocent people vs. the safety of a very small number of innocent people. [NEWLINE] [NEWLINE] The definition of "reasonable doubt" is usually phrased as : [NEWLINE] [NEWLINE] [STARTQ] "Reasonable doubt is not mere possible doubt. "It is that state of the case which, after the entire comparison and consideration of all the evidence leaves the minds of the jurors in that condition that they canot say they feel an abiding conviction to a moral certainty of the truth of the charge." [ENDQ] [NEWLINE] I think that you're right that death-penalty opponents have a lower chance of convicting someone of a capital charge. [NEWLINE] [NEWLINE] However, I think that's because they would use a standard of evidence higher than "beyond a reasonable doubt". Because of the worry that the criminal might be put to death, they would require certainty beyond *any* doubt, which isn't a standard that we can possibly use and still protect society against murderers. [NEWLINE] [NEWLINE] The question isn't who has a higher or lower standard of saying someone is guilty. It's "who will properly use the standard of guilt that we, as a society, have decided is appropriate". Anyone that has a strong moral reason not to use that standard of guilt isn't appropriate for a trial. [NEWLINE] [NEWLINE] Some kind of sadist who is all gung ho to kill people would be likely to use *less* than a standard of "beyond a reasonable doubt" in a capital trial, and they should be excluded. [NEWLINE] [NEWLINE] Someone who is morally opposed to killing anyone for any reason would be likely to use *more* than a standard of "beyond a reasonable doubt" and * also * should be excluded. [NEWLINE] [NEWLINE] Both should be excluded. [NEWLINE] [NEWLINE] Either that, or we should just change the standard to "beyond all doubt" and never convict anyone. [NEWLINE] [NEWLINE] Or even better, in my opinion, get rid of the death penalty, because it really does require a moral certainty that's far beyond "reasonable doubt". [NEWLINE] [NEWLINE] But as long as we have the death penalty, and that uniform standard of guilt, we can't allow anyone on the jury that
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Except there are certain religious concepts that are outside the realm of science. Can science prove or disprove the existence of a higher power? Can it prove or disprove the idea that everyone has a soul? Or the idea of an afterlife? [ENDQ] [NEWLINE] They have not *currently* been proven, not to say that they ever will be. All three questions you have confronted me with are based on blind faith. Can science prove we're not in the matrix? Can it prove or disprove the idea that everyone is psychic? Or the idea of the Silver Surfer? [NEWLINE] [NEWLINE] [STARTQ] Religious books are nothing more than stories and guidelines with countless different interpretations which have changed over time. The interpretation of religious texts evolves<mask> time goes on,<mask> the overall lessons that they teach are timeless. [ENDQ] [NEWLINE] That sounds good enough to CMV!<mask>.... this view is not held by many religious believers<mask> many tend to take these texts quite literally. [NEWLINE] [NEWLINE] [STARTQ] Maybe based on your interpretation of religious texts,<mask> there are many common interpretations which teach love and compassion for others. You<mask> mentioned that religion teaches that the only two options for people after they die are eternal happiness or eternal damnation. This simply isn't true. Some religions teach reincarnation,<mask> other religions don't say much about<mask> happens after we die at all and focus on<mask> happens<mask> we are alive instead. [ENDQ] [NEWLINE] <mask> you're referring to Eastern cultures like Buddhism or Hinduism, Reincarnation is damnation<mask> this world is a form of hell that one needs to overcome through meditation and spiritual enlightenment in order to reach Nirvana/Samsara which is essentially an eternal enlightened paradise. [NEWLINE] [NEWLINE] [STARTQ] It can, just like anything else can separate humanity. Politics separates humanity. Money separates humanity.<mask> people live separates humanity.<mask> type of car people drive separates humanity. Anything can separate humanity. [ENDQ] [NEWLINE] <mask><mask> not get rid of religion in order to limit these separations? [NEWLINE] [NEWLINE] [STARTQ] Human intellect has continued to grow throughout history<mask> the fact that religion has been a large part of society for<mask> long.<mask> religion is truly in the way of proper intellect and reasoning like you seem to think, then<mask> have we developed<mask> well to this point in history? [ENDQ] [NEWLINE] We have developed<mask> well by people who have thought outside the typical norms and through technological and human advancement.<mask> people became more free and logic was shared throughout the world we have been able to develop<mask> well,<mask> in many cases development has been interrupted by religion, such<mask> stem cell research.
Label encoding: <s> [STARTQ] Except there are certain religious concepts that are outside the realm of science. Can science prove or disprove the existence of a higher power? Can it prove or disprove the idea that everyone has a soul? Or the idea of an afterlife? [ENDQ] [NEWLINE] They have not *currently* been proven, not to say that they ever will be. All three questions you have confronted me with are based on blind faith. Can science prove we're not in the matrix? Can it prove or disprove the idea that everyone is psychic? Or the idea of the Silver Surfer? [NEWLINE] [NEWLINE] [STARTQ] Religious books are nothing more than stories and guidelines with countless different interpretations which have changed over time. The interpretation of religious texts evolves as time goes on, while the overall lessons that they teach are timeless. [ENDQ] [NEWLINE] That sounds good enough to CMV! But.... this view is not held by many religious believers as many tend to take these texts quite literally. [NEWLINE] [NEWLINE] [STARTQ] Maybe based on your interpretation of religious texts, but there are many common interpretations which teach love and compassion for others. You also mentioned that religion teaches that the only two options for people after they die are eternal happiness or eternal damnation. This simply isn't true. Some religions teach reincarnation, while other religions don't say much about what happens after we die at all and focus on what happens while we are alive instead. [ENDQ] [NEWLINE] If you're referring to Eastern cultures like Buddhism or Hinduism, Reincarnation is damnation as this world is a form of hell that one needs to overcome through meditation and spiritual enlightenment in order to reach Nirvana/Samsara which is essentially an eternal enlightened paradise. [NEWLINE] [NEWLINE] [STARTQ] It can, just like anything else can separate humanity. Politics separates humanity. Money separates humanity. Where people live separates humanity. What type of car people drive separates humanity. Anything can separate humanity. [ENDQ] [NEWLINE] So why not get rid of religion in order to limit these separations? [NEWLINE] [NEWLINE] [STARTQ] Human intellect has continued to grow throughout history despite the fact that religion has been a large part of society for so long. If religion is truly in the way of proper intellect and reasoning like you seem to think, then how have we developed so well to this point in history? [ENDQ] [NEWLINE] We have developed so well by people who have thought outside the typical norms and through technological and human advancement. As people became more free and logic was shared throughout the world we have been able to develop so well, but in many cases development has been interrupted by religion, such as stem cell research.
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's anything<mask> a succinct description. Really, everything except the "*Aeon of Strife*-styled" part is completely redundant; the rest of the letters are there to make the acronym offensive/silly. "Fortress Assault" could mean anything from *Dynasty Warriors* to *Heroes of Might and Magic* to classic RTS games. "Going on Two Sides" is not a necessary element of the genre at all (<mask><mask>, back in the days of Battle.net and Garena [three-sided AoS-like games]( [URL] /) were fairly popular). That last part<mask> sounds kind of weird to me,<mask> w/e, I'm not a native speaker. [NEWLINE] [NEWLINE] "*Aeon of Strife*-like",<mask>, doesn't really explain anything at all, unless you know<mask> *Aeon of Strife* actually is, and even<mask> you do, it's still somewhat confusing,<mask> modern variations of it are<mask> different from the original. <mask> we want our genre acronym to be descriptive, we need to think of<mask> distinguishes games in that genre from other games. For example, in a first-person shooter you shoot at things from a first-person perspective (duh),<mask> we know that *Half-Life* is an FPS,<mask> *Guitar Hero* isn't.<mask><mask> is it that is present in AoS, DotA, LoL, HoN, HoTS, *Awesomenauts*, etc., and doesn't exist in other games? [NEWLINE] [NEWLINE] It's pushing lanes. You can do without items (*Heroes of the Storm* does that,<mask> did the original *Aeon of Strife*, IIRC), jungles, skills; even the RTS-like overhead perspective and mouse control is more of a legacy of the *Warcraft* era than an actual feature of the genre.<mask>, lanes, towers and creeps are inseparable from this kind of game; without them, you get something different. [NEWLINE] [NEWLINE] <mask>,<mask> about LPG, that is, "Lane Pushing Game"? It's about<mask> precise a description<mask> you can get, and it has a fair bit of traction in the community, too. It's<mask> [PC Gamer]( [URL] /) tags such games, for instance. [NEWLINE] [NEWLINE] Honestly,<mask>,<mask><mask> we should just stick with MOBA. Yes, it's vague,<mask> thanks to the efforts of *Riot* everyone knows<mask>
Label encoding: <s>It's anything but a succinct description. Really, everything except the "*Aeon of Strife*-styled" part is completely redundant; the rest of the letters are there to make the acronym offensive/silly. "Fortress Assault" could mean anything from *Dynasty Warriors* to *Heroes of Might and Magic* to classic RTS games. "Going on Two Sides" is not a necessary element of the genre at all ( in fact, back in the days of Battle.net and Garena [three-sided AoS-like games]( [URL] /) were fairly popular). That last part also sounds kind of weird to me, but w/e, I'm not a native speaker. [NEWLINE] [NEWLINE] "*Aeon of Strife*-like", however, doesn't really explain anything at all, unless you know what *Aeon of Strife* actually is, and even if you do, it's still somewhat confusing, since modern variations of it are so different from the original.  If we want our genre acronym to be descriptive, we need to think of what distinguishes games in that genre from other games. For example, in a first-person shooter you shoot at things from a first-person perspective (duh), so we know that *Half-Life* is an FPS, while *Guitar Hero* isn't. So what is it that is present in AoS, DotA, LoL, HoN, HoTS, *Awesomenauts*, etc., and doesn't exist in other games? [NEWLINE] [NEWLINE] It's pushing lanes. You can do without items (*Heroes of the Storm* does that, as did the original *Aeon of Strife*, IIRC), jungles, skills; even the RTS-like overhead perspective and mouse control is more of a legacy of the *Warcraft* era than an actual feature of the genre. However, lanes, towers and creeps are inseparable from this kind of game; without them, you get something different. [NEWLINE] [NEWLINE] So, how about LPG, that is, "Lane Pushing Game"? It's about as precise a description as you can get, and it has a fair bit of traction in the community, too. It's what [PC Gamer]( [URL] /) tags such games, for instance. [NEWLINE] [NEWLINE] Honestly, though, I think we should just stick with MOBA. Yes, it's vague, but thanks to the efforts of *Riot* everyone knows what
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm talking about the ethical value of an action. Happiness is the standard of comparison, *not* the value of the deed. Making a baby smile is not "worth" 2 utils in happiness, whatever that may be or mean. (I'm not the one who tried to introduce units of measure into this.) Comparisons in happiness are, by nature, relative. There is no absolute measure such<mask> an "util", and<mask><mask> even for the sake of argument it is extremely obstructive and misleading to assume there was such a thing. [NEWLINE] [NEWLINE] Prioritarians don't trade off fairness for happiness, they use happiness<mask> the measure by which to judge fairness. Personally, I've always disliked the talk about "happiness" to begin with,<mask> it's not defined, and the only thing anybody agrees on about it is that it's too individual to be adequately assessed by anybody else than the individual in question. Before I go on, I should point out that from here on I will be talking about my personal worldview rather than "official prioritarian doctrine"<mask> you<mask> will.<mask> you want something more official, I recommend the relevant first chapters from Peter Singer's *Practical Ethics*. [NEWLINE] [NEWLINE] I'm not a philosopher of ethics. I identify<mask> a Humanist, and I've never given too much thought about wether or not there may be some hypothetical dilemma in which my prioritarian principles would fail me. I've consciously lived by these principles for a little over 10 years now, and<mask> far it has *worked*. That is not something you can say for very many ethical systems. [NEWLINE] [NEWLINE] <mask>, my personal doctrine differs slightly from the original utilitarianism in that I define it negatively: *Reduce suffering* is my ethical imperative, not *maximize happiness*. This has the practical benefit that suffering is much less individual than happiness, and<mask> much easier to assess for a third party. [NEWLINE] [NEWLINE] Being a Humanist<mask> means that there's other things that enter into my equations<mask> well.<mask> a fan of the idea of Human Rights, for example, I do not believe that killing somebody against their will can ever be justified, which means that I will never fucking care<mask> much pleasure the utility monster would derive from murdering person X, I'd still think that person X's interest in his own life will always outweigh the monster's interest in (more) personal happiness. [NEWLINE] [NEWLINE] Btw, this principle of "equality of interests" is pretty essential to modern (priority-)utilitarian philosophy<mask><mask><mask>
Label encoding: <s>I'm talking about the ethical value of an action. Happiness is the standard of comparison, *not* the value of the deed. Making a baby smile is not "worth" 2 utils in happiness, whatever that may be or mean. (I'm not the one who tried to introduce units of measure into this.) Comparisons in happiness are, by nature, relative. There is no absolute measure such as an "util", and I think even for the sake of argument it is extremely obstructive and misleading to assume there was such a thing. [NEWLINE] [NEWLINE] Prioritarians don't trade off fairness for happiness, they use happiness as the measure by which to judge fairness. Personally, I've always disliked the talk about "happiness" to begin with, because it's not defined, and the only thing anybody agrees on about it is that it's too individual to be adequately assessed by anybody else than the individual in question. Before I go on, I should point out that from here on I will be talking about my personal worldview rather than "official prioritarian doctrine" if you so will. If you want something more official, I recommend the relevant first chapters from Peter Singer's *Practical Ethics*. [NEWLINE] [NEWLINE] I'm not a philosopher of ethics. I identify as a Humanist, and I've never given too much thought about wether or not there may be some hypothetical dilemma in which my prioritarian principles would fail me. I've consciously lived by these principles for a little over 10 years now, and so far it has *worked*. That is not something you can say for very many ethical systems. [NEWLINE] [NEWLINE] Also, my personal doctrine differs slightly from the original utilitarianism in that I define it negatively: *Reduce suffering* is my ethical imperative, not *maximize happiness*. This has the practical benefit that suffering is much less individual than happiness, and thus much easier to assess for a third party. [NEWLINE] [NEWLINE] Being a Humanist also means that there's other things that enter into my equations as well. As a fan of the idea of Human Rights, for example, I do not believe that killing somebody against their will can ever be justified, which means that I will never fucking care how much pleasure the utility monster would derive from murdering person X, I'd still think that person X's interest in his own life will always outweigh the monster's interest in (more) personal happiness. [NEWLINE] [NEWLINE] Btw, this principle of "equality of interests" is pretty essential to modern (priority-)utilitarian philosophy as far as
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Most guns are bought legally then sold illegally or stolen. Take away the first step and it's incredibly difficult to get guns into the ***country***. Which brings me to your next point. [NEWLINE] [NEWLINE] [STARTQ] Tell that to Chicago and the state of Illinois. Or tell that to NYC. Or DC. [ENDQ] [NEWLINE] Ya,<mask> it doesn't work<mask> you can drive 5 miles away and buy a small armory no questions asked and then drive back without any risk. It only works<mask> every place within maintained borders agrees to enforce the same legislation. And then border patrol enforces it. You can go talk to Europe about that. Or Australia. [NEWLINE] [NEWLINE] [STARTQ] <mask> the hell the government or police or whatever is going to get that money [ENDQ] [NEWLINE] That's<mask> it's impractical on a large scale<mask> the argument against OPs point still stands.<mask> I'm fairly certain (know for a fact) that cities have done this in the past and it worked. [NEWLINE] [NEWLINE] [STARTQ] you'd better believe the criminals would know about it too.<mask> the police are offering $1000 for guns, a dealer would offer $1200,<mask> he knows that anyone who wants that gun would pay more anyway. [ENDQ] [NEWLINE]...<mask>? That is a ridiculous point. Could that happen? Maybe in a few select cases.<mask> you brought up a good point [NEWLINE] [NEWLINE] [STARTQ] <mask> the hell the ~~government or police or whatever~~ dealer is going to get that money [ENDQ] [NEWLINE] You're talking about a massive amount of capital to do this.<mask> you think a dealer buying guns no questions asked for even higher prices than the police isn't going to get noticed?<mask> long before the police get them on a sting? [NEWLINE] [NEWLINE] [STARTQ] the supply WOULD go down,<mask> never completely [ENDQ] [NEWLINE] Yes, obviously. I didn't think I needed to specify this. [NEWLINE] [NEWLINE] [STARTQ] There is no amount of money worth more than the ability to defend myself [ENDQ] [NEWLINE]... I'm not even going to get into the fact that except for military and police trained individuals you are far more likely to be killed<mask> you own a gun than<mask> you don't. Even<mask> Hollywood has convinced you otherwise.<mask> this is<mask> far from the argument it's almost ridiculous. [NEWLINE] [NEWLINE] [STARTQ] <mask> you're not allowed to own guns, and nobody is, selling it off will always be worth more than getting caught with it. [ENDQ] [NEWLINE]... No?<mask> owning the gun will help you commit crimes and make money than it is more valuable to keep it than get rid of it for free. That
Label encoding: <s>Most guns are bought legally then sold illegally or stolen. Take away the first step and it's incredibly difficult to get guns into the ***country***. Which brings me to your next point. [NEWLINE] [NEWLINE] [STARTQ] Tell that to Chicago and the state of Illinois. Or tell that to NYC. Or DC. [ENDQ] [NEWLINE] Ya, because it doesn't work if you can drive 5 miles away and buy a small armory no questions asked and then drive back without any risk. It only works if every place within maintained borders agrees to enforce the same legislation. And then border patrol enforces it. You can go talk to Europe about that. Or Australia. [NEWLINE] [NEWLINE] [STARTQ] where the hell the government or police or whatever is going to get that money [ENDQ] [NEWLINE] That's why it's impractical on a large scale but the argument against OPs point still stands. Although I'm fairly certain (know for a fact) that cities have done this in the past and it worked. [NEWLINE] [NEWLINE] [STARTQ] you'd better believe the criminals would know about it too. If the police are offering $1000 for guns, a dealer would offer $1200, because he knows that anyone who wants that gun would pay more anyway. [ENDQ] [NEWLINE]... What? That is a ridiculous point. Could that happen? Maybe in a few select cases. But you brought up a good point [NEWLINE] [NEWLINE] [STARTQ] where the hell the ~~government or police or whatever~~ dealer is going to get that money [ENDQ] [NEWLINE] You're talking about a massive amount of capital to do this. Besides you think a dealer buying guns no questions asked for even higher prices than the police isn't going to get noticed? How long before the police get them on a sting? [NEWLINE] [NEWLINE] [STARTQ] the supply WOULD go down, but never completely [ENDQ] [NEWLINE] Yes, obviously. I didn't think I needed to specify this. [NEWLINE] [NEWLINE] [STARTQ] There is no amount of money worth more than the ability to defend myself [ENDQ] [NEWLINE]... I'm not even going to get into the fact that except for military and police trained individuals you are far more likely to be killed if you own a gun than if you don't. Even if Hollywood has convinced you otherwise. Because this is so far from the argument it's almost ridiculous. [NEWLINE] [NEWLINE] [STARTQ] If you're not allowed to own guns, and nobody is, selling it off will always be worth more than getting caught with it. [ENDQ] [NEWLINE]... No? If owning the gun will help you commit crimes and make money than it is more valuable to keep it than get rid of it for free. That
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>,<mask> I mentioned, one can extend the terms of the 1 year debt to 30 years. The debt burdens are now the same. Opting for full forgiveness<mask> a less disruptive solution exists seems utterly arbitrary. [ENDQ] [NEWLINE] <mask> I noted before, I don't view debt<mask> a great positive thing to support, I'm not going to suggest the least disruptive solutions. There is clear reason to my suggestions, in that I am (obviously) not a huge fan of debt. [NEWLINE] [NEWLINE] [STARTQ] The rest explained<mask> they believed that lack of access to credit interferes with career mobility, which interferes with economic mobility, which interferes with social mobility. This is precisely<mask> you were getting at. [ENDQ] [NEWLINE] I was asking for evidence that this was a true assertion, not evidence that people believed it. [NEWLINE] [NEWLINE] [STARTQ] Your proposal doesn't ban loans. It just makes them un-enforceable in a wider variety of scenarios. This natural reduces the market for credit making it more difficult for people to borrow. [ENDQ] [NEWLINE] <mask> does the government forgiving all debts and limiting collection of debts make it harder for the government to make loans? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> loans are<mask> terrible<mask> you say<mask> would you care<mask> the quality business people get them? [ENDQ] [NEWLINE] More that loans are ok for most, terrible for a large minority, and great for a small minority. [NEWLINE] [NEWLINE] [STARTQ] I'm talking about Joe getting a small business loan to buy a truck and a ride on mower to start his lawn care business. [ENDQ] [NEWLINE] <mask> I have noted, I'd<mask> support an expansion of government loans to support such things. [NEWLINE] [NEWLINE] [STARTQ] In our current environment lenders have more incentive to roll the dice on Joe which ultimately gives a larger number of people the chance to move up economically. [ENDQ] [NEWLINE] Or it locks Joe into a cycle of debt from which he can never get out of. I am suggesting that this effect is more significant than potential positive events. Most businesses don't massively pick off. [NEWLINE] [NEWLINE] [STARTQ] <mask> do you believe the debt works differently in Guatemala and Georgia? [ENDQ] [NEWLINE] It was a microloaning project and I couldn't find the actual source, I don't know<mask> the source says. Microloans are very different from bank loans. [NEWLINE] [NEWLINE] [STARTQ] <mask> do you think the taxpayer is going to happily subsidize crappy loans that people don't have to pay back<mask> they have demonstrated time and again that they have no interest in subsidizing education (which they should). [ENDQ] [NEWLINE] The rich tax payer would
Label encoding: <s> [STARTQ] But, as I mentioned, one can extend the terms of the 1 year debt to 30 years. The debt burdens are now the same. Opting for full forgiveness when a less disruptive solution exists seems utterly arbitrary. [ENDQ] [NEWLINE] As I noted before, I don't view debt as a great positive thing to support, I'm not going to suggest the least disruptive solutions. There is clear reason to my suggestions, in that I am (obviously) not a huge fan of debt. [NEWLINE] [NEWLINE] [STARTQ] The rest explained how they believed that lack of access to credit interferes with career mobility, which interferes with economic mobility, which interferes with social mobility. This is precisely what you were getting at. [ENDQ] [NEWLINE] I was asking for evidence that this was a true assertion, not evidence that people believed it. [NEWLINE] [NEWLINE] [STARTQ] Your proposal doesn't ban loans. It just makes them un-enforceable in a wider variety of scenarios. This natural reduces the market for credit making it more difficult for people to borrow. [ENDQ] [NEWLINE] Why does the government forgiving all debts and limiting collection of debts make it harder for the government to make loans? [NEWLINE] [NEWLINE] [STARTQ] Also if loans are as terrible as you say why would you care if the quality business people get them? [ENDQ] [NEWLINE] More that loans are ok for most, terrible for a large minority, and great for a small minority. [NEWLINE] [NEWLINE] [STARTQ] I'm talking about Joe getting a small business loan to buy a truck and a ride on mower to start his lawn care business. [ENDQ] [NEWLINE] As I have noted, I'd also support an expansion of government loans to support such things. [NEWLINE] [NEWLINE] [STARTQ] In our current environment lenders have more incentive to roll the dice on Joe which ultimately gives a larger number of people the chance to move up economically. [ENDQ] [NEWLINE] Or it locks Joe into a cycle of debt from which he can never get out of. I am suggesting that this effect is more significant than potential positive events. Most businesses don't massively pick off. [NEWLINE] [NEWLINE] [STARTQ] Why do you believe the debt works differently in Guatemala and Georgia? [ENDQ] [NEWLINE] It was a microloaning project and I couldn't find the actual source, I don't know what the source says. Microloans are very different from bank loans. [NEWLINE] [NEWLINE] [STARTQ] Why do you think the taxpayer is going to happily subsidize crappy loans that people don't have to pay back when they have demonstrated time and again that they have no interest in subsidizing education (which they should). [ENDQ] [NEWLINE] The rich tax payer would
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0509, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0446, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A lot of games considered the best fall into the category of WRPG. Baldurs Gate, Planescape Torment, Dragon Age, Mass Effect, Skyrim, Morrowind, The Witcher etc. [NEWLINE] [NEWLINE] These games, and many other highly regarded games, have to rely on dialogue trees and manually created storyline branches for narration. This is a massive, massive limiting factor to<mask> good the gameplay of these highly dialogue-focused games can get. The characters within the game cannot react to anything that hasn't explicitly been programmed<mask> a reaction. You have limited dialogue options. You have limited ways of tackling the problems the game presents to you. In particular, romance or character-focused subplots tend to be incredibly stilted by this. [NEWLINE] [NEWLINE] Modern advances in hardware may be able to fix these issues using artificial general intelligence. [NEWLINE] [NEWLINE] Imagine an RPG (or other genres<mask> you wish)<mask> your party members can react to anything you say to them like a person would, with varying personalities. Party members that can form genuine reactions to the player, from adoration to begrudging acceptance to love, depending on<mask> you say AND<mask> you do. NPCs that can be debated with. Merchants you can negotiate prices with beyond the scope of a minigame. Enemies you can convince to switch sides through your own conversational skills. [NEWLINE] [NEWLINE] This can go even further<mask> you apply this intelligence to every inhabitant simultaneously across a huge world. No longer do developers have to hard-code the story; the story emerges out of the intelligent actions of hundreds of characters. Dynamic economies, cities, political intrigue and diplomacy can create a game world far more immersive than anything that currently exists. AGI game developers could even work behind the scenes<mask> the game runs to add new content, tuned to<mask> the player enjoys playing based upon their reaction to current content. [NEWLINE] [NEWLINE] In short, CK2 on steroids<mask> you have infinite possible interactions with every other character instead of fixed choices you click on. [NEWLINE] [NEWLINE] This is not possible at all now. The very small number of attempts to put NPCs you can converse with directly instead of via dialogue choices are<mask> poor<mask> to be comedic instead of realistic (Façade is an example of this). I am not even certain that<mask> I said above will happen,<mask> it's certainly within the realm of possibility. AGI would create games far superior to anything we currently have. This is<mask><mask><mask> a statement like "The best video game of all time has already been made" to be false; the potential
Label encoding: <s>A lot of games considered the best fall into the category of WRPG. Baldurs Gate, Planescape Torment, Dragon Age, Mass Effect, Skyrim, Morrowind, The Witcher etc. [NEWLINE] [NEWLINE] These games, and many other highly regarded games, have to rely on dialogue trees and manually created storyline branches for narration. This is a massive, massive limiting factor to how good the gameplay of these highly dialogue-focused games can get. The characters within the game cannot react to anything that hasn't explicitly been programmed as a reaction. You have limited dialogue options. You have limited ways of tackling the problems the game presents to you. In particular, romance or character-focused subplots tend to be incredibly stilted by this. [NEWLINE] [NEWLINE] Modern advances in hardware may be able to fix these issues using artificial general intelligence. [NEWLINE] [NEWLINE] Imagine an RPG (or other genres if you wish) where your party members can react to anything you say to them like a person would, with varying personalities. Party members that can form genuine reactions to the player, from adoration to begrudging acceptance to love, depending on what you say AND what you do. NPCs that can be debated with. Merchants you can negotiate prices with beyond the scope of a minigame. Enemies you can convince to switch sides through your own conversational skills. [NEWLINE] [NEWLINE] This can go even further if you apply this intelligence to every inhabitant simultaneously across a huge world. No longer do developers have to hard-code the story; the story emerges out of the intelligent actions of hundreds of characters. Dynamic economies, cities, political intrigue and diplomacy can create a game world far more immersive than anything that currently exists. AGI game developers could even work behind the scenes as the game runs to add new content, tuned to what the player enjoys playing based upon their reaction to current content. [NEWLINE] [NEWLINE] In short, CK2 on steroids because you have infinite possible interactions with every other character instead of fixed choices you click on. [NEWLINE] [NEWLINE] This is not possible at all now. The very small number of attempts to put NPCs you can converse with directly instead of via dialogue choices are so poor as to be comedic instead of realistic (Façade is an example of this). I am not even certain that what I said above will happen, but it's certainly within the realm of possibility. AGI would create games far superior to anything we currently have. This is why I think a statement like "The best video game of all time has already been made" to be false; the potential
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0426, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Not trying to start a flame war, I'm looking for a new tablet and I would consider an ipad,<mask> I just don't see the appeal just<mask>.<mask> on earth does it do that justifies a price tage 2-3x that of an android or windows tablet that performs the same functions? [NEWLINE] [NEWLINE] Clearly people are buying them,<mask> there's obviously something there I'm not seeing. [NEWLINE] [NEWLINE] I used my previous tablet (Nexus 7, rip) for [NEWLINE] [NEWLINE] * TV/Movies (kodi) [NEWLINE] * streaming said movies to my tv (MicroUSB- [STARTQ] HDMI out) [ENDQ] * ebooks ([pocketbook]( [URL].obreey.reader) ftw) [NEWLINE] * chrome (teh web) [NEWLINE] * dolphin browser (porn/flash videos) [NEWLINE] * reddit (relay) [NEWLINE] * google voice search (<mask> sometimes I can't be bothered to type shit out) [NEWLINE] * podcast/music streaming (8tracks,spotify,podcast addict) [NEWLINE] * some games (not a ton) [NEWLINE] * all the free apps [NEWLINE] * vpn (pia) [NEWLINE] [NEWLINE] Given the above, can someone tell me<mask> justifies paying an extra $360* for an ipad? (I could by 3 new nexus 7 tablets for the price of 1 ipad air 2 and have $60 left over*), wtf is going on? [NEWLINE] [NEWLINE] [NEWLINE] *figures compare this [n7]( [URL] ) to this [ipad2]( [URL] -dc_mtid_1870765e38482_pcrid_52243313890_&amp;cid=aos-us-kwg-pla-ipad-slid-) [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***.
Label encoding: <s>Not trying to start a flame war, I'm looking for a new tablet and I would consider an ipad, but I just don't see the appeal just yet. What on earth does it do that justifies a price tage 2-3x that of an android or windows tablet that performs the same functions? [NEWLINE] [NEWLINE] Clearly people are buying them, so there's obviously something there I'm not seeing. [NEWLINE] [NEWLINE] I used my previous tablet (Nexus 7, rip) for [NEWLINE] [NEWLINE] * TV/Movies (kodi) [NEWLINE] * streaming said movies to my tv (MicroUSB- [STARTQ] HDMI out) [ENDQ] * ebooks ([pocketbook]( [URL].obreey.reader) ftw) [NEWLINE] * chrome (teh web) [NEWLINE] * dolphin browser (porn/flash videos) [NEWLINE] * reddit (relay) [NEWLINE] * google voice search ( because sometimes I can't be bothered to type shit out) [NEWLINE] * podcast/music streaming (8tracks,spotify,podcast addict) [NEWLINE] * some games (not a ton) [NEWLINE] * all the free apps [NEWLINE] * vpn (pia) [NEWLINE] [NEWLINE] Given the above, can someone tell me what justifies paying an extra $360* for an ipad? (I could by 3 new nexus 7 tablets for the price of 1 ipad air 2 and have $60 left over*), wtf is going on? [NEWLINE] [NEWLINE] [NEWLINE] *figures compare this [n7]( [URL] ) to this [ipad2]( [URL] -dc_mtid_1870765e38482_pcrid_52243313890_&amp;cid=aos-us-kwg-pla-ipad-slid-) [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***.
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0503, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0389, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Statistics don't really make you feel safer, nor do they really give you a reason to feel safer. It may be the case for some people,<mask> not all. [NEWLINE] [NEWLINE] Take me for example. I'm afraid of flying on planes. Not terribly afraid,<mask> heights scare me and turbulence of any sort sets me into panic mode.<mask> I tell someone I'm afraid of flying, or that flying makes me nervous, I get the answer "Well, **statistically speaking**, you've got more of a chance of getting in a car wreck than you do a plane crash!" [NEWLINE] [NEWLINE] This is true,<mask>, it does NOT quell my fears any. No matter<mask> slim the chance is,<mask> you get nervous like that the only thing going through your mind is "I might make the news today". [NEWLINE] [NEWLINE] <mask> terrorism (generally speaking) isn't about planes. Terrorism is about something else. People who fear terrorism have probably been marked by the news or the media and are overly suspicious of people in public. I was only around eight years old<mask> the WTC was hit,<mask> I BARELY remember it.<mask>, thanks to the internet I've been exposed to all of the horrible things that happened afterward. I watched that one journalist's video of him running around the area, filming people running away, the clouds of dust, the towers collapsing, all that fun stuff. I can see<mask> people who were old enough to understand<mask> was happening at the time would feel scared. [NEWLINE] [NEWLINE] <mask> that feeling may have come to slowly fade and go away, I can imagine that a lot of people still remember that day vividly whether they were there or not. A lot of people were hurt and involved with 9/11. Lots of families lost loved ones to it. Plus, this happened in NY, and there's a whole lot of the US's population packed into the east coast.<mask> I wanted to, I could be in NY within the next 2 hours. It happened close to a LOT of us, and that's<mask> scares some of us. [NEWLINE] [NEWLINE] I can't speak for people in other countries, and I can't say that this is a REASONABLE reason to fear terrorism any more than it's reasonable for me to fear flying on planes,<mask> I can say that this is a reason<mask> people are afraid. It's something that happened close to them, and it's hard to shake a bit of paranoia<mask> something<mask> horrible and unimaginable happened in your lifetime, in one of the biggest cities in your
Label encoding: <s>Statistics don't really make you feel safer, nor do they really give you a reason to feel safer. It may be the case for some people, but not all. [NEWLINE] [NEWLINE] Take me for example. I'm afraid of flying on planes. Not terribly afraid, but heights scare me and turbulence of any sort sets me into panic mode. When I tell someone I'm afraid of flying, or that flying makes me nervous, I get the answer "Well, **statistically speaking**, you've got more of a chance of getting in a car wreck than you do a plane crash!" [NEWLINE] [NEWLINE] This is true, however, it does NOT quell my fears any. No matter how slim the chance is, when you get nervous like that the only thing going through your mind is "I might make the news today". [NEWLINE] [NEWLINE] But terrorism (generally speaking) isn't about planes. Terrorism is about something else. People who fear terrorism have probably been marked by the news or the media and are overly suspicious of people in public. I was only around eight years old when the WTC was hit, so I BARELY remember it. However, thanks to the internet I've been exposed to all of the horrible things that happened afterward. I watched that one journalist's video of him running around the area, filming people running away, the clouds of dust, the towers collapsing, all that fun stuff. I can see how people who were old enough to understand what was happening at the time would feel scared. [NEWLINE] [NEWLINE] While that feeling may have come to slowly fade and go away, I can imagine that a lot of people still remember that day vividly whether they were there or not. A lot of people were hurt and involved with 9/11. Lots of families lost loved ones to it. Plus, this happened in NY, and there's a whole lot of the US's population packed into the east coast. If I wanted to, I could be in NY within the next 2 hours. It happened close to a LOT of us, and that's what scares some of us. [NEWLINE] [NEWLINE] I can't speak for people in other countries, and I can't say that this is a REASONABLE reason to fear terrorism any more than it's reasonable for me to fear flying on planes, but I can say that this is a reason why people are afraid. It's something that happened close to them, and it's hard to shake a bit of paranoia when something so horrible and unimaginable happened in your lifetime, in one of the biggest cities in your
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] some brilliant some inane [ENDQ] [NEWLINE] Of course creativity can be used well or poorly,<mask> a definition that includes all of those wouldn't be the best place to find one,<mask> delimiting<mask> something does best is the best way to decide<mask> something really is. Otherwise the definition of football could be throwing one in your yard or being paid millions and the most watched program on television every year, not a very useful definition. [NEWLINE] [NEWLINE] [STARTQ] is fluid intelligence [ENDQ] [NEWLINE] Well yes, that was my point, that creativity and fluid intelligence and higher level problem solving skills aren't interchangeable<mask> they do all the same things and can even look the same,<mask> I purposefully picked the definition of creativity<mask> illustrating that. [NEWLINE] [NEWLINE] [STARTQ] rather odd definition [ENDQ] [NEWLINE] I'm not sure<mask> you noticed<mask><mask> psychologists commonly define fluid intelligence to mean the ability to solve new problems 'independent of acquired knowledge' and I define fluid intelligence<mask> anything in intelligence (people thinking or performing) that isn't the analysis of the'static state' which are things like 'knowledge already known' and the state the person is in to make good choices like '<mask> well they've practiced their creative capacity' then we're really saying the same thing. I didn't say fluid intelligence is solving 'new' problems directly<mask> that's<mask> I implied by getting at the static state. It appears my definition implies I'm saying fluid intelligence is solving old problems or new,<mask> I didn't say it was solving old problems<mask> that was handled by the'state a person is in'<mask><mask> they try to look at problems they have already come across<mask> 'old' and solve them the same old way they aren't utilizing their 'fluid intelligence' in that in the real world everything we come across is different somehow and maybe the difference is trivial to the solution<mask> treating it like new is<mask> counts to even be utilizing your fluid intelligence. [NEWLINE] [NEWLINE] [STARTQ] definition of creativity [ENDQ] [NEWLINE] <mask><mask> many things involved in intellectual quarters share the way they looked<mask> performed and share individual processes they do end up being benefitted by being defined using each other. '<mask> to think' and'meta' and 'creativity' and 'higher level problem solving skills' and 'creativity' and'skepticism' and 'theory' and 'autopilot' and all the rest,<mask> there are tons, are helped by being defined using each other. Not entirely of course,<mask> for delimiting a workable definition<mask> we know<mask> we're talking about in a discussion
Label encoding: <s> [STARTQ] some brilliant some inane [ENDQ] [NEWLINE] Of course creativity can be used well or poorly, so a definition that includes all of those wouldn't be the best place to find one, since delimiting what something does best is the best way to decide what something really is. Otherwise the definition of football could be throwing one in your yard or being paid millions and the most watched program on television every year, not a very useful definition. [NEWLINE] [NEWLINE] [STARTQ] is fluid intelligence [ENDQ] [NEWLINE] Well yes, that was my point, that creativity and fluid intelligence and higher level problem solving skills aren't interchangeable but they do all the same things and can even look the same, so I purposefully picked the definition of creativity while illustrating that. [NEWLINE] [NEWLINE] [STARTQ] rather odd definition [ENDQ] [NEWLINE] I'm not sure if you noticed but if psychologists commonly define fluid intelligence to mean the ability to solve new problems 'independent of acquired knowledge' and I define fluid intelligence as anything in intelligence (people thinking or performing) that isn't the analysis of the'static state' which are things like 'knowledge already known' and the state the person is in to make good choices like'how well they've practiced their creative capacity' then we're really saying the same thing. I didn't say fluid intelligence is solving 'new' problems directly because that's what I implied by getting at the static state. It appears my definition implies I'm saying fluid intelligence is solving old problems or new, but I didn't say it was solving old problems because that was handled by the'state a person is in' because if they try to look at problems they have already come across as 'old' and solve them the same old way they aren't utilizing their 'fluid intelligence' in that in the real world everything we come across is different somehow and maybe the difference is trivial to the solution but treating it like new is what counts to even be utilizing your fluid intelligence. [NEWLINE] [NEWLINE] [STARTQ] definition of creativity [ENDQ] [NEWLINE] Since so many things involved in intellectual quarters share the way they looked when performed and share individual processes they do end up being benefitted by being defined using each other.'How to think' and'meta' and 'creativity' and 'higher level problem solving skills' and 'creativity' and'skepticism' and 'theory' and 'autopilot' and all the rest, because there are tons, are helped by being defined using each other. Not entirely of course, but for delimiting a workable definition so we know what we're talking about in a discussion
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I appreciate the lack of pussyfooting, and I've seen a couple interesting responses from you throughout the thread. [NEWLINE] [NEWLINE] To give you some understanding of<mask> this subject bothers me enough to post about it: I don't become overly concerned with<mask> people think about me<mask> I do like to understand *<mask> * they think the way they do. [NEWLINE] [NEWLINE] <mask> someone thinks I'm an asshole it's not hard to figure out<mask> they think I'm an asshole. [NEWLINE] [NEWLINE] <mask> someone thinks I'm a nice guy it's not hard to figure out<mask> they think I'm a nice guy. [NEWLINE] [NEWLINE] <mask> two extremely intelligent women referred to me<mask> a misogynist on two separate occasions it bothered me<mask> I don't understand<mask> I would be considered a misogynist. I was offered very little explanation, and over the years each time it's mentioned I can't help<mask> wonder "<mask>?" [NEWLINE] [NEWLINE] There's a variety of potential reasons and<mask><mask> I've listed them all,<mask> who knows<mask> someone is going to ask a question that triggers a memory of something else.<mask> I'm looking for is input from the community. I don't *think* I'm a misogynist,<mask> I can't form an accurate conclusion based off the fact that<mask><mask> I'm not<mask> other people have told me that I am. People certainly don't run around saying "You're not a misogynist." [NEWLINE] [NEWLINE] You've made a valid point in stating that no matter<mask> at the conclusion of this thread *someone* is going to be convinced I'm a misogynist.<mask> I'm aiming for is an answer to that question for myself, and the input I get is going to help me come to that conclusion. Someone screaming "YOU'RE A MISOGYNIST" has<mask> little value to me<mask> someone screaming "YOU'RE NOT A MISOGYNIST." The people here in CMV are good for making me think, and that's<mask> I'm looking for. Even<mask> the conclusion is "Well yea, I'm a misogynist<mask> it's<mask> I am overly respectful to women and not<mask> I hate them" it's still an answer that I couldn't come up with on my own. It provides an explanation for *<mask>.* [NEWLINE] [NEWLINE] In summary, I'm looking for more input and all the gender issues floating around on CMV today made it seem like the ideal turbulent environment to surf through<mask> I can get some perspective from both sides and figure out<mask>'s sitting on the raft at the
Label encoding: <s>I appreciate the lack of pussyfooting, and I've seen a couple interesting responses from you throughout the thread. [NEWLINE] [NEWLINE] To give you some understanding of why this subject bothers me enough to post about it: I don't become overly concerned with what people think about me but I do like to understand * why * they think the way they do. [NEWLINE] [NEWLINE] If someone thinks I'm an asshole it's not hard to figure out why they think I'm an asshole. [NEWLINE] [NEWLINE] If someone thinks I'm a nice guy it's not hard to figure out why they think I'm a nice guy. [NEWLINE] [NEWLINE] When two extremely intelligent women referred to me as a misogynist on two separate occasions it bothered me because I don't understand why I would be considered a misogynist. I was offered very little explanation, and over the years each time it's mentioned I can't help but wonder " why?" [NEWLINE] [NEWLINE] There's a variety of potential reasons and I think I've listed them all, though who knows if someone is going to ask a question that triggers a memory of something else. What I'm looking for is input from the community. I don't *think* I'm a misogynist, but I can't form an accurate conclusion based off the fact that I think I'm not when other people have told me that I am. People certainly don't run around saying "You're not a misogynist." [NEWLINE] [NEWLINE] You've made a valid point in stating that no matter what at the conclusion of this thread *someone* is going to be convinced I'm a misogynist. What I'm aiming for is an answer to that question for myself, and the input I get is going to help me come to that conclusion. Someone screaming "YOU'RE A MISOGYNIST" has as little value to me as someone screaming "YOU'RE NOT A MISOGYNIST." The people here in CMV are good for making me think, and that's what I'm looking for. Even if the conclusion is "Well yea, I'm a misogynist but it's because I am overly respectful to women and not because I hate them" it's still an answer that I couldn't come up with on my own. It provides an explanation for * why.* [NEWLINE] [NEWLINE] In summary, I'm looking for more input and all the gender issues floating around on CMV today made it seem like the ideal turbulent environment to surf through so I can get some perspective from both sides and figure out what's sitting on the raft at the
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is like a reverse Fallacy of Composition. [NEWLINE] [NEWLINE] The fallacy of composition states that ''that<mask> one infers that something is true of the whole from the fact that it is true of some part of the whole (or even of every proper part)'' [NEWLINE] [NEWLINE] That makes it a bit complicated - consider the following:<mask> I stand up in the seatings at a match, I can see the game better.<mask>,<mask> everyone stands up, we can all see better. [NEWLINE] [NEWLINE] Or ''<mask> I live on benefits, I don't have to work.<mask>,<mask> we all live on benefits, we don't have to work.'' (Just<mask> it is true, for part does not mean it is true for the whole) [NEWLINE] [NEWLINE] <mask> individually your vote does not count,<mask> collectively a large group of people each individually believe their vote does not count,<mask> is inevitable, the collective vote *does* matter.<mask> Bob comes on this thread and is convinced of your reasoning, he will stop voting.<mask> then Bob's friends ask him<mask> he doesn't vote, he will explain their reasoning. Again some of his friends will not vote. And through this, collectively a large number of votes are removed. Yes, your vote does not matter,<mask> your lack of voting is not unique to you. A larger group of people too can choose not to vote - and<mask> affect the outcome in that way. [NEWLINE] [NEWLINE] Yes,<mask> 99% of the people voted, and 1 person didn't, that 1 person never could have made the difference.<mask> imagine<mask> out of that 99%, 34% wanted to vote for Party Red, rather than Party Green. They realize their vote doesn't matter<mask> the next year they decide not to vote<mask> their vote doesn't matter, they can't affect the outcome anyway. Now we have 65% voters, 35% people who want to vote to something else or feel they don't matter.<mask> the next year, 15% of the population dies/decides not to vote/feels their vote are unworthy, suddenly you have a 15% split. [NEWLINE] [NEWLINE] The reality is that you are not voting on your behalf, you are voting on behalf of a group you want to adequately represent you. And other people are voting for a group they want to represent them. And<mask> you want your group to come close to winning, you need to collectively vote. Mountains are made of pebbles. You are not the mountain,<mask> that doesn't reduce your role in creating the
Label encoding: <s>This is like a reverse Fallacy of Composition. [NEWLINE] [NEWLINE] The fallacy of composition states that ''that when one infers that something is true of the whole from the fact that it is true of some part of the whole (or even of every proper part)'' [NEWLINE] [NEWLINE] That makes it a bit complicated - consider the following: If I stand up in the seatings at a match, I can see the game better. Therefore, if everyone stands up, we can all see better. [NEWLINE] [NEWLINE] Or '' If I live on benefits, I don't have to work. Therefore, if we all live on benefits, we don't have to work.'' (Just because it is true, for part does not mean it is true for the whole) [NEWLINE] [NEWLINE] While individually your vote does not count, when collectively a large group of people each individually believe their vote does not count, as is inevitable, the collective vote *does* matter. If Bob comes on this thread and is convinced of your reasoning, he will stop voting. If then Bob's friends ask him why he doesn't vote, he will explain their reasoning. Again some of his friends will not vote. And through this, collectively a large number of votes are removed. Yes, your vote does not matter, but your lack of voting is not unique to you. A larger group of people too can choose not to vote - and thus affect the outcome in that way. [NEWLINE] [NEWLINE] Yes, if 99% of the people voted, and 1 person didn't, that 1 person never could have made the difference. But imagine if out of that 99%, 34% wanted to vote for Party Red, rather than Party Green. They realize their vote doesn't matter so the next year they decide not to vote since their vote doesn't matter, they can't affect the outcome anyway. Now we have 65% voters, 35% people who want to vote to something else or feel they don't matter. If the next year, 15% of the population dies/decides not to vote/feels their vote are unworthy, suddenly you have a 15% split. [NEWLINE] [NEWLINE] The reality is that you are not voting on your behalf, you are voting on behalf of a group you want to adequately represent you. And other people are voting for a group they want to represent them. And if you want your group to come close to winning, you need to collectively vote. Mountains are made of pebbles. You are not the mountain, but that doesn't reduce your role in creating the
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>edit: Thanks to some discussion with posters and links to recent changes in [NIH policy]( [URL] ) and [U.S. policy]( [URL] /) I've conceded that my "upon first publication point" may be too dramatic.<mask><mask> making the published research open access is still important,<mask> have conceded that it should be done after a reasonable time frame. NIH says 1 year, the U.S. government says 6 months. [NEWLINE] [NEWLINE] edit:<mask> /u/Maxzines (and others) pointed out [here]( [URL] ) I didn't make it clear that I meant private projects and research done at universities or colleges. Defense departments should be able to do research in secret. [NEWLINE] [NEWLINE] I am speaking purely from an American perspective<mask> I do genuinely feel that this issue should extend to all countries. [NEWLINE] [NEWLINE] My argument is<mask> follows: [NEWLINE] 1. Acceptance of funds from any single person or group makes that entity an investor in that research. [NEWLINE] 2. An investor in research should have access to the end results of that research. [NEWLINE] 3.<mask> a researcher accepts government grants to fund their research then that government, and by extension its citizens, have become an investor and should have free access to the published work. [NEWLINE] [NEWLINE] To clarify I am<mask> saying that acceptance of any amount of government funding should lead to free access to the published work,<mask><mask><mask> it is $1 or $100,000. Change my view,<mask> is wrong about this line of thought? [NEWLINE] [NEWLINE] edit: To further clarify I am talking purely about publication of scientific research. I am in a sense arguing for the legislation of a requirement to publish government funded research in an [open access journal]( [URL] ). [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!
Label encoding: <s>edit: Thanks to some discussion with posters and links to recent changes in [NIH policy]( [URL] ) and [U.S. policy]( [URL] /) I've conceded that my "upon first publication point" may be too dramatic. I think making the published research open access is still important, but have conceded that it should be done after a reasonable time frame. NIH says 1 year, the U.S. government says 6 months. [NEWLINE] [NEWLINE] edit: As /u/Maxzines (and others) pointed out [here]( [URL] ) I didn't make it clear that I meant private projects and research done at universities or colleges. Defense departments should be able to do research in secret. [NEWLINE] [NEWLINE] I am speaking purely from an American perspective but I do genuinely feel that this issue should extend to all countries. [NEWLINE] [NEWLINE] My argument is as follows: [NEWLINE] 1. Acceptance of funds from any single person or group makes that entity an investor in that research. [NEWLINE] 2. An investor in research should have access to the end results of that research. [NEWLINE] 3. If a researcher accepts government grants to fund their research then that government, and by extension its citizens, have become an investor and should have free access to the published work. [NEWLINE] [NEWLINE] To clarify I am indeed saying that acceptance of any amount of government funding should lead to free access to the published work, regardless of if it is $1 or $100,000. Change my view, what is wrong about this line of thought? [NEWLINE] [NEWLINE] edit: To further clarify I am talking purely about publication of scientific research. I am in a sense arguing for the legislation of a requirement to publish government funded research in an [open access journal]( [URL] ). [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0412, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0464, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Technically, the U.S. military could be classified<mask> a terrorist organization.<mask> the connotative understanding of terrorism has had one, big component that conflicts with your view more or less<mask> its birth: terrorists kill innocent people and use the resulting fear<mask> a means of political and/or social change.<mask>, unless you can prove that the U.S. did these attacks intentionally for the explicit purpose of spreading terror, it will conflict with the connotative definition of the word "terrorist group." And<mask> the denotative definition of the word is currently in a state of dispute, the connotative definition is all we have. [NEWLINE] [NEWLINE] Interestingly enough, the term actually *did* originate for a government. The French government, to be exact, during the Reign of Terror. The term quickly became a term of abuse, and about 90 years later, was a self-described term of Russian terrorist organizations that largely resemble organizations such<mask> Al-Quaeda in their tactics and revenge ideology. By the time Franz Ferdinand was shot, this connotation of the word was<mask> largely agreed upon that newspapers all over the world called Gabril Princip and his accomplices terrorists.<mask> by this time, the term has gained another big component: terrorists are the little guys trying to throw off some sort of oppression (real or perceived). This term largely stuck, and was later applied to organizations like the IRA, the Tamil Tigers, and<mask> on. [NEWLINE] [NEWLINE] Now, I'm sure that many of the Bosnians considered Austria to be a terrorist organization just<mask> you consider the U.S. military to be a terrorist organization.<mask> I hope I at least opened your eyes to the fact that a word's denotative definition is not its sole source of meaning. Sometimes the connotative definition holds even more solid ground. The countries of the world *still* argue over<mask> is and isn't considered terrorism,<mask> the connotation really hasn't changed that much. It is rooted in a history stretching back over 100 years, and<mask> there have always been people opposed to the thinking of the majority, the majority still rules<mask><mask><mask> word connotation goes. [NEWLINE] [NEWLINE] Sources: [NEWLINE] [NEWLINE] The world cannot come up with a consistent denotative definition of terrorism,<mask> exemplified by the varying definitions in terrorism legislation of different countries: [URL] [NEWLINE] [NEWLINE] The origin of the term (make sure to look into the different sources posted here, both for accuracy and some interesting reads): [URL] [NEWLINE] [NEWLINE] And another [URL] </s>
Label encoding: <s>Technically, the U.S. military could be classified as a terrorist organization. But the connotative understanding of terrorism has had one, big component that conflicts with your view more or less since its birth: terrorists kill innocent people and use the resulting fear as a means of political and/or social change. So, unless you can prove that the U.S. did these attacks intentionally for the explicit purpose of spreading terror, it will conflict with the connotative definition of the word "terrorist group." And since the denotative definition of the word is currently in a state of dispute, the connotative definition is all we have. [NEWLINE] [NEWLINE] Interestingly enough, the term actually *did* originate for a government. The French government, to be exact, during the Reign of Terror. The term quickly became a term of abuse, and about 90 years later, was a self-described term of Russian terrorist organizations that largely resemble organizations such as Al-Quaeda in their tactics and revenge ideology. By the time Franz Ferdinand was shot, this connotation of the word was so largely agreed upon that newspapers all over the world called Gabril Princip and his accomplices terrorists. So by this time, the term has gained another big component: terrorists are the little guys trying to throw off some sort of oppression (real or perceived). This term largely stuck, and was later applied to organizations like the IRA, the Tamil Tigers, and so on. [NEWLINE] [NEWLINE] Now, I'm sure that many of the Bosnians considered Austria to be a terrorist organization just as you consider the U.S. military to be a terrorist organization. But I hope I at least opened your eyes to the fact that a word's denotative definition is not its sole source of meaning. Sometimes the connotative definition holds even more solid ground. The countries of the world *still* argue over what is and isn't considered terrorism, but the connotation really hasn't changed that much. It is rooted in a history stretching back over 100 years, and while there have always been people opposed to the thinking of the majority, the majority still rules as far as word connotation goes. [NEWLINE] [NEWLINE] Sources: [NEWLINE] [NEWLINE] The world cannot come up with a consistent denotative definition of terrorism, as exemplified by the varying definitions in terrorism legislation of different countries: [URL] [NEWLINE] [NEWLINE] The origin of the term (make sure to look into the different sources posted here, both for accuracy and some interesting reads): [URL] [NEWLINE] [NEWLINE] And another [URL] </s>
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>&gt; That is not the purpose of academia, it is a place to learn things or to discover new things, not to learn<mask> to think (whatever it means to think correctly.) My view is that it is not worth it for the public to be paying to make progress in these fields. [NEWLINE] [NEWLINE] Aha... and here is the heart of the matter. [NEWLINE] [NEWLINE] I suppose this opinion is the inevitable eventuality for someone who views education primarily<mask> a means of consuming and retaining facts.  The problem, put simply, is that knowing everything in any given field only allows for<mask> much room for progress - and<mask><mask> can often be detrimental<mask> up against a problem that defies known parameters.  The easiest example is probably the theory of special relativity; plenty of learned physicists were nearly stopped in their tracks by the inherent contradictions brought about by the aether and the constancy of time. To finally resolve those inconsistencies, it took someone not only knowledgeable,<mask> curious enough to treat them<mask><mask> they were - assumptions, not facts - and throw them out the window. [NEWLINE] [NEWLINE] It is precisely this kind of thought which /u/yamsx1 refers to<mask> "thinking correctly" - the ability to approach, examine, and solve arbitrary problems not just logically,<mask><mask> creatively. <mask> to think, not just<mask> to know.  This is the skill of raw thought, and should be treated<mask> mathematics is - not only<mask> a worthy pursuit unto itself,<mask><mask> a fundamental building block of most other fields.  Given this, we can properly understand and appreciate fields like theology and philosophy for<mask> they are: logical and creative thought<mask> directly applied to God and society, respectively. [NEWLINE] [NEWLINE] Even<mask> "progress" in these applied fields in modern times is mostly restricted to the ivory towers, the fundamentals are and remain valuable skills for all to learn,<mask><mask> whether they are pursued for their own ends. <mask>?  For the same reason we all learn some level of mathematics; even<mask> you do not pursue a career which uses it, directly or indirectly, the fundamentals are used<mask> broadly in everyday life that it would be flatly irresponsible to *not* teach them.  It is no accident that the very first material taught to any student of philosophy is a crash course on logical relationships and fallacies. <mask> improved would, say, our political debates be<mask> everyone had learned<mask> to reason their way to the end of a simple syllogism right alongside their multiplication tables?</s>
Label encoding: <s>&gt; That is not the purpose of academia, it is a place to learn things or to discover new things, not to learn how to think (whatever it means to think correctly.) My view is that it is not worth it for the public to be paying to make progress in these fields. [NEWLINE] [NEWLINE] Aha... and here is the heart of the matter. [NEWLINE] [NEWLINE] I suppose this opinion is the inevitable eventuality for someone who views education primarily as a means of consuming and retaining facts.  The problem, put simply, is that knowing everything in any given field only allows for so much room for progress - and in fact can often be detrimental when up against a problem that defies known parameters.  The easiest example is probably the theory of special relativity; plenty of learned physicists were nearly stopped in their tracks by the inherent contradictions brought about by the aether and the constancy of time. To finally resolve those inconsistencies, it took someone not only knowledgeable, but curious enough to treat them as what they were - assumptions, not facts - and throw them out the window. [NEWLINE] [NEWLINE] It is precisely this kind of thought which /u/yamsx1 refers to as "thinking correctly" - the ability to approach, examine, and solve arbitrary problems not just logically, but also creatively.  How to think, not just what to know.  This is the skill of raw thought, and should be treated as mathematics is - not only as a worthy pursuit unto itself, but as a fundamental building block of most other fields.  Given this, we can properly understand and appreciate fields like theology and philosophy for what they are: logical and creative thought as directly applied to God and society, respectively. [NEWLINE] [NEWLINE] Even if "progress" in these applied fields in modern times is mostly restricted to the ivory towers, the fundamentals are and remain valuable skills for all to learn, regardless of whether they are pursued for their own ends.  Why?  For the same reason we all learn some level of mathematics; even if you do not pursue a career which uses it, directly or indirectly, the fundamentals are used so broadly in everyday life that it would be flatly irresponsible to *not* teach them.  It is no accident that the very first material taught to any student of philosophy is a crash course on logical relationships and fallacies.  How improved would, say, our political debates be if everyone had learned how to reason their way to the end of a simple syllogism right alongside their multiplication tables?</s>
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0471, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0348, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**West Germany v Austria (1982 FIFA World Cup)**]( [URL] %20Germany%20v%20Austria%20(1982%20FIFA%20World%20Cup\)): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] __West Germany versus Austria__ was a [1982 FIFA World Cup]( [URL] ) game played at the [El Molinón stadium]( [URL] %C3%B3n), [Gijón]( [URL] %C3%B3n), [Spain]( [URL] ), on 25 June 1982. The match was the last game of the first-round [Group 2]( [URL] #Group_2), with [Algeria]( [URL] ) and [Chile]( [URL] ) having played the day before. With the outcome of that match already decided, a win by one or two goals for [West Germany]( [URL] ) would result in both them and [Austria]( [URL] ) qualifying at the expense of Algeria, who had beaten West Germany in the first game. After 10 minutes, West Germany took the lead. Thereafter, neither team scored, and few scoring chances were created, along with much own-half passing and few tackles: with both sets of players flamboyantly missing with no clear attempt to guide the ball whenever they shot on goal. [ENDQ] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Germany ^national ^football ^team]( [URL] ) ^| [^Italy ^national ^football ^team]( [URL] ) ^| [^1982 ^FIFA ^World ^Cup]( [URL] ) ^| [^Algeria ^national ^football ^team]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+ciw4trm) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+ciw4trm)^. ^Will ^<mask> ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Label encoding: <s>#####&amp;#009; [NEWLINE] [NEWLINE] ######&amp;#009; [NEWLINE] [NEWLINE] ####&amp;#009; [NEWLINE] [**West Germany v Austria (1982 FIFA World Cup)**]( [URL] %20Germany%20v%20Austria%20(1982%20FIFA%20World%20Cup\)): [](#sfw) [NEWLINE] [NEWLINE] --- [NEWLINE] [NEWLINE] [STARTQ] __West Germany versus Austria__ was a [1982 FIFA World Cup]( [URL] ) game played at the [El Molinón stadium]( [URL] %C3%B3n), [Gijón]( [URL] %C3%B3n), [Spain]( [URL] ), on 25 June 1982. The match was the last game of the first-round [Group 2]( [URL] #Group_2), with [Algeria]( [URL] ) and [Chile]( [URL] ) having played the day before. With the outcome of that match already decided, a win by one or two goals for [West Germany]( [URL] ) would result in both them and [Austria]( [URL] ) qualifying at the expense of Algeria, who had beaten West Germany in the first game. After 10 minutes, West Germany took the lead. Thereafter, neither team scored, and few scoring chances were created, along with much own-half passing and few tackles: with both sets of players flamboyantly missing with no clear attempt to guide the ball whenever they shot on goal. [ENDQ] [NEWLINE] [STARTQ] [ENDQ] [NEWLINE] --- [NEWLINE] [NEWLINE] ^Interesting: [^Germany ^national ^football ^team]( [URL] ) ^| [^Italy ^national ^football ^team]( [URL] ) ^| [^1982 ^FIFA ^World ^Cup]( [URL] ) ^| [^Algeria ^national ^football ^team]( [URL] ) [NEWLINE] [NEWLINE] ^Parent ^commenter ^can [^toggle ^NSFW]( [URL] ;subject=AutoWikibot NSFW toggle&amp;message=%2Btoggle-nsfw+ciw4trm) ^or[](#or) [^delete]( [URL] ;subject=AutoWikibot Deletion&amp;message=%2Bdelete+ciw4trm)^. ^Will ^ also ^delete ^on ^comment ^score ^of ^-1 ^or ^less. ^| [^(FAQs)]( [URL] ) ^| [^Mods]( [URL] /) ^| [^Magic ^Words]( [URL] /)</s>
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Actually yes. The quality of life (and median income) in the developing world has increased massively in the past decade(s)<mask> of technological advancements, political stability, and the influence of the West. [ENDQ] [NEWLINE] Sure. This is all true, and very much a good thing.<mask> things are going to get better for the developing world until it is close to the developed world.<mask> happens<mask> they reach close to parity?<mask> the median income trend is a good sign then after that point the actual amount of improvement will  be small. [NEWLINE] [NEWLINE] [STARTQ] It feels weird for someone to actually recognize they're wrong change their opinion right before your eyes right? [ENDQ] [NEWLINE] Yes,<mask> quite refreshing. It is depressingly rare even in this subreddit. [NEWLINE] [NEWLINE] [STARTQ] History is written by the victors and Westerners // Americans have won it all. With that<mask>, we have an obligation to preserve other cultures, even more<mask>, incorporate them into our own [ENDQ] [NEWLINE] Sure we have such an obligation. Are we doing it? Not really. [NEWLINE] [NEWLINE] [STARTQ] Human exceptionalism is actually 100% rational,<mask><mask> I'm the only person that believes it - it just seems irrational. [ENDQ] [NEWLINE] I don't think you are the only person.<mask> anything, human exceptionalism was historically the default view. It is only in the last few centuries that the idea that humans might not be special has been common. Heck, many religions even believe that we're the pinnacle of creation. [NEWLINE] [NEWLINE] [STARTQ] &gt; every species faced with the threat of the Great Filter had some amount of "irrational [species here] exceptionalism" and that doesn't seem to have helped. [ENDQ] [NEWLINE] [STARTQ] We've seen species filtered by the Great Filter. They go extinct on our planet every day. They lost, we won. Look at a panda some time - they look<mask> defeated. It's almost<mask><mask> they've been imprisoned by the stronger species and refuse to continue reproducing. Weird right? [ENDQ] [NEWLINE] The Filter is specifically about the lack of species on a interstellar scale. It doesn't make sense to talk about the filtration in that sense of other species here except in<mask> far<mask> having a single highly intelligent species on a planet probably does cut off development of civilization by other smart species on the planet.<mask> corvids and elephants are now out of luck<mask> we got there first. The point<mask> is about the species that elsewhere in the universe tried to make civilizations and failed. </s>
Label encoding: <s> [STARTQ] Actually yes. The quality of life (and median income) in the developing world has increased massively in the past decade(s) because of technological advancements, political stability, and the influence of the West. [ENDQ] [NEWLINE] Sure. This is all true, and very much a good thing. So things are going to get better for the developing world until it is close to the developed world. What happens when they reach close to parity? If the median income trend is a good sign then after that point the actual amount of improvement will  be small. [NEWLINE] [NEWLINE] [STARTQ] It feels weird for someone to actually recognize they're wrong change their opinion right before your eyes right? [ENDQ] [NEWLINE] Yes, but quite refreshing. It is depressingly rare even in this subreddit. [NEWLINE] [NEWLINE] [STARTQ] History is written by the victors and Westerners // Americans have won it all. With that though, we have an obligation to preserve other cultures, even more so, incorporate them into our own [ENDQ] [NEWLINE] Sure we have such an obligation. Are we doing it? Not really. [NEWLINE] [NEWLINE] [STARTQ] Human exceptionalism is actually 100% rational, but since I'm the only person that believes it - it just seems irrational. [ENDQ] [NEWLINE] I don't think you are the only person. If anything, human exceptionalism was historically the default view. It is only in the last few centuries that the idea that humans might not be special has been common. Heck, many religions even believe that we're the pinnacle of creation. [NEWLINE] [NEWLINE] [STARTQ] &gt; every species faced with the threat of the Great Filter had some amount of "irrational [species here] exceptionalism" and that doesn't seem to have helped. [ENDQ] [NEWLINE] [STARTQ] We've seen species filtered by the Great Filter. They go extinct on our planet every day. They lost, we won. Look at a panda some time - they look so defeated. It's almost as if they've been imprisoned by the stronger species and refuse to continue reproducing. Weird right? [ENDQ] [NEWLINE] The Filter is specifically about the lack of species on a interstellar scale. It doesn't make sense to talk about the filtration in that sense of other species here except in so far as having a single highly intelligent species on a planet probably does cut off development of civilization by other smart species on the planet. So corvids and elephants are now out of luck because we got there first. The point though is about the species that elsewhere in the universe tried to make civilizations and failed. </s>
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0555, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0524, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think the need to protect children, even those children that are not our own, is a base instinct of humans whereas we feel less need to protect other adults. [NEWLINE] [NEWLINE] It really should not come<mask> any surprise then that there is a more visceral reaction to child rape than there is for adult-adult rape.<mask> the later happens, it is a crime. Crimes happen. People get mugged or beaten or even murdered for no reason. We have a judicial system which attempts to justly and humanely deal with this. <mask>, we<mask> humans feel strongly<mask> these same things happen to children. It's not just a crime at this point. It's a crime against a child, the future of our species.<mask> you rape a child, you may<mask> well not even be human. Hell, even murderers in prison treat child rapists (and to a large degree adult-adult rapists<mask> well)<mask><mask> they are not even human,<mask> an animal. [NEWLINE] [NEWLINE] Now,<mask> for rape fantasies I don't know<mask> that's about,<mask> many women have them and from<mask> I have gathered, it is<mask> rape fantasies are not rape at all<mask> there is consent and ultimately the woman has control of the situation,<mask> the primal sense of danger may linger which leads to a hormonal cocktail of adrenaline and flight-or-fight hormones<mask> well<mask> normal sex hormones.<mask><mask><mask> one partner dressing up like a child and mixing in the rape fantasy,<mask><mask> that's bending the spoon quite profusely,<mask> I would still say that it's two consenting adults. Maybe one of them likes age-play, it's not terribly uncommon and the thinking is that at the end of the day, everyone puts their pants on and goes on with their adult lives. [NEWLINE] [NEWLINE] <mask>, "gangs of people" is not really society<mask> a whole.<mask> you are trying to convince me that society<mask> a whole thinks rape is ok, then you would be hard pressed convincing me of that. The closest I can think of is middle eastern countries in backwoods fundamentalist areas who think<mask> a man rapes a woman that she deserves it for being alone, Ok yeah fuck those people. To be fair, they<mask> think apostasy deserves death<mask> it's obviously just a violent culture. India<mask> has its problems,<mask> most people even there think rape is wrong, just not everyone and that small violent and ignorant group of people should be put in prison. </s>
Label encoding: <s>I think the need to protect children, even those children that are not our own, is a base instinct of humans whereas we feel less need to protect other adults. [NEWLINE] [NEWLINE] It really should not come as any surprise then that there is a more visceral reaction to child rape than there is for adult-adult rape. When the later happens, it is a crime. Crimes happen. People get mugged or beaten or even murdered for no reason. We have a judicial system which attempts to justly and humanely deal with this.  However, we as humans feel strongly when these same things happen to children. It's not just a crime at this point. It's a crime against a child, the future of our species. If you rape a child, you may as well not even be human. Hell, even murderers in prison treat child rapists (and to a large degree adult-adult rapists as well) as if they are not even human, but an animal. [NEWLINE] [NEWLINE] Now, as for rape fantasies I don't know what that's about, but many women have them and from what I have gathered, it is because rape fantasies are not rape at all because there is consent and ultimately the woman has control of the situation, but the primal sense of danger may linger which leads to a hormonal cocktail of adrenaline and flight-or-fight hormones as well as normal sex hormones. As far as one partner dressing up like a child and mixing in the rape fantasy, I agree that's bending the spoon quite profusely, but I would still say that it's two consenting adults. Maybe one of them likes age-play, it's not terribly uncommon and the thinking is that at the end of the day, everyone puts their pants on and goes on with their adult lives. [NEWLINE] [NEWLINE] Lastly, "gangs of people" is not really society as a whole. If you are trying to convince me that society as a whole thinks rape is ok, then you would be hard pressed convincing me of that. The closest I can think of is middle eastern countries in backwoods fundamentalist areas who think if a man rapes a woman that she deserves it for being alone, Ok yeah fuck those people. To be fair, they also think apostasy deserves death so it's obviously just a violent culture. India also has its problems, but most people even there think rape is wrong, just not everyone and that small violent and ignorant group of people should be put in prison. </s>
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1. This is one of the few objections I can kind of see the logic of,<mask><mask> many foods<mask> soup dumplings fit into both categories of "can't be pierced" and "can't be picked up with fingers"? [NEWLINE] [NEWLINE] 2. Is that something you do? Just carry around chopsticks in your pocket? I'm not even sure that this is true. chopsticks are longer,<mask> they don't always fit into the same container your food is in. [NEWLINE] [NEWLINE] 3. This is undeniable. OP made the classic CMV mistake of overstating their view. [NEWLINE] [NEWLINE] 4. Are you saying you hold one chopstick in each hand, or four altogether?<mask> it's the former, I fail to see<mask> that's more efficient,<mask> the latter, you could do that just<mask> easily<mask> not more<mask> with two forks or two knives. [NEWLINE] [NEWLINE] 5. You know<mask> makes good tongs? Tongs. This one seems outside the scope of the discussion.<mask> you asked me which one makes a better makeshift arrow for a miniature bow, I would say chopsticks,<mask> that doesn't make me more likely to want a pair in my kitchen. [NEWLINE] [NEWLINE] 6. I guess I've never scrambled eggs with chopsticks,<mask> I find this hard to believe.<mask> I question<mask> you would eat scrambled eggs off of an easily-scratched dish<mask> not care enough to just use a whisk. [NEWLINE] [NEWLINE] 7.<mask> would I want to eat a single grain of rice? The fork is better for scooping the last few grains from an empty dish,<mask><mask> it can't discriminate between specific grains<mask> the dish is full. [NEWLINE] <mask>, the amount of skill and strength it would take to lift an entire steak with chopsticks highlights<mask> they're a terrible implement for that purpose. I don't want to gnaw a chunk of meat off a steak suspended in midair by my implausibly strong forearms, I just want to cut off a piece and take a bite. [NEWLINE] [NEWLINE] 8. Again, outside the scope. We're talking about whether a pair of finished chopsticks is a better eating tool than a set of finished silverware, not which one is easier to craft out of whatever's sitting within 3 feet. [NEWLINE] [NEWLINE] 9. Shouldn't mastery be a point against the chopsticks? I can do both, and let me tell you, the fork and knife are way easier.</s>
Label encoding: <s>1. This is one of the few objections I can kind of see the logic of, but how many foods besides soup dumplings fit into both categories of "can't be pierced" and "can't be picked up with fingers"? [NEWLINE] [NEWLINE] 2. Is that something you do? Just carry around chopsticks in your pocket? I'm not even sure that this is true. chopsticks are longer, so they don't always fit into the same container your food is in. [NEWLINE] [NEWLINE] 3. This is undeniable. OP made the classic CMV mistake of overstating their view. [NEWLINE] [NEWLINE] 4. Are you saying you hold one chopstick in each hand, or four altogether? If it's the former, I fail to see how that's more efficient, if the latter, you could do that just as easily if not more so with two forks or two knives. [NEWLINE] [NEWLINE] 5. You know what makes good tongs? Tongs. This one seems outside the scope of the discussion. If you asked me which one makes a better makeshift arrow for a miniature bow, I would say chopsticks, but that doesn't make me more likely to want a pair in my kitchen. [NEWLINE] [NEWLINE] 6. I guess I've never scrambled eggs with chopsticks, but I find this hard to believe. Also I question why you would eat scrambled eggs off of an easily-scratched dish but not care enough to just use a whisk. [NEWLINE] [NEWLINE] 7. Why would I want to eat a single grain of rice? The fork is better for scooping the last few grains from an empty dish, even though it can't discriminate between specific grains when the dish is full. [NEWLINE] Also, the amount of skill and strength it would take to lift an entire steak with chopsticks highlights how they're a terrible implement for that purpose. I don't want to gnaw a chunk of meat off a steak suspended in midair by my implausibly strong forearms, I just want to cut off a piece and take a bite. [NEWLINE] [NEWLINE] 8. Again, outside the scope. We're talking about whether a pair of finished chopsticks is a better eating tool than a set of finished silverware, not which one is easier to craft out of whatever's sitting within 3 feet. [NEWLINE] [NEWLINE] 9. Shouldn't mastery be a point against the chopsticks? I can do both, and let me tell you, the fork and knife are way easier.</s>
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0470, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask><mask> there is probably a great argument that you'll never break even on the net energy poured into the production of a modern EV,<mask> I want to ignore that aspect. Let's say an EV just materializes, dropped off by aliens, perhaps. With our current means of putting energy onto the electric grid, one may very well trade a gasoline powered vehicle for a coal powered one by choosing an EV, and that doesn't make much sense to me. [ENDQ] [NEWLINE] Quantitative lifecycle analyses have been done into the lifecycle energy costs of EVs, and come out in favour of EVs.  The [UCLA]( [URL].pdf), for example, found that a conventional car uses 858,154 megajoules of energy over its lifetime,<mask> an electric car uses 506,988 MJ.  That's factoring in manufacturing of the car and battery, transportation, operation, and disposal,<mask> well<mask> replacement of 50% of the battery pack over the car's lifetime.  All things considered, electric cars come out to using 40% less energy over their lifetime.  This is<mask> the electric motor is inherently more efficient than ICEs even after factoring in transmission losses, and there are a lot of hidden energy costs to turning petroleum into gasoline that are not immediately apparent to the driver. [NEWLINE] [NEWLINE] <mask> for emissions, they found that an electric car pollutes about 30% less over its lifetime based on the United States' aggregate power generation source mix, with that number changing to 50% for a relatively clean grid such<mask> California's.  Even with coal accounting for about 40% of the US aggregate generation mix, the relatively low pollution of the other 60% mean that electric cars come out ahead. [NEWLINE] [NEWLINE] The fact is, electric cars are already cleaner and use less energy over their lifetime, and will get even cleaner<mask> the percentage of coal in the generation grid continues to decline.  More to the point, electric cars can draw on a wide variety of energy sources - no matter<mask> hard you try, you'll never get a car to accept coal or nuclear power.  Our cheap oil has run out, and it's time to decouple oil from our transportation infrastructure.  Electric cars let us do just that, and by divorcing ourselves from an increasingly scarce commodity, it will ensure the continuity of automotive transportation, or at least mitigate the shock<mask> we transition to a less transport-intensive society.</s>
Label encoding: <s> [STARTQ] I think there is probably a great argument that you'll never break even on the net energy poured into the production of a modern EV, but I want to ignore that aspect. Let's say an EV just materializes, dropped off by aliens, perhaps. With our current means of putting energy onto the electric grid, one may very well trade a gasoline powered vehicle for a coal powered one by choosing an EV, and that doesn't make much sense to me. [ENDQ] [NEWLINE] Quantitative lifecycle analyses have been done into the lifecycle energy costs of EVs, and come out in favour of EVs.  The [UCLA]( [URL].pdf), for example, found that a conventional car uses 858,154 megajoules of energy over its lifetime, while an electric car uses 506,988 MJ.  That's factoring in manufacturing of the car and battery, transportation, operation, and disposal, as well as replacement of 50% of the battery pack over the car's lifetime.  All things considered, electric cars come out to using 40% less energy over their lifetime.  This is because the electric motor is inherently more efficient than ICEs even after factoring in transmission losses, and there are a lot of hidden energy costs to turning petroleum into gasoline that are not immediately apparent to the driver. [NEWLINE] [NEWLINE] As for emissions, they found that an electric car pollutes about 30% less over its lifetime based on the United States' aggregate power generation source mix, with that number changing to 50% for a relatively clean grid such as California's.  Even with coal accounting for about 40% of the US aggregate generation mix, the relatively low pollution of the other 60% mean that electric cars come out ahead. [NEWLINE] [NEWLINE] The fact is, electric cars are already cleaner and use less energy over their lifetime, and will get even cleaner as the percentage of coal in the generation grid continues to decline.  More to the point, electric cars can draw on a wide variety of energy sources - no matter how hard you try, you'll never get a car to accept coal or nuclear power.  Our cheap oil has run out, and it's time to decouple oil from our transportation infrastructure.  Electric cars let us do just that, and by divorcing ourselves from an increasingly scarce commodity, it will ensure the continuity of automotive transportation, or at least mitigate the shock as we transition to a less transport-intensive society.</s>
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While I will admit to not knowing much about the movement, I can't help<mask> feel like the paper you linked to doesn't support your interpretation of it. [NEWLINE] [NEWLINE] It does say there was an anti-feminist backlash in the 1970s, which spawned *an* anti feminist movement,<mask> I'm not entirely convinced that that carries to the modern one. That said, I have done no formal studies in gender or social movements; and<mask> such am going to assume you are right for the time being. [NEWLINE] [NEWLINE] Even<mask> this movement is the historical precursor of the modern one, I don't think that the original motivation of necessity dictates the goals 40 years down the line, we are literally talking about generations. I don't think anyone would<mask><mask> feminism, or any other social/human rights movement has remained completely static in its goals or means,<mask> should we expect a men's rights one to? [NEWLINE] [NEWLINE] <mask> for the quote, after reading the excerpt it appears pretty radical,<mask> in context is it less severe. It is sourced from an article that takes partiarchy theory (<mask> I understand it the idea that the legal and social system is set up to abuse women for the benefit of men) and applies the same set of logic to a feminst-inspired mirror image (<mask> the quote comes from)... Evidently the author believes that this "mirror image" is the reality of the situation. [NEWLINE] [NEWLINE] That's not to say you are wrong about avoiceformen, or at least the article's author, they do explicitly state that they see feminism<mask> a driver of inequality in society and lay out a mission statement to have society "acknowledge that *Feminism*[sic] is the reason these [discriminatory] laws exist." [NEWLINE] [NEWLINE] <mask> it appears that there is a sizable body in the subreddit in question (<mask> not the movement<mask> a whole) that blames feminism and partiarchy theory for their woes, and TBH I'm not sure they are *entirely* incorrect. I'm still not finding anything that explicitly hates on women<mask> a gender,<mask>. [NEWLINE] [NEWLINE] I have a followup question for you,<mask> you seem to have a better grasp of the gender-focused rights movements: Is there a movement that addresses the real issues that males and men face that you would say are good or that you support?<mask> not:<mask> do you think that is?</s>
Label encoding: <s>While I will admit to not knowing much about the movement, I can't help but feel like the paper you linked to doesn't support your interpretation of it. [NEWLINE] [NEWLINE] It does say there was an anti-feminist backlash in the 1970s, which spawned *an* anti feminist movement, but I'm not entirely convinced that that carries to the modern one. That said, I have done no formal studies in gender or social movements; and as such am going to assume you are right for the time being. [NEWLINE] [NEWLINE] Even if this movement is the historical precursor of the modern one, I don't think that the original motivation of necessity dictates the goals 40 years down the line, we are literally talking about generations. I don't think anyone would argue that feminism, or any other social/human rights movement has remained completely static in its goals or means, why should we expect a men's rights one to? [NEWLINE] [NEWLINE] As for the quote, after reading the excerpt it appears pretty radical, but in context is it less severe. It is sourced from an article that takes partiarchy theory ( as I understand it the idea that the legal and social system is set up to abuse women for the benefit of men) and applies the same set of logic to a feminst-inspired mirror image ( where the quote comes from)... Evidently the author believes that this "mirror image" is the reality of the situation. [NEWLINE] [NEWLINE] That's not to say you are wrong about avoiceformen, or at least the article's author, they do explicitly state that they see feminism as a driver of inequality in society and lay out a mission statement to have society "acknowledge that *Feminism*[sic] is the reason these [discriminatory] laws exist." [NEWLINE] [NEWLINE] So it appears that there is a sizable body in the subreddit in question ( if not the movement as a whole) that blames feminism and partiarchy theory for their woes, and TBH I'm not sure they are *entirely* incorrect. I'm still not finding anything that explicitly hates on women as a gender, though. [NEWLINE] [NEWLINE] I have a followup question for you, as you seem to have a better grasp of the gender-focused rights movements: Is there a movement that addresses the real issues that males and men face that you would say are good or that you support? If not: why do you think that is?</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your opening assertion is patently false.  Josephus and Tacitus both reference the death of Jesus and neither man's writing is included in the Bible, of course.  Your second assertion does not square even remotely with current scholarship on the gospels.  It's only mythicists who think we can trace all the gospels back to Mark.  At bare minimum you have Mark and Q.  Your remaining assertions are not terribly substantive. [NEWLINE] [NEWLINE] The bottom line is that we have multiple, largely independent sources referencing Jesus within a century of his death.  That is a reasonable body of evidence considering the status of the man (<mask> another Jewish troublemaker in a backwater of the Roman Empire) and the general lack of sources to survive from the period in general. [NEWLINE] [NEWLINE] **Christian Sources.** You have the gospels (and the sources lying behind the gospels like Q).  You have Paul (Galatians 1 being the strongest link, in which he mentions meeting with Peter and James in the mid 30s).  Both Paul and the extant gospels preserve earlier material, e.g. the hymn in Philippians or Matthew 12:1-8, likely surviving elements of earlier oral traditions. [NEWLINE] [NEWLINE] **Jewish Sources** Josephus in his Antiquities mentions Jesus twice, in books 18 and 20.  He<mask> mentions John the Baptist in Book 18.  The references to John and the reference to Jesus in Book 20 are considered authentic by the overwhelming majority of Josephus scholars.  The reference to Jesus in Book 18, the Testimonium, is obviously an interpolation<mask> there is widespread agreement among scholars that the passage contains an authentic core. [NEWLINE] [NEWLINE] **Roman Sources.** You have Tacitus, Suetonius and Pliny the Younger.  Of those three, Tacitus is the strongest and clearest reference. [NEWLINE] [NEWLINE] For the mythicist case to stand, you need to explain away all of those references.  You<mask> need to explain<mask> somebody in the first century would invent a supposed Jewish messiah who acts nothing like<mask> every Jew knew the messiah was supposed to be (hint: it wasn't dying in shame on the cross).  You need to explain<mask> Matthew and Luke take great pains to explain<mask> a man from Nazareth was actually from Bethlehem,<mask><mask> the man were a fiction they could simply have placed him in the right city to begin with.</s>
Label encoding: <s>Your opening assertion is patently false.  Josephus and Tacitus both reference the death of Jesus and neither man's writing is included in the Bible, of course.  Your second assertion does not square even remotely with current scholarship on the gospels.  It's only mythicists who think we can trace all the gospels back to Mark.  At bare minimum you have Mark and Q.  Your remaining assertions are not terribly substantive. [NEWLINE] [NEWLINE] The bottom line is that we have multiple, largely independent sources referencing Jesus within a century of his death.  That is a reasonable body of evidence considering the status of the man ( yet another Jewish troublemaker in a backwater of the Roman Empire) and the general lack of sources to survive from the period in general. [NEWLINE] [NEWLINE] **Christian Sources.** You have the gospels (and the sources lying behind the gospels like Q).  You have Paul (Galatians 1 being the strongest link, in which he mentions meeting with Peter and James in the mid 30s).  Both Paul and the extant gospels preserve earlier material, e.g. the hymn in Philippians or Matthew 12:1-8, likely surviving elements of earlier oral traditions. [NEWLINE] [NEWLINE] **Jewish Sources** Josephus in his Antiquities mentions Jesus twice, in books 18 and 20.  He also mentions John the Baptist in Book 18.  The references to John and the reference to Jesus in Book 20 are considered authentic by the overwhelming majority of Josephus scholars.  The reference to Jesus in Book 18, the Testimonium, is obviously an interpolation but there is widespread agreement among scholars that the passage contains an authentic core. [NEWLINE] [NEWLINE] **Roman Sources.** You have Tacitus, Suetonius and Pliny the Younger.  Of those three, Tacitus is the strongest and clearest reference. [NEWLINE] [NEWLINE] For the mythicist case to stand, you need to explain away all of those references.  You also need to explain why somebody in the first century would invent a supposed Jewish messiah who acts nothing like what every Jew knew the messiah was supposed to be (hint: it wasn't dying in shame on the cross).  You need to explain why Matthew and Luke take great pains to explain how a man from Nazareth was actually from Bethlehem, when if the man were a fiction they could simply have placed him in the right city to begin with.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>My mother came to this country with two kids, no money and couldn't speak a word in English. We originally lived in a roach infested apartment in a not-<mask> -great neighborhood. She worked night and day in a factory equivalent to a sweat shop for minimum wage. She moved up the ranks to become manager, with that, we managed to move to a better neighborhood and afford a car after three years in America. Soon after she started a business and we were able buy a summer house. Six years later, we moved again to an actual house and bought another car. She still works day and night, even on weekends. During this adventure, she only needed welfare for the first six months we moved to America. [NEWLINE] [NEWLINE] Now for my part, seeing my mother work this hard, it inspired me. I got a job<mask> soon<mask> I was of age and it only took me a week. I worked hard and got a few raises, several dollars above the minimum wage. I left the job<mask> I started to attend college<mask> worked other small jobs here and there.<mask> I needed money again, I went out looking and again I got a job within a week doing deliveries. I worked hard and received raises, earning a bit more than double the minimum wage.<mask> I looked for internships in my field, I got one within one month (with no connections). [NEWLINE] [NEWLINE] My entire family and friends have similar stories to my mother and I. We worked hard and got<mask> we deserved.<mask>, people are still on welfare and in the same position for couple of years. Some have no intent of making better of themselves.<mask> we were broke, I got robbed by someone who had better shoes than me and I only had a dollar on me. I have no sympathy to anyone who doesn't apply themselves.<mask> should I pay for their leisure? [NEWLINE] [NEWLINE] Some apply for welfare<mask> they have kids.<mask> should I pay for their kids<mask> I'm waiting to be financially fit to afford them<mask> they just didn't use a condom or think of the consequences? We all went to the same schools and had the same opportunities,<mask> I get punished for succeeding<mask> others haven't even tried? I go to school two hours away from me and some people aren't willing to do 30 min commutes. [NEWLINE] [NEWLINE] From my point of view, the world doesn't owe anybody a goddamn thing. CMV</s>
Label encoding: <s>My mother came to this country with two kids, no money and couldn't speak a word in English. We originally lived in a roach infested apartment in a not- so -great neighborhood. She worked night and day in a factory equivalent to a sweat shop for minimum wage. She moved up the ranks to become manager, with that, we managed to move to a better neighborhood and afford a car after three years in America. Soon after she started a business and we were able buy a summer house. Six years later, we moved again to an actual house and bought another car. She still works day and night, even on weekends. During this adventure, she only needed welfare for the first six months we moved to America. [NEWLINE] [NEWLINE] Now for my part, seeing my mother work this hard, it inspired me. I got a job as soon as I was of age and it only took me a week. I worked hard and got a few raises, several dollars above the minimum wage. I left the job when I started to attend college but worked other small jobs here and there. When I needed money again, I went out looking and again I got a job within a week doing deliveries. I worked hard and received raises, earning a bit more than double the minimum wage. When I looked for internships in my field, I got one within one month (with no connections). [NEWLINE] [NEWLINE] My entire family and friends have similar stories to my mother and I. We worked hard and got what we deserved. Yet, people are still on welfare and in the same position for couple of years. Some have no intent of making better of themselves. When we were broke, I got robbed by someone who had better shoes than me and I only had a dollar on me. I have no sympathy to anyone who doesn't apply themselves. Why should I pay for their leisure? [NEWLINE] [NEWLINE] Some apply for welfare when they have kids. Why should I pay for their kids when I'm waiting to be financially fit to afford them while they just didn't use a condom or think of the consequences? We all went to the same schools and had the same opportunities, yet I get punished for succeeding when others haven't even tried? I go to school two hours away from me and some people aren't willing to do 30 min commutes. [NEWLINE] [NEWLINE] From my point of view, the world doesn't owe anybody a goddamn thing. CMV</s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0581, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0642, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0414, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] wife and kids at home and a piece of ass on the side. [ENDQ] [NEWLINE] Personally I'm into polyamory, and do not do casual sex at all.  I build relationships with each partner, and I love each of the ones I've been with for a long period of time. [NEWLINE] [NEWLINE] There are others that do have casual flings on the side, with their primary/wife/spouse fully aware.  There are people that just love chasing NRE and have short relationships that end<mask> the NRE ends, which is fine<mask> you at least warn your new partner. [NEWLINE] [NEWLINE] [STARTQ] able to deny yourself from wanton needs? [ENDQ] [NEWLINE] I never understood<mask> relationships are built around denying yourself, that is some old-school religious mindmolding right there.  Ultimately open relationships go against that basic principle. <mask> you meet someone you are interested in, you can explore<mask> that means.   You can learn more about yourself with each new partner.  There are things you may never have known<mask> you didn't go out with a specific partner. [NEWLINE] [NEWLINE] Humans are extremely complex, and we change over time.  I co-parent with my wife and we get to finish raising our children together in the same house<mask><mask> the romantic side died down.  We are still both happy and have fulfilling lives, and we can fulfill our obligations<mask> parents without being miserable. [NEWLINE] [NEWLINE] And I know many couples in their 70's and 80's still in open relationships. [NEWLINE] [NEWLINE] For your scenario, hypothetically<mask><mask> his wife was fully aware and said "we aren't going to see each other much, go ahead and find someone out there".  Think of the fun you had together and take away the broken trust of lying to you. <mask> everyone is fully informed things are a lot different. [NEWLINE] [NEWLINE] Polyamorous relationships don't survive without complete honesty, even Don't ask/Don't tell are at least honest<mask><mask><mask> knowing stuff is going on, just not interested in details.  (Personally, I avoid DA/DT, it usually includes some drama). [NEWLINE] [NEWLINE] In short, relationships are complicated in general,<mask> you remove some of the restrictions they change<mask> is complicated about it,<mask> at least not constantly thinking "<mask><mask>?" which is a big issue with monogamous relationships that end up with cheating. [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] wife and kids at home and a piece of ass on the side. [ENDQ] [NEWLINE] Personally I'm into polyamory, and do not do casual sex at all.  I build relationships with each partner, and I love each of the ones I've been with for a long period of time. [NEWLINE] [NEWLINE] There are others that do have casual flings on the side, with their primary/wife/spouse fully aware.  There are people that just love chasing NRE and have short relationships that end when the NRE ends, which is fine if you at least warn your new partner. [NEWLINE] [NEWLINE] [STARTQ] able to deny yourself from wanton needs? [ENDQ] [NEWLINE] I never understood why relationships are built around denying yourself, that is some old-school religious mindmolding right there.  Ultimately open relationships go against that basic principle.  When you meet someone you are interested in, you can explore what that means.   You can learn more about yourself with each new partner.  There are things you may never have known if you didn't go out with a specific partner. [NEWLINE] [NEWLINE] Humans are extremely complex, and we change over time.  I co-parent with my wife and we get to finish raising our children together in the same house even though the romantic side died down.  We are still both happy and have fulfilling lives, and we can fulfill our obligations as parents without being miserable. [NEWLINE] [NEWLINE] And I know many couples in their 70's and 80's still in open relationships. [NEWLINE] [NEWLINE] For your scenario, hypothetically what if his wife was fully aware and said "we aren't going to see each other much, go ahead and find someone out there".  Think of the fun you had together and take away the broken trust of lying to you.  When everyone is fully informed things are a lot different. [NEWLINE] [NEWLINE] Polyamorous relationships don't survive without complete honesty, even Don't ask/Don't tell are at least honest as far as knowing stuff is going on, just not interested in details.  (Personally, I avoid DA/DT, it usually includes some drama). [NEWLINE] [NEWLINE] In short, relationships are complicated in general, if you remove some of the restrictions they change what is complicated about it, but at least not constantly thinking " what if?" which is a big issue with monogamous relationships that end up with cheating. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0339, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0431, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You are confusing yourself on<mask> selfish means. [NEWLINE] [NEWLINE] [STARTQ] <mask> the people in question were not selfish and genuinely cared [ENDQ] [NEWLINE] <mask><mask><mask>, the person doesnt have to genuinely care about the suicidal person. All the person has to do is genuinely care about the other people. [NEWLINE] [NEWLINE] An example: [NEWLINE] [NEWLINE] I steal a loaf of bread from a man. Now<mask> I did it for myself it would be selfish<mask> it's lacking in consideration for the man and is only in my interest.<mask> I stole the bread for my children I'm not selfish. My relationship with the baker doesnt matter in this equation. All that matters is<mask> Im doing it selfishly.<mask> IN FOR MYSELF! Selflessly would be the opposite of selfishly,<mask> in a eat none of the bread and give it all to the children. Again my relationship with the baker doesnt count for shit. [NEWLINE] [NEWLINE] Islamic suicide bombers are selfless. Are they morally in the right? No.<mask> they are selfless. Well actually, I'm not sure<mask> they are doing it in order to get into heaven. I revise my statement. Suicide bombing for your country is selfless. Who you bombed doesnt matter.<mask> it was a rapist or a child, the victim of your act doesnt influence your motive<mask> deciding whether the act is selfish or not, unless ofc you are doing it for the victem. Even then they really have no say in whether or not you are acting for their sakes. [NEWLINE] [NEWLINE] <mask> now that you have realized terrorist can be selfless, you can<mask> recognize that friends who plead with suiciders not to kill themselves can<mask> not be selfish. [NEWLINE] [NEWLINE] Oh btw your condition doesnt have anything to do with whether or not you are selfish.<mask> you are dying and you need a cure,<mask> a guy with a broken leg<mask> needs that cure to heal his leg, and you take that cure for yourself, you maybe justified in your actions<mask> you are selfish all the same. [NEWLINE] [NEWLINE] Same thing with suicide people who dont care about others. [NEWLINE] [NEWLINE] Being selfish here maybe the lesser evil<mask> and you should keep that in mind. Inflicting cruel conditions on a person(suicider) is a greater evil than being selfish.<mask> in no way is the pleader justified by being non-selfish.</s>
Label encoding: <s>You are confusing yourself on what selfish means. [NEWLINE] [NEWLINE] [STARTQ] If the people in question were not selfish and genuinely cared [ENDQ] [NEWLINE] First of all, the person doesnt have to genuinely care about the suicidal person. All the person has to do is genuinely care about the other people. [NEWLINE] [NEWLINE] An example: [NEWLINE] [NEWLINE] I steal a loaf of bread from a man. Now if I did it for myself it would be selfish since it's lacking in consideration for the man and is only in my interest. If I stole the bread for my children I'm not selfish. My relationship with the baker doesnt matter in this equation. All that matters is if Im doing it selfishly. AS IN FOR MYSELF! Selflessly would be the opposite of selfishly, as in a eat none of the bread and give it all to the children. Again my relationship with the baker doesnt count for shit. [NEWLINE] [NEWLINE] Islamic suicide bombers are selfless. Are they morally in the right? No. But they are selfless. Well actually, I'm not sure if they are doing it in order to get into heaven. I revise my statement. Suicide bombing for your country is selfless. Who you bombed doesnt matter. If it was a rapist or a child, the victim of your act doesnt influence your motive when deciding whether the act is selfish or not, unless ofc you are doing it for the victem. Even then they really have no say in whether or not you are acting for their sakes. [NEWLINE] [NEWLINE] So now that you have realized terrorist can be selfless, you can also recognize that friends who plead with suiciders not to kill themselves can also not be selfish. [NEWLINE] [NEWLINE] Oh btw your condition doesnt have anything to do with whether or not you are selfish. If you are dying and you need a cure, but a guy with a broken leg also needs that cure to heal his leg, and you take that cure for yourself, you maybe justified in your actions but you are selfish all the same. [NEWLINE] [NEWLINE] Same thing with suicide people who dont care about others. [NEWLINE] [NEWLINE] Being selfish here maybe the lesser evil though and you should keep that in mind. Inflicting cruel conditions on a person(suicider) is a greater evil than being selfish. So in no way is the pleader justified by being non-selfish.</s>
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Two points. One, a moral premise; the other, an empirical observation. [NEWLINE] [NEWLINE] *<mask>, in general, it isn't justifiable to force someone to help another except under certain dire circumstances.<mask> even then, it probably isn't justified *<mask> it isn't clear<mask> the action will help or harm.* For example,<mask> there was a child drowning in a pond and only you could save him/her, then I *might* be justified in forcing you to save the child (like by threatening to impose some punishment on you<mask> you chose not to do it, including imprisonment, which happens to be the penalty for not paying taxes for the welfare state.)<mask>, suppose it wasn't clear that you could save the child.<mask>, suppose it was plausible that you might actually end up knocking another child in the pond, resulting in its death, during your forced attempt to save the first. Would I still be justified in *forcing* you to help? I submit I would not be justified in doing such a thing. **TL;DR**:<mask> a general moral principle, you aren't justified in forcing someone to perform an action<mask> it is unclear whether that action would help or harm. [NEWLINE] *<mask>, many people see welfare<mask> a system which *helps* the poor.<mask>, this is highly debatable. I won't go into all the reasons for and against (<mask> I'll mention that poverty rates have stagnated after implementing our "War on Poverty", and they were drastically declining prior,)<mask> they are all very complicated, and there are thoughtful voices on both sides of the debate. The point I want to make is *that it isn't at all clear* that the welfare state helps the poor. See Charles Murray's [*Losing Ground: American Social Policy, 1950-1980*]( [URL] ;sr=&amp;qid=) for an academic criticism of the welfare state. **TL;DR**: It isn't clear that welfare actually helps the poor, and it may<mask><mask> harm them. [NEWLINE] [NEWLINE] It follows from these two points that the welfare state is immoral.<mask> you want to deny this conclusion, you have to deny one of the premises (the first being a moral principle, the second being an empirical matter.)<mask> both of the premises seem fairly plausible to me.</s><pad>
Label encoding: <s>Two points. One, a moral premise; the other, an empirical observation. [NEWLINE] [NEWLINE] * Firstly, in general, it isn't justifiable to force someone to help another except under certain dire circumstances. But even then, it probably isn't justified * when it isn't clear if the action will help or harm.* For example, if there was a child drowning in a pond and only you could save him/her, then I *might* be justified in forcing you to save the child (like by threatening to impose some punishment on you if you chose not to do it, including imprisonment, which happens to be the penalty for not paying taxes for the welfare state.) However, suppose it wasn't clear that you could save the child. Additionally, suppose it was plausible that you might actually end up knocking another child in the pond, resulting in its death, during your forced attempt to save the first. Would I still be justified in *forcing* you to help? I submit I would not be justified in doing such a thing. **TL;DR**: As a general moral principle, you aren't justified in forcing someone to perform an action if it is unclear whether that action would help or harm. [NEWLINE] * Secondly, many people see welfare as a system which *helps* the poor. However, this is highly debatable. I won't go into all the reasons for and against ( though I'll mention that poverty rates have stagnated after implementing our "War on Poverty", and they were drastically declining prior,) because they are all very complicated, and there are thoughtful voices on both sides of the debate. The point I want to make is *that it isn't at all clear* that the welfare state helps the poor. See Charles Murray's [*Losing Ground: American Social Policy, 1950-1980*]( [URL] ;sr=&amp;qid=) for an academic criticism of the welfare state. **TL;DR**: It isn't clear that welfare actually helps the poor, and it may in fact harm them. [NEWLINE] [NEWLINE] It follows from these two points that the welfare state is immoral. If you want to deny this conclusion, you have to deny one of the premises (the first being a moral principle, the second being an empirical matter.) But both of the premises seem fairly plausible to me.</s><pad>
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Fahrenheit was a very poorly-designed scale. Fahrenheit was created with the intent being "OK,<mask> 0 is the coldest it's ever been around here—about<mask> cold<mask> this brine mixture I made to represent it—and the human body is naturally at 100. OK, based on that scale, water freezes at 32 and boils at 212, interesting... that'll probably be easier to replicate than human body temperature." It turned out his brine wasn't fully mixed and would stabilize at -3°,<mask> eventually they had to recalibrate this based on water's freezing point instead. It reeks of bad craft,<mask> he was the guy who invented the mercury thermometer,<mask> that's<mask> it caught on. [NEWLINE] [NEWLINE] Celsius,<mask><mask><mask><mask>, was created with the intent being "0° is freezing, 100° is boiling, and both of these measurements are at one atmosphere of pressure." Fahrenheit never had that last part until it got a specific definition in relation to Celsius. It's a much better-designed standard that was much easier to accurately replicate. [NEWLINE] [NEWLINE] Kelvin? Lord Kelvin calculated<mask> absolute zero would be in Celsius and then recalibrated that<mask> zero. The conversion is easier—mere subtraction,<mask> opposed to a Fahrenheit conversion's three-operation conversion—<mask> more importantly, *a degree Celsius and a degree Kelvin have the same magnitude.* They are compatible! That's really useful! [NEWLINE] [NEWLINE] <mask> you want to bring usefulness into it? OK. Fahrenheit is probably the slightly better perceptual meteorological scale, I'll give you that. (Is it significantly better? Not really. 0° to 100° Fahrenheit is roughly analogous to -20° to 40° Celsius.)<mask> is it better than Celsius at cooking temperatures? Not really. Celsius' perceptual definitions are at round numbers, too, which is helpful. And even<mask> you just say "have a second system for scientific temperature measurements," compatibility is really useful—would you really want to have to buy custom thermometers or do clunky conversions just to be able to use your measurements with common scientific formulas? And it's better-designed, which made it easier to manufacture accurately at the time. Frankly, even<mask> someone who uses Fahrenheit every day, I'm glad Celsius got the nod from the SI.</s>
Label encoding: <s>Fahrenheit was a very poorly-designed scale. Fahrenheit was created with the intent being "OK, so 0 is the coldest it's ever been around here—about as cold as this brine mixture I made to represent it—and the human body is naturally at 100. OK, based on that scale, water freezes at 32 and boils at 212, interesting... that'll probably be easier to replicate than human body temperature." It turned out his brine wasn't fully mixed and would stabilize at -3°, so eventually they had to recalibrate this based on water's freezing point instead. It reeks of bad craft, but he was the guy who invented the mercury thermometer, so that's why it caught on. [NEWLINE] [NEWLINE] Celsius, on the other hand, was created with the intent being "0° is freezing, 100° is boiling, and both of these measurements are at one atmosphere of pressure." Fahrenheit never had that last part until it got a specific definition in relation to Celsius. It's a much better-designed standard that was much easier to accurately replicate. [NEWLINE] [NEWLINE] Kelvin? Lord Kelvin calculated what absolute zero would be in Celsius and then recalibrated that as zero. The conversion is easier—mere subtraction, as opposed to a Fahrenheit conversion's three-operation conversion— but more importantly, *a degree Celsius and a degree Kelvin have the same magnitude.* They are compatible! That's really useful! [NEWLINE] [NEWLINE] But you want to bring usefulness into it? OK. Fahrenheit is probably the slightly better perceptual meteorological scale, I'll give you that. (Is it significantly better? Not really. 0° to 100° Fahrenheit is roughly analogous to -20° to 40° Celsius.) But is it better than Celsius at cooking temperatures? Not really. Celsius' perceptual definitions are at round numbers, too, which is helpful. And even if you just say "have a second system for scientific temperature measurements," compatibility is really useful—would you really want to have to buy custom thermometers or do clunky conversions just to be able to use your measurements with common scientific formulas? And it's better-designed, which made it easier to manufacture accurately at the time. Frankly, even as someone who uses Fahrenheit every day, I'm glad Celsius got the nod from the SI.</s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm undecided<mask> I feel about flying the flag at public institutions,<mask> there a lot of problems I have issue with that I've heard often from a lot of people. [NEWLINE] [NEWLINE] People say the flag represents slavery even<mask> by default.<mask> does it seem this way? We started a war over slavery? Is this really<mask> you feel? There's nothing below the surface? Over a course of a few months people decided they were willing to lay their lives on the line<mask> they loved slavery<mask> much? [NEWLINE] [NEWLINE] The South's resentment for the North started before the civil war. And the Northern states were threatening the entire economy of the South, which were already poorer than the North. The southern states were agrarian and the North were much more industrial, had the circumstances been flipped (Southern states wanting to ban manufacturing I guess??) I'm certain the North would have done the same.<mask> someone is threatening your livelihood your family's livelihood, I hope you'd try to defend it to. [NEWLINE] [NEWLINE] Now of course we can sit on our high horses<mask> we've progressed<mask> a society and say "Slavery is<mask> awful I can't believe anyone would ever think it's okay!!!"<mask> don't forget most civilizations  have participated in slavery. Do you think Haiti's flag represents slavery<mask> about Jamaica? Those countries were practically founded upon slave exploitation. [NEWLINE] [NEWLINE] The way I see it is to state that the Confederate flag is a sign of slavery is an oversimplification and narrow-minded. Do you think a people on the battle field standing next to their bothers, and across from their bothers looked at the flag ready to lay their lives down just<mask> they can go home and own another person? There's<mask> much more to it than that<mask><mask>. [NEWLINE] [NEWLINE] Now about<mask> it's perceived today? Yeah a lot of people probably look at it in disgust "stupid rednecks" or "racists". And often they are probably right. Most people I know who fly it or have it hanging off their truck are usually the redneckiest racists and<mask><mask> that often reflects poorly and reinforces a negative image. It's unfortunate<mask> I truly believe that's not<mask> it represents. [NEWLINE] [NEWLINE] We don't let our ideals of our forefather be sullied<mask> they owned slaves. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I'm undecided how I feel about flying the flag at public institutions, but there a lot of problems I have issue with that I've heard often from a lot of people. [NEWLINE] [NEWLINE] People say the flag represents slavery even if by default. Why does it seem this way? We started a war over slavery? Is this really how you feel? There's nothing below the surface? Over a course of a few months people decided they were willing to lay their lives on the line because they loved slavery so much? [NEWLINE] [NEWLINE] The South's resentment for the North started before the civil war. And the Northern states were threatening the entire economy of the South, which were already poorer than the North. The southern states were agrarian and the North were much more industrial, had the circumstances been flipped (Southern states wanting to ban manufacturing I guess??) I'm certain the North would have done the same. If someone is threatening your livelihood your family's livelihood, I hope you'd try to defend it to. [NEWLINE] [NEWLINE] Now of course we can sit on our high horses because we've progressed as a society and say "Slavery is so awful I can't believe anyone would ever think it's okay!!!" but don't forget most civilizations  have participated in slavery. Do you think Haiti's flag represents slavery what about Jamaica? Those countries were practically founded upon slave exploitation. [NEWLINE] [NEWLINE] The way I see it is to state that the Confederate flag is a sign of slavery is an oversimplification and narrow-minded. Do you think a people on the battle field standing next to their bothers, and across from their bothers looked at the flag ready to lay their lives down just so they can go home and own another person? There's so much more to it than that I think. [NEWLINE] [NEWLINE] Now about how it's perceived today? Yeah a lot of people probably look at it in disgust "stupid rednecks" or "racists". And often they are probably right. Most people I know who fly it or have it hanging off their truck are usually the redneckiest racists and I think that often reflects poorly and reinforces a negative image. It's unfortunate because I truly believe that's not what it represents. [NEWLINE] [NEWLINE] We don't let our ideals of our forefather be sullied because they owned slaves. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are a couple of your ideas that need to be debunked. [NEWLINE] You assert: [NEWLINE] -that good art has a message from the artist, and that this message/intent/substance is important [NEWLINE] [NEWLINE] -that good art necessitates skilled technique [NEWLINE] [NEWLINE] -that good art is a pleasing image [NEWLINE] [NEWLINE] -that good art exists (there is good are and bad art) [NEWLINE] [NEWLINE] All of these things are both true and false. Let's start at the bottom<mask> it informs the rest.<mask> there is good art, then there is bad art. The problem is that there is no way to quantifiably assert whether art is good or bad. There is only opinion. Art gains strength from all forms of conversation. That<mask> it enters culture.<mask> it's completely useless to say, "Look at this art! It's bad!" It's useless<mask> you've just undermined yourself. You've introduced another person to this piece of bad art, started another conversations about this piece of bad art, further established this piece of bad art<mask> *art*. Once it's art it remains<mask> art.<mask> you don't like something then ignore it. There is way too much good art to linger on<mask> doesn't appeal to you. [NEWLINE] [NEWLINE] Which is<mask> the next two points about. Whether an image is pleasing or technically skilled is only relevant to the tastes of the observer. It's<mask> speaks to you that matters. [NEWLINE] [NEWLINE] <mask> then the last (first) point: art should have a message. This is<mask>'s holding you up. Most art has no message. I find that art that does have a message to be really really boring.<mask> this is<mask><mask><mask> that the best art has a sense of mystery—you don't quite understand<mask> or<mask> it's working. Political art (art with an explicit message) is kind of boring<mask> once I figure it out then it loses it's mystery. It's cool in the way that history is cool,<mask> not in the way that art is cool. At least to me. [NEWLINE] [NEWLINE] <mask> kind of art do you find interesting? I'm interested even<mask> you don't know that much.<mask> have you seen that really speaks to you—that moves you in a way that you can't really explain? That goes for anyone on this thread.</s>
Label encoding: <s>There are a couple of your ideas that need to be debunked. [NEWLINE] You assert: [NEWLINE] -that good art has a message from the artist, and that this message/intent/substance is important [NEWLINE] [NEWLINE] -that good art necessitates skilled technique [NEWLINE] [NEWLINE] -that good art is a pleasing image [NEWLINE] [NEWLINE] -that good art exists (there is good are and bad art) [NEWLINE] [NEWLINE] All of these things are both true and false. Let's start at the bottom because it informs the rest. If there is good art, then there is bad art. The problem is that there is no way to quantifiably assert whether art is good or bad. There is only opinion. Art gains strength from all forms of conversation. That how it enters culture. So it's completely useless to say, "Look at this art! It's bad!" It's useless because you've just undermined yourself. You've introduced another person to this piece of bad art, started another conversations about this piece of bad art, further established this piece of bad art as *art*. Once it's art it remains as art. If you don't like something then ignore it. There is way too much good art to linger on what doesn't appeal to you. [NEWLINE] [NEWLINE] Which is what the next two points about. Whether an image is pleasing or technically skilled is only relevant to the tastes of the observer. It's what speaks to you that matters. [NEWLINE] [NEWLINE] So then the last (first) point: art should have a message. This is what's holding you up. Most art has no message. I find that art that does have a message to be really really boring. But this is because I think that the best art has a sense of mystery—you don't quite understand why or how it's working. Political art (art with an explicit message) is kind of boring because once I figure it out then it loses it's mystery. It's cool in the way that history is cool, but not in the way that art is cool. At least to me. [NEWLINE] [NEWLINE] What kind of art do you find interesting? I'm interested even if you don't know that much. What have you seen that really speaks to you—that moves you in a way that you can't really explain? That goes for anyone on this thread.</s>
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0428, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Something I see come up a lot in music discussion threads, whether on Sputnik Music, /r/Music, or even a more niche sub like /r/PostHardcore, is that being very specific with genres is an annoying thing. This is an especially large joke in extreme metal discussions<mask> there are multitudes of different sub-genres. [NEWLINE] [NEWLINE] <mask><mask> these are great.<mask> you like a band with a specific sound and you want to find more bands like them your best bet is to search for bands in the same genre. Being vague or broad with genres is a great way to not find more bands like the ones you're looking for. [NEWLINE] [NEWLINE] For example, say I like Protest the Hero and want to find more bands like them. They have a lot of influences in their style<mask> could be chalked up into the umbrella genre of "metal."<mask> I search "metal" I will get a *ton* of bands that sound absolutely nothing like Protest the Hero. Even narrowing it down to "Progressive Metal" still gets a lot of bands that aren't very similar, like Opeth.<mask><mask> I search for "Mathcore" which is a lot more specific I can find bands that sound a lot like them. [NEWLINE] [NEWLINE] There's no reason for the disdain of genres. No one complains about<mask> movies get a million genre tags ([Example]( [URL].png))<mask><mask> bands do they get annoyed and I can't see<mask>. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>Something I see come up a lot in music discussion threads, whether on Sputnik Music, /r/Music, or even a more niche sub like /r/PostHardcore, is that being very specific with genres is an annoying thing. This is an especially large joke in extreme metal discussions where there are multitudes of different sub-genres. [NEWLINE] [NEWLINE] I think these are great. If you like a band with a specific sound and you want to find more bands like them your best bet is to search for bands in the same genre. Being vague or broad with genres is a great way to not find more bands like the ones you're looking for. [NEWLINE] [NEWLINE] For example, say I like Protest the Hero and want to find more bands like them. They have a lot of influences in their style but could be chalked up into the umbrella genre of "metal." If I search "metal" I will get a *ton* of bands that sound absolutely nothing like Protest the Hero. Even narrowing it down to "Progressive Metal" still gets a lot of bands that aren't very similar, like Opeth. But if I search for "Mathcore" which is a lot more specific I can find bands that sound a lot like them. [NEWLINE] [NEWLINE] There's no reason for the disdain of genres. No one complains about when movies get a million genre tags ([Example]( [URL].png)) but when bands do they get annoyed and I can't see why. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>First up, congrats on the weight loss! [NEWLINE] [NEWLINE] Second,<mask> our headspaces are pretty different, our situations are very similar. I'm<mask> a 26 year old woman who's lost a significant amount of weight (52kg, which<mask><mask> google is around 115lbs)<mask> I hope I can speak with some authority here [NEWLINE] [NEWLINE] Honestly, I don't think the skin sagging around your abdomen needs surgical intervention. It's not an ideal situation,<mask> it's really not that bad, and nowhere near<mask> bad<mask> i'm sure you think it is. Exercise can help in firming up the area,<mask> unfortunately the damage to the skin is done and it will never look like it would have had you not been overweight.<mask> I'm not saying you need to embrace it<mask> some kind of battle scar (<mask><mask> you can, go for it)<mask> accept that it's there, that many *many* women have similar arrangements going on and try not to obsess about it. [NEWLINE] [NEWLINE] <mask> for your breasts, please check out - [URL].php. It's a gallery of pictures of normal, natural, non-sexualised breasts of all shapes and sizes from women of all shapes and sizes and it can be really eye opening to see<mask> your only exposure to breasts has been to your own and to those in the media. [NEWLINE] [NEWLINE] The issue of whether men will find you sexually attractive really shouldn't be the main focus of your thinking around these issues. There's almost no one in the world who isn't physically attractive to at least someone - some people would have been attracted to you<mask> you were bigger, some people will be attracted to you now, some would be attracted to you<mask> you were smaller. You can only give you, and you should only have to give you. And at the moment, the emotional you, honestly, probably isn't helping in terms of attraction.<mask> you approach the world with a sense of inadequacy, people will pick up on that and may start searching for<mask> it is you feel is inadequate. [NEWLINE] [NEWLINE] Rather than beating yourself up about having 'damaged' your body in the first place, try to feel good about the positive life changes you've made.<mask> you're happy to be with yourself, others will be happier to be with you</s>
Label encoding: <s>First up, congrats on the weight loss! [NEWLINE] [NEWLINE] Second, while our headspaces are pretty different, our situations are very similar. I'm also a 26 year old woman who's lost a significant amount of weight (52kg, which according to google is around 115lbs) so I hope I can speak with some authority here [NEWLINE] [NEWLINE] Honestly, I don't think the skin sagging around your abdomen needs surgical intervention. It's not an ideal situation, but it's really not that bad, and nowhere near as bad as i'm sure you think it is. Exercise can help in firming up the area, but unfortunately the damage to the skin is done and it will never look like it would have had you not been overweight. While I'm not saying you need to embrace it as some kind of battle scar ( though if you can, go for it) but accept that it's there, that many *many* women have similar arrangements going on and try not to obsess about it. [NEWLINE] [NEWLINE] As for your breasts, please check out - [URL].php. It's a gallery of pictures of normal, natural, non-sexualised breasts of all shapes and sizes from women of all shapes and sizes and it can be really eye opening to see if your only exposure to breasts has been to your own and to those in the media. [NEWLINE] [NEWLINE] The issue of whether men will find you sexually attractive really shouldn't be the main focus of your thinking around these issues. There's almost no one in the world who isn't physically attractive to at least someone - some people would have been attracted to you when you were bigger, some people will be attracted to you now, some would be attracted to you if you were smaller. You can only give you, and you should only have to give you. And at the moment, the emotional you, honestly, probably isn't helping in terms of attraction. If you approach the world with a sense of inadequacy, people will pick up on that and may start searching for what it is you feel is inadequate. [NEWLINE] [NEWLINE] Rather than beating yourself up about having 'damaged' your body in the first place, try to feel good about the positive life changes you've made. When you're happy to be with yourself, others will be happier to be with you</s>
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0799, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The reason we ban defences is the same<mask> the reason for<mask> we ban the admission of certain forms of evidence: not<mask> they are necessarily wrong<mask><mask> we don't trust judges/juries to use them responsibly. [NEWLINE] [NEWLINE] <mask><mask><mask> I can tell the trans-panic defence falls into a broader umbrella of provocation defences (<mask> not then it would be closely analogous to them). In Canada (and the Common Law world generally) these defences have been restricted over time.<mask>, provocation defences are *not* absolute, they would only serve to reduce the charge to manslaughter (which<mask><mask> you suggested in another comment). [NEWLINE] [NEWLINE] The standard for these defences is highly restrictive: the shock needs to be sufficient to deprive an ordinary, unexcitable (and 'not pugnacious') person of self-control. It is not enough that you be shocked, insulted, betrayed, offended or repulsed. It must be serious enough to deprive you of responsibility for your own actions. [NEWLINE] [NEWLINE] Unfortunately, many people are prejudiced against trans-persons. This prejudice often manifests itself in the form of an unjustifed fear of sexual dishonesty (i.e. that the trans person will trick you into being attracted to them). This defence directly and unavoidably engages that prejudice such that it is unlikely that a judge or jury will assess it reasonably and with the appropriately high bar in mind. And, even<mask> some judges/juries would (and no doubt some would) it remains the case that it is incredibly difficult/impossible to tell in advance who would and it is further very difficult to change this sort of behaviour/prejudice. [NEWLINE] [NEWLINE] <mask>, the law in general, and the law of provocation defences in particular, play a normative role(The Alberta case [R v Tran]( [URL].lexum.com/scc-csc/scc-csc/en/item/7902/index.do) explains<mask> this question *must* have a normative dimension<mask> -ever it is asked). They endorse or condemn certain behaviours in recognition of the fact that these endorsements effect people's behaviour. Finding out someone is trans *should* never justify a crime and<mask> the courts/legislature declare that it *will* never be permitted.</s>
Label encoding: <s>The reason we ban defences is the same as the reason for why we ban the admission of certain forms of evidence: not because they are necessarily wrong but because we don't trust judges/juries to use them responsibly. [NEWLINE] [NEWLINE] As far as I can tell the trans-panic defence falls into a broader umbrella of provocation defences ( If not then it would be closely analogous to them). In Canada (and the Common Law world generally) these defences have been restricted over time. Also, provocation defences are *not* absolute, they would only serve to reduce the charge to manslaughter (which I think you suggested in another comment). [NEWLINE] [NEWLINE] The standard for these defences is highly restrictive: the shock needs to be sufficient to deprive an ordinary, unexcitable (and 'not pugnacious') person of self-control. It is not enough that you be shocked, insulted, betrayed, offended or repulsed. It must be serious enough to deprive you of responsibility for your own actions. [NEWLINE] [NEWLINE] Unfortunately, many people are prejudiced against trans-persons. This prejudice often manifests itself in the form of an unjustifed fear of sexual dishonesty (i.e. that the trans person will trick you into being attracted to them). This defence directly and unavoidably engages that prejudice such that it is unlikely that a judge or jury will assess it reasonably and with the appropriately high bar in mind. And, even if some judges/juries would (and no doubt some would) it remains the case that it is incredibly difficult/impossible to tell in advance who would and it is further very difficult to change this sort of behaviour/prejudice. [NEWLINE] [NEWLINE] Lastly, the law in general, and the law of provocation defences in particular, play a normative role(The Alberta case [R v Tran]( [URL].lexum.com/scc-csc/scc-csc/en/item/7902/index.do) explains why this question *must* have a normative dimension where -ever it is asked). They endorse or condemn certain behaviours in recognition of the fact that these endorsements effect people's behaviour. Finding out someone is trans *should* never justify a crime and therefore the courts/legislature declare that it *will* never be permitted.</s>
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I thought people were not allowed to disagree about whether something deserves to be posted on SRS prime? At least, they can't say<mask> in the same thread. [ENDQ] [NEWLINE] Rule X is more to preserve the spirit of the sub, and limited discussion, especially<mask> there's some ambiguity or misunderstanding, is usually tolerated. I've only seen like, two posts deleted for Rule X violations? [NEWLINE] [NEWLINE] [STARTQ] Ok, then please correct my own misunderstanding.<mask> makes SRS satire,<mask> that word alone being thrown out<mask> a common excuse for shitty or hateful content posted there ostensibly against shitty, hateful Redditors? [ENDQ] [NEWLINE] <mask><mask> it started out<mask> a satire I don't really think of SRS<mask> a satire, especially<mask>, honestly, a lot of the jokes aren't very creative (<mask><mask> they are they're lovely). [NEWLINE] [NEWLINE] My claim is that it's not hypocritical,<mask> wrt. "not getting"<mask>'s going on in the linked posts they either show some kind of subconscious bigotry, or use "satire" or "that's just<mask> it is" to screen their bigotry, or aren't even concerned about their bigotry, etc. Using reverse slurs--or whatever you feel like calling them--is not the same<mask> 1) none of us really believes those things, whereas the bigotry called out is often sincere in one form or another and 2) frankly, cis-scum will never hurt anyone<mask> much<mask> tr***y, or het<mask> much<mask> f****t, or honky<mask> much<mask> n****r. Like, for you to say that our activity is equivalent to that of whoever we're linking is to say there's parity between the two groups in tension, which isn't true. [NEWLINE] [NEWLINE] I would be happy to engage in good faith with the posters I have issue with,<mask> I mean, look<mask> happens? I appreciate that you're making points, even<mask><mask><mask> with them,<mask> most of my inbox is trolls right now. [NEWLINE] [NEWLINE] Finally,<mask><mask> that the ~General Reddit Ethos--rationally progressive, bold and iconoclastic, etc.--is too self-assured and needs to be seriously questioned.<mask><mask> that whole ego trip is exactly<mask> bigotries thrive in some corners of this site.</s>
Label encoding: <s> [STARTQ] I thought people were not allowed to disagree about whether something deserves to be posted on SRS prime? At least, they can't say so in the same thread. [ENDQ] [NEWLINE] Rule X is more to preserve the spirit of the sub, and limited discussion, especially when there's some ambiguity or misunderstanding, is usually tolerated. I've only seen like, two posts deleted for Rule X violations? [NEWLINE] [NEWLINE] [STARTQ] Ok, then please correct my own misunderstanding. What makes SRS satire, besides that word alone being thrown out as a common excuse for shitty or hateful content posted there ostensibly against shitty, hateful Redditors? [ENDQ] [NEWLINE] Even though it started out as a satire I don't really think of SRS as a satire, especially because, honestly, a lot of the jokes aren't very creative ( though when they are they're lovely). [NEWLINE] [NEWLINE] My claim is that it's not hypocritical, because wrt. "not getting" what's going on in the linked posts they either show some kind of subconscious bigotry, or use "satire" or "that's just how it is" to screen their bigotry, or aren't even concerned about their bigotry, etc. Using reverse slurs--or whatever you feel like calling them--is not the same because 1) none of us really believes those things, whereas the bigotry called out is often sincere in one form or another and 2) frankly, cis-scum will never hurt anyone as much as tr***y, or het as much as f****t, or honky as much as n****r. Like, for you to say that our activity is equivalent to that of whoever we're linking is to say there's parity between the two groups in tension, which isn't true. [NEWLINE] [NEWLINE] I would be happy to engage in good faith with the posters I have issue with, but I mean, look what happens? I appreciate that you're making points, even if I disagree with them, but most of my inbox is trolls right now. [NEWLINE] [NEWLINE] Finally, I think that the ~General Reddit Ethos--rationally progressive, bold and iconoclastic, etc.--is too self-assured and needs to be seriously questioned. I think that whole ego trip is exactly why bigotries thrive in some corners of this site.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There is. You are basing your ideas on the idea that the default position is to assume that there is no God, and that you have to comprehensively prove otherwise for this to change. For practical purposes, withholding judgement means neutrality.<mask> someone for personal reasons coming down on either side is not really making any assumptions, only stating personal preferences. (Note that is for the idea of a God, not specific religions, which is another matter.) [NEWLINE] [NEWLINE] <mask>. We know that the universe is specifically organized in a way that allows it's rules to make livable life sustainable. And that this is<mask> rather an unlikely combination, all thigns considered. This does not in itself prove that there's a God,<mask> there is only two realistic options. Either A: there is, or B: it means there is some kind of probably infinite multiverse, with many different rules for universes in them. We simply happen to live in one of the ones that makes sense for life.<mask>, the ISSUE with #2, is that it admits infinite things can exist,<mask> you could define "everything"<mask> a thing. And<mask><mask>,<mask> is stopping a God from existing, no less with infinite chances?<mask> you would have to demonstrate that such a thing is impossible to jsutify the idea that it is not there. [NEWLINE] [NEWLINE] That's just for the traditional western monotheistic idea of God. And only one slice of logic, that is a justification for belief by itself.<mask>, your idea of not believing in a god is western-centric. Across the universe there are many interpretations of god which range from philosophical outlooks on life, to ideas you have probably never heard of. Many which are much more likely than others.<mask><mask> not<mask> you are used to, some other god concepts may<mask> be real. There's a lot of room for looking at this, and the idea that there is some standard that you revert to upon rejecting one single idea comes form the idea that you are defining it<mask> "nothing"<mask><mask> that is not only not the case,<mask> by some philosophies counts anyways.<mask> such, it is a meaningless debate,<mask> you are trying to seem neutral by defining yourself by<mask> you are not, rather than by<mask> you are.</s>
Label encoding: <s>There is. You are basing your ideas on the idea that the default position is to assume that there is no God, and that you have to comprehensively prove otherwise for this to change. For practical purposes, withholding judgement means neutrality. So someone for personal reasons coming down on either side is not really making any assumptions, only stating personal preferences. (Note that is for the idea of a God, not specific religions, which is another matter.) [NEWLINE] [NEWLINE] However. We know that the universe is specifically organized in a way that allows it's rules to make livable life sustainable. And that this is also rather an unlikely combination, all thigns considered. This does not in itself prove that there's a God, but there is only two realistic options. Either A: there is, or B: it means there is some kind of probably infinite multiverse, with many different rules for universes in them. We simply happen to live in one of the ones that makes sense for life. However, the ISSUE with #2, is that it admits infinite things can exist, since you could define "everything" as a thing. And if so, what is stopping a God from existing, no less with infinite chances? So you would have to demonstrate that such a thing is impossible to jsutify the idea that it is not there. [NEWLINE] [NEWLINE] That's just for the traditional western monotheistic idea of God. And only one slice of logic, that is a justification for belief by itself. However, your idea of not believing in a god is western-centric. Across the universe there are many interpretations of god which range from philosophical outlooks on life, to ideas you have probably never heard of. Many which are much more likely than others. So if not what you are used to, some other god concepts may indeed be real. There's a lot of room for looking at this, and the idea that there is some standard that you revert to upon rejecting one single idea comes form the idea that you are defining it as "nothing" even though that is not only not the case, but by some philosophies counts anyways. As such, it is a meaningless debate, since you are trying to seem neutral by defining yourself by what you are not, rather than by what you are.</s>
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0553, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0486, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll address your first, second, and fourth points separately from your third point<mask> I really think your third point is a lot more persuasive (to me at least) than the other four. [NEWLINE] [NEWLINE] <mask> it comes to international relations, I subscribe to a school of thought called "Realism." Realism basically posits that international politics is driven by a struggle for power between nation-states in an anarchic world. [NEWLINE] [NEWLINE] <mask> a Realist, I tend not to think that individual leadership is necessarily all that important.<mask><mask> it's especially unimportant for a country like the United States. The US has<mask> much latent economic and military capacity that it is difficult to imagine any nation picking a fight with us. Simply put, we do not have a "peer competitor" on the horizon. Every other major geopolitical power is in a much less favorable demographic and/or geopolitical situation than the US. China is the only nation that is even close to becoming  a peer competitor for the US, and they are surrounded by hostile neighbors like India and Japan, and their population is expected to shrink dramatically in the coming decades<mask> well. Simply put, it doesn't especially matter<mask> Bernie Sanders has a difficult personality -- Richard Nixon had a similarly awkward personality<mask> the nation survived just fine. In my view, world politics is determined by broad geopolitical forces rather than forces of individual personality,<mask> I wouldn't be terribly concerned about American national security from that point of view. [NEWLINE] [NEWLINE] <mask><mask> your third point,<mask>, is a very good one. The foreign policy establishment in the US -- both Democratic and Republican -- tend to have an incredibly myopic view of the world. This is unfortunate,<mask> a wise foreign policy has to be based on an honest, unprejudiced understanding of the motivations of different governments around the world. Maybe President Sanders would have an easier time seeing things from perspectives other than his own in the arena of foreign policy than he does on domestic policy,<mask> his inability to appreciate<mask> anyone could think differently than he does could be a real problem. [NEWLINE] [NEWLINE] The Bush administration had a similarly difficult time appreciating the viewpoints of people and of governments which did not share its worldview. I don't think we need to rehash the problems that this has caused. </s>
Label encoding: <s>I'll address your first, second, and fourth points separately from your third point because I really think your third point is a lot more persuasive (to me at least) than the other four. [NEWLINE] [NEWLINE] When it comes to international relations, I subscribe to a school of thought called "Realism." Realism basically posits that international politics is driven by a struggle for power between nation-states in an anarchic world. [NEWLINE] [NEWLINE] As a Realist, I tend not to think that individual leadership is necessarily all that important. I think it's especially unimportant for a country like the United States. The US has so much latent economic and military capacity that it is difficult to imagine any nation picking a fight with us. Simply put, we do not have a "peer competitor" on the horizon. Every other major geopolitical power is in a much less favorable demographic and/or geopolitical situation than the US. China is the only nation that is even close to becoming  a peer competitor for the US, and they are surrounded by hostile neighbors like India and Japan, and their population is expected to shrink dramatically in the coming decades as well. Simply put, it doesn't especially matter if Bernie Sanders has a difficult personality -- Richard Nixon had a similarly awkward personality but the nation survived just fine. In my view, world politics is determined by broad geopolitical forces rather than forces of individual personality, so I wouldn't be terribly concerned about American national security from that point of view. [NEWLINE] [NEWLINE] I think your third point, however, is a very good one. The foreign policy establishment in the US -- both Democratic and Republican -- tend to have an incredibly myopic view of the world. This is unfortunate, because a wise foreign policy has to be based on an honest, unprejudiced understanding of the motivations of different governments around the world. Maybe President Sanders would have an easier time seeing things from perspectives other than his own in the arena of foreign policy than he does on domestic policy, but his inability to appreciate why anyone could think differently than he does could be a real problem. [NEWLINE] [NEWLINE] The Bush administration had a similarly difficult time appreciating the viewpoints of people and of governments which did not share its worldview. I don't think we need to rehash the problems that this has caused. </s>
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0444, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Ok, I totally get<mask> you're coming from. [NEWLINE] I've been to parties and have had to drive, and after a few hours you notice everyone acting like drunk idiots<mask> you just get tired and annoyed at not having<mask> much fun<mask> you'should' be having. [NEWLINE] [NEWLINE] I had a period<mask> I decided to row competitively. 7+ trainings a week, be in bed before midnight.  It was rough.<mask> after a<mask> I realized that I wasn't missing the alcohol at all. I felt good, in shape, and on top of the world. [NEWLINE] [NEWLINE] In another period<mask> my life I lost someone close to me and ended up being sober for months to sort of set an example for others who weren't taking it<mask> well and took to drinking their grief away. I did not miss alcohol then,<mask> I had grief that hurt more than anything ever hurt me, and the joy that alcohol brought felt insignificant and meaningless. [NEWLINE] [NEWLINE] The takeaway, I guess, is that sometimes we drink out of habit, and then the stuff acts<mask> a way of loosening up, a social lubricant that helps you drop boundaries. The alcohol is not the focus. It's something on the side that's not imperative. That's<mask> I like my alcohol to be and stay. [NEWLINE] [NEWLINE] In other times, alcohol is a token for something else. For many young insecure people, drinking is a symbol of rebellion, of physical toughness, of belonging to the group.  Look<mask> cool I am by intoxicating myself. It is this way that I've happily outgrown. [NEWLINE] [NEWLINE] I have crazy friends who can be sober all night and still be the epicenter of fun and spontaneity in the party.  I know musicians that will just jam during parties and not touch a drink.  I know friends who can talk deep inti the night about philosophy and politics. I'm a huge board game geek and I can have hours<mask> fun without a drop. [NEWLINE] Once you realize that there are other ways of being yourself,  confident, and have fun, you will be free from your problem. [NEWLINE] Stop trying to be part of it. Just have fun and ignore<mask> people think about<mask> liquids you put in your mouth.</s>
Label encoding: <s>Ok, I totally get where you're coming from. [NEWLINE] I've been to parties and have had to drive, and after a few hours you notice everyone acting like drunk idiots while you just get tired and annoyed at not having as much fun as you'should' be having. [NEWLINE] [NEWLINE] I had a period where I decided to row competitively. 7+ trainings a week, be in bed before midnight.  It was rough. But after a while I realized that I wasn't missing the alcohol at all. I felt good, in shape, and on top of the world. [NEWLINE] [NEWLINE] In another period if my life I lost someone close to me and ended up being sober for months to sort of set an example for others who weren't taking it as well and took to drinking their grief away. I did not miss alcohol then, but I had grief that hurt more than anything ever hurt me, and the joy that alcohol brought felt insignificant and meaningless. [NEWLINE] [NEWLINE] The takeaway, I guess, is that sometimes we drink out of habit, and then the stuff acts as a way of loosening up, a social lubricant that helps you drop boundaries. The alcohol is not the focus. It's something on the side that's not imperative. That's how I like my alcohol to be and stay. [NEWLINE] [NEWLINE] In other times, alcohol is a token for something else. For many young insecure people, drinking is a symbol of rebellion, of physical toughness, of belonging to the group.  Look how cool I am by intoxicating myself. It is this way that I've happily outgrown. [NEWLINE] [NEWLINE] I have crazy friends who can be sober all night and still be the epicenter of fun and spontaneity in the party.  I know musicians that will just jam during parties and not touch a drink.  I know friends who can talk deep inti the night about philosophy and politics. I'm a huge board game geek and I can have hours if fun without a drop. [NEWLINE] Once you realize that there are other ways of being yourself,  confident, and have fun, you will be free from your problem. [NEWLINE] Stop trying to be part of it. Just have fun and ignore what people think about what liquids you put in your mouth.</s>
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0439, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Men are not evil, there's no such thing<mask> the oft mentioned'male gaze', objectification is made up bullshit, and<mask> on. [ENDQ] [NEWLINE] This is the hyperbole. No one is saying men on a whole are evil (well, no one reasonable anyways)<mask> the "male gaze"<mask> a cultural problem does exist<mask> does objectification. You're lumping in one hyperbolic exaggeration with other things based on reality and calling it all bullshit. [NEWLINE] [NEWLINE] [STARTQ] It is quite hard to have a constructive discussion with anyone who believes the misandric fairytale that men somehow hurt women just by looking at them. [ENDQ] [NEWLINE] This just shows you don't actually know or understand the topic. The idea of the "male gaze" and objectification is a much more complex issue than just "men hurt women just by looking at them". You don't seem the type to actually listen or be open minded enough to try to understand<mask> someone explained it to you<mask>.<mask><mask>, I'm not the most equipped to explain it myself. I'm just not very good at explaining it. [NEWLINE] [NEWLINE] [STARTQ] I understand<mask> catcalling can make you nervous<mask> part of that nervousness is not caused by the catcalls themselves<mask> the paranoia you have [ENDQ] [NEWLINE] The problem is, not being a woman, you don't know<mask> it feels to be one in our society. This isn't an insult,<mask> I'm not a woman either. The problem is that<mask> you see<mask> "paranoia" is pretty justified fear based on experiences and observations of our culture. [NEWLINE] [NEWLINE] [STARTQ] <mask> you want to move forward in your life and have less nervousness, the solution (your personal solution) is not trying to shame 3.5 billion men into changing themselves to mitigate your irrational fears,<mask> growing some skin and realizing that catcalling is actually harmless. [ENDQ] [NEWLINE] Except catcalling isn't harmless. That's the point that you seem to refuse to get. You need to have some perspective taking.<mask> does the woman have to change to accomodate a man's behavior?<mask> can't the man being a jackass be called out and expected to change to accomodate the woman's feelings?</s>
Label encoding: <s> [STARTQ] Men are not evil, there's no such thing as the oft mentioned'male gaze', objectification is made up bullshit, and so on. [ENDQ] [NEWLINE] This is the hyperbole. No one is saying men on a whole are evil (well, no one reasonable anyways) however the "male gaze" as a cultural problem does exist as does objectification. You're lumping in one hyperbolic exaggeration with other things based on reality and calling it all bullshit. [NEWLINE] [NEWLINE] [STARTQ] It is quite hard to have a constructive discussion with anyone who believes the misandric fairytale that men somehow hurt women just by looking at them. [ENDQ] [NEWLINE] This just shows you don't actually know or understand the topic. The idea of the "male gaze" and objectification is a much more complex issue than just "men hurt women just by looking at them". You don't seem the type to actually listen or be open minded enough to try to understand if someone explained it to you though. In addition, I'm not the most equipped to explain it myself. I'm just not very good at explaining it. [NEWLINE] [NEWLINE] [STARTQ] I understand how catcalling can make you nervous but part of that nervousness is not caused by the catcalls themselves but the paranoia you have [ENDQ] [NEWLINE] The problem is, not being a woman, you don't know how it feels to be one in our society. This isn't an insult, as I'm not a woman either. The problem is that what you see as "paranoia" is pretty justified fear based on experiences and observations of our culture. [NEWLINE] [NEWLINE] [STARTQ] If you want to move forward in your life and have less nervousness, the solution (your personal solution) is not trying to shame 3.5 billion men into changing themselves to mitigate your irrational fears, but growing some skin and realizing that catcalling is actually harmless. [ENDQ] [NEWLINE] Except catcalling isn't harmless. That's the point that you seem to refuse to get. You need to have some perspective taking. Why does the woman have to change to accomodate a man's behavior? Why can't the man being a jackass be called out and expected to change to accomodate the woman's feelings?</s>
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0469, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Sea level rise. [ENDQ] [NEWLINE] That's just the first step, it will continue to rise and eventually most coastal cities and harbors, by definition built just<mask> the water meets the land, will be forced to relocate, drown or turtle up...<mask> their main economic engine, their harbor, is becoming dysfunctional. It will require absolutely nauseating amounts of infrastructure to be built or replaced. And the main problem is: we won't know<mask> it will stop and building a new harbor inland is pointless,<mask> we will have to make these expenses again and again and again. [NEWLINE] [NEWLINE] And that's ignoring the historical value of these cities. [NEWLINE] [NEWLINE] [STARTQ] Drought and other negative impacts on agricultural production. [ENDQ] [NEWLINE] It's amazing that you cite a single line from a 500-page report to<mask>othe your conscience and ignore the rest... In any case, there are thousands of assumptions that have to be made to be able to make that statement. [NEWLINE] Fact is that current agriculture is optimized for current climatologic circumstances and vegetation patterns. Changing those unpredictably will<mask> change agricultural output unpredictably.<mask> consider that we already have trouble distributing food appropriately, adapting to changing climate will at the very least discombobulate the supply chain. [NEWLINE] [NEWLINE] [STARTQ] Severe weather events. [ENDQ] [NEWLINE] 10% of world population lives within hurricane range from the coasts. [NEWLINE] [NEWLINE] Even looking at it strictly from a financial perspective, it simply makes no sense not to prevent climate change. It's like refusing to pay for an airbag and seatbelts in your car. Just check the expected costs related to climate change for insurance firms alone. [NEWLINE] [NEWLINE] [STARTQ] <mask>, I'm primarily concerned with the impact on humans.<mask> certainly we have to live in the natural environment, and so it matters that the environment is in decent shape, I see it<mask> more an instrumentality to human wellbeing, not an end in itself. [ENDQ] [NEWLINE] You're using an outdated economic model<mask> the environment is just one element that exists within the economy. It's the other way around: the human economy exists within the environment. Even<mask><mask><mask> all the growth of the human economy the environment still produces 2/3 of all goods and services we enjoy.</s>
Label encoding: <s> [STARTQ] Sea level rise. [ENDQ] [NEWLINE] That's just the first step, it will continue to rise and eventually most coastal cities and harbors, by definition built just where the water meets the land, will be forced to relocate, drown or turtle up... while their main economic engine, their harbor, is becoming dysfunctional. It will require absolutely nauseating amounts of infrastructure to be built or replaced. And the main problem is: we won't know where it will stop and building a new harbor inland is pointless, so we will have to make these expenses again and again and again. [NEWLINE] [NEWLINE] And that's ignoring the historical value of these cities. [NEWLINE] [NEWLINE] [STARTQ] Drought and other negative impacts on agricultural production. [ENDQ] [NEWLINE] It's amazing that you cite a single line from a 500-page report to soothe your conscience and ignore the rest... In any case, there are thousands of assumptions that have to be made to be able to make that statement. [NEWLINE] Fact is that current agriculture is optimized for current climatologic circumstances and vegetation patterns. Changing those unpredictably will also change agricultural output unpredictably. Also consider that we already have trouble distributing food appropriately, adapting to changing climate will at the very least discombobulate the supply chain. [NEWLINE] [NEWLINE] [STARTQ] Severe weather events. [ENDQ] [NEWLINE] 10% of world population lives within hurricane range from the coasts. [NEWLINE] [NEWLINE] Even looking at it strictly from a financial perspective, it simply makes no sense not to prevent climate change. It's like refusing to pay for an airbag and seatbelts in your car. Just check the expected costs related to climate change for insurance firms alone. [NEWLINE] [NEWLINE] [STARTQ] Also, I'm primarily concerned with the impact on humans. While certainly we have to live in the natural environment, and so it matters that the environment is in decent shape, I see it as more an instrumentality to human wellbeing, not an end in itself. [ENDQ] [NEWLINE] You're using an outdated economic model where the environment is just one element that exists within the economy. It's the other way around: the human economy exists within the environment. Even in spite of all the growth of the human economy the environment still produces 2/3 of all goods and services we enjoy.</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Bill O'Reilly has famously said "tide comes in, tide goes out, you can't explain that". It was<mask> stupid it became a meme. [ENDQ] [NEWLINE] Really? You're going to use this? Everybody on air says something stupid every<mask> often. It doesn't speak for a party or a news organization. Would you want everyone to judge liberals for everything Joe Biden says? I didn't think<mask>. [NEWLINE] [NEWLINE] [STARTQ] Journalists,<mask><mask><mask><mask>, try not to believe in that shit, are educated themselves and don't implicitly hate academia, etc. I don't think that qualifies<mask> liberal bias! [ENDQ] [NEWLINE] This is a blind, prejudiced, and incredibly flawed generalization in many ways. News flash: not all fundamentally religious people are stupid<mask> they don't agree with you. I don't agree with them either,<mask> that doesn't mean they're unintelligent.<mask><mask><mask> with him on most religious issues, one of the men I respect the most, a Young Earth creationist, is the most well-read man I've ever met, with lots of respect for academia. [NEWLINE] [NEWLINE] [STARTQ] Our party system deludes us into thinking that Democrats and Republicans both have valid positions on any given issue. That's not always the case. Republicans in general even support discrimination, even today. [ENDQ] [NEWLINE] I don't know<mask> you're talking about with the first part. People tend to easily pick one side or the other, especially with social issues<mask> of their outlooks on life. And the second part: wow. Examples?<mask> you're talking about the voter registration, that seriously is to prevent voter fraud. It was passed in both Arkansas and Virginia, and part of the law was to give free IDs to those that lacked it. Or did you throw that accusation around<mask> most Republicans didn't vote for Obama and he's black? That's equally ridiculous. [NEWLINE] [NEWLINE] I do think the media has a liberal bias, with the exception of FOX, which does have a conservative bias.<mask> it's to pander to their audience and get people watching with sensationalist news stories, which aren't hurt by throwing the blame all on one party or the other. [NEWLINE] </s>
Label encoding: <s> [STARTQ] Bill O'Reilly has famously said "tide comes in, tide goes out, you can't explain that". It was so stupid it became a meme. [ENDQ] [NEWLINE] Really? You're going to use this? Everybody on air says something stupid every so often. It doesn't speak for a party or a news organization. Would you want everyone to judge liberals for everything Joe Biden says? I didn't think so. [NEWLINE] [NEWLINE] [STARTQ] Journalists, on the other hand, try not to believe in that shit, are educated themselves and don't implicitly hate academia, etc. I don't think that qualifies as liberal bias! [ENDQ] [NEWLINE] This is a blind, prejudiced, and incredibly flawed generalization in many ways. News flash: not all fundamentally religious people are stupid because they don't agree with you. I don't agree with them either, but that doesn't mean they're unintelligent. Although I disagree with him on most religious issues, one of the men I respect the most, a Young Earth creationist, is the most well-read man I've ever met, with lots of respect for academia. [NEWLINE] [NEWLINE] [STARTQ] Our party system deludes us into thinking that Democrats and Republicans both have valid positions on any given issue. That's not always the case. Republicans in general even support discrimination, even today. [ENDQ] [NEWLINE] I don't know what you're talking about with the first part. People tend to easily pick one side or the other, especially with social issues because of their outlooks on life. And the second part: wow. Examples? If you're talking about the voter registration, that seriously is to prevent voter fraud. It was passed in both Arkansas and Virginia, and part of the law was to give free IDs to those that lacked it. Or did you throw that accusation around because most Republicans didn't vote for Obama and he's black? That's equally ridiculous. [NEWLINE] [NEWLINE] I do think the media has a liberal bias, with the exception of FOX, which does have a conservative bias. But it's to pander to their audience and get people watching with sensationalist news stories, which aren't hurt by throwing the blame all on one party or the other. [NEWLINE] </s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0425, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask>,<mask> it were the owners fault, wouldn't you prefer it made known<mask> that black customers would not patronize a business that treats them unfairly? Wouldn't you prefer the rest of us who realize that's wrong avoid the restaurant? [ENDQ] [NEWLINE] This is the part of your view I have to take issue with the strongest. Essentially I see it<mask> two-fold: [NEWLINE] [NEWLINE] 1) Discrimination based on intrinsic properties of citizens is fundamentally wrong. I do not believe there is any scenario<mask> a business owner is justified in saying: "I'm sorry, I just don't like _____ people." Store policies certainly have the right to filter which types of customers they serve based on optional parameters like: dress code, conduct, hygiene, etc.<mask><mask> it stands, all the protected classes covered by non-discrimination laws are by definition, non-elective properties of those citizens, and<mask> ought to be protected. [NEWLINE] [NEWLINE] 2) Having the combined forces of legal action for overt discrimination and social/economic forces for subtle, emergent discrimination makes for a powerful team that can limit the amount of discrimination people can get away with. [NEWLINE] [NEWLINE] <mask> we cannot agree on 1), I do not believe that there are any arguments that I can make to persuade you. [NEWLINE] [NEWLINE] Edit: To add to some of your other comments. [NEWLINE] [NEWLINE] I see that you have stressed the point that the social forces will serve to add weight against discrimination,<mask> you are forgetting that people who *agree* with the discrimination will be drawn to those businesses in support of that view. Look at<mask> many people flocked to Chic-Fil-A after that whole debacle. [NEWLINE] [NEWLINE] With your plan, only the market forces of a large number of people can alter discriminating behavior, and in small areas<mask> the worldview is likely to be homogeneous, this is nearly impossible<mask> the main view is discriminatory.<mask> it stands, it only takes one person who is getting the raw end of the stick to simply file a lawsuit and curb behaviour. With your plan, only the majority can alter the behavior, which<mask> all of history shows, is never an effective way of allowing discriminated minorities the opportunity to gain equal status.</s>
Label encoding: <s> [STARTQ] However, if it were the owners fault, wouldn't you prefer it made known so that black customers would not patronize a business that treats them unfairly? Wouldn't you prefer the rest of us who realize that's wrong avoid the restaurant? [ENDQ] [NEWLINE] This is the part of your view I have to take issue with the strongest. Essentially I see it as two-fold: [NEWLINE] [NEWLINE] 1) Discrimination based on intrinsic properties of citizens is fundamentally wrong. I do not believe there is any scenario where a business owner is justified in saying: "I'm sorry, I just don't like _____ people." Store policies certainly have the right to filter which types of customers they serve based on optional parameters like: dress code, conduct, hygiene, etc. But as it stands, all the protected classes covered by non-discrimination laws are by definition, non-elective properties of those citizens, and thus ought to be protected. [NEWLINE] [NEWLINE] 2) Having the combined forces of legal action for overt discrimination and social/economic forces for subtle, emergent discrimination makes for a powerful team that can limit the amount of discrimination people can get away with. [NEWLINE] [NEWLINE] If we cannot agree on 1), I do not believe that there are any arguments that I can make to persuade you. [NEWLINE] [NEWLINE] Edit: To add to some of your other comments. [NEWLINE] [NEWLINE] I see that you have stressed the point that the social forces will serve to add weight against discrimination, but you are forgetting that people who *agree* with the discrimination will be drawn to those businesses in support of that view. Look at how many people flocked to Chic-Fil-A after that whole debacle. [NEWLINE] [NEWLINE] With your plan, only the market forces of a large number of people can alter discriminating behavior, and in small areas where the worldview is likely to be homogeneous, this is nearly impossible if the main view is discriminatory. As it stands, it only takes one person who is getting the raw end of the stick to simply file a lawsuit and curb behaviour. With your plan, only the majority can alter the behavior, which as all of history shows, is never an effective way of allowing discriminated minorities the opportunity to gain equal status.</s>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0418, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The left/right political divide is, like all human behaviour, an artefact of evolution. [NEWLINE] [NEWLINE] The right believes that tougher circumstances make fitter varieties<mask> always want to boost competition and avoid protection (free markets, etc.) The strong survive and they like strength. [NEWLINE] [NEWLINE] The left believes in the power humans have to shape their own ecological niche: to create an environment better for human kind (now expanding to "all life" on the radical fringe of the movement). Bear in mind that the way im using the word "technology" can<mask> apply to an idea or social innovation. They believe in big, collective action more likely to disregard the individual for the sake of the whole, making the world a place<mask> its easier for lots of different types to survive (<mask> you rightly point out). [NEWLINE] [NEWLINE] Its two sides of our evolutionary coin. Get better at surviving in the environment by changing ourselves, or make the environment better for us<mask> we are. Its<mask> people on the right seem<mask> harsh to the left, and people on the left seem<mask> unrealistic to people on the most right. [NEWLINE] [NEWLINE] The truly weird part is that it isn't an either-or situation, it's both-and. At times we must adapt, other times we can innovate out of danger. Factionalizing the attitudes isn't a helpful step<mask><mask><mask>. [NEWLINE] [NEWLINE] Religion confuses the question. It preserves rules for survival well past their usefulness meaning they distort the evolutionary effects. [NEWLINE] [NEWLINE] Evidence: [NEWLINE] [NEWLINE] Take abortion: technology solving a human problem (look at crime stats post roe. V. Wade) at the expense of survival-based mechanisms (like your shitty parents abandoning you<mask> they can't invest in you instead of aborting you outright). [NEWLINE] [NEWLINE] Or healthcare: technology (medical, computational, organizational) making it harder to die<mask> you fall off a cliff. Of course someone more interested in environment-shaping-people would oppose this intervention. [NEWLINE] [NEWLINE] They all turn out this way investigated case-by-case, apart from a few awkward religious confusions. Recognizing that this is<mask> is REALLY going on would helpbour government make a LOT more sense.</s>
Label encoding: <s>The left/right political divide is, like all human behaviour, an artefact of evolution. [NEWLINE] [NEWLINE] The right believes that tougher circumstances make fitter varieties so always want to boost competition and avoid protection (free markets, etc.) The strong survive and they like strength. [NEWLINE] [NEWLINE] The left believes in the power humans have to shape their own ecological niche: to create an environment better for human kind (now expanding to "all life" on the radical fringe of the movement). Bear in mind that the way im using the word "technology" can also apply to an idea or social innovation. They believe in big, collective action more likely to disregard the individual for the sake of the whole, making the world a place where its easier for lots of different types to survive ( as you rightly point out). [NEWLINE] [NEWLINE] Its two sides of our evolutionary coin. Get better at surviving in the environment by changing ourselves, or make the environment better for us as we are. Its why people on the right seem so harsh to the left, and people on the left seem so unrealistic to people on the most right. [NEWLINE] [NEWLINE] The truly weird part is that it isn't an either-or situation, it's both-and. At times we must adapt, other times we can innovate out of danger. Factionalizing the attitudes isn't a helpful step in my opinion. [NEWLINE] [NEWLINE] Religion confuses the question. It preserves rules for survival well past their usefulness meaning they distort the evolutionary effects. [NEWLINE] [NEWLINE] Evidence: [NEWLINE] [NEWLINE] Take abortion: technology solving a human problem (look at crime stats post roe. V. Wade) at the expense of survival-based mechanisms (like your shitty parents abandoning you if they can't invest in you instead of aborting you outright). [NEWLINE] [NEWLINE] Or healthcare: technology (medical, computational, organizational) making it harder to die if you fall off a cliff. Of course someone more interested in environment-shaping-people would oppose this intervention. [NEWLINE] [NEWLINE] They all turn out this way investigated case-by-case, apart from a few awkward religious confusions. Recognizing that this is what is REALLY going on would helpbour government make a LOT more sense.</s>
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0358, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thanks for explaining. Please allow me to offer my idea and we can try to get to a solution together. [NEWLINE] [NEWLINE] In my experience, every single person in a position of power has abused it in some way. Some minor and some not<mask> minor abuses,<mask> the idea is that it's human nature to gather and use resources,<mask> resources were scarce until only relatively recently in human history. Power is a resource (or gives access to resources). [NEWLINE] [NEWLINE] <mask><mask> choose one person to have all the decision making power at each level? I believe something along the lines of a [Technocracy]( [URL] ) would be ideal; giving collective power to a group of experts in selected fields,<mask> making their work completely transparent for the public to review. That way each decision has checks and balances on several levels.<mask>, I believe the public should be able to submit business cases for improvements on individual plans, and<mask> they are "upvoted" enough by the public, the government has to give a detailed, time-limited, and researched response to the plan (<mask> opposed to the token paragraph they have to give for such supported plans now). Tying the submissions and their voting to an official online system would reduce trolling. [NEWLINE] [NEWLINE] Politicians should operate<mask> they were originally meant to by George Washington;<mask> spokespeople, with no real individual power. I believe there has to be a person "at the top", a face to associate with the nation/region,<mask> they should not have all veto power. The experts should have equal voting power, with some sort of mediation<mask> there are tight votes or disputes. [NEWLINE] [NEWLINE] This gives the people real power to influence society and change, and reduces the amount of corruption in the government<mask> everything (at least in public policy and spending) is 100% transparent.<mask><mask> this plan is actually quite liberal,<mask> I don't like to side with either party. I don't like left or right, I like answers. [NEWLINE] [NEWLINE] PS: that's<mask> I was too lazy to write<mask> I said my first comment, which I admit was fairly steredditypical. [NEWLINE] [NEWLINE] <mask><mask> that's a new word.</s>
Label encoding: <s>Thanks for explaining. Please allow me to offer my idea and we can try to get to a solution together. [NEWLINE] [NEWLINE] In my experience, every single person in a position of power has abused it in some way. Some minor and some not so minor abuses, but the idea is that it's human nature to gather and use resources, since resources were scarce until only relatively recently in human history. Power is a resource (or gives access to resources). [NEWLINE] [NEWLINE] So why choose one person to have all the decision making power at each level? I believe something along the lines of a [Technocracy]( [URL] ) would be ideal; giving collective power to a group of experts in selected fields, while making their work completely transparent for the public to review. That way each decision has checks and balances on several levels. Also, I believe the public should be able to submit business cases for improvements on individual plans, and if they are "upvoted" enough by the public, the government has to give a detailed, time-limited, and researched response to the plan ( as opposed to the token paragraph they have to give for such supported plans now). Tying the submissions and their voting to an official online system would reduce trolling. [NEWLINE] [NEWLINE] Politicians should operate as they were originally meant to by George Washington; as spokespeople, with no real individual power. I believe there has to be a person "at the top", a face to associate with the nation/region, but they should not have all veto power. The experts should have equal voting power, with some sort of mediation if there are tight votes or disputes. [NEWLINE] [NEWLINE] This gives the people real power to influence society and change, and reduces the amount of corruption in the government since everything (at least in public policy and spending) is 100% transparent. I think this plan is actually quite liberal, though I don't like to side with either party. I don't like left or right, I like answers. [NEWLINE] [NEWLINE] PS: that's what I was too lazy to write when I said my first comment, which I admit was fairly steredditypical. [NEWLINE] [NEWLINE] I think that's a new word.</s>
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm curious<mask> you characterize torturing and killing animals for "pure, sadistic pleasure"<mask> immoral, and then go on to suggest that making any moral judgment of another's treatment of animals is unfair<mask> the judge is in even some, circuitous way engaged in behavior which results in animal suffering. In<mask> doing, aren't you drawing<mask> arbitrary a line<mask> the judgmental vegans who bother you<mask> much? I'm about to use slavery<mask> an analogy, and I hope you don't think I'm implying I find your argument<mask> reprehensible<mask> that might make it seem,<mask> I've found it can help me make my argument more clearly. Say that someone, themselves a slave owner, argued that<mask> slaves ought not be mistreated for pleasure, owning and using them to one's benefit can't be argued to be immoral<mask> the abolitionists making that judgment engage in behavior which, after a drawn out and tenuous chain of cause and effect, results in slave suffering. Does an abolitionist's use of a railroad, the construction of which was achieved in part by slave labor, mean that their argument that slavery is immoral is invalid?<mask><mask> most of us would agree that it doesn't. Most of us would recognize that with an institution<mask> insidious and pervasive<mask> slavery was, the people of its era couldn't possibly help<mask> engage at some point in an activity that could be said to contribute to the suffering of slaves.<mask> percentage of an individual slave's suffering could you reasonably attribute to that one abolitionist? Thousandths of a percent? Maybe.<mask> you sit down to a meal of a whole rotisserie chicken,<mask> percentage of that chicken's suffering is attributable to you? It lived, likely in horrible conditions, and died, probably painfully,<mask> that you could consume its flesh. Subtracting the use of its less palatable tissues (head, feathers, feet, organs), would you say you purchased,<mask> not entirely consumed, about 85% of the bird? The point is to do<mask> little harm<mask> possible, and I'd say being about 85% responsible vs thousandths of a percent isn't even worth comparing. [NEWLINE] </s>
Label encoding: <s>I'm curious why you characterize torturing and killing animals for "pure, sadistic pleasure" as immoral, and then go on to suggest that making any moral judgment of another's treatment of animals is unfair if the judge is in even some, circuitous way engaged in behavior which results in animal suffering. In so doing, aren't you drawing as arbitrary a line as the judgmental vegans who bother you so much? I'm about to use slavery as an analogy, and I hope you don't think I'm implying I find your argument as reprehensible as that might make it seem, but I've found it can help me make my argument more clearly. Say that someone, themselves a slave owner, argued that while slaves ought not be mistreated for pleasure, owning and using them to one's benefit can't be argued to be immoral if the abolitionists making that judgment engage in behavior which, after a drawn out and tenuous chain of cause and effect, results in slave suffering. Does an abolitionist's use of a railroad, the construction of which was achieved in part by slave labor, mean that their argument that slavery is immoral is invalid? I think most of us would agree that it doesn't. Most of us would recognize that with an institution as insidious and pervasive as slavery was, the people of its era couldn't possibly help but engage at some point in an activity that could be said to contribute to the suffering of slaves. What percentage of an individual slave's suffering could you reasonably attribute to that one abolitionist? Thousandths of a percent? Maybe. When you sit down to a meal of a whole rotisserie chicken, what percentage of that chicken's suffering is attributable to you? It lived, likely in horrible conditions, and died, probably painfully, so that you could consume its flesh. Subtracting the use of its less palatable tissues (head, feathers, feet, organs), would you say you purchased, if not entirely consumed, about 85% of the bird? The point is to do as little harm as possible, and I'd say being about 85% responsible vs thousandths of a percent isn't even worth comparing. [NEWLINE] </s>
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0334, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0396, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a drug and alcohol abuse counselor it's probably worth pointing out that the rate of violence and acts of aggression<mask> under the influence is far more likely with alcohol than virtually any other drug in existence. [NEWLINE] [NEWLINE] [STARTQ] Boyum, D., and Kleiman, M. Alcohol and other drugs. In Wilson, J.Q., and Petersilia, J., eds. Crime. [ENDQ] San Francisco: ICS Press, 1995. [NEWLINE] [NEWLINE] In terms of dependency, it's one of very few drugs which creates a level of dependency in which the withdrawal of the chemical can actually kill the person discontinuing its use. The same can't be said of heroin, cocaine, or methamphetamines,<mask> there's generally a cultural bias that these drugs are "more addictive" than alcohol. [NEWLINE] [NEWLINE] I'm in the same camp<mask> OP in terms of not looking to regulate alcohol any more than it already is. The vast majority of people who use alcohol do use it responsibly.<mask> I am opposed to is the idea that alcohol is somehow "different than" other drugs. It's not. [NEWLINE] [NEWLINE] <mask> I used to work with middle school students and I would ask them to identify all the drugs they could think of, alcohol rarely made the list. Ironically, tobacco would almost always make that list,<mask> alcohol was frequently absent. Often,<mask> I would suggest (at the end of the activity) that alcohol might be added to the list, there would often be protest from someone arguing that alcohol was not,<mask><mask>, a drug at all.<mask> it generated some interesting conversation, it's worth noting that people do tend to mentally hold alcohol<mask> being set apart from other substances. [NEWLINE] [NEWLINE] That's anecdotal,<mask> take it with a grain of salt,<mask><mask><mask> it does say something worth looking at about<mask> we view drugs and alcohol (at least in America).<mask> I'm certainly open to less regulation in terms of most substances, I would much prefer living in a world and society that views policy in terms of the evidence, rather than our feelings, intuitions, history, or cultural biases about a drug (or anything else for that matter).</s>
Label encoding: <s>As a drug and alcohol abuse counselor it's probably worth pointing out that the rate of violence and acts of aggression while under the influence is far more likely with alcohol than virtually any other drug in existence. [NEWLINE] [NEWLINE] [STARTQ] Boyum, D., and Kleiman, M. Alcohol and other drugs. In Wilson, J.Q., and Petersilia, J., eds. Crime. [ENDQ] San Francisco: ICS Press, 1995. [NEWLINE] [NEWLINE] In terms of dependency, it's one of very few drugs which creates a level of dependency in which the withdrawal of the chemical can actually kill the person discontinuing its use. The same can't be said of heroin, cocaine, or methamphetamines, though there's generally a cultural bias that these drugs are "more addictive" than alcohol. [NEWLINE] [NEWLINE] I'm in the same camp as OP in terms of not looking to regulate alcohol any more than it already is. The vast majority of people who use alcohol do use it responsibly. What I am opposed to is the idea that alcohol is somehow "different than" other drugs. It's not. [NEWLINE] [NEWLINE] When I used to work with middle school students and I would ask them to identify all the drugs they could think of, alcohol rarely made the list. Ironically, tobacco would almost always make that list, but alcohol was frequently absent. Often, when I would suggest (at the end of the activity) that alcohol might be added to the list, there would often be protest from someone arguing that alcohol was not, in fact, a drug at all. While it generated some interesting conversation, it's worth noting that people do tend to mentally hold alcohol as being set apart from other substances. [NEWLINE] [NEWLINE] That's anecdotal, so take it with a grain of salt, but I think it does say something worth looking at about how we view drugs and alcohol (at least in America). While I'm certainly open to less regulation in terms of most substances, I would much prefer living in a world and society that views policy in terms of the evidence, rather than our feelings, intuitions, history, or cultural biases about a drug (or anything else for that matter).</s>
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0479, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When I see an argument like this, I like to throw out some numbers to play with and put things in perspective. [NEWLINE] [NEWLINE] <mask> the "The number of people you've slept with dictates<mask> we date" argument can be downright illogical : [NEWLINE] [NEWLINE] In High School you lost your virginity at 16 in a typically brief realtionship. Over the next two years you're rage full of hormones and by the time you graduate at 18, you've had a total of 5 partners. Dated two people for 4 months, a couple summer teen dream romances thrown in there, and a "whoops,<mask>?" in there for good measure, it's the age of mistakes after-all. [NEWLINE] [NEWLINE] From 18-20 you have yourself an on and off again realtionship. In-between the longest break you have a FWB. [NEWLINE] [NEWLINE] By 20, you've had 7 partners. [NEWLINE] [NEWLINE] [NEWLINE] At 21 the bar opens up many opportunities. From 21-24 you have no desire for a realtionship and enjoy occasional casual sex. You sleep with 12 people during this time period, or 1 person every 3 months. [NEWLINE] [NEWLINE] From 25-30 you're more interested in serious dating now. Maybe only one person a year at this point<mask> you're not just trying to hook up,<mask> genuinely date and find a good person. [NEWLINE] [NEWLINE] By 30 you've had 24 partners. A number that can sound really "high"<mask> actually isn't considering circumstances. [NEWLINE] [NEWLINE] - [NEWLINE] [NEWLINE] Of course people can have any reason to date or not date someone.<mask><mask> sometimes people with the "I don't date a high number" too quickly assume a certain "picked" number is high, dismiss a person based on chosen ignorance instead of actual circumstances. This is cheating themselves out of very possibly, perfectly good partners. [NEWLINE] [NEWLINE] [NEWLINE] The caveat is very extreme numbers to unbalanced age ratios which I can understand make people uncomfortable (ex. Start having sex at 16, by 21 has had 55 partners)<mask> it seems that sometimes people make a distinction of<mask> a "high" number really is without much thought behind it. </s>
Label encoding: <s>When I see an argument like this, I like to throw out some numbers to play with and put things in perspective. [NEWLINE] [NEWLINE] Why the "The number of people you've slept with dictates if we date" argument can be downright illogical : [NEWLINE] [NEWLINE] In High School you lost your virginity at 16 in a typically brief realtionship. Over the next two years you're rage full of hormones and by the time you graduate at 18, you've had a total of 5 partners. Dated two people for 4 months, a couple summer teen dream romances thrown in there, and a "whoops, why?" in there for good measure, it's the age of mistakes after-all. [NEWLINE] [NEWLINE] From 18-20 you have yourself an on and off again realtionship. In-between the longest break you have a FWB. [NEWLINE] [NEWLINE] By 20, you've had 7 partners. [NEWLINE] [NEWLINE] [NEWLINE] At 21 the bar opens up many opportunities. From 21-24 you have no desire for a realtionship and enjoy occasional casual sex. You sleep with 12 people during this time period, or 1 person every 3 months. [NEWLINE] [NEWLINE] From 25-30 you're more interested in serious dating now. Maybe only one person a year at this point as you're not just trying to hook up, but genuinely date and find a good person. [NEWLINE] [NEWLINE] By 30 you've had 24 partners. A number that can sound really "high" but actually isn't considering circumstances. [NEWLINE] [NEWLINE] - [NEWLINE] [NEWLINE] Of course people can have any reason to date or not date someone. I think sometimes people with the "I don't date a high number" too quickly assume a certain "picked" number is high, dismiss a person based on chosen ignorance instead of actual circumstances. This is cheating themselves out of very possibly, perfectly good partners. [NEWLINE] [NEWLINE] [NEWLINE] The caveat is very extreme numbers to unbalanced age ratios which I can understand make people uncomfortable (ex. Start having sex at 16, by 21 has had 55 partners) But it seems that sometimes people make a distinction of what a "high" number really is without much thought behind it. </s>
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>If you just thought someone was eating peas one by one than same thing holds<mask> you were clearly looking for an excuse.<mask> there is a legit misunderstanding it would be over something important and wouldn't fall into category of "dumb" reasons (for example,<mask> you get the wrong idea<mask> based on good<mask> misleading evidence that your partner is cheating, no one would tell you this is a dumb reason in the first place. Hopefully you will discuss it and see<mask> the misunderstanding comes from.) [NEWLINE] [NEWLINE] [STARTQ] And<mask><mask> the two have been married for 25 years, and have two kids? [ENDQ] [NEWLINE] Yes, like I said before, even more<mask>. Whereas feelings may oscillate<mask> just meeting someone,<mask> someone told me they want to leave their spouse over 25 years whom they know inside and out over the way they eat peas, I would make a safe bet that they don't love their spouse anymore and are just looking for a reason to make this feeling something concrete. i would think their subconsciousness knows<mask> all the practical reasons you mentioned (kids, house, etc) is making them not consciously acknowledge it<mask> they are feeling it through stuff like getting abnormally upset over peas. [NEWLINE] [NEWLINE] [STARTQ] And it's not like emotions are constant either. [ENDQ] [NEWLINE] <mask><mask> that emotions aren't constant and that is actually a good counter argument - sometimes you can feel you want to leave someone for the reasons that aren't lack of love<mask> some other insecurity or protective mechanism. I wouldn't say this changed everything<mask><mask><mask><mask> this is only true at very early stages of relationship<mask> of baggage and such. It is something to consider for the point 8.<mask>, something people should sit and think about before proceeding. Δ for that point,<mask> I would still keep most of the argument unchanged. [NEWLINE] [NEWLINE] People make mistakes part?<mask><mask> that's irrelevant.<mask> the relationship is worth it to you, you will forgive a mistake,<mask> it bothers you too much, you don't owe anyone a second chance. It's not punishment, it's simply not staying in a situation that isn't<mask> you want anymore, mistake or no mistake.</s>
Label encoding: <s>If you just thought someone was eating peas one by one than same thing holds because you were clearly looking for an excuse. If there is a legit misunderstanding it would be over something important and wouldn't fall into category of "dumb" reasons (for example, if you get the wrong idea but based on good but misleading evidence that your partner is cheating, no one would tell you this is a dumb reason in the first place. Hopefully you will discuss it and see where the misunderstanding comes from.) [NEWLINE] [NEWLINE] [STARTQ] And what if the two have been married for 25 years, and have two kids? [ENDQ] [NEWLINE] Yes, like I said before, even more so. Whereas feelings may oscillate when just meeting someone, if someone told me they want to leave their spouse over 25 years whom they know inside and out over the way they eat peas, I would make a safe bet that they don't love their spouse anymore and are just looking for a reason to make this feeling something concrete. i would think their subconsciousness knows but all the practical reasons you mentioned (kids, house, etc) is making them not consciously acknowledge it so they are feeling it through stuff like getting abnormally upset over peas. [NEWLINE] [NEWLINE] [STARTQ] And it's not like emotions are constant either. [ENDQ] [NEWLINE] I agree that emotions aren't constant and that is actually a good counter argument - sometimes you can feel you want to leave someone for the reasons that aren't lack of love but some other insecurity or protective mechanism. I wouldn't say this changed everything though because I think this is only true at very early stages of relationship because of baggage and such. It is something to consider for the point 8. though, something people should sit and think about before proceeding. Δ for that point, although I would still keep most of the argument unchanged. [NEWLINE] [NEWLINE] People make mistakes part? I think that's irrelevant. If the relationship is worth it to you, you will forgive a mistake, if it bothers you too much, you don't owe anyone a second chance. It's not punishment, it's simply not staying in a situation that isn't what you want anymore, mistake or no mistake.</s>
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0451, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] It does not improve service<mask> shown throughout every other country on earth. [ENDQ] [NEWLINE] That's debatable. [NEWLINE] [NEWLINE] [STARTQ] Tipping makes work dishonest<mask> it isnt regulated. [ENDQ] [NEWLINE] <mask> you mean honest<mask><mask><mask> taxes go, my system fixes that by making the tip a line item on the bill. Your system does nothing to control and regulate tips that the server receives on an optional basis and fixes nothing. [NEWLINE] [NEWLINE] [STARTQ] Tipping should not be based on the price of the bill. It should be based on service.<mask> we werent to abolish tipping completely. [ENDQ] [NEWLINE] Your system includes a 15% increase in the cost of the food. There is no difference in that and a 15% tip based on the cost of the food, except one goes in the hands of the operator and the other goes into the hands of the server. You are still paying for service based on the cost, you are just fooling yourself by making it invisible. At least in my system,<mask> the service is terrible, there is a way to negotiate the amount you are paying for service. With your system, you are stuck with it. [NEWLINE] [NEWLINE] [STARTQ] Employee tips can be affected by the quality and timliness of the cooks, whether the restaurant is understaffed, whether they have to share the tips with kitchen staff or with managemen. It should only be based on the work of the server [ENDQ] [NEWLINE] Most establishments require the server to share a percentage of their tips with the back of the house employees. They are compelled to make service<mask> good<mask> possible<mask><mask><mask> incentive.<mask> it's a management or staffing problem, my system mitigates that by including gratuity on all bills. The server isn't harmed,<mask><mask><mask> is compensated directly for having to increase productivity to make up for poor management, unless someone asks for a discount on the gratuity. In your system, the employer actually benefits from being understaffed. They get to collect the 15% increase on the price of the food and pay fewer servers than they need, making extra profits from managing poorly and making their servers work harder without extra compensation.</s>
Label encoding: <s> [STARTQ] It does not improve service as shown throughout every other country on earth. [ENDQ] [NEWLINE] That's debatable. [NEWLINE] [NEWLINE] [STARTQ] Tipping makes work dishonest as it isnt regulated. [ENDQ] [NEWLINE] If you mean honest as far as taxes go, my system fixes that by making the tip a line item on the bill. Your system does nothing to control and regulate tips that the server receives on an optional basis and fixes nothing. [NEWLINE] [NEWLINE] [STARTQ] Tipping should not be based on the price of the bill. It should be based on service. If we werent to abolish tipping completely. [ENDQ] [NEWLINE] Your system includes a 15% increase in the cost of the food. There is no difference in that and a 15% tip based on the cost of the food, except one goes in the hands of the operator and the other goes into the hands of the server. You are still paying for service based on the cost, you are just fooling yourself by making it invisible. At least in my system, if the service is terrible, there is a way to negotiate the amount you are paying for service. With your system, you are stuck with it. [NEWLINE] [NEWLINE] [STARTQ] Employee tips can be affected by the quality and timliness of the cooks, whether the restaurant is understaffed, whether they have to share the tips with kitchen staff or with managemen. It should only be based on the work of the server [ENDQ] [NEWLINE] Most establishments require the server to share a percentage of their tips with the back of the house employees. They are compelled to make service as good as possible because of this incentive. If it's a management or staffing problem, my system mitigates that by including gratuity on all bills. The server isn't harmed, but in fact is compensated directly for having to increase productivity to make up for poor management, unless someone asks for a discount on the gratuity. In your system, the employer actually benefits from being understaffed. They get to collect the 15% increase on the price of the food and pay fewer servers than they need, making extra profits from managing poorly and making their servers work harder without extra compensation.</s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Cyclists seem to be demanding the road be shared<mask> with special treatment. Essentially this is nothing more than radical feminism of the road...equality<mask> with special treatment. Compromise lies everywhere.<mask> you want to be treated like another vehicle on the road, you must do just that. [NEWLINE] [NEWLINE] Not only are your points hazardous to everyone else on the road, they are hazardous to your wellbeing. It doesn't make sense to say biking is better for your health<mask> you practice it in a manner that is unsafe and can kill you. That argument you make comes off<mask> extremely counterproductive. [NEWLINE] [NEWLINE] <mask> all of these changes you want seem to be for the sake of minor convenience<mask> opposed to Public Safety, by that logic joggers should have more special treatment and cyclists<mask> jogging is healthier than cycling and less convenient for travel that a bicycle is.<mask> does that mean that joggers should be able to jog in the middle of the road, get their own jogging lane<mask> a sidewalk is not available, be able to cut across traffic<mask><mask><mask> color the light is? Basically it's nothing more than "treat me special<mask> it's less convenient for me". [NEWLINE] [NEWLINE] With that said, to be blunt, it seems that you're doing nothing more than venting instead of wanting your view changed. people are giving good argument,<mask> repeated quite a bit<mask> everybody else shares common sense of safety. I'm sorry<mask> stopping on your bike at every stop sign or light is uncomfortable or inconvenient for you,<mask> that's the choice you made by biking. I'm sure there are different positions you can stand in or can just get off of your bike<mask> it's that uncomfortable for you to stand over it at a stop.  People on motorcycles need to stop and lean over on one leg just like you do and nobody complains about it. it's just the way your vehicle happens to work.<mask> you really don't like it that much,<mask> ridiculous<mask> it sounds, just get training wheels or a more durable kickstand and you can continue to sit in the same position upon your bike at a stop. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Cyclists seem to be demanding the road be shared but with special treatment. Essentially this is nothing more than radical feminism of the road...equality but with special treatment. Compromise lies everywhere. If you want to be treated like another vehicle on the road, you must do just that. [NEWLINE] [NEWLINE] Not only are your points hazardous to everyone else on the road, they are hazardous to your wellbeing. It doesn't make sense to say biking is better for your health when you practice it in a manner that is unsafe and can kill you. That argument you make comes off as extremely counterproductive. [NEWLINE] [NEWLINE] Since all of these changes you want seem to be for the sake of minor convenience as opposed to Public Safety, by that logic joggers should have more special treatment and cyclists because jogging is healthier than cycling and less convenient for travel that a bicycle is. so does that mean that joggers should be able to jog in the middle of the road, get their own jogging lane when a sidewalk is not available, be able to cut across traffic regardless of what color the light is? Basically it's nothing more than "treat me special because it's less convenient for me". [NEWLINE] [NEWLINE] With that said, to be blunt, it seems that you're doing nothing more than venting instead of wanting your view changed. people are giving good argument, although repeated quite a bit because everybody else shares common sense of safety. I'm sorry if stopping on your bike at every stop sign or light is uncomfortable or inconvenient for you, but that's the choice you made by biking. I'm sure there are different positions you can stand in or can just get off of your bike if it's that uncomfortable for you to stand over it at a stop.  People on motorcycles need to stop and lean over on one leg just like you do and nobody complains about it. it's just the way your vehicle happens to work. If you really don't like it that much, as ridiculous as it sounds, just get training wheels or a more durable kickstand and you can continue to sit in the same position upon your bike at a stop. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The *State* (entity)<mask> is understood in most democratic legal systems is a representation of the entire populace of a country (or political division or populace or whatever). It acts representing the collective will of that populace, by the means legislated within. [NEWLINE] [NEWLINE] The State should act<mask><mask><mask> the biggest amount of people want it to do. It is NOT  group of people that rule another,<mask> many people are selected to operate it. [NEWLINE] [NEWLINE] <mask> it embodies my will and acts on<mask> is truly the best for me, along that of many others I cannot be subject to the State's will, unless it corresponds to the general public's. [NEWLINE] [NEWLINE] I mean that a decision taken by a small number of people for no other reason<mask> to gain power or control, and in no way distilled from the collective will of said populace is more reminiscent of an Oligarchy, were appointed leaders get to have *all* the power and not a Democracy,<mask> many more appointed functionaries carry on the state's functions. [NEWLINE] [NEWLINE] The state should have no privacy,<mask> it is not a person.<mask> the state is a person, or many people instead of *all* people, then the whole power over the state (all the people) is compressed into a single or few individuals. [NEWLINE] [NEWLINE] This is not a trait of a democratical government. It is more reminiscent of a Monarchical or Oligarchical one. [NEWLINE] [NEWLINE] To say that the state needs *privacy* is to imply that the state is somebody other than me.<mask> I should be aware of any and every decisions or actions I take, right? [NEWLINE] [NEWLINE] And these sort of governments we live in (I'm not from the US<mask><mask> live in one) are Representative Democracies. In such, the State (entity) is acted by a cycling group of elected people that *represent* the entire populace's will and have to act<mask><mask> this will,<mask> have no significant power over the populace itself. [NEWLINE] [NEWLINE] <mask><mask> the state takes an action that I am not aware of, it is not representing me. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>The *State* (entity) as is understood in most democratic legal systems is a representation of the entire populace of a country (or political division or populace or whatever). It acts representing the collective will of that populace, by the means legislated within. [NEWLINE] [NEWLINE] The State should act according to what the biggest amount of people want it to do. It is NOT  group of people that rule another, although many people are selected to operate it. [NEWLINE] [NEWLINE] If it embodies my will and acts on what is truly the best for me, along that of many others I cannot be subject to the State's will, unless it corresponds to the general public's. [NEWLINE] [NEWLINE] I mean that a decision taken by a small number of people for no other reason as to gain power or control, and in no way distilled from the collective will of said populace is more reminiscent of an Oligarchy, were appointed leaders get to have *all* the power and not a Democracy, where many more appointed functionaries carry on the state's functions. [NEWLINE] [NEWLINE] The state should have no privacy, as it is not a person. If the state is a person, or many people instead of *all* people, then the whole power over the state (all the people) is compressed into a single or few individuals. [NEWLINE] [NEWLINE] This is not a trait of a democratical government. It is more reminiscent of a Monarchical or Oligarchical one. [NEWLINE] [NEWLINE] To say that the state needs *privacy* is to imply that the state is somebody other than me. Because I should be aware of any and every decisions or actions I take, right? [NEWLINE] [NEWLINE] And these sort of governments we live in (I'm not from the US but also live in one) are Representative Democracies. In such, the State (entity) is acted by a cycling group of elected people that *represent* the entire populace's will and have to act according to this will, but have no significant power over the populace itself. [NEWLINE] [NEWLINE] Thus if the state takes an action that I am not aware of, it is not representing me. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] it carries the implication that an entire class of people should be held accountable for the actions of a subset. [ENDQ] [NEWLINE] There are two senses of "accountable" that need to be separated: [NEWLINE] [NEWLINE] (1) Accountability<mask> deserving of punishment. [NEWLINE] [NEWLINE] (2) Accountability<mask> having an obligation to do something. [NEWLINE] [NEWLINE] You are correct that an entire class of people is not deserving of punishment for the injustices committed by a few.<mask> you think anybody disagrees with you about this, then you must really think that they are stupid and evil. Perhaps there are some people this stupid and evil--people who believe in collective punishment--<mask> you are radically, totally, mistaken<mask> you think that most people who talk about privilege endorse collective punishment. [NEWLINE] [NEWLINE] <mask>, even<mask> an entire class is not deserving of punishment, it might still be true that an entire class has an obligation to do something to fix the situation. For example, it's not my fault that my neighbor's house burned down last year,<mask> I shouldn't be punished for it.<mask>,<mask><mask> I'm not blameworthy, I still have an obligation to subsidize the fire department's expenses for putting out my neighbor's fire. Most taxes work like this. [NEWLINE] [NEWLINE] Many people discussing privilege are not trying to discuss either kind of responsibility. They're just trying to discuss people's psychology and the information that different people have differential access to. Some people, like me, who discuss privilege are trying to discuss (2)--<mask> obligations the privileged have. My house wasn't struck by lightning,<mask> I was "fire privileged" (oops, I'm a horrible person and a parasite on society again),<mask> I have an obligation to help out. Likewise, everybody--most of whom are hearing--need to subsidize anti-discrimination litigation from the Department of Justice to help your deaf friend. And there's more to do. [NEWLINE] [NEWLINE] <mask>, in a sense, we are all accountable for the actions of a subset. (Just like we all have to pay for the arrest and incarceration of criminals--<mask><mask> we're not all guilty.)</s>
Label encoding: <s> [STARTQ] it carries the implication that an entire class of people should be held accountable for the actions of a subset. [ENDQ] [NEWLINE] There are two senses of "accountable" that need to be separated: [NEWLINE] [NEWLINE] (1) Accountability as deserving of punishment. [NEWLINE] [NEWLINE] (2) Accountability as having an obligation to do something. [NEWLINE] [NEWLINE] You are correct that an entire class of people is not deserving of punishment for the injustices committed by a few. If you think anybody disagrees with you about this, then you must really think that they are stupid and evil. Perhaps there are some people this stupid and evil--people who believe in collective punishment-- but you are radically, totally, mistaken if you think that most people who talk about privilege endorse collective punishment. [NEWLINE] [NEWLINE] However, even if an entire class is not deserving of punishment, it might still be true that an entire class has an obligation to do something to fix the situation. For example, it's not my fault that my neighbor's house burned down last year, so I shouldn't be punished for it. However, even though I'm not blameworthy, I still have an obligation to subsidize the fire department's expenses for putting out my neighbor's fire. Most taxes work like this. [NEWLINE] [NEWLINE] Many people discussing privilege are not trying to discuss either kind of responsibility. They're just trying to discuss people's psychology and the information that different people have differential access to. Some people, like me, who discuss privilege are trying to discuss (2)-- what obligations the privileged have. My house wasn't struck by lightning, so I was "fire privileged" (oops, I'm a horrible person and a parasite on society again), but I have an obligation to help out. Likewise, everybody--most of whom are hearing--need to subsidize anti-discrimination litigation from the Department of Justice to help your deaf friend. And there's more to do. [NEWLINE] [NEWLINE] So, in a sense, we are all accountable for the actions of a subset. (Just like we all have to pay for the arrest and incarceration of criminals-- even though we're not all guilty.)</s>
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0409, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I've admitted that I am no poetry expert,<mask> I have been exposed to poetry that I have appreciated.<mask> I read "The Raven" in highschool I was particularly impressed. It seemed like Poe could effortlessly wrestle with the language and make and make it do his bidding. I was involved in a production of Romeo and Juliet last summer and was absolutely amazed at the wordplay. [ENDQ] [NEWLINE] This is really important for you to take to heart. Your understanding of poetry is very elementary, and I'm guessing you are too young to have a good understanding of the context of her writing. Essentially it's like some old guy claiming to have a valid opinion on contemporary rock music<mask> all they've been exposed to is "The Eagles Greatest Hits." A more helpful attitude for you at this point would involve suspending judgment, and collecting information about<mask> she's talking about, and<mask>. [NEWLINE] [NEWLINE] <mask> this sounds harsh, it<mask> I feel like I'm pretty much in the same boat<mask> you -- I don't read much poetry, and don't much care for most of<mask> I read, outside of song lyrics. I don't judge<mask>,<mask> I know the limitations of my knowledge. This is a *really* useful skill to develop -- knowing and acknowledging your own weaknesses. Concentrating on adapting *yourself* to understand and appreciate<mask> you see is a lovely growing process<mask>, and is a great way of opening doors to pleasurable things that might not present themselves readily without a bit of prior education. Acquired tastes are wonderful. [NEWLINE] [NEWLINE] <mask>,<mask> you want my two cents<mask> a fellow cellar-dweller,<mask><mask> you're working with an overly restricted view of<mask> an artist like Angelou did. The category "poet" does not accurately describe her cultural and historical placement in the same way that it describes Poe or Shakespeare. I believe the parts of her efforts that exist outside of the "poet" category are very relevant an understanding of<mask> she was going for. Perhaps a more in-depth study of those things will increase your ability to appreciate her work. </s>
Label encoding: <s> [STARTQ] I've admitted that I am no poetry expert, but I have been exposed to poetry that I have appreciated. When I read "The Raven" in highschool I was particularly impressed. It seemed like Poe could effortlessly wrestle with the language and make and make it do his bidding. I was involved in a production of Romeo and Juliet last summer and was absolutely amazed at the wordplay. [ENDQ] [NEWLINE] This is really important for you to take to heart. Your understanding of poetry is very elementary, and I'm guessing you are too young to have a good understanding of the context of her writing. Essentially it's like some old guy claiming to have a valid opinion on contemporary rock music when all they've been exposed to is "The Eagles Greatest Hits." A more helpful attitude for you at this point would involve suspending judgment, and collecting information about what she's talking about, and why. [NEWLINE] [NEWLINE] If this sounds harsh, it because I feel like I'm pretty much in the same boat as you -- I don't read much poetry, and don't much care for most of what I read, outside of song lyrics. I don't judge though, since I know the limitations of my knowledge. This is a *really* useful skill to develop -- knowing and acknowledging your own weaknesses. Concentrating on adapting *yourself* to understand and appreciate what you see is a lovely growing process however, and is a great way of opening doors to pleasurable things that might not present themselves readily without a bit of prior education. Acquired tastes are wonderful. [NEWLINE] [NEWLINE] Also, if you want my two cents as a fellow cellar-dweller, I think you're working with an overly restricted view of what an artist like Angelou did. The category "poet" does not accurately describe her cultural and historical placement in the same way that it describes Poe or Shakespeare. I believe the parts of her efforts that exist outside of the "poet" category are very relevant an understanding of what she was going for. Perhaps a more in-depth study of those things will increase your ability to appreciate her work. </s>
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>OK, devil's advocate here: [NEWLINE] [NEWLINE] 1) There are more women than men that take a significant amount of maternity leave, at least for a period of time after the child is born.<mask> the mother may go back to work at some point after the birth, the "safe bet" (discrimination lawsuits notwithstanding) for OP is to assume that a woman of child-rearing age would take the maximum allowed maternity leave before returning to work, which is at least 12 weeks (unpaid) in the US (sometimes more, depending on state laws);<mask> an employee decides to come back after those 12 weeks, the employer is required to make provisions for their continued employment with the company. [NEWLINE] [NEWLINE] 2) A father may elect to stay at home with his kids,<mask> OP has no obligation to pay said father unless they have committed to do<mask> in his contract,<mask> laws regarding paid maternity/paternity leave are largely decided at a state level. It is<mask> less likely that a father will take more than a few weeks/months off after a child birth than the equivalent female worker. [NEWLINE] [NEWLINE] 3) I'll admit I don't quite understand the sexual harassment angle. I am inclined to think that OP's company deals with a fair number of socially inept employees who do not quite "get" the concept of sexual harrassment, and<mask> such, would be prone to break the rules should the opportunity present itself. [NEWLINE] [NEWLINE] <mask> OP is perpetuating the problem, he really didn't create it.<mask> there were more equitable laws regarding parental leave in the US, it wouldn't matter whether he hires a man or a woman; the fact of the matter is that<mask> the Family and Medical Leave Act of 1993 only requires unpaid leave allowances, and is geared mainly toward those activities a mother is more likely to perform (childbirth, pregnancy, family health issues, etc.), it essentially puts women at a disadvantage through the hiring process,<mask> employers go into interviews with that knowledge.<mask>, OP is doing<mask> he feels is best for his company,<mask> putting potential employment candidates at a deliberate disadvantage.</s>
Label encoding: <s>OK, devil's advocate here: [NEWLINE] [NEWLINE] 1) There are more women than men that take a significant amount of maternity leave, at least for a period of time after the child is born. While the mother may go back to work at some point after the birth, the "safe bet" (discrimination lawsuits notwithstanding) for OP is to assume that a woman of child-rearing age would take the maximum allowed maternity leave before returning to work, which is at least 12 weeks (unpaid) in the US (sometimes more, depending on state laws); if an employee decides to come back after those 12 weeks, the employer is required to make provisions for their continued employment with the company. [NEWLINE] [NEWLINE] 2) A father may elect to stay at home with his kids, but OP has no obligation to pay said father unless they have committed to do so in his contract, although laws regarding paid maternity/paternity leave are largely decided at a state level. It is also less likely that a father will take more than a few weeks/months off after a child birth than the equivalent female worker. [NEWLINE] [NEWLINE] 3) I'll admit I don't quite understand the sexual harassment angle. I am inclined to think that OP's company deals with a fair number of socially inept employees who do not quite "get" the concept of sexual harrassment, and as such, would be prone to break the rules should the opportunity present itself. [NEWLINE] [NEWLINE] While OP is perpetuating the problem, he really didn't create it. If there were more equitable laws regarding parental leave in the US, it wouldn't matter whether he hires a man or a woman; the fact of the matter is that since the Family and Medical Leave Act of 1993 only requires unpaid leave allowances, and is geared mainly toward those activities a mother is more likely to perform (childbirth, pregnancy, family health issues, etc.), it essentially puts women at a disadvantage through the hiring process, because employers go into interviews with that knowledge. Thus, OP is doing what he feels is best for his company, while putting potential employment candidates at a deliberate disadvantage.</s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm not a very spiritual person. And I don't believe in a lot 'outside' of<mask> we can percieve. Yes I believe there's more to the world,<mask> I don't believe in anything beyond biological sex. Of course it's not binary. Intersex people very obviously exist, that's just scientific fact. [ENDQ] [NEWLINE] I'm having trouble understanding<mask> exactly your view is,<mask> I'm sure you'll understand seeing<mask> you acknowledged that it's rather vague and complicated.<mask> your use of the word'spiritual' is intriguing and<mask><mask> that might be<mask> the 'problem' (<mask> I can use that word) with your view is. [NEWLINE] [NEWLINE] Your use of'spiritual' seems to imply that you see gender (<mask> opposed to biological sex,<mask><mask><mask> with that distinction for other reasons)<mask> a *metaphysical* concept. This isn't the case, at least in most academic circles: it's a *social* concept. [NEWLINE] [NEWLINE] You seem to have empiricist leanings,<mask> I feel like you've taken that a step beyond<mask> is reasonable:<mask> there is something inherently anti-empirical about metaphysical or spiritual claims, and<mask> an empiricist would be sceptical of them, you've extended this scepticism to all non-physical claims, which isn't reasonable<mask> there's clearly plenty of social and mental phenomena which are empirical<mask> not physical. Take art, for instance. It's pretty clear that there is no essential physical property to<mask> we call 'art',<mask> that doesn't mean we can't study art and make objective statements about art beyond its physical attributes (we can talk about the role of art in society, or the history of art, or<mask> art is influential etc.) [NEWLINE] [NEWLINE] Art, then, is clearly an abstract concept in that it's not a physical concept. This doesn't make it 'not real',<mask>. Do you see<mask> there's a parallel with gender there? Or have I missed your view/this doesn't actually touch on<mask> I assumed you were assuming? [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] I'm not a very spiritual person. And I don't believe in a lot 'outside' of what we can percieve. Yes I believe there's more to the world, but I don't believe in anything beyond biological sex. Of course it's not binary. Intersex people very obviously exist, that's just scientific fact. [ENDQ] [NEWLINE] I'm having trouble understanding what exactly your view is, as I'm sure you'll understand seeing as you acknowledged that it's rather vague and complicated. But your use of the word'spiritual' is intriguing and I think that might be where the 'problem' ( if I can use that word) with your view is. [NEWLINE] [NEWLINE] Your use of'spiritual' seems to imply that you see gender ( as opposed to biological sex, although I disagree with that distinction for other reasons) as a *metaphysical* concept. This isn't the case, at least in most academic circles: it's a *social* concept. [NEWLINE] [NEWLINE] You seem to have empiricist leanings, but I feel like you've taken that a step beyond what is reasonable: although there is something inherently anti-empirical about metaphysical or spiritual claims, and so an empiricist would be sceptical of them, you've extended this scepticism to all non-physical claims, which isn't reasonable since there's clearly plenty of social and mental phenomena which are empirical but not physical. Take art, for instance. It's pretty clear that there is no essential physical property to what we call 'art', but that doesn't mean we can't study art and make objective statements about art beyond its physical attributes (we can talk about the role of art in society, or the history of art, or what art is influential etc.) [NEWLINE] [NEWLINE] Art, then, is clearly an abstract concept in that it's not a physical concept. This doesn't make it 'not real', though. Do you see how there's a parallel with gender there? Or have I missed your view/this doesn't actually touch on what I assumed you were assuming? [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Almost ALL laws are based on extremist expectations. [NEWLINE] Regular people don't murder others. <mask><mask> have laws against murder? [NEWLINE] Regular people do not rob banks. <mask><mask> have laws that make bank robbery illegal? [NEWLINE] People don't look for reasons to hate each other. <mask> they DO look for ways to feel superior to others.  And discrimination is a part of that feeling of superiority.  Or do exclusive clubs, fashion label clothing, specific car brands, and all the other cache that comes with being in that exclusive group just a bunch of hogwash?  Advertising today tells us it is very much a human motivation. <mask> does religious persecution, pogroms, nationalism, and most other -ism in the past. <mask> you are not part of the In Club, you suffer the consequences. [NEWLINE] And you're missing the point of my post.  The fact that CFA has been boycotted in some areas and is booming in others encourages those who hold the same belief to EXPAND their actions to other avenues.  It doesn't START<mask> extremism. <mask><mask> it builds to it. [NEWLINE] In Wisconsin, a pharmacist refused to give a woman a prescription for her birth control<mask> HE disagreed with her having it.<mask> she asked for the prescription back, he refused.  There's now a state law granting him immunity,<mask> the fact that the pharmacist now exercises veto power over the private decisions of a woman and her doctor. <mask> THE PHARMACIST'S religious beliefs are protected. [NEWLINE] Your example of the community being able to tell the business to fuck right off ignores a simple issue. [NEWLINE] That group may not be the majority. [NEWLINE] And we're right back to<mask> I started.  A (minority) group can suffer consequences from a different (majority) group's displeasure in business.  And that is the foundation of oppression.  Restricting the free access of the minority to the expression of their life, liberty, and pursuit of happiness. [NEWLINE] And that does not serve society.  It harms it.  </s>
Label encoding: <s>Almost ALL laws are based on extremist expectations. [NEWLINE] Regular people don't murder others.  So why have laws against murder? [NEWLINE] Regular people do not rob banks.  So why have laws that make bank robbery illegal? [NEWLINE] People don't look for reasons to hate each other.  But they DO look for ways to feel superior to others.  And discrimination is a part of that feeling of superiority.  Or do exclusive clubs, fashion label clothing, specific car brands, and all the other cache that comes with being in that exclusive group just a bunch of hogwash?  Advertising today tells us it is very much a human motivation.  As does religious persecution, pogroms, nationalism, and most other -ism in the past.  If you are not part of the In Club, you suffer the consequences. [NEWLINE] And you're missing the point of my post.  The fact that CFA has been boycotted in some areas and is booming in others encourages those who hold the same belief to EXPAND their actions to other avenues.  It doesn't START as extremism.  BUt it builds to it. [NEWLINE] In Wisconsin, a pharmacist refused to give a woman a prescription for her birth control because HE disagreed with her having it. When she asked for the prescription back, he refused.  There's now a state law granting him immunity, despite the fact that the pharmacist now exercises veto power over the private decisions of a woman and her doctor.  But THE PHARMACIST'S religious beliefs are protected. [NEWLINE] Your example of the community being able to tell the business to fuck right off ignores a simple issue. [NEWLINE] That group may not be the majority. [NEWLINE] And we're right back to where I started.  A (minority) group can suffer consequences from a different (majority) group's displeasure in business.  And that is the foundation of oppression.  Restricting the free access of the minority to the expression of their life, liberty, and pursuit of happiness. [NEWLINE] And that does not serve society.  It harms it.  </s>
Loss: tensor(0.0392, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm fighting for gay rights. They need their own movement. [ENDQ] [NEWLINE] Stonewall was full of trans* people both before and during the raid that led to the riots.  The movement<mask> we know it wouldn't have have started<mask> it weren't for trans* people.  In the 80s, trans* people were<mask> fighting for AIDS activism.  In the 90s, they were told to support the push for marriage, that trans* issues would be included<mask> that fight was won.  We donated.  We campaigned.  We picketed, even<mask> we were told we were not welcome.  And now we're seeing the gay lobby seemingly abandoning us, telling us to get our own movement, to get in a hole with a deeper stigma and less money and political power after we have given<mask> much to help you get married. [NEWLINE] [NEWLINE] Not to mention that most of us have spent time being gay, either before or after transition--sometimes both.  Throughout many cultures and many time periods, alternative sexualities and alternative expressions of gender were strongly linked.  Even<mask> you want to make the case that they don't necessarily need to be, an appeal to history like the one you make needs to be taken into account--<mask> a full history that includes the birth of the LGBT movement in the abuse of Stonewall that included a disproportionate number of trans* people. [NEWLINE] [NEWLINE] And, well, in the eyes of the straight, cis people who make life shitty for both of us, the two groups are linked.  Trying to push us away *now* after we have fought with and for you feels like you're trying to wash your hands of something dirty after you got<mask> you wanted out of us. [NEWLINE] [NEWLINE] It makes me<mask> sad to see you take this attitude, and it goes deeper than a sense of betrayal.  It just makes me... sorrowful.  *We want you to succeed, too.* <mask> do you feel a sense of hate or scorn that's<mask> strong you need to actively work against trans* people?</s>
Label encoding: <s> [STARTQ] I'm fighting for gay rights. They need their own movement. [ENDQ] [NEWLINE] Stonewall was full of trans* people both before and during the raid that led to the riots.  The movement as we know it wouldn't have have started if it weren't for trans* people.  In the 80s, trans* people were also fighting for AIDS activism.  In the 90s, they were told to support the push for marriage, that trans* issues would be included when that fight was won.  We donated.  We campaigned.  We picketed, even when we were told we were not welcome.  And now we're seeing the gay lobby seemingly abandoning us, telling us to get our own movement, to get in a hole with a deeper stigma and less money and political power after we have given so much to help you get married. [NEWLINE] [NEWLINE] Not to mention that most of us have spent time being gay, either before or after transition--sometimes both.  Throughout many cultures and many time periods, alternative sexualities and alternative expressions of gender were strongly linked.  Even if you want to make the case that they don't necessarily need to be, an appeal to history like the one you make needs to be taken into account-- but a full history that includes the birth of the LGBT movement in the abuse of Stonewall that included a disproportionate number of trans* people. [NEWLINE] [NEWLINE] And, well, in the eyes of the straight, cis people who make life shitty for both of us, the two groups are linked.  Trying to push us away *now* after we have fought with and for you feels like you're trying to wash your hands of something dirty after you got what you wanted out of us. [NEWLINE] [NEWLINE] It makes me so sad to see you take this attitude, and it goes deeper than a sense of betrayal.  It just makes me... sorrowful.  *We want you to succeed, too.*  Why do you feel a sense of hate or scorn that's so strong you need to actively work against trans* people?</s>
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As a gay man, I can be extremely wary of other people for my own protection. I've always felt that homophobia is mainly a symptom of the willingness to hate other groups of people they have never met or gotten to know.<mask><mask> I hear someone say "I used to be homophobic until [insert "my son came out," or "my friend came out" or etc.]" I assume they still are willing to hate groups of people they haven't met, it's just now that they know a gay person "gay people" no longer fall under that category. [NEWLINE] [NEWLINE] <mask> I am outwardly supportive<mask> I run into these folks, inwardly I am uncomfortable to the point of being repulsed. Obviously it wouldn't be good to push for gay rights and then effectively punish homophobes for changing their minds towards being supportive,<mask> it's hard to think in any way<mask> "these people are still severely prejudiced or hateful, just not specifically towards me anymore, avoid avoid avoid!" Please CMV! [NEWLINE] [NEWLINE] Edit: In order to address one of the rules in this sub, I should say I am at work, and<mask> replying to people will probably have to wait 4-5 hours,<mask> it will happen I promise :) [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>As a gay man, I can be extremely wary of other people for my own protection. I've always felt that homophobia is mainly a symptom of the willingness to hate other groups of people they have never met or gotten to know. Thus when I hear someone say "I used to be homophobic until [insert "my son came out," or "my friend came out" or etc.]" I assume they still are willing to hate groups of people they haven't met, it's just now that they know a gay person "gay people" no longer fall under that category. [NEWLINE] [NEWLINE] While I am outwardly supportive when I run into these folks, inwardly I am uncomfortable to the point of being repulsed. Obviously it wouldn't be good to push for gay rights and then effectively punish homophobes for changing their minds towards being supportive, but it's hard to think in any way besides "these people are still severely prejudiced or hateful, just not specifically towards me anymore, avoid avoid avoid!" Please CMV! [NEWLINE] [NEWLINE] Edit: In order to address one of the rules in this sub, I should say I am at work, and so replying to people will probably have to wait 4-5 hours, but it will happen I promise :) [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Neither the act of earning a low wage nor the act of "having" money has much impact on society, and there's little you can conclude about a person's economic or societal value based on those facts in a vacuum. [NEWLINE] [NEWLINE] <mask>, it's not unreasonable to make assumptions about someone's behavior and attributes based on his income/wealth -- that's<mask> you're doing here.  For example, you're<mask><mask> a low-wage earner is likely to spend his money.  That's probably a supportable assumption that would be validated by almost any dataset you collect. [NEWLINE] [NEWLINE] <mask> there are other correlations we can posit, too, that would probably turn out to be true<mask> we examined the data.   Compared to the guy making $10k, the guy making or poessessing $10m: [NEWLINE] [NEWLINE] * Probably has more education [NEWLINE] * Probably has more "natural" cognitive ability (i.e., the heritable/innate component of IQ unaffected by environment) [NEWLINE] * Is probably a better parent whose kids are less likley to end up in jail [NEWLINE] * Is, himself, less likely to commit crimes (white collar, violent or otherwise) [NEWLINE] * Is more likely to make a valuable, impactful contribution to science, art, public policy, medicine, law, or basically any important field.  In part this is<mask> he's likelier to have a track record of already having made such contributions. [NEWLINE] * In the same vein<mask> the above -- he's more likely to generate value for his employer or shareholders that meets or exceeds whatever comp he's paid (or was paid in the past before retiring rich).  Even<mask> his wealth is inherited, the advantages he was raised with make him far likelier to end up<mask> an elite societal contributor than the offpsring of your average janitor. [NEWLINE] [NEWLINE] Make no mistake:<mask> you could magically push a button and delete, from the universe, either 1000 of America's richest or 1000 of America's poorest adults, deleting the rich would be much likelier to have an adverse effect. </s>
Label encoding: <s>Neither the act of earning a low wage nor the act of "having" money has much impact on society, and there's little you can conclude about a person's economic or societal value based on those facts in a vacuum. [NEWLINE] [NEWLINE] However, it's not unreasonable to make assumptions about someone's behavior and attributes based on his income/wealth -- that's what you're doing here.  For example, you're assuming that a low-wage earner is likely to spend his money.  That's probably a supportable assumption that would be validated by almost any dataset you collect. [NEWLINE] [NEWLINE] But there are other correlations we can posit, too, that would probably turn out to be true if we examined the data.   Compared to the guy making $10k, the guy making or poessessing $10m: [NEWLINE] [NEWLINE] * Probably has more education [NEWLINE] * Probably has more "natural" cognitive ability (i.e., the heritable/innate component of IQ unaffected by environment) [NEWLINE] * Is probably a better parent whose kids are less likley to end up in jail [NEWLINE] * Is, himself, less likely to commit crimes (white collar, violent or otherwise) [NEWLINE] * Is more likely to make a valuable, impactful contribution to science, art, public policy, medicine, law, or basically any important field.  In part this is because he's likelier to have a track record of already having made such contributions. [NEWLINE] * In the same vein as the above -- he's more likely to generate value for his employer or shareholders that meets or exceeds whatever comp he's paid (or was paid in the past before retiring rich).  Even if his wealth is inherited, the advantages he was raised with make him far likelier to end up as an elite societal contributor than the offpsring of your average janitor. [NEWLINE] [NEWLINE] Make no mistake: if you could magically push a button and delete, from the universe, either 1000 of America's richest or 1000 of America's poorest adults, deleting the rich would be much likelier to have an adverse effect. </s>
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0408, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay thanks, I see<mask> you are saying now. [NEWLINE] [NEWLINE] <mask><mask> with your point. I do think that *some* of the responsibility lies with the victim, even<mask> the police officer is totally incorrect. Even in the most recent case,<mask> the officer is clearly in the wrong, I believe the victim would still be alive<mask> they had remained in their vehicle<mask> the officer instructed. You can't say that the victim did "nothing wrong",<mask> disobeying the order to remain in their vehicle was wrong. [NEWLINE] [NEWLINE] Take your jaywalking example,<mask> an officer simply shot me for jaywalking without issuing any commands obviously that officer just murdered me.<mask>, it wouldn't had happened<mask> I had not been jaywalking. This doesn't mean that reform with the police isn't needed,<mask> it does mean that for my own personal safety it is better to not jaywalk than to jaywalk in most circumstances (clearly<mask> I was like running away from danger or something jaywalking would be understandable). [NEWLINE] [NEWLINE] Specific to jaywalking, this is outside the point of the thread<mask> worth mentioning. People in the city don't seem to realize<mask> impactful jaywalking can be.<mask> I begin to cross the street<mask> the sign has changed from "walk" to blinking red and it is perpendicular to a road<mask> cars are trying to make right turns I'm having an extreme effect on traffic. Instead of being able to make a right turn on green, now all of the cars have to wait for every pedestrian to finish. By the time they do finish, the light is turning red and they will never be clear to take a right turn.<mask><mask><mask>, only one car can turn right at a time, and the cars on the road they are turning into have to wait. This causes compounding delays until eventually traffic is delayed 20+ minutes more than it should be. [NEWLINE] [NEWLINE] <mask> a commuter, this is *extremely* annoying and I wish more officers would be at busy intersections writing tickets to people that do this.</s>
Label encoding: <s>Okay thanks, I see what you are saying now. [NEWLINE] [NEWLINE] I disagree with your point. I do think that *some* of the responsibility lies with the victim, even if the police officer is totally incorrect. Even in the most recent case, where the officer is clearly in the wrong, I believe the victim would still be alive if they had remained in their vehicle as the officer instructed. You can't say that the victim did "nothing wrong", because disobeying the order to remain in their vehicle was wrong. [NEWLINE] [NEWLINE] Take your jaywalking example, if an officer simply shot me for jaywalking without issuing any commands obviously that officer just murdered me. However, it wouldn't had happened if I had not been jaywalking. This doesn't mean that reform with the police isn't needed, but it does mean that for my own personal safety it is better to not jaywalk than to jaywalk in most circumstances (clearly if I was like running away from danger or something jaywalking would be understandable). [NEWLINE] [NEWLINE] Specific to jaywalking, this is outside the point of the thread but worth mentioning. People in the city don't seem to realize how impactful jaywalking can be. If I begin to cross the street when the sign has changed from "walk" to blinking red and it is perpendicular to a road where cars are trying to make right turns I'm having an extreme effect on traffic. Instead of being able to make a right turn on green, now all of the cars have to wait for every pedestrian to finish. By the time they do finish, the light is turning red and they will never be clear to take a right turn. As a result, only one car can turn right at a time, and the cars on the road they are turning into have to wait. This causes compounding delays until eventually traffic is delayed 20+ minutes more than it should be. [NEWLINE] [NEWLINE] As a commuter, this is *extremely* annoying and I wish more officers would be at busy intersections writing tickets to people that do this.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0322, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Depends on<mask> you would describe<mask> "not<mask> bad". Having to hide your status<mask> Hibakusha<mask> your kids and grandkids might not find anyone to marry is bad in my eyes. It's not like Japan would be nice to people who are deemed to be different. Even Burakumin are discriminated against and that is really stupid and pointless. [NEWLINE] [NEWLINE] Looking at these incidence rates doesn't really work unless it's dramatically horrible. I mean, thousands over thousands of people get cancer every year. It's nothing special to have cancer.<mask>, even<mask> they had a increased rate of cancer by 5-10%, that might be "natural" fluctuations. Across country you probably have these fluctuations due to all kinds of things. Unless you get insane numbers of cancers (i.e. people dying like flies) you won't be able to prove anything. Even something on the scale of Chernobyl is pretty much not measurable: [NEWLINE] [NEWLINE] "On the death toll of the accident, the report states that twenty-eight emergency workers ("liquidators") died from acute radiation syndrome including beta burns and 15 patients died from thyroid cancer in the following years, and it roughly estimated that cancer deaths caused by Chernobyl may reach a total of about 4,000 among the 5 million persons residing in the contaminated areas, the report projected cancer mortality "increases of less than one per cent" (~0.3%) on a time span of 80 years, cautioning that this estimate was "speculative"<mask> at this time only a few cancer deaths are linked to the Chernobyl disaster.[132] The report says it is impossible to reliably predict the number of fatal cancers arising from the incident<mask> small differences in assumptions can result in large differences in the estimated health costs. The report says it represents the consensus view of the eight UN organisations." [NEWLINE] [NEWLINE] <mask> that doesn't mean nothing ever happened. Decades later you can't eat venison and mushrooms in certain parts of Europe, thanks to radiation.<mask> of course, no effects on the people living there. Same for Fukushima....</s>
Label encoding: <s>Depends on what you would describe as "not so bad". Having to hide your status as Hibakusha because your kids and grandkids might not find anyone to marry is bad in my eyes. It's not like Japan would be nice to people who are deemed to be different. Even Burakumin are discriminated against and that is really stupid and pointless. [NEWLINE] [NEWLINE] Looking at these incidence rates doesn't really work unless it's dramatically horrible. I mean, thousands over thousands of people get cancer every year. It's nothing special to have cancer. So, even if they had a increased rate of cancer by 5-10%, that might be "natural" fluctuations. Across country you probably have these fluctuations due to all kinds of things. Unless you get insane numbers of cancers (i.e. people dying like flies) you won't be able to prove anything. Even something on the scale of Chernobyl is pretty much not measurable: [NEWLINE] [NEWLINE] "On the death toll of the accident, the report states that twenty-eight emergency workers ("liquidators") died from acute radiation syndrome including beta burns and 15 patients died from thyroid cancer in the following years, and it roughly estimated that cancer deaths caused by Chernobyl may reach a total of about 4,000 among the 5 million persons residing in the contaminated areas, the report projected cancer mortality "increases of less than one per cent" (~0.3%) on a time span of 80 years, cautioning that this estimate was "speculative" since at this time only a few cancer deaths are linked to the Chernobyl disaster.[132] The report says it is impossible to reliably predict the number of fatal cancers arising from the incident as small differences in assumptions can result in large differences in the estimated health costs. The report says it represents the consensus view of the eight UN organisations." [NEWLINE] [NEWLINE] But that doesn't mean nothing ever happened. Decades later you can't eat venison and mushrooms in certain parts of Europe, thanks to radiation. But of course, no effects on the people living there. Same for Fukushima....</s>
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm totally with you on the whole Santa thing. We don't lie to our kids either.<mask> I actually don't want them going around telling other kids there is no Santa and<mask><mask><mask><mask> I believe that the kids who are lied to need their parents to be the ones to break the news to them. I know not all parents sit their kids down and tell them that Santa isn't real,<mask><mask><mask> that they should, I hope that they do, and<mask><mask> it is more damaging to the parent-child relationship<mask> someone else reveals to them that their parents have been lying. Part of<mask> made it less traumatic for me was the fact that my parents sat me down, told me, and explained to me<mask>. It still freaking sucked to find out,<mask> at least they were able to give me a loving explanation that helped me understand. [NEWLINE] [NEWLINE] <mask>, I<mask> don't believe in telling my kids to lie to others.<mask> I explain the whole Santa IS real, he lived a long time ago, and he's a fun character around Christmas. All those things are true, and<mask> they said them to other kids, it wouldn't ruin the lie. I<mask> tell them that people like to pretend that he comes to their house<mask><mask> my nieces and nephews say anything about Santa coming that they will think they're just pretending (which is<mask> the truth<mask> it all is pretend).<mask> they get older and can understand things better (right now my oldest is 5. He's having a hard time understanding that Santa isn't real,<mask> we have never lied to him and have made a big effort to explain that he doesn't ACTUALLY come here<mask><mask> he believes it, it's not<mask> of us.<mask><mask> he'll better understand it next year). Then<mask> he's older I will explain that parents lie to their kids and that we shouldn't be the ones to tell them the truth,<mask> to ignore it<mask> kids talk about it and don't tell them lies. I hope that makes sense. </s>
Label encoding: <s>I'm totally with you on the whole Santa thing. We don't lie to our kids either. But I actually don't want them going around telling other kids there is no Santa and the reason is because I believe that the kids who are lied to need their parents to be the ones to break the news to them. I know not all parents sit their kids down and tell them that Santa isn't real, but I think that they should, I hope that they do, and I think it is more damaging to the parent-child relationship when someone else reveals to them that their parents have been lying. Part of what made it less traumatic for me was the fact that my parents sat me down, told me, and explained to me why. It still freaking sucked to find out, but at least they were able to give me a loving explanation that helped me understand. [NEWLINE] [NEWLINE] But, I also don't believe in telling my kids to lie to others. So I explain the whole Santa IS real, he lived a long time ago, and he's a fun character around Christmas. All those things are true, and if they said them to other kids, it wouldn't ruin the lie. I also tell them that people like to pretend that he comes to their house so if my nieces and nephews say anything about Santa coming that they will think they're just pretending (which is also the truth because it all is pretend). When they get older and can understand things better (right now my oldest is 5. He's having a hard time understanding that Santa isn't real, but we have never lied to him and have made a big effort to explain that he doesn't ACTUALLY come here so if he believes it, it's not because of us. I think he'll better understand it next year). Then when he's older I will explain that parents lie to their kids and that we shouldn't be the ones to tell them the truth, but to ignore it when kids talk about it and don't tell them lies. I hope that makes sense. </s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0364, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0630, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0467, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> your morality cannot exist in a vacuum either.  you make the argument that people are obligated to give up any money that they dont need.  needs are ONLY food, water, shelter.  thats it.  those are classified<mask> basic human needs. <mask> anyone earning more money than that has to give it up in order to support healthcare. [NEWLINE] [NEWLINE] <mask> in order for heathcare to work, we need doctors.  most people think that doctors make money hand over fist, and that can be the case. <mask> doctors<mask> have really crappy lives, generally.  they work like 20 hour days every day, are on call all the time, sleep at the hospital cause the just dont have time to get home.  its not a good life. that (combined with the cost of their education and our need for them) is<mask> we<mask> a society pay them<mask> much. [NEWLINE] [NEWLINE] but,<mask> you are talking about doing will change the whole system. most people will not be willing to be a doctor (and live that terrible life)<mask> they arent even going to get money for it.  sure some people will still do it for the passion of it. <mask> most people wont.  and people will burn out a lot faster too<mask> the passion fades. now, education wouldnt really be an issue,<mask> taxes are<mask> high and the government will just pay for it. <mask> the life itself would see to it that a lot of doctors wouldnt want to do the job. [NEWLINE] [NEWLINE] suddenly, healthcare is just a giant pill vending machine and thats all it can do.  there is very little reason to put in significant extra work with zero return. and this would apply to everything.  art would disapear, videogames, movies, music would disappear<mask> no one could afford it (i can only buy water, food and shelter). [NEWLINE] [NEWLINE] you arent just talking about upping ACA funding. you are talking about completly changing the way that the entire world works.  </s>
Label encoding: <s>but your morality cannot exist in a vacuum either.  you make the argument that people are obligated to give up any money that they dont need.  needs are ONLY food, water, shelter.  thats it.  those are classified as basic human needs.  so anyone earning more money than that has to give it up in order to support healthcare. [NEWLINE] [NEWLINE] but in order for heathcare to work, we need doctors.  most people think that doctors make money hand over fist, and that can be the case.  but doctors also have really crappy lives, generally.  they work like 20 hour days every day, are on call all the time, sleep at the hospital cause the just dont have time to get home.  its not a good life. that (combined with the cost of their education and our need for them) is why we as a society pay them so much. [NEWLINE] [NEWLINE] but, what you are talking about doing will change the whole system. most people will not be willing to be a doctor (and live that terrible life) if they arent even going to get money for it.  sure some people will still do it for the passion of it.  but most people wont.  and people will burn out a lot faster too as the passion fades. now, education wouldnt really be an issue, because taxes are so high and the government will just pay for it.  but the life itself would see to it that a lot of doctors wouldnt want to do the job. [NEWLINE] [NEWLINE] suddenly, healthcare is just a giant pill vending machine and thats all it can do.  there is very little reason to put in significant extra work with zero return. and this would apply to everything.  art would disapear, videogames, movies, music would disappear because no one could afford it (i can only buy water, food and shelter). [NEWLINE] [NEWLINE] you arent just talking about upping ACA funding. you are talking about completly changing the way that the entire world works.  </s>
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> Scientology survived 1300 years then it wouldn't seem that crazy. [NEWLINE] [NEWLINE] I mean consider that historically leaving Islam was (and still is in some parts) a death sentence, isn't that different to their disconnection policy, the space opera is<mask> crazy<mask> the Buraq tale (the flying horse) or the transparent virgins in Muslim heaven. [NEWLINE] [NEWLINE] The idea of engrams messing with humanity is no more silly than the idea of the holy spirit or the Devil influencing humanity.  The idea of Jesus resurrecting is<mask> daft<mask> the idea of clear souls etc. [NEWLINE] [NEWLINE] Confession is<mask> you give your secrets ("sins") to a priest to be forgiven, add some rudimentary galvanic skin response stuff and wham you have auditing [NEWLINE] [NEWLINE] Practices like Disconnection displayed by groups like Jehovah's Witnesses is very similar to the Scientology practice of it. The Sea Org isn't a world away from Mormon Missionary work [NEWLINE] [NEWLINE] Then you have the founders, both LRon and Joesph Smith were conmen, the first pope wanted Christianity<mask> a power tool same goes for Muhammed [NEWLINE] [NEWLINE] <mask> Scientology survives for 1300 years I bet it would be seen the same<mask> mainstream religion today [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s> if Scientology survived 1300 years then it wouldn't seem that crazy. [NEWLINE] [NEWLINE] I mean consider that historically leaving Islam was (and still is in some parts) a death sentence, isn't that different to their disconnection policy, the space opera is as crazy as the Buraq tale (the flying horse) or the transparent virgins in Muslim heaven. [NEWLINE] [NEWLINE] The idea of engrams messing with humanity is no more silly than the idea of the holy spirit or the Devil influencing humanity.  The idea of Jesus resurrecting is as daft as the idea of clear souls etc. [NEWLINE] [NEWLINE] Confession is when you give your secrets ("sins") to a priest to be forgiven, add some rudimentary galvanic skin response stuff and wham you have auditing [NEWLINE] [NEWLINE] Practices like Disconnection displayed by groups like Jehovah's Witnesses is very similar to the Scientology practice of it. The Sea Org isn't a world away from Mormon Missionary work [NEWLINE] [NEWLINE] Then you have the founders, both LRon and Joesph Smith were conmen, the first pope wanted Christianity as a power tool same goes for Muhammed [NEWLINE] [NEWLINE] If Scientology survives for 1300 years I bet it would be seen the same as mainstream religion today [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] The tax<mask> falls under the larger umbrella of fat-shaming, along with placing undue financial burden on statistically more poverty-stricken individuals. This<mask> introduces a higher level of stress into the family's life. [ENDQ] [NEWLINE] [STARTQ] First off, fat-shaming campaigns have the exact opposite effect for which they are intended. They simply don't work and are a massive detriment to the general public. This is really heartbreaking<mask> well<mask> these campaigns are<mask> well-intentioned.<mask> there comes a point that<mask> you have the evidence staring you in the face, you need to accept it and look at a different solution. [ENDQ] [NEWLINE] It's not fat-shaming at all. It's a punitive measure to rescue children from obesity. No one cares that families targeted for a health intervention might feel ashamed of themselves. Do not change the topic. [NEWLINE] [NEWLINE] [STARTQ] Education and emotional support help people get healthy, not only those affected by obesity of course,<mask> recovery from obesity is dramatically more successful with these two assets. [ENDQ] [NEWLINE] Did you forget about the other part of the plan?<mask> do you think the school program is,<mask> not educational and supportive? [NEWLINE] [NEWLINE] [STARTQ] Second, the undue financial burden you're placing on people that are statistically more poverty-stricken introduces stress into the lives of these people. Not only are you shaming them, you're making it harder to afford decent food. [ENDQ] [NEWLINE] The government can provide aid to help families afford healthy meals. [NEWLINE] [NEWLINE] [STARTQ] This would be tough to handle in America and it would be absolutely devastating in a third-world country like Puerto Rico. [ENDQ] [NEWLINE] **<mask> the shit?** [NEWLINE] [NEWLINE] Puerto Rico is a US territory, [with a Human Development Index of 0.905]( [URL] #Countries_missing_from_latest_report)! For reference, the US proper has an HDI of 0.914. [NEWLINE] [NEWLINE] [STARTQ] Stress does not make it easier to lose weight<mask> you're obese [ENDQ] [NEWLINE] Healthy living helps you lose weight. The stress is merely incidental.</s>
Label encoding: <s> [STARTQ] The tax also falls under the larger umbrella of fat-shaming, along with placing undue financial burden on statistically more poverty-stricken individuals. This also introduces a higher level of stress into the family's life. [ENDQ] [NEWLINE] [STARTQ] First off, fat-shaming campaigns have the exact opposite effect for which they are intended. They simply don't work and are a massive detriment to the general public. This is really heartbreaking as well because these campaigns are so well-intentioned. But there comes a point that when you have the evidence staring you in the face, you need to accept it and look at a different solution. [ENDQ] [NEWLINE] It's not fat-shaming at all. It's a punitive measure to rescue children from obesity. No one cares that families targeted for a health intervention might feel ashamed of themselves. Do not change the topic. [NEWLINE] [NEWLINE] [STARTQ] Education and emotional support help people get healthy, not only those affected by obesity of course, but recovery from obesity is dramatically more successful with these two assets. [ENDQ] [NEWLINE] Did you forget about the other part of the plan? What do you think the school program is, if not educational and supportive? [NEWLINE] [NEWLINE] [STARTQ] Second, the undue financial burden you're placing on people that are statistically more poverty-stricken introduces stress into the lives of these people. Not only are you shaming them, you're making it harder to afford decent food. [ENDQ] [NEWLINE] The government can provide aid to help families afford healthy meals. [NEWLINE] [NEWLINE] [STARTQ] This would be tough to handle in America and it would be absolutely devastating in a third-world country like Puerto Rico. [ENDQ] [NEWLINE] ** What the shit?** [NEWLINE] [NEWLINE] Puerto Rico is a US territory, [with a Human Development Index of 0.905]( [URL] #Countries_missing_from_latest_report)! For reference, the US proper has an HDI of 0.914. [NEWLINE] [NEWLINE] [STARTQ] Stress does not make it easier to lose weight when you're obese [ENDQ] [NEWLINE] Healthy living helps you lose weight. The stress is merely incidental.</s>
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0450, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Genetic variation is important, it is the fuel for evolution that makes species robust. Human species is still lacking variation thanks to population bottlenecks and almost going extinct. [NEWLINE] [NEWLINE] The patient you are about to kill may hold mutation that would allow our species to survive in the future. [NEWLINE] [NEWLINE] Thanks to sex and genetic recombination genes are selected individually and get reshuffled in each generation. [NEWLINE] <mask> the one good gene from the patient can get combined with the one good gene from you, and the harmful genes can disappear. And nobody needs to die for the genes to disappear. Small variations in the procreation rate over generations will erase harmful genes. [NEWLINE] [NEWLINE] It does not matter<mask> many children you get 1-50, the bad genes will disappear over the future generations,<mask> they won't maintain each the same child rate, and they will each have different combinations, each of them is missing 50% of your genes, each of their children is missing 75% of your genes. [NEWLINE] [NEWLINE] Our sexual procreation erases genes quickly,<mask> only half of your genes get passed on to a child. [NEWLINE] <mask> you have 1 child, 50% of your genes are immediately gone, after 4 generations with 1 child in each only 0.5^4 = 6% of your genes are left on average. [NEWLINE] [NEWLINE] <mask> you have 2 children, on average 25% of your genes are immediately gone, and after 4 generations with 2 children for each child only 0.75^4 = 31% of your genes are left on average. [NEWLINE] [NEWLINE] <mask> even good and slightly beneficial genes will disappear. Slightly harmful genes will disappear even faster. The surviving genes need to be exceptionally good, or exceptionally lucky. [NEWLINE] [NEWLINE] And<mask> the genes aren't harmful anymore, there is no point trying to erase them anyway. [NEWLINE] The tiny issues they may cause today, probably shrink further in the future [NEWLINE] For example there are many genes now that would have been harmful in the past,<mask> are good or neutral now. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Genetic variation is important, it is the fuel for evolution that makes species robust. Human species is still lacking variation thanks to population bottlenecks and almost going extinct. [NEWLINE] [NEWLINE] The patient you are about to kill may hold mutation that would allow our species to survive in the future. [NEWLINE] [NEWLINE] Thanks to sex and genetic recombination genes are selected individually and get reshuffled in each generation. [NEWLINE] So the one good gene from the patient can get combined with the one good gene from you, and the harmful genes can disappear. And nobody needs to die for the genes to disappear. Small variations in the procreation rate over generations will erase harmful genes. [NEWLINE] [NEWLINE] It does not matter how many children you get 1-50, the bad genes will disappear over the future generations, because they won't maintain each the same child rate, and they will each have different combinations, each of them is missing 50% of your genes, each of their children is missing 75% of your genes. [NEWLINE] [NEWLINE] Our sexual procreation erases genes quickly, because only half of your genes get passed on to a child. [NEWLINE] If you have 1 child, 50% of your genes are immediately gone, after 4 generations with 1 child in each only 0.5^4 = 6% of your genes are left on average. [NEWLINE] [NEWLINE] If you have 2 children, on average 25% of your genes are immediately gone, and after 4 generations with 2 children for each child only 0.75^4 = 31% of your genes are left on average. [NEWLINE] [NEWLINE] So even good and slightly beneficial genes will disappear. Slightly harmful genes will disappear even faster. The surviving genes need to be exceptionally good, or exceptionally lucky. [NEWLINE] [NEWLINE] And if the genes aren't harmful anymore, there is no point trying to erase them anyway. [NEWLINE] The tiny issues they may cause today, probably shrink further in the future [NEWLINE] For example there are many genes now that would have been harmful in the past, but are good or neutral now. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0382, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Wittgenstein, in *Philosophical Investigations*, offered a critique of words<mask> corresponding to the traditional notion of categories. That is, we<mask> humans use language in such a way that many words do not have a particular feature that unifies all of them. [NEWLINE] [NEWLINE] [STARTQ] Look for example at board-games, with their multifarious relationships. Now pass to card-games; here you find many correspondences with the first group,<mask> many common features drop out, and others appear. <mask> we pass next to ball-games, much that is common is retained,<mask> much is lost. [ENDQ] [NEWLINE] [STARTQ] —Are they all 'amusing'? Compare chess with noughts and crosses. Or is there always winning and losing, or competition between players? Think of patience. In ball games there is winning and losing;<mask><mask> a child throws his ball at the wall and catches it again, this feature has disappeared. Look at the parts played by skill and luck; and at the difference between skill in chess and skill in tennis. [ENDQ] [NEWLINE] [STARTQ] Think now of games like ring-a-ring-a-roses; here is the element of amusement,<mask><mask> many other characteristic features have disappeared! [ENDQ] [NEWLINE] The very same treatment can be applied to sports.<mask> must a "sport" be<mask><mask> your criteria that it has to be won by objective means? Is it agreed upon by dictionary makers? By ESPN? By the Olympics committee? By the general English-speaking public? For others, a sport is simply an activity that is physically demanding and competitive. [NEWLINE] [NEWLINE] This is the nature of words. Some would say that the Dallas Cowboys cheerleaders aren't engaging in sports,<mask> teams that compete in national championships are. Perhaps it's that cheerleading has taken on more of a competitive element, that there are regulatory bodies, that it takes much more training and skill to perform. Your definition of "sport" may be related<mask> different than the definitions of others.<mask> makes you right and others wrong?</s>
Label encoding: <s>Wittgenstein, in *Philosophical Investigations*, offered a critique of words as corresponding to the traditional notion of categories. That is, we as humans use language in such a way that many words do not have a particular feature that unifies all of them. [NEWLINE] [NEWLINE] [STARTQ] Look for example at board-games, with their multifarious relationships. Now pass to card-games; here you find many correspondences with the first group, but many common features drop out, and others appear.  When we pass next to ball-games, much that is common is retained, but much is lost. [ENDQ] [NEWLINE] [STARTQ] —Are they all 'amusing'? Compare chess with noughts and crosses. Or is there always winning and losing, or competition between players? Think of patience. In ball games there is winning and losing; but when a child throws his ball at the wall and catches it again, this feature has disappeared. Look at the parts played by skill and luck; and at the difference between skill in chess and skill in tennis. [ENDQ] [NEWLINE] [STARTQ] Think now of games like ring-a-ring-a-roses; here is the element of amusement, but how many other characteristic features have disappeared! [ENDQ] [NEWLINE] The very same treatment can be applied to sports. Why must a "sport" be according to your criteria that it has to be won by objective means? Is it agreed upon by dictionary makers? By ESPN? By the Olympics committee? By the general English-speaking public? For others, a sport is simply an activity that is physically demanding and competitive. [NEWLINE] [NEWLINE] This is the nature of words. Some would say that the Dallas Cowboys cheerleaders aren't engaging in sports, but teams that compete in national championships are. Perhaps it's that cheerleading has taken on more of a competitive element, that there are regulatory bodies, that it takes much more training and skill to perform. Your definition of "sport" may be related but different than the definitions of others. What makes you right and others wrong?</s>
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Who pays for the schools and roads and police now? TAXPAYERS. The government doesn't actually pay for anything. It cannot create wealth, it can only print money, collect taxes, and borrow. [NEWLINE] [NEWLINE] <mask> most people think public schools are a good idea? let them contribute to it voluntarily.? Don't like public schools, well then you don't have to support them. Perhaps without the government, we wouldn't have<mask> many roads that aren't economically viable - IE they don't help enough business to make them viable. In a voluntary society, that road would not be built, rather than having taxpayers be forced to waste money on it. [NEWLINE] [NEWLINE] The free market absolutely has the interests of the people in mind.<mask> you want to form a successful business, you MUST give customers<mask> they want.<mask> you sell a product to a consumer and they give you money for it voluntarily, it is win-win. The government doesn't have to produce a profit, or be efficient,<mask> they don't have to make money. They just take it. The government has zero incentive to produce a cost-efficient product. [NEWLINE] [NEWLINE] It's not just a free market. It's a free market combined with respect for property rights, and it can do amazing things. [NEWLINE] [NEWLINE] <mask> the government is doing such a great job,<mask> do you we have a system of politicians that can't decide<mask> is the best interest?<mask> do we go to war for the profit of the few?<mask> do we lock up people who commit victimless "crimes?"<mask> do we let corrupt politicians be bought out by donations, and then pass legislation to limit business competition, hurting consumers? [NEWLINE] [NEWLINE] Letting people make up their own minds is the best way to form a moral society. Almost every good and service could be provided without coercion by the free market<mask> enough people want to support it. The free market cannot solve every problem,<mask> then again, neither can government, and we'd all have a lot less waste without the government.</s><pad><pad><pad>
Label encoding: <s>Who pays for the schools and roads and police now? TAXPAYERS. The government doesn't actually pay for anything. It cannot create wealth, it can only print money, collect taxes, and borrow. [NEWLINE] [NEWLINE] So most people think public schools are a good idea? let them contribute to it voluntarily.? Don't like public schools, well then you don't have to support them. Perhaps without the government, we wouldn't have as many roads that aren't economically viable - IE they don't help enough business to make them viable. In a voluntary society, that road would not be built, rather than having taxpayers be forced to waste money on it. [NEWLINE] [NEWLINE] The free market absolutely has the interests of the people in mind. If you want to form a successful business, you MUST give customers what they want. When you sell a product to a consumer and they give you money for it voluntarily, it is win-win. The government doesn't have to produce a profit, or be efficient, because they don't have to make money. They just take it. The government has zero incentive to produce a cost-efficient product. [NEWLINE] [NEWLINE] It's not just a free market. It's a free market combined with respect for property rights, and it can do amazing things. [NEWLINE] [NEWLINE] If the government is doing such a great job, why do you we have a system of politicians that can't decide what is the best interest? Why do we go to war for the profit of the few? Why do we lock up people who commit victimless "crimes?" Why do we let corrupt politicians be bought out by donations, and then pass legislation to limit business competition, hurting consumers? [NEWLINE] [NEWLINE] Letting people make up their own minds is the best way to form a moral society. Almost every good and service could be provided without coercion by the free market if enough people want to support it. The free market cannot solve every problem, but then again, neither can government, and we'd all have a lot less waste without the government.</s><pad><pad><pad>
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0397, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Thank you for your input. I can sometimes be bad at phrasing things,<mask> I suppose I am speaking specifically of some specific Atheists I know who argue constantly with people who ARE religious. I do agree that Atheists change their views and incorporate new truths,<mask> I<mask> think that,<mask> based on different "evidence," Religions change their views over time<mask> well based on<mask> they believe to be true at that point in time - look at the many different ways the Hebrew Bible, the Koran and Christian texts have been interpreted. Even Atheists who believe only in scientific proof argue over<mask> the Earth was created. [NEWLINE] [NEWLINE] Lack of evidence suggests that there is not a creator,<mask> there is lack of evidence to suggest there is not.<mask> the universe did<mask> have a beginning (<mask> supported by the big bang), by the simple logic of cause and effect, there had to be an agent – separate and apart from the effect – that caused it. We don't know that it is or is not a higher being. [NEWLINE] [NEWLINE] And yes, Religion does cause many conflicts,<mask><mask> do political, ethnic, and nationalistic beliefs.<mask> per your example of WWII, that was a result of a Nazi regime which was based on Hitler's political and ethnic ideologies - he killed Jews<mask> a race, not a religion. He believed that the Aryan race was superior to all others and argued that<mask> not for ethnic cleansing, the human species would not survive. In Mein Kampf he wrote: At this point someone or other may laugh,<mask> this planet once moved through the ether for millions of years without human beings and it can do<mask> again some day<mask> men forget that they owe their higher existence, not to the ideas of a few crazy ideologists,<mask> to the knowledge and ruthless application of Nature's stern and rigid laws." [NEWLINE] [NEWLINE] Hitler believed in the "stern and rigid laws" of nature. [NEWLINE] [NEWLINE] North Korea is an Atheist state and is committing mass genocide<mask> we speak.</s>
Label encoding: <s>Thank you for your input. I can sometimes be bad at phrasing things, so I suppose I am speaking specifically of some specific Atheists I know who argue constantly with people who ARE religious. I do agree that Atheists change their views and incorporate new truths, but I also think that, although based on different "evidence," Religions change their views over time as well based on what they believe to be true at that point in time - look at the many different ways the Hebrew Bible, the Koran and Christian texts have been interpreted. Even Atheists who believe only in scientific proof argue over how the Earth was created. [NEWLINE] [NEWLINE] Lack of evidence suggests that there is not a creator, yet there is lack of evidence to suggest there is not. If the universe did indeed have a beginning ( as supported by the big bang), by the simple logic of cause and effect, there had to be an agent – separate and apart from the effect – that caused it. We don't know that it is or is not a higher being. [NEWLINE] [NEWLINE] And yes, Religion does cause many conflicts, but so do political, ethnic, and nationalistic beliefs. As per your example of WWII, that was a result of a Nazi regime which was based on Hitler's political and ethnic ideologies - he killed Jews as a race, not a religion. He believed that the Aryan race was superior to all others and argued that if not for ethnic cleansing, the human species would not survive. In Mein Kampf he wrote: At this point someone or other may laugh, but this planet once moved through the ether for millions of years without human beings and it can do so again some day if men forget that they owe their higher existence, not to the ideas of a few crazy ideologists, but to the knowledge and ruthless application of Nature's stern and rigid laws." [NEWLINE] [NEWLINE] Hitler believed in the "stern and rigid laws" of nature. [NEWLINE] [NEWLINE] North Korea is an Atheist state and is committing mass genocide as we speak.</s>
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0002, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>"<mask> a pencil has been in loads of pencil sharpeners, it’s probably a short pencil that wears out really quickly and should be thrown away!<mask><mask> a pencil sharpener has sharpened lots of pencils, it must be a pretty good pencil sharpener." [NEWLINE] [NEWLINE] "A man who can ride any horse is a cowboy; a horse anyone can ride is a good horse." [NEWLINE] [NEWLINE] "A function that can be computed by any machine is a simple function; a machine that can compute any computable function is a Universal Turing Machine." [NEWLINE] [NEWLINE] "A man who’ll dance with anyone is fun at a party; A woman who’ll dance with anyone is fun at a party." [NEWLINE] [NEWLINE] The problem with analogies is that they encourage us to believe things without actually addressing the issue, and it's really easy to create one that supports just about any point you want to make. You say that you aren't sure<mask> you believe there is truth to the message. And<mask><mask> that that lack of understanding your belief is important. Just about everything in our culture (from pithy analogies trotted out<mask> common sense to mainstream media's depiction of "slutty" women) encourages the belief you hold. [NEWLINE] [NEWLINE] For the sake of discussion, let's agree that women and men are discrete categories that are biologically and socially different. There's still no reason to assume that sexually active women *should* be viewed<mask> sluts<mask> sexually active men *should* be viewed<mask> studs. Difference and hierarchy aren't the same thing. To say they are different is not to say<mask> one should be revered and the other reviled. [NEWLINE] [NEWLINE] TL;DR Analogies are like pugs. Some people think they're cute and others don't,<mask> we should all agree that you sure<mask> hell can't build anything structurally sound out of them.<mask>, biologically/socially different is not a justification for women being shamed and men being lauded.</s>
Label encoding: <s>" If a pencil has been in loads of pencil sharpeners, it’s probably a short pencil that wears out really quickly and should be thrown away! But if a pencil sharpener has sharpened lots of pencils, it must be a pretty good pencil sharpener." [NEWLINE] [NEWLINE] "A man who can ride any horse is a cowboy; a horse anyone can ride is a good horse." [NEWLINE] [NEWLINE] "A function that can be computed by any machine is a simple function; a machine that can compute any computable function is a Universal Turing Machine." [NEWLINE] [NEWLINE] "A man who’ll dance with anyone is fun at a party; A woman who’ll dance with anyone is fun at a party." [NEWLINE] [NEWLINE] The problem with analogies is that they encourage us to believe things without actually addressing the issue, and it's really easy to create one that supports just about any point you want to make. You say that you aren't sure why you believe there is truth to the message. And I think that that lack of understanding your belief is important. Just about everything in our culture (from pithy analogies trotted out as common sense to mainstream media's depiction of "slutty" women) encourages the belief you hold. [NEWLINE] [NEWLINE] For the sake of discussion, let's agree that women and men are discrete categories that are biologically and socially different. There's still no reason to assume that sexually active women *should* be viewed as sluts while sexually active men *should* be viewed as studs. Difference and hierarchy aren't the same thing. To say they are different is not to say why one should be revered and the other reviled. [NEWLINE] [NEWLINE] TL;DR Analogies are like pugs. Some people think they're cute and others don't, but we should all agree that you sure as hell can't build anything structurally sound out of them. Also, biologically/socially different is not a justification for women being shamed and men being lauded.</s>
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Do you have a source for that? [ENDQ] [NEWLINE] I already explained this.<mask> you think I'm wrong, the burden of proof is on you. I am not obligated to prove that something has never happened. You must prove it has. [NEWLINE] [NEWLINE] In case you still don't understand: [NEWLINE] [NEWLINE] <mask> I claim no human can read other people's minds, I don't have to show a source. You have to prove I'm wrong<mask> you disagree. [NEWLINE] [NEWLINE] [STARTQ] One could say that<mask> the actual purpose of sex is to create children (whether you or<mask><mask> with that or not) or at the very least that unprotected sex has the possibility of pregnancy, that<mask> the man is only allowed to change his mind before conception that he has already made his decisions by choosing to have sex in the first place. [ENDQ] [NEWLINE] Choosing to have sex is not deciding to have a child, in reality. And legally, it should not be the case either - for men or women. [NEWLINE] [NEWLINE] [STARTQ] <mask> : my point about abortions becoming increasingly difficult to provide in some states is to emphasize the importance of<mask> the decision to make it legal was always about bodily autonomy, and giving women the impression that men are able to cut loose from something they had a part in creating puts an incredible pressure to abort [ENDQ] [NEWLINE] The only difference between current reality and the proposed is money - and unless the father is quite rich, the amount of money is perhaps $200-$500 a month. [NEWLINE] [NEWLINE] <mask> taking away that money makes a woman decide to abort - then that is a good thing.  No one should become a parent unless they are quite prepared to be, emotionally and financially. [NEWLINE] [NEWLINE] [STARTQ] <mask> it was a matter of not being able to change your mind after conception [ENDQ] [NEWLINE] Sure. Which is<mask> financial abortion should occur pre-conception. [NEWLINE] [NEWLINE] Men should exist in a default state of non-consent to raising a child. They would have to explicitly agree to it, in order to have parental obligations. [NEWLINE] </s><pad>
Label encoding: <s> [STARTQ] Do you have a source for that? [ENDQ] [NEWLINE] I already explained this. If you think I'm wrong, the burden of proof is on you. I am not obligated to prove that something has never happened. You must prove it has. [NEWLINE] [NEWLINE] In case you still don't understand: [NEWLINE] [NEWLINE] If I claim no human can read other people's minds, I don't have to show a source. You have to prove I'm wrong if you disagree. [NEWLINE] [NEWLINE] [STARTQ] One could say that because the actual purpose of sex is to create children (whether you or I agree with that or not) or at the very least that unprotected sex has the possibility of pregnancy, that if the man is only allowed to change his mind before conception that he has already made his decisions by choosing to have sex in the first place. [ENDQ] [NEWLINE] Choosing to have sex is not deciding to have a child, in reality. And legally, it should not be the case either - for men or women. [NEWLINE] [NEWLINE] [STARTQ] Also : my point about abortions becoming increasingly difficult to provide in some states is to emphasize the importance of how the decision to make it legal was always about bodily autonomy, and giving women the impression that men are able to cut loose from something they had a part in creating puts an incredible pressure to abort [ENDQ] [NEWLINE] The only difference between current reality and the proposed is money - and unless the father is quite rich, the amount of money is perhaps $200-$500 a month. [NEWLINE] [NEWLINE] If taking away that money makes a woman decide to abort - then that is a good thing.  No one should become a parent unless they are quite prepared to be, emotionally and financially. [NEWLINE] [NEWLINE] [STARTQ] If it was a matter of not being able to change your mind after conception [ENDQ] [NEWLINE] Sure. Which is why financial abortion should occur pre-conception. [NEWLINE] [NEWLINE] Men should exist in a default state of non-consent to raising a child. They would have to explicitly agree to it, in order to have parental obligations. [NEWLINE] </s><pad>
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>[The death penalty has no deterrent effect on crime, murder or otherwise.]( [URL] ) [NEWLINE] [NEWLINE] <mask><mask>, [states that still have the death penalty see *higher* murder rates than states that have banned it.]( [URL] #stateswithvwithout) [NEWLINE] [NEWLINE] <mask>, [the death penalty is actually tremendously more expensive to the state than legal alternatives like life in prison.]( [URL] #financialfacts) For example, in one study on the California justice system: [NEWLINE] [NEWLINE] [STARTQ] The authors calculated that,<mask> the Governor commuted the sentences of those remaining on death row to life without parole, it would result in an immediate savings of $170 million per year, with a savings of $5 billion over the next 20 years. [ENDQ] [NEWLINE] <mask> there's pretty solid evidence that the death penalty doesn't really have any of the advantages that its supporters often assume it has, and the only reasonable alternative is life in prison without possibility of parole. It's not perfect, I won't<mask><mask>.<mask> it accomplishes everything the death penalty does, and it does<mask> more cheaply and with a lower risk of catastrophic errors like innocent people being murdered. [NEWLINE] [NEWLINE] <mask> I want to tackle your last point.<mask> are you not concerned about the execution of the innocent? DNA evidence is more widespread,<mask> guess<mask> -- the only reason we're aware we're killing<mask> many innocent people is<mask> of DNA evidence that sometimes doesn't become available until after we've murdered them. [NEWLINE] [NEWLINE] Think about it:<mask> someone is convicted of a crime and sentenced to death, and DNA evidence that exonerates them doesn't surface until *after* they've been executed, then<mask> does that help them? [NEWLINE] [NEWLINE] Here is a simple,<mask> difficult question:<mask> would you feel<mask> a loved one of yours was wrongfully executed by the state?<mask> you stood at their grave, awash in profound grief, would you find comfort in the trite, empty statement, "The needs of the many outweigh the needs of the few?"</s>
Label encoding: <s>[The death penalty has no deterrent effect on crime, murder or otherwise.]( [URL] ) [NEWLINE] [NEWLINE] In fact, [states that still have the death penalty see *higher* murder rates than states that have banned it.]( [URL] #stateswithvwithout) [NEWLINE] [NEWLINE] Additionally, [the death penalty is actually tremendously more expensive to the state than legal alternatives like life in prison.]( [URL] #financialfacts) For example, in one study on the California justice system: [NEWLINE] [NEWLINE] [STARTQ] The authors calculated that, if the Governor commuted the sentences of those remaining on death row to life without parole, it would result in an immediate savings of $170 million per year, with a savings of $5 billion over the next 20 years. [ENDQ] [NEWLINE] So there's pretty solid evidence that the death penalty doesn't really have any of the advantages that its supporters often assume it has, and the only reasonable alternative is life in prison without possibility of parole. It's not perfect, I won't argue that. But it accomplishes everything the death penalty does, and it does so more cheaply and with a lower risk of catastrophic errors like innocent people being murdered. [NEWLINE] [NEWLINE] But I want to tackle your last point. Why are you not concerned about the execution of the innocent? DNA evidence is more widespread, but guess what -- the only reason we're aware we're killing so many innocent people is because of DNA evidence that sometimes doesn't become available until after we've murdered them. [NEWLINE] [NEWLINE] Think about it: if someone is convicted of a crime and sentenced to death, and DNA evidence that exonerates them doesn't surface until *after* they've been executed, then how does that help them? [NEWLINE] [NEWLINE] Here is a simple, if difficult question: how would you feel if a loved one of yours was wrongfully executed by the state? As you stood at their grave, awash in profound grief, would you find comfort in the trite, empty statement, "The needs of the many outweigh the needs of the few?"</s>
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You had me until this point.<mask><mask> this is like the "chicken-and-the-egg" problem, unless you have some verifiable documentation one came first? [ENDQ] [NEWLINE] Sorry,<mask> I was typing that I realized that the phrasing implies the wrong thing. My point was that<mask><mask> which came first, religion is the one that caused the values to become commonplace. Even<mask> it was somehow subconsciously believed by humans that homosexuality is wrong, the influence of religion thought history is far more concrete in terms of influencing society. [NEWLINE] [NEWLINE] [STARTQ] That's one heck of a generalization; can you prove there really is "no doubt"? [ENDQ] [NEWLINE] No, of course not.<mask>, given both human instinct and<mask> hypocritical humans are in private, I'd say that the number is far larger than even<mask> we can predict with surveys. Guilt is really the only thing that those who openly oppose it carry with them, unlike everyone else. [NEWLINE] [NEWLINE] [STARTQ] Another generalization presented without proof. It looks to me<mask><mask> you have hit the point of overreaching and the phraseology used suggests you might have some sort of unresolved hostility and "paranoia and hysteria regarding" "[c]onservative societal norms". May I suggest rewording your argument to appear less vitriolic and to present proof? [ENDQ] [NEWLINE] I'm actually on mobile at the moment<mask> unfortunately I can't present direct evidence for my claims. Then again, neither can he,<mask> I saw it<mask> a bit of a leveled playing field. You're right, I am being too literal,<mask> really it isn't that hard to figure out. Everyone is a product of their environment (I hope I dont need to provide proof for that)<mask> there really isn't any way that EVERYONE could share the exact same values (other than innate ones based on empathy)<mask> it weren't for those values being the norm at the time of their upbringing. You can see that just from looking at the world's cultures and throughout history.</s>
Label encoding: <s> [STARTQ] You had me until this point. I think this is like the "chicken-and-the-egg" problem, unless you have some verifiable documentation one came first? [ENDQ] [NEWLINE] Sorry, as I was typing that I realized that the phrasing implies the wrong thing. My point was that regardless of which came first, religion is the one that caused the values to become commonplace. Even if it was somehow subconsciously believed by humans that homosexuality is wrong, the influence of religion thought history is far more concrete in terms of influencing society. [NEWLINE] [NEWLINE] [STARTQ] That's one heck of a generalization; can you prove there really is "no doubt"? [ENDQ] [NEWLINE] No, of course not. But, given both human instinct and how hypocritical humans are in private, I'd say that the number is far larger than even what we can predict with surveys. Guilt is really the only thing that those who openly oppose it carry with them, unlike everyone else. [NEWLINE] [NEWLINE] [STARTQ] Another generalization presented without proof. It looks to me as if you have hit the point of overreaching and the phraseology used suggests you might have some sort of unresolved hostility and "paranoia and hysteria regarding" "[c]onservative societal norms". May I suggest rewording your argument to appear less vitriolic and to present proof? [ENDQ] [NEWLINE] I'm actually on mobile at the moment so unfortunately I can't present direct evidence for my claims. Then again, neither can he, so I saw it as a bit of a leveled playing field. You're right, I am being too literal, but really it isn't that hard to figure out. Everyone is a product of their environment (I hope I dont need to provide proof for that) so there really isn't any way that EVERYONE could share the exact same values (other than innate ones based on empathy) if it weren't for those values being the norm at the time of their upbringing. You can see that just from looking at the world's cultures and throughout history.</s>
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I grew up in a Baptist Church in Georgia that was actually kicked out of the Southern Baptist Convention over this issue. [NEWLINE] [NEWLINE] The first thing I would say is Jesus never said anything on the topic. Obviously that does not mean a lot on its own,<mask> it does leave a lot more room for debate than has been traditionally considered. [NEWLINE] [NEWLINE] I appreciate your belief that God's will is definite and knowable,<mask> I would disagree that God gives much consideration to the gender of our loving relationships.<mask> I believe God cares about is dividedness within ourselves and between our brothers and sisters. Jesus clearly spoke at length (and said some pretty good things) on that question. [NEWLINE] [NEWLINE] A minister I know and love often says that Christianity tends to get in the way of people being Christian, and that Jesus likely never intended to start a new religious order.<mask><mask> the writers who ultimately crafted the New Testament were writing to a specific audience to address specific social concerns in their day. And in Greek, not English. [NEWLINE] [NEWLINE] <mask> then, the translation of these texts has been similarly employed to suit an agenda that reinforces Christianity<mask> a status quo instead of a radical subversive movement. It is a lot easier to hate on women and gays and call yourself righteous than it is to feed the poor and visit prisoners and work on your own relationship with God. [NEWLINE] [NEWLINE] Pope Francis gives me a lot of hope for Christianity,<mask> he seems to understand<mask> the most important role of the Church should be. Whether the Bible says little about homosexuality or Dogma says a lot, it all misses the point of<mask>  Jesus actually spent his short life trying to show. [NEWLINE] [NEWLINE] Jesus chose to spend his time washing the feet of the rejects of his society, prostitutes, lepers, tax collectors, etc. and even shared his last meal with those who would deny knowing him and the one he knew would betray him to torture and death. [NEWLINE] [NEWLINE] You really think he would draw the line<mask> a man was in love with another man?</s>
Label encoding: <s>I grew up in a Baptist Church in Georgia that was actually kicked out of the Southern Baptist Convention over this issue. [NEWLINE] [NEWLINE] The first thing I would say is Jesus never said anything on the topic. Obviously that does not mean a lot on its own, but it does leave a lot more room for debate than has been traditionally considered. [NEWLINE] [NEWLINE] I appreciate your belief that God's will is definite and knowable, but I would disagree that God gives much consideration to the gender of our loving relationships. What I believe God cares about is dividedness within ourselves and between our brothers and sisters. Jesus clearly spoke at length (and said some pretty good things) on that question. [NEWLINE] [NEWLINE] A minister I know and love often says that Christianity tends to get in the way of people being Christian, and that Jesus likely never intended to start a new religious order. I think the writers who ultimately crafted the New Testament were writing to a specific audience to address specific social concerns in their day. And in Greek, not English. [NEWLINE] [NEWLINE] Since then, the translation of these texts has been similarly employed to suit an agenda that reinforces Christianity as a status quo instead of a radical subversive movement. It is a lot easier to hate on women and gays and call yourself righteous than it is to feed the poor and visit prisoners and work on your own relationship with God. [NEWLINE] [NEWLINE] Pope Francis gives me a lot of hope for Christianity, since he seems to understand what the most important role of the Church should be. Whether the Bible says little about homosexuality or Dogma says a lot, it all misses the point of what  Jesus actually spent his short life trying to show. [NEWLINE] [NEWLINE] Jesus chose to spend his time washing the feet of the rejects of his society, prostitutes, lepers, tax collectors, etc. and even shared his last meal with those who would deny knowing him and the one he knew would betray him to torture and death. [NEWLINE] [NEWLINE] You really think he would draw the line where a man was in love with another man?</s>
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The HPV vaccine should have a controversy.  This is actually super interesting, and an excellent thing to dig into. (I study stats and machine learning) [NEWLINE] [NEWLINE] 1) it only protects against a couple strains (<mask> in many people it does nothing.)<mask> lets be generous, and say the two strains constitute 70% of the popular strains out there: [NEWLINE] [ [URL] ] [NEWLINE] [NEWLINE] [NEWLINE] 2) Untreated, the HPV infection *disappears* in around 90% of the affected women(!) [NEWLINE] [NEWLINE] 3)<mask><mask>,<mask> for the last 10%, most can easily be treated. In the realistic maximum, 5 women, out of 100,000, actually die of cancer(!). Thats' a death risk of 0.005% here. (actually, it's<mask> low<mask> 0.0025%,<mask> I'm ballparking a  high number for the sake of argument!) [NEWLINE] [ [URL] ] [NEWLINE] [NEWLINE] 4) it turns out that serious side affects happen in about 0.0046% of vaccinations, leading to long term hospitalization and/or death. [NEWLINE] [ [URL]?s_cid=mm6229a4_w] [NEWLINE] [NEWLINE] [NEWLINE] TL;DR: [NEWLINE] [NEWLINE] <mask> you crunch the numbers, this all means that, untreated, around 0.005% of women will have complications due to HPV. (I mean they die) [NEWLINE] [NEWLINE] <mask> EVERYBODY takes the shot, assuming the 70% safe-rate is true we get the result of 0.7*( 0.0046%) + 0.3*(0.005% + 0.0046%) = 0.0061% ( a higher percent die!!) [NEWLINE] [NEWLINE] <mask> the numbers used are correct, vaccinating women for HPV actually makes them worse off,<mask> the rate of benefit is lower than the rate of (life threatening) side effects! [NEWLINE] [NEWLINE] I LOVE politics in healthcare,<mask><mask> everything is private and rushed through the door we end off with risk. [NEWLINE] </s><pad><pad>
Label encoding: <s>The HPV vaccine should have a controversy.  This is actually super interesting, and an excellent thing to dig into. (I study stats and machine learning) [NEWLINE] [NEWLINE] 1) it only protects against a couple strains ( so in many people it does nothing.) So lets be generous, and say the two strains constitute 70% of the popular strains out there: [NEWLINE] [ [URL] ] [NEWLINE] [NEWLINE] [NEWLINE] 2) Untreated, the HPV infection *disappears* in around 90% of the affected women(!) [NEWLINE] [NEWLINE] 3) lastly, as for the last 10%, most can easily be treated. In the realistic maximum, 5 women, out of 100,000, actually die of cancer(!). Thats' a death risk of 0.005% here. (actually, it's as low as 0.0025%, but I'm ballparking a  high number for the sake of argument!) [NEWLINE] [ [URL] ] [NEWLINE] [NEWLINE] 4) it turns out that serious side affects happen in about 0.0046% of vaccinations, leading to long term hospitalization and/or death. [NEWLINE] [ [URL]?s_cid=mm6229a4_w] [NEWLINE] [NEWLINE] [NEWLINE] TL;DR: [NEWLINE] [NEWLINE] If you crunch the numbers, this all means that, untreated, around 0.005% of women will have complications due to HPV. (I mean they die) [NEWLINE] [NEWLINE] If EVERYBODY takes the shot, assuming the 70% safe-rate is true we get the result of 0.7*( 0.0046%) + 0.3*(0.005% + 0.0046%) = 0.0061% ( a higher percent die!!) [NEWLINE] [NEWLINE] If the numbers used are correct, vaccinating women for HPV actually makes them worse off, because the rate of benefit is lower than the rate of (life threatening) side effects! [NEWLINE] [NEWLINE] I LOVE politics in healthcare, because if everything is private and rushed through the door we end off with risk. [NEWLINE] </s><pad><pad>
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0370, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I believe his for a few reasons: [NEWLINE] [NEWLINE] 1) Putin will not care.  He is dedicated to the idea of a greater Russia and deeply distrusts the West already.  By imposing sanctions on him, we are reinforcing his view of an evil west. [NEWLINE] [NEWLINE] 2) America has no prior obligations to Ukraine.  They are not in MATO and we have no other treaties with them.  Intervening into the affairs of a random nation is both imperialist and exactly<mask> we are deriding Russia for. [NEWLINE] [NEWLINE] 3) The Russian people should not be punished for the actions of their government.  By harming the economy of Russia we hurting the populace, who,<mask> in an ostensibly democratic state, would have a hard time voting Putin out of office anyway. [NEWLINE] [NEWLINE] 4) Imposing sanctions strains our relationship with Russia, which is a strong geopolitical power and would be troublesome to have<mask> an enemy. [NEWLINE] [NEWLINE] 5) Crimea *voted* to join Russia.  By intervening, we are going against the very "will of the people" we value<mask> highly, and it<mask> makes us look ridiculous in the eyes of the international community. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>I believe his for a few reasons: [NEWLINE] [NEWLINE] 1) Putin will not care.  He is dedicated to the idea of a greater Russia and deeply distrusts the West already.  By imposing sanctions on him, we are reinforcing his view of an evil west. [NEWLINE] [NEWLINE] 2) America has no prior obligations to Ukraine.  They are not in MATO and we have no other treaties with them.  Intervening into the affairs of a random nation is both imperialist and exactly what we are deriding Russia for. [NEWLINE] [NEWLINE] 3) The Russian people should not be punished for the actions of their government.  By harming the economy of Russia we hurting the populace, who, while in an ostensibly democratic state, would have a hard time voting Putin out of office anyway. [NEWLINE] [NEWLINE] 4) Imposing sanctions strains our relationship with Russia, which is a strong geopolitical power and would be troublesome to have as an enemy. [NEWLINE] [NEWLINE] 5) Crimea *voted* to join Russia.  By intervening, we are going against the very "will of the people" we value so highly, and it also makes us look ridiculous in the eyes of the international community. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Interestingly enough: [NEWLINE] [NEWLINE] [STARTQ] World War III: This seems the least likely,<mask> the repercussions of North Korea's potential idiocy from a couple moths ago proved that it might not be that unlikely<mask> one country makes one wrong move. [ENDQ] [NEWLINE] This seems the most likely to me.  I guess that is probably<mask> the other examples you listed are too unlikely for me. [NEWLINE] [NEWLINE] [STARTQ] Climate change: The possibility of famines and other food supply-disrupting weather, shrinking landmasses = competition for resources, scarcity, destabilization of society. This is practically a certainty, the only question is<mask> fast will it happen and<mask> well will we handle it. [ENDQ] [NEWLINE] It won't be during your lifetime.  Even<mask> you are a genius five year old.  Even the most aggressive projection puts any significant life threatening situations a 100 years out. [NEWLINE] [NEWLINE] [STARTQ] Telecom/electrical grid failure or sabotage: Our society would fall into chaos without these services, and<mask> any were interrupted long enough, bad things would happen. Just really think about<mask> would happen<mask> the electrical grid throughout the US and Canada was just turned off for a couple of months. [ENDQ] [NEWLINE] Which is<mask> the government spend a lot of resources to making sure that doesn't happen.  We have been relying on telecom/electrical grid for the last 100 years or<mask>.  Never had we had the situation you are talking about, not even for a day. <mask> makes you think we would have that for months. [NEWLINE] [NEWLINE] [STARTQ] I'm not a paranoid person, I don't lose sleep over this, I've just accepted it<mask> a fact of life for maybe the last 5 years. Do CMV,<mask><mask> I do think about it it's depressing<mask> hell. I'm just saying, society is very very fragile and we're currently barreling towards a number of potential disasters. [ENDQ] [NEWLINE] This might just be your young brain justifying not being responsible for your own future.   </s>
Label encoding: <s>Interestingly enough: [NEWLINE] [NEWLINE] [STARTQ] World War III: This seems the least likely, but the repercussions of North Korea's potential idiocy from a couple moths ago proved that it might not be that unlikely if one country makes one wrong move. [ENDQ] [NEWLINE] This seems the most likely to me.  I guess that is probably because the other examples you listed are too unlikely for me. [NEWLINE] [NEWLINE] [STARTQ] Climate change: The possibility of famines and other food supply-disrupting weather, shrinking landmasses = competition for resources, scarcity, destabilization of society. This is practically a certainty, the only question is how fast will it happen and how well will we handle it. [ENDQ] [NEWLINE] It won't be during your lifetime.  Even if you are a genius five year old.  Even the most aggressive projection puts any significant life threatening situations a 100 years out. [NEWLINE] [NEWLINE] [STARTQ] Telecom/electrical grid failure or sabotage: Our society would fall into chaos without these services, and if any were interrupted long enough, bad things would happen. Just really think about what would happen if the electrical grid throughout the US and Canada was just turned off for a couple of months. [ENDQ] [NEWLINE] Which is why the government spend a lot of resources to making sure that doesn't happen.  We have been relying on telecom/electrical grid for the last 100 years or so.  Never had we had the situation you are talking about, not even for a day.  What makes you think we would have that for months. [NEWLINE] [NEWLINE] [STARTQ] I'm not a paranoid person, I don't lose sleep over this, I've just accepted it as a fact of life for maybe the last 5 years. Do CMV, because when I do think about it it's depressing as hell. I'm just saying, society is very very fragile and we're currently barreling towards a number of potential disasters. [ENDQ] [NEWLINE] This might just be your young brain justifying not being responsible for your own future.   </s>
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0475, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Before you label me<mask> a feminist or even a sycophant, know that I am a young, white, mildly religious male who often has difficulty in relating to female issues. I'm just looking for answers. [NEWLINE] [NEWLINE] <mask>,<mask> a male, I do not feel qualified to form an opinion, let alone conscientiously vote, on anything regarding abortion, and I don't feel any male should have any say in<mask> happens to a woman's body. There might be a simple/logical reason we (males) are allowed to - that's<mask> I'm here. [NEWLINE] [NEWLINE] I've considered some counterpoints,<mask> can't think of anything that's not an appeal to emotion, which I admittedly have difficulty considering. I've never been a father,<mask> I still don't see that having any weight in the matter. [NEWLINE] [NEWLINE] Edit: Alright, you all have<mask> changed my view. Here's<mask> you did it: [NEWLINE] [NEWLINE] [STARTQ] I find arguments which dismiss the opinions of males solely<mask> they come from males,<mask> well<mask> those arguments that simply deal with female autonomy alone, to be highly unsatisfactory and completely ignoring the concerns of the other side. [ENDQ] [NEWLINE] And<mask> : [NEWLINE] [NEWLINE] [STARTQ] There's disparate impact across a number of issues and voters are overwhelmingly ignorant<mask> it comes to just about anything we can vote on.<mask>, usually, voters don't directly vote on issues. Yes, there are referendums,<mask> we typically vote for people to represent us in legislatures. This means that men could conceivably be voting for women representatives who would cast the direct vote on some abortion law, which should be a digression<mask> I guess I'll bring it up to illustrate<mask> much more complex voter choice is. [ENDQ] [NEWLINE] [STARTQ] The point is that<mask> removal from a circumstance or knowledge/empathy of an issue is the proxy for being able to vote, very few people should be allowed to cast a ballot, and that's fundamentally undemocratic. [ENDQ] [NEWLINE] </s>
Label encoding: <s>Before you label me as a feminist or even a sycophant, know that I am a young, white, mildly religious male who often has difficulty in relating to female issues. I'm just looking for answers. [NEWLINE] [NEWLINE] So, as a male, I do not feel qualified to form an opinion, let alone conscientiously vote, on anything regarding abortion, and I don't feel any male should have any say in what happens to a woman's body. There might be a simple/logical reason we (males) are allowed to - that's why I'm here. [NEWLINE] [NEWLINE] I've considered some counterpoints, but can't think of anything that's not an appeal to emotion, which I admittedly have difficulty considering. I've never been a father, but I still don't see that having any weight in the matter. [NEWLINE] [NEWLINE] Edit: Alright, you all have indeed changed my view. Here's how you did it: [NEWLINE] [NEWLINE] [STARTQ] I find arguments which dismiss the opinions of males solely because they come from males, as well as those arguments that simply deal with female autonomy alone, to be highly unsatisfactory and completely ignoring the concerns of the other side. [ENDQ] [NEWLINE] And also : [NEWLINE] [NEWLINE] [STARTQ] There's disparate impact across a number of issues and voters are overwhelmingly ignorant when it comes to just about anything we can vote on. Although, usually, voters don't directly vote on issues. Yes, there are referendums, but we typically vote for people to represent us in legislatures. This means that men could conceivably be voting for women representatives who would cast the direct vote on some abortion law, which should be a digression but I guess I'll bring it up to illustrate how much more complex voter choice is. [ENDQ] [NEWLINE] [STARTQ] The point is that if removal from a circumstance or knowledge/empathy of an issue is the proxy for being able to vote, very few people should be allowed to cast a ballot, and that's fundamentally undemocratic. [ENDQ] [NEWLINE] </s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0572, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Abortion is the procedure of terminating a pregnancy, not an already-born child.<mask>, abortion cannot be an option after birth. I'd appreciate<mask> we just focus on pre-birth situations. [ENDQ] [NEWLINE] The important point here is that to a pro-life person, killing an unborn fetus is equivalent to killing a baby. [NEWLINE] [NEWLINE] *It does not matter that you disagree, no matter<mask> factually right you believe that disagreement is.* [NEWLINE] [NEWLINE] <mask> you are going to<mask><mask> someone is a hypocrite, you have to demonstrate that their beliefs and actions are inconsistent. You cannot demonstrate that<mask> you are going to disregard their beliefs. [NEWLINE] [NEWLINE] <mask> you think that anti-abortion advocates should sign up for to be a foster parent, then following the pro-life belief system it should follow that anyone against the killing of a child should sign up to be foster parent<mask> the two scenarios are morally equivalent. [NEWLINE] [NEWLINE] I assume you believe that people against child killing do not have a duty to sign up to be a foster parent and that you're okay with that.<mask> that is the case, whatever your argument you make for<mask> you don't have a duty to yourself up for foster parenting can<mask> be used to argue<mask> pro-lifers don't have a duty to sign up for foster care. [NEWLINE] [NEWLINE] [STARTQ] To reiterate from my first edit: a fetus is not alive until 7 months into the pregnancy. You cant kill something that was never alive to begin with. No, this is not my belief/opinion, it is fact. [ENDQ] [NEWLINE] You're making two errors here: You're disregarding the pro-life viewpoint. There's nothing hypocritical about being factually wrong about something.<mask>, the biological definition of life does not necessarily hold moral weight. I say this<mask> a pro-choice person. Whatever definition science has chosen for<mask> it means for something to be alive is an artificial construct meant to be useful to scientists, not to those judging morality. </s>
Label encoding: <s> [STARTQ] Abortion is the procedure of terminating a pregnancy, not an already-born child. Therefore, abortion cannot be an option after birth. I'd appreciate if we just focus on pre-birth situations. [ENDQ] [NEWLINE] The important point here is that to a pro-life person, killing an unborn fetus is equivalent to killing a baby. [NEWLINE] [NEWLINE] *It does not matter that you disagree, no matter how factually right you believe that disagreement is.* [NEWLINE] [NEWLINE] If you are going to argue that someone is a hypocrite, you have to demonstrate that their beliefs and actions are inconsistent. You cannot demonstrate that if you are going to disregard their beliefs. [NEWLINE] [NEWLINE] If you think that anti-abortion advocates should sign up for to be a foster parent, then following the pro-life belief system it should follow that anyone against the killing of a child should sign up to be foster parent because the two scenarios are morally equivalent. [NEWLINE] [NEWLINE] I assume you believe that people against child killing do not have a duty to sign up to be a foster parent and that you're okay with that. If that is the case, whatever your argument you make for why you don't have a duty to yourself up for foster parenting can also be used to argue why pro-lifers don't have a duty to sign up for foster care. [NEWLINE] [NEWLINE] [STARTQ] To reiterate from my first edit: a fetus is not alive until 7 months into the pregnancy. You cant kill something that was never alive to begin with. No, this is not my belief/opinion, it is fact. [ENDQ] [NEWLINE] You're making two errors here: You're disregarding the pro-life viewpoint. There's nothing hypocritical about being factually wrong about something. Secondly, the biological definition of life does not necessarily hold moral weight. I say this as a pro-choice person. Whatever definition science has chosen for what it means for something to be alive is an artificial construct meant to be useful to scientists, not to those judging morality. </s>
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0406, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0492, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Hey I thought I'd chime in<mask> I'm nearly a doctor (6 more months). All doctors are trained to screen their patients for depression by default by looking for certain physical symptoms and to pay attention to a patients demeanor during an encounter. [NEWLINE] [NEWLINE] <mask> from a purely medical standpoint [NEWLINE] 1. Depression is defined by objective criteria in the Diagnostics Service Manual V.<mask> doctors diagnosis depression they use these criteria. Other wise they would not be able to bill and convince insurance to cover any treatment. Depending on<mask> criteria is met<mask> influences treatment. [NEWLINE] [NEWLINE] 2. It has been well proven that medications aimed at increasing serotonin levels lead to an alleviation of the symptoms of depression. There is no literature that suggests looking at levels of these chemicals to screen or make a diagnosis, the diagnosis is made by criteria in the DSM V mentioned above. [NEWLINE] [NEWLINE] 3. Agreed, no treatment for any disease should be forces upon a paient. Doctors are huge advocates of this, that is until the disease may lead to harm of themselves or others, ie. Suicidal or homicidal ideations, or more tradiional things such<mask> tuberculosis [NEWLINE] [NEWLINE] 4. My only comment here is that depression is treated with relatively cheap drugs in conjunction with non pharmaceutical treatments such<mask> cognitive behavioral therapy. [NEWLINE] [NEWLINE] 5. Sounds like an interesting read [NEWLINE] [NEWLINE] <mask> for the original CMV like I stated previously most every doctor screens for depression in any patient interaction, many physical symptoms of depression can mimic other diseases such<mask> hypothyroidism. We are<mask> all trained in medical school quick screens we can do<mask> we suspect a depressed patient ( google "Sig E Caps") and just to reiterate it is the part of the physician to help treat disease,<mask> to never force treatment on others unless it may lead to danger to that patient or others (I'd write more and probably make more sense<mask> I'm on my phone about to take off on a plane<mask> forgive any errors thaks)</s>
Label encoding: <s>Hey I thought I'd chime in as I'm nearly a doctor (6 more months). All doctors are trained to screen their patients for depression by default by looking for certain physical symptoms and to pay attention to a patients demeanor during an encounter. [NEWLINE] [NEWLINE] Also from a purely medical standpoint [NEWLINE] 1. Depression is defined by objective criteria in the Diagnostics Service Manual V. When doctors diagnosis depression they use these criteria. Other wise they would not be able to bill and convince insurance to cover any treatment. Depending on what criteria is met also influences treatment. [NEWLINE] [NEWLINE] 2. It has been well proven that medications aimed at increasing serotonin levels lead to an alleviation of the symptoms of depression. There is no literature that suggests looking at levels of these chemicals to screen or make a diagnosis, the diagnosis is made by criteria in the DSM V mentioned above. [NEWLINE] [NEWLINE] 3. Agreed, no treatment for any disease should be forces upon a paient. Doctors are huge advocates of this, that is until the disease may lead to harm of themselves or others, ie. Suicidal or homicidal ideations, or more tradiional things such as tuberculosis [NEWLINE] [NEWLINE] 4. My only comment here is that depression is treated with relatively cheap drugs in conjunction with non pharmaceutical treatments such as cognitive behavioral therapy. [NEWLINE] [NEWLINE] 5. Sounds like an interesting read [NEWLINE] [NEWLINE] As for the original CMV like I stated previously most every doctor screens for depression in any patient interaction, many physical symptoms of depression can mimic other diseases such as hypothyroidism. We are also all trained in medical school quick screens we can do if we suspect a depressed patient ( google "Sig E Caps") and just to reiterate it is the part of the physician to help treat disease, but to never force treatment on others unless it may lead to danger to that patient or others (I'd write more and probably make more sense but I'm on my phone about to take off on a plane so forgive any errors thaks)</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0545, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Finally,<mask><mask> I understand your point! [NEWLINE] [NEWLINE] Okay,<mask><mask>, there is no worldwide feminist organization like the UN<mask> all the feminists in the world get together and talk about world issues and man-hating or size 0's or whatever. It just doesn't exist. [NEWLINE] [NEWLINE] <mask>, the term "rebranding" assumes a corporate structure (trust me, I've done a few rebrandings in my life). You're assuming the grassroots feminism movement is an organized, well-funded, worldwide corporate machine that can make decisions, unless I'm reading your CMV wrong. You're using terms like "charter" and "rebranding" and "global announcement", "big feminist meeting". These things just don't exist. You're assuming a level of organization that just isn't there. [NEWLINE] [NEWLINE] I, personally like the idea of a unified, "Equality Party" or something that could fight for LGBT, women's, men's, minority's rights,<mask><mask><mask> it's fairly naive to think that those groups will be able to all work together. In this case, it's the men's and women's rights groups- especially<mask> the men's rights groups are<mask> reactionary against feminism (particularly the extreme neo-feminism that you seem to dislike, and for good reason). [NEWLINE] [NEWLINE] I don't think this is possible<mask> several of the MRA's are *violently* anti-feminist. They are<mask> reactionary against feminism that they've gone the opposite way and<mask> have many extreme neo-feminists. I just don't know<mask> it's possible to bring the two sides together at all. They all have their own issues and you've said it yourself that most extreme feminists are polarizing. Most feminists can deal with the extremists. MRA's can't. MRA's have extremists,<mask> well. Feminists won't be able to deal with them. Equalism wouldn't happen<mask> of that. </s><pad><pad>
Label encoding: <s>Finally, I think I understand your point! [NEWLINE] [NEWLINE] Okay, firstly, there is no worldwide feminist organization like the UN where all the feminists in the world get together and talk about world issues and man-hating or size 0's or whatever. It just doesn't exist. [NEWLINE] [NEWLINE] Secondly, the term "rebranding" assumes a corporate structure (trust me, I've done a few rebrandings in my life). You're assuming the grassroots feminism movement is an organized, well-funded, worldwide corporate machine that can make decisions, unless I'm reading your CMV wrong. You're using terms like "charter" and "rebranding" and "global announcement", "big feminist meeting". These things just don't exist. You're assuming a level of organization that just isn't there. [NEWLINE] [NEWLINE] I, personally like the idea of a unified, "Equality Party" or something that could fight for LGBT, women's, men's, minority's rights, but I think it's fairly naive to think that those groups will be able to all work together. In this case, it's the men's and women's rights groups- especially since the men's rights groups are so reactionary against feminism (particularly the extreme neo-feminism that you seem to dislike, and for good reason). [NEWLINE] [NEWLINE] I don't think this is possible because several of the MRA's are *violently* anti-feminist. They are so reactionary against feminism that they've gone the opposite way and so have many extreme neo-feminists. I just don't know if it's possible to bring the two sides together at all. They all have their own issues and you've said it yourself that most extreme feminists are polarizing. Most feminists can deal with the extremists. MRA's can't. MRA's have extremists, as well. Feminists won't be able to deal with them. Equalism wouldn't happen because of that. </s><pad><pad>
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>One of the major policy appeals behind affirmative action is the creation of a diverse environment. A completely homogeneous environment is unlikely to be<mask> intellectually fulfilling<mask> a diverse one. Universities have done an excellent job in creating an environment that has tons of racial, sexual, ethnic and religious diversity.<mask>, it still lacks intellectual and political diversity. People with conservative or right-wing views are [woefully unrepresented]( [URL].pdf) in academia, particularly in the social sciences. Not only does this make people with conservative political views feel unwelcome and unappreciated in the university environment, it threatens the perceived validity of the research done. [This is even recognized by some leftists themselves]( [URL] /).<mask> are you supposed to have valid research, let alone critical thinking,<mask> there is an absolute hegemony of leftist views, which no one dares challenge due to their overwhelmingly disproportionate power and influence on campus?<mask>, to make university a more diverse environment and encourage critical thinking, we should prioritize student and faculty applicants who are affiliated with right wing organizations, and ensure that peer review panels have at least one right wing professor on them CMV. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>One of the major policy appeals behind affirmative action is the creation of a diverse environment. A completely homogeneous environment is unlikely to be as intellectually fulfilling as a diverse one. Universities have done an excellent job in creating an environment that has tons of racial, sexual, ethnic and religious diversity. However, it still lacks intellectual and political diversity. People with conservative or right-wing views are [woefully unrepresented]( [URL].pdf) in academia, particularly in the social sciences. Not only does this make people with conservative political views feel unwelcome and unappreciated in the university environment, it threatens the perceived validity of the research done. [This is even recognized by some leftists themselves]( [URL] /). How are you supposed to have valid research, let alone critical thinking, when there is an absolute hegemony of leftist views, which no one dares challenge due to their overwhelmingly disproportionate power and influence on campus? Therefore, to make university a more diverse environment and encourage critical thinking, we should prioritize student and faculty applicants who are affiliated with right wing organizations, and ensure that peer review panels have at least one right wing professor on them CMV. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0027, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There's nothing inherently wrong with putting a penis in a relative's vagina, any more than there's anything inherently wrong with pouring poison into water. The consequences flow out from the actions often enough<mask> that the government can make it illegal. [ENDQ] [NEWLINE] This is the answer I was looking for.<mask> you acknowledge that the act of the pouring isn't the wrong, then<mask> do you justify an opposition to non-reproductive incest? I doubt the poisoning of water reference is an accident,<mask> I assume you mean that the crime,<mask> to speak, is in the damage to the gene pool and the future others. After all, the crime wouldn't really lie in pouring poison into water,<mask> causing water that is used by others to be poisonous,<mask> you catch my meaning,<mask> the analogy breaks down<mask> there's nobody to poison.<mask> you poor toxic chemicals into a bottle of water and you put that bottle in a safe and cover that safe in concrete, that's something different from poisoning a river. Absent the intent to, fact of, or possibility of reproduction,<mask> does this argument go? [NEWLINE] [NEWLINE] It's extremely easy to imagine cases of incest that will not or cannot cause any of the problems you want to prevent in reproduction, rape, molestation, and romantic/family breakdowns.<mask> the government is making illegal that which can't or won't produce the consequences they seek to prevent by enacting a law, then that law is flawed, no? None of those bad results are inseparable from incest or flow naturally from it, and to severely understate the case: not everything with the mere theoretical potential for harm, or with which some kind of harm is generally associated (those harms being the result of acts which are distinct from the crime, and themselves illegal), is against the law.<mask> it were, incest would just be one thing among countless others, most of them commonplace, unremarkable, and unqualifiedly acceptable to nearly all people.</s><pad>
Label encoding: <s> [STARTQ] There's nothing inherently wrong with putting a penis in a relative's vagina, any more than there's anything inherently wrong with pouring poison into water. The consequences flow out from the actions often enough though that the government can make it illegal. [ENDQ] [NEWLINE] This is the answer I was looking for. If you acknowledge that the act of the pouring isn't the wrong, then how do you justify an opposition to non-reproductive incest? I doubt the poisoning of water reference is an accident, so I assume you mean that the crime, so to speak, is in the damage to the gene pool and the future others. After all, the crime wouldn't really lie in pouring poison into water, but causing water that is used by others to be poisonous, if you catch my meaning, so the analogy breaks down when there's nobody to poison. If you poor toxic chemicals into a bottle of water and you put that bottle in a safe and cover that safe in concrete, that's something different from poisoning a river. Absent the intent to, fact of, or possibility of reproduction, where does this argument go? [NEWLINE] [NEWLINE] It's extremely easy to imagine cases of incest that will not or cannot cause any of the problems you want to prevent in reproduction, rape, molestation, and romantic/family breakdowns. If the government is making illegal that which can't or won't produce the consequences they seek to prevent by enacting a law, then that law is flawed, no? None of those bad results are inseparable from incest or flow naturally from it, and to severely understate the case: not everything with the mere theoretical potential for harm, or with which some kind of harm is generally associated (those harms being the result of acts which are distinct from the crime, and themselves illegal), is against the law. If it were, incest would just be one thing among countless others, most of them commonplace, unremarkable, and unqualifiedly acceptable to nearly all people.</s><pad>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] No. Seeking fun and its monetary value are completely different. I like to swim and I like to play soccer with my friends,<mask> I don't believe that it is worth anything to the society. [ENDQ] [NEWLINE] Worth to society and monetary value are<mask> are different. Worth to society is debatable; the issue is<mask> people are willing to pay. You can establish a Brave New World style totalitarian government<mask> everyone is only allowed a soccer ball and a hackeysack, and that's fun, or you can give people the freedom to seek out the pleasures and muses they desire. Naturally, industries and corporations are going to rise up and exploit people and generate massive revenue<mask><mask><mask>. That is the inevitability of a capitalist society -<mask><mask><mask> people are willing to pay, people will profit highly. The only way to achieve the goal you are proposing is to create a totalitarian government that strictly and forcefully controls "fun." You said earlier you are not calling for entertainment to be banned,<mask> in a way you are,<mask> you are saying that people shouldn't participate in entertainment in the way they are currently. [NEWLINE] [NEWLINE] Who are you to say<mask> people do with their money? People like to have fun. Some people's idea of fun is to watch sports. That's all there is to it. [NEWLINE] [NEWLINE] Truth be told, the reason<mask> sports generates<mask> much revenue is mostly advertising. You lament the fact that doctors do not profit<mask> much<mask> sports players,<mask><mask>, medicine is, like sports entertainment,<mask> a giant industry founded on invasive advertising. Some people actually think advertising is *detrimental* to health care. Some people even want health care to be *free.* (shocker). Seems like health care would improve greatly<mask> LESS money were involved. Perhaps money is not the marker of cultural and societal value - perhaps it is quite the contrary. I doubt you want hospitals run like sports organizations anymore than they already are.</s>
Label encoding: <s> [STARTQ] No. Seeking fun and its monetary value are completely different. I like to swim and I like to play soccer with my friends, but I don't believe that it is worth anything to the society. [ENDQ] [NEWLINE] Worth to society and monetary value are what are different. Worth to society is debatable; the issue is what people are willing to pay. You can establish a Brave New World style totalitarian government where everyone is only allowed a soccer ball and a hackeysack, and that's fun, or you can give people the freedom to seek out the pleasures and muses they desire. Naturally, industries and corporations are going to rise up and exploit people and generate massive revenue as a result. That is the inevitability of a capitalist society - as long as people are willing to pay, people will profit highly. The only way to achieve the goal you are proposing is to create a totalitarian government that strictly and forcefully controls "fun." You said earlier you are not calling for entertainment to be banned, but in a way you are, because you are saying that people shouldn't participate in entertainment in the way they are currently. [NEWLINE] [NEWLINE] Who are you to say what people do with their money? People like to have fun. Some people's idea of fun is to watch sports. That's all there is to it. [NEWLINE] [NEWLINE] Truth be told, the reason why sports generates so much revenue is mostly advertising. You lament the fact that doctors do not profit as much as sports players, even though, medicine is, like sports entertainment, also a giant industry founded on invasive advertising. Some people actually think advertising is *detrimental* to health care. Some people even want health care to be *free.* (shocker). Seems like health care would improve greatly if LESS money were involved. Perhaps money is not the marker of cultural and societal value - perhaps it is quite the contrary. I doubt you want hospitals run like sports organizations anymore than they already are.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0340, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I use Google voice and couldn't send or receive MMS messages for a couple years. (<mask><mask> they finally offered MMS through select carriers) [ENDQ] [NEWLINE] You used a service that could only exist due to government involvement in the market. Okay, not seeing the relevance. [NEWLINE] [NEWLINE] [STARTQ] For your example:<mask> I were an Americaphone user, at some point, the costs (overcharging, dropped calls, bad customer service) would outweigh the benefits (calling friends) [ENDQ] [NEWLINE] Only<mask> they went beyond abusing their monopoly and into abusing their customers. At which point you might drop off the network and do without phone service at all. [NEWLINE] [NEWLINE] <mask> would Americaphone go that far<mask>? They're not idiots, they want high prices that you're still willing to pay. They want prices<mask> you'd go to competition<mask> there was any...<mask> there isn't. [NEWLINE] [NEWLINE] Monopoly pricing is much higher than competitive pricing,<mask> it's not infinite. [NEWLINE] [NEWLINE] [STARTQ] <mask>, competition is an on-going, multifaceted process. Your example:<mask>, Boston gets a local company very similar to myphone, Baltimore gets one too, and New York does, too. [ENDQ] [NEWLINE] That requires those companies to get off the ground. Which doesn't make much sense. [NEWLINE] [NEWLINE] <mask> would they do that?<mask> would they persuade me to go "Oh, heh, I'm going to get a phoneline that can't be used to call any of my friends or family; nor to make (or receive) business calls with any companies I have accounts with. Oh, and I don't need to be able to use the internet through it." [NEWLINE] [NEWLINE] You're going to need to explain<mask> you manage to get Myphone into the first 100,000 people's houses,<mask> it's more expensive to supply than Americaphone (they have the network set up already) and doesn't actually let you phone most people.</s>
Label encoding: <s> [STARTQ] I use Google voice and couldn't send or receive MMS messages for a couple years. ( I think they finally offered MMS through select carriers) [ENDQ] [NEWLINE] You used a service that could only exist due to government involvement in the market. Okay, not seeing the relevance. [NEWLINE] [NEWLINE] [STARTQ] For your example: If I were an Americaphone user, at some point, the costs (overcharging, dropped calls, bad customer service) would outweigh the benefits (calling friends) [ENDQ] [NEWLINE] Only if they went beyond abusing their monopoly and into abusing their customers. At which point you might drop off the network and do without phone service at all. [NEWLINE] [NEWLINE] Why would Americaphone go that far though? They're not idiots, they want high prices that you're still willing to pay. They want prices where you'd go to competition if there was any... but there isn't. [NEWLINE] [NEWLINE] Monopoly pricing is much higher than competitive pricing, but it's not infinite. [NEWLINE] [NEWLINE] [STARTQ] Also, competition is an on-going, multifaceted process. Your example: So, Boston gets a local company very similar to myphone, Baltimore gets one too, and New York does, too. [ENDQ] [NEWLINE] That requires those companies to get off the ground. Which doesn't make much sense. [NEWLINE] [NEWLINE] How would they do that? How would they persuade me to go "Oh, heh, I'm going to get a phoneline that can't be used to call any of my friends or family; nor to make (or receive) business calls with any companies I have accounts with. Oh, and I don't need to be able to use the internet through it." [NEWLINE] [NEWLINE] You're going to need to explain how you manage to get Myphone into the first 100,000 people's houses, when it's more expensive to supply than Americaphone (they have the network set up already) and doesn't actually let you phone most people.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think we need to break feminism down into two categories: mainstream and radical.  Other commenters have addressed the mainstream rather well,<mask> I'll address the radicals. [NEWLINE] [NEWLINE] Radical feminists are out their.  They are vocal and disproportionately influential.  They are sexist, often homophobic and transphibic.  They will actively fight against, suppress and oppress anyone advocating for men's issues.  They idealize a society<mask> women are dominant over men or even<mask> men do not exist.  They often advocate the use of violence to achieve their goals.  Even the more mellow radicals will support blatantly sexist legislation. [NEWLINE] [NEWLINE] Now<mask><mask> we can both agree that these people are hateful and seek to harm men.  The problem is that these hateful people are femenists just<mask> much<mask> the intelligent ones you seam to run with.  This means that a part of the feminist movement seeks to actively harm men.  Is it unreasonable or 'illegitimate' for another movement to form to stop such a hateful and vicious group? [NEWLINE] [NEWLINE] Please do not construe my statements<mask> "all femenists are bad".  Merely a subset is<mask> that subset must be vehemently opposed. [NEWLINE] [NEWLINE] <mask> : [NEWLINE] [NEWLINE] [STARTQ] The men's rights movement (which I truly hope only exists on the internet) is unnecessary. Men have held positions of power for millennia - is it really necessary for us to feel better about being a man? [ENDQ] [NEWLINE] This really bothers me.  It gets to a major problem with both feminism and the men's rights movement.  They lump people together by gender<mask> opposed to acknowledging them<mask> individuals.  Should the poor homeless man be okay with being refused charity<mask> he is a man just<mask> other men are in positions of power?  I say no.  No more<mask> than a woman should be okay with earning a lower wage for her work<mask> other women are getting women only scholarships. [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I think we need to break feminism down into two categories: mainstream and radical.  Other commenters have addressed the mainstream rather well, so I'll address the radicals. [NEWLINE] [NEWLINE] Radical feminists are out their.  They are vocal and disproportionately influential.  They are sexist, often homophobic and transphibic.  They will actively fight against, suppress and oppress anyone advocating for men's issues.  They idealize a society where women are dominant over men or even where men do not exist.  They often advocate the use of violence to achieve their goals.  Even the more mellow radicals will support blatantly sexist legislation. [NEWLINE] [NEWLINE] Now I think we can both agree that these people are hateful and seek to harm men.  The problem is that these hateful people are femenists just as much as the intelligent ones you seam to run with.  This means that a part of the feminist movement seeks to actively harm men.  Is it unreasonable or 'illegitimate' for another movement to form to stop such a hateful and vicious group? [NEWLINE] [NEWLINE] Please do not construe my statements as "all femenists are bad".  Merely a subset is but that subset must be vehemently opposed. [NEWLINE] [NEWLINE] Also : [NEWLINE] [NEWLINE] [STARTQ] The men's rights movement (which I truly hope only exists on the internet) is unnecessary. Men have held positions of power for millennia - is it really necessary for us to feel better about being a man? [ENDQ] [NEWLINE] This really bothers me.  It gets to a major problem with both feminism and the men's rights movement.  They lump people together by gender as opposed to acknowledging them as individuals.  Should the poor homeless man be okay with being refused charity because he is a man just because other men are in positions of power?  I say no.  No more so than a woman should be okay with earning a lower wage for her work because other women are getting women only scholarships. [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0521, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>As I said in another comment, I don't think it's prosecutable, nor do I necessarily think it should be.<mask>, by no means does that make it ok and yes, it is rape. Using someone's mental state against them is a form of force. Coercion, with clear intent is rape. You are convincing someone to do something they don't want to do by dehumanizing them. [NEWLINE] [NEWLINE] <mask> I appreciate your view it borders a bit on victim blaming. I'm not talking about a person consenting and then coming back later<mask> 'they didn't enjoy the sex'. Mental duress is still duress. Telling a person that they are somehow less of a person<mask> they won't have sex with you, telling them that they owe you sex for some reason, saying that you are too far gone and that they have to finish or they are a bad person, IS duress. It's emotional manipulation and it is rape. [NEWLINE] [NEWLINE] There are a lot of scenarios, like the one I discussed in an earlier comment in which there may have been no malicious intent and those scenarios are a very grey area in which effective communication can resolve the issue. Outside of those,<mask> a person is pressuring someone persistently, and using malicious means to get<mask> they want it certainly is rape and in this case we may have to agree to disagree. [NEWLINE] [NEWLINE] We have to stop nitpicking these scenarios to death in lieu of doing anything. By going back and forth in the technicality of<mask> rape is we are losing the ultimate goal which is to make sure everyone knows<mask> their rights are,<mask> to enforce them and<mask> to both protect themselves and their partners. [NEWLINE] [NEWLINE] And yes, coercion of any kind is absolutely rape. We don't coerce people we care about. We want both parties to be having a good time. The second someone is not is the second we have to step back and evaluate. </s>
Label encoding: <s>As I said in another comment, I don't think it's prosecutable, nor do I necessarily think it should be. However, by no means does that make it ok and yes, it is rape. Using someone's mental state against them is a form of force. Coercion, with clear intent is rape. You are convincing someone to do something they don't want to do by dehumanizing them. [NEWLINE] [NEWLINE] While I appreciate your view it borders a bit on victim blaming. I'm not talking about a person consenting and then coming back later because 'they didn't enjoy the sex'. Mental duress is still duress. Telling a person that they are somehow less of a person because they won't have sex with you, telling them that they owe you sex for some reason, saying that you are too far gone and that they have to finish or they are a bad person, IS duress. It's emotional manipulation and it is rape. [NEWLINE] [NEWLINE] There are a lot of scenarios, like the one I discussed in an earlier comment in which there may have been no malicious intent and those scenarios are a very grey area in which effective communication can resolve the issue. Outside of those, if a person is pressuring someone persistently, and using malicious means to get what they want it certainly is rape and in this case we may have to agree to disagree. [NEWLINE] [NEWLINE] We have to stop nitpicking these scenarios to death in lieu of doing anything. By going back and forth in the technicality of what rape is we are losing the ultimate goal which is to make sure everyone knows what their rights are, how to enforce them and how to both protect themselves and their partners. [NEWLINE] [NEWLINE] And yes, coercion of any kind is absolutely rape. We don't coerce people we care about. We want both parties to be having a good time. The second someone is not is the second we have to step back and evaluate. </s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Source for edits to the other books (especially the Quran<mask> iirc muslims think the book is unchanged in its arabic form)? [ENDQ] [NEWLINE] [Book of Mormon]( [URL] ) [NEWLINE] [NEWLINE] [Quran]( [URL] %27a_manuscript) [NEWLINE] [NEWLINE] [STARTQ] The Bible was changed...<mask> you partially admitted that yourself. [ENDQ] [NEWLINE] Sorry, I hadn't meant to infer that there were no *copies* of the Bible with errors in them.  My intent was to show that modern bible translations are derived in a different way than other Holy books. [NEWLINE] The Book of Mormon and the Quran were originally conceived<mask> a singular document which was later edited. <mask> these edits were produced iteratively,<mask> we are left with today is the most recent version, which differes dramatically from the original. [NEWLINE] [NEWLINE] Modern Bible Translations, by contrast, are produced by going back<mask> closely<mask> possible to the original source material. Of course there have been many, many variant copies and fragments of the Bible over the years,<mask> great efforts are taken to both [determine which variants were original, and to be open about the process]( [URL] ). [NEWLINE] [NEWLINE] [STARTQ] You ignore the possibility of exaggerations and lies in a timeframe of at least 50 years. [ENDQ] [NEWLINE] The possibility is there for sure,<mask> multiple books by different authors seem to complement each other's accounts quite well, and I don't see any historical reason<mask> the authors would have benefited from lying. [NEWLINE] [NEWLINE] [STARTQ] Scripture not being altered wouldn't make it true. [ENDQ] [NEWLINE] <mask><mask> with this. [NEWLINE] [NEWLINE] [STARTQ] <mask> has King James maybe being gay to do with anything? [ENDQ] [NEWLINE] It's a 'fun fact'.<mask> the most used Bible version in history bears his name, and given Christianity's stance on homosexuality, it's an interesting TILish bit of information, no? [NEWLINE] [NEWLINE] I like your username, by the way :-) [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s> [STARTQ] Source for edits to the other books (especially the Quran since iirc muslims think the book is unchanged in its arabic form)? [ENDQ] [NEWLINE] [Book of Mormon]( [URL] ) [NEWLINE] [NEWLINE] [Quran]( [URL] %27a_manuscript) [NEWLINE] [NEWLINE] [STARTQ] The Bible was changed... Also you partially admitted that yourself. [ENDQ] [NEWLINE] Sorry, I hadn't meant to infer that there were no *copies* of the Bible with errors in them.  My intent was to show that modern bible translations are derived in a different way than other Holy books. [NEWLINE] The Book of Mormon and the Quran were originally conceived as a singular document which was later edited.  Since these edits were produced iteratively, what we are left with today is the most recent version, which differes dramatically from the original. [NEWLINE] [NEWLINE] Modern Bible Translations, by contrast, are produced by going back as closely as possible to the original source material. Of course there have been many, many variant copies and fragments of the Bible over the years, but great efforts are taken to both [determine which variants were original, and to be open about the process]( [URL] ). [NEWLINE] [NEWLINE] [STARTQ] You ignore the possibility of exaggerations and lies in a timeframe of at least 50 years. [ENDQ] [NEWLINE] The possibility is there for sure, but multiple books by different authors seem to complement each other's accounts quite well, and I don't see any historical reason why the authors would have benefited from lying. [NEWLINE] [NEWLINE] [STARTQ] Scripture not being altered wouldn't make it true. [ENDQ] [NEWLINE] I agree with this. [NEWLINE] [NEWLINE] [STARTQ] What has King James maybe being gay to do with anything? [ENDQ] [NEWLINE] It's a 'fun fact'. Because the most used Bible version in history bears his name, and given Christianity's stance on homosexuality, it's an interesting TILish bit of information, no? [NEWLINE] [NEWLINE] I like your username, by the way :-) [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s><mask> there is one God who was the author of all creation (including right &amp; wrong) &amp; who decides to save or condemn, who is a human to judge the criteria by which that decision is made? Sure it may seem unfair to me, it may seem unfair to you,<mask><mask> should God be bound by our sense of fairness?<mask> should God save any of us at all? &amp;<mask> is it any more arbitrary to pick the saved based on their attitude to him than it is to pick them by<mask> they treated their own, or<mask> colour hair they had or whether they had children or not, or whether they ate meat? [NEWLINE] [NEWLINE] If such a God exists then the criteria for salvation he picks are the criteria, there'd be no point in deciding they were unfair &amp; condemning others by choosing not to give them the opportunity to meet them. You'd hardly call it fair to put your own moral vanity ahead of other people's souls; I don't think it's fair that ships sometimes sink,<mask> that doesn't mean I'd be justified in refusing to hand out life-jackets<mask> I was present at a ship-wreck &amp; happened to have some. In the kind of world-view we're talking about, damnation &amp; salvation are just simply facts, exactly the same<mask> any other danger. [NEWLINE] [NEWLINE] Now, I may be reading too much into<mask> your saying in this next paragraph,<mask> ignore this<mask> I am,<mask> I'm getting the impression that you're saying that the realisation that the god one believes in is unfair should be reason for doubt. That would be a complete nonsequitur<mask> whether or not something is fair does not have any impact on whether or not it's real,<mask> to leave one's religion<mask> one thinks damnation is wrong would be<mask> pointless<mask> giving up a healthy lifestyle<mask> one believes sickness is wrong.</s>
Label encoding: <s>If there is one God who was the author of all creation (including right &amp; wrong) &amp; who decides to save or condemn, who is a human to judge the criteria by which that decision is made? Sure it may seem unfair to me, it may seem unfair to you, but why should God be bound by our sense of fairness? Why should God save any of us at all? &amp; how is it any more arbitrary to pick the saved based on their attitude to him than it is to pick them by how they treated their own, or what colour hair they had or whether they had children or not, or whether they ate meat? [NEWLINE] [NEWLINE] If such a God exists then the criteria for salvation he picks are the criteria, there'd be no point in deciding they were unfair &amp; condemning others by choosing not to give them the opportunity to meet them. You'd hardly call it fair to put your own moral vanity ahead of other people's souls; I don't think it's fair that ships sometimes sink, but that doesn't mean I'd be justified in refusing to hand out life-jackets if I was present at a ship-wreck &amp; happened to have some. In the kind of world-view we're talking about, damnation &amp; salvation are just simply facts, exactly the same as any other danger. [NEWLINE] [NEWLINE] Now, I may be reading too much into what your saying in this next paragraph, so ignore this if I am, but I'm getting the impression that you're saying that the realisation that the god one believes in is unfair should be reason for doubt. That would be a complete nonsequitur since whether or not something is fair does not have any impact on whether or not it's real, so to leave one's religion because one thinks damnation is wrong would be as pointless as giving up a healthy lifestyle because one believes sickness is wrong.</s>
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0436, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0461, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0480, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] There is no such thing<mask> a "patriarchy" in the western world. [ENDQ] [NEWLINE] <mask> you described about men being unable to open up about their abuse at the hands of women is an example of the patriarchy. To quote Allan G. Johnson's definition of a patriarchy "Aspects of society and personal attributes that are highly valued are associated with men,<mask> devalued attributes and social activities are associated with women." Does that sound familiar? [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> pointing out double standards such<mask> "<mask> can women hit men with impunity<mask> men hitting women **in self defense** are seen<mask> abusers" is taking an active step to shuck the traditional gender norms of the past. [ENDQ] [NEWLINE] You would be mistaken. An active step would be pushing for better representations of men in the media, creating a culture in which male vulnerability isn't dismissed or laughed at<mask> something only the weak do, and denouncing the use of the term"man up" whenever it is brought up<mask><mask> the context. Laughing at, cheering on, or otherwise encouraging violence towards women<mask><mask> a perceived justification is a step in the wrong direction. [NEWLINE] [NEWLINE] The reason society cheers<mask> a woman stands up for herself is<mask> in western society female empowerment is still a relatively new concept. 1920 is<mask> women were granted the right to vote in the US, that is less than 100 years ago. One woman standing up for herself serves<mask> an example for all women. For men<mask>, standing up for ourselves is a given.<mask> you stated,<mask> men we are expected to defend ourselves and<mask> we don't we are ridiculed and seen<mask> lesser. A man who stands up for himself teaches me nothing that I have not already learned from a patriarchal society,<mask> a woman standing up for herself does<mask><mask> teach girls that it okay to stand up and defend themselves. [NEWLINE] [NEWLINE] "Boy babies get blue socks; girl babies get pink socks."</s>
Label encoding: <s> [STARTQ] There is no such thing as a "patriarchy" in the western world. [ENDQ] [NEWLINE] What you described about men being unable to open up about their abuse at the hands of women is an example of the patriarchy. To quote Allan G. Johnson's definition of a patriarchy "Aspects of society and personal attributes that are highly valued are associated with men, while devalued attributes and social activities are associated with women." Does that sound familiar? [NEWLINE] [NEWLINE] [STARTQ] I think pointing out double standards such as " why can women hit men with impunity but men hitting women **in self defense** are seen as abusers" is taking an active step to shuck the traditional gender norms of the past. [ENDQ] [NEWLINE] You would be mistaken. An active step would be pushing for better representations of men in the media, creating a culture in which male vulnerability isn't dismissed or laughed at as something only the weak do, and denouncing the use of the term"man up" whenever it is brought up regardless of the context. Laughing at, cheering on, or otherwise encouraging violence towards women regardless of a perceived justification is a step in the wrong direction. [NEWLINE] [NEWLINE] The reason society cheers when a woman stands up for herself is because in western society female empowerment is still a relatively new concept. 1920 is when women were granted the right to vote in the US, that is less than 100 years ago. One woman standing up for herself serves as an example for all women. For men however, standing up for ourselves is a given. As you stated, as men we are expected to defend ourselves and when we don't we are ridiculed and seen as lesser. A man who stands up for himself teaches me nothing that I have not already learned from a patriarchal society, yet a woman standing up for herself does in fact teach girls that it okay to stand up and defend themselves. [NEWLINE] [NEWLINE] "Boy babies get blue socks; girl babies get pink socks."</s>
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>But your response doesn't adequately address the issue of 'interest or knowledge.' [NEWLINE] [NEWLINE] Compelling action does not mean that action will have quality.<mask> anything, it becomes an unpleasant chore to be discharged with the least effort possible. You don't reduce the cost of becoming informed<mask> you compel a ballot to be cast; you only increase the cost of not voting. [NEWLINE] [NEWLINE] ~~Your lunch example is flawed; you say that a person must pay a cost (vote/pay),<mask> paying slightly more (by becoming informed) is easier. Except that you're really proposing that a person must pay a cost (vote/pay) and that not satisfying a required action results in a much higher cost(pay); that gives them the incentive to do the least necessary to satisfy the requirement and<mask> take the lower cost (voting). They have no incentive to take additional action to make that vote an informed one.~~ [NEWLINE] [NEWLINE] I posit that compelling a ballot to be cast will not provide an incentive to learn anything beforehand,<mask> instead reduces the cost to the ignorant, who would otherwise abstain, from voting based on whatever foolish notions they might have. I posit they would be far more likely to make poor electoral decisions based on those foolish notions (<mask> that would be perceived<mask> an 'accomplishment') than they would to vote 'undecided' or 'none of the above,'<mask> traveling somewhere to void their own vote places a perceived cost on the act of voting. [NEWLINE] [NEWLINE] Edit:<mask> I reread this, my take down of your lunch example makes no fracking sense.<mask> I'm trying to (inarticulately) say is, you've reduced the relative cost of being an informed voter,<mask> all you've required is being a voter; there is no reason for anyone to take that additional cost to become properly informed and in all likelyhood, they will act on ignorance and believe it wisdom.</s>
Label encoding: <s>But your response doesn't adequately address the issue of 'interest or knowledge.' [NEWLINE] [NEWLINE] Compelling action does not mean that action will have quality. If anything, it becomes an unpleasant chore to be discharged with the least effort possible. You don't reduce the cost of becoming informed when you compel a ballot to be cast; you only increase the cost of not voting. [NEWLINE] [NEWLINE] ~~Your lunch example is flawed; you say that a person must pay a cost (vote/pay), so paying slightly more (by becoming informed) is easier. Except that you're really proposing that a person must pay a cost (vote/pay) and that not satisfying a required action results in a much higher cost(pay); that gives them the incentive to do the least necessary to satisfy the requirement and thus take the lower cost (voting). They have no incentive to take additional action to make that vote an informed one.~~ [NEWLINE] [NEWLINE] I posit that compelling a ballot to be cast will not provide an incentive to learn anything beforehand, but instead reduces the cost to the ignorant, who would otherwise abstain, from voting based on whatever foolish notions they might have. I posit they would be far more likely to make poor electoral decisions based on those foolish notions ( because that would be perceived as an 'accomplishment') than they would to vote 'undecided' or 'none of the above,' as traveling somewhere to void their own vote places a perceived cost on the act of voting. [NEWLINE] [NEWLINE] Edit: as I reread this, my take down of your lunch example makes no fracking sense. What I'm trying to (inarticulately) say is, you've reduced the relative cost of being an informed voter, but all you've required is being a voter; there is no reason for anyone to take that additional cost to become properly informed and in all likelyhood, they will act on ignorance and believe it wisdom.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0342, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0008, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0387, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0343, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0447, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I don't think there's much anyone can say to change your view on the sexualizeation of women,<mask> I'm going to focus on whether or not it detracts from the quality of Reddit. [NEWLINE] [NEWLINE] The reason Reddit is good is<mask> it's a wholly user controlled experience. The quality of reddit varies from user to user. Every subreddit is made up of users who vote on the material<mask> that the best rises to the top and gains visibility.<mask> a user, you have a tremendous role in<mask> you see.<mask><mask> to upvoting content you enjoy and downvoting content you don't within a subreddit, you are<mask> in complete control of which subreddits you let affect your front page.<mask> you're unhappy with the posts that are being upvoted in a subreddit, then that probably means you're subscribed to a subreddit that, collectively, is interested in different content than you. The Reddit experience is not universal, and<mask> you want a quality experience, then you have to unsubscribe from those subreddits. [NEWLINE] [NEWLINE] Many of the subreddits you cite<mask> being overly sexualized are massive subreddits with millions of subscribers and very lax rules on submission standards. There are numerous smaller subreddits that will provide a much more focused experience that is more likely to fit your interests.<mask>, with a tool like multireddits, you can combine related subreddits to make it even easier to control<mask> content you want to see. [NEWLINE] [NEWLINE] It isn't possible for content that doesn't break any rules to detract from the quality of reddit. By it's very nature,<mask> it's on the front page<mask> of its merit, then it is a high quality post. [NEWLINE] [NEWLINE] tl;dr:<mask> you're seeing sexualized images of women, it's<mask> you're subscribed to a subreddit that views that sort of thing<mask> quality.<mask> you don't like it, then unsubscribe. </s>
Label encoding: <s>I don't think there's much anyone can say to change your view on the sexualizeation of women, so I'm going to focus on whether or not it detracts from the quality of Reddit. [NEWLINE] [NEWLINE] The reason Reddit is good is because it's a wholly user controlled experience. The quality of reddit varies from user to user. Every subreddit is made up of users who vote on the material so that the best rises to the top and gains visibility. As a user, you have a tremendous role in what you see. In addition to upvoting content you enjoy and downvoting content you don't within a subreddit, you are also in complete control of which subreddits you let affect your front page. If you're unhappy with the posts that are being upvoted in a subreddit, then that probably means you're subscribed to a subreddit that, collectively, is interested in different content than you. The Reddit experience is not universal, and if you want a quality experience, then you have to unsubscribe from those subreddits. [NEWLINE] [NEWLINE] Many of the subreddits you cite as being overly sexualized are massive subreddits with millions of subscribers and very lax rules on submission standards. There are numerous smaller subreddits that will provide a much more focused experience that is more likely to fit your interests. Additionally, with a tool like multireddits, you can combine related subreddits to make it even easier to control what content you want to see. [NEWLINE] [NEWLINE] It isn't possible for content that doesn't break any rules to detract from the quality of reddit. By it's very nature, if it's on the front page because of its merit, then it is a high quality post. [NEWLINE] [NEWLINE] tl;dr: if you're seeing sexualized images of women, it's because you're subscribed to a subreddit that views that sort of thing as quality. If you don't like it, then unsubscribe. </s>
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's hard to argue with facts<mask><mask> it's well known that the supply of diamonds is deliberately restricted, and effective marketing has ensured there is no re-sale market. [NEWLINE] [NEWLINE] Having said that: there is a lot to be said for human expression. Even the expression of love through material items. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> that people are brainwashed into thinking that this is a symbol of love. [ENDQ] [NEWLINE] They're not brainwashed and you're not some amazing free thinker. It's a social convention like many others. It is a symbol of love<mask> some people said<mask> and other people agreed. People need symbols of love and fidelity. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> the money could be much better spent on a a great holiday, or to help start your married life together. [ENDQ] [NEWLINE] Sure. The holiday is great fun and of course you don't want to start your married life in debt.<mask> your partner is going to wear that ring every day for perhaps 60 or<mask> years?<mask> even<mask> you don't want to buy a diamond ring you best buy something that lasts. Diamond and gold don't blemish or rust. [NEWLINE] [NEWLINE] [STARTQ] The main reason I don't want to buy an engagement ring is money. They are ridiculusly overpriced. I'm not willing to pay a couple of grand for a shiny rock. [ENDQ] [NEWLINE] Depends<mask> much you earn.<mask> sure. My diamond ring was about £300 and I wouldn't trade it for millions. Would I like it more<mask> it cost £3,000 - of course not. It's a symbol. Nothing more. Nothing less. It might strike an starving person<mask> ridiculously extravagant; it might strike a millionairess<mask> beyond under-stated. [NEWLINE] [NEWLINE] <mask> it's mine. Somebody who loves me and will love me for the rest of my life placed it on my finger. It means the world...</s>
Label encoding: <s>It's hard to argue with facts given that it's well known that the supply of diamonds is deliberately restricted, and effective marketing has ensured there is no re-sale market. [NEWLINE] [NEWLINE] Having said that: there is a lot to be said for human expression. Even the expression of love through material items. [NEWLINE] [NEWLINE] [STARTQ] I think that people are brainwashed into thinking that this is a symbol of love. [ENDQ] [NEWLINE] They're not brainwashed and you're not some amazing free thinker. It's a social convention like many others. It is a symbol of love because some people said so and other people agreed. People need symbols of love and fidelity. [NEWLINE] [NEWLINE] [STARTQ] I think the money could be much better spent on a a great holiday, or to help start your married life together. [ENDQ] [NEWLINE] Sure. The holiday is great fun and of course you don't want to start your married life in debt. But your partner is going to wear that ring every day for perhaps 60 or so years? So even if you don't want to buy a diamond ring you best buy something that lasts. Diamond and gold don't blemish or rust. [NEWLINE] [NEWLINE] [STARTQ] The main reason I don't want to buy an engagement ring is money. They are ridiculusly overpriced. I'm not willing to pay a couple of grand for a shiny rock. [ENDQ] [NEWLINE] Depends how much you earn. But sure. My diamond ring was about £300 and I wouldn't trade it for millions. Would I like it more if it cost £3,000 - of course not. It's a symbol. Nothing more. Nothing less. It might strike an starving person as ridiculously extravagant; it might strike a millionairess as beyond under-stated. [NEWLINE] [NEWLINE] But it's mine. Somebody who loves me and will love me for the rest of my life placed it on my finger. It means the world...</s>
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>A few examples<mask> it becomes "some of your goddamn business" [NEWLINE] [NEWLINE] * Public Transportation. Overweight people take up more space on public transit. *Especially* planes. I'm a thin guy, it's incredibly frustrating<mask> another person is overflowing into my seat, that I paid for. I have to take a bus too and from my parking spot at my university and the same thing happens with large people, they overflow into my seat and touch me unnecessarily. [NEWLINE] [NEWLINE] * Health care: Overweight people use health care more often,<mask> more and more Americans push for universal health care this will become an even bigger issue. It's my business<mask> we're all paying for them to continue living an unhealthy life-style. I work out, I eat well, I overall take good care of myself. I'm far less likely to use a universal health-care system than an overweight individual. Same thing with smokers, I don't smoke<mask> it becomes my business<mask> others do<mask> the public is subsidizing their lifestyle. [NEWLINE] [NEWLINE] *<mask> people get fatter and become a larger portion of the population things become worse for thin people. Toilets are getting larger, hell there's a restaurant near me that has those new larger toilets<mask> I almost fell into the toilet with the seat down. Car seats are getting larger for fat people making them less comfortable and supportive for thin people. [NEWLINE] [NEWLINE] * Fat people raise fat kids: We<mask> a society have decided that the welfare of children, too a degree, is a public concern. This is evident with the existence of Child Protective Services.<mask>, it is some of our goddamn business<mask> there is an increase in overweight children. We're seeing an increase in childhood diabetes<mask> of their overweight parents raising overweight children, this ties into my Health Care point. [NEWLINE] [NEWLINE] There are plenty of reasons that a persons weight could be others goddamn business. </s>
Label encoding: <s>A few examples where it becomes "some of your goddamn business" [NEWLINE] [NEWLINE] * Public Transportation. Overweight people take up more space on public transit. *Especially* planes. I'm a thin guy, it's incredibly frustrating when another person is overflowing into my seat, that I paid for. I have to take a bus too and from my parking spot at my university and the same thing happens with large people, they overflow into my seat and touch me unnecessarily. [NEWLINE] [NEWLINE] * Health care: Overweight people use health care more often, as more and more Americans push for universal health care this will become an even bigger issue. It's my business when we're all paying for them to continue living an unhealthy life-style. I work out, I eat well, I overall take good care of myself. I'm far less likely to use a universal health-care system than an overweight individual. Same thing with smokers, I don't smoke but it becomes my business if others do if the public is subsidizing their lifestyle. [NEWLINE] [NEWLINE] * As people get fatter and become a larger portion of the population things become worse for thin people. Toilets are getting larger, hell there's a restaurant near me that has those new larger toilets where I almost fell into the toilet with the seat down. Car seats are getting larger for fat people making them less comfortable and supportive for thin people. [NEWLINE] [NEWLINE] * Fat people raise fat kids: We as a society have decided that the welfare of children, too a degree, is a public concern. This is evident with the existence of Child Protective Services. Therefore, it is some of our goddamn business if there is an increase in overweight children. We're seeing an increase in childhood diabetes because of their overweight parents raising overweight children, this ties into my Health Care point. [NEWLINE] [NEWLINE] There are plenty of reasons that a persons weight could be others goddamn business. </s>
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0411, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think you misunderstand the concept of science. The term'scientific authority' is an oxymoron. Science is not about authority, it's about experimentation and deduction. It's about empirical evidence, open review, critical analysis, and reproducible results. [NEWLINE] [NEWLINE] Of course scientific expertise is valuable. Athletic ability is valuable too, it might help an athlete run a race more quickly and more adeptly than anyone else.<mask> I,<mask> I am a bit overweight and out of shape, can still make it down the track and cross the finish line, even<mask> I'm the last to get there, and panting and wheezing<mask> I arrive.<mask> I had no arms or legs, I might have to have help to make it, I might need someone to wheel me to the finish line,<mask> I could still potentially get there. Sure, I could just let the professional athlete do all the finish-line crossing and just watch from the bleachers.<mask> maybe I want to see<mask> it looks like from the vantage point of the finish line myself, and not just take his word for it. [NEWLINE] [NEWLINE] Scientific expertise gives a person an advantage in their field, not a monopoly on it. They might be the best of the best,<mask> they still make errors, and they still have limits. And more than that, they have the potential to be subjective, to be biased, and in some circumstances to be corruptible. Real science is about asking questions and not being satisfied with the answers you are given until it has been proven to you. Taking an expert's opinion in lieu of evidence, and letting them think for you instead of thinking for yourself, is all too reminiscent of the dogmatic system that scientific method was created to displace. [NEWLINE] [NEWLINE] Authority is *not* science. [NEWLINE] [NEWLINE] Oh, and nano-thermite [is a thing]( [URL] ).</s>
Label encoding: <s>I think you misunderstand the concept of science. The term'scientific authority' is an oxymoron. Science is not about authority, it's about experimentation and deduction. It's about empirical evidence, open review, critical analysis, and reproducible results. [NEWLINE] [NEWLINE] Of course scientific expertise is valuable. Athletic ability is valuable too, it might help an athlete run a race more quickly and more adeptly than anyone else. But I, though I am a bit overweight and out of shape, can still make it down the track and cross the finish line, even if I'm the last to get there, and panting and wheezing as I arrive. If I had no arms or legs, I might have to have help to make it, I might need someone to wheel me to the finish line, but I could still potentially get there. Sure, I could just let the professional athlete do all the finish-line crossing and just watch from the bleachers. But maybe I want to see what it looks like from the vantage point of the finish line myself, and not just take his word for it. [NEWLINE] [NEWLINE] Scientific expertise gives a person an advantage in their field, not a monopoly on it. They might be the best of the best, but they still make errors, and they still have limits. And more than that, they have the potential to be subjective, to be biased, and in some circumstances to be corruptible. Real science is about asking questions and not being satisfied with the answers you are given until it has been proven to you. Taking an expert's opinion in lieu of evidence, and letting them think for you instead of thinking for yourself, is all too reminiscent of the dogmatic system that scientific method was created to displace. [NEWLINE] [NEWLINE] Authority is *not* science. [NEWLINE] [NEWLINE] Oh, and nano-thermite [is a thing]( [URL] ).</s>
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0319, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Even<mask> they were cool with it, all that means is they changed the definition of the word nigger to mean something other than a racist insult. [ENDQ] [NEWLINE] Ta da! You do realize that's<mask> colloquial, playful slang connotation means right? Maybe now we're finally on the same page here.<mask> like I said, this is not an important aspect of the argument. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> nowhere did I claim it's possible to do blackface without being racist. [ENDQ] [NEWLINE] Then I misunderstood the point of the CMV. [NEWLINE] [NEWLINE] [STARTQ] All I'm demanding here is self consistency. [ENDQ] [NEWLINE] Then<mask> aren't you being consistent? You said nigger couldn't be not racist<mask> of it's definition<mask> the same doesn't appear to apply to black face? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> it's possible for blacks to put on white makeup in such a way, then<mask> isn't it possible for the reverse to be true? One would have to hold a double-standard. [ENDQ] [NEWLINE] [NEWLINE] It is not a double standard<mask> it is not the same. There is no history of black people using white face<mask> part of an act of racial oppression. It simply isn't tasteful. It's the same reason that we mention a black man's race<mask> he does something great.<mask> a black man land's on the moon we will likely point that out.<mask> there is a history of his race being oppressed it makes sense to mention that civil rights have overcome the oppression and allowed him to<mask> accomplish this feat. [NEWLINE] [NEWLINE] <mask>, we did not say Neil Armstrong was the first white man on the moon.<mask>? Is that a double standard? No.<mask> there was no history of racial oppression that made it worth noting. [NEWLINE] [NEWLINE] Same with white and black face. There is no racial oppression associated with white face. It is not a double standard. </s>
Label encoding: <s> [STARTQ] Even if they were cool with it, all that means is they changed the definition of the word nigger to mean something other than a racist insult. [ENDQ] [NEWLINE] Ta da! You do realize that's what colloquial, playful slang connotation means right? Maybe now we're finally on the same page here. But like I said, this is not an important aspect of the argument. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] Because nowhere did I claim it's possible to do blackface without being racist. [ENDQ] [NEWLINE] Then I misunderstood the point of the CMV. [NEWLINE] [NEWLINE] [STARTQ] All I'm demanding here is self consistency. [ENDQ] [NEWLINE] Then why aren't you being consistent? You said nigger couldn't be not racist because of it's definition but the same doesn't appear to apply to black face? [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] If it's possible for blacks to put on white makeup in such a way, then why isn't it possible for the reverse to be true? One would have to hold a double-standard. [ENDQ] [NEWLINE] [NEWLINE] It is not a double standard because it is not the same. There is no history of black people using white face as part of an act of racial oppression. It simply isn't tasteful. It's the same reason that we mention a black man's race when he does something great. When a black man land's on the moon we will likely point that out. Because there is a history of his race being oppressed it makes sense to mention that civil rights have overcome the oppression and allowed him to also accomplish this feat. [NEWLINE] [NEWLINE] However, we did not say Neil Armstrong was the first white man on the moon. Why? Is that a double standard? No. Because there was no history of racial oppression that made it worth noting. [NEWLINE] [NEWLINE] Same with white and black face. There is no racial oppression associated with white face. It is not a double standard. </s>
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0306, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0386, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0325, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0416, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0357, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0604, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Gifted student programs aren't to give students more help<mask> to cover more material, quicker and provide a deeper understanding of the topic. A classic example would be in a Calculus curriculum. Almost every student (engineers included)does not need to know<mask> limits really work and<mask> the theorems work the way they do. Doing the epsilon-delta (something that will be in a accelerated course<mask> not in a regular one) definition would be a waste of time, not only<mask> it does not add any practical value<mask><mask> those not good in math would not be able to understand it. For that reason it was only covered at the highest level in my high school. Some people could spend every hour of every day attempting to understanding it, and still not be able to comprehend it. Gifted programs are by nature more difficult (for the vast majority of students)<mask> of the material (some exceptions are students who do not work in a non-accelerated program and do poorly simply<mask> they are bored). By the same token, taking a gifted student outside an accelerated program would make them horribly bored and not develop any sort of work ethic (much much worse). The bottom levels  should receive most of the attention,<mask> those in the top need some relevant material to be occupied. In an ideal world, those in the bottom will receive the most help<mask> those on top will receive the most difficult and interesting problems in order to keep busy. [NEWLINE] [NEWLINE] <mask> for judging people's ability, a teacher should have little trouble determining those with talent and those without it.<mask> they can't, that is a problem with that teachers competence and not with class organizational structure. [NEWLINE] [NEWLINE] It should be noted that I am unfamiliar with this program and<mask> effective it is (or is not).<mask> for biases, I am considered to be a gifted student. </s>
Label encoding: <s>Gifted student programs aren't to give students more help but to cover more material, quicker and provide a deeper understanding of the topic. A classic example would be in a Calculus curriculum. Almost every student (engineers included)does not need to know how limits really work and why the theorems work the way they do. Doing the epsilon-delta (something that will be in a accelerated course but not in a regular one) definition would be a waste of time, not only because it does not add any practical value but because those not good in math would not be able to understand it. For that reason it was only covered at the highest level in my high school. Some people could spend every hour of every day attempting to understanding it, and still not be able to comprehend it. Gifted programs are by nature more difficult (for the vast majority of students) because of the material (some exceptions are students who do not work in a non-accelerated program and do poorly simply because they are bored). By the same token, taking a gifted student outside an accelerated program would make them horribly bored and not develop any sort of work ethic (much much worse). The bottom levels  should receive most of the attention, but those in the top need some relevant material to be occupied. In an ideal world, those in the bottom will receive the most help while those on top will receive the most difficult and interesting problems in order to keep busy. [NEWLINE] [NEWLINE] As for judging people's ability, a teacher should have little trouble determining those with talent and those without it. If they can't, that is a problem with that teachers competence and not with class organizational structure. [NEWLINE] [NEWLINE] It should be noted that I am unfamiliar with this program and how effective it is (or is not). As for biases, I am considered to be a gifted student. </s>
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You don't have to make a game for everyone. Being all inclusive and all encompassing leads to watered down games.<mask> a consumer I don't want to enjoy an experience EVERYONE enjoys I want an experience I enjoy. The issue you're referring to is completely transparent with the "Casual vs Hardcore" argument. Anita Sarkeesian is a huge input consultant for Mirror's edge two and her mainstay criticism against the game is it's control scheme.<mask> is something that worked for gamers by and large in the first game being tampered with to be more inclusive for anything other than profits?<mask> EA's goal here is to improve their bottom line by creating a demographic shift, I don't really feel I have room to complain.<mask><mask> they're doing it for any other reason than that I feel it's an injustice to the audience that built the gaming industry in the first place. It's a businesses prerogative to do<mask> is profitable, and<mask> the 18-30 male demographic is the majority of sales and it's<mask> is going to maintain profitability long term there really isn't room to complain. Nobody likes the games that you're talking about.<mask> it were the case they would already be more popular. Money talks. I'm at the point<mask> AAA games are<mask> bland that I vehemently hate them,<mask> they're trying to cater to everyone in the name of money. I'd rather have something crude and crass and not progressive<mask> it meant that people like OP got to be passionate, load their games up with unique and interesting features and characters and not have their creativity stifled<mask> it meant hurting some feelings.<mask> A game company says that Chain-mail tops confer the same protection<mask> the male equivalent and they tastefully explain they're made with magic or better materials my suspension of disbelief will handle the rest. It's really not that hard.</s>
Label encoding: <s>You don't have to make a game for everyone. Being all inclusive and all encompassing leads to watered down games. As a consumer I don't want to enjoy an experience EVERYONE enjoys I want an experience I enjoy. The issue you're referring to is completely transparent with the "Casual vs Hardcore" argument. Anita Sarkeesian is a huge input consultant for Mirror's edge two and her mainstay criticism against the game is it's control scheme. Why is something that worked for gamers by and large in the first game being tampered with to be more inclusive for anything other than profits? If EA's goal here is to improve their bottom line by creating a demographic shift, I don't really feel I have room to complain. But if they're doing it for any other reason than that I feel it's an injustice to the audience that built the gaming industry in the first place. It's a businesses prerogative to do what is profitable, and if the 18-30 male demographic is the majority of sales and it's what is going to maintain profitability long term there really isn't room to complain. Nobody likes the games that you're talking about. If it were the case they would already be more popular. Money talks. I'm at the point where AAA games are so bland that I vehemently hate them, because they're trying to cater to everyone in the name of money. I'd rather have something crude and crass and not progressive if it meant that people like OP got to be passionate, load their games up with unique and interesting features and characters and not have their creativity stifled if it meant hurting some feelings. If A game company says that Chain-mail tops confer the same protection as the male equivalent and they tastefully explain they're made with magic or better materials my suspension of disbelief will handle the rest. It's really not that hard.</s>
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'll preface this by saying I don't believe it was an "inside job." I'm going to base<mask> I say on the movie, *Loose Change 9/11*, which<mask><mask> you would benefit from watching (<mask><mask> it addressed some of the questions you raised). Let's start. [NEWLINE] [NEWLINE] [STARTQ] <mask> exactly was the benefit of these unknown insiders destroying building 7? [ENDQ] [NEWLINE] The movie claimed that there was an entire department in that building with information of value to the investigations. Maybe someone can remind me of<mask>,<mask> the point is that you shouldn't dismiss the claim on this basis. [NEWLINE] [NEWLINE] [STARTQ] The buildings fell at free fall speed<mask> they were being piledriven by all the building on top of it. [ENDQ] [NEWLINE] From<mask> I know, it wasn't buildings. It was one building, which was number seven. It was the only one of the other centers to completely collapse. The other ones should have<mask> been covered in material,<mask> this should raise some suspicion. [NEWLINE] [NEWLINE] [STARTQ] <mask> would these noises mean? [ENDQ] [NEWLINE] Not just noises,<mask> explosions. There's a big difference<mask> witnesses claim they saw/heard evidence of explosions. [NEWLINE] [NEWLINE] [STARTQ] A building falling with such force pulverized almost everything in it's path, including rusty stuff and stuff made out of aluminum. [ENDQ] [NEWLINE] I'm not sure<mask> this addresses the claim that nano-thermite was found in the debris. It could be small,<mask> it may be seen or detected with instruments. Omega37 above,<mask>, seems to have evidence that this entire claim has been refuted. [NEWLINE] [NEWLINE] [STARTQ] <mask> could they have possibly gotten away with it? [ENDQ] [NEWLINE] Maybe they didn't, in a full sense. You can see who the movie claims to be part of the "inside" and might come to the conclusion that its in their best interest to keep quiet.</s>
Label encoding: <s>I'll preface this by saying I don't believe it was an "inside job." I'm going to base what I say on the movie, *Loose Change 9/11*, which I think you would benefit from watching ( I think it addressed some of the questions you raised). Let's start. [NEWLINE] [NEWLINE] [STARTQ] What exactly was the benefit of these unknown insiders destroying building 7? [ENDQ] [NEWLINE] The movie claimed that there was an entire department in that building with information of value to the investigations. Maybe someone can remind me of what, but the point is that you shouldn't dismiss the claim on this basis. [NEWLINE] [NEWLINE] [STARTQ] The buildings fell at free fall speed because they were being piledriven by all the building on top of it. [ENDQ] [NEWLINE] From what I know, it wasn't buildings. It was one building, which was number seven. It was the only one of the other centers to completely collapse. The other ones should have also been covered in material, so this should raise some suspicion. [NEWLINE] [NEWLINE] [STARTQ] What would these noises mean? [ENDQ] [NEWLINE] Not just noises, but explosions. There's a big difference when witnesses claim they saw/heard evidence of explosions. [NEWLINE] [NEWLINE] [STARTQ] A building falling with such force pulverized almost everything in it's path, including rusty stuff and stuff made out of aluminum. [ENDQ] [NEWLINE] I'm not sure how this addresses the claim that nano-thermite was found in the debris. It could be small, but it may be seen or detected with instruments. Omega37 above, though, seems to have evidence that this entire claim has been refuted. [NEWLINE] [NEWLINE] [STARTQ] How could they have possibly gotten away with it? [ENDQ] [NEWLINE] Maybe they didn't, in a full sense. You can see who the movie claims to be part of the "inside" and might come to the conclusion that its in their best interest to keep quiet.</s>
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Humanity going vegetarian or carnivorous is not primordial,<mask> is, is the betterment of animal industry practices, and increasing efficiency and sustainability. Going vegetarian<mask> a social statement is useless; it’s not even an effective way to solve all the problems it addresses. [NEWLINE] [NEWLINE] The only way it would be of utter importance, is<mask> our omnivorous diet were endangering to our species. You say the industry is unsustainable. That’s right, it is not eating meat per se<mask> ’s unsustainable,<mask> the industry.<mask> then stop eating meat? Fight<mask> ’s actually wrong. [NEWLINE] [NEWLINE] Here are some points for you to ponder<mask> becoming a vegetarian to solve anything is flawed and ineffective: [NEWLINE] [NEWLINE] Environmental impact: [NEWLINE] - Scenario: would you still be a vegetarian<mask> you could actually do more good to the environment by eating meat? Someone mentioned eating pests such<mask> rabbits. [NEWLINE] - Agriculture’s impact on the environment is still big.<mask>, depending solely on it would be endangering to the human species, don’t you think<mask>? (Think climate change). Of course, these problems can be solved, by technology. [NEWLINE] [NEWLINE] Morality: [NEWLINE] - Scenario: would you still be a vegetarian<mask> animals were farmed humanely.<mask><mask> you were absolutely sure they felt nothing<mask> killed?  (search for: “growing brain-dead chicken”) [NEWLINE] - Have you considered that these animals<mask> species are better having this kind of relationship with us than<mask> they were in the wild? [NEWLINE] - Would you eat meat<mask> it was grown in a lab? [NEWLINE] [NEWLINE] <mask><mask> vegetarianism<mask> a solution deviates us from progress. That’s<mask> you should instead advocate for technology, for improvement of the industry, not for vegetarianism.<mask><mask>. [NEWLINE] </s>
Label encoding: <s>Humanity going vegetarian or carnivorous is not primordial, what is, is the betterment of animal industry practices, and increasing efficiency and sustainability. Going vegetarian as a social statement is useless; it’s not even an effective way to solve all the problems it addresses. [NEWLINE] [NEWLINE] The only way it would be of utter importance, is if our omnivorous diet were endangering to our species. You say the industry is unsustainable. That’s right, it is not eating meat per se what ’s unsustainable, but the industry. Why then stop eating meat? Fight what ’s actually wrong. [NEWLINE] [NEWLINE] Here are some points for you to ponder why becoming a vegetarian to solve anything is flawed and ineffective: [NEWLINE] [NEWLINE] Environmental impact: [NEWLINE] - Scenario: would you still be a vegetarian if you could actually do more good to the environment by eating meat? Someone mentioned eating pests such as rabbits. [NEWLINE] - Agriculture’s impact on the environment is still big. Moreover, depending solely on it would be endangering to the human species, don’t you think so? (Think climate change). Of course, these problems can be solved, by technology. [NEWLINE] [NEWLINE] Morality: [NEWLINE] - Scenario: would you still be a vegetarian if animals were farmed humanely. What if you were absolutely sure they felt nothing when killed?  (search for: “growing brain-dead chicken”) [NEWLINE] - Have you considered that these animals as species are better having this kind of relationship with us than if they were in the wild? [NEWLINE] - Would you eat meat if it was grown in a lab? [NEWLINE] [NEWLINE] I think vegetarianism as a solution deviates us from progress. That’s why you should instead advocate for technology, for improvement of the industry, not for vegetarianism. IMO. [NEWLINE] </s>
Loss: tensor(0.0459, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Poetry is like a ship fairing across a raw stormy ocean. [NEWLINE] [NEWLINE] _____________________________________________________________________________ [NEWLINE] [NEWLINE] Unlike essays which rely on stable water [NEWLINE] [NEWLINE] Poetry is free from restriction [NEWLINE] [NEWLINE] There is no need for such organized fodder. [NEWLINE] _____________________________________________________________________________ [NEWLINE] Poetry is<mask> the most open in terms of creative control [NEWLINE] [NEWLINE] The writer is able to influence or try to deny the reader's comprehension [NEWLINE] [NEWLINE] Stories and papers require some basic structure in order to be considered for apprehension [NEWLINE] [NEWLINE] Poetry<mask><mask><mask><mask> can fit into no such hole [NEWLINE] _____________________________________________________________________________ [NEWLINE] A poem can be classy [NEWLINE] [NEWLINE] Perhaps outright pompous [NEWLINE] [NEWLINE] A poem can messy [NEWLINE] [NEWLINE] Or angry enough to whack you in your rumpus [NEWLINE] [NEWLINE] $ [NEWLINE] [NEWLINE] Some people think that poetry may be pointless [NEWLINE] [NEWLINE] <mask> I beg to differ [NEWLINE] [NEWLINE] I've never seen any other writing form cheesy enough [NEWLINE] [NEWLINE] To make a girl smile just before you kiss her [NEWLINE] [NEWLINE] $ [NEWLINE] [NEWLINE] Poetry is meant to express raw emotion [NEWLINE] [NEWLINE] The only restrictions present are those placed by the poet [NEWLINE] [NEWLINE] The wide variety of form and styles makes the lack of poetic originally a silly notion [NEWLINE] [NEWLINE] And spoken word poets try to make you feel their passions<mask> that you too can know it [NEWLINE] [NEWLINE] $ [NEWLINE] [NEWLINE] <mask><mask> that you should give poetry a second chance [NEWLINE] [NEWLINE] Rappers, historians, and teachers  disagree with the view expressed from your first glance. [NEWLINE] [NEWLINE] ________ [NEWLINE] [NEWLINE] Long story short [NEWLINE] [NEWLINE] There isn't such a need for memorization [NEWLINE] [NEWLINE] <mask> I find that appreciation comes from comprehension of poetry [NEWLINE] [NEWLINE] and the only way to keep a poem in ones mind involves memorization [NEWLINE] [NEWLINE] <mask> one can ponder for a long period of time [NEWLINE] _____ [NEWLINE] [NEWLINE] <mask><mask> : I'm a poet and I know it. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Label encoding: <s>Poetry is like a ship fairing across a raw stormy ocean. [NEWLINE] [NEWLINE] _____________________________________________________________________________ [NEWLINE] [NEWLINE] Unlike essays which rely on stable water [NEWLINE] [NEWLINE] Poetry is free from restriction [NEWLINE] [NEWLINE] There is no need for such organized fodder. [NEWLINE] _____________________________________________________________________________ [NEWLINE] Poetry is also the most open in terms of creative control [NEWLINE] [NEWLINE] The writer is able to influence or try to deny the reader's comprehension [NEWLINE] [NEWLINE] Stories and papers require some basic structure in order to be considered for apprehension [NEWLINE] [NEWLINE] Poetry on the other hand can fit into no such hole [NEWLINE] _____________________________________________________________________________ [NEWLINE] A poem can be classy [NEWLINE] [NEWLINE] Perhaps outright pompous [NEWLINE] [NEWLINE] A poem can messy [NEWLINE] [NEWLINE] Or angry enough to whack you in your rumpus [NEWLINE] [NEWLINE] $ [NEWLINE] [NEWLINE] Some people think that poetry may be pointless [NEWLINE] [NEWLINE] But I beg to differ [NEWLINE] [NEWLINE] I've never seen any other writing form cheesy enough [NEWLINE] [NEWLINE] To make a girl smile just before you kiss her [NEWLINE] [NEWLINE] $ [NEWLINE] [NEWLINE] Poetry is meant to express raw emotion [NEWLINE] [NEWLINE] The only restrictions present are those placed by the poet [NEWLINE] [NEWLINE] The wide variety of form and styles makes the lack of poetic originally a silly notion [NEWLINE] [NEWLINE] And spoken word poets try to make you feel their passions so that you too can know it [NEWLINE] [NEWLINE] $ [NEWLINE] [NEWLINE] I think that you should give poetry a second chance [NEWLINE] [NEWLINE] Rappers, historians, and teachers  disagree with the view expressed from your first glance. [NEWLINE] [NEWLINE] ________ [NEWLINE] [NEWLINE] Long story short [NEWLINE] [NEWLINE] There isn't such a need for memorization [NEWLINE] [NEWLINE] But I find that appreciation comes from comprehension of poetry [NEWLINE] [NEWLINE] and the only way to keep a poem in ones mind involves memorization [NEWLINE] [NEWLINE] So one can ponder for a long period of time [NEWLINE] _____ [NEWLINE] [NEWLINE] TLDR : I'm a poet and I know it. [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0478, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not going to say that Batman's moral code is "right" or anything. I'm going to<mask><mask> it's not Batman responsibility to kill the Joker. [NEWLINE] [NEWLINE] Let me put it this way:<mask> is it the Batman's job to kill the Joker? [NEWLINE] [NEWLINE] Batman does his thing<mask> a service to Gotham. He takes down criminals and super-villains and hands them off to the police. That's<mask> he ends his line of duty. A duty that he assigned himself. You can't really blame him for not doing more. [NEWLINE] [NEWLINE] <mask> you want to,<mask> not blame every cop that has ever had the Joker under arrest? They could have shot him at any time, saying he was resisting arrest, and everyone would LOVE them.<mask> not blame Gordon, who was given the opportunity by Batman to kill the Joker?<mask> any of the people working at Arkham?<mask> not the judges and people of Gotham that haven't just given him the death sentence? [NEWLINE] [NEWLINE] Batman is being heroic by putting himself out there, every night, risking death by taking down criminals. He doesn't want to kill, that's his decision. He's brought Joker to the police, to Arkham multiple times.<mask> many, many, many other individuals have had the opportunity and ability to kill the Joker,<mask> is Batman, the guy who's been responsible for stopping the Joker time after time, taking the blame for not killing him too? [NEWLINE] [NEWLINE] (A secondary point is that Batman keeps his code to retain the trust of the Gothamites (and the GCPD). Yes, he's a crazy man who goes around beating up thugs.<mask> he's never crossed the line and killed anyone. That demonstrates to the Gothamites (and the GDPD) that no matter<mask> crazy shit this person does, he has restraint to never, ever kill.)</s>
Label encoding: <s>I'm not going to say that Batman's moral code is "right" or anything. I'm going to argue that it's not Batman responsibility to kill the Joker. [NEWLINE] [NEWLINE] Let me put it this way: Why is it the Batman's job to kill the Joker? [NEWLINE] [NEWLINE] Batman does his thing as a service to Gotham. He takes down criminals and super-villains and hands them off to the police. That's where he ends his line of duty. A duty that he assigned himself. You can't really blame him for not doing more. [NEWLINE] [NEWLINE] If you want to, why not blame every cop that has ever had the Joker under arrest? They could have shot him at any time, saying he was resisting arrest, and everyone would LOVE them. Why not blame Gordon, who was given the opportunity by Batman to kill the Joker? Why any of the people working at Arkham? Why not the judges and people of Gotham that haven't just given him the death sentence? [NEWLINE] [NEWLINE] Batman is being heroic by putting himself out there, every night, risking death by taking down criminals. He doesn't want to kill, that's his decision. He's brought Joker to the police, to Arkham multiple times. If many, many, many other individuals have had the opportunity and ability to kill the Joker, why is Batman, the guy who's been responsible for stopping the Joker time after time, taking the blame for not killing him too? [NEWLINE] [NEWLINE] (A secondary point is that Batman keeps his code to retain the trust of the Gothamites (and the GCPD). Yes, he's a crazy man who goes around beating up thugs. But he's never crossed the line and killed anyone. That demonstrates to the Gothamites (and the GDPD) that no matter what crazy shit this person does, he has restraint to never, ever kill.)</s>
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0014, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] A bit of both. Obamacare is better than<mask> we had before,<mask> still not good enough.<mask><mask> the next move is state-by-state single-payer. Vermont, lead the way. Perhaps California next? [ENDQ] [NEWLINE] I know I'm not going to change your mind on this,<mask> ObamaCare<mask> it currently stands is vastly worse than<mask> we had before.  Our health care system is disintegrating, quickly.  I understand that you do not<mask> see this,<mask> from the perspective of healthcare providers, the end is now. [NEWLINE] [NEWLINE] <mask> you are curious at all about this, I can discuss at length,<mask> the stories are starting to show up.  It is a nightmare. [NEWLINE] [NEWLINE] That's not to say it can't be fixed,<mask> none of the proposals I have seen address the really serious problems. [NEWLINE] [NEWLINE] [STARTQ] Given the choice between control by an interested minority and control by the majority, I'll take the latter. [ENDQ] [NEWLINE] Have you ever heard of Edward Bernays?  He combined the theories of Freud with crowd psychology to create the modern field of public relations.  He personally was responsible for women smoking, the CIA-orchestrated fall of Guatemala in the 50's (and the term "banana republic"), the popularity of bacon, the popularity of water fluoridation, and he assisted with public opinion management during the Vietnam War.  His models were used by Goebbels against the Jews, and today is used by major corporations in fields ranging from health care to pharmaceuticals to energy. [NEWLINE] [NEWLINE] Edward Bernays ensured that we will never have control by a majority.  He singlehandedly invented a system that allows an interested minority to control a majority.  His books are available inexpensively on Amazon.com, and -- should you choose to read them -- will change your world.</s>
Label encoding: <s> [STARTQ] A bit of both. Obamacare is better than what we had before, but still not good enough. I think the next move is state-by-state single-payer. Vermont, lead the way. Perhaps California next? [ENDQ] [NEWLINE] I know I'm not going to change your mind on this, but ObamaCare as it currently stands is vastly worse than what we had before.  Our health care system is disintegrating, quickly.  I understand that you do not yet see this, but from the perspective of healthcare providers, the end is now. [NEWLINE] [NEWLINE] If you are curious at all about this, I can discuss at length, but the stories are starting to show up.  It is a nightmare. [NEWLINE] [NEWLINE] That's not to say it can't be fixed, but none of the proposals I have seen address the really serious problems. [NEWLINE] [NEWLINE] [STARTQ] Given the choice between control by an interested minority and control by the majority, I'll take the latter. [ENDQ] [NEWLINE] Have you ever heard of Edward Bernays?  He combined the theories of Freud with crowd psychology to create the modern field of public relations.  He personally was responsible for women smoking, the CIA-orchestrated fall of Guatemala in the 50's (and the term "banana republic"), the popularity of bacon, the popularity of water fluoridation, and he assisted with public opinion management during the Vietnam War.  His models were used by Goebbels against the Jews, and today is used by major corporations in fields ranging from health care to pharmaceuticals to energy. [NEWLINE] [NEWLINE] Edward Bernays ensured that we will never have control by a majority.  He singlehandedly invented a system that allows an interested minority to control a majority.  His books are available inexpensively on Amazon.com, and -- should you choose to read them -- will change your world.</s>
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0345, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>PRELIMINARY NOTE: I may or may not have made several unjustified assumptions in writing this.<mask> none of it is relevant to you, please disregard it. :) [NEWLINE] [NEWLINE] My first thought: get off the internet. [NEWLINE] [NEWLINE] My second thought: use your new free time to go volunteer somewhere. [NEWLINE] [NEWLINE] <mask> people are capable of some pretty nasty things by themselves, the internet is a mess. I don't know<mask>. Maybe it amplifies the negative, maybe it's the anonymity, maybe it's all the emotional context that gets lost in text, who knows.<mask> in my experience it's a lot easier to be cynical about people<mask> you don't interact with real people. [NEWLINE] [NEWLINE] <mask> for volunteering, it may seem kind of random<mask> I can think of several ways in which it could help. [NEWLINE] [NEWLINE] * You will spend time with people who spend their free time doing helpful, constructive things. These people are usually pretty cool. [NEWLINE] [NEWLINE] * You yourself will be a person who does helpful things in your free time. I've always found it easy to be optimistic about other people<mask> I feel good about<mask> I'm doing. [NEWLINE] [NEWLINE] * Depending on<mask> you volunteer you will serve people, and<mask><mask> it's never easier to love people than<mask> you're trying to serve them, even<mask> they're kind of jerks. I got punched in the face outside a soup kitchen once and I still feel positively about all my time there. [NEWLINE] [NEWLINE] <mask><mask> I say volunteer I don't necessarily mean humanitarian service. Help do tech for a community play, help out at a bicycle collective. Just getting involved regularly with other people in a constructive way *in real life*. [NEWLINE] [NEWLINE] It may not change the nature of humanity,<mask> it may help change your view of which side tends to dominate in real people.</s>
Label encoding: <s>PRELIMINARY NOTE: I may or may not have made several unjustified assumptions in writing this. If none of it is relevant to you, please disregard it. :) [NEWLINE] [NEWLINE] My first thought: get off the internet. [NEWLINE] [NEWLINE] My second thought: use your new free time to go volunteer somewhere. [NEWLINE] [NEWLINE] While people are capable of some pretty nasty things by themselves, the internet is a mess. I don't know why. Maybe it amplifies the negative, maybe it's the anonymity, maybe it's all the emotional context that gets lost in text, who knows. But in my experience it's a lot easier to be cynical about people when you don't interact with real people. [NEWLINE] [NEWLINE] As for volunteering, it may seem kind of random but I can think of several ways in which it could help. [NEWLINE] [NEWLINE] * You will spend time with people who spend their free time doing helpful, constructive things. These people are usually pretty cool. [NEWLINE] [NEWLINE] * You yourself will be a person who does helpful things in your free time. I've always found it easy to be optimistic about other people when I feel good about what I'm doing. [NEWLINE] [NEWLINE] * Depending on where you volunteer you will serve people, and I think it's never easier to love people than when you're trying to serve them, even if they're kind of jerks. I got punched in the face outside a soup kitchen once and I still feel positively about all my time there. [NEWLINE] [NEWLINE] Also when I say volunteer I don't necessarily mean humanitarian service. Help do tech for a community play, help out at a bicycle collective. Just getting involved regularly with other people in a constructive way *in real life*. [NEWLINE] [NEWLINE] It may not change the nature of humanity, but it may help change your view of which side tends to dominate in real people.</s>
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0337, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0538, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0522, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0366, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0744, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0290, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0331, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well contraception is contraception and abortion is abortion. Both have to do with a woman's reproductive system.<mask><mask> I'm going to go ahead and call the umbrella issues are different. [NEWLINE] [NEWLINE] Abortion is more or less the union of two issues: Bodily autonomy, and church vs state (again, more or less). Under topics relating to bodily autonomy you might find issues such<mask>... whether you can sell your liver after you die, or whether you can consume drugs in your own home. Under church vs state, you can find issues such<mask>... whether to include "under god" in the pledge of allegiance. Here, these issues boil down to a woman's control over her own body, and<mask> does it mean to be a human. Point being, these "umbrella issues" aren't the whole story,<mask> arguments for or against it usually fall under them. [NEWLINE] [NEWLINE] <mask><mask><mask><mask>, contraception availability is certainly not an issue about bodily autonomy or church vs state. It is an issue about<mask> to spend taxpayer dollars, the role of the government, and<mask> constitutes adverse selection. I can be pro or against contraception,<mask> for reasons having nothing to do with whether I am against abortion.<mask> I believe that requiring abortions<mask> backup is due to insufficient contraception, fine.<mask> I might not believe in abortions,<mask><mask> not believe in, I don't know, spending taxpayer dollars in a way that encourages sexual activity, spending taxpayer dollars on only one sex, or spending taxpayer dollars on giving people handouts at all. People like this aren't terribly difficult to find, by the way, and<mask> you should take away from this is that<mask> they are definitely connected, they are fundamentally different issues. Given your role on abortions, there are still many things that can swing your view one way or the other.</s>
Label encoding: <s>Well contraception is contraception and abortion is abortion. Both have to do with a woman's reproductive system. But what I'm going to go ahead and call the umbrella issues are different. [NEWLINE] [NEWLINE] Abortion is more or less the union of two issues: Bodily autonomy, and church vs state (again, more or less). Under topics relating to bodily autonomy you might find issues such as... whether you can sell your liver after you die, or whether you can consume drugs in your own home. Under church vs state, you can find issues such as... whether to include "under god" in the pledge of allegiance. Here, these issues boil down to a woman's control over her own body, and what does it mean to be a human. Point being, these "umbrella issues" aren't the whole story, but arguments for or against it usually fall under them. [NEWLINE] [NEWLINE] On the other hand, contraception availability is certainly not an issue about bodily autonomy or church vs state. It is an issue about where to spend taxpayer dollars, the role of the government, and what constitutes adverse selection. I can be pro or against contraception, but for reasons having nothing to do with whether I am against abortion. If I believe that requiring abortions as backup is due to insufficient contraception, fine. But I might not believe in abortions, but also not believe in, I don't know, spending taxpayer dollars in a way that encourages sexual activity, spending taxpayer dollars on only one sex, or spending taxpayer dollars on giving people handouts at all. People like this aren't terribly difficult to find, by the way, and what you should take away from this is that while they are definitely connected, they are fundamentally different issues. Given your role on abortions, there are still many things that can swing your view one way or the other.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I've seen some discussion about this topic,<mask> many kids and adults may call people something they aren't to try and insult them. For example: 'That's<mask> gay' and 'You throw like a girl' type insults. Many people come to the conclusion that these type of insults are homophobic and misogynistic,<mask> I don't believe that's necessarily the case at all. [NEWLINE] [NEWLINE] <mask><mask> the most insulting thing about being called a girl for a boy is the affront to his own identity. He is being called something he is not. He doesn't think its insulting<mask> he hates girls and thinks they're inferior,<mask> rather<mask> his identity<mask> a boy is being taken away. [NEWLINE] [NEWLINE] Women in general<mask> hate being associated with masculinity. They tend to want to be smaller than their partners<mask> they feel masculine<mask> they are taller. They religiously pluck facial hair and shave away body hair<mask> that's manly.<mask> they probably don't hate men. [NEWLINE] [NEWLINE] Similarly, being called gay is a huge insult to many guys not<mask> they are all homophobic,<mask><mask> they are being called something they are not. [NEWLINE] [NEWLINE] <mask><mask> the best way to see it from my point of view is looking at<mask> offensive it is for trans people<mask> they are mis-gendered (Purposely). It is a deliberate refusal of their identity. No one says a trans woman is a misandrist<mask> she hates being referred to<mask> 'he'. The people calling them 'he' are<mask> unlikely to hate men,<mask> say it to hurt the trans person. [NEWLINE] [NEWLINE] <mask> CMV, insults such<mask> those above aren't necessarily homophobic or misogynist. They sting<mask> they are an affront to identity more<mask> than<mask> being gay or a woman is a bad thing. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] </s>
Label encoding: <s>I've seen some discussion about this topic, where many kids and adults may call people something they aren't to try and insult them. For example: 'That's so gay' and 'You throw like a girl' type insults. Many people come to the conclusion that these type of insults are homophobic and misogynistic, but I don't believe that's necessarily the case at all. [NEWLINE] [NEWLINE] I think the most insulting thing about being called a girl for a boy is the affront to his own identity. He is being called something he is not. He doesn't think its insulting because he hates girls and thinks they're inferior, but rather because his identity as a boy is being taken away. [NEWLINE] [NEWLINE] Women in general also hate being associated with masculinity. They tend to want to be smaller than their partners because they feel masculine if they are taller. They religiously pluck facial hair and shave away body hair because that's manly. But they probably don't hate men. [NEWLINE] [NEWLINE] Similarly, being called gay is a huge insult to many guys not because they are all homophobic, but because they are being called something they are not. [NEWLINE] [NEWLINE] I think the best way to see it from my point of view is looking at how offensive it is for trans people when they are mis-gendered (Purposely). It is a deliberate refusal of their identity. No one says a trans woman is a misandrist if she hates being referred to as 'he'. The people calling them 'he' are also unlikely to hate men, but say it to hurt the trans person. [NEWLINE] [NEWLINE] So CMV, insults such as those above aren't necessarily homophobic or misogynist. They sting because they are an affront to identity more so than because being gay or a woman is a bad thing. [NEWLINE] [NEWLINE] _____ [NEWLINE] [NEWLINE] </s>
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0006, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0407, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0020, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Oh. Statement of Intent. Sorry, I shouldn't be taking part in the ideological debate before I've fully read the materials. There's a term called statement of intent I read today. Basically the same thing<mask> your middle paragraph. [NEWLINE] [NEWLINE] <mask> always use a Statement of Intent at some point. I understand. [NEWLINE] [NEWLINE] I believe that, like anything, this knowledge can and will be used in both a morally responsible and morally irresponsible way.  I should've learned<mask> to socialize,<mask> my school and home life were both nightmares. [NEWLINE] [NEWLINE] This world is not, and never will be perfect. I, and<mask> many like me have fallen through the cracks, and are broken<mask> of it. After you take the gun away from your head you decide to pick up the pieces and do<mask> you can to fix it. [NEWLINE] [NEWLINE] The first thing this man's guide instructs you to do is build self worth, and find out<mask> you want. It doesn't just give platitudes and tell us to "be ourselves". It tells us to get real. That women don't want a man who doesn't stay fit, dress well, and have good hygiene. That you need to value your own opinion and actually fucking have one. That you need to find out<mask> your standards for a women are beyond their looks and adhere to those standards. [NEWLINE] [NEWLINE] I'm sorry for<mask> that bastard did to you. That may not mean anything coming from<mask> you might picture me<mask>,<mask> I don't want to be that guy.<mask> is popularly known<mask> Game has evolved<mask> more and different men have adopted and adapted the ideology. I'll make sure to mention over at Seddit: **Always make a statement of Intent before heavy physical escalation.**<mask> you've got more to add, please do. </s>
Label encoding: <s>Oh. Statement of Intent. Sorry, I shouldn't be taking part in the ideological debate before I've fully read the materials. There's a term called statement of intent I read today. Basically the same thing as your middle paragraph. [NEWLINE] [NEWLINE] So always use a Statement of Intent at some point. I understand. [NEWLINE] [NEWLINE] I believe that, like anything, this knowledge can and will be used in both a morally responsible and morally irresponsible way.  I should've learned how to socialize, but my school and home life were both nightmares. [NEWLINE] [NEWLINE] This world is not, and never will be perfect. I, and so many like me have fallen through the cracks, and are broken because of it. After you take the gun away from your head you decide to pick up the pieces and do what you can to fix it. [NEWLINE] [NEWLINE] The first thing this man's guide instructs you to do is build self worth, and find out what you want. It doesn't just give platitudes and tell us to "be ourselves". It tells us to get real. That women don't want a man who doesn't stay fit, dress well, and have good hygiene. That you need to value your own opinion and actually fucking have one. That you need to find out what your standards for a women are beyond their looks and adhere to those standards. [NEWLINE] [NEWLINE] I'm sorry for what that bastard did to you. That may not mean anything coming from what you might picture me as, but I don't want to be that guy. What is popularly known as Game has evolved as more and different men have adopted and adapted the ideology. I'll make sure to mention over at Seddit: **Always make a statement of Intent before heavy physical escalation.** If you've got more to add, please do. </s>
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0308, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>1.  Okay,<mask> reports that sex took place would need to be specifically able to prove that the sex that happened involved mucous membranes or fluids touching other mucous membranes. <mask> would that happen? [NEWLINE] [NEWLINE] 2.  The results *published*?  You really see no problem with that? <mask>, you've said that it lessens your objection,<mask> you haven't answered the question:<mask> is it valid for you to tell people not to risk STIs by having sex,<mask> not valid for you to tell people not to go swing-dancing or rock-climbing or skiing, physical activities with a chance of injury? [NEWLINE] [NEWLINE] 3.  You haven't answered the question.  Do you think making extramarital sex illegal would be a significant deterrent to extramarital sex? [NEWLINE] [NEWLINE] 4. <mask> under your system, this would be illegal. [NEWLINE] [NEWLINE] 5.  This was not a study on *forced* abortion, and you haven't answered the question.  I am a doctor and an abortionist, and I'm very troubled by<mask> okay you are with forcing people to undergo procedures.  I, for one, would sooner cut off my own right hand than perform a procedure on somebody who didn't consent. [NEWLINE] [NEWLINE] <mask> please, don't be cowardly about it by misdirecting.  Answer the question.  Somebody is caught at twelve weeks pregnant, too late for a medication abortion.  Would you be willing to hold them down? <mask> they screamed and said, "no, I want to be pregnant, I want to have this baby!"  You would hold them down<mask> they could be anesthetized for a procedure? [NEWLINE] [NEWLINE] (For 2, 3, and 5, you did not answer the questions.)</s>
Label encoding: <s>1.  Okay, so reports that sex took place would need to be specifically able to prove that the sex that happened involved mucous membranes or fluids touching other mucous membranes.  How would that happen? [NEWLINE] [NEWLINE] 2.  The results *published*?  You really see no problem with that?  Also, you've said that it lessens your objection, but you haven't answered the question: why is it valid for you to tell people not to risk STIs by having sex, but not valid for you to tell people not to go swing-dancing or rock-climbing or skiing, physical activities with a chance of injury? [NEWLINE] [NEWLINE] 3.  You haven't answered the question.  Do you think making extramarital sex illegal would be a significant deterrent to extramarital sex? [NEWLINE] [NEWLINE] 4.  But under your system, this would be illegal. [NEWLINE] [NEWLINE] 5.  This was not a study on *forced* abortion, and you haven't answered the question.  I am a doctor and an abortionist, and I'm very troubled by how okay you are with forcing people to undergo procedures.  I, for one, would sooner cut off my own right hand than perform a procedure on somebody who didn't consent. [NEWLINE] [NEWLINE] So please, don't be cowardly about it by misdirecting.  Answer the question.  Somebody is caught at twelve weeks pregnant, too late for a medication abortion.  Would you be willing to hold them down?  As they screamed and said, "no, I want to be pregnant, I want to have this baby!"  You would hold them down so they could be anesthetized for a procedure? [NEWLINE] [NEWLINE] (For 2, 3, and 5, you did not answer the questions.)</s>
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0362, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0380, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It's all about the perspectives on sexuality that *both* partners have for their relationships.<mask> there is a huge mismatch between the two partners, in any direction, that's not a solid foundation for a relationship. A person who has tons of hookups with strangers and outside of relationships has a completely different view on sexuality than someone who sticks to people they are committed to or friends with. That fundamental perspective you have about sexuality is a very valid thing to discriminate against a partner on. It's not slut shaming, it's sexual compatibility. [NEWLINE] [NEWLINE] Here's my real world experience: I'm not hookup type person, I've only ever slept with people I already felt close to. That<mask> means I typically have some very long dry spells between relationships. I dated someone who was a very hookup type person. My experience was that she felt like she could invest less into the relationship<mask> losing it was less of a big deal for her, she could find someone else at the drop of a hat,<mask> I had a reason to stay and invest<mask> I knew that<mask> I left I would likely be alone for a long time afterward. She ended up being very controlling and abusive toward me until I finally snapped out of it and left. She met her current fiance about two weeks later, I'm still single four years later. [NEWLINE] [NEWLINE] On the flip side, I've<mask> seen the same type of dynamic in relationships between very promiscuous dudes and less-partnered ladies. This dynamic is considered a lot more normal and desirable in this direction,<mask> in my view it's still very unhealthy.<mask> you're a woman who mostly only sleeps with people you're close to, don't date a guy who's done a lot of hooking up, for exactly the reason my relationship with my ex was unhealthy.</s>
Label encoding: <s>It's all about the perspectives on sexuality that *both* partners have for their relationships. If there is a huge mismatch between the two partners, in any direction, that's not a solid foundation for a relationship. A person who has tons of hookups with strangers and outside of relationships has a completely different view on sexuality than someone who sticks to people they are committed to or friends with. That fundamental perspective you have about sexuality is a very valid thing to discriminate against a partner on. It's not slut shaming, it's sexual compatibility. [NEWLINE] [NEWLINE] Here's my real world experience: I'm not hookup type person, I've only ever slept with people I already felt close to. That also means I typically have some very long dry spells between relationships. I dated someone who was a very hookup type person. My experience was that she felt like she could invest less into the relationship because losing it was less of a big deal for her, she could find someone else at the drop of a hat, while I had a reason to stay and invest because I knew that if I left I would likely be alone for a long time afterward. She ended up being very controlling and abusive toward me until I finally snapped out of it and left. She met her current fiance about two weeks later, I'm still single four years later. [NEWLINE] [NEWLINE] On the flip side, I've also seen the same type of dynamic in relationships between very promiscuous dudes and less-partnered ladies. This dynamic is considered a lot more normal and desirable in this direction, but in my view it's still very unhealthy. If you're a woman who mostly only sleeps with people you're close to, don't date a guy who's done a lot of hooking up, for exactly the reason my relationship with my ex was unhealthy.</s>
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0320, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0423, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0361, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Getting angry<mask> people don't is a waste of energy. [ENDQ] [NEWLINE] It depends on your goals. It's possible many people said the same thing about movements against the sexist nature of US culture years ago. It's not a waste of energy<mask> there are more and more people getting angry all over the place. Eventually, behaviors, habits, and language can change. It doesn't<mask> people give up getting angry. [NEWLINE] [NEWLINE] I personally use "they" a lot b/c I don't like to imply gender and I'm not particularly interested in learning new pronouns.<mask> there are some set ones,<mask>, that I don't have to go out of my way to learn, then I'll probably start using those<mask> well. I can adapt and it matters to me that I don't offend people<mask> I can help it. I try to choose my words such that I won't offend anyone,<mask><mask> I do, I don't mind people correcting me.<mask> they get mad, I understand they are more angry at the culture that has instilled<mask> they see<mask> an injustice in me, rather than angry at me. [NEWLINE] [NEWLINE] <mask><mask> not hurting people's feelings are important. Not everyone will have thick skins, and even those that do can have had a rough day/week/month and maybe they just need someone to be nice to them. Suicide rates are great enough within the trans community.<mask> many may think it's just natural selection working it's course,<mask><mask> we've proven<mask> a species that sometimes<mask> seems like the weakest of us can do much good for our society<mask> given the proper environment. In the end, thinking about my words a little in the off chance that this individual could possibly affect society in a positive way (rather than killing themselves) is worth it.</s>
Label encoding: <s> [STARTQ] Getting angry when people don't is a waste of energy. [ENDQ] [NEWLINE] It depends on your goals. It's possible many people said the same thing about movements against the sexist nature of US culture years ago. It's not a waste of energy when there are more and more people getting angry all over the place. Eventually, behaviors, habits, and language can change. It doesn't when people give up getting angry. [NEWLINE] [NEWLINE] I personally use "they" a lot b/c I don't like to imply gender and I'm not particularly interested in learning new pronouns. When there are some set ones, however, that I don't have to go out of my way to learn, then I'll probably start using those as well. I can adapt and it matters to me that I don't offend people if I can help it. I try to choose my words such that I won't offend anyone, but if I do, I don't mind people correcting me. If they get mad, I understand they are more angry at the culture that has instilled what they see as an injustice in me, rather than angry at me. [NEWLINE] [NEWLINE] I think not hurting people's feelings are important. Not everyone will have thick skins, and even those that do can have had a rough day/week/month and maybe they just need someone to be nice to them. Suicide rates are great enough within the trans community. While many may think it's just natural selection working it's course, I think we've proven as a species that sometimes what seems like the weakest of us can do much good for our society if given the proper environment. In the end, thinking about my words a little in the off chance that this individual could possibly affect society in a positive way (rather than killing themselves) is worth it.</s>
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0018, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0301, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0391, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0298, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0280, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0435, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0313, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0501, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That is correct pricing for the cost of the device itself. [NEWLINE] [NEWLINE] <mask> you need to add to that the cost of two doctor's visits. They require them two weeks apart to be sure you aren't pregnant. You take a pregnancy test the first time, they do an exam and make sure you are able to have the device and still want it. Then they insert it<mask> you are on your period in the second appointment. *(ideally, or you take another pregnancy test and deal with not having a slightly more relaxed cervix)* [NEWLINE] [NEWLINE] <mask> a procedure/test/device is not covered by insurance, neither is the appointment itself.  It is probably easy,<mask> morally questionable, to get the first appointment disguised or coinciding with the yearly gyno exam.<mask> you would have to time that right.<mask> the second you would no doubt have to pay out of pocket. [NEWLINE] [NEWLINE] And that is going to run anywhere from 300-1000 depending on who your doctor is and<mask> kind of facility they are located in. [NEWLINE] [NEWLINE] It is<mask> possible to try to work around this by going to a free clinic, and then I believe you pay only for the device itself.<mask>, some clinics will only give IUDs to women who have already had children. Certainly not all,<mask> it was hit and miss last time I checked. Seemed to be mainly geographic, some areas all free clinics had the restrictions, in others none did. [NEWLINE] [NEWLINE] And<mask><mask> you are very right, about it being the cheapest over time, it takes time to save up that money to pay for it in the first place. To a low income person, it could very well take years, depending on the kindness of fate and<mask> other large expenses they need to save up for.</s>
Label encoding: <s>That is correct pricing for the cost of the device itself. [NEWLINE] [NEWLINE] But you need to add to that the cost of two doctor's visits. They require them two weeks apart to be sure you aren't pregnant. You take a pregnancy test the first time, they do an exam and make sure you are able to have the device and still want it. Then they insert it while you are on your period in the second appointment. *(ideally, or you take another pregnancy test and deal with not having a slightly more relaxed cervix)* [NEWLINE] [NEWLINE] When a procedure/test/device is not covered by insurance, neither is the appointment itself.  It is probably easy, but morally questionable, to get the first appointment disguised or coinciding with the yearly gyno exam. Although you would have to time that right. But the second you would no doubt have to pay out of pocket. [NEWLINE] [NEWLINE] And that is going to run anywhere from 300-1000 depending on who your doctor is and what kind of facility they are located in. [NEWLINE] [NEWLINE] It is also possible to try to work around this by going to a free clinic, and then I believe you pay only for the device itself. However, some clinics will only give IUDs to women who have already had children. Certainly not all, but it was hit and miss last time I checked. Seemed to be mainly geographic, some areas all free clinics had the restrictions, in others none did. [NEWLINE] [NEWLINE] And even though you are very right, about it being the cheapest over time, it takes time to save up that money to pay for it in the first place. To a low income person, it could very well take years, depending on the kindness of fate and what other large expenses they need to save up for.</s>
Loss: tensor(0.0294, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0379, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0363, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0532, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you're counting the O2 and N2<mask> pennies, then sure, there are ten thousand.<mask> that would have no relationship to reality at all<mask> they don't absorb infrared. [ENDQ] [NEWLINE] I'm counting all air<mask> ten thousand pennies. Four of those pennies equate to CO2, up from three pennies at the start of the industrial revolution. At that time, the logarithmic nature of CO2<mask> a greenhouse gas meant that it was already nearly opaque to capturing all infra red radiation at the few bands that it absorbs (which partially overlap with the bands H2O absorb). At that point all heat in those infra red bands have to either be reflected or retransmitted either up or down. Simple<mask> that. We can throw another thirty pennies on the pile and it won't matter much at this point now that we're past the logarithmic "knee".<mask> you don't understand that, then you don't understand basic mathematics. The CO2 can convect into higher altitudes to retransmit, or the heat must be absorbed by the earth or held in the atmosphere.<mask> the atmosphere doesn't retain it by fact that we have not measured the atmosphere to retain it for the last eighteen years, the heat either already re-radiated back into space, or it has been absorbed by the earth. Either that earth heat is in the crust or it is in the ocean. Prove which one (or mixture) of the three places that heat went. You can't! We simply do not currently have the instrumentation to do that. Even the new Argo system can only measure down to 2000 meters of ocean. All other possibilities are purely speculative, and certainly your reddit facts are meaningless.</s>
Label encoding: <s> [STARTQ] If you're counting the O2 and N2 as pennies, then sure, there are ten thousand. But that would have no relationship to reality at all because they don't absorb infrared. [ENDQ] [NEWLINE] I'm counting all air as ten thousand pennies. Four of those pennies equate to CO2, up from three pennies at the start of the industrial revolution. At that time, the logarithmic nature of CO2 as a greenhouse gas meant that it was already nearly opaque to capturing all infra red radiation at the few bands that it absorbs (which partially overlap with the bands H2O absorb). At that point all heat in those infra red bands have to either be reflected or retransmitted either up or down. Simple as that. We can throw another thirty pennies on the pile and it won't matter much at this point now that we're past the logarithmic "knee". If you don't understand that, then you don't understand basic mathematics. The CO2 can convect into higher altitudes to retransmit, or the heat must be absorbed by the earth or held in the atmosphere. Since the atmosphere doesn't retain it by fact that we have not measured the atmosphere to retain it for the last eighteen years, the heat either already re-radiated back into space, or it has been absorbed by the earth. Either that earth heat is in the crust or it is in the ocean. Prove which one (or mixture) of the three places that heat went. You can't! We simply do not currently have the instrumentation to do that. Even the new Argo system can only measure down to 2000 meters of ocean. All other possibilities are purely speculative, and certainly your reddit facts are meaningless.</s>
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I won't argue about taste,<mask>... who gives a shit really. A CMV about taste will get nobody anywhere, is boring, and useless.<mask>, it seems like the same type of justification<mask> many things<mask><mask>. The reason for taste is more important. [NEWLINE] [NEWLINE] Do you like beer?<mask> you like natty light and hate any and all beers with real texture and flavor, you probably don't really like beer. [NEWLINE] [NEWLINE] Do you like whiskey?<mask> you like a wallop of jack daniels and despise the taste of a glenfiddich 30 or glenlivit 21... you probably don't really like whiskey. [NEWLINE] [NEWLINE] <mask> you like a fully cooked steak... and despise a medium rare/rare, you probably don't like steak. [NEWLINE] [NEWLINE] In all these cases, it's more likely to me that you simply don't really like the 'essence' (meh... cliche<mask> whatever, the point is there) of<mask> something actually is. You like a bastard version of it.... which really isn't<mask> it *is*. [NEWLINE] [NEWLINE] <mask> I'd say, there is something wrong with claiming they are on equal footing.<mask> you only like something<mask> it's been bastardized into something completely unrecognizable, then there is something wrong with that. I'm quite sure it won't stop you from ordering it or attempting to refine your beer or whiskey or steak tastes obviously. [NEWLINE] [NEWLINE] <mask> like myself, I'll drink a natty light, and I'm perfectly happy with that, it tastes pretty good<mask> I feel like having one.<mask> I don't really like beer obviously. I'll stick with my glenfiddich 30 and my rare steaks for<mask> I want to enjoy things in that refined manner.</s>
Label encoding: <s>I won't argue about taste, because... who gives a shit really. A CMV about taste will get nobody anywhere, is boring, and useless. However, it seems like the same type of justification as many things I think. The reason for taste is more important. [NEWLINE] [NEWLINE] Do you like beer? If you like natty light and hate any and all beers with real texture and flavor, you probably don't really like beer. [NEWLINE] [NEWLINE] Do you like whiskey? If you like a wallop of jack daniels and despise the taste of a glenfiddich 30 or glenlivit 21... you probably don't really like whiskey. [NEWLINE] [NEWLINE] If you like a fully cooked steak... and despise a medium rare/rare, you probably don't like steak. [NEWLINE] [NEWLINE] In all these cases, it's more likely to me that you simply don't really like the 'essence' (meh... cliche but whatever, the point is there) of what something actually is. You like a bastard version of it.... which really isn't what it *is*. [NEWLINE] [NEWLINE] So I'd say, there is something wrong with claiming they are on equal footing. If you only like something because it's been bastardized into something completely unrecognizable, then there is something wrong with that. I'm quite sure it won't stop you from ordering it or attempting to refine your beer or whiskey or steak tastes obviously. [NEWLINE] [NEWLINE] But like myself, I'll drink a natty light, and I'm perfectly happy with that, it tastes pretty good when I feel like having one. But I don't really like beer obviously. I'll stick with my glenfiddich 30 and my rare steaks for when I want to enjoy things in that refined manner.</s>
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0338, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I completely agree: it's a pretty despicable, brutally Capitalistic idea,<mask><mask><mask>. It comes, I suppose, in part from a sense of encouraging people to earn their pay. One of those wonderful delusions of '<mask> you work real hard, you could become a millionaire on your tips alone, kid!' which has never (and will never) come to fruition. [NEWLINE] [NEWLINE] In practice, it's just exploitation, and—again, *<mask><mask><mask> *,—damages the whole service relationship<mask> there's a pretty explicit sense of pretence.<mask><mask> it undermines good service *for its own sake*. [NEWLINE] [NEWLINE] This extract from [an Esquire article]( [URL] /) is quite pertinent: [NEWLINE] [NEWLINE] [STARTQ] You are not technically stealing<mask> you don't tip the customary 15 to 20 percent,<mask> that's probably the best that can be said of you. The tip you pay is a sort of wage: federal law allows tips to be used to make up the difference between a server's salary and minimum wage, meaning they can make<mask> little<mask> £1.28 ($2) to £1.93 ($3) per hour from their restaurant employer. Tips are absolutely depended upon to make up the shortfall. [ENDQ] [NEWLINE] <mask> you leave a bad tip, you are docking a person's wages. This may either be<mask> you're confused about<mask>'s expected or<mask> you're an asshole, and you really believe that your sea bass arriving lukewarm is justly punishable by making it a little harder for the guy who brought it to you to pay his rent. [NEWLINE] [NEWLINE] In terms of changing your view,<mask><mask> not-liking tipping is fair enough.<mask>, in the US, at least it's something I feel you're morally obliged to take part in.</s>
Label encoding: <s>I completely agree: it's a pretty despicable, brutally Capitalistic idea, in my opinion. It comes, I suppose, in part from a sense of encouraging people to earn their pay. One of those wonderful delusions of'if you work real hard, you could become a millionaire on your tips alone, kid!' which has never (and will never) come to fruition. [NEWLINE] [NEWLINE] In practice, it's just exploitation, and—again, * in my opinion *,—damages the whole service relationship as there's a pretty explicit sense of pretence. I think it undermines good service *for its own sake*. [NEWLINE] [NEWLINE] This extract from [an Esquire article]( [URL] /) is quite pertinent: [NEWLINE] [NEWLINE] [STARTQ] You are not technically stealing if you don't tip the customary 15 to 20 percent, but that's probably the best that can be said of you. The tip you pay is a sort of wage: federal law allows tips to be used to make up the difference between a server's salary and minimum wage, meaning they can make as little as £1.28 ($2) to £1.93 ($3) per hour from their restaurant employer. Tips are absolutely depended upon to make up the shortfall. [ENDQ] [NEWLINE] When you leave a bad tip, you are docking a person's wages. This may either be because you're confused about what's expected or because you're an asshole, and you really believe that your sea bass arriving lukewarm is justly punishable by making it a little harder for the guy who brought it to you to pay his rent. [NEWLINE] [NEWLINE] In terms of changing your view, I think not-liking tipping is fair enough. But, in the US, at least it's something I feel you're morally obliged to take part in.</s>
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0384, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0372, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0395, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0278, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>It seems like you want a moral argument: [NEWLINE] [NEWLINE] Your definition of consent that you outlined is the absence of refusal,<mask> that's not<mask> consent is. There are various cases, for example, in which boyfriends have coerced their girlfriends into having sex,<mask> that the girls don't give permission<mask> don't or can't necessarily outright refuse, even<mask> they would not like to have sex. Most people would consider this<mask> a form of rape. This problem is exacerbated<mask> it comes to animals. There is not really any way to clearly communicate between animals and humans to do any consenting. Sure, maybe you can try to "assume" the animal is consenting<mask> it isn't resisting,<mask> some animals just stay really still out of fear. Even<mask> there's no readily apparent physical harm, there can be psychological harm, which is much harder to gauge. I've seen dogs that are very submissive<mask> they've gone through abuse,<mask> it's often hard to differentiate between those who are naturally submissive or forced into submission, especially<mask> it happened at a very early age. I suppose you could<mask><mask> the animals themselves might be the ones initiating the sex,<mask><mask> remember that there is a huge power imbalance. Animals are not humans, even<mask> they may have some human characteristics, and they do not think or respond<mask> we would. We are *much* smarter than animals, even more<mask> than children. We can sometimes literally condition them into<mask> we want them to be (a more extreme version of teachers or relatives "conditioning" children to love them in a sexual way).<mask> unless there comes a different species that is equally<mask> smart<mask> us and makes similar thought and behavior processes<mask> us, sex with animals will never be moral.</s>
Label encoding: <s>It seems like you want a moral argument: [NEWLINE] [NEWLINE] Your definition of consent that you outlined is the absence of refusal, but that's not what consent is. There are various cases, for example, in which boyfriends have coerced their girlfriends into having sex, so that the girls don't give permission but don't or can't necessarily outright refuse, even if they would not like to have sex. Most people would consider this as a form of rape. This problem is exacerbated when it comes to animals. There is not really any way to clearly communicate between animals and humans to do any consenting. Sure, maybe you can try to "assume" the animal is consenting because it isn't resisting, but some animals just stay really still out of fear. Even if there's no readily apparent physical harm, there can be psychological harm, which is much harder to gauge. I've seen dogs that are very submissive because they've gone through abuse, but it's often hard to differentiate between those who are naturally submissive or forced into submission, especially when it happened at a very early age. I suppose you could argue that the animals themselves might be the ones initiating the sex, but also remember that there is a huge power imbalance. Animals are not humans, even if they may have some human characteristics, and they do not think or respond as we would. We are *much* smarter than animals, even more so than children. We can sometimes literally condition them into what we want them to be (a more extreme version of teachers or relatives "conditioning" children to love them in a sexual way). So unless there comes a different species that is equally as smart as us and makes similar thought and behavior processes as us, sex with animals will never be moral.</s>
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0349, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I sort of agree with you<mask><mask><mask> the conclusion you draw is incorrect. [NEWLINE] [NEWLINE] In college, I was a resident advisor, and they made us all stand in a line. Then they read out random privileges ("My parents paid for my college" and such), and<mask> it applied to us we were supposed to take a step forward.<mask> not, a step back. After the set of questions was read, we were rather widely distributed. It was effective at demonstrating the point, that many of us are more privileged than others. [NEWLINE] [NEWLINE] **<mask>,** I was rather upset by the whole ordeal (not<mask> I was labeled<mask> more or less privileged -- I was somewhere around the middle/back),<mask><mask> it seemed to force on us the idea that we aren't all equals. I really, really didn't like the execution of the exercise,<mask> I was left thinking, "<mask>,<mask>, am I supposed to pity the people below me and be jealous of those above me? I'd prefer to just view them<mask> my peers and treat them all equally." [NEWLINE] [NEWLINE] <mask>, I certainly understand that emotional response that you have to these sorts of labels.<mask> I say you're wrong,<mask>, is that drawing the distinction is always bad (i.e. that people who do<mask> are horrible and parasites). No,<mask> uncomfortable<mask> it is to draw these distinctions, it's necessary<mask> we ever want to level the playing field, to give everyone an equal shot. Think of it less<mask> a retrospective thing and more<mask> a prospective thing -- it's not for the sake of guilting you for<mask> happened to you,<mask> rather for the sake of improving the circumstances for those who might otherwise be at a severe disadvantage in the future.</s>
Label encoding: <s>I sort of agree with you but I think the conclusion you draw is incorrect. [NEWLINE] [NEWLINE] In college, I was a resident advisor, and they made us all stand in a line. Then they read out random privileges ("My parents paid for my college" and such), and if it applied to us we were supposed to take a step forward. If not, a step back. After the set of questions was read, we were rather widely distributed. It was effective at demonstrating the point, that many of us are more privileged than others. [NEWLINE] [NEWLINE] ** However,** I was rather upset by the whole ordeal (not because I was labeled as more or less privileged -- I was somewhere around the middle/back), but because it seemed to force on us the idea that we aren't all equals. I really, really didn't like the execution of the exercise, because I was left thinking, " so, what, am I supposed to pity the people below me and be jealous of those above me? I'd prefer to just view them as my peers and treat them all equally." [NEWLINE] [NEWLINE] So, I certainly understand that emotional response that you have to these sorts of labels. Where I say you're wrong, though, is that drawing the distinction is always bad (i.e. that people who do so are horrible and parasites). No, as uncomfortable as it is to draw these distinctions, it's necessary if we ever want to level the playing field, to give everyone an equal shot. Think of it less as a retrospective thing and more as a prospective thing -- it's not for the sake of guilting you for what happened to you, but rather for the sake of improving the circumstances for those who might otherwise be at a severe disadvantage in the future.</s>
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0272, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> is the ugliness and imperfection then? [ENDQ] [NEWLINE] Satan or the Devil of course! (Sort of tounge in cheek here) [NEWLINE] [NEWLINE] [STARTQ] You say that the world is not chaotic, not arbitrary, not random and<mask><mask> much of it is. [ENDQ] [NEWLINE] Yes, chaos exists too.<mask> is obviously not the dominant force.<mask> it were, I'd imagine very little would exist<mask> chaos is self-destructive/self-defeating. [NEWLINE] [NEWLINE] <mask> go deeper a level to the Law of Identity, that whatever a thing is, it is that thing and not "not that thing". We live in a universe<mask> Jupiter doesn't turn into a Squirrel,<mask> remains Jupiter. Underlying the apparent chaos and randomness is a startling amount or order. [NEWLINE] [NEWLINE] Even death isn't an arbitrary thing - it exists<mask> an evolved and natural solution to the problem of<mask> life can keep existing overall. And horrors like cancer and "random bus accidents" and horns that curl into your own skull aren't actually *causeless* - they have reasons they occurred, causes we may one day discover using science and logic, and one day defeat! That's the battle of the good guys - the lab assistant that discovers a cure, the engineer who invents pedestrian sensors, the geneticist who turns off a gene. [NEWLINE] [NEWLINE] Another option is to see the randomness itself<mask> causal, and angst upon the existential insanity of that premise -<mask> that would be giving up. And in religious terminology, allowing evil to win. [NEWLINE] [NEWLINE] The truth is everything happens for a reason, in that everything has a logical cause. "Randomness" is simply an apparent effect of "lacking the knowledge"<mask> to which cup hides the ball. </s>
Label encoding: <s> [STARTQ] What is the ugliness and imperfection then? [ENDQ] [NEWLINE] Satan or the Devil of course! (Sort of tounge in cheek here) [NEWLINE] [NEWLINE] [STARTQ] You say that the world is not chaotic, not arbitrary, not random and yet so much of it is. [ENDQ] [NEWLINE] Yes, chaos exists too. But is obviously not the dominant force. If it were, I'd imagine very little would exist as chaos is self-destructive/self-defeating. [NEWLINE] [NEWLINE] But go deeper a level to the Law of Identity, that whatever a thing is, it is that thing and not "not that thing". We live in a universe where Jupiter doesn't turn into a Squirrel, but remains Jupiter. Underlying the apparent chaos and randomness is a startling amount or order. [NEWLINE] [NEWLINE] Even death isn't an arbitrary thing - it exists as an evolved and natural solution to the problem of how life can keep existing overall. And horrors like cancer and "random bus accidents" and horns that curl into your own skull aren't actually *causeless* - they have reasons they occurred, causes we may one day discover using science and logic, and one day defeat! That's the battle of the good guys - the lab assistant that discovers a cure, the engineer who invents pedestrian sensors, the geneticist who turns off a gene. [NEWLINE] [NEWLINE] Another option is to see the randomness itself as causal, and angst upon the existential insanity of that premise - but that would be giving up. And in religious terminology, allowing evil to win. [NEWLINE] [NEWLINE] The truth is everything happens for a reason, in that everything has a logical cause. "Randomness" is simply an apparent effect of "lacking the knowledge" as to which cup hides the ball. </s>
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0309, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> should you even bother knowing another system of measurments, such<mask> the imperial, to quantify length, area,volume and weight<mask> again? [ENDQ] [NEWLINE] <mask> in this corner of the world, there *is no choice<mask> to know it*<mask> you are working on anything that may have been built before 1995 or<mask>, before people started switching over. My 2011 Ford Fiesta has metric parts,<mask><mask> I understand it, that's<mask> it was designed in Germany and built in Mexico.<mask> on my '97 Mustang, everything was in Imperial. [NEWLINE] [NEWLINE] The thing is, I can, and have, worked in both Imperial and Metric. I learned both side by side, and they each have their strengths and weaknesses.<mask> I am building some sort of fluid container, I like to use metric<mask> knowing<mask> big my sides are will give me a pretty easy conversion to volume, whereas SI units are just fucked in that respect.<mask><mask> I'm building something<mask> I need to know a lot of proportions, I like Imperial<mask> it typically divides into common numbers easier, and allows me to do things on the fly, in my head, and be able to eyeball it easier. Most of the time,<mask>,<mask> system I use depends on<mask> parts I happen to be building with.<mask> it's completely freeform, I use inches<mask> I recognize that I have that bias in thinking and in culture; I mentally think of length and mass in Imperial units. Just like<mask> I listen and think, I can understand French (I took 3 years in High School and then subsequently rarely practiced,<mask> I'm *very* rusty),<mask> I don't naturally look at a blue house and thing "aah, le chateau bleu"</s>
Label encoding: <s> [STARTQ] why should you even bother knowing another system of measurments, such as the imperial, to quantify length, area,volume and weight yet again? [ENDQ] [NEWLINE] Because in this corner of the world, there *is no choice but to know it* if you are working on anything that may have been built before 1995 or so, before people started switching over. My 2011 Ford Fiesta has metric parts, but as I understand it, that's because it was designed in Germany and built in Mexico. But on my '97 Mustang, everything was in Imperial. [NEWLINE] [NEWLINE] The thing is, I can, and have, worked in both Imperial and Metric. I learned both side by side, and they each have their strengths and weaknesses. If I am building some sort of fluid container, I like to use metric because knowing how big my sides are will give me a pretty easy conversion to volume, whereas SI units are just fucked in that respect. But if I'm building something where I need to know a lot of proportions, I like Imperial because it typically divides into common numbers easier, and allows me to do things on the fly, in my head, and be able to eyeball it easier. Most of the time, though, what system I use depends on what parts I happen to be building with. If it's completely freeform, I use inches because I recognize that I have that bias in thinking and in culture; I mentally think of length and mass in Imperial units. Just like if I listen and think, I can understand French (I took 3 years in High School and then subsequently rarely practiced, so I'm *very* rusty), but I don't naturally look at a blue house and thing "aah, le chateau bleu"</s>
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0346, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Your title is not related to or backed up by your argument. You claim to disagree with "Secular Jewish culture"<mask> by the end of the argument express a desire to claim that you "Don't like the Jewish people". These are two very different things. The first would be a reasonable criticism of a culture<mask> the second, being generalized to an entire race, would equate to racism. And yes, I'd say the same about the Japanese or Hispanics etc [NEWLINE] [NEWLINE] Your first argument is that the Jews have a history of perpetrating genocide. This is 100% incorrect.<mask><mask> that in ancient times they were involved in wars of conquest<mask> never genocide, and there is no single example of this<mask>. [NEWLINE] [NEWLINE] Moving on to the current war. Personally<mask><mask> that<mask> many Israelis being personally racist, and far too many civilians are dying, there is no way that<mask> the IDF is doing could be considered a genocide. [NEWLINE] [NEWLINE] We can argue this all day,<mask> the real issue that I have with your argument is that there is no reasonable correlation between the actions of the IDF and secular Jewish culture. Secular Jewish culture is about community, education and family, not at all about conflict. Please explain<mask> you think the culture of Jews, most of whom live outside of Israel, is somehow related to genocide. There are no books, festivals or rituals that in any way promote aggressive war or genocide. [NEWLINE] [NEWLINE] Your second point focused on religious practices such<mask> circumcision.<mask> these are religious not secular. A large percentage of secular Jews don't keep kosher and no secular Jew would have a moil suck their child's penis. Your arguments against these rituals are reasonable<mask> they don't excuse a dislike for the entire Jewish people<mask> a secular race.</s>
Label encoding: <s>Your title is not related to or backed up by your argument. You claim to disagree with "Secular Jewish culture" but by the end of the argument express a desire to claim that you "Don't like the Jewish people". These are two very different things. The first would be a reasonable criticism of a culture while the second, being generalized to an entire race, would equate to racism. And yes, I'd say the same about the Japanese or Hispanics etc [NEWLINE] [NEWLINE] Your first argument is that the Jews have a history of perpetrating genocide. This is 100% incorrect. I agree that in ancient times they were involved in wars of conquest but never genocide, and there is no single example of this since. [NEWLINE] [NEWLINE] Moving on to the current war. Personally I think that despite many Israelis being personally racist, and far too many civilians are dying, there is no way that what the IDF is doing could be considered a genocide. [NEWLINE] [NEWLINE] We can argue this all day, but the real issue that I have with your argument is that there is no reasonable correlation between the actions of the IDF and secular Jewish culture. Secular Jewish culture is about community, education and family, not at all about conflict. Please explain how you think the culture of Jews, most of whom live outside of Israel, is somehow related to genocide. There are no books, festivals or rituals that in any way promote aggressive war or genocide. [NEWLINE] [NEWLINE] Your second point focused on religious practices such as circumcision. But these are religious not secular. A large percentage of secular Jews don't keep kosher and no secular Jew would have a moil suck their child's penis. Your arguments against these rituals are reasonable but they don't excuse a dislike for the entire Jewish people as a secular race.</s>
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0493, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0314, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0369, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0440, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I think /u/sunnyEl-ahrairah gave the best advice you're going to get in this thread,<mask> I'll throw in my $0.02: [NEWLINE] [NEWLINE] **<mask> Would Jesus Do?** [NEWLINE] [NEWLINE] Just<mask> you no longer believe that Jesus is the Son of God doesn't mean you can't still have him<mask> a role model. Even<mask> Jesus is completely made up, there's no denying that he's a very wise person or character. He was a visionary way ahead of his time. I'm a pantheist, and<mask><mask> Jesus is awesome. Many Buddhists consider Jesus an enlightened being, just like the Buddha himself. [NEWLINE] [NEWLINE] <mask> perhaps,<mask> you still think that his teachings are worth following, you should try to imagine<mask> he'd do in your shoes. Or think of another figure you admire. A saint, a philosopher, anyone who you feel would be able to navigate a crisis of faith with confidence.<mask> Would Mister Rogers Do? [NEWLINE] [NEWLINE] For<mask> it's worth, I was a Roman Catholic and my crisis of faith lasted for several years. I came out the other end with a completely different worldview, one that I'm confident in, that gives me comfort and fills me with awe at<mask> wonderful life is. It's worth facing this struggle head-on,<mask> don't try to steer it in any particular direction,<mask> that won't work. Simply keep an open mind and honestly try to find something that resonates with you. [NEWLINE] [NEWLINE] TO THE MODS: I know that this comment is not strictly a challenge on the OP's views<mask> per the rules,<mask> it *is* an honest attempt to help OP to reach an answer. I hope it's acceptable.</s>
Label encoding: <s>I think /u/sunnyEl-ahrairah gave the best advice you're going to get in this thread, but I'll throw in my $0.02: [NEWLINE] [NEWLINE] ** What Would Jesus Do?** [NEWLINE] [NEWLINE] Just because you no longer believe that Jesus is the Son of God doesn't mean you can't still have him as a role model. Even if Jesus is completely made up, there's no denying that he's a very wise person or character. He was a visionary way ahead of his time. I'm a pantheist, and I think Jesus is awesome. Many Buddhists consider Jesus an enlightened being, just like the Buddha himself. [NEWLINE] [NEWLINE] So perhaps, if you still think that his teachings are worth following, you should try to imagine what he'd do in your shoes. Or think of another figure you admire. A saint, a philosopher, anyone who you feel would be able to navigate a crisis of faith with confidence. What Would Mister Rogers Do? [NEWLINE] [NEWLINE] For what it's worth, I was a Roman Catholic and my crisis of faith lasted for several years. I came out the other end with a completely different worldview, one that I'm confident in, that gives me comfort and fills me with awe at how wonderful life is. It's worth facing this struggle head-on, but don't try to steer it in any particular direction, because that won't work. Simply keep an open mind and honestly try to find something that resonates with you. [NEWLINE] [NEWLINE] TO THE MODS: I know that this comment is not strictly a challenge on the OP's views as per the rules, but it *is* an honest attempt to help OP to reach an answer. I hope it's acceptable.</s>
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0327, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0417, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0038, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Yes,<mask><mask> does that have to do with<mask> you are saying? You said a stereotype needs to be an accurate generalisation. I gave you an example<mask> that is clearly false. [ENDQ] [NEWLINE] I consider that stereotype dying and of little concern. It's on its way to becoming extinct<mask> of<mask> I was originally implying: basically, showing that it's not true. It's still believed by many people<mask> information does not travel to every person instantaneously.<mask><mask> a lot of stereotypes are in that phase. My original thought process centered around stereotypes that are based in some kind of measurable statistic, with a truth value to them. Earlier another person made me realize that was unfair for me to do<mask> making sweeping statements about stereotypes. [NEWLINE] [NEWLINE] [STARTQ] <mask><mask> are you saying then? [ENDQ] [NEWLINE] In a case like that, I'm saying that victims are not only victims of circumstance and stereotypes, they can<mask> be victims of their own choices.<mask> I believe a black man does have more control over the outcome of an interaction with a police officer than<mask> he's given credit for. I believe there's a universal standard for dealing with police officers that stereotyped victims foolishly disregard<mask> of emotion (obviously every type of person can disregard that standard,<mask> black people get victimized more for not respecting it, which is a different topic).<mask> they disregard the appropriate way to deal with such a situation, the outcome is much worse than it would have been otherwise. I'm not saying "black people just need to be nice to cops, then everything will be fine!" I'm saying the chances of black men dying or being wrongfully arrested will be drastically lowered<mask> they break the aggressive black male stereotype that police officers deal with every single day.</s>
Label encoding: <s> [STARTQ] Yes, but what does that have to do with what you are saying? You said a stereotype needs to be an accurate generalisation. I gave you an example where that is clearly false. [ENDQ] [NEWLINE] I consider that stereotype dying and of little concern. It's on its way to becoming extinct because of what I was originally implying: basically, showing that it's not true. It's still believed by many people because information does not travel to every person instantaneously. I think a lot of stereotypes are in that phase. My original thought process centered around stereotypes that are based in some kind of measurable statistic, with a truth value to them. Earlier another person made me realize that was unfair for me to do when making sweeping statements about stereotypes. [NEWLINE] [NEWLINE] [STARTQ] So what are you saying then? [ENDQ] [NEWLINE] In a case like that, I'm saying that victims are not only victims of circumstance and stereotypes, they can also be victims of their own choices. So I believe a black man does have more control over the outcome of an interaction with a police officer than what he's given credit for. I believe there's a universal standard for dealing with police officers that stereotyped victims foolishly disregard because of emotion (obviously every type of person can disregard that standard, but black people get victimized more for not respecting it, which is a different topic). When they disregard the appropriate way to deal with such a situation, the outcome is much worse than it would have been otherwise. I'm not saying "black people just need to be nice to cops, then everything will be fine!" I'm saying the chances of black men dying or being wrongfully arrested will be drastically lowered if they break the aggressive black male stereotype that police officers deal with every single day.</s>
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0333, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yes, that's all true.<mask> you have to understand that 4chan is buy and large full of people who have a use for the anonymity it offers. Some people use it to call people names online without consequence, some people use it to ERP, some people use it to complain about<mask> much they hate their life and want to die.<mask> a lot of these things are simply expressed here<mask> much<mask> they can't express them anywhere else safely. I don't think that the people on 4chan are really all that bad in real life. On /vg/, I frequent League of Legends generals and I often play with people from those generals. League of Legends general is a place<mask> we can't go six posts without arguing with eachother about which waifus are shit or which tripfag deserves to be decapitated or<mask> much we want to stick a fork in eachothers eyes for building the wrong item on the wrong character.<mask><mask> I get into Skype with these people to play games with them, they're totally normal. They're just regular people who blow off steam by calling other people cuckolds on 4chan.<mask> anything, I'd say that 4chan's environment is pretty healthy. There's something to be said about a place<mask> you can go to blow off steam and say basically whatever you want without consequence.<mask> you go into the experience thinking "I wanna fit in with everyone else here on 4chan and be angry and upset all the time!", then yeah, your life will change for the worse.<mask> I don't think 4chan did this to any of these people.<mask><mask> they're just letting go of baggage they've been carrying long before arriving here, and can now finally drop.</s>
Label encoding: <s>Yes, that's all true. But you have to understand that 4chan is buy and large full of people who have a use for the anonymity it offers. Some people use it to call people names online without consequence, some people use it to ERP, some people use it to complain about how much they hate their life and want to die. But a lot of these things are simply expressed here so much because they can't express them anywhere else safely. I don't think that the people on 4chan are really all that bad in real life. On /vg/, I frequent League of Legends generals and I often play with people from those generals. League of Legends general is a place where we can't go six posts without arguing with eachother about which waifus are shit or which tripfag deserves to be decapitated or how much we want to stick a fork in eachothers eyes for building the wrong item on the wrong character. But when I get into Skype with these people to play games with them, they're totally normal. They're just regular people who blow off steam by calling other people cuckolds on 4chan. If anything, I'd say that 4chan's environment is pretty healthy. There's something to be said about a place where you can go to blow off steam and say basically whatever you want without consequence. If you go into the experience thinking "I wanna fit in with everyone else here on 4chan and be angry and upset all the time!", then yeah, your life will change for the worse. But I don't think 4chan did this to any of these people. I think they're just letting go of baggage they've been carrying long before arriving here, and can now finally drop.</s>
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Here's the thing. <mask><mask>, based on my life experiences (first-hand, book, radio, etc.) that it is untrue that "Most people need a dual income to procure and afford a mortgage, nevermind the cost it takes to raise a child."  You, obviously, disagree.  Truthfully,<mask>, neither of us is using data,<mask> I'm afraid we're just going to keep talking at one another.  Further note,<mask>,<mask> does it say that owning your own home is a necessity to be a parent? [NEWLINE] [NEWLINE] The "just move" argument is actually very sound *financial advice.* <mask> you check out r/frugal and r/personalfinance, you will see over and over that housing is one of everyone' s biggest expenses and many people (especially in a suburb) already live somewhat near another town with slightly lower cost of living.  Admittedly, the "just move" argument is *terrible* social advice.  It would obviously be really hard to alter your social relationships and support networks.  It is clearly not an options for everyone.  A friend of mine,<mask>, moved from our town into a nicer bigger house in a nicer town at approximately the same time<mask> making this "we can't afford for me to stay home" comment.  Sure, I may be making an assumption,<mask> it seems a pretty safe assumption that they could have staying in the cheaper town &amp; house and saved quite a bit of money.  The new town does have better schools and lower crime,<mask><mask> they had said, "I went back to work<mask> the kids could go to nicer schools and be safer," I would have no argument.</s>
Label encoding: <s>Here's the thing.  I think, based on my life experiences (first-hand, book, radio, etc.) that it is untrue that "Most people need a dual income to procure and afford a mortgage, nevermind the cost it takes to raise a child."  You, obviously, disagree.  Truthfully, though, neither of us is using data, so I'm afraid we're just going to keep talking at one another.  Further note, though, where does it say that owning your own home is a necessity to be a parent? [NEWLINE] [NEWLINE] The "just move" argument is actually very sound *financial advice.*  If you check out r/frugal and r/personalfinance, you will see over and over that housing is one of everyone' s biggest expenses and many people (especially in a suburb) already live somewhat near another town with slightly lower cost of living.  Admittedly, the "just move" argument is *terrible* social advice.  It would obviously be really hard to alter your social relationships and support networks.  It is clearly not an options for everyone.  A friend of mine, however, moved from our town into a nicer bigger house in a nicer town at approximately the same time as making this "we can't afford for me to stay home" comment.  Sure, I may be making an assumption, but it seems a pretty safe assumption that they could have staying in the cheaper town &amp; house and saved quite a bit of money.  The new town does have better schools and lower crime, so if they had said, "I went back to work so the kids could go to nicer schools and be safer," I would have no argument.</s>
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I don't believe there is a better sex [ENDQ] [NEWLINE] Feminism doesn't either,<mask> it has acknowledged a societal preference for one sex over another. [NEWLINE] [NEWLINE] [STARTQ] Your genitalia is not indicative of your role in society, your actions and decisions are [ENDQ] [NEWLINE] You can partly thank feminist thought for popularizing this sentiment,<mask> I'm sure making dick jokes has been a common human endeavor<mask> time immemorial. [NEWLINE] [NEWLINE] [STARTQ] feminism is a movement for women, not equality for all sexes [ENDQ] [NEWLINE] <mask> feminism is primarily concerned with the plight of women, it has resulted in theory and policies that benefit men, children, minorities, homosexuals, and transgenders alike. For example, the issue of the patriarchy (a much maligned term here on reddit) has influenced men<mask> the 70s to seek [male liberation]( [URL] %27s_liberation_movement) from traditional gender roles, much in the same way women sought to emancipate themselves from gender roles. (Men's liberation should not be confused with men's rights activism, an anti-feminist movement). [NEWLINE] [NEWLINE] [STARTQ] special interest group [ENDQ] [NEWLINE] Feminists are not a special interest group. There may be certain organizations that have a pro-feminist agenda,<mask> feminism<mask> a whole isn't an interest group.<mask> you have issues with certain organizations, then you should name them. For instance, gun-owners are not a special interest group,<mask> the National Rifle Association (NRA) is. [NEWLINE] [NEWLINE] <mask><mask> I'm on the NRA,<mask> don't we change that name too? After all, the NRA isn't just about teaching rifle marksmanship anymore. They are concerned with handguns, assault rifles, national background checks...</s>
Label encoding: <s> [STARTQ] I don't believe there is a better sex [ENDQ] [NEWLINE] Feminism doesn't either, but it has acknowledged a societal preference for one sex over another. [NEWLINE] [NEWLINE] [STARTQ] Your genitalia is not indicative of your role in society, your actions and decisions are [ENDQ] [NEWLINE] You can partly thank feminist thought for popularizing this sentiment, though I'm sure making dick jokes has been a common human endeavor since time immemorial. [NEWLINE] [NEWLINE] [STARTQ] feminism is a movement for women, not equality for all sexes [ENDQ] [NEWLINE] While feminism is primarily concerned with the plight of women, it has resulted in theory and policies that benefit men, children, minorities, homosexuals, and transgenders alike. For example, the issue of the patriarchy (a much maligned term here on reddit) has influenced men since the 70s to seek [male liberation]( [URL] %27s_liberation_movement) from traditional gender roles, much in the same way women sought to emancipate themselves from gender roles. (Men's liberation should not be confused with men's rights activism, an anti-feminist movement). [NEWLINE] [NEWLINE] [STARTQ] special interest group [ENDQ] [NEWLINE] Feminists are not a special interest group. There may be certain organizations that have a pro-feminist agenda, but feminism as a whole isn't an interest group. If you have issues with certain organizations, then you should name them. For instance, gun-owners are not a special interest group, but the National Rifle Association (NRA) is. [NEWLINE] [NEWLINE] So while I'm on the NRA, why don't we change that name too? After all, the NRA isn't just about teaching rifle marksmanship anymore. They are concerned with handguns, assault rifles, national background checks...</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0530, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] my biggest issue with evolution is odds. The statistics that any sort of anything, even a single molecule, could be created by accident, ex nihilo, is just preposterous to me.<mask><mask> to that, the sheer size and complexity of the universe exponentially decreases these odds, and I find it baffling anyone could believe something<mask> unlikely,<mask> impossible, could even happen. [ENDQ] [NEWLINE] The chances of reality being exactly the way it is,<mask><mask> whether it was "random" or created, are very tiny.<mask><mask> you roll a bazillion-sided die, you are guaranteed to get a result than only has a 1-in-bazillion chance of happening.<mask> no matter<mask> life or the universe is, the way it turned out is expected to be unlikely compared to all other (infinite?) options. [NEWLINE] [NEWLINE] <mask> this is unrelated to whether or not it was created by a God.<mask> things are unlikely to "randomly" be the way they are, they are equally<mask> unlikely to have been created *exactly* this way by a God, who could have chosen infinite other possibilities.<mask> does that mean it should be "baffling" that anyone could believe something<mask> unlikely<mask> a God who happens to have the exact characteristics to have created exactly this universe? Let's just agree that our existence is baffling<mask><mask> evolution and/or God. [NEWLINE] [NEWLINE] <mask> for your original question about teaching ID alongside evolution; I do think both should be taught<mask> not *alongside* each other. Evolution should be taught in a biology class, and intelligent design should be taught in philosophy and/or religion classes.<mask><mask> we don't have enough variety of subjects throughout school.</s>
Label encoding: <s> [STARTQ] my biggest issue with evolution is odds. The statistics that any sort of anything, even a single molecule, could be created by accident, ex nihilo, is just preposterous to me. In addition to that, the sheer size and complexity of the universe exponentially decreases these odds, and I find it baffling anyone could believe something so unlikely, so impossible, could even happen. [ENDQ] [NEWLINE] The chances of reality being exactly the way it is, regardless of whether it was "random" or created, are very tiny. But if you roll a bazillion-sided die, you are guaranteed to get a result than only has a 1-in-bazillion chance of happening. So no matter how life or the universe is, the way it turned out is expected to be unlikely compared to all other (infinite?) options. [NEWLINE] [NEWLINE] But this is unrelated to whether or not it was created by a God. If things are unlikely to "randomly" be the way they are, they are equally as unlikely to have been created *exactly* this way by a God, who could have chosen infinite other possibilities. So does that mean it should be "baffling" that anyone could believe something so unlikely as a God who happens to have the exact characteristics to have created exactly this universe? Let's just agree that our existence is baffling regardless of evolution and/or God. [NEWLINE] [NEWLINE] As for your original question about teaching ID alongside evolution; I do think both should be taught but not *alongside* each other. Evolution should be taught in a biology class, and intelligent design should be taught in philosophy and/or religion classes. IMO we don't have enough variety of subjects throughout school.</s>
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0304, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0402, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0276, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0360, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0445, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0398, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0353, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0525, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0410, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0262, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0371, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0330, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0507, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0017, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0022, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0195, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0263, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0388, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>On my mobile,<mask> I'll be short,<mask> it would be ironic<mask> I wasn't. [NEWLINE] [NEWLINE] <mask> you can't make your point in a paragraph then you're not going to convince anyone. [NEWLINE] [NEWLINE] 1. Attention spans are too short. [NEWLINE] 2.<mask> you can't distill your point into a few well chosen sentences you probably don't have enough mastery over the subject to convince anyone. [NEWLINE] [NEWLINE] My point of view is more about rhetoric than being rigorously correct. I'm aware that there exists concepts that cannot be expressed in short form. I'm saying that communicating these concepts cannot reliably survive the medium of the Reddit post and still be effective at changing views. [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] EDIT_1: Thanks to everyone who has responded<mask> far.  I've had a few minor deltas and at this point now I've made the following modifications to my position. [NEWLINE] [NEWLINE] 1. The optimal length doesn't have to be a paragraph. <mask>, I contend that there is definite diminishing returns and negative returns associated with making points overly long. [NEWLINE] 2. For ideas complex enough to require long explanations, I still believe that the internet is a much less effective medium for this discussion than other options. <mask>, I will concede that it is not **ineffective**.  That was an overstatement. [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] EDIT_2: For clarity, I am intending on awarding deltas to a handful of users,<mask> not until my position is settled. [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] EDIT_3: I am going AFK for a few hours. <mask> I return I'll address any points I am able that I've missed and award deltas.</s>
Label encoding: <s>On my mobile, so I'll be short, also it would be ironic if I wasn't. [NEWLINE] [NEWLINE] If you can't make your point in a paragraph then you're not going to convince anyone. [NEWLINE] [NEWLINE] 1. Attention spans are too short. [NEWLINE] 2. If you can't distill your point into a few well chosen sentences you probably don't have enough mastery over the subject to convince anyone. [NEWLINE] [NEWLINE] My point of view is more about rhetoric than being rigorously correct. I'm aware that there exists concepts that cannot be expressed in short form. I'm saying that communicating these concepts cannot reliably survive the medium of the Reddit post and still be effective at changing views. [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] EDIT_1: Thanks to everyone who has responded thus far.  I've had a few minor deltas and at this point now I've made the following modifications to my position. [NEWLINE] [NEWLINE] 1. The optimal length doesn't have to be a paragraph.  However, I contend that there is definite diminishing returns and negative returns associated with making points overly long. [NEWLINE] 2. For ideas complex enough to require long explanations, I still believe that the internet is a much less effective medium for this discussion than other options.  However, I will concede that it is not **ineffective**.  That was an overstatement. [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] EDIT_2: For clarity, I am intending on awarding deltas to a handful of users, but not until my position is settled. [NEWLINE] [NEWLINE] ---- [NEWLINE] [NEWLINE] EDIT_3: I am going AFK for a few hours.  When I return I'll address any points I am able that I've missed and award deltas.</s>
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0543, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0326, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0274, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I lived in Provo, Utah for a<mask>. Mormon Capital of the world. Two of my coworkers were against gay marriage<mask> marriage is between a man and a woman. Not two men or two women. I had purposed to them that<mask> we changed the legal definition a<mask> back to something else and left each religion to decide<mask> marriage means to them,<mask> would they feel. They loved the idea<mask> they had a real problem with it. Now these two were your everyday a little religious mormons. Not crazy religious<mask> still very faithful.<mask> I'm betting that is probably a good portion of the mormon community(It's a very faithful and very religious religion from<mask> much I meet). [NEWLINE] [NEWLINE] Now people would call it marriage,<mask> that is society changing the definition compared to the government changing it which is the problem in my view. Or at least I can see the problem people are having. [NEWLINE] [NEWLINE] Religions are free to call their form of marriage whatever they want. That is fine. This is the government word marriage that is changing.<mask><mask> that people attribute the government and religious versions of the word marriage to be the similar and<mask> causing the issue I am talking about. [NEWLINE] [NEWLINE] Sources on you first bit about marriage<mask> a sacrament? Well the daily life stuff you talk about marriage especially tax returns, green cards and stuff like it are kind of recent(last 200 years) type of thing. They probably weren't there<mask> marriages first started happening. I have heard from another person that inheritance played a huge role in the creation of the idea of marriage and the church was used to keep records<mask> they were good at that. Still waiting on sources for that info.  </s>
Label encoding: <s>I lived in Provo, Utah for a while. Mormon Capital of the world. Two of my coworkers were against gay marriage but marriage is between a man and a woman. Not two men or two women. I had purposed to them that if we changed the legal definition a while back to something else and left each religion to decide what marriage means to them, how would they feel. They loved the idea because they had a real problem with it. Now these two were your everyday a little religious mormons. Not crazy religious but still very faithful. So I'm betting that is probably a good portion of the mormon community(It's a very faithful and very religious religion from how much I meet). [NEWLINE] [NEWLINE] Now people would call it marriage, but that is society changing the definition compared to the government changing it which is the problem in my view. Or at least I can see the problem people are having. [NEWLINE] [NEWLINE] Religions are free to call their form of marriage whatever they want. That is fine. This is the government word marriage that is changing. I think that people attribute the government and religious versions of the word marriage to be the similar and thus causing the issue I am talking about. [NEWLINE] [NEWLINE] Sources on you first bit about marriage as a sacrament? Well the daily life stuff you talk about marriage especially tax returns, green cards and stuff like it are kind of recent(last 200 years) type of thing. They probably weren't there when marriages first started happening. I have heard from another person that inheritance played a huge role in the creation of the idea of marriage and the church was used to keep records because they were good at that. Still waiting on sources for that info.  </s>
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>There are otherkin who are disruptive to the general public. I have a co-worker who identifies herself<mask> an otherkin/furry. Normally, I would not care about<mask> she does with her outside life, except it has begun to affect myself, my coworkers, and my customers. [NEWLINE] [NEWLINE] This girl, who is a "shapeshifting" otherkin, is one of my co-workers at a pet store. She believes that<mask> she has the soul of an animal that she is able to interact with any animal peacefully. [NEWLINE] [NEWLINE] She is the only employee who has been maliciously attacked multiple times by animals (both ours and customers). She even had to be sent to a hospital once<mask> she approached an aggressive pit-bull (even after being told it was aggressive and not to go near it *by the owners*) and the animal tore open her arm. [NEWLINE] [NEWLINE] And even after that, she is still under the impression that<mask> she is an animal, she did nothing wrong. It was the owners fault. She,<mask> an otherkin, could never have a bad experience with an animal. She claims she can communicate with them. [NEWLINE] [NEWLINE] Then she got bit by a snake. Then a bird. Then another dog. Hell, I'm sure the fish would attack her<mask> they could. And still, she did nothing wrong. [NEWLINE] [NEWLINE] I believe that she is being disruptive to both herself and to the general public. She is constantly attacked by animals and a few of our customers refuse to come into the store<mask> she is working. [NEWLINE] [NEWLINE] Of course, this is not true of all otherkin (<mask><mask> ). She is the only one I've met in real life.</s>
Label encoding: <s>There are otherkin who are disruptive to the general public. I have a co-worker who identifies herself as an otherkin/furry. Normally, I would not care about what she does with her outside life, except it has begun to affect myself, my coworkers, and my customers. [NEWLINE] [NEWLINE] This girl, who is a "shapeshifting" otherkin, is one of my co-workers at a pet store. She believes that because she has the soul of an animal that she is able to interact with any animal peacefully. [NEWLINE] [NEWLINE] She is the only employee who has been maliciously attacked multiple times by animals (both ours and customers). She even had to be sent to a hospital once because she approached an aggressive pit-bull (even after being told it was aggressive and not to go near it *by the owners*) and the animal tore open her arm. [NEWLINE] [NEWLINE] And even after that, she is still under the impression that because she is an animal, she did nothing wrong. It was the owners fault. She, as an otherkin, could never have a bad experience with an animal. She claims she can communicate with them. [NEWLINE] [NEWLINE] Then she got bit by a snake. Then a bird. Then another dog. Hell, I'm sure the fish would attack her if they could. And still, she did nothing wrong. [NEWLINE] [NEWLINE] I believe that she is being disruptive to both herself and to the general public. She is constantly attacked by animals and a few of our customers refuse to come into the store if she is working. [NEWLINE] [NEWLINE] Of course, this is not true of all otherkin ( I think ). She is the only one I've met in real life.</s>
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0255, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>youre missing two things: [NEWLINE] [NEWLINE] The first is that omniscience includes self knowledge; any omniscient being must have perfect knowledge of themself otherwise they have a hole in their knowledge and are not truly omniscient.<mask> they have perfect knowledge of themselves, they have perfect knowledge of<mask> they will act in every situation.<mask><mask> which situations occur, their choice is already determined by their knowledge of themselves. [NEWLINE] [NEWLINE] The second is that this proof doesnt require the existence of god, or,<mask>, any being that is omniscient or  omnipotent to exist. It merely requires that it is *possible* for such a being to exist.<mask> omniscience is possible, *free will fundamentally cannot be possible*. [NEWLINE] [NEWLINE] Even<mask> an omnipotent being operates outside our time, it cannot exist outside its *own* time; that doesnt really make sense<mask> a thing to say,<mask> even<mask> it did, then the portion of it outside its own time would<mask> be omniscient etc,<mask> we have a recursive decent into infinite levels of omniscient beings knowing<mask> it would act<mask> it were acting on any form of time. [NEWLINE] [NEWLINE] it<mask> works in reverse.<mask> true randomness exists, strict omnipotence and omniscience cannot:<mask> something is truly random it is unknowable before it is generated;<mask> it is unknowable, it falls outside of the knowledge of every being, and<mask> nothing can be omniscient <mask> the random results are holes in the knowledge.<mask> nothing is capable of being omniscient, nothing is capable of being omnipotent<mask> it cannot make itself omniscient. The terms are incompatible with each other.</s>
Label encoding: <s>youre missing two things: [NEWLINE] [NEWLINE] The first is that omniscience includes self knowledge; any omniscient being must have perfect knowledge of themself otherwise they have a hole in their knowledge and are not truly omniscient. If they have perfect knowledge of themselves, they have perfect knowledge of how they will act in every situation. Regardless of which situations occur, their choice is already determined by their knowledge of themselves. [NEWLINE] [NEWLINE] The second is that this proof doesnt require the existence of god, or, indeed, any being that is omniscient or  omnipotent to exist. It merely requires that it is *possible* for such a being to exist. If omniscience is possible, *free will fundamentally cannot be possible*. [NEWLINE] [NEWLINE] Even if an omnipotent being operates outside our time, it cannot exist outside its *own* time; that doesnt really make sense as a thing to say, but even if it did, then the portion of it outside its own time would also be omniscient etc, so we have a recursive decent into infinite levels of omniscient beings knowing how it would act if it were acting on any form of time. [NEWLINE] [NEWLINE] it also works in reverse. if true randomness exists, strict omnipotence and omniscience cannot: If something is truly random it is unknowable before it is generated; If it is unknowable, it falls outside of the knowledge of every being, and therefore nothing can be omniscient  because the random results are holes in the knowledge. If nothing is capable of being omniscient, nothing is capable of being omnipotent because it cannot make itself omniscient. The terms are incompatible with each other.</s>
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Generally<mask> you own a business you *can* refuse to do business with anyone. Lots of shops post signs "we reserve the right to refuse service", and they do. I'm pretty sure<mask> a print shop you can refuse to serve WBC, or your ex-girlfriend. [NEWLINE] [NEWLINE] <mask>, for certain classes of businesses, called common carriers, they don't have this right to refuse<mask> we decided the economy<mask> a whole depends on them serving everyone. Examples of this are hotels, transport companies, and restaurants,<mask> you can always ask people to leave for making a disturbance or bothering people (<mask> your ex comes in just to be rude, for example).<mask> places to eat and sleep can refuse people at will it supposedly hurts interstate commerce,<mask> it makes travel harder. Most businesses (print shops, tattoo artists, etc.) can turn down clients without having to explain themselves,<mask> the ones that can't are very visible. [NEWLINE] [NEWLINE] The other thing is there are certain protected classes<mask> we look at discrimination, such<mask> race, orientation, and gender,<mask> the court will be more actively suspicious<mask> you refuse to do business with lots of people who have the same (blank). I don't think political beliefs or your personal relationship with someone have special protection.<mask> I run a consultancy, I'm free to turn down all democrat/republican/green party/neo-nazi/ex-girlfriend clients<mask> I want to. [NEWLINE] [NEWLINE] TL;DR: Businesses have tremendous control over who they work with and who they refuse to serve,<mask> this control is limited for certain kinds of businesses called common carriers, and for certain protected classes of customers who have historically been discriminated against.</s>
Label encoding: <s>Generally if you own a business you *can* refuse to do business with anyone. Lots of shops post signs "we reserve the right to refuse service", and they do. I'm pretty sure as a print shop you can refuse to serve WBC, or your ex-girlfriend. [NEWLINE] [NEWLINE] BUT, for certain classes of businesses, called common carriers, they don't have this right to refuse because we decided the economy as a whole depends on them serving everyone. Examples of this are hotels, transport companies, and restaurants, although you can always ask people to leave for making a disturbance or bothering people ( if your ex comes in just to be rude, for example). If places to eat and sleep can refuse people at will it supposedly hurts interstate commerce, because it makes travel harder. Most businesses (print shops, tattoo artists, etc.) can turn down clients without having to explain themselves, but the ones that can't are very visible. [NEWLINE] [NEWLINE] The other thing is there are certain protected classes when we look at discrimination, such as race, orientation, and gender, where the court will be more actively suspicious if you refuse to do business with lots of people who have the same (blank). I don't think political beliefs or your personal relationship with someone have special protection. If I run a consultancy, I'm free to turn down all democrat/republican/green party/neo-nazi/ex-girlfriend clients if I want to. [NEWLINE] [NEWLINE] TL;DR: Businesses have tremendous control over who they work with and who they refuse to serve, but this control is limited for certain kinds of businesses called common carriers, and for certain protected classes of customers who have historically been discriminated against.</s>
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0383, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0324, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0286, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] <mask> you haven't encountered the kind of people that<mask><mask> OP is talking about, then you are...privileged [ENDQ] [NEWLINE] Add it to the list :) I mean, to some extent you could say I was "playing dumb" in that regard. I can *guess* about<mask> sorts of conversations OP has had to lead him to his current views,<mask> I'd rather try to get it from him than to put make assumptions / put words in his mouth.<mask> without having any information, its easy to guess several possible explanations for the perception that the people are "throwing around privilege"<mask> some kind of callous insult. [NEWLINE] [NEWLINE] 1. They're reasonable people,<mask> aren't good at articulating themselves.<mask><mask> is meant to be a reasonable argument comes off overly hostile or lacking nuance. This is they're fault,<mask> I do suspect that it's unfair to call them terrible people or parasites on account of this failing. [NEWLINE] 2. They're smart people making smart arguments,<mask> for whatever reason you or the OP misunderstood them, heard it out of context, or whatever. I wouldn't jump to this conclusion,<mask> it could be an unwarranted insult to you or the OP without enough information,<mask> its a possibility. [NEWLINE] 3. They're dumb / crazy. I don't like to jump to this either,<mask> there are all sorts of folks on the internet. [NEWLINE] 4. Something else. [NEWLINE] [NEWLINE] <mask> the point is,<mask><mask> broad sweeping statements like the ones OP is making are a mistake. I suspect that<mask> we could really drill deeper into OP's interactions, there would be a more reasonable explanation,<mask> I'm reluctant to make too many assumptions.</s>
Label encoding: <s> [STARTQ] if you haven't encountered the kind of people that I think OP is talking about, then you are...privileged [ENDQ] [NEWLINE] Add it to the list :) I mean, to some extent you could say I was "playing dumb" in that regard. I can *guess* about what sorts of conversations OP has had to lead him to his current views, but I'd rather try to get it from him than to put make assumptions / put words in his mouth. But without having any information, its easy to guess several possible explanations for the perception that the people are "throwing around privilege" as some kind of callous insult. [NEWLINE] [NEWLINE] 1. They're reasonable people, but aren't good at articulating themselves. So what is meant to be a reasonable argument comes off overly hostile or lacking nuance. This is they're fault, but I do suspect that it's unfair to call them terrible people or parasites on account of this failing. [NEWLINE] 2. They're smart people making smart arguments, but for whatever reason you or the OP misunderstood them, heard it out of context, or whatever. I wouldn't jump to this conclusion, as it could be an unwarranted insult to you or the OP without enough information, but its a possibility. [NEWLINE] 3. They're dumb / crazy. I don't like to jump to this either, but there are all sorts of folks on the internet. [NEWLINE] 4. Something else. [NEWLINE] [NEWLINE] But the point is, I think broad sweeping statements like the ones OP is making are a mistake. I suspect that if we could really drill deeper into OP's interactions, there would be a more reasonable explanation, but I'm reluctant to make too many assumptions.</s>
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0288, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0011, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Our society is geared towards white men. Part of that gearing is that their ideas and opinions are celebrated and encouraged, and<mask> a consequence others might feel uncomfortable sharing their ideas and opinions in spaces dominated by straight white men. [NEWLINE] [NEWLINE] You can observe this phenomenon yourself! Are you in school, university, or employed with a wide variety of people? Observe<mask> often women and people of color speak up - particularly in spaces<mask> they are a) new and/or b) in a numerical minority. I'm willing to bet it's significantly less than the white dudes speak up. [NEWLINE] [NEWLINE] The other consequence of this, of course, is that<mask> there's a straight white male opinion on an issue you better fucking believe *everyone* has heard it,<mask> the straight<mask> male opinions are the mainstream opinions. They're the op eds in your newspapers, they're the SCOTUS decisions, they're the authors of your magazine articles, etc. Sometimes it's nice to have a space<mask> this particular viewpoint isn't belabored past the point<mask> it's useful. [NEWLINE] [NEWLINE] Finally,'safe spaces' can<mask> foster a discussion between a group of people who are all on the same page. No movement would ever get anywhere<mask> it had the first hour of any given meeting trying to explain its core tenets to newcomers.<mask> I want to hash out the specifics of intersectional feminism, I am most likely to accomplish that goal with a group of intersectional feminists. I *ought*,<mask> I'm a good intersectional feminist, to have a good logical justification for intersectional feminism in general,<mask> sometimes I don't want to have to *give* that justification every time I talk about it.</s><pad>
Label encoding: <s>Our society is geared towards white men. Part of that gearing is that their ideas and opinions are celebrated and encouraged, and as a consequence others might feel uncomfortable sharing their ideas and opinions in spaces dominated by straight white men. [NEWLINE] [NEWLINE] You can observe this phenomenon yourself! Are you in school, university, or employed with a wide variety of people? Observe how often women and people of color speak up - particularly in spaces where they are a) new and/or b) in a numerical minority. I'm willing to bet it's significantly less than the white dudes speak up. [NEWLINE] [NEWLINE] The other consequence of this, of course, is that if there's a straight white male opinion on an issue you better fucking believe *everyone* has heard it, because the straight while male opinions are the mainstream opinions. They're the op eds in your newspapers, they're the SCOTUS decisions, they're the authors of your magazine articles, etc. Sometimes it's nice to have a space where this particular viewpoint isn't belabored past the point where it's useful. [NEWLINE] [NEWLINE] Finally,'safe spaces' can also foster a discussion between a group of people who are all on the same page. No movement would ever get anywhere if it had the first hour of any given meeting trying to explain its core tenets to newcomers. If I want to hash out the specifics of intersectional feminism, I am most likely to accomplish that goal with a group of intersectional feminists. I *ought*, if I'm a good intersectional feminist, to have a good logical justification for intersectional feminism in general, but sometimes I don't want to have to *give* that justification every time I talk about it.</s><pad>
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0332, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0016, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm sure the vast majority of gay people would prefer not to be around homophobic people - who wants to be around people who dislike you? [NEWLINE] [NEWLINE] <mask> I specifically referred to people who keep their beliefs to themselves, and I'm<mask> certain that there are plenty of homophobic people who never express those opinions to gay people or treat them differently. Anecdotal evidence isn't evidence of a general rule,<mask> I have known gay people and 'closeted' homophobes. Some of them were actually good friends with each other<mask> well,<mask> it does happen, to at least some degree. [NEWLINE] [NEWLINE] Let's take a step back and think about<mask> we're talking about here. It's one thing to hold an opposing opinion, vote a certain way, or even have respectful discussions about issues like this with friends, family or acquaintances.<mask><mask><mask> there's a perception that someone's beliefs make them a better or worse person depending on<mask> they are; I'm merely arguing that actions and overall character ought to be at least equally important. In my eyes, they're more important. I'm not giving a free pass to bigots or condemning people for standing firm by their beliefs.<mask> part of the reason<mask> our national dialogue is<mask> toxic is that we're willing to pass judgment on others not on<mask> they behave or treat people<mask> based on beliefs that we have contempt for.<mask> yeah - a nice person is a nice person<mask><mask><mask> their contemptible beliefs don't become contemptible actions and a self-righteous jackass is a jackass even<mask> we happen to agree with them. We need to be critical of everyone's conduct, especially<mask> we agree with them.</s>
Label encoding: <s>I'm sure the vast majority of gay people would prefer not to be around homophobic people - who wants to be around people who dislike you? [NEWLINE] [NEWLINE] But I specifically referred to people who keep their beliefs to themselves, and I'm also certain that there are plenty of homophobic people who never express those opinions to gay people or treat them differently. Anecdotal evidence isn't evidence of a general rule, but I have known gay people and 'closeted' homophobes. Some of them were actually good friends with each other as well, so it does happen, to at least some degree. [NEWLINE] [NEWLINE] Let's take a step back and think about what we're talking about here. It's one thing to hold an opposing opinion, vote a certain way, or even have respectful discussions about issues like this with friends, family or acquaintances. But I think there's a perception that someone's beliefs make them a better or worse person depending on what they are; I'm merely arguing that actions and overall character ought to be at least equally important. In my eyes, they're more important. I'm not giving a free pass to bigots or condemning people for standing firm by their beliefs. But part of the reason why our national dialogue is so toxic is that we're willing to pass judgment on others not on how they behave or treat people but based on beliefs that we have contempt for. So yeah - a nice person is a nice person so long as their contemptible beliefs don't become contemptible actions and a self-righteous jackass is a jackass even if we happen to agree with them. We need to be critical of everyone's conduct, especially when we agree with them.</s>
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0015, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0404, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Well first, your OP has a nasty case of cherry picking. Just<mask> a guy who wrote a book called "Dysgenics" thinks that IQ is highly heritable and the Flynn effect can be attributed to nutrition does not make it scientific consensus.<mask>, from the wikipedia article you linked to "prove" that low IQ individuals have more children: *<mask> many demographic studies have been performed, there is no conclusive evidence of a positive or negative correlation between human intelligence and fertility rate.* [NEWLINE] [NEWLINE] With that said, let's go ahead and say that nature vs. nurture has been settled, and it turns out nature wins. Let's say it is possible to use eugenics to significantly increase the IQ and empathy of the population.<mask> would the downsides be? [NEWLINE] [NEWLINE] 1. People with high IQ are more likely to abuse drugs and suffer from mental illness [\[source\]]( [URL] ) [NEWLINE] [NEWLINE] 2. Empathy can cloud decision-making. For example your eugenics plan is unlikely to be supported by a person with high empathy for the people affected by it. Funnily enough, it seems like using eugenics to increase empathy would be like one of those machines that exists only to turn itself off. [NEWLINE] [NEWLINE] 3. (Your claim, not mine.) People with high IQ are less likely to have children.<mask><mask> you say about overpopulation, birth rates in the first world are declining, and actually birth rates in the US are currently too low to sustain our current population [\[source\]]( [URL] ). Making more intelligent people will lower the birth rate even further, potentially driving our population dangerously low and even leading to extinction. </s>
Label encoding: <s>Well first, your OP has a nasty case of cherry picking. Just because a guy who wrote a book called "Dysgenics" thinks that IQ is highly heritable and the Flynn effect can be attributed to nutrition does not make it scientific consensus. Also, from the wikipedia article you linked to "prove" that low IQ individuals have more children: * Though many demographic studies have been performed, there is no conclusive evidence of a positive or negative correlation between human intelligence and fertility rate.* [NEWLINE] [NEWLINE] With that said, let's go ahead and say that nature vs. nurture has been settled, and it turns out nature wins. Let's say it is possible to use eugenics to significantly increase the IQ and empathy of the population. What would the downsides be? [NEWLINE] [NEWLINE] 1. People with high IQ are more likely to abuse drugs and suffer from mental illness [\[source\]]( [URL] ) [NEWLINE] [NEWLINE] 2. Empathy can cloud decision-making. For example your eugenics plan is unlikely to be supported by a person with high empathy for the people affected by it. Funnily enough, it seems like using eugenics to increase empathy would be like one of those machines that exists only to turn itself off. [NEWLINE] [NEWLINE] 3. (Your claim, not mine.) People with high IQ are less likely to have children. Despite what you say about overpopulation, birth rates in the first world are declining, and actually birth rates in the US are currently too low to sustain our current population [\[source\]]( [URL] ). Making more intelligent people will lower the birth rate even further, potentially driving our population dangerously low and even leading to extinction. </s>
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>They're all. Directly or indirectly. Knowingly or not. Trough violence, intimidation, shaming, peer pressure or education. [NEWLINE] [NEWLINE] This is not normal to have one half of your population walking hidden under a sheet, no matter<mask> you turn it. [NEWLINE] [NEWLINE] Women here struggled for their rights and for equity, we aren't going to let some dark age zealots erode this building. [NEWLINE] [NEWLINE] Muslim women with full veil are very few actually compared to the muslim population, even the lightest veils are scarce in this regard. [NEWLINE] [NEWLINE] Americans are too soft on religions, still way too much under its influence, just see all the creationism bullshit,<mask> they try and invade schools (our children's minds) with their lies and deceptions. See<mask> your presidents must swear on the bible (wtf is that a joke?) [NEWLINE] [NEWLINE] Religions are not "peaceful gatherings of simple folks", religions are organized, paralell political powers that need to be kept in close check, they're utterly anti democratic. We crushed them during the revolution for a reason (the clergies are, historically -check it-  ALWAYS the current tyrant's best friend). [NEWLINE] [NEWLINE] Yes, we aren't soft with religions. And that's a thing future humanity will uphold<mask> a human progress. [NEWLINE] [NEWLINE] Now,<mask> france was a catholic country for<mask> long, the catholics are still predominant<mask> they have some more latitudes,<mask> not many and they're losing them. And that's fortunate. [NEWLINE] [NEWLINE] tl;dr: religions aren't your friends, they're parallel powers invariably trying to take over<mask> they're designed around this objective. </s><pad>
Label encoding: <s>They're all. Directly or indirectly. Knowingly or not. Trough violence, intimidation, shaming, peer pressure or education. [NEWLINE] [NEWLINE] This is not normal to have one half of your population walking hidden under a sheet, no matter how you turn it. [NEWLINE] [NEWLINE] Women here struggled for their rights and for equity, we aren't going to let some dark age zealots erode this building. [NEWLINE] [NEWLINE] Muslim women with full veil are very few actually compared to the muslim population, even the lightest veils are scarce in this regard. [NEWLINE] [NEWLINE] Americans are too soft on religions, still way too much under its influence, just see all the creationism bullshit, how they try and invade schools (our children's minds) with their lies and deceptions. See how your presidents must swear on the bible (wtf is that a joke?) [NEWLINE] [NEWLINE] Religions are not "peaceful gatherings of simple folks", religions are organized, paralell political powers that need to be kept in close check, they're utterly anti democratic. We crushed them during the revolution for a reason (the clergies are, historically -check it-  ALWAYS the current tyrant's best friend). [NEWLINE] [NEWLINE] Yes, we aren't soft with religions. And that's a thing future humanity will uphold as a human progress. [NEWLINE] [NEWLINE] Now, because france was a catholic country for so long, the catholics are still predominant so they have some more latitudes, but not many and they're losing them. And that's fortunate. [NEWLINE] [NEWLINE] tl;dr: religions aren't your friends, they're parallel powers invariably trying to take over because they're designed around this objective. </s><pad>
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0165, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0217, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0271, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0307, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0260, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0048, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>That sounds to me like an agnostic position, rather than an atheistic one.  There's nothing wrong with being agnostic.  There's a lot of gradations here,<mask> it is possible to be an agnostic atheist, too. <mask><mask><mask><mask> the answer of whether a god exists or not is a provable thing.  I tend to fall on the side that the notion of a god doesn't exist,<mask> for me, it's more than that. <mask><mask> to doubting the existence of a superbeing that created the universe, I don't buy into the concept of deity worship. [NEWLINE] [NEWLINE] Let's say a superbeing were to appear tomorrow and state that it was responsible for the creation of the universe and demand worship.  First off,<mask> would you know that the religion you had chosen was the correct one?  Second,<mask> would a superbeing that was both all powerful and all seeing need to create lesser creatures simply for the purpose of being worshiped?  Would that being not already be aware of all of its creations thoughts and feelings? <mask> is worship, and<mask> does that make sense<mask> the 'end product' of creation.  And<mask> on. [NEWLINE] [NEWLINE] <mask> the answer of whether a being created the universe or not is unknowable with our current level of knowledge.  You can either choose to use logic, reason, and empiricism to determine your direction, or be convinced by some other person who tells you<mask> they believe to be the case.  It's up to you. <mask><mask> you've got something inside you telling you that the religious answer isn't real, you may just find you're right.</s>
Label encoding: <s>That sounds to me like an agnostic position, rather than an atheistic one.  There's nothing wrong with being agnostic.  There's a lot of gradations here, but it is possible to be an agnostic atheist, too.  I do not think the answer of whether a god exists or not is a provable thing.  I tend to fall on the side that the notion of a god doesn't exist, but for me, it's more than that.  In addition to doubting the existence of a superbeing that created the universe, I don't buy into the concept of deity worship. [NEWLINE] [NEWLINE] Let's say a superbeing were to appear tomorrow and state that it was responsible for the creation of the universe and demand worship.  First off, how would you know that the religion you had chosen was the correct one?  Second, why would a superbeing that was both all powerful and all seeing need to create lesser creatures simply for the purpose of being worshiped?  Would that being not already be aware of all of its creations thoughts and feelings?  What is worship, and why does that make sense as the 'end product' of creation.  And so on. [NEWLINE] [NEWLINE] But the answer of whether a being created the universe or not is unknowable with our current level of knowledge.  You can either choose to use logic, reason, and empiricism to determine your direction, or be convinced by some other person who tells you what they believe to be the case.  It's up to you.  But if you've got something inside you telling you that the religious answer isn't real, you may just find you're right.</s>
Loss: tensor(0.0299, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0019, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0292, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0336, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0174, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0047, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0176, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I'm lead to believe that the inhaler was confiscated<mask> the protocol was not observed. [ENDQ] [STARTQ] <mask>? [ENDQ] [NEWLINE] The GAWKER article you posted said this, [NEWLINE] [NEWLINE] "Gibbons would regularly receive a call from the school asking her to pick up Ryan's inhaler after he was "caught" carrying it around in his backpack." [NEWLINE] [NEWLINE] The mother neglected to pick the inhaler up after they did not follow the proper procedure. The article even says the mother received regular calls. This is neglectful. [NEWLINE] [NEWLINE] Again, I don't know<mask> it works in Canada<mask> I suspect it is the same<mask> the United States. I'm not saying inhalers shouldn't be allowed to remain in the students possession throughout the day. I am saying that there was some (a small amount) neglect on the part of the mother and that this is not a valid example of<mask> public schools do not meet the health needs of students. [NEWLINE] [NEWLINE] I thought the second article you posted was the CBC news one and didn't see the ABC one before. That is a much better quality article. The thing that I noticed was that it said the majority of states have passed laws that say a child must be allowed to have their inhaler at all times. This is an example of the public school system working. Sure it may work slowly at times<mask> nonetheless it works. The article<mask> mentions that the parents can create an individualized Health Plan<mask> they do not like the school's policy in places<mask> students aren't allowed their inhalers. The mother in the example you gave chose to break school rules rather than follow the proper steps and her son is dead<mask> of that.</s>
Label encoding: <s> [STARTQ] I'm lead to believe that the inhaler was confiscated because the protocol was not observed. [ENDQ] [STARTQ] Why? [ENDQ] [NEWLINE] The GAWKER article you posted said this, [NEWLINE] [NEWLINE] "Gibbons would regularly receive a call from the school asking her to pick up Ryan's inhaler after he was "caught" carrying it around in his backpack." [NEWLINE] [NEWLINE] The mother neglected to pick the inhaler up after they did not follow the proper procedure. The article even says the mother received regular calls. This is neglectful. [NEWLINE] [NEWLINE] Again, I don't know how it works in Canada but I suspect it is the same as the United States. I'm not saying inhalers shouldn't be allowed to remain in the students possession throughout the day. I am saying that there was some (a small amount) neglect on the part of the mother and that this is not a valid example of how public schools do not meet the health needs of students. [NEWLINE] [NEWLINE] I thought the second article you posted was the CBC news one and didn't see the ABC one before. That is a much better quality article. The thing that I noticed was that it said the majority of states have passed laws that say a child must be allowed to have their inhaler at all times. This is an example of the public school system working. Sure it may work slowly at times but nonetheless it works. The article also mentions that the parents can create an individualized Health Plan if they do not like the school's policy in places where students aren't allowed their inhalers. The mother in the example you gave chose to break school rules rather than follow the proper steps and her son is dead because of that.</s>
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0376, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0302, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0520, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0070, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Overpopulation is fixing itself. Hunger is now a political disaster, occurring only<mask> food shipments are actively disrupted. War is less common now and claims a smaller percentage of the population<mask> causalities than at any point in human history. I would<mask><mask> the problems you are citing aren't the problems our children will have to struggle with. [NEWLINE] [NEWLINE] I would<mask><mask> any effort to incentivize desirable traits is doomed to failure. First, we can't accurately measure our positive qualities. It's all we have is a horrible argument<mask> you're talking about wasting billions or trillions of dollars and stigmatizing people.<mask> you aren't actually incentivizing intelligence, then you aren't going to have a more intelligent population in the end.<mask> your test allows for people to eschew actual intelligence for some sort of "teaching to the test" then the latter will incentivized more than the former. [NEWLINE] [NEWLINE] <mask>, **there is no mechanical difference between a tax break for the smart and a tax on the dumb**. Things are measured on a relative scale, not an objective one. Problems are caused by the fact that there is a difference, not the details of<mask> the difference is created. [NEWLINE] [NEWLINE] I would<mask><mask><mask> this would decrease our ability to combat complex challenges. After all, people who are encouraged with "you are smart"<mask> opposed to "you did well" don't accomplish<mask> much,<mask> success due to an inborn trait doesn't have a pay off and failure holds a big risk. [Studies prove this]( [URL] /). We need people who work hard, even<mask> they aren't the smartest, more than we need people who are smart to begin with.</s>
Label encoding: <s>Overpopulation is fixing itself. Hunger is now a political disaster, occurring only where food shipments are actively disrupted. War is less common now and claims a smaller percentage of the population as causalities than at any point in human history. I would argue that the problems you are citing aren't the problems our children will have to struggle with. [NEWLINE] [NEWLINE] I would argue that any effort to incentivize desirable traits is doomed to failure. First, we can't accurately measure our positive qualities. It's all we have is a horrible argument when you're talking about wasting billions or trillions of dollars and stigmatizing people. If you aren't actually incentivizing intelligence, then you aren't going to have a more intelligent population in the end. If your test allows for people to eschew actual intelligence for some sort of "teaching to the test" then the latter will incentivized more than the former. [NEWLINE] [NEWLINE] Also, **there is no mechanical difference between a tax break for the smart and a tax on the dumb**. Things are measured on a relative scale, not an objective one. Problems are caused by the fact that there is a difference, not the details of how the difference is created. [NEWLINE] [NEWLINE] I would also argue that this would decrease our ability to combat complex challenges. After all, people who are encouraged with "you are smart" as opposed to "you did well" don't accomplish as much, because success due to an inborn trait doesn't have a pay off and failure holds a big risk. [Studies prove this]( [URL] /). We need people who work hard, even if they aren't the smartest, more than we need people who are smart to begin with.</s>
Loss: tensor(0.0463, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0289, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0265, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0421, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0341, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0012, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Okay, from a semiotics/representation perspective: We agree that black can be used to easily mark evil characters in a movie.<mask>,<mask> you place the symbol of evil onto the canvas of the human body, you marry the two concepts: dark skin becomes the symbol of evil. I do not<mask><mask> this is intentional racism, I<mask><mask> it is a lazy, harmful choice (with possibly well-intentioned underlying design decisions) that teaches children to view race in an unexamined, stereotypical way. [NEWLINE] [NEWLINE] [STARTQ] They all have dark colors and a dark vibe,<mask> not really black features or any other racial indicator. [ENDQ] [NEWLINE] It's true, Scar doesn't fit this example.<mask>, the Hyenas are voiced by people of color, and<mask> do have racially defining characteristics.<mask>, Jafar is definitely the darkest character in the movie and has racially defining characteristics. That said, you do raise a good point that each character needs to be thought of independently. [NEWLINE] [NEWLINE] --- [NEWLINE] P.S. I studied Journalism (lots of media theory), International Relations (lots of racial/political theory) and Digital Media (again, lots of media theory),<mask> I come from a pretty biased background on this topic. More than once, I have been exposed to arguments that some Disney movies are racist (Aladdin is really the undisputed one in my mind).<mask>, I am a web designer and I have studied color theory a number of times. Honestly, I usually get frustrated with the endless lists of<mask> color evokes<mask> sense... we all know<mask> the colors mean in our culture, or we are not being very aware. [NEWLINE] </s>
Label encoding: <s>Okay, from a semiotics/representation perspective: We agree that black can be used to easily mark evil characters in a movie. However, when you place the symbol of evil onto the canvas of the human body, you marry the two concepts: dark skin becomes the symbol of evil. I do not argue that this is intentional racism, I argue that it is a lazy, harmful choice (with possibly well-intentioned underlying design decisions) that teaches children to view race in an unexamined, stereotypical way. [NEWLINE] [NEWLINE] [STARTQ] They all have dark colors and a dark vibe, but not really black features or any other racial indicator. [ENDQ] [NEWLINE] It's true, Scar doesn't fit this example. However, the Hyenas are voiced by people of color, and so do have racially defining characteristics. Also, Jafar is definitely the darkest character in the movie and has racially defining characteristics. That said, you do raise a good point that each character needs to be thought of independently. [NEWLINE] [NEWLINE] --- [NEWLINE] P.S. I studied Journalism (lots of media theory), International Relations (lots of racial/political theory) and Digital Media (again, lots of media theory), so I come from a pretty biased background on this topic. More than once, I have been exposed to arguments that some Disney movies are racist (Aladdin is really the undisputed one in my mind). However, I am a web designer and I have studied color theory a number of times. Honestly, I usually get frustrated with the endless lists of what color evokes what sense... we all know what the colors mean in our culture, or we are not being very aware. [NEWLINE] </s>
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0316, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0257, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0275, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0311, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, there was a famous philosophical argument written along similar lines. [NEWLINE] [NEWLINE] [URL] #The_Violinist [NEWLINE] [NEWLINE] [STARTQ] In A Defense of Abortion, Thomson grants for the sake of argument that the fetus has a right to life,<mask> defends the permissibility of abortion by appeal to a thought experiment: [ENDQ] [NEWLINE] [STARTQ] You wake up in the morning and find yourself back to back in bed with an unconscious violinist. A famous unconscious violinist. He has been found to have a fatal kidney ailment, and the Society of Music Lovers has canvassed all the available medical records and found that you alone have the right blood type to help. They have<mask> kidnapped you, and last night the violinist's circulatory system was plugged into yours,<mask> that your kidneys can be used to extract poisons from his blood<mask> well<mask> your own. [<mask> he is unplugged from you now, he will die;<mask> ] in nine months he will have recovered from his ailment, and can safely be unplugged from you.[4] [ENDQ] [NEWLINE] &gt;Thomson takes it that you may now permissibly unplug yourself from the violinist<mask><mask> this will cause his death: the right to life, Thomson says, does not entail the right to use another person's body, and<mask> by unplugging the violinist you do not violate his right to life<mask> merely deprive him of something—the use of your body—to which he has no right. "[I]f you do allow him to go on using your kidneys, this is a kindness on your part, and not something he can claim from you<mask> his due."[5]</s>
Label encoding: <s>Yeah, there was a famous philosophical argument written along similar lines. [NEWLINE] [NEWLINE] [URL] #The_Violinist [NEWLINE] [NEWLINE] [STARTQ] In A Defense of Abortion, Thomson grants for the sake of argument that the fetus has a right to life, but defends the permissibility of abortion by appeal to a thought experiment: [ENDQ] [NEWLINE] [STARTQ] You wake up in the morning and find yourself back to back in bed with an unconscious violinist. A famous unconscious violinist. He has been found to have a fatal kidney ailment, and the Society of Music Lovers has canvassed all the available medical records and found that you alone have the right blood type to help. They have therefore kidnapped you, and last night the violinist's circulatory system was plugged into yours, so that your kidneys can be used to extract poisons from his blood as well as your own. [ If he is unplugged from you now, he will die; but ] in nine months he will have recovered from his ailment, and can safely be unplugged from you.[4] [ENDQ] [NEWLINE] &gt;Thomson takes it that you may now permissibly unplug yourself from the violinist even though this will cause his death: the right to life, Thomson says, does not entail the right to use another person's body, and so by unplugging the violinist you do not violate his right to life but merely deprive him of something—the use of your body—to which he has no right. "[I]f you do allow him to go on using your kidneys, this is a kindness on your part, and not something he can claim from you as his due."[5]</s>
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0552, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0438, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0056, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] Credit to the poor.<mask> there is any question that lending to a person might be a problem that loan will not be made. Credit cards for the poor and lower middle income people would immediately be revoked<mask> suddenly we can't count on bankruptcy and asset liquidation<mask> backstopping their debts. [ENDQ] [NEWLINE] Do you have evidence that without bankruptcy and asset liquidation banks would no longer be profitable? Perhaps a breakdown of their profits on various customers? [NEWLINE] [NEWLINE] [STARTQ] Instead of a bank lending you the money to buy a house (a mortgage) the bank would buy the house and rent it to you with a promise to sell it to you under certain terms. [ENDQ] [NEWLINE] <mask> banks make new, predatory loans, I am fine with the government stepping in. [NEWLINE] [NEWLINE] [STARTQ] Loan sharks. You know who doesn't need collection agencies or to worry about the niceties of laws? criminals. Their security is breaking your legs<mask> you don't pay them.<mask> people need money and can't borrow it they turn to loan sharks. [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Feels to me like this is the current situation. I'd really prefer government loans to the poor. Wiping out the (legal) debts of these loan sharks is part of<mask> I support this. [NEWLINE] [NEWLINE] [STARTQ] Punishes savers. [ENDQ] [NEWLINE] <mask> I noted, my priority isn't protecting the rich who have large savings.<mask> my family wouldn't be affected, my parents and me only make ethical investments into companies which avoid success through slave labour, debt imprisonment and things like that.<mask> I view the things debters do with minimal love I don't have much sympathy for those who invested their savings into debt.</s>
Label encoding: <s> [STARTQ] Credit to the poor. If there is any question that lending to a person might be a problem that loan will not be made. Credit cards for the poor and lower middle income people would immediately be revoked because suddenly we can't count on bankruptcy and asset liquidation as backstopping their debts. [ENDQ] [NEWLINE] Do you have evidence that without bankruptcy and asset liquidation banks would no longer be profitable? Perhaps a breakdown of their profits on various customers? [NEWLINE] [NEWLINE] [STARTQ] Instead of a bank lending you the money to buy a house (a mortgage) the bank would buy the house and rent it to you with a promise to sell it to you under certain terms. [ENDQ] [NEWLINE] If banks make new, predatory loans, I am fine with the government stepping in. [NEWLINE] [NEWLINE] [STARTQ] Loan sharks. You know who doesn't need collection agencies or to worry about the niceties of laws? criminals. Their security is breaking your legs if you don't pay them. When people need money and can't borrow it they turn to loan sharks. [ENDQ] [NEWLINE] [URL] [NEWLINE] [NEWLINE] Feels to me like this is the current situation. I'd really prefer government loans to the poor. Wiping out the (legal) debts of these loan sharks is part of why I support this. [NEWLINE] [NEWLINE] [STARTQ] Punishes savers. [ENDQ] [NEWLINE] As I noted, my priority isn't protecting the rich who have large savings. Also my family wouldn't be affected, my parents and me only make ethical investments into companies which avoid success through slave labour, debt imprisonment and things like that. Since I view the things debters do with minimal love I don't have much sympathy for those who invested their savings into debt.</s>
Loss: tensor(0.0220, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0089, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0405, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0213, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>You've already changed your mind on them being "evil",<mask> I want to push it a bit further. [NEWLINE] [NEWLINE] [STARTQ] I've seen a lot of people want to abort their kids<mask> they'd be born with Down Syndrome or autism, or similar things. [ENDQ] [NEWLINE] <mask> they're aborted they're not *kids*. They're *fetuses*, which I consider of no intrinsic value. You are already a person, which arguably has intrinsic value. [NEWLINE] [NEWLINE] In a sense, *you weren't born*.<mask> was born was not you, even<mask> it *became you*, and we consider it "you" in common speech.<mask> your desires, your personality, your goals, your interests, your memories… those weren't there at first. [NEWLINE] [NEWLINE] Personally I don't think "they'll have a hard life" is a good argument. Life is sometimes hard, that's<mask> it's<mask> great. I've seen stories of people with harlequin ichthyiosis who have made it to adolescence or adulthood, and far from desiring nonexistence they really love life. And even<mask> people with such fucked up diseases can be happy, then people with worse ones can. [NEWLINE] [NEWLINE] <mask>, I am strongly pro-choice, on the basis of bodily autonomy (you can't force continued organ donation) paired with the lack of intrinsic value of the fetus.<mask> a woman has an abortion, she is (generally) wronging no one,<mask> there is no "one" to be wronged in the first place.<mask> she doesn't want to take care of a kid with some disorder, *she doesn't even need to give any explanations*.</s>
Label encoding: <s>You've already changed your mind on them being "evil", but I want to push it a bit further. [NEWLINE] [NEWLINE] [STARTQ] I've seen a lot of people want to abort their kids because they'd be born with Down Syndrome or autism, or similar things. [ENDQ] [NEWLINE] If they're aborted they're not *kids*. They're *fetuses*, which I consider of no intrinsic value. You are already a person, which arguably has intrinsic value. [NEWLINE] [NEWLINE] In a sense, *you weren't born*. What was born was not you, even if it *became you*, and we consider it "you" in common speech. But your desires, your personality, your goals, your interests, your memories… those weren't there at first. [NEWLINE] [NEWLINE] Personally I don't think "they'll have a hard life" is a good argument. Life is sometimes hard, that's why it's also great. I've seen stories of people with harlequin ichthyiosis who have made it to adolescence or adulthood, and far from desiring nonexistence they really love life. And even if people with such fucked up diseases can be happy, then people with worse ones can. [NEWLINE] [NEWLINE] However, I am strongly pro-choice, on the basis of bodily autonomy (you can't force continued organ donation) paired with the lack of intrinsic value of the fetus. If a woman has an abortion, she is (generally) wronging no one, because there is no "one" to be wronged in the first place. If she doesn't want to take care of a kid with some disorder, *she doesn't even need to give any explanations*.</s>
Loss: tensor(0.0118, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0099, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0344, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0422, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0029, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0009, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0285, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0269, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0277, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Yeah, I consider myself an egalitarian and I *really* despise this idea that feminists (or any other group) have a monopoly on equality. [NEWLINE] [NEWLINE] The label "feminism" comes with a huge load of historic and linguistic baggage (<mask> OP pointed out) which I don't have the desire *or* energy to combat. It comes with fringe groups which make me roll my eyes every time they speak - and that's the *best*-case scenario. More to the point than that,<mask>, it comes with a number of ideological elements with which I either emphatically disagree, or which I find to be too lacking in nuance to be useful. [NEWLINE] [NEWLINE] The term is now too vague to be in any way descriptive whatsoever, and the movement itself is too diffuse to even *be* a single movement. To say that you're a feminist is *in no way indicative* of your position on issues like parental leave, equality of opportunity vs. equality of outcomes, misandry, transgender issues, etc. People who self-identify<mask> "feminists" run the gamut on these issues -<mask> they do on just about *every* issue I've ever heard discussed in a feminist context. [NEWLINE] [NEWLINE] I personally see it<mask> a useless label - worse than useless,<mask> it's an *inflammatory* label which draws attention away from core issues and toward the "Who Is A Real Feminist" side-track. [NEWLINE] [NEWLINE] And I really, *really* resent people who tell me that *they* get to label me or my ideology. It's<mask><mask> it's impossible for people with somewhat different labels to work together for a common goal.</s>
Label encoding: <s>Yeah, I consider myself an egalitarian and I *really* despise this idea that feminists (or any other group) have a monopoly on equality. [NEWLINE] [NEWLINE] The label "feminism" comes with a huge load of historic and linguistic baggage ( as OP pointed out) which I don't have the desire *or* energy to combat. It comes with fringe groups which make me roll my eyes every time they speak - and that's the *best*-case scenario. More to the point than that, however, it comes with a number of ideological elements with which I either emphatically disagree, or which I find to be too lacking in nuance to be useful. [NEWLINE] [NEWLINE] The term is now too vague to be in any way descriptive whatsoever, and the movement itself is too diffuse to even *be* a single movement. To say that you're a feminist is *in no way indicative* of your position on issues like parental leave, equality of opportunity vs. equality of outcomes, misandry, transgender issues, etc. People who self-identify as "feminists" run the gamut on these issues - as they do on just about *every* issue I've ever heard discussed in a feminist context. [NEWLINE] [NEWLINE] I personally see it as a useless label - worse than useless, because it's an *inflammatory* label which draws attention away from core issues and toward the "Who Is A Real Feminist" side-track. [NEWLINE] [NEWLINE] And I really, *really* resent people who tell me that *they* get to label me or my ideology. It's as though it's impossible for people with somewhat different labels to work together for a common goal.</s>
Loss: tensor(0.0279, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0356, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0023, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0230, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0222, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0283, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0005, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I'm not the biggest expert on this,<mask><mask><mask> you're confusing the difference between sex and gender. Sex refers to the biological and physiological characteristics that define men and women (ie. genitalia). Gender refers to the socially-constructed roles, behaviors and attributes associated with men and women (ie. women wear dresses, speak in a higher voice, have longer hair, men have shorter hair, participate in "masculine" activities etc). One way to think of it is a difference in sex is "male and female" and a difference in gender is "masculine and feminine." There are intersex people who are born in between the two poles of sex,<mask> in general this is the difference. Basically the idea is that it's a spectrum. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] <mask> I don't understand is those who decide they want to be considered a man after being born a woman.<mask> would someone decide they want to be "treated like a man"<mask> all my life I've been told women should be treated the same? Does that mean we're losing the battle for gender equality, and trans folks are just quitters? Please get this negative thought out of my mind and CMV. [ENDQ] [NEWLINE] Folks say men and women *should be* treated equally, not that they *are*. For example, men and women have very different expectations placed upon them, and have different socially-defined roles. Someone may feel that they are "born the wrong sex" and<mask> redefine their gender role to fit that identity. They may undergo reassignment surgery and take hormones and thereby actually alter their sex. [NEWLINE] [NEWLINE] Hopefully this makes sense. [NEWLINE] </s>
Label encoding: <s>I'm not the biggest expert on this, but I think you're confusing the difference between sex and gender. Sex refers to the biological and physiological characteristics that define men and women (ie. genitalia). Gender refers to the socially-constructed roles, behaviors and attributes associated with men and women (ie. women wear dresses, speak in a higher voice, have longer hair, men have shorter hair, participate in "masculine" activities etc). One way to think of it is a difference in sex is "male and female" and a difference in gender is "masculine and feminine." There are intersex people who are born in between the two poles of sex, but in general this is the difference. Basically the idea is that it's a spectrum. [NEWLINE] [NEWLINE] [NEWLINE] [STARTQ] What I don't understand is those who decide they want to be considered a man after being born a woman. Why would someone decide they want to be "treated like a man" when all my life I've been told women should be treated the same? Does that mean we're losing the battle for gender equality, and trans folks are just quitters? Please get this negative thought out of my mind and CMV. [ENDQ] [NEWLINE] Folks say men and women *should be* treated equally, not that they *are*. For example, men and women have very different expectations placed upon them, and have different socially-defined roles. Someone may feel that they are "born the wrong sex" and therefore redefine their gender role to fit that identity. They may undergo reassignment surgery and take hormones and thereby actually alter their sex. [NEWLINE] [NEWLINE] Hopefully this makes sense. [NEWLINE] </s>
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I can confirm the following lowdown of Sweden: [NEWLINE] [NEWLINE] [STARTQ] * Darkness, we get barely any sunlight during winter. And the winter is loooooong. It gets quite depressing even thought I'm born here. [ENDQ] [NEWLINE] [STARTQ] * High taxes can be seen<mask> a downside to some,<mask> I prefer paying more tax than having less educated voters. [ENDQ] [NEWLINE] [STARTQ] * People are a bit more shy than say the US,<mask> you can talk to anyone and they will tell you their life story etc. On the flipside, I speculate that there is less gossip and trash talk behind your back and it's easier to know your real friends (<mask> that is just me speculating).<mask> I know a few south-european-foreign-exchange students that got too depressed due to the culture shock and never wished to come back. [ENDQ] [NEWLINE] [STARTQ] * Similarly I've find that the few NON-scandinavians friends I have are extremely hospitable. I mean they offer me to crash at their place<mask> visiting without me even asking.<mask> a swede, I unfortunately would feel uncomfortable doing the same and with a few exceptions most swedes I know would feel similarly. [ENDQ] [NEWLINE] [STARTQ] * Beer is expensive<mask> I honestly don't care. Knowing some alcoholics,<mask><mask> it's actually a good thing. [ENDQ] [NEWLINE] [[Source]]( [URL] /) [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] I would like to emphasis my two biggest issues: [NEWLINE] [NEWLINE] 1. The lack of sunlight, which has caused me depression on a number of occasions. [NEWLINE] [NEWLINE] 1. The disinterest and inhospitableness we show towards each other.</s>
Label encoding: <s>I can confirm the following lowdown of Sweden: [NEWLINE] [NEWLINE] [STARTQ] * Darkness, we get barely any sunlight during winter. And the winter is loooooong. It gets quite depressing even thought I'm born here. [ENDQ] [NEWLINE] [STARTQ] * High taxes can be seen as a downside to some, but I prefer paying more tax than having less educated voters. [ENDQ] [NEWLINE] [STARTQ] * People are a bit more shy than say the US, where you can talk to anyone and they will tell you their life story etc. On the flipside, I speculate that there is less gossip and trash talk behind your back and it's easier to know your real friends ( but that is just me speculating). But I know a few south-european-foreign-exchange students that got too depressed due to the culture shock and never wished to come back. [ENDQ] [NEWLINE] [STARTQ] * Similarly I've find that the few NON-scandinavians friends I have are extremely hospitable. I mean they offer me to crash at their place when visiting without me even asking. As a swede, I unfortunately would feel uncomfortable doing the same and with a few exceptions most swedes I know would feel similarly. [ENDQ] [NEWLINE] [STARTQ] * Beer is expensive but I honestly don't care. Knowing some alcoholics, I think it's actually a good thing. [ENDQ] [NEWLINE] [[Source]]( [URL] /) [NEWLINE] [NEWLINE] [NEWLINE] &amp;nbsp; [NEWLINE] [NEWLINE] I would like to emphasis my two biggest issues: [NEWLINE] [NEWLINE] 1. The lack of sunlight, which has caused me depression on a number of occasions. [NEWLINE] [NEWLINE] 1. The disinterest and inhospitableness we show towards each other.</s>
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0121, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0291, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0191, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0037, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0484, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0190, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0235, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0036, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0040, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0091, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0273, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0429, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] You sound like someone who rarely<mask> ever has to actually defend themselves. [ENDQ] [NEWLINE] <mask><mask><mask>, my work places me in a position<mask> I am required to defend myself quite regularly from violent individuals. That being said, I<mask> have the experience to do<mask> with a minimum of violence in return<mask>, there have been times<mask> it has been necessary for me to escalate beyond that. [NEWLINE] [NEWLINE] [STARTQ] <mask> an obviously weaker person is attacking you and you retaliate in'self defense'<mask> it wasnt necessary for your self defense then a court will not count it<mask> self-defense<mask> it's not. [ENDQ] [NEWLINE] You obviously do not understand<mask> constitutes self defence. [NEWLINE] [NEWLINE] [STARTQ] <mask> you're small and weak enough that the average woman can actually do serious damage to you then by all means defend yourself its just that most men arent that small and weak. [ENDQ] [NEWLINE] Without protecting oneself, anybody,<mask><mask> size, is capable of causing significant damage to your person. I was recently attacked by a girl, who was about 5 foot 3 who felt absolutely entitled to punch me in the face (I'm 6 foot 2)<mask> I refused to let her into a secured area. Had I not stopped her after the first swing, she would have continued to throw punches at me.<mask> it was, she was shocked (SHOCKED!) that I dared lay a hand on her to keep her from hitting me further. [NEWLINE] [NEWLINE] [STARTQ] the average male can easily restrain the average female without punching her in the face. [ENDQ] [NEWLINE] And sometimes you can't. the option to escalate necessarily must exist<mask><mask><mask> the action taken is a reasonable one under the circumstances.</s>
Label encoding: <s> [STARTQ] You sound like someone who rarely if ever has to actually defend themselves. [ENDQ] [NEWLINE] On the contrary, my work places me in a position where I am required to defend myself quite regularly from violent individuals. That being said, I also have the experience to do so with a minimum of violence in return although, there have been times when it has been necessary for me to escalate beyond that. [NEWLINE] [NEWLINE] [STARTQ] If an obviously weaker person is attacking you and you retaliate in'self defense' when it wasnt necessary for your self defense then a court will not count it as self-defense because it's not. [ENDQ] [NEWLINE] You obviously do not understand what constitutes self defence. [NEWLINE] [NEWLINE] [STARTQ] If you're small and weak enough that the average woman can actually do serious damage to you then by all means defend yourself its just that most men arent that small and weak. [ENDQ] [NEWLINE] Without protecting oneself, anybody, regardless of size, is capable of causing significant damage to your person. I was recently attacked by a girl, who was about 5 foot 3 who felt absolutely entitled to punch me in the face (I'm 6 foot 2) because I refused to let her into a secured area. Had I not stopped her after the first swing, she would have continued to throw punches at me. As it was, she was shocked (SHOCKED!) that I dared lay a hand on her to keep her from hitting me further. [NEWLINE] [NEWLINE] [STARTQ] the average male can easily restrain the average female without punching her in the face. [ENDQ] [NEWLINE] And sometimes you can't. the option to escalate necessarily must exist as long as the action taken is a reasonable one under the circumstances.</s>
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0130, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0253, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Be careful<mask> you take your examples of patterns from. It is obvious to associate the arrangement bottles in a glass recycling bin to human activity<mask> we know that only humans recycle glass bottles, and some humans do it neatly in rows. [NEWLINE] [NEWLINE] Same for any phenomena that we regularly associate with humans: sure, all paintings have a creator, and not even necessarily a [human one]( [URL].jpg),<mask><mask> we can at least provide an explanation for<mask> that phenomena came to be, then we can (at least hypothetically) explain its existence. We know a [watch found on the beach has a creator]( [URL] ),<mask> we know humans make watches. Does life have a creator? Well,<mask> do you make life? A supernatural hypothesis "explains everything"<mask> it posits no testable mechanism<mask> to<mask>,<mask> it is scientifically useless: we can put no confidence in one explanation that couldn't be<mask> easily and trivially explained by another slightly (or wholly) different supernatural ad hoc hypothesis. [NEWLINE] [NEWLINE] A natural explanation of [abiogenesis]( [URL] ) is still in development<mask> there are many hypotheses and models undergoing experimentation. You are hesitant to believe in the haphazard and chaotic,<mask> remember that naturalism has one big rule: the universe follows rules. For example, atoms don't combine randomly, they actually have a tremendous list of rules and exceptions that dictate their behavior (chemistry). Rocks will order themselves by size on a stream bed. Planets and solar systems coalesce by gravity.<mask> you seem to be truly impressed with is [emergence]( [URL] ): the appearance of complicated things in a system of simple rules.</s>
Label encoding: <s>Be careful where you take your examples of patterns from. It is obvious to associate the arrangement bottles in a glass recycling bin to human activity because we know that only humans recycle glass bottles, and some humans do it neatly in rows. [NEWLINE] [NEWLINE] Same for any phenomena that we regularly associate with humans: sure, all paintings have a creator, and not even necessarily a [human one]( [URL].jpg), but if we can at least provide an explanation for HOW that phenomena came to be, then we can (at least hypothetically) explain its existence. We know a [watch found on the beach has a creator]( [URL] ), because we know humans make watches. Does life have a creator? Well, how do you make life? A supernatural hypothesis "explains everything" but it posits no testable mechanism as to how, thus it is scientifically useless: we can put no confidence in one explanation that couldn't be as easily and trivially explained by another slightly (or wholly) different supernatural ad hoc hypothesis. [NEWLINE] [NEWLINE] A natural explanation of [abiogenesis]( [URL] ) is still in development but there are many hypotheses and models undergoing experimentation. You are hesitant to believe in the haphazard and chaotic, but remember that naturalism has one big rule: the universe follows rules. For example, atoms don't combine randomly, they actually have a tremendous list of rules and exceptions that dictate their behavior (chemistry). Rocks will order themselves by size on a stream bed. Planets and solar systems coalesce by gravity. What you seem to be truly impressed with is [emergence]( [URL] ): the appearance of complicated things in a system of simple rules.</s>
Loss: tensor(0.0238, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0028, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0287, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0236, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0315, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0032, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0178, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0111, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0321, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0399, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0065, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>well i mean, that's the thing with religion isn't it? It doesn't seem justified or fair unless you believe in that religion in that way. [NEWLINE] [NEWLINE] For people (predominantly christians in the US) who are against gay marriage it's perfectly logical<mask> they're believing and interpreting the Bible in certain ways. God creates marriage between a man and woman, there are a few mentions of<mask> bad it is to be gay (<mask><mask> even Jesus says it). Part of the Christian faith is spreading the word, and trying to keep the whole world moral -<mask> it's perfectly logical for them to be against gay marriage. [NEWLINE] [NEWLINE] Right or acceptable in this day and age - No. Logical from an outside perspective - No.<mask> perfectly logical within their interpretation of that faith. In the end<mask> you have that faith and that interpretation, doing anything against it is very illogical<mask> you'll go to hell. [NEWLINE] [NEWLINE] Now<mask> you are looking for arguments that do follow more general logic, I've heard taxes mentioned multiple times. Essentially there's a theory that gay couples are less likely to have children,<mask> will be better off,<mask> will be getting tax breaks for nothing. I don't know the finer points or<mask> accurate any of it is,<mask><mask> any of it's true then that's a fairly logical point to make -<mask> gay couples would be benefiting from marriage in a way that the tax breaks are not designed for, that's a reason to at least delay it until the kinks and wordings of the tax breaks can be more universal.<mask> I don't actually know<mask> any of that is true. </s>
Label encoding: <s>well i mean, that's the thing with religion isn't it? It doesn't seem justified or fair unless you believe in that religion in that way. [NEWLINE] [NEWLINE] For people (predominantly christians in the US) who are against gay marriage it's perfectly logical if they're believing and interpreting the Bible in certain ways. God creates marriage between a man and woman, there are a few mentions of how bad it is to be gay ( I think even Jesus says it). Part of the Christian faith is spreading the word, and trying to keep the whole world moral - so it's perfectly logical for them to be against gay marriage. [NEWLINE] [NEWLINE] Right or acceptable in this day and age - No. Logical from an outside perspective - No. But perfectly logical within their interpretation of that faith. In the end if you have that faith and that interpretation, doing anything against it is very illogical because you'll go to hell. [NEWLINE] [NEWLINE] Now if you are looking for arguments that do follow more general logic, I've heard taxes mentioned multiple times. Essentially there's a theory that gay couples are less likely to have children, so will be better off, so will be getting tax breaks for nothing. I don't know the finer points or how accurate any of it is, but if any of it's true then that's a fairly logical point to make - if gay couples would be benefiting from marriage in a way that the tax breaks are not designed for, that's a reason to at least delay it until the kinks and wordings of the tax breaks can be more universal. However I don't actually know if any of that is true. </s>
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0133, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0177, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0214, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0143, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0328, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0317, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0161, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0216, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0248, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0081, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0198, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0046, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0083, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0138, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0080, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0208, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0186, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0049, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] invasion, nukes, traditional combat.. [ENDQ] [NEWLINE] Or bribery and gerrymandering, extortion, assassination,<mask> is the current order of the day. You've got a lot to learn about war, and even about violence<mask> you really believe that military action cannot commence against any party just<mask> they wield an impressive arsenal. Did you see the United States nuking Al-Qaeda after 9/11? [NEWLINE] [NEWLINE] Violence is state. Organized crime, drug cartels, the mafia, these are nothing more nor less than interstitial feudal states. The Mafia will extort taxes from you the same<mask> the US government will, and find themselves in an equal position to protect you from the domestic violence that local confederated governments obviously can't (themselves yes, and other mobs competing for their turf. They may even act to shield you from local laws or place pressure on your competitors that are not<mask> on their payroll). They will exploit black (underserved and/or illicit) markets to connect consumers to contraband goods and services. It is apt to say they are a cancerous parasite; an entirely separate and independent organism feeding of of the environment inside of the host. [NEWLINE] [NEWLINE] I always find it strange<mask> libertarians expect nature to provide a fair playing field. You say "<mask> everything was reset to dirt,<mask> productive advantage could a business gain from raising an army"<mask><mask> you've never watched Mad Max before.<mask> you reset everything to dirt then you'd damned well better be protected by an army or any wealth you accrue will be wrested from you by whomever is more apt at violence than you are.</s>
Label encoding: <s> [STARTQ] invasion, nukes, traditional combat.. [ENDQ] [NEWLINE] Or bribery and gerrymandering, extortion, assassination, as is the current order of the day. You've got a lot to learn about war, and even about violence if you really believe that military action cannot commence against any party just because they wield an impressive arsenal. Did you see the United States nuking Al-Qaeda after 9/11? [NEWLINE] [NEWLINE] Violence is state. Organized crime, drug cartels, the mafia, these are nothing more nor less than interstitial feudal states. The Mafia will extort taxes from you the same as the US government will, and find themselves in an equal position to protect you from the domestic violence that local confederated governments obviously can't (themselves yes, and other mobs competing for their turf. They may even act to shield you from local laws or place pressure on your competitors that are not also on their payroll). They will exploit black (underserved and/or illicit) markets to connect consumers to contraband goods and services. It is apt to say they are a cancerous parasite; an entirely separate and independent organism feeding of of the environment inside of the host. [NEWLINE] [NEWLINE] I always find it strange how libertarians expect nature to provide a fair playing field. You say " if everything was reset to dirt, what productive advantage could a business gain from raising an army" as though you've never watched Mad Max before. If you reset everything to dirt then you'd damned well better be protected by an army or any wealth you accrue will be wrested from you by whomever is more apt at violence than you are.</s>
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0368, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0159, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0079, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0033, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0378, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0132, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0025, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0197, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0135, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0243, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0241, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0267, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>Building a house is a great example.<mask> you have seen that a contractor builds houses on shaky foundations, it calls into question the quality and stability of any house they build. Likewise,<mask> a researcher bases conclusions on questionable findings that are unsubstantiated, I tend to doubt there entire concept<mask> the researched was probably biased. The Denver Airport theories are good examples. It's weird. Yep. There's some strange stuff that tells me something is up with that place.<mask> many of the points they make (the backwards barbed wire... that's standard for all airports, the murals... the artist was not influenced by anything other than his own political leanings, etc...) can be easily dismissed. Now there are points they make that are solid. Like billionaires and world leaders such<mask> the royal family of Britain buying property adjacent to the airport. That's weird and definitely suggests something is up.<mask><mask> I'm trying to figure out the truth, I'm never going to trust someone who laces real accusations with things that are obviously wrong. The problem is that research should always be unbiased. You want to learn<mask> the truth is. Conspiracy theorists tend to go about it backwards. They decided that the airport was built and funded by the New World Order<mask> a bunker for themselves and their rich friends. Then their "research" is only to back up their pre-decided opinion. You never see any of the facts that would weaken their opinion or problems with their theory. Ask a republican<mask> everyone should be a republican and<mask> you take only their word for it and don't ask anyone else, I guarantee, you'll agree. </s>
Label encoding: <s>Building a house is a great example. If you have seen that a contractor builds houses on shaky foundations, it calls into question the quality and stability of any house they build. Likewise, if a researcher bases conclusions on questionable findings that are unsubstantiated, I tend to doubt there entire concept as the researched was probably biased. The Denver Airport theories are good examples. It's weird. Yep. There's some strange stuff that tells me something is up with that place. But many of the points they make (the backwards barbed wire... that's standard for all airports, the murals... the artist was not influenced by anything other than his own political leanings, etc...) can be easily dismissed. Now there are points they make that are solid. Like billionaires and world leaders such as the royal family of Britain buying property adjacent to the airport. That's weird and definitely suggests something is up. But if I'm trying to figure out the truth, I'm never going to trust someone who laces real accusations with things that are obviously wrong. The problem is that research should always be unbiased. You want to learn what the truth is. Conspiracy theorists tend to go about it backwards. They decided that the airport was built and funded by the New World Order as a bunker for themselves and their rich friends. Then their "research" is only to back up their pre-decided opinion. You never see any of the facts that would weaken their opinion or problems with their theory. Ask a republican why everyone should be a republican and if you take only their word for it and don't ask anyone else, I guarantee, you'll agree. </s>
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>I have specialized surgical training.  For example I perform surgeries that no one else in a several hundred mile radius does.  Sometimes I get a call in the middle of the night asking<mask> I'm willing to do it for some patient (often uninsured, which is to say I'd be doing it for free). [NEWLINE] [NEWLINE] Sometimes I say yes, sometimes I say no. <mask> that patient has a "right to healthcare"<mask> does that work in your scenario, exactly?  Am I obligated to say yes every time? <mask><mask> I have other 8 other cases the next morning that would benefit from me being well rested?  Am I ever allowed to retire? [NEWLINE] [NEWLINE] <mask> the government were to raise taxes and offer more money that might be an incentive for a<mask>,<mask> frankly, I make plenty of money, and sometimes I just don't feel like doing emergency surgery in the middle of the night, for any price.  Is the government obliged to increase rates until the service I provide becomes available?  Is there any limit to this? [NEWLINE] [NEWLINE] <mask> this hypothetical patient doesn't end up getting surgery, is that an infrigement of his rights? <mask> would you fix that? [NEWLINE] [NEWLINE] [STARTQ] He<mask> owes something to... the guy who designed the oven and the factory workers that put it together [ENDQ] [NEWLINE] <mask>? <mask>?  Presumablely these people were paid, no? <mask> does he owe them even more, on top of<mask> they already agreed they would accept for their work?  Can you put a price on<mask> much he actually owes them?  Or is this just feel good nonsense?</s>
Label encoding: <s>I have specialized surgical training.  For example I perform surgeries that no one else in a several hundred mile radius does.  Sometimes I get a call in the middle of the night asking if I'm willing to do it for some patient (often uninsured, which is to say I'd be doing it for free). [NEWLINE] [NEWLINE] Sometimes I say yes, sometimes I say no.  If that patient has a "right to healthcare" how does that work in your scenario, exactly?  Am I obligated to say yes every time?  What if I have other 8 other cases the next morning that would benefit from me being well rested?  Am I ever allowed to retire? [NEWLINE] [NEWLINE] If the government were to raise taxes and offer more money that might be an incentive for a while, but frankly, I make plenty of money, and sometimes I just don't feel like doing emergency surgery in the middle of the night, for any price.  Is the government obliged to increase rates until the service I provide becomes available?  Is there any limit to this? [NEWLINE] [NEWLINE] If this hypothetical patient doesn't end up getting surgery, is that an infrigement of his rights?  How would you fix that? [NEWLINE] [NEWLINE] [STARTQ] He also owes something to... the guy who designed the oven and the factory workers that put it together [ENDQ] [NEWLINE] What?  Why?  Presumablely these people were paid, no?  Why does he owe them even more, on top of what they already agreed they would accept for their work?  Can you put a price on how much he actually owes them?  Or is this just feel good nonsense?</s>
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0201, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0354, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0050, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0305, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0219, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0526, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0030, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0010, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0188, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0127, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0147, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0212, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0187, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0295, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0207, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0223, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0125, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0137, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0160, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0103, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0192, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0215, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0093, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0087, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0254, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0293, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0335, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0225, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0021, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0168, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0128, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0227, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0062, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0261, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0053, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0226, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0281, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0069, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0142, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0351, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0239, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0221, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0107, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>The US Constitution was written at a time<mask> states believed they needed a check on federal power. Requiring a body that represented all states equally to pass any legislation seemed to preserve those rights. [NEWLINE] [NEWLINE] <mask>, this attitude is mostly archaic today. Representing every state equally allows small states disproportionate power in matters pertaining to the whole country.<mask> should they be able to block the rest of the country's people from deciding on federal matters? It's essentially saying that voters in North Dakota, Vermont and Alaska are more important than elsewhere. [NEWLINE] [NEWLINE] <mask> you can find me a purpose for the US Senate<mask> an actual decision-making body in today's day and age, and not just<mask> a vastly unrepresentative body that has the power to completely kill legislation that most Americans support, please CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things.<mask>, please remember to* ***[read through our rules]( [URL] )***. *<mask> you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****!<mask> you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Label encoding: <s>The US Constitution was written at a time when states believed they needed a check on federal power. Requiring a body that represented all states equally to pass any legislation seemed to preserve those rights. [NEWLINE] [NEWLINE] However, this attitude is mostly archaic today. Representing every state equally allows small states disproportionate power in matters pertaining to the whole country. Why should they be able to block the rest of the country's people from deciding on federal matters? It's essentially saying that voters in North Dakota, Vermont and Alaska are more important than elsewhere. [NEWLINE] [NEWLINE] If you can find me a purpose for the US Senate as an actual decision-making body in today's day and age, and not just as a vastly unrepresentative body that has the power to completely kill legislation that most Americans support, please CMV. [NEWLINE] _____ [NEWLINE] [NEWLINE] &gt; *Hello, users of CMV! This is a footnote from your moderators. We'd just like to remind you of a couple of things. Firstly, please remember to* ***[read through our rules]( [URL] )***. * If you see a comment that has broken one, it is more effective to report it than downvote it. Speaking of which,* ***[downvotes don't change views]( [URL] #wiki_upvoting.2Fdownvoting)****! If you are thinking about submitting a CMV yourself, please have a look through our* ***[popular topics wiki]( [URL] )*** *first. Any questions or concerns? Feel free to* ***[message us]( [URL] /r/changemyview)***. *Happy CMVing!*</s>
Loss: tensor(0.0034, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0184, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0350, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0246, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s> [STARTQ] I personally do not know enough about boxing or MMA to say for sure<mask><mask><mask> is regarding them,<mask> isn't the loser determined by who gets knocked out or cannot fight any longer. [ENDQ] [NEWLINE] Often no one is knocked out, and at the end of the match the judges determine the winner. [NEWLINE] [NEWLINE] [STARTQ] For your second point, no, that does not mean it isn't a sport,<mask> there is a rule that states that<mask> no team can score more points than the other, then there is a tie. [ENDQ] [NEWLINE] <mask> a written rule is all that it takes to make it a sport? [NEWLINE] [NEWLINE] [STARTQ] For your last point, look at my comment above. It should address more reasons on<mask> I don't consider cheerleading and dance a sport. [ENDQ] [NEWLINE] I did.  you clearly stated: *Even<mask> you don't accept this definition, there is a clear difference between artistic activities and athletic activities. Athletic events require mostly physical fitness and coordination, and have a clear winner.* [NEWLINE] [NEWLINE] I gave you examples of both cases<mask> physical activity was required<mask> there wasn't a clear winner.  I<mask> gave an example of people who are very physically fit,<mask> much<mask> that they wear practically nothing (and it would be easy to see who is not fit), and are more fit than many of the people on the football field in which they cheer and dance for.  You clearly stated that football is a sport<mask> of a rule, and the cheerleaders and dance team members are not part of a sport......just<mask>. [NEWLINE] [NEWLINE] <mask> the double standard? <mask><mask> the lack of explanation?</s>
Label encoding: <s> [STARTQ] I personally do not know enough about boxing or MMA to say for sure what my opinion is regarding them, but isn't the loser determined by who gets knocked out or cannot fight any longer. [ENDQ] [NEWLINE] Often no one is knocked out, and at the end of the match the judges determine the winner. [NEWLINE] [NEWLINE] [STARTQ] For your second point, no, that does not mean it isn't a sport, since there is a rule that states that if no team can score more points than the other, then there is a tie. [ENDQ] [NEWLINE] So a written rule is all that it takes to make it a sport? [NEWLINE] [NEWLINE] [STARTQ] For your last point, look at my comment above. It should address more reasons on why I don't consider cheerleading and dance a sport. [ENDQ] [NEWLINE] I did.  you clearly stated: *Even if you don't accept this definition, there is a clear difference between artistic activities and athletic activities. Athletic events require mostly physical fitness and coordination, and have a clear winner.* [NEWLINE] [NEWLINE] I gave you examples of both cases where physical activity was required but there wasn't a clear winner.  I also gave an example of people who are very physically fit, so much so that they wear practically nothing (and it would be easy to see who is not fit), and are more fit than many of the people on the football field in which they cheer and dance for.  You clearly stated that football is a sport because of a rule, and the cheerleaders and dance team members are not part of a sport......just because. [NEWLINE] [NEWLINE] Why the double standard?  Also why the lack of explanation?</s>
Loss: tensor(0.0146, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>For me, the most compelling thing that illustrates that "rape culture" is a thing are the statistics. [NEWLINE] [NEWLINE] [STARTQ] Is 1 in 5 American women surviving rape or attempted rape considered a cultural norm? Is 1 in 6 men being abused before the age of 18 a cultural norm? ([TIME]( [URL] /)) [ENDQ] [NEWLINE] [Here is a much longer list of statistics<mask> you are interested.]( [URL] /) [NEWLINE] [NEWLINE] <mask><mask> you want some compelling evidence that there is a "rape culture"--that is to say that it is a subliminal cultural norm that rape is okay in certain circumstances, then you just have too look at<mask> the children say: [NEWLINE] [NEWLINE] [STARTQ] In a survey of high school students, 56% of the girls and 76% of the boys believed forced sex was acceptable under some circumstances. (ref 5) [ENDQ] [NEWLINE] [STARTQ] A survey of 11-to-14 year-olds found: [ENDQ] [NEWLINE] [STARTQ] 51% of the boys and 41% of the girls said forced sex was acceptable<mask> the boy, "spent a lot of money" on the girl; [ENDQ] [NEWLINE] [STARTQ] 31% of the boys and 32% of the girls said it was acceptable for a man to rape a woman with past sexual experience; [ENDQ] [NEWLINE] [STARTQ] 87% of boys and 79% of girls said sexual assault was acceptable<mask> the man and the woman were married; [ENDQ] [NEWLINE] [STARTQ] 65% of the boys and 47% of the girls said it was acceptable for a boy to rape a girl<mask> they had been dating for more than six months. ( [URL] ) [ENDQ] [NEWLINE] <mask> yeah, it seems to exist in very significant ways.</s>
Label encoding: <s>For me, the most compelling thing that illustrates that "rape culture" is a thing are the statistics. [NEWLINE] [NEWLINE] [STARTQ] Is 1 in 5 American women surviving rape or attempted rape considered a cultural norm? Is 1 in 6 men being abused before the age of 18 a cultural norm? ([TIME]( [URL] /)) [ENDQ] [NEWLINE] [Here is a much longer list of statistics if you are interested.]( [URL] /) [NEWLINE] [NEWLINE] But if you want some compelling evidence that there is a "rape culture"--that is to say that it is a subliminal cultural norm that rape is okay in certain circumstances, then you just have too look at what the children say: [NEWLINE] [NEWLINE] [STARTQ] In a survey of high school students, 56% of the girls and 76% of the boys believed forced sex was acceptable under some circumstances. (ref 5) [ENDQ] [NEWLINE] [STARTQ] A survey of 11-to-14 year-olds found: [ENDQ] [NEWLINE] [STARTQ] 51% of the boys and 41% of the girls said forced sex was acceptable if the boy, "spent a lot of money" on the girl; [ENDQ] [NEWLINE] [STARTQ] 31% of the boys and 32% of the girls said it was acceptable for a man to rape a woman with past sexual experience; [ENDQ] [NEWLINE] [STARTQ] 87% of boys and 79% of girls said sexual assault was acceptable if the man and the woman were married; [ENDQ] [NEWLINE] [STARTQ] 65% of the boys and 47% of the girls said it was acceptable for a boy to rape a girl if they had been dating for more than six months. ( [URL] ) [ENDQ] [NEWLINE] So yeah, it seems to exist in very significant ways.</s>
Loss: tensor(0.0123, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>When discussing people in altered states, including those brought about through organic pathology or "mental" illnesses like schizophrenia and psychotic depression, it is inappropriate and false to state such individuals are less in contact with reality. [NEWLINE] [NEWLINE] <mask><mask><mask> that nobody experiences reality directly.  By this, I mean that you see, hear, taste, smell and feel nothing that is not filtered, processed or otherwise distorted by your brain. [NEWLINE] [NEWLINE] <mask><mask> does your brain filter, process and distort sensory information? <mask>  your perceptual faculties are shaped by evolutionary pressures.  There is a selection advantage to accentuate survival maximising data that your senses pick up, and<mask> to minimise sense-data that does not confer some sort of advantage, be it survival and/or reproductive.  That is to say that nobody's brain is selected for directly perceiving reality.  Only selected for perceiving elements of reality that relate to survival and promotion of genes. [NEWLINE] [NEWLINE] <mask>, I don't understand<mask> it is appropriate to say that<mask> someone who is psychotic, or under the effect of hallucinogens, has less legitimate contact with reality. <mask><mask>, with psilocybin, brain regions that limit one's perception are temporarily inhibited and<mask> anything, a more direct way of perceiving reality may be achieved.  (I am not a neuroscientist) [NEWLINE] [NEWLINE] The only sense in which it is appropriate to say that someone that deviates from normal mental functioning is less in contact with reality is to say that they are less in contact with '*consensual reality*', which is by no means identical to actual reality. </s>
Label encoding: <s>When discussing people in altered states, including those brought about through organic pathology or "mental" illnesses like schizophrenia and psychotic depression, it is inappropriate and false to state such individuals are less in contact with reality. [NEWLINE] [NEWLINE] The reason is that nobody experiences reality directly.  By this, I mean that you see, hear, taste, smell and feel nothing that is not filtered, processed or otherwise distorted by your brain. [NEWLINE] [NEWLINE] So why does your brain filter, process and distort sensory information?  Because  your perceptual faculties are shaped by evolutionary pressures.  There is a selection advantage to accentuate survival maximising data that your senses pick up, and also to minimise sense-data that does not confer some sort of advantage, be it survival and/or reproductive.  That is to say that nobody's brain is selected for directly perceiving reality.  Only selected for perceiving elements of reality that relate to survival and promotion of genes. [NEWLINE] [NEWLINE] So, I don't understand why it is appropriate to say that because someone who is psychotic, or under the effect of hallucinogens, has less legitimate contact with reality.  In fact, with psilocybin, brain regions that limit one's perception are temporarily inhibited and if anything, a more direct way of perceiving reality may be achieved.  (I am not a neuroscientist) [NEWLINE] [NEWLINE] The only sense in which it is appropriate to say that someone that deviates from normal mental functioning is less in contact with reality is to say that they are less in contact with '*consensual reality*', which is by no means identical to actual reality. </s>
Loss: tensor(0.0113, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0169, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0318, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0244, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0026, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0100, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0122, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0312, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0259, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0355, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0375, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0268, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0078, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0042, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0112, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0158, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0211, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0120, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0116, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0256, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0329, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0024, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0296, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0071, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0085, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0252, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0144, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0164, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0163, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0218, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0109, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0229, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0074, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0090, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0061, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0151, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0180, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0051, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0242, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0250, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0152, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0088, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0258, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0297, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0096, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0185, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>While the "pure" form may be alike, the thing that makes this scenario interesting is the deontological vs teleological ethical reasoning of the problem. [NEWLINE] [NEWLINE] Before we go on, 2 quick (and off the cuff, you can find better) definitions: [NEWLINE] [NEWLINE] deontological ethics- an action is either right or wrong in and of itself,<mask><mask> the consequences.  It is wrong to steal, and<mask> one should never steal bread to feed a starving person. [NEWLINE] [NEWLINE] teleological ethics- an action is right or wrong based on its consequences.  The ends justify the means.  It is ok to lie to save a life. [NEWLINE] [NEWLINE] The trolley problem, in all its forms, is interesting<mask> it places those ethical systems at odds. [NEWLINE] [NEWLINE] You can see the deontological side of things tested in each scenario.  It is fairly light in the original version (pull a lever, kill one person,<mask> save five).  It is tested a little harder in the fat man version,<mask> that is sending a person to a violent death, which is regarded<mask> "more wrong" than the original, more clinical death. [NEWLINE] [NEWLINE] The transplant scenario pushed the deontological barrier even further.  Now a doctor, whom people place their trust in to keep them well, betrays that trust and plans and executes a first degree murder (the "worst kind" of murder) and then covers her tracks, all to save five people. [NEWLINE] [NEWLINE] The transplant scenario involves many more violations of deontological ethics than the original trolley problem, making the ethical calculus much more difficult.</s>
Label encoding: <s>While the "pure" form may be alike, the thing that makes this scenario interesting is the deontological vs teleological ethical reasoning of the problem. [NEWLINE] [NEWLINE] Before we go on, 2 quick (and off the cuff, you can find better) definitions: [NEWLINE] [NEWLINE] deontological ethics- an action is either right or wrong in and of itself, regardless of the consequences.  It is wrong to steal, and therefore one should never steal bread to feed a starving person. [NEWLINE] [NEWLINE] teleological ethics- an action is right or wrong based on its consequences.  The ends justify the means.  It is ok to lie to save a life. [NEWLINE] [NEWLINE] The trolley problem, in all its forms, is interesting because it places those ethical systems at odds. [NEWLINE] [NEWLINE] You can see the deontological side of things tested in each scenario.  It is fairly light in the original version (pull a lever, kill one person, but save five).  It is tested a little harder in the fat man version, because that is sending a person to a violent death, which is regarded as "more wrong" than the original, more clinical death. [NEWLINE] [NEWLINE] The transplant scenario pushed the deontological barrier even further.  Now a doctor, whom people place their trust in to keep them well, betrays that trust and plans and executes a first degree murder (the "worst kind" of murder) and then covers her tracks, all to save five people. [NEWLINE] [NEWLINE] The transplant scenario involves many more violations of deontological ethics than the original trolley problem, making the ethical calculus much more difficult.</s>
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0247, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0449, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0064, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0063, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0013, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0323, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0162, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0359, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0282, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0300, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0139, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0092, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0234, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0045, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0140, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0098, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0157, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0270, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0182, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0115, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0136, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0052, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0393, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0420, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0413, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0007, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0055, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0067, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0347, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0172, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0068, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0209, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0043, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0124, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0210, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0129, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0237, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0114, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0232, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0117, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0206, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0126, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0150, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0106, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0076, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0196, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0073, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0041, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0365, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0183, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0084, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0203, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0145, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0059, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0154, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0149, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0054, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0134, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0101, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0110, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0179, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0251, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0193, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0194, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0264, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0228, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0166, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0245, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0173, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0141, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0148, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0066, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0433, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0240, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0233, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0108, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0303, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0167, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0039, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0202, device='cuda:0', grad_fn=<NllLossBackward>)
Masked encoding: <s>This is a really misunderstood concept in general and not something taught to anyone except law students and police officers. Police officers have more authority,<mask> barely more. In the US,<mask> a person witnesses a felony, that person can effect a citizens arrest of the perpetrator and not face any criminal or civil liability<mask><mask><mask> there is no excessive force. That means everyone.<mask> it requires you to witness it first hand. The police are authorized to make arrests in that same situation or have *probable cause* that the person being arrested committed a serious misdemeanor or felony. Otherwise, the officers would need an arrest warrant issued by a judge or magistrate. Then they're not making the decision, the court is, and the officers are simply enforcing the order. [NEWLINE] [NEWLINE] Similarly, the police need a warrant to enter a place that the target enjoys a reasonable expectation of privacy unless a specific exception to the warrant requirement exists. Again,<mask> there is a warrant, the police are simply executing the court's order.<mask> there is no warrant, the search is only valid<mask> an exception applies. [NEWLINE] [NEWLINE] The police are given more authority regarding the weapons they carry and<mask> those weapons can be carried.<mask> the sidearms you see officers carry are the same models anyone could purchase. Same with the shotguns that are held in the police cars. Even the weapons the SWAT or other specialized units carry are available to the public for the most part. [NEWLINE] [NEWLINE] The cops are not the judge and they are not the jury. They are charged with enforcing court orders and patrolling and observing the streets to watch for and respond to dangerous situations. </s>
Label encoding: <s>This is a really misunderstood concept in general and not something taught to anyone except law students and police officers. Police officers have more authority, but barely more. In the US, if a person witnesses a felony, that person can effect a citizens arrest of the perpetrator and not face any criminal or civil liability as long as there is no excessive force. That means everyone. But it requires you to witness it first hand. The police are authorized to make arrests in that same situation or have *probable cause* that the person being arrested committed a serious misdemeanor or felony. Otherwise, the officers would need an arrest warrant issued by a judge or magistrate. Then they're not making the decision, the court is, and the officers are simply enforcing the order. [NEWLINE] [NEWLINE] Similarly, the police need a warrant to enter a place that the target enjoys a reasonable expectation of privacy unless a specific exception to the warrant requirement exists. Again, if there is a warrant, the police are simply executing the court's order. If there is no warrant, the search is only valid if an exception applies. [NEWLINE] [NEWLINE] The police are given more authority regarding the weapons they carry and where those weapons can be carried. But the sidearms you see officers carry are the same models anyone could purchase. Same with the shotguns that are held in the police cars. Even the weapons the SWAT or other specialized units carry are available to the public for the most part. [NEWLINE] [NEWLINE] The cops are not the judge and they are not the jury. They are charged with enforcing court orders and patrolling and observing the streets to watch for and respond to dangerous situations. </s>
Loss: tensor(0.0181, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0189, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0266, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0224, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0060, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0284, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0153, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0155, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0204, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0170, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0119, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0171, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0097, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0131, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0086, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0156, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0057, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0310, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0205, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0102, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0058, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0105, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0200, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0031, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0199, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0095, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0077, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0104, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0094, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0075, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0044, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0175, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0082, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0072, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0035, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0249, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: tensor(0.0231, device='cuda:0', grad_fn=<NllLossBackward>)
Loss: 